uboot/arch/powerpc/cpu/mpc85xx/start.S
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0+ */
   2/*
   3 * Copyright 2004, 2007-2012 Freescale Semiconductor, Inc.
   4 * Copyright (C) 2003  Motorola,Inc.
   5 */
   6
   7/* U-Boot Startup Code for Motorola 85xx PowerPC based Embedded Boards
   8 *
   9 * The processor starts at 0xfffffffc and the code is first executed in the
  10 * last 4K page(0xfffff000-0xffffffff) in flash/rom.
  11 *
  12 */
  13
  14#include <asm-offsets.h>
  15#include <config.h>
  16#include <mpc85xx.h>
  17#include <version.h>
  18
  19#include <ppc_asm.tmpl>
  20#include <ppc_defs.h>
  21
  22#include <asm/cache.h>
  23#include <asm/mmu.h>
  24
  25#undef  MSR_KERNEL
  26#define MSR_KERNEL ( MSR_ME )   /* Machine Check */
  27
  28#define LAW_EN          0x80000000
  29
  30#if defined(CONFIG_NAND_SPL) || \
  31        (defined(CONFIG_SPL_BUILD) && defined(CONFIG_SPL_INIT_MINIMAL))
  32#define MINIMAL_SPL
  33#endif
  34
  35#if !defined(CONFIG_SPL) && !defined(CONFIG_SYS_RAMBOOT) && \
  36        !defined(CONFIG_SECURE_BOOT) && !defined(CONFIG_SRIO_PCIE_BOOT_SLAVE)
  37#define NOR_BOOT
  38#endif
  39
  40/*
  41 * Set up GOT: Global Offset Table
  42 *
  43 * Use r12 to access the GOT
  44 */
  45        START_GOT
  46        GOT_ENTRY(_GOT2_TABLE_)
  47        GOT_ENTRY(_FIXUP_TABLE_)
  48
  49#ifndef MINIMAL_SPL
  50        GOT_ENTRY(_start)
  51        GOT_ENTRY(_start_of_vectors)
  52        GOT_ENTRY(_end_of_vectors)
  53        GOT_ENTRY(transfer_to_handler)
  54#endif
  55
  56        GOT_ENTRY(__init_end)
  57        GOT_ENTRY(__bss_end)
  58        GOT_ENTRY(__bss_start)
  59        END_GOT
  60
  61/*
  62 * e500 Startup -- after reset only the last 4KB of the effective
  63 * address space is mapped in the MMU L2 TLB1 Entry0. The .bootpg
  64 * section is located at THIS LAST page and basically does three
  65 * things: clear some registers, set up exception tables and
  66 * add more TLB entries for 'larger spaces'(e.g. the boot rom) to
  67 * continue the boot procedure.
  68
  69 * Once the boot rom is mapped by TLB entries we can proceed
  70 * with normal startup.
  71 *
  72 */
  73
  74        .section .bootpg,"ax"
  75        .globl _start_e500
  76
  77_start_e500:
  78/* Enable debug exception */
  79        li      r1,MSR_DE
  80        mtmsr   r1
  81
  82        /*
  83         * If we got an ePAPR device tree pointer passed in as r3, we need that
  84         * later in cpu_init_early_f(). Save it to a safe register before we
  85         * clobber it so that we can fetch it from there later.
  86         */
  87        mr      r24, r3
  88
  89#ifdef CONFIG_SYS_FSL_ERRATUM_A004510
  90        mfspr   r3,SPRN_SVR
  91        rlwinm  r3,r3,0,0xff
  92        li      r4,CONFIG_SYS_FSL_ERRATUM_A004510_SVR_REV
  93        cmpw    r3,r4
  94        beq     1f
  95
  96#ifdef CONFIG_SYS_FSL_ERRATUM_A004510_SVR_REV2
  97        li      r4,CONFIG_SYS_FSL_ERRATUM_A004510_SVR_REV2
  98        cmpw    r3,r4
  99        beq     1f
 100#endif
 101
 102        /* Not a supported revision affected by erratum */
 103        li      r27,0
 104        b       2f
 105
 1061:      li      r27,1   /* Remember for later that we have the erratum */
 107        /* Erratum says set bits 55:60 to 001001 */
 108        msync
 109        isync
 110        mfspr   r3,SPRN_HDBCR0
 111        li      r4,0x48
 112        rlwimi  r3,r4,0,0x1f8
 113        mtspr   SPRN_HDBCR0,r3
 114        isync
 1152:
 116#endif
 117#ifdef CONFIG_SYS_FSL_ERRATUM_A005125
 118        msync
 119        isync
 120        mfspr   r3, SPRN_HDBCR0
 121        oris    r3, r3, 0x0080
 122        mtspr   SPRN_HDBCR0, r3
 123#endif
 124
 125
 126#if defined(CONFIG_SECURE_BOOT) && defined(CONFIG_E500MC) && \
 127        !defined(CONFIG_E6500)
 128        /* ISBC uses L2 as stack.
 129         * Disable L2 cache here so that u-boot can enable it later
 130         * as part of it's normal flow
 131        */
 132
 133        /* Check if L2 is enabled */
 134        mfspr   r3, SPRN_L2CSR0
 135        lis     r2, L2CSR0_L2E@h
 136        ori     r2, r2, L2CSR0_L2E@l
 137        and.    r4, r3, r2
 138        beq     l2_disabled
 139
 140        mfspr r3, SPRN_L2CSR0
 141        /* Flush L2 cache */
 142        lis     r2,(L2CSR0_L2FL)@h
 143        ori     r2, r2, (L2CSR0_L2FL)@l
 144        or      r3, r2, r3
 145        sync
 146        isync
 147        mtspr   SPRN_L2CSR0,r3
 148        isync
 1491:
 150        mfspr r3, SPRN_L2CSR0
 151        and. r1, r3, r2
 152        bne 1b
 153
 154        mfspr r3, SPRN_L2CSR0
 155        lis r2, L2CSR0_L2E@h
 156        ori r2, r2, L2CSR0_L2E@l
 157        andc r4, r3, r2
 158        sync
 159        isync
 160        mtspr SPRN_L2CSR0,r4
 161        isync
 162
 163l2_disabled:
 164#endif
 165
 166/* clear registers/arrays not reset by hardware */
 167
 168        /* L1 */
 169        li      r0,2
 170        mtspr   L1CSR0,r0       /* invalidate d-cache */
 171        mtspr   L1CSR1,r0       /* invalidate i-cache */
 172
 173        mfspr   r1,DBSR
 174        mtspr   DBSR,r1         /* Clear all valid bits */
 175
 176
 177        .macro  create_tlb1_entry esel ts tsize epn wimg rpn perm phy_high scratch
 178        lis     \scratch, FSL_BOOKE_MAS0(1, \esel, 0)@h
 179        ori     \scratch, \scratch, FSL_BOOKE_MAS0(1, \esel, 0)@l
 180        mtspr   MAS0, \scratch
 181        lis     \scratch, FSL_BOOKE_MAS1(1, 1, 0, \ts, \tsize)@h
 182        ori     \scratch, \scratch, FSL_BOOKE_MAS1(1, 1, 0, \ts, \tsize)@l
 183        mtspr   MAS1, \scratch
 184        lis     \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@h
 185        ori     \scratch, \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@l
 186        mtspr   MAS2, \scratch
 187        lis     \scratch, FSL_BOOKE_MAS3(\rpn, 0, \perm)@h
 188        ori     \scratch, \scratch, FSL_BOOKE_MAS3(\rpn, 0, \perm)@l
 189        mtspr   MAS3, \scratch
 190        lis     \scratch, \phy_high@h
 191        ori     \scratch, \scratch, \phy_high@l
 192        mtspr   MAS7, \scratch
 193        isync
 194        msync
 195        tlbwe
 196        isync
 197        .endm
 198
 199        .macro  create_tlb0_entry esel ts tsize epn wimg rpn perm phy_high scratch
 200        lis     \scratch, FSL_BOOKE_MAS0(0, \esel, 0)@h
 201        ori     \scratch, \scratch, FSL_BOOKE_MAS0(0, \esel, 0)@l
 202        mtspr   MAS0, \scratch
 203        lis     \scratch, FSL_BOOKE_MAS1(1, 0, 0, \ts, \tsize)@h
 204        ori     \scratch, \scratch, FSL_BOOKE_MAS1(1, 0, 0, \ts, \tsize)@l
 205        mtspr   MAS1, \scratch
 206        lis     \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@h
 207        ori     \scratch, \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@l
 208        mtspr   MAS2, \scratch
 209        lis     \scratch, FSL_BOOKE_MAS3(\rpn, 0, \perm)@h
 210        ori     \scratch, \scratch, FSL_BOOKE_MAS3(\rpn, 0, \perm)@l
 211        mtspr   MAS3, \scratch
 212        lis     \scratch, \phy_high@h
 213        ori     \scratch, \scratch, \phy_high@l
 214        mtspr   MAS7, \scratch
 215        isync
 216        msync
 217        tlbwe
 218        isync
 219        .endm
 220
 221        .macro  delete_tlb1_entry esel scratch
 222        lis     \scratch, FSL_BOOKE_MAS0(1, \esel, 0)@h
 223        ori     \scratch, \scratch, FSL_BOOKE_MAS0(1, \esel, 0)@l
 224        mtspr   MAS0, \scratch
 225        li      \scratch, 0
 226        mtspr   MAS1, \scratch
 227        isync
 228        msync
 229        tlbwe
 230        isync
 231        .endm
 232
 233        .macro  delete_tlb0_entry esel epn wimg scratch
 234        lis     \scratch, FSL_BOOKE_MAS0(0, \esel, 0)@h
 235        ori     \scratch, \scratch, FSL_BOOKE_MAS0(0, \esel, 0)@l
 236        mtspr   MAS0, \scratch
 237        li      \scratch, 0
 238        mtspr   MAS1, \scratch
 239        lis     \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@h
 240        ori     \scratch, \scratch, FSL_BOOKE_MAS2(\epn, \wimg)@l
 241        mtspr   MAS2, \scratch
 242        isync
 243        msync
 244        tlbwe
 245        isync
 246        .endm
 247
 248/* Interrupt vectors do not fit in minimal SPL. */
 249#if !defined(MINIMAL_SPL)
 250        /* Setup interrupt vectors */
 251        lis     r1,CONFIG_SYS_MONITOR_BASE@h
 252        mtspr   IVPR,r1
 253
 254        li      r4,CriticalInput@l
 255        mtspr   IVOR0,r4        /* 0: Critical input */
 256        li      r4,MachineCheck@l
 257        mtspr   IVOR1,r4        /* 1: Machine check */
 258        li      r4,DataStorage@l
 259        mtspr   IVOR2,r4        /* 2: Data storage */
 260        li      r4,InstStorage@l
 261        mtspr   IVOR3,r4        /* 3: Instruction storage */
 262        li      r4,ExtInterrupt@l
 263        mtspr   IVOR4,r4        /* 4: External interrupt */
 264        li      r4,Alignment@l
 265        mtspr   IVOR5,r4        /* 5: Alignment */
 266        li      r4,ProgramCheck@l
 267        mtspr   IVOR6,r4        /* 6: Program check */
 268        li      r4,FPUnavailable@l
 269        mtspr   IVOR7,r4        /* 7: floating point unavailable */
 270        li      r4,SystemCall@l
 271        mtspr   IVOR8,r4        /* 8: System call */
 272        /* 9: Auxiliary processor unavailable(unsupported) */
 273        li      r4,Decrementer@l
 274        mtspr   IVOR10,r4       /* 10: Decrementer */
 275        li      r4,IntervalTimer@l
 276        mtspr   IVOR11,r4       /* 11: Interval timer */
 277        li      r4,WatchdogTimer@l
 278        mtspr   IVOR12,r4       /* 12: Watchdog timer */
 279        li      r4,DataTLBError@l
 280        mtspr   IVOR13,r4       /* 13: Data TLB error */
 281        li      r4,InstructionTLBError@l
 282        mtspr   IVOR14,r4       /* 14: Instruction TLB error */
 283        li      r4,DebugBreakpoint@l
 284        mtspr   IVOR15,r4       /* 15: Debug */
 285#endif
 286
 287        /* Clear and set up some registers. */
 288        li      r0,0x0000
 289        lis     r1,0xffff
 290        mtspr   DEC,r0                  /* prevent dec exceptions */
 291        mttbl   r0                      /* prevent fit & wdt exceptions */
 292        mttbu   r0
 293        mtspr   TSR,r1                  /* clear all timer exception status */
 294        mtspr   TCR,r0                  /* disable all */
 295        mtspr   ESR,r0                  /* clear exception syndrome register */
 296        mtspr   MCSR,r0                 /* machine check syndrome register */
 297        mtxer   r0                      /* clear integer exception register */
 298
 299#ifdef CONFIG_SYS_BOOK3E_HV
 300        mtspr   MAS8,r0                 /* make sure MAS8 is clear */
 301#endif
 302
 303        /* Enable Time Base and Select Time Base Clock */
 304        lis     r0,HID0_EMCP@h          /* Enable machine check */
 305#if defined(CONFIG_ENABLE_36BIT_PHYS)
 306        ori     r0,r0,HID0_ENMAS7@l     /* Enable MAS7 */
 307#endif
 308#ifndef CONFIG_E500MC
 309        ori     r0,r0,HID0_TBEN@l       /* Enable Timebase */
 310#endif
 311        mtspr   HID0,r0
 312
 313#if !defined(CONFIG_E500MC) && !defined(CONFIG_ARCH_QEMU_E500)
 314        li      r0,(HID1_ASTME|HID1_ABE)@l      /* Addr streaming & broadcast */
 315        mfspr   r3,PVR
 316        andi.   r3,r3, 0xff
 317        cmpwi   r3,0x50@l       /* if we are rev 5.0 or greater set MBDD */
 318        blt 1f
 319        /* Set MBDD bit also */
 320        ori r0, r0, HID1_MBDD@l
 3211:
 322        mtspr   HID1,r0
 323#endif
 324
 325#ifdef CONFIG_SYS_FSL_ERRATUM_CPU_A003999
 326        mfspr   r3,SPRN_HDBCR1
 327        oris    r3,r3,0x0100
 328        mtspr   SPRN_HDBCR1,r3
 329#endif
 330
 331        /* Enable Branch Prediction */
 332#if defined(CONFIG_BTB)
 333        lis     r0,BUCSR_ENABLE@h
 334        ori     r0,r0,BUCSR_ENABLE@l
 335        mtspr   SPRN_BUCSR,r0
 336#endif
 337
 338#if defined(CONFIG_SYS_INIT_DBCR)
 339        lis     r1,0xffff
 340        ori     r1,r1,0xffff
 341        mtspr   DBSR,r1                 /* Clear all status bits */
 342        lis     r0,CONFIG_SYS_INIT_DBCR@h       /* DBCR0[IDM] must be set */
 343        ori     r0,r0,CONFIG_SYS_INIT_DBCR@l
 344        mtspr   DBCR0,r0
 345#endif
 346
 347#ifdef CONFIG_ARCH_MPC8569
 348#define CONFIG_SYS_LBC_ADDR (CONFIG_SYS_CCSRBAR_DEFAULT + 0x5000)
 349#define CONFIG_SYS_LBCR_ADDR (CONFIG_SYS_LBC_ADDR + 0xd0)
 350
 351        /* MPC8569 Rev.0 silcon needs to set bit 13 of LBCR to allow elBC to
 352         * use address space which is more than 12bits, and it must be done in
 353         * the 4K boot page. So we set this bit here.
 354         */
 355
 356        /* create a temp mapping TLB0[0] for LBCR  */
 357        create_tlb0_entry 0, \
 358                0, BOOKE_PAGESZ_4K, \
 359                CONFIG_SYS_LBC_ADDR, MAS2_I|MAS2_G, \
 360                CONFIG_SYS_LBC_ADDR, MAS3_SW|MAS3_SR, \
 361                0, r6
 362
 363        /* Set LBCR register */
 364        lis     r4,CONFIG_SYS_LBCR_ADDR@h
 365        ori     r4,r4,CONFIG_SYS_LBCR_ADDR@l
 366
 367        lis     r5,CONFIG_SYS_LBC_LBCR@h
 368        ori     r5,r5,CONFIG_SYS_LBC_LBCR@l
 369        stw     r5,0(r4)
 370        isync
 371
 372        /* invalidate this temp TLB */
 373        lis     r4,CONFIG_SYS_LBC_ADDR@h
 374        ori     r4,r4,CONFIG_SYS_LBC_ADDR@l
 375        tlbivax 0,r4
 376        isync
 377
 378#endif /* CONFIG_ARCH_MPC8569 */
 379
 380/*
 381 * Search for the TLB that covers the code we're executing, and shrink it
 382 * so that it covers only this 4K page.  That will ensure that any other
 383 * TLB we create won't interfere with it.  We assume that the TLB exists,
 384 * which is why we don't check the Valid bit of MAS1.  We also assume
 385 * it is in TLB1.
 386 *
 387 * This is necessary, for example, when booting from the on-chip ROM,
 388 * which (oddly) creates a single 4GB TLB that covers CCSR and DDR.
 389 */
 390        bl      nexti           /* Find our address */
 391nexti:  mflr    r1              /* R1 = our PC */
 392        li      r2, 0
 393        mtspr   MAS6, r2        /* Assume the current PID and AS are 0 */
 394        isync
 395        msync
 396        tlbsx   0, r1           /* This must succeed */
 397
 398        mfspr   r14, MAS0       /* Save ESEL for later */
 399        rlwinm  r14, r14, 16, 0xfff
 400
 401        /* Set the size of the TLB to 4KB */
 402        mfspr   r3, MAS1
 403        li      r2, 0xF80
 404        andc    r3, r3, r2      /* Clear the TSIZE bits */
 405        ori     r3, r3, MAS1_TSIZE(BOOKE_PAGESZ_4K)@l
 406        oris    r3, r3, MAS1_IPROT@h
 407        mtspr   MAS1, r3
 408
 409        /*
 410         * Set the base address of the TLB to our PC.  We assume that
 411         * virtual == physical.  We also assume that MAS2_EPN == MAS3_RPN.
 412         */
 413        lis     r3, MAS2_EPN@h
 414        ori     r3, r3, MAS2_EPN@l      /* R3 = MAS2_EPN */
 415
 416        and     r1, r1, r3      /* Our PC, rounded down to the nearest page */
 417
 418        mfspr   r2, MAS2
 419        andc    r2, r2, r3
 420        or      r2, r2, r1
 421#ifdef CONFIG_SYS_FSL_ERRATUM_A004510
 422        cmpwi   r27,0
 423        beq     1f
 424        andi.   r15, r2, MAS2_I|MAS2_G /* save the old I/G for later */
 425        rlwinm  r2, r2, 0, ~MAS2_I
 426        ori     r2, r2, MAS2_G
 4271:
 428#endif
 429        mtspr   MAS2, r2        /* Set the EPN to our PC base address */
 430
 431        mfspr   r2, MAS3
 432        andc    r2, r2, r3
 433        or      r2, r2, r1
 434        mtspr   MAS3, r2        /* Set the RPN to our PC base address */
 435
 436        isync
 437        msync
 438        tlbwe
 439
 440/*
 441 * Clear out any other TLB entries that may exist, to avoid conflicts.
 442 * Our TLB entry is in r14.
 443 */
 444        li      r0, TLBIVAX_ALL | TLBIVAX_TLB0
 445        tlbivax 0, r0
 446        tlbsync
 447
 448        mfspr   r4, SPRN_TLB1CFG
 449        rlwinm  r4, r4, 0, TLBnCFG_NENTRY_MASK
 450
 451        li      r3, 0
 452        mtspr   MAS1, r3
 4531:      cmpw    r3, r14
 454        rlwinm  r5, r3, 16, MAS0_ESEL_MSK
 455        addi    r3, r3, 1
 456        beq     2f              /* skip the entry we're executing from */
 457
 458        oris    r5, r5, MAS0_TLBSEL(1)@h
 459        mtspr   MAS0, r5
 460
 461        isync
 462        tlbwe
 463        isync
 464        msync
 465
 4662:      cmpw    r3, r4
 467        blt     1b
 468
 469#if defined(CONFIG_SYS_PPC_E500_DEBUG_TLB) && !defined(MINIMAL_SPL) && \
 470        !defined(CONFIG_SECURE_BOOT)
 471/*
 472 * TLB entry for debuggging in AS1
 473 * Create temporary TLB entry in AS0 to handle debug exception
 474 * As on debug exception MSR is cleared i.e. Address space is changed
 475 * to 0. A TLB entry (in AS0) is required to handle debug exception generated
 476 * in AS1.
 477 */
 478
 479#ifdef NOR_BOOT
 480/*
 481 * TLB entry is created for IVPR + IVOR15 to map on valid OP code address
 482 * bacause flash's virtual address maps to 0xff800000 - 0xffffffff.
 483 * and this window is outside of 4K boot window.
 484 */
 485        create_tlb1_entry CONFIG_SYS_PPC_E500_DEBUG_TLB, \
 486                0, BOOKE_PAGESZ_4M, \
 487                CONFIG_SYS_MONITOR_BASE & 0xffc00000,  MAS2_I|MAS2_G, \
 488                0xffc00000, MAS3_SX|MAS3_SW|MAS3_SR, \
 489                0, r6
 490
 491#else
 492/*
 493 * TLB entry is created for IVPR + IVOR15 to map on valid OP code address
 494 * because "nexti" will resize TLB to 4K
 495 */
 496        create_tlb1_entry CONFIG_SYS_PPC_E500_DEBUG_TLB, \
 497                0, BOOKE_PAGESZ_256K, \
 498                CONFIG_SYS_MONITOR_BASE & 0xfffc0000, MAS2_I, \
 499                CONFIG_SYS_MONITOR_BASE & 0xfffc0000, MAS3_SX|MAS3_SW|MAS3_SR, \
 500                0, r6
 501#endif
 502#endif
 503
 504/*
 505 * Relocate CCSR, if necessary.  We relocate CCSR if (obviously) the default
 506 * location is not where we want it.  This typically happens on a 36-bit
 507 * system, where we want to move CCSR to near the top of 36-bit address space.
 508 *
 509 * To move CCSR, we create two temporary TLBs, one for the old location, and
 510 * another for the new location.  On CoreNet systems, we also need to create
 511 * a special, temporary LAW.
 512 *
 513 * As a general rule, TLB0 is used for short-term TLBs, and TLB1 is used for
 514 * long-term TLBs, so we use TLB0 here.
 515 */
 516#if (CONFIG_SYS_CCSRBAR_DEFAULT != CONFIG_SYS_CCSRBAR_PHYS)
 517
 518#if !defined(CONFIG_SYS_CCSRBAR_PHYS_HIGH) || !defined(CONFIG_SYS_CCSRBAR_PHYS_LOW)
 519#error "CONFIG_SYS_CCSRBAR_PHYS_HIGH and CONFIG_SYS_CCSRBAR_PHYS_LOW) must be defined."
 520#endif
 521
 522create_ccsr_new_tlb:
 523        /*
 524         * Create a TLB for the new location of CCSR.  Register R8 is reserved
 525         * for the virtual address of this TLB (CONFIG_SYS_CCSRBAR).
 526         */
 527        lis     r8, CONFIG_SYS_CCSRBAR@h
 528        ori     r8, r8, CONFIG_SYS_CCSRBAR@l
 529        lis     r9, (CONFIG_SYS_CCSRBAR + 0x1000)@h
 530        ori     r9, r9, (CONFIG_SYS_CCSRBAR + 0x1000)@l
 531        create_tlb0_entry 0, \
 532                0, BOOKE_PAGESZ_4K, \
 533                CONFIG_SYS_CCSRBAR, MAS2_I|MAS2_G, \
 534                CONFIG_SYS_CCSRBAR_PHYS_LOW, MAS3_SW|MAS3_SR, \
 535                CONFIG_SYS_CCSRBAR_PHYS_HIGH, r3
 536        /*
 537         * Create a TLB for the current location of CCSR.  Register R9 is reserved
 538         * for the virtual address of this TLB (CONFIG_SYS_CCSRBAR + 0x1000).
 539         */
 540create_ccsr_old_tlb:
 541        create_tlb0_entry 1, \
 542                0, BOOKE_PAGESZ_4K, \
 543                CONFIG_SYS_CCSRBAR + 0x1000, MAS2_I|MAS2_G, \
 544                CONFIG_SYS_CCSRBAR_DEFAULT, MAS3_SW|MAS3_SR, \
 545                0, r3 /* The default CCSR address is always a 32-bit number */
 546
 547
 548        /*
 549         * We have a TLB for what we think is the current (old) CCSR.  Let's
 550         * verify that, otherwise we won't be able to move it.
 551         * CONFIG_SYS_CCSRBAR_DEFAULT is always a 32-bit number, so we only
 552         * need to compare the lower 32 bits of CCSRBAR on CoreNet systems.
 553         */
 554verify_old_ccsr:
 555        lis     r0, CONFIG_SYS_CCSRBAR_DEFAULT@h
 556        ori     r0, r0, CONFIG_SYS_CCSRBAR_DEFAULT@l
 557#ifdef CONFIG_FSL_CORENET
 558        lwz     r1, 4(r9)               /* CCSRBARL */
 559#else
 560        lwz     r1, 0(r9)               /* CCSRBAR, shifted right by 12 */
 561        slwi    r1, r1, 12
 562#endif
 563
 564        cmpl    0, r0, r1
 565
 566        /*
 567         * If the value we read from CCSRBARL is not what we expect, then
 568         * enter an infinite loop.  This will at least allow a debugger to
 569         * halt execution and examine TLBs, etc.  There's no point in going
 570         * on.
 571         */
 572infinite_debug_loop:
 573        bne     infinite_debug_loop
 574
 575#ifdef CONFIG_FSL_CORENET
 576
 577#define CCSR_LAWBARH0   (CONFIG_SYS_CCSRBAR + 0x1000)
 578#define LAW_SIZE_4K     0xb
 579#define CCSRBAR_LAWAR   (LAW_EN | (0x1e << 20) | LAW_SIZE_4K)
 580#define CCSRAR_C        0x80000000      /* Commit */
 581
 582create_temp_law:
 583        /*
 584         * On CoreNet systems, we create the temporary LAW using a special LAW
 585         * target ID of 0x1e.  LAWBARH is at offset 0xc00 in CCSR.
 586         */
 587        lis     r0, CONFIG_SYS_CCSRBAR_PHYS_HIGH@h
 588        ori     r0, r0, CONFIG_SYS_CCSRBAR_PHYS_HIGH@l
 589        lis     r1, CONFIG_SYS_CCSRBAR_PHYS_LOW@h
 590        ori     r1, r1, CONFIG_SYS_CCSRBAR_PHYS_LOW@l
 591        lis     r2, CCSRBAR_LAWAR@h
 592        ori     r2, r2, CCSRBAR_LAWAR@l
 593
 594        stw     r0, 0xc00(r9)   /* LAWBARH0 */
 595        stw     r1, 0xc04(r9)   /* LAWBARL0 */
 596        sync
 597        stw     r2, 0xc08(r9)   /* LAWAR0 */
 598
 599        /*
 600         * Read back from LAWAR to ensure the update is complete.  e500mc
 601         * cores also require an isync.
 602         */
 603        lwz     r0, 0xc08(r9)   /* LAWAR0 */
 604        isync
 605
 606        /*
 607         * Read the current CCSRBARH and CCSRBARL using load word instructions.
 608         * Follow this with an isync instruction. This forces any outstanding
 609         * accesses to configuration space to completion.
 610         */
 611read_old_ccsrbar:
 612        lwz     r0, 0(r9)       /* CCSRBARH */
 613        lwz     r0, 4(r9)       /* CCSRBARL */
 614        isync
 615
 616        /*
 617         * Write the new values for CCSRBARH and CCSRBARL to their old
 618         * locations.  The CCSRBARH has a shadow register. When the CCSRBARH
 619         * has a new value written it loads a CCSRBARH shadow register. When
 620         * the CCSRBARL is written, the CCSRBARH shadow register contents
 621         * along with the CCSRBARL value are loaded into the CCSRBARH and
 622         * CCSRBARL registers, respectively.  Follow this with a sync
 623         * instruction.
 624         */
 625write_new_ccsrbar:
 626        lis     r0, CONFIG_SYS_CCSRBAR_PHYS_HIGH@h
 627        ori     r0, r0, CONFIG_SYS_CCSRBAR_PHYS_HIGH@l
 628        lis     r1, CONFIG_SYS_CCSRBAR_PHYS_LOW@h
 629        ori     r1, r1, CONFIG_SYS_CCSRBAR_PHYS_LOW@l
 630        lis     r2, CCSRAR_C@h
 631        ori     r2, r2, CCSRAR_C@l
 632
 633        stw     r0, 0(r9)       /* Write to CCSRBARH */
 634        sync                    /* Make sure we write to CCSRBARH first */
 635        stw     r1, 4(r9)       /* Write to CCSRBARL */
 636        sync
 637
 638        /*
 639         * Write a 1 to the commit bit (C) of CCSRAR at the old location.
 640         * Follow this with a sync instruction.
 641         */
 642        stw     r2, 8(r9)
 643        sync
 644
 645        /* Delete the temporary LAW */
 646delete_temp_law:
 647        li      r1, 0
 648        stw     r1, 0xc08(r8)
 649        sync
 650        stw     r1, 0xc00(r8)
 651        stw     r1, 0xc04(r8)
 652        sync
 653
 654#else /* #ifdef CONFIG_FSL_CORENET */
 655
 656write_new_ccsrbar:
 657        /*
 658         * Read the current value of CCSRBAR using a load word instruction
 659         * followed by an isync. This forces all accesses to configuration
 660         * space to complete.
 661         */
 662        sync
 663        lwz     r0, 0(r9)
 664        isync
 665
 666/* CONFIG_SYS_CCSRBAR_PHYS right shifted by 12 */
 667#define CCSRBAR_PHYS_RS12 ((CONFIG_SYS_CCSRBAR_PHYS_HIGH << 20) | \
 668                           (CONFIG_SYS_CCSRBAR_PHYS_LOW >> 12))
 669
 670        /* Write the new value to CCSRBAR. */
 671        lis     r0, CCSRBAR_PHYS_RS12@h
 672        ori     r0, r0, CCSRBAR_PHYS_RS12@l
 673        stw     r0, 0(r9)
 674        sync
 675
 676        /*
 677         * The manual says to perform a load of an address that does not
 678         * access configuration space or the on-chip SRAM using an existing TLB,
 679         * but that doesn't appear to be necessary.  We will do the isync,
 680         * though.
 681         */
 682        isync
 683
 684        /*
 685         * Read the contents of CCSRBAR from its new location, followed by
 686         * another isync.
 687         */
 688        lwz     r0, 0(r8)
 689        isync
 690
 691#endif  /* #ifdef CONFIG_FSL_CORENET */
 692
 693        /* Delete the temporary TLBs */
 694delete_temp_tlbs:
 695        delete_tlb0_entry 0, CONFIG_SYS_CCSRBAR, MAS2_I|MAS2_G, r3
 696        delete_tlb0_entry 1, CONFIG_SYS_CCSRBAR + 0x1000, MAS2_I|MAS2_G, r3
 697
 698#endif /* #if (CONFIG_SYS_CCSRBAR_DEFAULT != CONFIG_SYS_CCSRBAR_PHYS) */
 699
 700#if defined(CONFIG_SYS_FSL_QORIQ_CHASSIS2) && defined(CONFIG_E6500)
 701create_ccsr_l2_tlb:
 702        /*
 703         * Create a TLB for the MMR location of CCSR
 704         * to access L2CSR0 register
 705         */
 706        create_tlb0_entry 0, \
 707                0, BOOKE_PAGESZ_4K, \
 708                CONFIG_SYS_CCSRBAR + 0xC20000, MAS2_I|MAS2_G, \
 709                CONFIG_SYS_CCSRBAR_PHYS_LOW + 0xC20000, MAS3_SW|MAS3_SR, \
 710                CONFIG_SYS_CCSRBAR_PHYS_HIGH, r3
 711
 712enable_l2_cluster_l2:
 713        /* enable L2 cache */
 714        lis     r3, (CONFIG_SYS_CCSRBAR + 0xC20000)@h
 715        ori     r3, r3, (CONFIG_SYS_CCSRBAR + 0xC20000)@l
 716        li      r4, 33  /* stash id */
 717        stw     r4, 4(r3)
 718        lis     r4, (L2CSR0_L2FI|L2CSR0_L2LFC)@h
 719        ori     r4, r4, (L2CSR0_L2FI|L2CSR0_L2LFC)@l
 720        sync
 721        stw     r4, 0(r3)       /* invalidate L2 */
 722        /* Poll till the bits are cleared */
 7231:      sync
 724        lwz     r0, 0(r3)
 725        twi     0, r0, 0
 726        isync
 727        and.    r1, r0, r4
 728        bne     1b
 729
 730        /* L2PE must be set before L2 cache is enabled */
 731        lis     r4, (L2CSR0_L2PE)@h
 732        ori     r4, r4, (L2CSR0_L2PE)@l
 733        sync
 734        stw     r4, 0(r3)       /* enable L2 parity/ECC error checking */
 735        /* Poll till the bit is set */
 7361:      sync
 737        lwz     r0, 0(r3)
 738        twi     0, r0, 0
 739        isync
 740        and.    r1, r0, r4
 741        beq     1b
 742
 743        lis     r4, (L2CSR0_L2E|L2CSR0_L2PE)@h
 744        ori     r4, r4, (L2CSR0_L2REP_MODE)@l
 745        sync
 746        stw     r4, 0(r3)       /* enable L2 */
 747        /* Poll till the bit is set */
 7481:      sync
 749        lwz     r0, 0(r3)
 750        twi     0, r0, 0
 751        isync
 752        and.    r1, r0, r4
 753        beq     1b
 754
 755delete_ccsr_l2_tlb:
 756        delete_tlb0_entry 0, CONFIG_SYS_CCSRBAR + 0xC20000, MAS2_I|MAS2_G, r3
 757#endif
 758
 759        /*
 760         * Enable the L1. On e6500, this has to be done
 761         * after the L2 is up.
 762         */
 763
 764#ifdef CONFIG_SYS_CACHE_STASHING
 765        /* set stash id to (coreID) * 2 + 32 + L1 CT (0) */
 766        li      r2,(32 + 0)
 767        mtspr   L1CSR2,r2
 768#endif
 769
 770        /* Enable/invalidate the I-Cache */
 771        lis     r2,(L1CSR1_ICFI|L1CSR1_ICLFR)@h
 772        ori     r2,r2,(L1CSR1_ICFI|L1CSR1_ICLFR)@l
 773        mtspr   SPRN_L1CSR1,r2
 7741:
 775        mfspr   r3,SPRN_L1CSR1
 776        and.    r1,r3,r2
 777        bne     1b
 778
 779        lis     r3,(L1CSR1_CPE|L1CSR1_ICE)@h
 780        ori     r3,r3,(L1CSR1_CPE|L1CSR1_ICE)@l
 781        mtspr   SPRN_L1CSR1,r3
 782        isync
 7832:
 784        mfspr   r3,SPRN_L1CSR1
 785        andi.   r1,r3,L1CSR1_ICE@l
 786        beq     2b
 787
 788        /* Enable/invalidate the D-Cache */
 789        lis     r2,(L1CSR0_DCFI|L1CSR0_DCLFR)@h
 790        ori     r2,r2,(L1CSR0_DCFI|L1CSR0_DCLFR)@l
 791        mtspr   SPRN_L1CSR0,r2
 7921:
 793        mfspr   r3,SPRN_L1CSR0
 794        and.    r1,r3,r2
 795        bne     1b
 796
 797        lis     r3,(L1CSR0_CPE|L1CSR0_DCE)@h
 798        ori     r3,r3,(L1CSR0_CPE|L1CSR0_DCE)@l
 799        mtspr   SPRN_L1CSR0,r3
 800        isync
 8012:
 802        mfspr   r3,SPRN_L1CSR0
 803        andi.   r1,r3,L1CSR0_DCE@l
 804        beq     2b
 805#ifdef CONFIG_SYS_FSL_ERRATUM_A004510
 806#define DCSR_LAWBARH0   (CONFIG_SYS_CCSRBAR + 0x1000)
 807#define LAW_SIZE_1M     0x13
 808#define DCSRBAR_LAWAR   (LAW_EN | (0x1d << 20) | LAW_SIZE_1M)
 809
 810        cmpwi   r27,0
 811        beq     9f
 812
 813        /*
 814         * Create a TLB entry for CCSR
 815         *
 816         * We're executing out of TLB1 entry in r14, and that's the only
 817         * TLB entry that exists.  To allocate some TLB entries for our
 818         * own use, flip a bit high enough that we won't flip it again
 819         * via incrementing.
 820         */
 821
 822        xori    r8, r14, 32
 823        lis     r0, MAS0_TLBSEL(1)@h
 824        rlwimi  r0, r8, 16, MAS0_ESEL_MSK
 825        lis     r1, FSL_BOOKE_MAS1(1, 1, 0, 0, BOOKE_PAGESZ_16M)@h
 826        ori     r1, r1, FSL_BOOKE_MAS1(1, 1, 0, 0, BOOKE_PAGESZ_16M)@l
 827        lis     r7, CONFIG_SYS_CCSRBAR@h
 828        ori     r7, r7, CONFIG_SYS_CCSRBAR@l
 829        ori     r2, r7, MAS2_I|MAS2_G
 830        lis     r3, FSL_BOOKE_MAS3(CONFIG_SYS_CCSRBAR_PHYS_LOW, 0, (MAS3_SW|MAS3_SR))@h
 831        ori     r3, r3, FSL_BOOKE_MAS3(CONFIG_SYS_CCSRBAR_PHYS_LOW, 0, (MAS3_SW|MAS3_SR))@l
 832        lis     r4, CONFIG_SYS_CCSRBAR_PHYS_HIGH@h
 833        ori     r4, r4, CONFIG_SYS_CCSRBAR_PHYS_HIGH@l
 834        mtspr   MAS0, r0
 835        mtspr   MAS1, r1
 836        mtspr   MAS2, r2
 837        mtspr   MAS3, r3
 838        mtspr   MAS7, r4
 839        isync
 840        tlbwe
 841        isync
 842        msync
 843
 844        /* Map DCSR temporarily to physical address zero */
 845        li      r0, 0
 846        lis     r3, DCSRBAR_LAWAR@h
 847        ori     r3, r3, DCSRBAR_LAWAR@l
 848
 849        stw     r0, 0xc00(r7)   /* LAWBARH0 */
 850        stw     r0, 0xc04(r7)   /* LAWBARL0 */
 851        sync
 852        stw     r3, 0xc08(r7)   /* LAWAR0 */
 853
 854        /* Read back from LAWAR to ensure the update is complete. */
 855        lwz     r3, 0xc08(r7)   /* LAWAR0 */
 856        isync
 857
 858        /* Create a TLB entry for DCSR at zero */
 859
 860        addi    r9, r8, 1
 861        lis     r0, MAS0_TLBSEL(1)@h
 862        rlwimi  r0, r9, 16, MAS0_ESEL_MSK
 863        lis     r1, FSL_BOOKE_MAS1(1, 1, 0, 0, BOOKE_PAGESZ_1M)@h
 864        ori     r1, r1, FSL_BOOKE_MAS1(1, 1, 0, 0, BOOKE_PAGESZ_1M)@l
 865        li      r6, 0   /* DCSR effective address */
 866        ori     r2, r6, MAS2_I|MAS2_G
 867        li      r3, MAS3_SW|MAS3_SR
 868        li      r4, 0
 869        mtspr   MAS0, r0
 870        mtspr   MAS1, r1
 871        mtspr   MAS2, r2
 872        mtspr   MAS3, r3
 873        mtspr   MAS7, r4
 874        isync
 875        tlbwe
 876        isync
 877        msync
 878
 879        /* enable the timebase */
 880#define CTBENR  0xe2084
 881        li      r3, 1
 882        addis   r4, r7, CTBENR@ha
 883        stw     r3, CTBENR@l(r4)
 884        lwz     r3, CTBENR@l(r4)
 885        twi     0,r3,0
 886        isync
 887
 888        .macro  erratum_set_ccsr offset value
 889        addis   r3, r7, \offset@ha
 890        lis     r4, \value@h
 891        addi    r3, r3, \offset@l
 892        ori     r4, r4, \value@l
 893        bl      erratum_set_value
 894        .endm
 895
 896        .macro  erratum_set_dcsr offset value
 897        addis   r3, r6, \offset@ha
 898        lis     r4, \value@h
 899        addi    r3, r3, \offset@l
 900        ori     r4, r4, \value@l
 901        bl      erratum_set_value
 902        .endm
 903
 904        erratum_set_dcsr 0xb0e08 0xe0201800
 905        erratum_set_dcsr 0xb0e18 0xe0201800
 906        erratum_set_dcsr 0xb0e38 0xe0400000
 907        erratum_set_dcsr 0xb0008 0x00900000
 908        erratum_set_dcsr 0xb0e40 0xe00a0000
 909        erratum_set_ccsr 0x18600 CONFIG_SYS_FSL_CORENET_SNOOPVEC_COREONLY
 910#ifdef  CONFIG_RAMBOOT_PBL
 911        erratum_set_ccsr 0x10f00 0x495e5000
 912#else
 913        erratum_set_ccsr 0x10f00 0x415e5000
 914#endif
 915        erratum_set_ccsr 0x11f00 0x415e5000
 916
 917        /* Make temp mapping uncacheable again, if it was initially */
 918        bl      2f
 9192:      mflr    r3
 920        tlbsx   0, r3
 921        mfspr   r4, MAS2
 922        rlwimi  r4, r15, 0, MAS2_I
 923        rlwimi  r4, r15, 0, MAS2_G
 924        mtspr   MAS2, r4
 925        isync
 926        tlbwe
 927        isync
 928        msync
 929
 930        /* Clear the cache */
 931        lis     r3,(L1CSR1_ICFI|L1CSR1_ICLFR)@h
 932        ori     r3,r3,(L1CSR1_ICFI|L1CSR1_ICLFR)@l
 933        sync
 934        isync
 935        mtspr   SPRN_L1CSR1,r3
 936        isync
 9372:      sync
 938        mfspr   r4,SPRN_L1CSR1
 939        and.    r4,r4,r3
 940        bne     2b
 941
 942        lis     r3,(L1CSR1_CPE|L1CSR1_ICE)@h
 943        ori     r3,r3,(L1CSR1_CPE|L1CSR1_ICE)@l
 944        sync
 945        isync
 946        mtspr   SPRN_L1CSR1,r3
 947        isync
 9482:      sync
 949        mfspr   r4,SPRN_L1CSR1
 950        and.    r4,r4,r3
 951        beq     2b
 952
 953        /* Remove temporary mappings */
 954        lis     r0, MAS0_TLBSEL(1)@h
 955        rlwimi  r0, r9, 16, MAS0_ESEL_MSK
 956        li      r3, 0
 957        mtspr   MAS0, r0
 958        mtspr   MAS1, r3
 959        isync
 960        tlbwe
 961        isync
 962        msync
 963
 964        li      r3, 0
 965        stw     r3, 0xc08(r7)   /* LAWAR0 */
 966        lwz     r3, 0xc08(r7)
 967        isync
 968
 969        lis     r0, MAS0_TLBSEL(1)@h
 970        rlwimi  r0, r8, 16, MAS0_ESEL_MSK
 971        li      r3, 0
 972        mtspr   MAS0, r0
 973        mtspr   MAS1, r3
 974        isync
 975        tlbwe
 976        isync
 977        msync
 978
 979        b       9f
 980
 981        /* r3 = addr, r4 = value, clobbers r5, r11, r12 */
 982erratum_set_value:
 983        /* Lock two cache lines into I-Cache */
 984        sync
 985        mfspr   r11, SPRN_L1CSR1
 986        rlwinm  r11, r11, 0, ~L1CSR1_ICUL
 987        sync
 988        isync
 989        mtspr   SPRN_L1CSR1, r11
 990        isync
 991
 992        mflr    r12
 993        bl      5f
 9945:      mflr    r5
 995        addi    r5, r5, 2f - 5b
 996        icbtls  0, 0, r5
 997        addi    r5, r5, 64
 998
 999        sync
1000        mfspr   r11, SPRN_L1CSR1
10013:      andi.   r11, r11, L1CSR1_ICUL
1002        bne     3b
1003
1004        icbtls  0, 0, r5
1005        addi    r5, r5, 64
1006
1007        sync
1008        mfspr   r11, SPRN_L1CSR1
10093:      andi.   r11, r11, L1CSR1_ICUL
1010        bne     3b
1011
1012        b       2f
1013        .align  6
1014        /* Inside a locked cacheline, wait a while, write, then wait a while */
10152:      sync
1016
1017        mfspr   r5, SPRN_TBRL
1018        addis   r11, r5, 0x10000@h /* wait 65536 timebase ticks */
10194:      mfspr   r5, SPRN_TBRL
1020        subf.   r5, r5, r11
1021        bgt     4b
1022
1023        stw     r4, 0(r3)
1024
1025        mfspr   r5, SPRN_TBRL
1026        addis   r11, r5, 0x10000@h /* wait 65536 timebase ticks */
10274:      mfspr   r5, SPRN_TBRL
1028        subf.   r5, r5, r11
1029        bgt     4b
1030
1031        sync
1032
1033        /*
1034         * Fill out the rest of this cache line and the next with nops,
1035         * to ensure that nothing outside the locked area will be
1036         * fetched due to a branch.
1037         */
1038        .rept 19
1039        nop
1040        .endr
1041
1042        sync
1043        mfspr   r11, SPRN_L1CSR1
1044        rlwinm  r11, r11, 0, ~L1CSR1_ICUL
1045        sync
1046        isync
1047        mtspr   SPRN_L1CSR1, r11
1048        isync
1049
1050        mtlr    r12
1051        blr
1052
10539:
1054#endif
1055
1056create_init_ram_area:
1057        lis     r6,FSL_BOOKE_MAS0(1, 15, 0)@h
1058        ori     r6,r6,FSL_BOOKE_MAS0(1, 15, 0)@l
1059
1060#ifdef NOR_BOOT
1061        /* create a temp mapping in AS=1 to the 4M boot window */
1062        create_tlb1_entry 15, \
1063                1, BOOKE_PAGESZ_4M, \
1064                CONFIG_SYS_MONITOR_BASE & 0xffc00000, MAS2_I|MAS2_G, \
1065                0xffc00000, MAS3_SX|MAS3_SW|MAS3_SR, \
1066                0, r6
1067
1068#elif !defined(CONFIG_SYS_RAMBOOT) && defined(CONFIG_SECURE_BOOT)
1069        /* create a temp mapping in AS = 1 for Flash mapping
1070         * created by PBL for ISBC code
1071         */
1072        create_tlb1_entry 15, \
1073                1, BOOKE_PAGESZ_1M, \
1074                CONFIG_SYS_MONITOR_BASE & 0xfff00000, MAS2_I|MAS2_G, \
1075                CONFIG_SYS_PBI_FLASH_WINDOW & 0xfff00000, MAS3_SX|MAS3_SW|MAS3_SR, \
1076                0, r6
1077
1078/*
1079 * For Targets without CONFIG_SPL like P3, P5
1080 * and for targets with CONFIG_SPL like T1, T2, T4, only for
1081 * u-boot-spl i.e. CONFIG_SPL_BUILD
1082 */
1083#elif defined(CONFIG_RAMBOOT_PBL) && defined(CONFIG_SECURE_BOOT) && \
1084        (!defined(CONFIG_SPL) || defined(CONFIG_SPL_BUILD))
1085        /* create a temp mapping in AS = 1 for mapping CONFIG_SYS_MONITOR_BASE
1086         * to L3 Address configured by PBL for ISBC code
1087         */
1088        create_tlb1_entry 15, \
1089                1, BOOKE_PAGESZ_1M, \
1090                CONFIG_SYS_MONITOR_BASE & 0xfff00000, MAS2_I|MAS2_G, \
1091                CONFIG_SYS_INIT_L3_ADDR & 0xfff00000, MAS3_SX|MAS3_SW|MAS3_SR, \
1092                0, r6
1093
1094#else
1095        /*
1096         * create a temp mapping in AS=1 to the 1M CONFIG_SYS_MONITOR_BASE space, the main
1097         * image has been relocated to CONFIG_SYS_MONITOR_BASE on the second stage.
1098         */
1099        create_tlb1_entry 15, \
1100                1, BOOKE_PAGESZ_1M, \
1101                CONFIG_SYS_MONITOR_BASE & 0xfff00000, MAS2_I|MAS2_G, \
1102                CONFIG_SYS_MONITOR_BASE & 0xfff00000, MAS3_SX|MAS3_SW|MAS3_SR, \
1103                0, r6
1104#endif
1105
1106        /* create a temp mapping in AS=1 to the stack */
1107#if defined(CONFIG_SYS_INIT_RAM_ADDR_PHYS_LOW) && \
1108    defined(CONFIG_SYS_INIT_RAM_ADDR_PHYS_HIGH)
1109        create_tlb1_entry 14, \
1110                1, BOOKE_PAGESZ_16K, \
1111                CONFIG_SYS_INIT_RAM_ADDR, 0, \
1112                CONFIG_SYS_INIT_RAM_ADDR_PHYS_LOW, MAS3_SX|MAS3_SW|MAS3_SR, \
1113                CONFIG_SYS_INIT_RAM_ADDR_PHYS_HIGH, r6
1114
1115#else
1116        create_tlb1_entry 14, \
1117                1, BOOKE_PAGESZ_16K, \
1118                CONFIG_SYS_INIT_RAM_ADDR, 0, \
1119                CONFIG_SYS_INIT_RAM_ADDR, MAS3_SX|MAS3_SW|MAS3_SR, \
1120                0, r6
1121#endif
1122
1123        lis     r6,MSR_IS|MSR_DS|MSR_DE@h
1124        ori     r6,r6,MSR_IS|MSR_DS|MSR_DE@l
1125        lis     r7,switch_as@h
1126        ori     r7,r7,switch_as@l
1127
1128        mtspr   SPRN_SRR0,r7
1129        mtspr   SPRN_SRR1,r6
1130        rfi
1131
1132switch_as:
1133/* L1 DCache is used for initial RAM */
1134
1135        /* Allocate Initial RAM in data cache.
1136         */
1137        lis     r3,CONFIG_SYS_INIT_RAM_ADDR@h
1138        ori     r3,r3,CONFIG_SYS_INIT_RAM_ADDR@l
1139        mfspr   r2, L1CFG0
1140        andi.   r2, r2, 0x1ff
1141        /* cache size * 1024 / (2 * L1 line size) */
1142        slwi    r2, r2, (10 - 1 - L1_CACHE_SHIFT)
1143        mtctr   r2
1144        li      r0,0
11451:
1146        dcbz    r0,r3
1147#ifdef CONFIG_E6500     /* Lock/unlock L2 cache long with L1 */
1148        dcbtls  2, r0, r3
1149        dcbtls  0, r0, r3
1150#else
1151        dcbtls  0, r0, r3
1152#endif
1153        addi    r3,r3,CONFIG_SYS_CACHELINE_SIZE
1154        bdnz    1b
1155
1156        /* Jump out the last 4K page and continue to 'normal' start */
1157#if defined(CONFIG_SYS_RAMBOOT) || defined(CONFIG_SPL)
1158        /* We assume that we're already running at the address we're linked at */
1159        b       _start_cont
1160#else
1161        /* Calculate absolute address in FLASH and jump there           */
1162        /*--------------------------------------------------------------*/
1163        lis     r3,CONFIG_SYS_MONITOR_BASE@h
1164        ori     r3,r3,CONFIG_SYS_MONITOR_BASE@l
1165        addi    r3,r3,_start_cont - _start
1166        mtlr    r3
1167        blr
1168#endif
1169
1170        .text
1171        .globl  _start
1172_start:
1173        .long   0x27051956              /* U-BOOT Magic Number */
1174        .globl  version_string
1175version_string:
1176        .ascii U_BOOT_VERSION_STRING, "\0"
1177
1178        .align  4
1179        .globl  _start_cont
1180_start_cont:
1181        /* Setup the stack in initial RAM,could be L2-as-SRAM or L1 dcache*/
1182        lis     r3,(CONFIG_SYS_INIT_RAM_ADDR)@h
1183        ori     r3,r3,((CONFIG_SYS_INIT_SP_OFFSET-16)&~0xf)@l /* Align to 16 */
1184
1185#if CONFIG_VAL(SYS_MALLOC_F_LEN)
1186#if CONFIG_VAL(SYS_MALLOC_F_LEN) + GENERATED_GBL_DATA_SIZE > CONFIG_SYS_INIT_RAM_SIZE
1187#error "SYS_MALLOC_F_LEN too large to fit into initial RAM."
1188#endif
1189
1190        /* Leave 16+ byte for back chain termination and NULL return address */
1191        subi    r3,r3,((CONFIG_VAL(SYS_MALLOC_F_LEN)+16+15)&~0xf)
1192#endif
1193
1194        /* End of RAM */
1195        lis     r4,(CONFIG_SYS_INIT_RAM_ADDR)@h
1196        ori     r4,r4,(CONFIG_SYS_INIT_RAM_SIZE)@l
1197
1198        li      r0,0
1199
12001:      subi    r4,r4,4
1201        stw     r0,0(r4)
1202        cmplw   r4,r3
1203        bne     1b
1204
1205#if CONFIG_VAL(SYS_MALLOC_F_LEN)
1206        lis     r4,(CONFIG_SYS_INIT_RAM_ADDR)@h
1207        ori     r4,r4,(CONFIG_SYS_GBL_DATA_OFFSET)@l
1208
1209        addi    r3,r3,16        /* Pre-relocation malloc area */
1210        stw     r3,GD_MALLOC_BASE(r4)
1211        subi    r3,r3,16
1212#endif
1213        li      r0,0
1214        stw     r0,0(r3)        /* Terminate Back Chain */
1215        stw     r0,+4(r3)       /* NULL return address. */
1216        mr      r1,r3           /* Transfer to SP(r1) */
1217
1218        GET_GOT
1219
1220        /* Pass our potential ePAPR device tree pointer to cpu_init_early_f */
1221        mr      r3, r24
1222
1223        bl      cpu_init_early_f
1224
1225        /* switch back to AS = 0 */
1226        lis     r3,(MSR_CE|MSR_ME|MSR_DE)@h
1227        ori     r3,r3,(MSR_CE|MSR_ME|MSR_DE)@l
1228        mtmsr   r3
1229        isync
1230
1231        bl      cpu_init_f      /* return boot_flag for calling board_init_f */
1232        bl      board_init_f
1233        isync
1234
1235        /* NOTREACHED - board_init_f() does not return */
1236
1237#ifndef MINIMAL_SPL
1238        .globl  _start_of_vectors
1239_start_of_vectors:
1240
1241/* Critical input. */
1242        CRIT_EXCEPTION(0x0100, CriticalInput, CritcalInputException)
1243
1244/* Machine check */
1245        MCK_EXCEPTION(0x200, MachineCheck, MachineCheckException)
1246
1247/* Data Storage exception. */
1248        STD_EXCEPTION(0x0300, DataStorage, UnknownException)
1249
1250/* Instruction Storage exception. */
1251        STD_EXCEPTION(0x0400, InstStorage, UnknownException)
1252
1253/* External Interrupt exception. */
1254        STD_EXCEPTION(0x0500, ExtInterrupt, ExtIntException)
1255
1256/* Alignment exception. */
1257Alignment:
1258        EXCEPTION_PROLOG(SRR0, SRR1)
1259        mfspr   r4,DAR
1260        stw     r4,_DAR(r21)
1261        mfspr   r5,DSISR
1262        stw     r5,_DSISR(r21)
1263        addi    r3,r1,STACK_FRAME_OVERHEAD
1264        EXC_XFER_TEMPLATE(0x600, Alignment, AlignmentException,
1265                MSR_KERNEL, COPY_EE)
1266
1267/* Program check exception */
1268ProgramCheck:
1269        EXCEPTION_PROLOG(SRR0, SRR1)
1270        addi    r3,r1,STACK_FRAME_OVERHEAD
1271        EXC_XFER_TEMPLATE(0x700, ProgramCheck, ProgramCheckException,
1272                MSR_KERNEL, COPY_EE)
1273
1274        /* No FPU on MPC85xx.  This exception is not supposed to happen.
1275        */
1276        STD_EXCEPTION(0x0800, FPUnavailable, UnknownException)
1277        STD_EXCEPTION(0x0900, SystemCall, UnknownException)
1278        STD_EXCEPTION(0x0a00, Decrementer, timer_interrupt)
1279        STD_EXCEPTION(0x0b00, IntervalTimer, UnknownException)
1280        STD_EXCEPTION(0x0c00, WatchdogTimer, UnknownException)
1281
1282        STD_EXCEPTION(0x0d00, DataTLBError, UnknownException)
1283        STD_EXCEPTION(0x0e00, InstructionTLBError, UnknownException)
1284
1285        CRIT_EXCEPTION(0x0f00, DebugBreakpoint, DebugException )
1286
1287        .globl  _end_of_vectors
1288_end_of_vectors:
1289
1290
1291        . = . + (0x100 - ( . & 0xff ))  /* align for debug */
1292
1293/*
1294 * This code finishes saving the registers to the exception frame
1295 * and jumps to the appropriate handler for the exception.
1296 * Register r21 is pointer into trap frame, r1 has new stack pointer.
1297 * r23 is the address of the handler.
1298 */
1299        .globl  transfer_to_handler
1300transfer_to_handler:
1301        SAVE_GPR(7, r21)
1302        SAVE_4GPRS(8, r21)
1303        SAVE_8GPRS(12, r21)
1304        SAVE_8GPRS(24, r21)
1305
1306        li      r22,0
1307        stw     r22,RESULT(r21)
1308        mtspr   SPRG2,r22               /* r1 is now kernel sp */
1309
1310        mtctr   r23                     /* virtual address of handler */
1311        mtmsr   r20
1312        bctrl
1313
1314int_return:
1315        mfmsr   r28             /* Disable interrupts */
1316        li      r4,0
1317        ori     r4,r4,MSR_EE
1318        andc    r28,r28,r4
1319        SYNC                    /* Some chip revs need this... */
1320        mtmsr   r28
1321        SYNC
1322        lwz     r2,_CTR(r1)
1323        lwz     r0,_LINK(r1)
1324        mtctr   r2
1325        mtlr    r0
1326        lwz     r2,_XER(r1)
1327        lwz     r0,_CCR(r1)
1328        mtspr   XER,r2
1329        mtcrf   0xFF,r0
1330        REST_10GPRS(3, r1)
1331        REST_10GPRS(13, r1)
1332        REST_8GPRS(23, r1)
1333        REST_GPR(31, r1)
1334        lwz     r2,_NIP(r1)     /* Restore environment */
1335        lwz     r0,_MSR(r1)
1336        mtspr   SRR0,r2
1337        mtspr   SRR1,r0
1338        lwz     r0,GPR0(r1)
1339        lwz     r2,GPR2(r1)
1340        lwz     r1,GPR1(r1)
1341        SYNC
1342        rfi
1343
1344/* Cache functions.
1345*/
1346.globl flush_icache
1347flush_icache:
1348.globl invalidate_icache
1349invalidate_icache:
1350        mfspr   r0,L1CSR1
1351        ori     r0,r0,L1CSR1_ICFI
1352        msync
1353        isync
1354        mtspr   L1CSR1,r0
1355        isync
1356        blr                             /* entire I cache */
1357
1358.globl invalidate_dcache
1359invalidate_dcache:
1360        mfspr   r0,L1CSR0
1361        ori     r0,r0,L1CSR0_DCFI
1362        msync
1363        isync
1364        mtspr   L1CSR0,r0
1365        isync
1366        blr
1367
1368        .globl  icache_enable
1369icache_enable:
1370        mflr    r8
1371        bl      invalidate_icache
1372        mtlr    r8
1373        isync
1374        mfspr   r4,L1CSR1
1375        ori     r4,r4,(L1CSR1_CPE | L1CSR1_ICE)@l
1376        oris    r4,r4,(L1CSR1_CPE | L1CSR1_ICE)@h
1377        mtspr   L1CSR1,r4
1378        isync
1379        blr
1380
1381        .globl  icache_disable
1382icache_disable:
1383        mfspr   r0,L1CSR1
1384        lis     r3,0
1385        ori     r3,r3,L1CSR1_ICE
1386        andc    r0,r0,r3
1387        mtspr   L1CSR1,r0
1388        isync
1389        blr
1390
1391        .globl  icache_status
1392icache_status:
1393        mfspr   r3,L1CSR1
1394        andi.   r3,r3,L1CSR1_ICE
1395        blr
1396
1397        .globl  dcache_enable
1398dcache_enable:
1399        mflr    r8
1400        bl      invalidate_dcache
1401        mtlr    r8
1402        isync
1403        mfspr   r0,L1CSR0
1404        ori     r0,r0,(L1CSR0_CPE |  L1CSR0_DCE)@l
1405        oris    r0,r0,(L1CSR0_CPE |  L1CSR0_DCE)@h
1406        msync
1407        isync
1408        mtspr   L1CSR0,r0
1409        isync
1410        blr
1411
1412        .globl  dcache_disable
1413dcache_disable:
1414        mfspr   r3,L1CSR0
1415        lis     r4,0
1416        ori     r4,r4,L1CSR0_DCE
1417        andc    r3,r3,r4
1418        mtspr   L1CSR0,r3
1419        isync
1420        blr
1421
1422        .globl  dcache_status
1423dcache_status:
1424        mfspr   r3,L1CSR0
1425        andi.   r3,r3,L1CSR0_DCE
1426        blr
1427
1428/*------------------------------------------------------------------------------- */
1429/* Function:     in8 */
1430/* Description:  Input 8 bits */
1431/*------------------------------------------------------------------------------- */
1432        .globl  in8
1433in8:
1434        lbz     r3,0x0000(r3)
1435        blr
1436
1437/*------------------------------------------------------------------------------- */
1438/* Function:     out8 */
1439/* Description:  Output 8 bits */
1440/*------------------------------------------------------------------------------- */
1441        .globl  out8
1442out8:
1443        stb     r4,0x0000(r3)
1444        sync
1445        blr
1446
1447/*------------------------------------------------------------------------------- */
1448/* Function:     out16 */
1449/* Description:  Output 16 bits */
1450/*------------------------------------------------------------------------------- */
1451        .globl  out16
1452out16:
1453        sth     r4,0x0000(r3)
1454        sync
1455        blr
1456
1457/*------------------------------------------------------------------------------- */
1458/* Function:     out16r */
1459/* Description:  Byte reverse and output 16 bits */
1460/*------------------------------------------------------------------------------- */
1461        .globl  out16r
1462out16r:
1463        sthbrx  r4,r0,r3
1464        sync
1465        blr
1466
1467/*------------------------------------------------------------------------------- */
1468/* Function:     out32 */
1469/* Description:  Output 32 bits */
1470/*------------------------------------------------------------------------------- */
1471        .globl  out32
1472out32:
1473        stw     r4,0x0000(r3)
1474        sync
1475        blr
1476
1477/*------------------------------------------------------------------------------- */
1478/* Function:     out32r */
1479/* Description:  Byte reverse and output 32 bits */
1480/*------------------------------------------------------------------------------- */
1481        .globl  out32r
1482out32r:
1483        stwbrx  r4,r0,r3
1484        sync
1485        blr
1486
1487/*------------------------------------------------------------------------------- */
1488/* Function:     in16 */
1489/* Description:  Input 16 bits */
1490/*------------------------------------------------------------------------------- */
1491        .globl  in16
1492in16:
1493        lhz     r3,0x0000(r3)
1494        blr
1495
1496/*------------------------------------------------------------------------------- */
1497/* Function:     in16r */
1498/* Description:  Input 16 bits and byte reverse */
1499/*------------------------------------------------------------------------------- */
1500        .globl  in16r
1501in16r:
1502        lhbrx   r3,r0,r3
1503        blr
1504
1505/*------------------------------------------------------------------------------- */
1506/* Function:     in32 */
1507/* Description:  Input 32 bits */
1508/*------------------------------------------------------------------------------- */
1509        .globl  in32
1510in32:
1511        lwz     3,0x0000(3)
1512        blr
1513
1514/*------------------------------------------------------------------------------- */
1515/* Function:     in32r */
1516/* Description:  Input 32 bits and byte reverse */
1517/*------------------------------------------------------------------------------- */
1518        .globl  in32r
1519in32r:
1520        lwbrx   r3,r0,r3
1521        blr
1522#endif  /* !MINIMAL_SPL */
1523
1524/*------------------------------------------------------------------------------*/
1525
1526/*
1527 * void write_tlb(mas0, mas1, mas2, mas3, mas7)
1528 */
1529        .globl  write_tlb
1530write_tlb:
1531        mtspr   MAS0,r3
1532        mtspr   MAS1,r4
1533        mtspr   MAS2,r5
1534        mtspr   MAS3,r6
1535#ifdef CONFIG_ENABLE_36BIT_PHYS
1536        mtspr   MAS7,r7
1537#endif
1538        li      r3,0
1539#ifdef CONFIG_SYS_BOOK3E_HV
1540        mtspr   MAS8,r3
1541#endif
1542        isync
1543        tlbwe
1544        msync
1545        isync
1546        blr
1547
1548/*
1549 * void relocate_code (addr_sp, gd, addr_moni)
1550 *
1551 * This "function" does not return, instead it continues in RAM
1552 * after relocating the monitor code.
1553 *
1554 * r3 = dest
1555 * r4 = src
1556 * r5 = length in bytes
1557 * r6 = cachelinesize
1558 */
1559        .globl  relocate_code
1560relocate_code:
1561        mr      r1,r3           /* Set new stack pointer                */
1562        mr      r9,r4           /* Save copy of Init Data pointer       */
1563        mr      r10,r5          /* Save copy of Destination Address     */
1564
1565        GET_GOT
1566#ifndef CONFIG_SPL_SKIP_RELOCATE
1567        mr      r3,r5                           /* Destination Address  */
1568        lis     r4,CONFIG_SYS_MONITOR_BASE@h            /* Source      Address  */
1569        ori     r4,r4,CONFIG_SYS_MONITOR_BASE@l
1570        lwz     r5,GOT(__init_end)
1571        sub     r5,r5,r4
1572        li      r6,CONFIG_SYS_CACHELINE_SIZE            /* Cache Line Size      */
1573
1574        /*
1575         * Fix GOT pointer:
1576         *
1577         * New GOT-PTR = (old GOT-PTR - CONFIG_SYS_MONITOR_BASE) + Destination Address
1578         *
1579         * Offset:
1580         */
1581        sub     r15,r10,r4
1582
1583        /* First our own GOT */
1584        add     r12,r12,r15
1585        /* the the one used by the C code */
1586        add     r30,r30,r15
1587
1588        /*
1589         * Now relocate code
1590         */
1591
1592        cmplw   cr1,r3,r4
1593        addi    r0,r5,3
1594        srwi.   r0,r0,2
1595        beq     cr1,4f          /* In place copy is not necessary       */
1596        beq     7f              /* Protect against 0 count              */
1597        mtctr   r0
1598        bge     cr1,2f
1599
1600        la      r8,-4(r4)
1601        la      r7,-4(r3)
16021:      lwzu    r0,4(r8)
1603        stwu    r0,4(r7)
1604        bdnz    1b
1605        b       4f
1606
16072:      slwi    r0,r0,2
1608        add     r8,r4,r0
1609        add     r7,r3,r0
16103:      lwzu    r0,-4(r8)
1611        stwu    r0,-4(r7)
1612        bdnz    3b
1613
1614/*
1615 * Now flush the cache: note that we must start from a cache aligned
1616 * address. Otherwise we might miss one cache line.
1617 */
16184:      cmpwi   r6,0
1619        add     r5,r3,r5
1620        beq     7f              /* Always flush prefetch queue in any case */
1621        subi    r0,r6,1
1622        andc    r3,r3,r0
1623        mr      r4,r3
16245:      dcbst   0,r4
1625        add     r4,r4,r6
1626        cmplw   r4,r5
1627        blt     5b
1628        sync                    /* Wait for all dcbst to complete on bus */
1629        mr      r4,r3
16306:      icbi    0,r4
1631        add     r4,r4,r6
1632        cmplw   r4,r5
1633        blt     6b
16347:      sync                    /* Wait for all icbi to complete on bus */
1635        isync
1636
1637/*
1638 * We are done. Do not return, instead branch to second part of board
1639 * initialization, now running from RAM.
1640 */
1641
1642        addi    r0,r10,in_ram - _start
1643
1644        /*
1645         * As IVPR is going to point RAM address,
1646         * Make sure IVOR15 has valid opcode to support debugger
1647         */
1648        mtspr   IVOR15,r0
1649
1650        /*
1651         * Re-point the IVPR at RAM
1652         */
1653        mtspr   IVPR,r10
1654
1655        mtlr    r0
1656        blr                             /* NEVER RETURNS! */
1657#endif
1658        .globl  in_ram
1659in_ram:
1660
1661        /*
1662         * Relocation Function, r12 point to got2+0x8000
1663         *
1664         * Adjust got2 pointers, no need to check for 0, this code
1665         * already puts a few entries in the table.
1666         */
1667        li      r0,__got2_entries@sectoff@l
1668        la      r3,GOT(_GOT2_TABLE_)
1669        lwz     r11,GOT(_GOT2_TABLE_)
1670        mtctr   r0
1671        sub     r11,r3,r11
1672        addi    r3,r3,-4
16731:      lwzu    r0,4(r3)
1674        cmpwi   r0,0
1675        beq-    2f
1676        add     r0,r0,r11
1677        stw     r0,0(r3)
16782:      bdnz    1b
1679
1680        /*
1681         * Now adjust the fixups and the pointers to the fixups
1682         * in case we need to move ourselves again.
1683         */
1684        li      r0,__fixup_entries@sectoff@l
1685        lwz     r3,GOT(_FIXUP_TABLE_)
1686        cmpwi   r0,0
1687        mtctr   r0
1688        addi    r3,r3,-4
1689        beq     4f
16903:      lwzu    r4,4(r3)
1691        lwzux   r0,r4,r11
1692        cmpwi   r0,0
1693        add     r0,r0,r11
1694        stw     r4,0(r3)
1695        beq-    5f
1696        stw     r0,0(r4)
16975:      bdnz    3b
16984:
1699clear_bss:
1700        /*
1701         * Now clear BSS segment
1702         */
1703        lwz     r3,GOT(__bss_start)
1704        lwz     r4,GOT(__bss_end)
1705
1706        cmplw   0,r3,r4
1707        beq     6f
1708
1709        li      r0,0
17105:
1711        stw     r0,0(r3)
1712        addi    r3,r3,4
1713        cmplw   0,r3,r4
1714        blt     5b
17156:
1716
1717        mr      r3,r9           /* Init Data pointer            */
1718        mr      r4,r10          /* Destination Address          */
1719        bl      board_init_r
1720
1721#ifndef MINIMAL_SPL
1722        /*
1723         * Copy exception vector code to low memory
1724         *
1725         * r3: dest_addr
1726         * r7: source address, r8: end address, r9: target address
1727         */
1728        .globl  trap_init
1729trap_init:
1730        mflr    r11
1731        bl      _GLOBAL_OFFSET_TABLE_-4
1732        mflr    r12
1733
1734        /* Update IVORs as per relocation */
1735        mtspr   IVPR,r3
1736
1737        lwz     r4,CriticalInput@got(r12)
1738        mtspr   IVOR0,r4        /* 0: Critical input */
1739        lwz     r4,MachineCheck@got(r12)
1740        mtspr   IVOR1,r4        /* 1: Machine check */
1741        lwz     r4,DataStorage@got(r12)
1742        mtspr   IVOR2,r4        /* 2: Data storage */
1743        lwz     r4,InstStorage@got(r12)
1744        mtspr   IVOR3,r4        /* 3: Instruction storage */
1745        lwz     r4,ExtInterrupt@got(r12)
1746        mtspr   IVOR4,r4        /* 4: External interrupt */
1747        lwz     r4,Alignment@got(r12)
1748        mtspr   IVOR5,r4        /* 5: Alignment */
1749        lwz     r4,ProgramCheck@got(r12)
1750        mtspr   IVOR6,r4        /* 6: Program check */
1751        lwz     r4,FPUnavailable@got(r12)
1752        mtspr   IVOR7,r4        /* 7: floating point unavailable */
1753        lwz     r4,SystemCall@got(r12)
1754        mtspr   IVOR8,r4        /* 8: System call */
1755        /* 9: Auxiliary processor unavailable(unsupported) */
1756        lwz     r4,Decrementer@got(r12)
1757        mtspr   IVOR10,r4       /* 10: Decrementer */
1758        lwz     r4,IntervalTimer@got(r12)
1759        mtspr   IVOR11,r4       /* 11: Interval timer */
1760        lwz     r4,WatchdogTimer@got(r12)
1761        mtspr   IVOR12,r4       /* 12: Watchdog timer */
1762        lwz     r4,DataTLBError@got(r12)
1763        mtspr   IVOR13,r4       /* 13: Data TLB error */
1764        lwz     r4,InstructionTLBError@got(r12)
1765        mtspr   IVOR14,r4       /* 14: Instruction TLB error */
1766        lwz     r4,DebugBreakpoint@got(r12)
1767        mtspr   IVOR15,r4       /* 15: Debug */
1768
1769        mtlr    r11
1770        blr
1771
1772.globl unlock_ram_in_cache
1773unlock_ram_in_cache:
1774        /* invalidate the INIT_RAM section */
1775        lis     r3,(CONFIG_SYS_INIT_RAM_ADDR & ~(CONFIG_SYS_CACHELINE_SIZE-1))@h
1776        ori     r3,r3,(CONFIG_SYS_INIT_RAM_ADDR & ~(CONFIG_SYS_CACHELINE_SIZE-1))@l
1777        mfspr   r4,L1CFG0
1778        andi.   r4,r4,0x1ff
1779        slwi    r4,r4,(10 - 1 - L1_CACHE_SHIFT)
1780        mtctr   r4
17811:      dcbi    r0,r3
1782#ifdef CONFIG_E6500     /* lock/unlock L2 cache long with L1 */
1783        dcblc   2, r0, r3
1784        dcblc   0, r0, r3
1785#else
1786        dcblc   r0,r3
1787#endif
1788        addi    r3,r3,CONFIG_SYS_CACHELINE_SIZE
1789        bdnz    1b
1790        sync
1791
1792        /* Invalidate the TLB entries for the cache */
1793        lis     r3,CONFIG_SYS_INIT_RAM_ADDR@h
1794        ori     r3,r3,CONFIG_SYS_INIT_RAM_ADDR@l
1795        tlbivax 0,r3
1796        addi    r3,r3,0x1000
1797        tlbivax 0,r3
1798        addi    r3,r3,0x1000
1799        tlbivax 0,r3
1800        addi    r3,r3,0x1000
1801        tlbivax 0,r3
1802        isync
1803        blr
1804
1805.globl flush_dcache
1806flush_dcache:
1807        mfspr   r3,SPRN_L1CFG0
1808
1809        rlwinm  r5,r3,9,3       /* Extract cache block size */
1810        twlgti  r5,1            /* Only 32 and 64 byte cache blocks
1811                                 * are currently defined.
1812                                 */
1813        li      r4,32
1814        subfic  r6,r5,2         /* r6 = log2(1KiB / cache block size) -
1815                                 *      log2(number of ways)
1816                                 */
1817        slw     r5,r4,r5        /* r5 = cache block size */
1818
1819        rlwinm  r7,r3,0,0xff    /* Extract number of KiB in the cache */
1820        mulli   r7,r7,13        /* An 8-way cache will require 13
1821                                 * loads per set.
1822                                 */
1823        slw     r7,r7,r6
1824
1825        /* save off HID0 and set DCFA */
1826        mfspr   r8,SPRN_HID0
1827        ori     r9,r8,HID0_DCFA@l
1828        mtspr   SPRN_HID0,r9
1829        isync
1830
1831        lis     r4,0
1832        mtctr   r7
1833
18341:      lwz     r3,0(r4)        /* Load... */
1835        add     r4,r4,r5
1836        bdnz    1b
1837
1838        msync
1839        lis     r4,0
1840        mtctr   r7
1841
18421:      dcbf    0,r4            /* ...and flush. */
1843        add     r4,r4,r5
1844        bdnz    1b
1845
1846        /* restore HID0 */
1847        mtspr   SPRN_HID0,r8
1848        isync
1849
1850        blr
1851#endif /* !MINIMAL_SPL */
1852