linux/arch/arm/mm/flush.c
<<
>>
Prefs
   1/*
   2 *  linux/arch/arm/mm/flush.c
   3 *
   4 *  Copyright (C) 1995-2002 Russell King
   5 *
   6 * This program is free software; you can redistribute it and/or modify
   7 * it under the terms of the GNU General Public License version 2 as
   8 * published by the Free Software Foundation.
   9 */
  10#include <linux/module.h>
  11#include <linux/mm.h>
  12#include <linux/pagemap.h>
  13
  14#include <asm/cacheflush.h>
  15#include <asm/cachetype.h>
  16#include <asm/system.h>
  17#include <asm/tlbflush.h>
  18
  19#include "mm.h"
  20
  21#ifdef CONFIG_CPU_CACHE_VIPT
  22
  23#define ALIAS_FLUSH_START       0xffff4000
  24
  25static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr)
  26{
  27        unsigned long to = ALIAS_FLUSH_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
  28        const int zero = 0;
  29
  30        set_pte_ext(TOP_PTE(to), pfn_pte(pfn, PAGE_KERNEL), 0);
  31        flush_tlb_kernel_page(to);
  32
  33        asm(    "mcrr   p15, 0, %1, %0, c14\n"
  34        "       mcr     p15, 0, %2, c7, c10, 4"
  35            :
  36            : "r" (to), "r" (to + PAGE_SIZE - L1_CACHE_BYTES), "r" (zero)
  37            : "cc");
  38        __flush_icache_all();
  39}
  40
  41void flush_cache_mm(struct mm_struct *mm)
  42{
  43        if (cache_is_vivt()) {
  44                if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
  45                        __cpuc_flush_user_all();
  46                return;
  47        }
  48
  49        if (cache_is_vipt_aliasing()) {
  50                asm(    "mcr    p15, 0, %0, c7, c14, 0\n"
  51                "       mcr     p15, 0, %0, c7, c10, 4"
  52                    :
  53                    : "r" (0)
  54                    : "cc");
  55                __flush_icache_all();
  56        }
  57}
  58
  59void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end)
  60{
  61        if (cache_is_vivt()) {
  62                if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm)))
  63                        __cpuc_flush_user_range(start & PAGE_MASK, PAGE_ALIGN(end),
  64                                                vma->vm_flags);
  65                return;
  66        }
  67
  68        if (cache_is_vipt_aliasing()) {
  69                asm(    "mcr    p15, 0, %0, c7, c14, 0\n"
  70                "       mcr     p15, 0, %0, c7, c10, 4"
  71                    :
  72                    : "r" (0)
  73                    : "cc");
  74                __flush_icache_all();
  75        }
  76}
  77
  78void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn)
  79{
  80        if (cache_is_vivt()) {
  81                if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm))) {
  82                        unsigned long addr = user_addr & PAGE_MASK;
  83                        __cpuc_flush_user_range(addr, addr + PAGE_SIZE, vma->vm_flags);
  84                }
  85                return;
  86        }
  87
  88        if (cache_is_vipt_aliasing())
  89                flush_pfn_alias(pfn, user_addr);
  90}
  91
  92void flush_ptrace_access(struct vm_area_struct *vma, struct page *page,
  93                         unsigned long uaddr, void *kaddr,
  94                         unsigned long len, int write)
  95{
  96        if (cache_is_vivt()) {
  97                if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm))) {
  98                        unsigned long addr = (unsigned long)kaddr;
  99                        __cpuc_coherent_kern_range(addr, addr + len);
 100                }
 101                return;
 102        }
 103
 104        if (cache_is_vipt_aliasing()) {
 105                flush_pfn_alias(page_to_pfn(page), uaddr);
 106                return;
 107        }
 108
 109        /* VIPT non-aliasing cache */
 110        if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm)) &&
 111            vma->vm_flags & VM_EXEC) {
 112                unsigned long addr = (unsigned long)kaddr;
 113                /* only flushing the kernel mapping on non-aliasing VIPT */
 114                __cpuc_coherent_kern_range(addr, addr + len);
 115        }
 116}
 117#else
 118#define flush_pfn_alias(pfn,vaddr)      do { } while (0)
 119#endif
 120
 121void __flush_dcache_page(struct address_space *mapping, struct page *page)
 122{
 123        /*
 124         * Writeback any data associated with the kernel mapping of this
 125         * page.  This ensures that data in the physical page is mutually
 126         * coherent with the kernels mapping.
 127         */
 128#ifdef CONFIG_HIGHMEM
 129        /*
 130         * kmap_atomic() doesn't set the page virtual address, and
 131         * kunmap_atomic() takes care of cache flushing already.
 132         */
 133        if (page_address(page))
 134#endif
 135                __cpuc_flush_dcache_page(page_address(page));
 136
 137        /*
 138         * If this is a page cache page, and we have an aliasing VIPT cache,
 139         * we only need to do one flush - which would be at the relevant
 140         * userspace colour, which is congruent with page->index.
 141         */
 142        if (mapping && cache_is_vipt_aliasing())
 143                flush_pfn_alias(page_to_pfn(page),
 144                                page->index << PAGE_CACHE_SHIFT);
 145}
 146
 147static void __flush_dcache_aliases(struct address_space *mapping, struct page *page)
 148{
 149        struct mm_struct *mm = current->active_mm;
 150        struct vm_area_struct *mpnt;
 151        struct prio_tree_iter iter;
 152        pgoff_t pgoff;
 153
 154        /*
 155         * There are possible user space mappings of this page:
 156         * - VIVT cache: we need to also write back and invalidate all user
 157         *   data in the current VM view associated with this page.
 158         * - aliasing VIPT: we only need to find one mapping of this page.
 159         */
 160        pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT);
 161
 162        flush_dcache_mmap_lock(mapping);
 163        vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) {
 164                unsigned long offset;
 165
 166                /*
 167                 * If this VMA is not in our MM, we can ignore it.
 168                 */
 169                if (mpnt->vm_mm != mm)
 170                        continue;
 171                if (!(mpnt->vm_flags & VM_MAYSHARE))
 172                        continue;
 173                offset = (pgoff - mpnt->vm_pgoff) << PAGE_SHIFT;
 174                flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page));
 175        }
 176        flush_dcache_mmap_unlock(mapping);
 177}
 178
 179/*
 180 * Ensure cache coherency between kernel mapping and userspace mapping
 181 * of this page.
 182 *
 183 * We have three cases to consider:
 184 *  - VIPT non-aliasing cache: fully coherent so nothing required.
 185 *  - VIVT: fully aliasing, so we need to handle every alias in our
 186 *          current VM view.
 187 *  - VIPT aliasing: need to handle one alias in our current VM view.
 188 *
 189 * If we need to handle aliasing:
 190 *  If the page only exists in the page cache and there are no user
 191 *  space mappings, we can be lazy and remember that we may have dirty
 192 *  kernel cache lines for later.  Otherwise, we assume we have
 193 *  aliasing mappings.
 194 *
 195 * Note that we disable the lazy flush for SMP.
 196 */
 197void flush_dcache_page(struct page *page)
 198{
 199        struct address_space *mapping = page_mapping(page);
 200
 201#ifndef CONFIG_SMP
 202        if (!PageHighMem(page) && mapping && !mapping_mapped(mapping))
 203                set_bit(PG_dcache_dirty, &page->flags);
 204        else
 205#endif
 206        {
 207                __flush_dcache_page(mapping, page);
 208                if (mapping && cache_is_vivt())
 209                        __flush_dcache_aliases(mapping, page);
 210                else if (mapping)
 211                        __flush_icache_all();
 212        }
 213}
 214EXPORT_SYMBOL(flush_dcache_page);
 215
 216/*
 217 * Flush an anonymous page so that users of get_user_pages()
 218 * can safely access the data.  The expected sequence is:
 219 *
 220 *  get_user_pages()
 221 *    -> flush_anon_page
 222 *  memcpy() to/from page
 223 *  if written to page, flush_dcache_page()
 224 */
 225void __flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr)
 226{
 227        unsigned long pfn;
 228
 229        /* VIPT non-aliasing caches need do nothing */
 230        if (cache_is_vipt_nonaliasing())
 231                return;
 232
 233        /*
 234         * Write back and invalidate userspace mapping.
 235         */
 236        pfn = page_to_pfn(page);
 237        if (cache_is_vivt()) {
 238                flush_cache_page(vma, vmaddr, pfn);
 239        } else {
 240                /*
 241                 * For aliasing VIPT, we can flush an alias of the
 242                 * userspace address only.
 243                 */
 244                flush_pfn_alias(pfn, vmaddr);
 245        }
 246
 247        /*
 248         * Invalidate kernel mapping.  No data should be contained
 249         * in this mapping of the page.  FIXME: this is overkill
 250         * since we actually ask for a write-back and invalidate.
 251         */
 252        __cpuc_flush_dcache_page(page_address(page));
 253}
 254