linux/arch/arm/include/asm/tlb.h
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0-only */
   2/*
   3 *  arch/arm/include/asm/tlb.h
   4 *
   5 *  Copyright (C) 2002 Russell King
   6 *
   7 *  Experimentation shows that on a StrongARM, it appears to be faster
   8 *  to use the "invalidate whole tlb" rather than "invalidate single
   9 *  tlb" for this.
  10 *
  11 *  This appears true for both the process fork+exit case, as well as
  12 *  the munmap-large-area case.
  13 */
  14#ifndef __ASMARM_TLB_H
  15#define __ASMARM_TLB_H
  16
  17#include <asm/cacheflush.h>
  18
  19#ifndef CONFIG_MMU
  20
  21#include <linux/pagemap.h>
  22
  23#define tlb_flush(tlb)  ((void) tlb)
  24
  25#include <asm-generic/tlb.h>
  26
  27#else /* !CONFIG_MMU */
  28
  29#include <linux/swap.h>
  30#include <asm/pgalloc.h>
  31#include <asm/tlbflush.h>
  32
  33static inline void __tlb_remove_table(void *_table)
  34{
  35        free_page_and_swap_cache((struct page *)_table);
  36}
  37
  38#include <asm-generic/tlb.h>
  39
  40static inline void
  41__pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte, unsigned long addr)
  42{
  43        pgtable_pte_page_dtor(pte);
  44
  45#ifndef CONFIG_ARM_LPAE
  46        /*
  47         * With the classic ARM MMU, a pte page has two corresponding pmd
  48         * entries, each covering 1MB.
  49         */
  50        addr = (addr & PMD_MASK) + SZ_1M;
  51        __tlb_adjust_range(tlb, addr - PAGE_SIZE, 2 * PAGE_SIZE);
  52#endif
  53
  54        tlb_remove_table(tlb, pte);
  55}
  56
  57static inline void
  58__pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmdp, unsigned long addr)
  59{
  60#ifdef CONFIG_ARM_LPAE
  61        struct page *page = virt_to_page(pmdp);
  62
  63        tlb_remove_table(tlb, page);
  64#endif
  65}
  66
  67#endif /* CONFIG_MMU */
  68#endif
  69