Blame view

kernel/linux-rt-4.4.41/arch/hexagon/include/asm/tlbflush.h 2.08 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
  /*
   * TLB flush support for Hexagon
   *
   * Copyright (c) 2010-2011, The Linux Foundation. All rights reserved.
   *
   * This program is free software; you can redistribute it and/or modify
   * it under the terms of the GNU General Public License version 2 and
   * only version 2 as published by the Free Software Foundation.
   *
   * This program is distributed in the hope that it will be useful,
   * but WITHOUT ANY WARRANTY; without even the implied warranty of
   * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
   * GNU General Public License for more details.
   *
   * You should have received a copy of the GNU General Public License
   * along with this program; if not, write to the Free Software
   * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
   * 02110-1301, USA.
   */
  
  #ifndef _ASM_TLBFLUSH_H
  #define _ASM_TLBFLUSH_H
  
  #include <linux/mm.h>
  #include <asm/processor.h>
  
  /*
   * TLB flushing -- in "SMP", these routines get defined to be the
   * ones from smp.c, else they are some local flavors.
   */
  
  /*
   * These functions are commonly macros, but in the interests of
   * VM vs. native implementation and code size, we simply declare
   * the function prototypes here.
   */
  extern void tlb_flush_all(void);
  extern void flush_tlb_mm(struct mm_struct *mm);
  extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr);
  extern void flush_tlb_range(struct vm_area_struct *vma,
  				unsigned long start, unsigned long end);
  extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
  extern void flush_tlb_one(unsigned long);
  
  /*
   * "This is called in munmap when we have freed up some page-table pages.
   * We don't need to do anything here..."
   *
   * The VM kernel doesn't walk page tables, and they are passed to the VMM
   * by logical address. There doesn't seem to be any possibility that they
   * could be referenced by the VM kernel based on a stale mapping, since
   * they would only be located by consulting the mm structure, and they
   * will have been purged from that structure by the munmap.  Seems like
   * a noop on HVM as well.
   */
  #define flush_tlb_pgtables(mm, start, end)
  
  #endif