Blame view

kernel/linux-rt-4.4.41/arch/powerpc/include/asm/mmu-40x.h 1.9 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
  #ifndef _ASM_POWERPC_MMU_40X_H_
  #define _ASM_POWERPC_MMU_40X_H_
  
  /*
   * PPC40x support
   */
  
  #define PPC40X_TLB_SIZE 64
  
  /*
   * TLB entries are defined by a "high" tag portion and a "low" data
   * portion.  On all architectures, the data portion is 32-bits.
   *
   * TLB entries are managed entirely under software control by reading,
   * writing, and searchoing using the 4xx-specific tlbre, tlbwr, and tlbsx
   * instructions.
   */
  
  #define	TLB_LO          1
  #define	TLB_HI          0
  
  #define	TLB_DATA        TLB_LO
  #define	TLB_TAG         TLB_HI
  
  /* Tag portion */
  
  #define TLB_EPN_MASK    0xFFFFFC00      /* Effective Page Number */
  #define TLB_PAGESZ_MASK 0x00000380
  #define TLB_PAGESZ(x)   (((x) & 0x7) << 7)
  #define   PAGESZ_1K		0
  #define   PAGESZ_4K             1
  #define   PAGESZ_16K            2
  #define   PAGESZ_64K            3
  #define   PAGESZ_256K           4
  #define   PAGESZ_1M             5
  #define   PAGESZ_4M             6
  #define   PAGESZ_16M            7
  #define TLB_VALID       0x00000040      /* Entry is valid */
  
  /* Data portion */
  
  #define TLB_RPN_MASK    0xFFFFFC00      /* Real Page Number */
  #define TLB_PERM_MASK   0x00000300
  #define TLB_EX          0x00000200      /* Instruction execution allowed */
  #define TLB_WR          0x00000100      /* Writes permitted */
  #define TLB_ZSEL_MASK   0x000000F0
  #define TLB_ZSEL(x)     (((x) & 0xF) << 4)
  #define TLB_ATTR_MASK   0x0000000F
  #define TLB_W           0x00000008      /* Caching is write-through */
  #define TLB_I           0x00000004      /* Caching is inhibited */
  #define TLB_M           0x00000002      /* Memory is coherent */
  #define TLB_G           0x00000001      /* Memory is guarded from prefetch */
  
  #ifndef __ASSEMBLY__
  
  typedef struct {
  	unsigned int	id;
  	unsigned int	active;
  	unsigned long	vdso_base;
  } mm_context_t;
  
  #endif /* !__ASSEMBLY__ */
  
  #define mmu_virtual_psize	MMU_PAGE_4K
  #define mmu_linear_psize	MMU_PAGE_256M
  
  #endif /* _ASM_POWERPC_MMU_40X_H_ */