6b13f685e
김민수
BSP 최초 추가
|
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
|
#include <linux/init.h>
#include <linux/mm.h>
#include <linux/highmem.h>
#include <asm/pgtable.h>
#include <asm/tlbflush.h>
#include <asm/cacheflush.h>
#include "mm.h"
#define minicache_pgprot __pgprot(L_PTE_PRESENT | L_PTE_YOUNG | \
L_PTE_MT_MINICACHE)
static DEFINE_RAW_SPINLOCK(minicache_lock);
static void __naked
mc_copy_user_page(void *from, void *to)
{
asm volatile(
"stmfd sp!, {r4, r5, lr}
\
mov lr, %2
\
pld [r0, #0]
\
pld [r0, #32]
\
pld [r1, #0]
\
pld [r1, #32]
\
1: pld [r0, #64]
\
pld [r0, #96]
\
pld [r1, #64]
\
pld [r1, #96]
\
2: ldrd r2, [r0], #8
\
ldrd r4, [r0], #8
\
mov ip, r1
\
strd r2, [r1], #8
\
ldrd r2, [r0], #8
\
strd r4, [r1], #8
\
ldrd r4, [r0], #8
\
strd r2, [r1], #8
\
strd r4, [r1], #8
\
mcr p15, 0, ip, c7, c10, 1 @ clean D line
\
ldrd r2, [r0], #8
\
mcr p15, 0, ip, c7, c6, 1 @ invalidate D line
\
ldrd r4, [r0], #8
\
mov ip, r1
\
strd r2, [r1], #8
\
ldrd r2, [r0], #8
\
strd r4, [r1], #8
\
ldrd r4, [r0], #8
\
strd r2, [r1], #8
\
strd r4, [r1], #8
\
mcr p15, 0, ip, c7, c10, 1 @ clean D line
\
subs lr, lr, #1
\
mcr p15, 0, ip, c7, c6, 1 @ invalidate D line
\
bgt 1b
\
beq 2b
\
ldmfd sp!, {r4, r5, pc} "
:
: "r" (from), "r" (to), "I" (PAGE_SIZE / 64 - 1));
}
void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
unsigned long vaddr, struct vm_area_struct *vma)
{
void *kto = kmap_atomic(to);
if (!test_and_set_bit(PG_dcache_clean, &from->flags))
__flush_dcache_page(page_mapping(from), from);
raw_spin_lock(&minicache_lock);
set_top_pte(COPYPAGE_MINICACHE, mk_pte(from, minicache_pgprot));
mc_copy_user_page((void *)COPYPAGE_MINICACHE, kto);
raw_spin_unlock(&minicache_lock);
kunmap_atomic(kto);
}
void
xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
{
void *ptr, *kaddr = kmap_atomic(page);
asm volatile(
"mov r1, %2
\
mov r2, #0
\
mov r3, #0
\
1: mov ip, %0
\
strd r2, [%0], #8
\
strd r2, [%0], #8
\
strd r2, [%0], #8
\
strd r2, [%0], #8
\
mcr p15, 0, ip, c7, c10, 1 @ clean D line
\
subs r1, r1, #1
\
mcr p15, 0, ip, c7, c6, 1 @ invalidate D line
\
bne 1b"
: "=r" (ptr)
: "0" (kaddr), "I" (PAGE_SIZE / 32)
: "r1", "r2", "r3", "ip");
kunmap_atomic(kaddr);
}
struct cpu_user_fns xscale_mc_user_fns __initdata = {
.cpu_clear_user_highpage = xscale_mc_clear_user_highpage,
.cpu_copy_user_highpage = xscale_mc_copy_user_highpage,
};
|