1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21#include "qemu/osdep.h"
22#include "cpu.h"
23#include "exec/exec-all.h"
24
25static unsigned int tlb_decode_size(unsigned int f)
26{
27 static const unsigned int sizes[] = {
28 1 * 1024, 4 * 1024, 16 * 1024, 64 * 1024, 256 * 1024,
29 1 * 1024 * 1024, 4 * 1024 * 1024, 16 * 1024 * 1024
30 };
31 assert(f < ARRAY_SIZE(sizes));
32 return sizes[f];
33}
34
35static void mmu_flush_idx(CPUMBState *env, unsigned int idx)
36{
37 CPUState *cs = CPU(mb_env_get_cpu(env));
38 struct microblaze_mmu *mmu = &env->mmu;
39 unsigned int tlb_size;
40 uint32_t tlb_tag, end, t;
41
42 t = mmu->rams[RAM_TAG][idx];
43 if (!(t & TLB_VALID))
44 return;
45
46 tlb_tag = t & TLB_EPN_MASK;
47 tlb_size = tlb_decode_size((t & TLB_PAGESZ_MASK) >> 7);
48 end = tlb_tag + tlb_size;
49
50 while (tlb_tag < end) {
51 tlb_flush_page(cs, tlb_tag);
52 tlb_tag += TARGET_PAGE_SIZE;
53 }
54}
55
56static void mmu_change_pid(CPUMBState *env, unsigned int newpid)
57{
58 struct microblaze_mmu *mmu = &env->mmu;
59 unsigned int i;
60 uint32_t t;
61
62 if (newpid & ~0xff)
63 qemu_log_mask(LOG_GUEST_ERROR, "Illegal rpid=%x\n", newpid);
64
65 for (i = 0; i < ARRAY_SIZE(mmu->rams[RAM_TAG]); i++) {
66
67 t = mmu->rams[RAM_TAG][i];
68 if (t & TLB_VALID) {
69 if (mmu->tids[i] && ((mmu->regs[MMU_R_PID] & 0xff) == mmu->tids[i]))
70 mmu_flush_idx(env, i);
71 }
72 }
73}
74
75
76unsigned int mmu_translate(struct microblaze_mmu *mmu,
77 struct microblaze_mmu_lookup *lu,
78 target_ulong vaddr, int rw, int mmu_idx)
79{
80 unsigned int i, hit = 0;
81 unsigned int tlb_ex = 0, tlb_wr = 0, tlb_zsel;
82 uint64_t tlb_tag, tlb_rpn, mask;
83 uint32_t tlb_size, t0;
84
85 lu->err = ERR_MISS;
86 for (i = 0; i < ARRAY_SIZE(mmu->rams[RAM_TAG]); i++) {
87 uint64_t t, d;
88
89
90 t = mmu->rams[RAM_TAG][i];
91 if (t & TLB_VALID) {
92 tlb_size = tlb_decode_size((t & TLB_PAGESZ_MASK) >> 7);
93 if (tlb_size < TARGET_PAGE_SIZE) {
94 qemu_log_mask(LOG_UNIMP, "%d pages not supported\n", tlb_size);
95 abort();
96 }
97
98 mask = ~((uint64_t)tlb_size - 1);
99 tlb_tag = t & TLB_EPN_MASK;
100 if ((vaddr & mask) != (tlb_tag & mask)) {
101 continue;
102 }
103 if (mmu->tids[i]
104 && ((mmu->regs[MMU_R_PID] & 0xff) != mmu->tids[i])) {
105 continue;
106 }
107
108
109 d = mmu->rams[RAM_DATA][i];
110 tlb_ex = d & TLB_EX;
111 tlb_wr = d & TLB_WR;
112
113
114 tlb_zsel = (d >> 4) & 0xf;
115 t0 = mmu->regs[MMU_R_ZPR] >> (30 - (tlb_zsel * 2));
116 t0 &= 0x3;
117
118 if (tlb_zsel > mmu->c_mmu_zones) {
119 qemu_log_mask(LOG_GUEST_ERROR,
120 "tlb zone select out of range! %d\n", tlb_zsel);
121 t0 = 1;
122 }
123
124 if (mmu->c_mmu == 1) {
125 t0 = 1;
126 }
127
128 switch (t0) {
129 case 0:
130 if (mmu_idx == MMU_USER_IDX)
131 continue;
132 break;
133 case 2:
134 if (mmu_idx != MMU_USER_IDX) {
135 tlb_ex = 1;
136 tlb_wr = 1;
137 }
138 break;
139 case 3:
140 tlb_ex = 1;
141 tlb_wr = 1;
142 break;
143 default: break;
144 }
145
146 lu->err = ERR_PROT;
147 lu->prot = PAGE_READ;
148 if (tlb_wr)
149 lu->prot |= PAGE_WRITE;
150 else if (rw == 1)
151 goto done;
152 if (tlb_ex)
153 lu->prot |=PAGE_EXEC;
154 else if (rw == 2) {
155 goto done;
156 }
157
158 tlb_rpn = d & TLB_RPN_MASK;
159
160 lu->vaddr = tlb_tag;
161 lu->paddr = tlb_rpn & mmu->c_addr_mask;
162 lu->size = tlb_size;
163 lu->err = ERR_HIT;
164 lu->idx = i;
165 hit = 1;
166 goto done;
167 }
168 }
169done:
170 qemu_log_mask(CPU_LOG_MMU,
171 "MMU vaddr=%" PRIx64 " rw=%d tlb_wr=%d tlb_ex=%d hit=%d\n",
172 vaddr, rw, tlb_wr, tlb_ex, hit);
173 return hit;
174}
175
176
177uint32_t mmu_read(CPUMBState *env, bool ext, uint32_t rn)
178{
179 unsigned int i;
180 uint32_t r = 0;
181
182 if (env->mmu.c_mmu < 2 || !env->mmu.c_mmu_tlb_access) {
183 qemu_log_mask(LOG_GUEST_ERROR, "MMU access on MMU-less system\n");
184 return 0;
185 }
186 if (ext && rn != MMU_R_TLBLO) {
187 qemu_log_mask(LOG_GUEST_ERROR, "Extended access only to TLBLO.\n");
188 return 0;
189 }
190
191 switch (rn) {
192
193 case MMU_R_TLBLO:
194 case MMU_R_TLBHI:
195 if (!(env->mmu.c_mmu_tlb_access & 1)) {
196 qemu_log_mask(LOG_GUEST_ERROR,
197 "Invalid access to MMU reg %d\n", rn);
198 return 0;
199 }
200
201 i = env->mmu.regs[MMU_R_TLBX] & 0xff;
202 r = extract64(env->mmu.rams[rn & 1][i], ext * 32, 32);
203 if (rn == MMU_R_TLBHI)
204 env->mmu.regs[MMU_R_PID] = env->mmu.tids[i];
205 break;
206 case MMU_R_PID:
207 case MMU_R_ZPR:
208 if (!(env->mmu.c_mmu_tlb_access & 1)) {
209 qemu_log_mask(LOG_GUEST_ERROR,
210 "Invalid access to MMU reg %d\n", rn);
211 return 0;
212 }
213 r = env->mmu.regs[rn];
214 break;
215 case MMU_R_TLBX:
216 r = env->mmu.regs[rn];
217 break;
218 case MMU_R_TLBSX:
219 qemu_log_mask(LOG_GUEST_ERROR, "TLBSX is write-only.\n");
220 break;
221 default:
222 qemu_log_mask(LOG_GUEST_ERROR, "Invalid MMU register %d.\n", rn);
223 break;
224 }
225 qemu_log_mask(CPU_LOG_MMU, "%s rn=%d=%x\n", __func__, rn, r);
226 return r;
227}
228
229void mmu_write(CPUMBState *env, bool ext, uint32_t rn, uint32_t v)
230{
231 MicroBlazeCPU *cpu = mb_env_get_cpu(env);
232 uint64_t tmp64;
233 unsigned int i;
234 qemu_log_mask(CPU_LOG_MMU,
235 "%s rn=%d=%x old=%x\n", __func__, rn, v, env->mmu.regs[rn]);
236
237 if (env->mmu.c_mmu < 2 || !env->mmu.c_mmu_tlb_access) {
238 qemu_log_mask(LOG_GUEST_ERROR, "MMU access on MMU-less system\n");
239 return;
240 }
241 if (ext && rn != MMU_R_TLBLO) {
242 qemu_log_mask(LOG_GUEST_ERROR, "Extended access only to TLBLO.\n");
243 return;
244 }
245
246 switch (rn) {
247
248 case MMU_R_TLBLO:
249 case MMU_R_TLBHI:
250 i = env->mmu.regs[MMU_R_TLBX] & 0xff;
251 if (rn == MMU_R_TLBHI) {
252 if (i < 3 && !(v & TLB_VALID) && qemu_loglevel_mask(~0))
253 qemu_log_mask(LOG_GUEST_ERROR,
254 "invalidating index %x at pc=%" PRIx64 "\n",
255 i, env->sregs[SR_PC]);
256 env->mmu.tids[i] = env->mmu.regs[MMU_R_PID] & 0xff;
257 mmu_flush_idx(env, i);
258 }
259 tmp64 = env->mmu.rams[rn & 1][i];
260 env->mmu.rams[rn & 1][i] = deposit64(tmp64, ext * 32, 32, v);
261 break;
262 case MMU_R_ZPR:
263 if (env->mmu.c_mmu_tlb_access <= 1) {
264 qemu_log_mask(LOG_GUEST_ERROR,
265 "Invalid access to MMU reg %d\n", rn);
266 return;
267 }
268
269
270
271 if (v != env->mmu.regs[rn]) {
272 tlb_flush(CPU(cpu));
273 }
274 env->mmu.regs[rn] = v;
275 break;
276 case MMU_R_PID:
277 if (env->mmu.c_mmu_tlb_access <= 1) {
278 qemu_log_mask(LOG_GUEST_ERROR,
279 "Invalid access to MMU reg %d\n", rn);
280 return;
281 }
282
283 if (v != env->mmu.regs[rn]) {
284 mmu_change_pid(env, v);
285 env->mmu.regs[rn] = v;
286 }
287 break;
288 case MMU_R_TLBX:
289
290 env->mmu.regs[rn] = deposit32(env->mmu.regs[rn], 0, 31, v);
291 break;
292 case MMU_R_TLBSX:
293 {
294 struct microblaze_mmu_lookup lu;
295 int hit;
296
297 if (env->mmu.c_mmu_tlb_access <= 1) {
298 qemu_log_mask(LOG_GUEST_ERROR,
299 "Invalid access to MMU reg %d\n", rn);
300 return;
301 }
302
303 hit = mmu_translate(&env->mmu, &lu,
304 v & TLB_EPN_MASK, 0, cpu_mmu_index(env, false));
305 if (hit) {
306 env->mmu.regs[MMU_R_TLBX] = lu.idx;
307 } else {
308 env->mmu.regs[MMU_R_TLBX] |= R_TBLX_MISS_MASK;
309 }
310 break;
311 }
312 default:
313 qemu_log_mask(LOG_GUEST_ERROR, "Invalid MMU register %d.\n", rn);
314 break;
315 }
316}
317
318void mmu_init(struct microblaze_mmu *mmu)
319{
320 int i;
321 for (i = 0; i < ARRAY_SIZE(mmu->regs); i++) {
322 mmu->regs[i] = 0;
323 }
324}
325