1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21#include "qemu/osdep.h"
22#include "cpu.h"
23#include "exec/exec-all.h"
24#include "qemu/host-utils.h"
25#include "exec/log.h"
26
27#ifndef CONFIG_USER_ONLY
28static bool mb_cpu_access_is_secure(MicroBlazeCPU *cpu,
29 MMUAccessType access_type)
30{
31 if (access_type == MMU_INST_FETCH) {
32 return !cpu->ns_axi_ip;
33 } else {
34 return !cpu->ns_axi_dp;
35 }
36}
37
38bool mb_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
39 MMUAccessType access_type, int mmu_idx,
40 bool probe, uintptr_t retaddr)
41{
42 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
43 CPUMBState *env = &cpu->env;
44 MicroBlazeMMULookup lu;
45 unsigned int hit;
46 int prot;
47 MemTxAttrs attrs = {};
48
49 attrs.secure = mb_cpu_access_is_secure(cpu, access_type);
50
51 if (mmu_idx == MMU_NOMMU_IDX) {
52
53 address &= TARGET_PAGE_MASK;
54 prot = PAGE_BITS;
55 tlb_set_page_with_attrs(cs, address, address, attrs, prot, mmu_idx,
56 TARGET_PAGE_SIZE);
57 return true;
58 }
59
60 hit = mmu_translate(cpu, &lu, address, access_type, mmu_idx);
61 if (likely(hit)) {
62 uint32_t vaddr = address & TARGET_PAGE_MASK;
63 uint32_t paddr = lu.paddr + vaddr - lu.vaddr;
64
65 qemu_log_mask(CPU_LOG_MMU, "MMU map mmu=%d v=%x p=%x prot=%x\n",
66 mmu_idx, vaddr, paddr, lu.prot);
67 tlb_set_page_with_attrs(cs, vaddr, paddr, attrs, lu.prot, mmu_idx,
68 TARGET_PAGE_SIZE);
69 return true;
70 }
71
72
73 if (probe) {
74 return false;
75 }
76
77 qemu_log_mask(CPU_LOG_MMU, "mmu=%d miss v=%" VADDR_PRIx "\n",
78 mmu_idx, address);
79
80 env->ear = address;
81 switch (lu.err) {
82 case ERR_PROT:
83 env->esr = access_type == MMU_INST_FETCH ? 17 : 16;
84 env->esr |= (access_type == MMU_DATA_STORE) << 10;
85 break;
86 case ERR_MISS:
87 env->esr = access_type == MMU_INST_FETCH ? 19 : 18;
88 env->esr |= (access_type == MMU_DATA_STORE) << 10;
89 break;
90 default:
91 abort();
92 }
93
94 if (cs->exception_index == EXCP_MMU) {
95 cpu_abort(cs, "recursive faults\n");
96 }
97
98
99 cs->exception_index = EXCP_MMU;
100 cpu_loop_exit_restore(cs, retaddr);
101}
102
103void mb_cpu_do_interrupt(CPUState *cs)
104{
105 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
106 CPUMBState *env = &cpu->env;
107 uint32_t t, msr = mb_cpu_read_msr(env);
108 bool set_esr;
109
110
111 assert((env->iflags & (D_FLAG | IMM_FLAG)) != (D_FLAG | IMM_FLAG));
112
113 assert((env->iflags & (D_FLAG | BIMM_FLAG)) != BIMM_FLAG);
114
115 assert(!(env->iflags & (DRTI_FLAG | DRTE_FLAG | DRTB_FLAG)));
116
117 switch (cs->exception_index) {
118 case EXCP_HW_EXCP:
119 if (!(cpu->cfg.pvr_regs[0] & PVR0_USE_EXC_MASK)) {
120 qemu_log_mask(LOG_GUEST_ERROR,
121 "Exception raised on system without exceptions!\n");
122 return;
123 }
124
125 qemu_log_mask(CPU_LOG_INT,
126 "INT: HWE at pc=%08x msr=%08x iflags=%x\n",
127 env->pc, msr, env->iflags);
128
129
130 set_esr = true;
131 env->esr &= ~D_FLAG;
132 if (env->iflags & D_FLAG) {
133 env->esr |= D_FLAG;
134 env->btr = env->btarget;
135 }
136
137
138 msr |= MSR_EIP;
139 env->regs[17] = env->pc + 4;
140 env->pc = cpu->cfg.base_vectors + 0x20;
141 break;
142
143 case EXCP_MMU:
144 qemu_log_mask(CPU_LOG_INT,
145 "INT: MMU at pc=%08x msr=%08x "
146 "ear=%" PRIx64 " iflags=%x\n",
147 env->pc, msr, env->ear, env->iflags);
148
149
150 set_esr = true;
151 env->esr &= ~D_FLAG;
152 if (env->iflags & D_FLAG) {
153 env->esr |= D_FLAG;
154 env->btr = env->btarget;
155
156 env->regs[17] = env->pc - (env->iflags & BIMM_FLAG ? 8 : 4);
157 } else if (env->iflags & IMM_FLAG) {
158
159 env->regs[17] = env->pc - 4;
160 } else {
161 env->regs[17] = env->pc;
162 }
163
164
165 msr |= MSR_EIP;
166 env->pc = cpu->cfg.base_vectors + 0x20;
167 break;
168
169 case EXCP_IRQ:
170 assert(!(msr & (MSR_EIP | MSR_BIP)));
171 assert(msr & MSR_IE);
172 assert(!(env->iflags & (D_FLAG | IMM_FLAG)));
173
174 qemu_log_mask(CPU_LOG_INT,
175 "INT: DEV at pc=%08x msr=%08x iflags=%x\n",
176 env->pc, msr, env->iflags);
177 set_esr = false;
178
179
180 msr &= ~MSR_IE;
181 env->regs[14] = env->pc;
182 env->pc = cpu->cfg.base_vectors + 0x10;
183 break;
184
185 case EXCP_HW_BREAK:
186 assert(!(env->iflags & (D_FLAG | IMM_FLAG)));
187
188 qemu_log_mask(CPU_LOG_INT,
189 "INT: BRK at pc=%08x msr=%08x iflags=%x\n",
190 env->pc, msr, env->iflags);
191 set_esr = false;
192
193
194 msr |= MSR_BIP;
195 env->regs[16] = env->pc;
196 env->pc = cpu->cfg.base_vectors + 0x18;
197 break;
198
199 default:
200 cpu_abort(cs, "unhandled exception type=%d\n", cs->exception_index);
201
202 }
203
204
205 t = (msr & (MSR_VM | MSR_UM)) << 1;
206 msr &= ~(MSR_VMS | MSR_UMS | MSR_VM | MSR_UM);
207 msr |= t;
208 mb_cpu_write_msr(env, msr);
209
210 env->res_addr = RES_ADDR_NONE;
211 env->iflags = 0;
212
213 if (!set_esr) {
214 qemu_log_mask(CPU_LOG_INT,
215 " to pc=%08x msr=%08x\n", env->pc, msr);
216 } else if (env->esr & D_FLAG) {
217 qemu_log_mask(CPU_LOG_INT,
218 " to pc=%08x msr=%08x esr=%04x btr=%08x\n",
219 env->pc, msr, env->esr, env->btr);
220 } else {
221 qemu_log_mask(CPU_LOG_INT,
222 " to pc=%08x msr=%08x esr=%04x\n",
223 env->pc, msr, env->esr);
224 }
225}
226
227hwaddr mb_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
228 MemTxAttrs *attrs)
229{
230 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
231 CPUMBState *env = &cpu->env;
232 target_ulong vaddr, paddr = 0;
233 MicroBlazeMMULookup lu;
234 int mmu_idx = cpu_mmu_index(env, false);
235 unsigned int hit;
236
237
238 *attrs = (MemTxAttrs) {};
239 attrs->secure = mb_cpu_access_is_secure(cpu, MMU_DATA_LOAD);
240
241 if (mmu_idx != MMU_NOMMU_IDX) {
242 hit = mmu_translate(cpu, &lu, addr, 0, 0);
243 if (hit) {
244 vaddr = addr & TARGET_PAGE_MASK;
245 paddr = lu.paddr + vaddr - lu.vaddr;
246 } else
247 paddr = 0;
248 } else
249 paddr = addr & TARGET_PAGE_MASK;
250
251 return paddr;
252}
253
254bool mb_cpu_exec_interrupt(CPUState *cs, int interrupt_request)
255{
256 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
257 CPUMBState *env = &cpu->env;
258
259 if ((interrupt_request & CPU_INTERRUPT_HARD)
260 && (env->msr & MSR_IE)
261 && !(env->msr & (MSR_EIP | MSR_BIP))
262 && !(env->iflags & (D_FLAG | IMM_FLAG))) {
263 cs->exception_index = EXCP_IRQ;
264 mb_cpu_do_interrupt(cs);
265 return true;
266 }
267 return false;
268}
269
270#endif
271
272void mb_cpu_do_unaligned_access(CPUState *cs, vaddr addr,
273 MMUAccessType access_type,
274 int mmu_idx, uintptr_t retaddr)
275{
276 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
277 uint32_t esr, iflags;
278
279
280 cpu_restore_state(cs, retaddr, true);
281 iflags = cpu->env.iflags;
282
283 qemu_log_mask(CPU_LOG_INT,
284 "Unaligned access addr=" TARGET_FMT_lx " pc=%x iflags=%x\n",
285 (target_ulong)addr, cpu->env.pc, iflags);
286
287 esr = ESR_EC_UNALIGNED_DATA;
288 if (likely(iflags & ESR_ESS_FLAG)) {
289 esr |= iflags & ESR_ESS_MASK;
290 } else {
291 qemu_log_mask(LOG_UNIMP, "Unaligned access without ESR_ESS_FLAG\n");
292 }
293
294 cpu->env.ear = addr;
295 cpu->env.esr = esr;
296 cs->exception_index = EXCP_HW_EXCP;
297 cpu_loop_exit(cs);
298}
299