1 |
|
/* |
2 |
|
* sheepshaver_glue.cpp - Glue Kheperix CPU to SheepShaver CPU engine interface |
3 |
|
* |
4 |
< |
* SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig |
4 |
> |
* SheepShaver (C) 1997-2004 Christian Bauer and Marc Hellwig |
5 |
|
* |
6 |
|
* This program is free software; you can redistribute it and/or modify |
7 |
|
* it under the terms of the GNU General Public License as published by |
30 |
|
#include "sigsegv.h" |
31 |
|
#include "cpu/ppc/ppc-cpu.hpp" |
32 |
|
#include "cpu/ppc/ppc-operations.hpp" |
33 |
+ |
#include "cpu/ppc/ppc-instructions.hpp" |
34 |
+ |
#include "thunks.h" |
35 |
|
|
36 |
|
// Used for NativeOp trampolines |
37 |
|
#include "video.h" |
38 |
|
#include "name_registry.h" |
39 |
|
#include "serial.h" |
40 |
|
#include "ether.h" |
41 |
+ |
#include "timer.h" |
42 |
|
|
43 |
|
#include <stdio.h> |
44 |
+ |
#include <stdlib.h> |
45 |
|
|
46 |
|
#if ENABLE_MON |
47 |
|
#include "mon.h" |
75 |
|
#endif |
76 |
|
} |
77 |
|
|
78 |
+ |
// From main_*.cpp |
79 |
+ |
extern uintptr SignalStackBase(); |
80 |
+ |
|
81 |
+ |
// From rsrc_patches.cpp |
82 |
+ |
extern "C" void check_load_invoc(uint32 type, int16 id, uint32 h); |
83 |
+ |
|
84 |
+ |
// PowerPC EmulOp to exit from emulation looop |
85 |
+ |
const uint32 POWERPC_EXEC_RETURN = POWERPC_EMUL_OP | 1; |
86 |
+ |
|
87 |
|
// Enable multicore (main/interrupts) cpu emulation? |
88 |
|
#define MULTICORE_CPU (ASYNC_IRQ ? 1 : 0) |
89 |
|
|
90 |
+ |
// Enable interrupt routine safety checks? |
91 |
+ |
#define SAFE_INTERRUPT_PPC 1 |
92 |
+ |
|
93 |
|
// Enable Execute68k() safety checks? |
94 |
|
#define SAFE_EXEC_68K 1 |
95 |
|
|
102 |
|
// Interrupts in native mode? |
103 |
|
#define INTERRUPTS_IN_NATIVE_MODE 1 |
104 |
|
|
105 |
+ |
// Enable native EMUL_OPs to be run without a mode switch |
106 |
+ |
#define ENABLE_NATIVE_EMUL_OP 1 |
107 |
+ |
|
108 |
|
// Pointer to Kernel Data |
109 |
|
static KernelData * const kernel_data = (KernelData *)KERNEL_DATA_BASE; |
110 |
|
|
111 |
|
// SIGSEGV handler |
112 |
|
static sigsegv_return_t sigsegv_handler(sigsegv_address_t, sigsegv_address_t); |
113 |
|
|
114 |
+ |
#if PPC_ENABLE_JIT && PPC_REENTRANT_JIT |
115 |
+ |
// Special trampolines for EmulOp and NativeOp |
116 |
+ |
static uint8 *emul_op_trampoline; |
117 |
+ |
static uint8 *native_op_trampoline; |
118 |
+ |
#endif |
119 |
+ |
|
120 |
+ |
// JIT Compiler enabled? |
121 |
+ |
static inline bool enable_jit_p() |
122 |
+ |
{ |
123 |
+ |
return PrefsFindBool("jit"); |
124 |
+ |
} |
125 |
+ |
|
126 |
|
|
127 |
|
/** |
128 |
|
* PowerPC emulator glue with special 'sheep' opcodes |
129 |
|
**/ |
130 |
|
|
131 |
+ |
enum { |
132 |
+ |
PPC_I(SHEEP) = PPC_I(MAX), |
133 |
+ |
PPC_I(SHEEP_MAX) |
134 |
+ |
}; |
135 |
+ |
|
136 |
|
class sheepshaver_cpu |
137 |
|
: public powerpc_cpu |
138 |
|
{ |
139 |
|
void init_decoder(); |
140 |
|
void execute_sheep(uint32 opcode); |
141 |
|
|
142 |
+ |
// Filter out EMUL_OP routines that only call native code |
143 |
+ |
bool filter_execute_emul_op(uint32 emul_op); |
144 |
+ |
|
145 |
+ |
// "Native" EMUL_OP routines |
146 |
+ |
void execute_emul_op_microseconds(); |
147 |
+ |
void execute_emul_op_idle_time_1(); |
148 |
+ |
void execute_emul_op_idle_time_2(); |
149 |
+ |
|
150 |
+ |
// CPU context to preserve on interrupt |
151 |
+ |
class interrupt_context { |
152 |
+ |
uint32 gpr[32]; |
153 |
+ |
uint32 pc; |
154 |
+ |
uint32 lr; |
155 |
+ |
uint32 ctr; |
156 |
+ |
uint32 cr; |
157 |
+ |
uint32 xer; |
158 |
+ |
sheepshaver_cpu *cpu; |
159 |
+ |
const char *where; |
160 |
+ |
public: |
161 |
+ |
interrupt_context(sheepshaver_cpu *_cpu, const char *_where); |
162 |
+ |
~interrupt_context(); |
163 |
+ |
}; |
164 |
+ |
|
165 |
|
public: |
166 |
|
|
167 |
|
// Constructor |
168 |
|
sheepshaver_cpu(); |
169 |
|
|
170 |
< |
// Condition Register accessors |
170 |
> |
// CR & XER accessors |
171 |
|
uint32 get_cr() const { return cr().get(); } |
172 |
|
void set_cr(uint32 v) { cr().set(v); } |
173 |
+ |
uint32 get_xer() const { return xer().get(); } |
174 |
+ |
void set_xer(uint32 v) { xer().set(v); } |
175 |
+ |
|
176 |
+ |
// Execute NATIVE_OP routine |
177 |
+ |
void execute_native_op(uint32 native_op); |
178 |
|
|
179 |
< |
// Execution loop |
180 |
< |
void execute(uint32 entry, bool enable_cache = false); |
179 |
> |
// Execute EMUL_OP routine |
180 |
> |
void execute_emul_op(uint32 emul_op); |
181 |
|
|
182 |
|
// Execute 68k routine |
183 |
|
void execute_68k(uint32 entry, M68kRegisters *r); |
188 |
|
// Execute MacOS/PPC code |
189 |
|
uint32 execute_macos_code(uint32 tvect, int nargs, uint32 const *args); |
190 |
|
|
191 |
+ |
// Compile one instruction |
192 |
+ |
virtual int compile1(codegen_context_t & cg_context); |
193 |
+ |
|
194 |
|
// Resource manager thunk |
195 |
|
void get_resource(uint32 old_get_resource); |
196 |
|
|
198 |
|
void interrupt(uint32 entry); |
199 |
|
void handle_interrupt(); |
200 |
|
|
134 |
– |
// Lazy memory allocator (one item at a time) |
135 |
– |
void *operator new(size_t size) |
136 |
– |
{ return allocator_helper< sheepshaver_cpu, lazy_allocator >::allocate(); } |
137 |
– |
void operator delete(void *p) |
138 |
– |
{ allocator_helper< sheepshaver_cpu, lazy_allocator >::deallocate(p); } |
139 |
– |
// FIXME: really make surre array allocation fail at link time? |
140 |
– |
void *operator new[](size_t); |
141 |
– |
void operator delete[](void *p); |
142 |
– |
|
201 |
|
// Make sure the SIGSEGV handler can access CPU registers |
202 |
|
friend sigsegv_return_t sigsegv_handler(sigsegv_address_t, sigsegv_address_t); |
203 |
|
}; |
204 |
|
|
205 |
< |
lazy_allocator< sheepshaver_cpu > allocator_helper< sheepshaver_cpu, lazy_allocator >::allocator; |
205 |
> |
// Memory allocator returning areas aligned on 16-byte boundaries |
206 |
> |
void *operator new(size_t size) |
207 |
> |
{ |
208 |
> |
void *p; |
209 |
> |
|
210 |
> |
#if defined(HAVE_POSIX_MEMALIGN) |
211 |
> |
if (posix_memalign(&p, 16, size) != 0) |
212 |
> |
throw std::bad_alloc(); |
213 |
> |
#elif defined(HAVE_MEMALIGN) |
214 |
> |
p = memalign(16, size); |
215 |
> |
#elif defined(HAVE_VALLOC) |
216 |
> |
p = valloc(size); // page-aligned! |
217 |
> |
#else |
218 |
> |
/* XXX: handle padding ourselves */ |
219 |
> |
p = malloc(size); |
220 |
> |
#endif |
221 |
> |
|
222 |
> |
return p; |
223 |
> |
} |
224 |
> |
|
225 |
> |
void operator delete(void *p) |
226 |
> |
{ |
227 |
> |
#if defined(HAVE_MEMALIGN) || defined(HAVE_VALLOC) |
228 |
> |
#if defined(__GLIBC__) |
229 |
> |
// this is known to work only with GNU libc |
230 |
> |
free(p); |
231 |
> |
#endif |
232 |
> |
#else |
233 |
> |
free(p); |
234 |
> |
#endif |
235 |
> |
} |
236 |
|
|
237 |
|
sheepshaver_cpu::sheepshaver_cpu() |
238 |
< |
: powerpc_cpu() |
238 |
> |
: powerpc_cpu(enable_jit_p()) |
239 |
|
{ |
240 |
|
init_decoder(); |
241 |
|
} |
242 |
|
|
243 |
|
void sheepshaver_cpu::init_decoder() |
244 |
|
{ |
157 |
– |
#ifndef PPC_NO_STATIC_II_INDEX_TABLE |
158 |
– |
static bool initialized = false; |
159 |
– |
if (initialized) |
160 |
– |
return; |
161 |
– |
initialized = true; |
162 |
– |
#endif |
163 |
– |
|
245 |
|
static const instr_info_t sheep_ii_table[] = { |
246 |
|
{ "sheep", |
247 |
|
(execute_pmf)&sheepshaver_cpu::execute_sheep, |
248 |
|
NULL, |
249 |
+ |
PPC_I(SHEEP), |
250 |
|
D_form, 6, 0, CFLOW_JUMP | CFLOW_TRAP |
251 |
|
} |
252 |
|
}; |
260 |
|
} |
261 |
|
} |
262 |
|
|
181 |
– |
// Forward declaration for native opcode handler |
182 |
– |
static void NativeOp(int selector); |
183 |
– |
|
263 |
|
/* NativeOp instruction format: |
264 |
< |
+------------+--------------------------+--+----------+------------+ |
265 |
< |
| 6 | |FN| OP | 2 | |
266 |
< |
+------------+--------------------------+--+----------+------------+ |
267 |
< |
0 5 |6 19 20 21 25 26 31 |
264 |
> |
+------------+-------------------------+--+-----------+------------+ |
265 |
> |
| 6 | |FN| OP | 2 | |
266 |
> |
+------------+-------------------------+--+-----------+------------+ |
267 |
> |
0 5 |6 18 19 20 25 26 31 |
268 |
|
*/ |
269 |
|
|
270 |
< |
typedef bit_field< 20, 20 > FN_field; |
271 |
< |
typedef bit_field< 21, 25 > NATIVE_OP_field; |
270 |
> |
typedef bit_field< 19, 19 > FN_field; |
271 |
> |
typedef bit_field< 20, 25 > NATIVE_OP_field; |
272 |
|
typedef bit_field< 26, 31 > EMUL_OP_field; |
273 |
|
|
274 |
+ |
// "Native" EMUL_OP routines |
275 |
+ |
#define GPR_A(REG) gpr(16 + (REG)) |
276 |
+ |
#define GPR_D(REG) gpr( 8 + (REG)) |
277 |
+ |
|
278 |
+ |
void sheepshaver_cpu::execute_emul_op_microseconds() |
279 |
+ |
{ |
280 |
+ |
Microseconds(GPR_A(0), GPR_D(0)); |
281 |
+ |
} |
282 |
+ |
|
283 |
+ |
void sheepshaver_cpu::execute_emul_op_idle_time_1() |
284 |
+ |
{ |
285 |
+ |
// Sleep if no events pending |
286 |
+ |
if (ReadMacInt32(0x14c) == 0) |
287 |
+ |
Delay_usec(16667); |
288 |
+ |
GPR_A(0) = ReadMacInt32(0x2b6); |
289 |
+ |
} |
290 |
+ |
|
291 |
+ |
void sheepshaver_cpu::execute_emul_op_idle_time_2() |
292 |
+ |
{ |
293 |
+ |
// Sleep if no events pending |
294 |
+ |
if (ReadMacInt32(0x14c) == 0) |
295 |
+ |
Delay_usec(16667); |
296 |
+ |
GPR_D(0) = (uint32)-2; |
297 |
+ |
} |
298 |
+ |
|
299 |
+ |
// Filter out EMUL_OP routines that only call native code |
300 |
+ |
bool sheepshaver_cpu::filter_execute_emul_op(uint32 emul_op) |
301 |
+ |
{ |
302 |
+ |
switch (emul_op) { |
303 |
+ |
case OP_MICROSECONDS: |
304 |
+ |
execute_emul_op_microseconds(); |
305 |
+ |
return true; |
306 |
+ |
case OP_IDLE_TIME: |
307 |
+ |
execute_emul_op_idle_time_1(); |
308 |
+ |
return true; |
309 |
+ |
case OP_IDLE_TIME_2: |
310 |
+ |
execute_emul_op_idle_time_2(); |
311 |
+ |
return true; |
312 |
+ |
} |
313 |
+ |
return false; |
314 |
+ |
} |
315 |
+ |
|
316 |
+ |
// Execute EMUL_OP routine |
317 |
+ |
void sheepshaver_cpu::execute_emul_op(uint32 emul_op) |
318 |
+ |
{ |
319 |
+ |
#if ENABLE_NATIVE_EMUL_OP |
320 |
+ |
// First, filter out EMUL_OPs that can be executed without a mode switch |
321 |
+ |
if (filter_execute_emul_op(emul_op)) |
322 |
+ |
return; |
323 |
+ |
#endif |
324 |
+ |
|
325 |
+ |
M68kRegisters r68; |
326 |
+ |
WriteMacInt32(XLM_68K_R25, gpr(25)); |
327 |
+ |
WriteMacInt32(XLM_RUN_MODE, MODE_EMUL_OP); |
328 |
+ |
for (int i = 0; i < 8; i++) |
329 |
+ |
r68.d[i] = gpr(8 + i); |
330 |
+ |
for (int i = 0; i < 7; i++) |
331 |
+ |
r68.a[i] = gpr(16 + i); |
332 |
+ |
r68.a[7] = gpr(1); |
333 |
+ |
uint32 saved_cr = get_cr() & CR_field<2>::mask(); |
334 |
+ |
uint32 saved_xer = get_xer(); |
335 |
+ |
EmulOp(&r68, gpr(24), emul_op); |
336 |
+ |
set_cr(saved_cr); |
337 |
+ |
set_xer(saved_xer); |
338 |
+ |
for (int i = 0; i < 8; i++) |
339 |
+ |
gpr(8 + i) = r68.d[i]; |
340 |
+ |
for (int i = 0; i < 7; i++) |
341 |
+ |
gpr(16 + i) = r68.a[i]; |
342 |
+ |
gpr(1) = r68.a[7]; |
343 |
+ |
WriteMacInt32(XLM_RUN_MODE, MODE_68K); |
344 |
+ |
} |
345 |
+ |
|
346 |
|
// Execute SheepShaver instruction |
347 |
|
void sheepshaver_cpu::execute_sheep(uint32 opcode) |
348 |
|
{ |
359 |
|
break; |
360 |
|
|
361 |
|
case 2: // EXEC_NATIVE |
362 |
< |
NativeOp(NATIVE_OP_field::extract(opcode)); |
362 |
> |
execute_native_op(NATIVE_OP_field::extract(opcode)); |
363 |
|
if (FN_field::test(opcode)) |
364 |
|
pc() = lr(); |
365 |
|
else |
366 |
|
pc() += 4; |
367 |
|
break; |
368 |
|
|
369 |
< |
default: { // EMUL_OP |
370 |
< |
M68kRegisters r68; |
220 |
< |
WriteMacInt32(XLM_68K_R25, gpr(25)); |
221 |
< |
WriteMacInt32(XLM_RUN_MODE, MODE_EMUL_OP); |
222 |
< |
for (int i = 0; i < 8; i++) |
223 |
< |
r68.d[i] = gpr(8 + i); |
224 |
< |
for (int i = 0; i < 7; i++) |
225 |
< |
r68.a[i] = gpr(16 + i); |
226 |
< |
r68.a[7] = gpr(1); |
227 |
< |
EmulOp(&r68, gpr(24), EMUL_OP_field::extract(opcode) - 3); |
228 |
< |
for (int i = 0; i < 8; i++) |
229 |
< |
gpr(8 + i) = r68.d[i]; |
230 |
< |
for (int i = 0; i < 7; i++) |
231 |
< |
gpr(16 + i) = r68.a[i]; |
232 |
< |
gpr(1) = r68.a[7]; |
233 |
< |
WriteMacInt32(XLM_RUN_MODE, MODE_68K); |
369 |
> |
default: // EMUL_OP |
370 |
> |
execute_emul_op(EMUL_OP_field::extract(opcode) - 3); |
371 |
|
pc() += 4; |
372 |
|
break; |
373 |
|
} |
374 |
+ |
} |
375 |
+ |
|
376 |
+ |
// Compile one instruction |
377 |
+ |
int sheepshaver_cpu::compile1(codegen_context_t & cg_context) |
378 |
+ |
{ |
379 |
+ |
#if PPC_ENABLE_JIT |
380 |
+ |
const instr_info_t *ii = cg_context.instr_info; |
381 |
+ |
if (ii->mnemo != PPC_I(SHEEP)) |
382 |
+ |
return COMPILE_FAILURE; |
383 |
+ |
|
384 |
+ |
int status = COMPILE_FAILURE; |
385 |
+ |
powerpc_dyngen & dg = cg_context.codegen; |
386 |
+ |
uint32 opcode = cg_context.opcode; |
387 |
+ |
|
388 |
+ |
switch (opcode & 0x3f) { |
389 |
+ |
case 0: // EMUL_RETURN |
390 |
+ |
dg.gen_invoke(QuitEmulator); |
391 |
+ |
status = COMPILE_CODE_OK; |
392 |
+ |
break; |
393 |
+ |
|
394 |
+ |
case 1: // EXEC_RETURN |
395 |
+ |
dg.gen_spcflags_set(SPCFLAG_CPU_EXEC_RETURN); |
396 |
+ |
// Don't check for pending interrupts, we do know we have to |
397 |
+ |
// get out of this block ASAP |
398 |
+ |
dg.gen_exec_return(); |
399 |
+ |
status = COMPILE_EPILOGUE_OK; |
400 |
+ |
break; |
401 |
+ |
|
402 |
+ |
case 2: { // EXEC_NATIVE |
403 |
+ |
uint32 selector = NATIVE_OP_field::extract(opcode); |
404 |
+ |
switch (selector) { |
405 |
+ |
#if !PPC_REENTRANT_JIT |
406 |
+ |
// Filter out functions that may invoke Execute68k() or |
407 |
+ |
// CallMacOS(), this would break reentrancy as they could |
408 |
+ |
// invalidate the translation cache and even overwrite |
409 |
+ |
// continuation code when we are done with them. |
410 |
+ |
case NATIVE_PATCH_NAME_REGISTRY: |
411 |
+ |
dg.gen_invoke(DoPatchNameRegistry); |
412 |
+ |
status = COMPILE_CODE_OK; |
413 |
+ |
break; |
414 |
+ |
case NATIVE_VIDEO_INSTALL_ACCEL: |
415 |
+ |
dg.gen_invoke(VideoInstallAccel); |
416 |
+ |
status = COMPILE_CODE_OK; |
417 |
+ |
break; |
418 |
+ |
case NATIVE_VIDEO_VBL: |
419 |
+ |
dg.gen_invoke(VideoVBL); |
420 |
+ |
status = COMPILE_CODE_OK; |
421 |
+ |
break; |
422 |
+ |
case NATIVE_GET_RESOURCE: |
423 |
+ |
case NATIVE_GET_1_RESOURCE: |
424 |
+ |
case NATIVE_GET_IND_RESOURCE: |
425 |
+ |
case NATIVE_GET_1_IND_RESOURCE: |
426 |
+ |
case NATIVE_R_GET_RESOURCE: { |
427 |
+ |
static const uint32 get_resource_ptr[] = { |
428 |
+ |
XLM_GET_RESOURCE, |
429 |
+ |
XLM_GET_1_RESOURCE, |
430 |
+ |
XLM_GET_IND_RESOURCE, |
431 |
+ |
XLM_GET_1_IND_RESOURCE, |
432 |
+ |
XLM_R_GET_RESOURCE |
433 |
+ |
}; |
434 |
+ |
uint32 old_get_resource = ReadMacInt32(get_resource_ptr[selector - NATIVE_GET_RESOURCE]); |
435 |
+ |
typedef void (*func_t)(dyngen_cpu_base, uint32); |
436 |
+ |
func_t func = (func_t)nv_mem_fun(&sheepshaver_cpu::get_resource).ptr(); |
437 |
+ |
dg.gen_invoke_CPU_im(func, old_get_resource); |
438 |
+ |
status = COMPILE_CODE_OK; |
439 |
+ |
break; |
440 |
+ |
} |
441 |
+ |
case NATIVE_CHECK_LOAD_INVOC: |
442 |
+ |
dg.gen_load_T0_GPR(3); |
443 |
+ |
dg.gen_load_T1_GPR(4); |
444 |
+ |
dg.gen_se_16_32_T1(); |
445 |
+ |
dg.gen_load_T2_GPR(5); |
446 |
+ |
dg.gen_invoke_T0_T1_T2((void (*)(uint32, uint32, uint32))check_load_invoc); |
447 |
+ |
status = COMPILE_CODE_OK; |
448 |
+ |
break; |
449 |
+ |
#endif |
450 |
+ |
case NATIVE_DISABLE_INTERRUPT: |
451 |
+ |
dg.gen_invoke(DisableInterrupt); |
452 |
+ |
status = COMPILE_CODE_OK; |
453 |
+ |
break; |
454 |
+ |
case NATIVE_ENABLE_INTERRUPT: |
455 |
+ |
dg.gen_invoke(EnableInterrupt); |
456 |
+ |
status = COMPILE_CODE_OK; |
457 |
+ |
break; |
458 |
+ |
case NATIVE_BITBLT: |
459 |
+ |
dg.gen_load_T0_GPR(3); |
460 |
+ |
dg.gen_invoke_T0((void (*)(uint32))NQD_bitblt); |
461 |
+ |
status = COMPILE_CODE_OK; |
462 |
+ |
break; |
463 |
+ |
case NATIVE_INVRECT: |
464 |
+ |
dg.gen_load_T0_GPR(3); |
465 |
+ |
dg.gen_invoke_T0((void (*)(uint32))NQD_invrect); |
466 |
+ |
status = COMPILE_CODE_OK; |
467 |
+ |
break; |
468 |
+ |
case NATIVE_FILLRECT: |
469 |
+ |
dg.gen_load_T0_GPR(3); |
470 |
+ |
dg.gen_invoke_T0((void (*)(uint32))NQD_fillrect); |
471 |
+ |
status = COMPILE_CODE_OK; |
472 |
+ |
break; |
473 |
+ |
} |
474 |
+ |
// Could we fully translate this NativeOp? |
475 |
+ |
if (FN_field::test(opcode)) { |
476 |
+ |
if (status != COMPILE_FAILURE) { |
477 |
+ |
dg.gen_load_A0_LR(); |
478 |
+ |
dg.gen_set_PC_A0(); |
479 |
+ |
} |
480 |
+ |
cg_context.done_compile = true; |
481 |
+ |
break; |
482 |
+ |
} |
483 |
+ |
else if (status != COMPILE_FAILURE) { |
484 |
+ |
cg_context.done_compile = false; |
485 |
+ |
break; |
486 |
+ |
} |
487 |
+ |
#if PPC_REENTRANT_JIT |
488 |
+ |
// Try to execute NativeOp trampoline |
489 |
+ |
dg.gen_set_PC_im(cg_context.pc + 4); |
490 |
+ |
dg.gen_mov_32_T0_im(selector); |
491 |
+ |
dg.gen_jmp(native_op_trampoline); |
492 |
+ |
cg_context.done_compile = true; |
493 |
+ |
status = COMPILE_EPILOGUE_OK; |
494 |
+ |
break; |
495 |
+ |
#endif |
496 |
+ |
// Invoke NativeOp handler |
497 |
+ |
typedef void (*func_t)(dyngen_cpu_base, uint32); |
498 |
+ |
func_t func = (func_t)nv_mem_fun(&sheepshaver_cpu::execute_native_op).ptr(); |
499 |
+ |
dg.gen_invoke_CPU_im(func, selector); |
500 |
+ |
cg_context.done_compile = false; |
501 |
+ |
status = COMPILE_CODE_OK; |
502 |
+ |
break; |
503 |
+ |
} |
504 |
+ |
|
505 |
+ |
default: { // EMUL_OP |
506 |
+ |
uint32 emul_op = EMUL_OP_field::extract(opcode) - 3; |
507 |
+ |
#if ENABLE_NATIVE_EMUL_OP |
508 |
+ |
typedef void (*emul_op_func_t)(dyngen_cpu_base); |
509 |
+ |
emul_op_func_t emul_op_func = 0; |
510 |
+ |
switch (emul_op) { |
511 |
+ |
case OP_MICROSECONDS: |
512 |
+ |
emul_op_func = (emul_op_func_t)nv_mem_fun(&sheepshaver_cpu::execute_emul_op_microseconds).ptr(); |
513 |
+ |
break; |
514 |
+ |
case OP_IDLE_TIME: |
515 |
+ |
emul_op_func = (emul_op_func_t)nv_mem_fun(&sheepshaver_cpu::execute_emul_op_idle_time_1).ptr(); |
516 |
+ |
break; |
517 |
+ |
case OP_IDLE_TIME_2: |
518 |
+ |
emul_op_func = (emul_op_func_t)nv_mem_fun(&sheepshaver_cpu::execute_emul_op_idle_time_2).ptr(); |
519 |
+ |
break; |
520 |
+ |
} |
521 |
+ |
if (emul_op_func) { |
522 |
+ |
dg.gen_invoke_CPU(emul_op_func); |
523 |
+ |
cg_context.done_compile = false; |
524 |
+ |
status = COMPILE_CODE_OK; |
525 |
+ |
break; |
526 |
+ |
} |
527 |
+ |
#endif |
528 |
+ |
#if PPC_REENTRANT_JIT |
529 |
+ |
// Try to execute EmulOp trampoline |
530 |
+ |
dg.gen_set_PC_im(cg_context.pc + 4); |
531 |
+ |
dg.gen_mov_32_T0_im(emul_op); |
532 |
+ |
dg.gen_jmp(emul_op_trampoline); |
533 |
+ |
cg_context.done_compile = true; |
534 |
+ |
status = COMPILE_EPILOGUE_OK; |
535 |
+ |
break; |
536 |
+ |
#endif |
537 |
+ |
// Invoke EmulOp handler |
538 |
+ |
typedef void (*func_t)(dyngen_cpu_base, uint32); |
539 |
+ |
func_t func = (func_t)nv_mem_fun(&sheepshaver_cpu::execute_emul_op).ptr(); |
540 |
+ |
dg.gen_invoke_CPU_im(func, emul_op); |
541 |
+ |
cg_context.done_compile = false; |
542 |
+ |
status = COMPILE_CODE_OK; |
543 |
+ |
break; |
544 |
|
} |
545 |
+ |
} |
546 |
+ |
return status; |
547 |
+ |
#endif |
548 |
+ |
return COMPILE_FAILURE; |
549 |
+ |
} |
550 |
+ |
|
551 |
+ |
// CPU context to preserve on interrupt |
552 |
+ |
sheepshaver_cpu::interrupt_context::interrupt_context(sheepshaver_cpu *_cpu, const char *_where) |
553 |
+ |
{ |
554 |
+ |
#if SAFE_INTERRUPT_PPC >= 2 |
555 |
+ |
cpu = _cpu; |
556 |
+ |
where = _where; |
557 |
+ |
|
558 |
+ |
// Save interrupt context |
559 |
+ |
memcpy(&gpr[0], &cpu->gpr(0), sizeof(gpr)); |
560 |
+ |
pc = cpu->pc(); |
561 |
+ |
lr = cpu->lr(); |
562 |
+ |
ctr = cpu->ctr(); |
563 |
+ |
cr = cpu->get_cr(); |
564 |
+ |
xer = cpu->get_xer(); |
565 |
+ |
#endif |
566 |
|
} |
567 |
|
|
568 |
< |
// Execution loop |
241 |
< |
void sheepshaver_cpu::execute(uint32 entry, bool enable_cache) |
568 |
> |
sheepshaver_cpu::interrupt_context::~interrupt_context() |
569 |
|
{ |
570 |
< |
powerpc_cpu::execute(entry, enable_cache); |
570 |
> |
#if SAFE_INTERRUPT_PPC >= 2 |
571 |
> |
// Check whether CPU context was preserved by interrupt |
572 |
> |
if (memcmp(&gpr[0], &cpu->gpr(0), sizeof(gpr)) != 0) { |
573 |
> |
printf("FATAL: %s: interrupt clobbers registers\n", where); |
574 |
> |
for (int i = 0; i < 32; i++) |
575 |
> |
if (gpr[i] != cpu->gpr(i)) |
576 |
> |
printf(" r%d: %08x -> %08x\n", i, gpr[i], cpu->gpr(i)); |
577 |
> |
} |
578 |
> |
if (pc != cpu->pc()) |
579 |
> |
printf("FATAL: %s: interrupt clobbers PC\n", where); |
580 |
> |
if (lr != cpu->lr()) |
581 |
> |
printf("FATAL: %s: interrupt clobbers LR\n", where); |
582 |
> |
if (ctr != cpu->ctr()) |
583 |
> |
printf("FATAL: %s: interrupt clobbers CTR\n", where); |
584 |
> |
if (cr != cpu->get_cr()) |
585 |
> |
printf("FATAL: %s: interrupt clobbers CR\n", where); |
586 |
> |
if (xer != cpu->get_xer()) |
587 |
> |
printf("FATAL: %s: interrupt clobbers XER\n", where); |
588 |
> |
#endif |
589 |
|
} |
590 |
|
|
591 |
|
// Handle MacOS interrupt |
596 |
|
const clock_t interrupt_start = clock(); |
597 |
|
#endif |
598 |
|
|
599 |
+ |
#if SAFE_INTERRUPT_PPC |
600 |
+ |
static int depth = 0; |
601 |
+ |
if (depth != 0) |
602 |
+ |
printf("FATAL: sheepshaver_cpu::interrupt() called more than once: %d\n", depth); |
603 |
+ |
depth++; |
604 |
+ |
#endif |
605 |
+ |
|
606 |
|
#if !MULTICORE_CPU |
607 |
|
// Save program counters and branch registers |
608 |
|
uint32 saved_pc = pc(); |
612 |
|
#endif |
613 |
|
|
614 |
|
// Initialize stack pointer to SheepShaver alternate stack base |
615 |
< |
gpr(1) = SheepStack1Base - 64; |
615 |
> |
gpr(1) = SignalStackBase() - 64; |
616 |
|
|
617 |
|
// Build trampoline to return from interrupt |
618 |
< |
uint32 trampoline[] = { htonl(POWERPC_EMUL_OP | 1) }; |
618 |
> |
SheepVar32 trampoline = POWERPC_EXEC_RETURN; |
619 |
|
|
620 |
|
// Prepare registers for nanokernel interrupt routine |
621 |
|
kernel_data->v[0x004 >> 2] = htonl(gpr(1)); |
634 |
|
gpr(1) = KernelDataAddr; |
635 |
|
gpr(7) = ntohl(kernel_data->v[0x660 >> 2]); |
636 |
|
gpr(8) = 0; |
637 |
< |
gpr(10) = (uint32)trampoline; |
638 |
< |
gpr(12) = (uint32)trampoline; |
637 |
> |
gpr(10) = trampoline.addr(); |
638 |
> |
gpr(12) = trampoline.addr(); |
639 |
|
gpr(13) = get_cr(); |
640 |
|
|
641 |
|
// rlwimi. r7,r7,8,0,0 |
660 |
|
#if EMUL_TIME_STATS |
661 |
|
interrupt_time += (clock() - interrupt_start); |
662 |
|
#endif |
663 |
+ |
|
664 |
+ |
#if SAFE_INTERRUPT_PPC |
665 |
+ |
depth--; |
666 |
+ |
#endif |
667 |
|
} |
668 |
|
|
669 |
|
// Execute 68k routine |
776 |
|
uint32 saved_ctr= ctr(); |
777 |
|
|
778 |
|
// Build trampoline with EXEC_RETURN |
779 |
< |
uint32 trampoline[] = { htonl(POWERPC_EMUL_OP | 1) }; |
780 |
< |
lr() = (uint32)trampoline; |
779 |
> |
SheepVar32 trampoline = POWERPC_EXEC_RETURN; |
780 |
> |
lr() = trampoline.addr(); |
781 |
|
|
782 |
|
gpr(1) -= 64; // Create stack frame |
783 |
|
uint32 proc = ReadMacInt32(tvect); // Get routine address |
821 |
|
// Save branch registers |
822 |
|
uint32 saved_lr = lr(); |
823 |
|
|
824 |
< |
const uint32 trampoline[] = { htonl(POWERPC_EMUL_OP | 1) }; |
825 |
< |
lr() = (uint32)trampoline; |
824 |
> |
SheepVar32 trampoline = POWERPC_EXEC_RETURN; |
825 |
> |
WriteMacInt32(trampoline.addr(), POWERPC_EXEC_RETURN); |
826 |
> |
lr() = trampoline.addr(); |
827 |
|
|
828 |
|
execute(entry); |
829 |
|
|
832 |
|
} |
833 |
|
|
834 |
|
// Resource Manager thunk |
478 |
– |
extern "C" void check_load_invoc(uint32 type, int16 id, uint32 h); |
479 |
– |
|
835 |
|
inline void sheepshaver_cpu::get_resource(uint32 old_get_resource) |
836 |
|
{ |
837 |
|
uint32 type = gpr(3); |
941 |
|
else if (pc == ROM_BASE + 0x4a10a0 && (cpu->gpr(20) == 0xf3012002 || cpu->gpr(20) == 0xf3012000)) |
942 |
|
return SIGSEGV_RETURN_SKIP_INSTRUCTION; |
943 |
|
|
944 |
+ |
// Ignore writes to the zero page |
945 |
+ |
else if ((uint32)(addr - SheepMem::ZeroPage()) < (uint32)SheepMem::PageSize()) |
946 |
+ |
return SIGSEGV_RETURN_SKIP_INSTRUCTION; |
947 |
+ |
|
948 |
|
// Ignore all other faults, if requested |
949 |
|
if (PrefsFindBool("ignoresegv")) |
950 |
|
return SIGSEGV_RETURN_SKIP_INSTRUCTION; |
970 |
|
// Initialize main CPU emulator |
971 |
|
main_cpu = new sheepshaver_cpu(); |
972 |
|
main_cpu->set_register(powerpc_registers::GPR(3), any_register((uint32)ROM_BASE + 0x30d000)); |
973 |
+ |
main_cpu->set_register(powerpc_registers::GPR(4), any_register(KernelDataAddr + 0x1000)); |
974 |
|
WriteMacInt32(XLM_RUN_MODE, MODE_68K); |
975 |
|
|
976 |
|
#if MULTICORE_CPU |
1028 |
|
#endif |
1029 |
|
} |
1030 |
|
|
1031 |
+ |
#if PPC_ENABLE_JIT && PPC_REENTRANT_JIT |
1032 |
+ |
// Initialize EmulOp trampolines |
1033 |
+ |
void init_emul_op_trampolines(basic_dyngen & dg) |
1034 |
+ |
{ |
1035 |
+ |
typedef void (*func_t)(dyngen_cpu_base, uint32); |
1036 |
+ |
func_t func; |
1037 |
+ |
|
1038 |
+ |
// EmulOp |
1039 |
+ |
emul_op_trampoline = dg.gen_start(); |
1040 |
+ |
func = (func_t)nv_mem_fun(&sheepshaver_cpu::execute_emul_op).ptr(); |
1041 |
+ |
dg.gen_invoke_CPU_T0(func); |
1042 |
+ |
dg.gen_exec_return(); |
1043 |
+ |
dg.gen_end(); |
1044 |
+ |
|
1045 |
+ |
// NativeOp |
1046 |
+ |
native_op_trampoline = dg.gen_start(); |
1047 |
+ |
func = (func_t)nv_mem_fun(&sheepshaver_cpu::execute_native_op).ptr(); |
1048 |
+ |
dg.gen_invoke_CPU_T0(func); |
1049 |
+ |
dg.gen_exec_return(); |
1050 |
+ |
dg.gen_end(); |
1051 |
+ |
|
1052 |
+ |
D(bug("EmulOp trampoline: %p\n", emul_op_trampoline)); |
1053 |
+ |
D(bug("NativeOp trampoline: %p\n", native_op_trampoline)); |
1054 |
+ |
} |
1055 |
+ |
#endif |
1056 |
+ |
|
1057 |
|
/* |
1058 |
|
* Emulation loop |
1059 |
|
*/ |
1061 |
|
void emul_ppc(uint32 entry) |
1062 |
|
{ |
1063 |
|
current_cpu = main_cpu; |
1064 |
< |
#if DEBUG |
1064 |
> |
#if 0 |
1065 |
|
current_cpu->start_log(); |
1066 |
|
#endif |
1067 |
|
// start emulation loop and enable code translation or caching |
1068 |
< |
current_cpu->execute(entry, true); |
1068 |
> |
current_cpu->execute(entry); |
1069 |
|
} |
1070 |
|
|
1071 |
|
/* |
1100 |
|
if (InterruptFlags == 0) |
1101 |
|
return; |
1102 |
|
|
1103 |
+ |
// Current interrupt nest level |
1104 |
+ |
static int interrupt_depth = 0; |
1105 |
+ |
++interrupt_depth; |
1106 |
+ |
|
1107 |
|
// Disable MacOS stack sniffer |
1108 |
|
WriteMacInt32(0x110, 0); |
1109 |
|
|
1120 |
|
case MODE_NATIVE: |
1121 |
|
// 68k emulator inactive, in nanokernel? |
1122 |
|
assert(current_cpu == main_cpu); |
1123 |
< |
if (gpr(1) != KernelDataAddr) { |
1123 |
> |
if (gpr(1) != KernelDataAddr && interrupt_depth == 1) { |
1124 |
> |
interrupt_context ctx(this, "PowerPC mode"); |
1125 |
> |
|
1126 |
|
// Prepare for 68k interrupt level 1 |
1127 |
|
WriteMacInt16(tswap32(kernel_data->v[0x67c >> 2]), 1); |
1128 |
|
WriteMacInt32(tswap32(kernel_data->v[0x658 >> 2]) + 0xdc, |
1145 |
|
case MODE_EMUL_OP: |
1146 |
|
// 68k emulator active, within EMUL_OP routine, execute 68k interrupt routine directly when interrupt level is 0 |
1147 |
|
if ((ReadMacInt32(XLM_68K_R25) & 7) == 0) { |
1148 |
+ |
interrupt_context ctx(this, "68k mode"); |
1149 |
|
#if 1 |
1150 |
|
// Execute full 68k interrupt routine |
1151 |
|
M68kRegisters r; |
1167 |
|
if (InterruptFlags & INTFLAG_VIA) { |
1168 |
|
ClearInterruptFlag(INTFLAG_VIA); |
1169 |
|
ADBInterrupt(); |
1170 |
< |
ExecutePPC(VideoVBL); |
1170 |
> |
ExecuteNative(NATIVE_VIDEO_VBL); |
1171 |
|
} |
1172 |
|
} |
1173 |
|
#endif |
1175 |
|
break; |
1176 |
|
#endif |
1177 |
|
} |
785 |
– |
} |
786 |
– |
|
787 |
– |
/* |
788 |
– |
* Execute NATIVE_OP opcode (called by PowerPC emulator) |
789 |
– |
*/ |
790 |
– |
|
791 |
– |
#define POWERPC_NATIVE_OP_INIT(LR, OP) \ |
792 |
– |
tswap32(POWERPC_EMUL_OP | ((LR) << 11) | (((uint32)OP) << 6) | 2) |
1178 |
|
|
1179 |
< |
// FIXME: Make sure 32-bit relocations are used |
1180 |
< |
const uint32 NativeOpTable[NATIVE_OP_MAX] = { |
1181 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_PATCH_NAME_REGISTRY), |
797 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_VIDEO_INSTALL_ACCEL), |
798 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_VIDEO_VBL), |
799 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_VIDEO_DO_DRIVER_IO), |
800 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_IRQ), |
801 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_INIT), |
802 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_TERM), |
803 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_OPEN), |
804 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_CLOSE), |
805 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_WPUT), |
806 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_ETHER_RSRV), |
807 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_NOTHING), |
808 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_OPEN), |
809 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_PRIME_IN), |
810 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_PRIME_OUT), |
811 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_CONTROL), |
812 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_STATUS), |
813 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_SERIAL_CLOSE), |
814 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_GET_RESOURCE), |
815 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_GET_1_RESOURCE), |
816 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_GET_IND_RESOURCE), |
817 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_GET_1_IND_RESOURCE), |
818 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_R_GET_RESOURCE), |
819 |
< |
POWERPC_NATIVE_OP_INIT(0, NATIVE_DISABLE_INTERRUPT), |
820 |
< |
POWERPC_NATIVE_OP_INIT(0, NATIVE_ENABLE_INTERRUPT), |
821 |
< |
POWERPC_NATIVE_OP_INIT(1, NATIVE_MAKE_EXECUTABLE), |
822 |
< |
}; |
1179 |
> |
// We are done with this interrupt |
1180 |
> |
--interrupt_depth; |
1181 |
> |
} |
1182 |
|
|
1183 |
|
static void get_resource(void); |
1184 |
|
static void get_1_resource(void); |
1186 |
|
static void get_1_ind_resource(void); |
1187 |
|
static void r_get_resource(void); |
1188 |
|
|
1189 |
< |
#define GPR(REG) current_cpu->gpr(REG) |
1190 |
< |
|
832 |
< |
static void NativeOp(int selector) |
1189 |
> |
// Execute NATIVE_OP routine |
1190 |
> |
void sheepshaver_cpu::execute_native_op(uint32 selector) |
1191 |
|
{ |
1192 |
|
#if EMUL_TIME_STATS |
1193 |
|
native_exec_count++; |
1205 |
|
VideoVBL(); |
1206 |
|
break; |
1207 |
|
case NATIVE_VIDEO_DO_DRIVER_IO: |
1208 |
< |
GPR(3) = (int32)(int16)VideoDoDriverIO((void *)GPR(3), (void *)GPR(4), |
1209 |
< |
(void *)GPR(5), GPR(6), GPR(7)); |
1208 |
> |
gpr(3) = (int32)(int16)VideoDoDriverIO((void *)gpr(3), (void *)gpr(4), |
1209 |
> |
(void *)gpr(5), gpr(6), gpr(7)); |
1210 |
|
break; |
1211 |
|
#ifdef WORDS_BIGENDIAN |
1212 |
|
case NATIVE_ETHER_IRQ: |
1213 |
|
EtherIRQ(); |
1214 |
|
break; |
1215 |
|
case NATIVE_ETHER_INIT: |
1216 |
< |
GPR(3) = InitStreamModule((void *)GPR(3)); |
1216 |
> |
gpr(3) = InitStreamModule((void *)gpr(3)); |
1217 |
|
break; |
1218 |
|
case NATIVE_ETHER_TERM: |
1219 |
|
TerminateStreamModule(); |
1220 |
|
break; |
1221 |
|
case NATIVE_ETHER_OPEN: |
1222 |
< |
GPR(3) = ether_open((queue_t *)GPR(3), (void *)GPR(4), GPR(5), GPR(6), (void*)GPR(7)); |
1222 |
> |
gpr(3) = ether_open((queue_t *)gpr(3), (void *)gpr(4), gpr(5), gpr(6), (void*)gpr(7)); |
1223 |
|
break; |
1224 |
|
case NATIVE_ETHER_CLOSE: |
1225 |
< |
GPR(3) = ether_close((queue_t *)GPR(3), GPR(4), (void *)GPR(5)); |
1225 |
> |
gpr(3) = ether_close((queue_t *)gpr(3), gpr(4), (void *)gpr(5)); |
1226 |
|
break; |
1227 |
|
case NATIVE_ETHER_WPUT: |
1228 |
< |
GPR(3) = ether_wput((queue_t *)GPR(3), (mblk_t *)GPR(4)); |
1228 |
> |
gpr(3) = ether_wput((queue_t *)gpr(3), (mblk_t *)gpr(4)); |
1229 |
|
break; |
1230 |
|
case NATIVE_ETHER_RSRV: |
1231 |
< |
GPR(3) = ether_rsrv((queue_t *)GPR(3)); |
1231 |
> |
gpr(3) = ether_rsrv((queue_t *)gpr(3)); |
1232 |
|
break; |
1233 |
|
#else |
1234 |
|
case NATIVE_ETHER_INIT: |
1235 |
|
// FIXME: needs more complicated thunks |
1236 |
< |
GPR(3) = false; |
1236 |
> |
gpr(3) = false; |
1237 |
|
break; |
1238 |
|
#endif |
1239 |
+ |
case NATIVE_SYNC_HOOK: |
1240 |
+ |
gpr(3) = NQD_sync_hook(gpr(3)); |
1241 |
+ |
break; |
1242 |
+ |
case NATIVE_BITBLT_HOOK: |
1243 |
+ |
gpr(3) = NQD_bitblt_hook(gpr(3)); |
1244 |
+ |
break; |
1245 |
+ |
case NATIVE_BITBLT: |
1246 |
+ |
NQD_bitblt(gpr(3)); |
1247 |
+ |
break; |
1248 |
+ |
case NATIVE_FILLRECT_HOOK: |
1249 |
+ |
gpr(3) = NQD_fillrect_hook(gpr(3)); |
1250 |
+ |
break; |
1251 |
+ |
case NATIVE_INVRECT: |
1252 |
+ |
NQD_invrect(gpr(3)); |
1253 |
+ |
break; |
1254 |
+ |
case NATIVE_FILLRECT: |
1255 |
+ |
NQD_fillrect(gpr(3)); |
1256 |
+ |
break; |
1257 |
|
case NATIVE_SERIAL_NOTHING: |
1258 |
|
case NATIVE_SERIAL_OPEN: |
1259 |
|
case NATIVE_SERIAL_PRIME_IN: |
1271 |
|
SerialStatus, |
1272 |
|
SerialClose |
1273 |
|
}; |
1274 |
< |
GPR(3) = serial_callbacks[selector - NATIVE_SERIAL_NOTHING](GPR(3), GPR(4)); |
1274 |
> |
gpr(3) = serial_callbacks[selector - NATIVE_SERIAL_NOTHING](gpr(3), gpr(4)); |
1275 |
|
break; |
1276 |
|
} |
1277 |
|
case NATIVE_GET_RESOURCE: |
1281 |
|
case NATIVE_R_GET_RESOURCE: { |
1282 |
|
typedef void (*GetResourceCallback)(void); |
1283 |
|
static const GetResourceCallback get_resource_callbacks[] = { |
1284 |
< |
get_resource, |
1285 |
< |
get_1_resource, |
1286 |
< |
get_ind_resource, |
1287 |
< |
get_1_ind_resource, |
1288 |
< |
r_get_resource |
1284 |
> |
::get_resource, |
1285 |
> |
::get_1_resource, |
1286 |
> |
::get_ind_resource, |
1287 |
> |
::get_1_ind_resource, |
1288 |
> |
::r_get_resource |
1289 |
|
}; |
1290 |
|
get_resource_callbacks[selector - NATIVE_GET_RESOURCE](); |
1291 |
|
break; |
1297 |
|
EnableInterrupt(); |
1298 |
|
break; |
1299 |
|
case NATIVE_MAKE_EXECUTABLE: |
1300 |
< |
MakeExecutable(0, (void *)GPR(4), GPR(5)); |
1300 |
> |
MakeExecutable(0, (void *)gpr(4), gpr(5)); |
1301 |
> |
break; |
1302 |
> |
case NATIVE_CHECK_LOAD_INVOC: |
1303 |
> |
check_load_invoc(gpr(3), gpr(4), gpr(5)); |
1304 |
|
break; |
1305 |
|
default: |
1306 |
|
printf("FATAL: NATIVE_OP called with bogus selector %d\n", selector); |
1314 |
|
} |
1315 |
|
|
1316 |
|
/* |
938 |
– |
* Execute native subroutine (LR must contain return address) |
939 |
– |
*/ |
940 |
– |
|
941 |
– |
void ExecuteNative(int selector) |
942 |
– |
{ |
943 |
– |
uint32 tvect[2]; |
944 |
– |
tvect[0] = tswap32(POWERPC_NATIVE_OP_FUNC(selector)); |
945 |
– |
tvect[1] = 0; // Fake TVECT |
946 |
– |
RoutineDescriptor desc = BUILD_PPC_ROUTINE_DESCRIPTOR(0, tvect); |
947 |
– |
M68kRegisters r; |
948 |
– |
Execute68k((uint32)&desc, &r); |
949 |
– |
} |
950 |
– |
|
951 |
– |
/* |
1317 |
|
* Execute 68k subroutine (must be ended with EXEC_RETURN) |
1318 |
|
* This must only be called by the emul_thread when in EMUL_OP mode |
1319 |
|
* r->a[7] is unused, the routine runs on the caller's stack |
1331 |
|
|
1332 |
|
void Execute68kTrap(uint16 trap, M68kRegisters *r) |
1333 |
|
{ |
1334 |
< |
uint16 proc[2]; |
1335 |
< |
proc[0] = htons(trap); |
1336 |
< |
proc[1] = htons(M68K_RTS); |
1337 |
< |
Execute68k((uint32)proc, r); |
1334 |
> |
SheepVar proc_var(4); |
1335 |
> |
uint32 proc = proc_var.addr(); |
1336 |
> |
WriteMacInt16(proc, trap); |
1337 |
> |
WriteMacInt16(proc + 2, M68K_RTS); |
1338 |
> |
Execute68k(proc, r); |
1339 |
|
} |
1340 |
|
|
1341 |
|
/* |