ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.2 by cebix, 2002-02-21T15:12:12Z vs.
Revision 1.49 by gbeauche, 2005-11-27T16:20:17Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2005 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 41 | Line 41
41   #include <assert.h>
42   #include <stdio.h>
43   #include <stdlib.h>
44 + #include <stddef.h>
45   #include <string.h>
46   #include <signal.h>
47  
48 + #ifdef HAVE_PTHREADS
49 + # include <pthread.h>
50 + #endif
51 +
52   #ifdef HAVE_FCNTL_H
53   # include <fcntl.h>
54   #endif
# Line 59 | Line 64
64   # endif
65   #endif
66  
67 < // Are we using a PPC emulator or the real thing?
68 < #ifdef __powerpc__
69 < #define EMULATED_PPC 0
67 > // Fix offsetof() on FreeBSD and GCC >= 3.4
68 > #if defined(__FreeBSD__) && defined(__cplusplus)
69 > #undef offsetof
70 > /* The cast to "char &" below avoids problems with user-defined
71 >   "operator &", which can appear in a POD type.  */
72 > #define offsetof(TYPE, MEMBER)                          \
73 >  (__offsetof__ (reinterpret_cast <size_t>              \
74 >                 (&reinterpret_cast <char &>            \
75 >                  (static_cast<TYPE *> (0)->MEMBER))))
76 > #endif
77 >
78 > // Define for external components
79 > #define SHEEPSHAVER 1
80 >
81 > // Always use Real Addressing mode on native architectures
82 > // Otherwise, use Direct Addressing mode if NATMEM_OFFSET is set
83 > #if !defined(EMULATED_PPC)
84 > #define REAL_ADDRESSING 1
85 > #include "ppc_asm.tmpl"
86 > #elif defined(NATMEM_OFFSET)
87 > #define DIRECT_ADDRESSING 1
88   #else
89 < #define EMULATED_PPC 1
89 > #define REAL_ADDRESSING 1
90 > #endif
91 >
92 > // Always use the complete non-stubs Ethernet driver
93 > #if DIRECT_ADDRESSING
94 > #define USE_ETHER_FULL_DRIVER 1
95   #endif
96  
97   #define POWERPC_ROM 1
98  
99 + #if EMULATED_PPC
100 + // Mac ROM is write protected when banked memory is used
101 + #if REAL_ADDRESSING || DIRECT_ADDRESSING
102 + # define ROM_IS_WRITE_PROTECTED 0
103 + # define USE_SCRATCHMEM_SUBTERFUGE 1
104 + #else
105 + # define ROM_IS_WRITE_PROTECTED 1
106 + #endif
107 + // Configure PowerPC emulator
108 + #define PPC_REENTRANT_JIT 1
109 + #define PPC_CHECK_INTERRUPTS 1
110 + #define PPC_DECODE_CACHE 1
111 + #define PPC_FLIGHT_RECORDER 1
112 + #define PPC_PROFILE_COMPILE_TIME 0
113 + #define PPC_PROFILE_GENERIC_CALLS 0
114 + #define KPX_MAX_CPUS 1
115 + #if ENABLE_DYNGEN
116 + #define PPC_ENABLE_JIT 1
117 + #endif
118 + #if defined(__i386__)
119 + #define DYNGEN_ASM_OPTS 1
120 + #endif
121 + #else
122 + // Mac ROM is write protected
123 + #define ROM_IS_WRITE_PROTECTED 1
124 + #define USE_SCRATCHMEM_SUBTERFUGE 0
125 + #endif
126 +
127   // Data types
128   typedef unsigned char uint8;
129   typedef signed char int8;
# Line 92 | Line 148 | typedef long int32;
148   #if SIZEOF_LONG == 8
149   typedef unsigned long uint64;
150   typedef long int64;
151 + #define VAL64(a) (a ## l)
152 + #define UVAL64(a) (a ## ul)
153   #elif SIZEOF_LONG_LONG == 8
154   typedef unsigned long long uint64;
155   typedef long long int64;
156 + #define VAL64(a) (a ## LL)
157 + #define UVAL64(a) (a ## uLL)
158   #else
159   #error "No 8 byte type, you lose."
160   #endif
161 + #if SIZEOF_VOID_P == 4
162 + typedef uint32 uintptr;
163 + typedef int32 intptr;
164 + #elif SIZEOF_VOID_P == 8
165 + typedef uint64 uintptr;
166 + typedef int64 intptr;
167 + #else
168 + #error "Unsupported size of pointer"
169 + #endif
170 +
171 + /**
172 + *              Helper functions to byteswap data
173 + **/
174 +
175 + #if defined(__GNUC__)
176 + #if defined(__x86_64__) || defined(__i386__)
177 + // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
178 + #define opt_bswap_32 do_opt_bswap_32
179 + static inline uint32 do_opt_bswap_32(uint32 x)
180 + {
181 +  uint32 v;
182 +  __asm__ __volatile__ ("bswap %0" : "=r" (v) : "0" (x));
183 +  return v;
184 + }
185 + #endif
186 + #endif
187 +
188 + #ifdef HAVE_BYTESWAP_H
189 + #include <byteswap.h>
190 + #endif
191 +
192 + #ifdef  opt_bswap_16
193 + #undef  bswap_16
194 + #define bswap_16 opt_bswap_16
195 + #endif
196 + #ifndef bswap_16
197 + #define bswap_16 generic_bswap_16
198 + #endif
199 +
200 + static inline uint16 generic_bswap_16(uint16 x)
201 + {
202 +  return ((x & 0xff) << 8) | ((x >> 8) & 0xff);
203 + }
204 +
205 + #ifdef  opt_bswap_32
206 + #undef  bswap_32
207 + #define bswap_32 opt_bswap_32
208 + #endif
209 + #ifndef bswap_32
210 + #define bswap_32 generic_bswap_32
211 + #endif
212 +
213 + static inline uint32 generic_bswap_32(uint32 x)
214 + {
215 +  return (((x & 0xff000000) >> 24) |
216 +                  ((x & 0x00ff0000) >>  8) |
217 +                  ((x & 0x0000ff00) <<  8) |
218 +                  ((x & 0x000000ff) << 24) );
219 + }
220 +
221 + #if defined(__i386__)
222 + #define opt_bswap_64 do_opt_bswap_64
223 + static inline uint64 do_opt_bswap_64(uint64 x)
224 + {
225 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
226 + }
227 + #endif
228 +
229 + #ifdef  opt_bswap_64
230 + #undef  bswap_64
231 + #define bswap_64 opt_bswap_64
232 + #endif
233 + #ifndef bswap_64
234 + #define bswap_64 generic_bswap_64
235 + #endif
236 +
237 + static inline uint64 generic_bswap_64(uint64 x)
238 + {
239 +  return (((x & UVAL64(0xff00000000000000)) >> 56) |
240 +                  ((x & UVAL64(0x00ff000000000000)) >> 40) |
241 +                  ((x & UVAL64(0x0000ff0000000000)) >> 24) |
242 +                  ((x & UVAL64(0x000000ff00000000)) >>  8) |
243 +                  ((x & UVAL64(0x00000000ff000000)) <<  8) |
244 +                  ((x & UVAL64(0x0000000000ff0000)) << 24) |
245 +                  ((x & UVAL64(0x000000000000ff00)) << 40) |
246 +                  ((x & UVAL64(0x00000000000000ff)) << 56) );
247 + }
248 +
249 + #ifdef WORDS_BIGENDIAN
250 + static inline uint16 tswap16(uint16 x) { return x; }
251 + static inline uint32 tswap32(uint32 x) { return x; }
252 + static inline uint64 tswap64(uint64 x) { return x; }
253 + #else
254 + static inline uint16 tswap16(uint16 x) { return bswap_16(x); }
255 + static inline uint32 tswap32(uint32 x) { return bswap_32(x); }
256 + static inline uint64 tswap64(uint64 x) { return bswap_64(x); }
257 + #endif
258 +
259 + // spin locks
260 + #ifdef __GNUC__
261 +
262 + #if defined(__powerpc__) || defined(__ppc__)
263 + #define HAVE_TEST_AND_SET 1
264 + static inline int testandset(volatile int *p)
265 + {
266 +        int ret;
267 +        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
268 +                                                 "      xor.    %0,%3,%0\n"
269 +                                                 "      bne             1f\n"
270 +                                                 "      stwcx.  %2,0,%1\n"
271 +                                                 "      bne-    0b\n"
272 +                                                 "1:    "
273 +                                                 : "=&r" (ret)
274 +                                                 : "r" (p), "r" (1), "r" (0)
275 +                                                 : "cr0", "memory");
276 +        return ret;
277 + }
278 + #endif
279 +
280 + #if defined(__i386__) || defined(__x86_64__)
281 + #define HAVE_TEST_AND_SET 1
282 + static inline int testandset(volatile int *p)
283 + {
284 +        long int ret;
285 +        /* Note: the "xchg" instruction does not need a "lock" prefix */
286 +        __asm__ __volatile__("xchgl %k0, %1"
287 +                                                 : "=r" (ret), "=m" (*p)
288 +                                                 : "0" (1), "m" (*p)
289 +                                                 : "memory");
290 +        return ret;
291 + }
292 + #endif
293 +
294 + #ifdef __s390__
295 + #define HAVE_TEST_AND_SET 1
296 + static inline int testandset(volatile int *p)
297 + {
298 +        int ret;
299 +
300 +        __asm__ __volatile__("0: cs    %0,%1,0(%2)\n"
301 +                                                 "   jl    0b"
302 +                                                 : "=&d" (ret)
303 +                                                 : "r" (1), "a" (p), "0" (*p)
304 +                                                 : "cc", "memory" );
305 +        return ret;
306 + }
307 + #endif
308 +
309 + #ifdef __alpha__
310 + #define HAVE_TEST_AND_SET 1
311 + static inline int testandset(volatile int *p)
312 + {
313 +        int ret;
314 +        unsigned long one;
315 +
316 +        __asm__ __volatile__("0:        mov 1,%2\n"
317 +                                                 "      ldl_l %0,%1\n"
318 +                                                 "      stl_c %2,%1\n"
319 +                                                 "      beq %2,1f\n"
320 +                                                 ".subsection 2\n"
321 +                                                 "1:    br 0b\n"
322 +                                                 ".previous"
323 +                                                 : "=r" (ret), "=m" (*p), "=r" (one)
324 +                                                 : "m" (*p));
325 +        return ret;
326 + }
327 + #endif
328 +
329 + #ifdef __sparc__
330 + #define HAVE_TEST_AND_SET 1
331 + static inline int testandset(volatile int *p)
332 + {
333 +        int ret;
334 +
335 +        __asm__ __volatile__("ldstub    [%1], %0"
336 +                                                 : "=r" (ret)
337 +                                                 : "r" (p)
338 +                                                 : "memory");
339 +
340 +        return (ret ? 1 : 0);
341 + }
342 + #endif
343 +
344 + #ifdef __arm__
345 + #define HAVE_TEST_AND_SET 1
346 + static inline int testandset(volatile int *p)
347 + {
348 +        register unsigned int ret;
349 +        __asm__ __volatile__("swp %0, %1, [%2]"
350 +                                                 : "=r"(ret)
351 +                                                 : "0"(1), "r"(p));
352 +        
353 +        return ret;
354 + }
355 + #endif
356 +
357 + #endif /* __GNUC__ */
358 +
359 + typedef volatile int spinlock_t;
360 +
361 + static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
362 +
363 + #if defined(HAVE_TEST_AND_SET) && defined(HAVE_PTHREADS)
364 + // There is nothing to lock if we are not in an multithreaded environment
365 + #define HAVE_SPINLOCKS 1
366 + static inline void spin_lock(spinlock_t *lock)
367 + {
368 +        while (testandset(lock));
369 + }
370 +
371 + static inline void spin_unlock(spinlock_t *lock)
372 + {
373 +        *lock = 0;
374 + }
375 +
376 + static inline int spin_trylock(spinlock_t *lock)
377 + {
378 +        return !testandset(lock);
379 + }
380 + #else
381 + static inline void spin_lock(spinlock_t *lock)
382 + {
383 + }
384 +
385 + static inline void spin_unlock(spinlock_t *lock)
386 + {
387 + }
388 +
389 + static inline int spin_trylock(spinlock_t *lock)
390 + {
391 +        return 1;
392 + }
393 + #endif
394  
395   // Time data type for Time Manager emulation
396   #ifdef HAVE_CLOCK_GETTIME
# Line 106 | Line 399 | typedef struct timespec tm_time_t;
399   typedef struct timeval tm_time_t;
400   #endif
401  
402 + /* Define codes for all the float formats that we know of.
403 + * Though we only handle IEEE format.  */
404 + #define UNKNOWN_FLOAT_FORMAT 0
405 + #define IEEE_FLOAT_FORMAT 1
406 + #define VAX_FLOAT_FORMAT 2
407 + #define IBM_FLOAT_FORMAT 3
408 + #define C4X_FLOAT_FORMAT 4
409 +
410 + // High-precision timing
411 + #if defined(HAVE_PTHREADS) && defined(HAVE_CLOCK_NANOSLEEP)
412 + #define PRECISE_TIMING 1
413 + #define PRECISE_TIMING_POSIX 1
414 + #endif
415 +
416 + // Timing functions
417 + extern uint64 GetTicks_usec(void);
418 + extern void Delay_usec(uint32 usec);
419 +
420 + #ifdef HAVE_PTHREADS
421   // Setup pthread attributes
422   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
423 + #endif
424  
425   // Various definitions
426   typedef struct rgb_color {
# Line 117 | Line 430 | typedef struct rgb_color {
430          uint8           alpha;
431   } rgb_color;
432  
433 + // X11 display fast locks
434 + #if defined(HAVE_PTHREADS)
435 + #define X11_LOCK_TYPE pthread_mutex_t
436 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
437 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
438 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
439 + #elif defined(HAVE_SPINLOCKS)
440 + #define X11_LOCK_TYPE spinlock_t
441 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
442 + #define XDisplayLock() spin_lock(&x_display_lock)
443 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
444 + #else
445 + #define XDisplayLock()
446 + #define XDisplayUnlock()
447 + #endif
448 + #ifdef X11_LOCK_TYPE
449 + extern X11_LOCK_TYPE x_display_lock;
450 + #endif
451 +
452   // Macro for calling MacOS routines
453 < #define CallMacOS(type, tvect) call_macos((uint32)tvect)
454 < #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)
455 < #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uint32)tvect, (uint32)arg1, (uint32)arg2)
456 < #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3)
457 < #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4)
458 < #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5)
459 < #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6)
460 < #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6, (uint32)arg7)
461 <
462 < extern "C" uint32 call_macos(uint32 tvect);
463 < extern "C" uint32 call_macos1(uint32 tvect, uint32 arg1);
464 < extern "C" uint32 call_macos2(uint32 tvect, uint32 arg1, uint32 arg2);
465 < extern "C" uint32 call_macos3(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3);
466 < extern "C" uint32 call_macos4(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4);
467 < extern "C" uint32 call_macos5(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5);
468 < extern "C" uint32 call_macos6(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5, uint32 arg6);
469 < extern "C" uint32 call_macos7(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5, uint32 arg6, uint32 arg7);
453 > #define CallMacOS(type, tvect) call_macos((uintptr)tvect)
454 > #define CallMacOS1(type, tvect, arg1) call_macos1((uintptr)tvect, (uintptr)arg1)
455 > #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uintptr)tvect, (uintptr)arg1, (uintptr)arg2)
456 > #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3)
457 > #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4)
458 > #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5)
459 > #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6)
460 > #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6, (uintptr)arg7)
461 >
462 > #ifdef __cplusplus
463 > extern "C" {
464 > #endif
465 > extern uint32 call_macos(uint32 tvect);
466 > extern uint32 call_macos1(uint32 tvect, uint32 arg1);
467 > extern uint32 call_macos2(uint32 tvect, uint32 arg1, uint32 arg2);
468 > extern uint32 call_macos3(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3);
469 > extern uint32 call_macos4(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4);
470 > extern uint32 call_macos5(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5);
471 > extern uint32 call_macos6(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5, uint32 arg6);
472 > extern uint32 call_macos7(uint32 tvect, uint32 arg1, uint32 arg2, uint32 arg3, uint32 arg4, uint32 arg5, uint32 arg6, uint32 arg7);
473 > #ifdef __cplusplus
474 > }
475 > #endif
476  
477   #endif

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines