ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.18 by gbeauche, 2003-12-28T17:56:03Z vs.
Revision 1.50 by gbeauche, 2005-11-30T07:24:53Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2005 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 41 | Line 41
41   #include <assert.h>
42   #include <stdio.h>
43   #include <stdlib.h>
44 + #include <stddef.h>
45   #include <string.h>
46   #include <signal.h>
47  
48 + #ifdef HAVE_PTHREADS
49 + # include <pthread.h>
50 + #endif
51 +
52   #ifdef HAVE_FCNTL_H
53   # include <fcntl.h>
54   #endif
# Line 59 | Line 64
64   # endif
65   #endif
66  
67 + // Fix offsetof() on FreeBSD and GCC >= 3.4
68 + #if defined(__FreeBSD__) && defined(__cplusplus)
69 + #undef offsetof
70 + /* The cast to "char &" below avoids problems with user-defined
71 +   "operator &", which can appear in a POD type.  */
72 + #define offsetof(TYPE, MEMBER)                          \
73 +  (__offsetof__ (reinterpret_cast <size_t>              \
74 +                 (&reinterpret_cast <char &>            \
75 +                  (static_cast<TYPE *> (0)->MEMBER))))
76 + #endif
77 +
78   // Define for external components
79   #define SHEEPSHAVER 1
80  
81 < // Mac and host address space are the same
81 > // Always use Real Addressing mode on native architectures
82 > // Otherwise, use Direct Addressing mode if NATMEM_OFFSET is set
83 > #if !defined(EMULATED_PPC)
84   #define REAL_ADDRESSING 1
85 + #include "ppc_asm.tmpl"
86 + #elif defined(NATMEM_OFFSET)
87 + #define DIRECT_ADDRESSING 1
88 + #else
89 + #define REAL_ADDRESSING 1
90 + #endif
91 +
92 + // Always use the complete non-stubs Ethernet driver
93 + #define USE_ETHER_FULL_DRIVER 1
94  
95   #define POWERPC_ROM 1
96  
97   #if EMULATED_PPC
71 // Handle interrupts asynchronously?
72 #define ASYNC_IRQ 0
98   // Mac ROM is write protected when banked memory is used
99   #if REAL_ADDRESSING || DIRECT_ADDRESSING
100   # define ROM_IS_WRITE_PROTECTED 0
# Line 78 | Line 103
103   # define ROM_IS_WRITE_PROTECTED 1
104   #endif
105   // Configure PowerPC emulator
106 < #define PPC_CHECK_INTERRUPTS (ASYNC_IRQ ? 0 : 1)
106 > #define PPC_REENTRANT_JIT 1
107 > #define PPC_CHECK_INTERRUPTS 1
108   #define PPC_DECODE_CACHE 1
109   #define PPC_FLIGHT_RECORDER 1
110   #define PPC_PROFILE_COMPILE_TIME 0
111   #define PPC_PROFILE_GENERIC_CALLS 0
112   #define KPX_MAX_CPUS 1
113 + #if ENABLE_DYNGEN
114 + #define PPC_ENABLE_JIT 1
115 + #endif
116 + #if defined(__i386__)
117 + #define DYNGEN_ASM_OPTS 1
118 + #endif
119   #else
120   // Mac ROM is write protected
121   #define ROM_IS_WRITE_PROTECTED 1
# Line 139 | Line 171 | typedef int64 intptr;
171   **/
172  
173   #if defined(__GNUC__)
174 < #if defined(__x86_64__)
174 > #if defined(__x86_64__) || defined(__i386__)
175   // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
176   #define opt_bswap_32 do_opt_bswap_32
177   static inline uint32 do_opt_bswap_32(uint32 x)
# Line 184 | Line 216 | static inline uint32 generic_bswap_32(ui
216                    ((x & 0x000000ff) << 24) );
217   }
218  
219 + #if defined(__i386__)
220 + #define opt_bswap_64 do_opt_bswap_64
221 + static inline uint64 do_opt_bswap_64(uint64 x)
222 + {
223 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
224 + }
225 + #endif
226 +
227   #ifdef  opt_bswap_64
228   #undef  bswap_64
229   #define bswap_64 opt_bswap_64
# Line 217 | Line 257 | static inline uint64 tswap64(uint64 x) {
257   // spin locks
258   #ifdef __GNUC__
259  
260 < #ifdef __powerpc__
260 > #if defined(__powerpc__) || defined(__ppc__)
261   #define HAVE_TEST_AND_SET 1
262 < static inline int testandset(int *p)
262 > static inline int testandset(volatile int *p)
263   {
264          int ret;
265 <        __asm__ __volatile__("0:    lwarx %0,0,%1 ;"
266 <                                                 "      xor. %0,%3,%0;"
267 <                                                 "      bne 1f;"
268 <                                                 "      stwcx. %2,0,%1;"
269 <                                                 "      bne- 0b;"
265 >        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
266 >                                                 "      xor.    %0,%3,%0\n"
267 >                                                 "      bne             1f\n"
268 >                                                 "      stwcx.  %2,0,%1\n"
269 >                                                 "      bne-    0b\n"
270                                                   "1:    "
271                                                   : "=&r" (ret)
272                                                   : "r" (p), "r" (1), "r" (0)
# Line 235 | Line 275 | static inline int testandset(int *p)
275   }
276   #endif
277  
278 < #ifdef __i386__
278 > #if defined(__i386__) || defined(__x86_64__)
279   #define HAVE_TEST_AND_SET 1
280 < static inline int testandset(int *p)
280 > static inline int testandset(volatile int *p)
281   {
282 <        char ret;
283 <        long int readval;
284 <        
285 <        __asm__ __volatile__("lock; cmpxchgl %3, %1; sete %0"
286 <                                                 : "=q" (ret), "=m" (*p), "=a" (readval)
247 <                                                 : "r" (1), "m" (*p), "a" (0)
282 >        long int ret;
283 >        /* Note: the "xchg" instruction does not need a "lock" prefix */
284 >        __asm__ __volatile__("xchgl %k0, %1"
285 >                                                 : "=r" (ret), "=m" (*p)
286 >                                                 : "0" (1), "m" (*p)
287                                                   : "memory");
288          return ret;
289   }
# Line 252 | Line 291 | static inline int testandset(int *p)
291  
292   #ifdef __s390__
293   #define HAVE_TEST_AND_SET 1
294 < static inline int testandset(int *p)
294 > static inline int testandset(volatile int *p)
295   {
296          int ret;
297  
# Line 267 | Line 306 | static inline int testandset(int *p)
306  
307   #ifdef __alpha__
308   #define HAVE_TEST_AND_SET 1
309 < static inline int testandset(int *p)
309 > static inline int testandset(volatile int *p)
310   {
311          int ret;
312          unsigned long one;
# Line 287 | Line 326 | static inline int testandset(int *p)
326  
327   #ifdef __sparc__
328   #define HAVE_TEST_AND_SET 1
329 < static inline int testandset(int *p)
329 > static inline int testandset(volatile int *p)
330   {
331          int ret;
332  
# Line 302 | Line 341 | static inline int testandset(int *p)
341  
342   #ifdef __arm__
343   #define HAVE_TEST_AND_SET 1
344 < static inline int testandset(int *p)
344 > static inline int testandset(volatile int *p)
345   {
346          register unsigned int ret;
347          __asm__ __volatile__("swp %0, %1, [%2]"
# Line 315 | Line 354 | static inline int testandset(int *p)
354  
355   #endif /* __GNUC__ */
356  
357 < #if HAVE_TEST_AND_SET
319 < #define HAVE_SPINLOCKS 1
320 < typedef int spinlock_t;
357 > typedef volatile int spinlock_t;
358  
359   static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
360  
361 + #if defined(HAVE_TEST_AND_SET) && defined(HAVE_PTHREADS)
362 + // There is nothing to lock if we are not in an multithreaded environment
363 + #define HAVE_SPINLOCKS 1
364   static inline void spin_lock(spinlock_t *lock)
365   {
366          while (testandset(lock));
# Line 335 | Line 375 | static inline int spin_trylock(spinlock_
375   {
376          return !testandset(lock);
377   }
378 + #else
379 + static inline void spin_lock(spinlock_t *lock)
380 + {
381 + }
382 +
383 + static inline void spin_unlock(spinlock_t *lock)
384 + {
385 + }
386 +
387 + static inline int spin_trylock(spinlock_t *lock)
388 + {
389 +        return 1;
390 + }
391   #endif
392  
393   // Time data type for Time Manager emulation
# Line 344 | Line 397 | typedef struct timespec tm_time_t;
397   typedef struct timeval tm_time_t;
398   #endif
399  
400 + /* Define codes for all the float formats that we know of.
401 + * Though we only handle IEEE format.  */
402 + #define UNKNOWN_FLOAT_FORMAT 0
403 + #define IEEE_FLOAT_FORMAT 1
404 + #define VAX_FLOAT_FORMAT 2
405 + #define IBM_FLOAT_FORMAT 3
406 + #define C4X_FLOAT_FORMAT 4
407 +
408 + // High-precision timing
409 + #if defined(HAVE_PTHREADS) && defined(HAVE_CLOCK_NANOSLEEP)
410 + #define PRECISE_TIMING 1
411 + #define PRECISE_TIMING_POSIX 1
412 + #endif
413 +
414   // Timing functions
415   extern uint64 GetTicks_usec(void);
416   extern void Delay_usec(uint32 usec);
417  
418 + #ifdef HAVE_PTHREADS
419   // Setup pthread attributes
420   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
421 + #endif
422  
423   // Various definitions
424   typedef struct rgb_color {
# Line 359 | Line 428 | typedef struct rgb_color {
428          uint8           alpha;
429   } rgb_color;
430  
431 + // X11 display fast locks
432 + #if defined(HAVE_PTHREADS)
433 + #define X11_LOCK_TYPE pthread_mutex_t
434 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
435 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
436 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
437 + #elif defined(HAVE_SPINLOCKS)
438 + #define X11_LOCK_TYPE spinlock_t
439 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
440 + #define XDisplayLock() spin_lock(&x_display_lock)
441 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
442 + #else
443 + #define XDisplayLock()
444 + #define XDisplayUnlock()
445 + #endif
446 + #ifdef X11_LOCK_TYPE
447 + extern X11_LOCK_TYPE x_display_lock;
448 + #endif
449 +
450   // Macro for calling MacOS routines
451 < #define CallMacOS(type, tvect) call_macos((uint32)tvect)
452 < #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)
453 < #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uint32)tvect, (uint32)arg1, (uint32)arg2)
454 < #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3)
455 < #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4)
456 < #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5)
457 < #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6)
458 < #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6, (uint32)arg7)
451 > #define CallMacOS(type, tvect) call_macos((uintptr)tvect)
452 > #define CallMacOS1(type, tvect, arg1) call_macos1((uintptr)tvect, (uintptr)arg1)
453 > #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uintptr)tvect, (uintptr)arg1, (uintptr)arg2)
454 > #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3)
455 > #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4)
456 > #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5)
457 > #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6)
458 > #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6, (uintptr)arg7)
459  
460   #ifdef __cplusplus
461   extern "C" {

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines