ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.11 by gbeauche, 2003-10-12T05:44:14Z vs.
Revision 1.47 by gbeauche, 2005-07-03T13:39:06Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2005 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 41 | Line 41
41   #include <assert.h>
42   #include <stdio.h>
43   #include <stdlib.h>
44 + #include <stddef.h>
45   #include <string.h>
46   #include <signal.h>
47  
48 + #ifdef HAVE_PTHREADS
49 + # include <pthread.h>
50 + #endif
51 +
52   #ifdef HAVE_FCNTL_H
53   # include <fcntl.h>
54   #endif
# Line 59 | Line 64
64   # endif
65   #endif
66  
67 + // Fix offsetof() on FreeBSD and GCC >= 3.4
68 + #if defined(__FreeBSD__) && defined(__cplusplus)
69 + #undef offsetof
70 + /* The cast to "char &" below avoids problems with user-defined
71 +   "operator &", which can appear in a POD type.  */
72 + #define offsetof(TYPE, MEMBER)                          \
73 +  (__offsetof__ (reinterpret_cast <size_t>              \
74 +                 (&reinterpret_cast <char &>            \
75 +                  (static_cast<TYPE *> (0)->MEMBER))))
76 + #endif
77 +
78   // Define for external components
79   #define SHEEPSHAVER 1
80  
81 < // Mac and host address space are the same
81 > // Always use Real Addressing mode on native architectures
82 > // Otherwise, use Direct Addressing mode if NATMEM_OFFSET is set
83 > #if !defined(EMULATED_PPC)
84 > #define REAL_ADDRESSING 1
85 > #include "ppc_asm.tmpl"
86 > #elif defined(NATMEM_OFFSET)
87 > #define DIRECT_ADDRESSING 1
88 > #else
89   #define REAL_ADDRESSING 1
90 + #endif
91  
92   #define POWERPC_ROM 1
93  
94   #if EMULATED_PPC
71 // Handle interrupts asynchronously?
72 #define ASYNC_IRQ 0
95   // Mac ROM is write protected when banked memory is used
96   #if REAL_ADDRESSING || DIRECT_ADDRESSING
97   # define ROM_IS_WRITE_PROTECTED 0
# Line 78 | Line 100
100   # define ROM_IS_WRITE_PROTECTED 1
101   #endif
102   // Configure PowerPC emulator
103 < #define PPC_NO_LAZY_PC_UPDATE 1
104 < #define PPC_NO_DECODE_CACHE 1
103 > #define PPC_REENTRANT_JIT 1
104 > #define PPC_CHECK_INTERRUPTS 1
105 > #define PPC_DECODE_CACHE 1
106   #define PPC_FLIGHT_RECORDER 1
107 + #define PPC_PROFILE_COMPILE_TIME 0
108 + #define PPC_PROFILE_GENERIC_CALLS 0
109 + #define KPX_MAX_CPUS 1
110 + #if ENABLE_DYNGEN
111 + // Don't bother with predecode cache when using JIT
112 + #define PPC_ENABLE_JIT 1
113 + #undef  PPC_DECODE_CACHE
114 + #endif
115 + #if defined(__i386__)
116 + #define DYNGEN_ASM_OPTS 1
117 + #endif
118   #else
119   // Mac ROM is write protected
120   #define ROM_IS_WRITE_PROTECTED 1
# Line 131 | Line 165 | typedef int64 intptr;
165   #error "Unsupported size of pointer"
166   #endif
167  
168 < // Helper functions to byteswap data
168 > /**
169 > *              Helper functions to byteswap data
170 > **/
171 >
172 > #if defined(__GNUC__)
173 > #if defined(__x86_64__) || defined(__i386__)
174 > // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
175 > #define opt_bswap_32 do_opt_bswap_32
176 > static inline uint32 do_opt_bswap_32(uint32 x)
177 > {
178 >  uint32 v;
179 >  __asm__ __volatile__ ("bswap %0" : "=r" (v) : "0" (x));
180 >  return v;
181 > }
182 > #endif
183 > #endif
184 >
185   #ifdef HAVE_BYTESWAP_H
186   #include <byteswap.h>
187   #endif
188  
189 + #ifdef  opt_bswap_16
190 + #undef  bswap_16
191 + #define bswap_16 opt_bswap_16
192 + #endif
193   #ifndef bswap_16
194   #define bswap_16 generic_bswap_16
195   #endif
# Line 145 | Line 199 | static inline uint16 generic_bswap_16(ui
199    return ((x & 0xff) << 8) | ((x >> 8) & 0xff);
200   }
201  
202 + #ifdef  opt_bswap_32
203 + #undef  bswap_32
204 + #define bswap_32 opt_bswap_32
205 + #endif
206   #ifndef bswap_32
207   #define bswap_32 generic_bswap_32
208   #endif
# Line 157 | Line 215 | static inline uint32 generic_bswap_32(ui
215                    ((x & 0x000000ff) << 24) );
216   }
217  
218 + #if defined(__i386__)
219 + #define opt_bswap_64 do_opt_bswap_64
220 + static inline uint64 do_opt_bswap_64(uint64 x)
221 + {
222 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
223 + }
224 + #endif
225 +
226 + #ifdef  opt_bswap_64
227 + #undef  bswap_64
228 + #define bswap_64 opt_bswap_64
229 + #endif
230   #ifndef bswap_64
231   #define bswap_64 generic_bswap_64
232   #endif
# Line 186 | Line 256 | static inline uint64 tswap64(uint64 x) {
256   // spin locks
257   #ifdef __GNUC__
258  
259 < #ifdef __powerpc__
259 > #if defined(__powerpc__) || defined(__ppc__)
260   #define HAVE_TEST_AND_SET 1
261 < static inline int testandset(int *p)
261 > static inline int testandset(volatile int *p)
262   {
263          int ret;
264 <        __asm__ __volatile__("0:    lwarx %0,0,%1 ;"
265 <                                                 "      xor. %0,%3,%0;"
266 <                                                 "      bne 1f;"
267 <                                                 "      stwcx. %2,0,%1;"
268 <                                                 "      bne- 0b;"
264 >        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
265 >                                                 "      xor.    %0,%3,%0\n"
266 >                                                 "      bne             1f\n"
267 >                                                 "      stwcx.  %2,0,%1\n"
268 >                                                 "      bne-    0b\n"
269                                                   "1:    "
270                                                   : "=&r" (ret)
271                                                   : "r" (p), "r" (1), "r" (0)
# Line 204 | Line 274 | static inline int testandset(int *p)
274   }
275   #endif
276  
277 < #ifdef __i386__
277 > #if defined(__i386__) || defined(__x86_64__)
278   #define HAVE_TEST_AND_SET 1
279 < static inline int testandset(int *p)
279 > static inline int testandset(volatile int *p)
280   {
281 <        char ret;
282 <        long int readval;
283 <        
284 <        __asm__ __volatile__("lock; cmpxchgl %3, %1; sete %0"
285 <                                                 : "=q" (ret), "=m" (*p), "=a" (readval)
216 <                                                 : "r" (1), "m" (*p), "a" (0)
281 >        long int ret;
282 >        /* Note: the "xchg" instruction does not need a "lock" prefix */
283 >        __asm__ __volatile__("xchgl %k0, %1"
284 >                                                 : "=r" (ret), "=m" (*p)
285 >                                                 : "0" (1), "m" (*p)
286                                                   : "memory");
287          return ret;
288   }
# Line 221 | Line 290 | static inline int testandset(int *p)
290  
291   #ifdef __s390__
292   #define HAVE_TEST_AND_SET 1
293 < static inline int testandset(int *p)
293 > static inline int testandset(volatile int *p)
294   {
295          int ret;
296  
# Line 236 | Line 305 | static inline int testandset(int *p)
305  
306   #ifdef __alpha__
307   #define HAVE_TEST_AND_SET 1
308 < static inline int testandset(int *p)
308 > static inline int testandset(volatile int *p)
309   {
310          int ret;
311          unsigned long one;
# Line 256 | Line 325 | static inline int testandset(int *p)
325  
326   #ifdef __sparc__
327   #define HAVE_TEST_AND_SET 1
328 < static inline int testandset(int *p)
328 > static inline int testandset(volatile int *p)
329   {
330          int ret;
331  
# Line 271 | Line 340 | static inline int testandset(int *p)
340  
341   #ifdef __arm__
342   #define HAVE_TEST_AND_SET 1
343 < static inline int testandset(int *p)
343 > static inline int testandset(volatile int *p)
344   {
345          register unsigned int ret;
346          __asm__ __volatile__("swp %0, %1, [%2]"
# Line 284 | Line 353 | static inline int testandset(int *p)
353  
354   #endif /* __GNUC__ */
355  
356 < #if HAVE_TEST_AND_SET
288 < #define HAVE_SPINLOCKS 1
289 < typedef int spinlock_t;
356 > typedef volatile int spinlock_t;
357  
358   static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
359  
360 + #if defined(HAVE_TEST_AND_SET) && defined(HAVE_PTHREADS)
361 + // There is nothing to lock if we are not in an multithreaded environment
362 + #define HAVE_SPINLOCKS 1
363   static inline void spin_lock(spinlock_t *lock)
364   {
365          while (testandset(lock));
# Line 304 | Line 374 | static inline int spin_trylock(spinlock_
374   {
375          return !testandset(lock);
376   }
377 + #else
378 + static inline void spin_lock(spinlock_t *lock)
379 + {
380 + }
381 +
382 + static inline void spin_unlock(spinlock_t *lock)
383 + {
384 + }
385 +
386 + static inline int spin_trylock(spinlock_t *lock)
387 + {
388 +        return 1;
389 + }
390   #endif
391  
392   // Time data type for Time Manager emulation
# Line 313 | Line 396 | typedef struct timespec tm_time_t;
396   typedef struct timeval tm_time_t;
397   #endif
398  
399 + /* Define codes for all the float formats that we know of.
400 + * Though we only handle IEEE format.  */
401 + #define UNKNOWN_FLOAT_FORMAT 0
402 + #define IEEE_FLOAT_FORMAT 1
403 + #define VAX_FLOAT_FORMAT 2
404 + #define IBM_FLOAT_FORMAT 3
405 + #define C4X_FLOAT_FORMAT 4
406 +
407 + // High-precision timing
408 + #if defined(HAVE_PTHREADS) && defined(HAVE_CLOCK_NANOSLEEP)
409 + #define PRECISE_TIMING 1
410 + #define PRECISE_TIMING_POSIX 1
411 + #endif
412 +
413 + // Timing functions
414 + extern uint64 GetTicks_usec(void);
415 + extern void Delay_usec(uint32 usec);
416 +
417 + #ifdef HAVE_PTHREADS
418   // Setup pthread attributes
419   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
420 + #endif
421  
422   // Various definitions
423   typedef struct rgb_color {
# Line 324 | Line 427 | typedef struct rgb_color {
427          uint8           alpha;
428   } rgb_color;
429  
430 + // X11 display fast locks
431 + #if defined(HAVE_PTHREADS)
432 + #define X11_LOCK_TYPE pthread_mutex_t
433 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
434 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
435 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
436 + #elif defined(HAVE_SPINLOCKS)
437 + #define X11_LOCK_TYPE spinlock_t
438 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
439 + #define XDisplayLock() spin_lock(&x_display_lock)
440 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
441 + #else
442 + #define XDisplayLock()
443 + #define XDisplayUnlock()
444 + #endif
445 + #ifdef X11_LOCK_TYPE
446 + extern X11_LOCK_TYPE x_display_lock;
447 + #endif
448 +
449   // Macro for calling MacOS routines
450 < #define CallMacOS(type, tvect) call_macos((uint32)tvect)
451 < #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)
452 < #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uint32)tvect, (uint32)arg1, (uint32)arg2)
453 < #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3)
454 < #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4)
455 < #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5)
456 < #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6)
457 < #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6, (uint32)arg7)
450 > #define CallMacOS(type, tvect) call_macos((uintptr)tvect)
451 > #define CallMacOS1(type, tvect, arg1) call_macos1((uintptr)tvect, (uintptr)arg1)
452 > #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uintptr)tvect, (uintptr)arg1, (uintptr)arg2)
453 > #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3)
454 > #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4)
455 > #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5)
456 > #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6)
457 > #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6, (uintptr)arg7)
458  
459   #ifdef __cplusplus
460   extern "C" {

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines