ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.7 by gbeauche, 2003-09-29T15:46:07Z vs.
Revision 1.40 by gbeauche, 2005-03-05T19:07:35Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2005 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 41 | Line 41
41   #include <assert.h>
42   #include <stdio.h>
43   #include <stdlib.h>
44 + #include <stddef.h>
45   #include <string.h>
46   #include <signal.h>
47  
48 + #ifdef HAVE_PTHREADS
49 + # include <pthread.h>
50 + #endif
51 +
52   #ifdef HAVE_FCNTL_H
53   # include <fcntl.h>
54   #endif
# Line 59 | Line 64
64   # endif
65   #endif
66  
67 + // Fix offsetof() on FreeBSD and GCC >= 3.4
68 + #if defined(__FreeBSD__) && defined(__cplusplus)
69 + #undef offsetof
70 + /* The cast to "char &" below avoids problems with user-defined
71 +   "operator &", which can appear in a POD type.  */
72 + #define offsetof(TYPE, MEMBER)                          \
73 +  (__offsetof__ (reinterpret_cast <size_t>              \
74 +                 (&reinterpret_cast <char &>            \
75 +                  (static_cast<TYPE *> (0)->MEMBER))))
76 + #endif
77 +
78   // Define for external components
79   #define SHEEPSHAVER 1
80  
81 < // Mac and host address space are the same
81 > // Always use Real Addressing mode on native architectures
82 > // Otherwise, use Direct Addressing mode if NATMEM_OFFSET is set
83 > #if !defined(EMULATED_PPC)
84 > #define REAL_ADDRESSING 1
85 > #include "ppc_asm.tmpl"
86 > #elif defined(__CYGWIN__)
87 > #define DIRECT_ADDRESSING 1
88 > #define DIRECT_ADDRESSING_HACK 1
89 > /*
90 >  The following address translation functions were empirically
91 >  determined on a Windows XP system running Cygwin 1.5.12-1 so
92 >  that RAM size can be maximized (up to 960 MB) and avoiding
93 >  the use of a TLB. This also takes into account reduced address
94 >  space available when the Cygwin runtime is used.
95 > */
96 > #define DIRECT_ADDRESSING_VIRT2PHYS(ADDR) \
97 >        ((ADDR) + (((ADDR)  < 0x41000000) ? 0x39000000 : 0xcf800000))
98 > #define DIRECT_ADDRESSING_PHYS2VIRT(ADDR) \
99 >        ((ADDR) - (((ADDR) >= 0x39000000) ? 0x39000000 : 0xcf800000))
100 > #elif defined(NATMEM_OFFSET)
101 > #define DIRECT_ADDRESSING 1
102 > #else
103   #define REAL_ADDRESSING 1
104 + #endif
105  
106   #define POWERPC_ROM 1
107  
108   #if EMULATED_PPC
71 // Handle interrupts asynchronously?
72 #define ASYNC_IRQ 0
109   // Mac ROM is write protected when banked memory is used
110   #if REAL_ADDRESSING || DIRECT_ADDRESSING
111   # define ROM_IS_WRITE_PROTECTED 0
# Line 77 | Line 113
113   #else
114   # define ROM_IS_WRITE_PROTECTED 1
115   #endif
116 + // Configure PowerPC emulator
117 + #define PPC_REENTRANT_JIT 1
118 + #define PPC_CHECK_INTERRUPTS 1
119 + #define PPC_DECODE_CACHE 1
120 + #define PPC_FLIGHT_RECORDER 1
121 + #define PPC_PROFILE_COMPILE_TIME 0
122 + #define PPC_PROFILE_GENERIC_CALLS 0
123 + #define KPX_MAX_CPUS 1
124 + #if ENABLE_DYNGEN
125 + // Don't bother with predecode cache when using JIT
126 + #define PPC_ENABLE_JIT 1
127 + #undef  PPC_DECODE_CACHE
128 + #endif
129 + #if defined(__i386__)
130 + #define DYNGEN_ASM_OPTS 1
131 + #endif
132   #else
133   // Mac ROM is write protected
134   #define ROM_IS_WRITE_PROTECTED 1
# Line 127 | Line 179 | typedef int64 intptr;
179   #error "Unsupported size of pointer"
180   #endif
181  
182 < // Helper functions to byteswap data
182 > /**
183 > *              Helper functions to byteswap data
184 > **/
185 >
186 > #if defined(__GNUC__)
187 > #if defined(__x86_64__) || defined(__i386__)
188 > // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
189 > #define opt_bswap_32 do_opt_bswap_32
190 > static inline uint32 do_opt_bswap_32(uint32 x)
191 > {
192 >  uint32 v;
193 >  __asm__ __volatile__ ("bswap %0" : "=r" (v) : "0" (x));
194 >  return v;
195 > }
196 > #endif
197 > #endif
198 >
199   #ifdef HAVE_BYTESWAP_H
200   #include <byteswap.h>
201   #endif
202  
203 + #ifdef  opt_bswap_16
204 + #undef  bswap_16
205 + #define bswap_16 opt_bswap_16
206 + #endif
207   #ifndef bswap_16
208   #define bswap_16 generic_bswap_16
209   #endif
# Line 141 | Line 213 | static inline uint16 generic_bswap_16(ui
213    return ((x & 0xff) << 8) | ((x >> 8) & 0xff);
214   }
215  
216 + #ifdef  opt_bswap_32
217 + #undef  bswap_32
218 + #define bswap_32 opt_bswap_32
219 + #endif
220   #ifndef bswap_32
221   #define bswap_32 generic_bswap_32
222   #endif
# Line 153 | Line 229 | static inline uint32 generic_bswap_32(ui
229                    ((x & 0x000000ff) << 24) );
230   }
231  
232 + #if defined(__i386__)
233 + #define opt_bswap_64 do_opt_bswap_64
234 + static inline uint64 do_opt_bswap_64(uint64 x)
235 + {
236 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
237 + }
238 + #endif
239 +
240 + #ifdef  opt_bswap_64
241 + #undef  bswap_64
242 + #define bswap_64 opt_bswap_64
243 + #endif
244   #ifndef bswap_64
245   #define bswap_64 generic_bswap_64
246   #endif
# Line 182 | Line 270 | static inline uint64 tswap64(uint64 x) {
270   // spin locks
271   #ifdef __GNUC__
272  
273 < #ifdef __powerpc__
273 > #if defined(__powerpc__) || defined(__ppc__)
274   #define HAVE_TEST_AND_SET 1
275 < static inline int testandset(int *p)
275 > static inline int testandset(volatile int *p)
276   {
277          int ret;
278 <        __asm__ __volatile__("0:    lwarx %0,0,%1 ;"
279 <                                                 "      xor. %0,%3,%0;"
280 <                                                 "      bne 1f;"
281 <                                                 "      stwcx. %2,0,%1;"
282 <                                                 "      bne- 0b;"
278 >        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
279 >                                                 "      xor.    %0,%3,%0\n"
280 >                                                 "      bne             1f\n"
281 >                                                 "      stwcx.  %2,0,%1\n"
282 >                                                 "      bne-    0b\n"
283                                                   "1:    "
284                                                   : "=&r" (ret)
285                                                   : "r" (p), "r" (1), "r" (0)
# Line 200 | Line 288 | static inline int testandset(int *p)
288   }
289   #endif
290  
291 < #ifdef __i386__
291 > /* FIXME: SheepShaver occasionnally hangs with those locks */
292 > #if 0 && (defined(__i386__) || defined(__x86_64__))
293   #define HAVE_TEST_AND_SET 1
294 < static inline int testandset(int *p)
294 > static inline int testandset(volatile int *p)
295   {
296 <        char ret;
297 <        long int readval;
298 <        
299 <        __asm__ __volatile__("lock; cmpxchgl %3, %1; sete %0"
300 <                                                 : "=q" (ret), "=m" (*p), "=a" (readval)
212 <                                                 : "r" (1), "m" (*p), "a" (0)
296 >        long int ret;
297 >        /* Note: the "xchg" instruction does not need a "lock" prefix */
298 >        __asm__ __volatile__("xchgl %k0, %1"
299 >                                                 : "=r" (ret), "=m" (*p)
300 >                                                 : "0" (1), "m" (*p)
301                                                   : "memory");
302          return ret;
303   }
# Line 217 | Line 305 | static inline int testandset(int *p)
305  
306   #ifdef __s390__
307   #define HAVE_TEST_AND_SET 1
308 < static inline int testandset(int *p)
308 > static inline int testandset(volatile int *p)
309   {
310          int ret;
311  
# Line 232 | Line 320 | static inline int testandset(int *p)
320  
321   #ifdef __alpha__
322   #define HAVE_TEST_AND_SET 1
323 < static inline int testandset(int *p)
323 > static inline int testandset(volatile int *p)
324   {
325          int ret;
326          unsigned long one;
# Line 252 | Line 340 | static inline int testandset(int *p)
340  
341   #ifdef __sparc__
342   #define HAVE_TEST_AND_SET 1
343 < static inline int testandset(int *p)
343 > static inline int testandset(volatile int *p)
344   {
345          int ret;
346  
# Line 267 | Line 355 | static inline int testandset(int *p)
355  
356   #ifdef __arm__
357   #define HAVE_TEST_AND_SET 1
358 < static inline int testandset(int *p)
358 > static inline int testandset(volatile int *p)
359   {
360          register unsigned int ret;
361          __asm__ __volatile__("swp %0, %1, [%2]"
# Line 280 | Line 368 | static inline int testandset(int *p)
368  
369   #endif /* __GNUC__ */
370  
371 < #if HAVE_TEST_AND_SET
284 < #define HAVE_SPINLOCKS 1
285 < typedef int spinlock_t;
371 > typedef volatile int spinlock_t;
372  
373 < const spinlock_t SPIN_LOCK_UNLOCKED = 0;
373 > static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
374  
375 + #if HAVE_TEST_AND_SET
376 + #define HAVE_SPINLOCKS 1
377   static inline void spin_lock(spinlock_t *lock)
378   {
379          while (testandset(lock));
# Line 300 | Line 388 | static inline int spin_trylock(spinlock_
388   {
389          return !testandset(lock);
390   }
391 + #else
392 + static inline void spin_lock(spinlock_t *lock)
393 + {
394 + }
395 +
396 + static inline void spin_unlock(spinlock_t *lock)
397 + {
398 + }
399 +
400 + static inline int spin_trylock(spinlock_t *lock)
401 + {
402 +        return 1;
403 + }
404   #endif
405  
406   // Time data type for Time Manager emulation
# Line 309 | Line 410 | typedef struct timespec tm_time_t;
410   typedef struct timeval tm_time_t;
411   #endif
412  
413 + // High-precision timing
414 + #if defined(HAVE_PTHREADS) && defined(HAVE_CLOCK_NANOSLEEP)
415 + #define PRECISE_TIMING 1
416 + #define PRECISE_TIMING_POSIX 1
417 + #endif
418 +
419 + // Timing functions
420 + extern uint64 GetTicks_usec(void);
421 + extern void Delay_usec(uint32 usec);
422 +
423 + #if defined(HAVE_PTHREADS) || ((defined(__linux__) || defined(__NetBSD__)) && defined(__powerpc__))
424   // Setup pthread attributes
425   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
426 + #endif
427  
428   // Various definitions
429   typedef struct rgb_color {
# Line 320 | Line 433 | typedef struct rgb_color {
433          uint8           alpha;
434   } rgb_color;
435  
436 + // X11 display fast locks
437 + #ifdef HAVE_SPINLOCKS
438 + #define X11_LOCK_TYPE spinlock_t
439 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
440 + #define XDisplayLock() spin_lock(&x_display_lock)
441 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
442 + #elif defined(HAVE_PTHREADS)
443 + #define X11_LOCK_TYPE pthread_mutex_t
444 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
445 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
446 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
447 + #else
448 + #define XDisplayLock()
449 + #define XDisplayUnlock()
450 + #endif
451 + #ifdef X11_LOCK_TYPE
452 + extern X11_LOCK_TYPE x_display_lock;
453 + #endif
454 +
455   // Macro for calling MacOS routines
456 < #define CallMacOS(type, tvect) call_macos((uint32)tvect)
457 < #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)
458 < #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uint32)tvect, (uint32)arg1, (uint32)arg2)
459 < #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3)
460 < #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4)
461 < #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5)
462 < #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6)
463 < #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6, (uint32)arg7)
456 > #define CallMacOS(type, tvect) call_macos((uintptr)tvect)
457 > #define CallMacOS1(type, tvect, arg1) call_macos1((uintptr)tvect, (uintptr)arg1)
458 > #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uintptr)tvect, (uintptr)arg1, (uintptr)arg2)
459 > #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3)
460 > #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4)
461 > #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5)
462 > #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6)
463 > #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6, (uintptr)arg7)
464  
465   #ifdef __cplusplus
466   extern "C" {

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines