ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.14 by gbeauche, 2003-11-11T11:32:24Z vs.
Revision 1.31 by gbeauche, 2004-05-20T17:46:49Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2004 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 44 | Line 44
44   #include <string.h>
45   #include <signal.h>
46  
47 + #ifdef HAVE_PTHREADS
48 + # include <pthread.h>
49 + #endif
50 +
51   #ifdef HAVE_FCNTL_H
52   # include <fcntl.h>
53   #endif
# Line 68 | Line 72
72   #define POWERPC_ROM 1
73  
74   #if EMULATED_PPC
71 // Handle interrupts asynchronously?
72 #define ASYNC_IRQ 0
75   // Mac ROM is write protected when banked memory is used
76   #if REAL_ADDRESSING || DIRECT_ADDRESSING
77   # define ROM_IS_WRITE_PROTECTED 0
# Line 78 | Line 80
80   # define ROM_IS_WRITE_PROTECTED 1
81   #endif
82   // Configure PowerPC emulator
83 < #define PPC_CHECK_INTERRUPTS (ASYNC_IRQ ? 0 : 1)
83 > #define PPC_REENTRANT_JIT 1
84 > #define PPC_CHECK_INTERRUPTS 1
85   #define PPC_DECODE_CACHE 1
86   #define PPC_FLIGHT_RECORDER 1
87 + #define PPC_PROFILE_COMPILE_TIME 0
88 + #define PPC_PROFILE_GENERIC_CALLS 0
89 + #define KPX_MAX_CPUS 1
90 + #if defined(__i386__)
91 + #define DYNGEN_ASM_OPTS 1
92 + #endif
93   #else
94   // Mac ROM is write protected
95   #define ROM_IS_WRITE_PROTECTED 1
# Line 131 | Line 140 | typedef int64 intptr;
140   #error "Unsupported size of pointer"
141   #endif
142  
143 < // Helper functions to byteswap data
143 > /**
144 > *              Helper functions to byteswap data
145 > **/
146 >
147 > #if defined(__GNUC__)
148 > #if defined(__x86_64__) || defined(__i386__)
149 > // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
150 > #define opt_bswap_32 do_opt_bswap_32
151 > static inline uint32 do_opt_bswap_32(uint32 x)
152 > {
153 >  uint32 v;
154 >  __asm__ __volatile__ ("bswap %0" : "=r" (v) : "0" (x));
155 >  return v;
156 > }
157 > #endif
158 > #endif
159 >
160   #ifdef HAVE_BYTESWAP_H
161   #include <byteswap.h>
162   #endif
163  
164 + #ifdef  opt_bswap_16
165 + #undef  bswap_16
166 + #define bswap_16 opt_bswap_16
167 + #endif
168   #ifndef bswap_16
169   #define bswap_16 generic_bswap_16
170   #endif
# Line 145 | Line 174 | static inline uint16 generic_bswap_16(ui
174    return ((x & 0xff) << 8) | ((x >> 8) & 0xff);
175   }
176  
177 + #ifdef  opt_bswap_32
178 + #undef  bswap_32
179 + #define bswap_32 opt_bswap_32
180 + #endif
181   #ifndef bswap_32
182   #define bswap_32 generic_bswap_32
183   #endif
# Line 157 | Line 190 | static inline uint32 generic_bswap_32(ui
190                    ((x & 0x000000ff) << 24) );
191   }
192  
193 + #if defined(__i386__)
194 + #define opt_bswap_64 do_opt_bswap_64
195 + static inline uint64 do_opt_bswap_64(uint64 x)
196 + {
197 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
198 + }
199 + #endif
200 +
201 + #ifdef  opt_bswap_64
202 + #undef  bswap_64
203 + #define bswap_64 opt_bswap_64
204 + #endif
205   #ifndef bswap_64
206   #define bswap_64 generic_bswap_64
207   #endif
# Line 186 | Line 231 | static inline uint64 tswap64(uint64 x) {
231   // spin locks
232   #ifdef __GNUC__
233  
234 < #ifdef __powerpc__
234 > #if defined(__powerpc__) || defined(__ppc__)
235   #define HAVE_TEST_AND_SET 1
236 < static inline int testandset(int *p)
236 > static inline int testandset(volatile int *p)
237   {
238          int ret;
239 <        __asm__ __volatile__("0:    lwarx %0,0,%1 ;"
240 <                                                 "      xor. %0,%3,%0;"
241 <                                                 "      bne 1f;"
242 <                                                 "      stwcx. %2,0,%1;"
243 <                                                 "      bne- 0b;"
239 >        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
240 >                                                 "      xor.    %0,%3,%0\n"
241 >                                                 "      bne             1f\n"
242 >                                                 "      stwcx.  %2,0,%1\n"
243 >                                                 "      bne-    0b\n"
244                                                   "1:    "
245                                                   : "=&r" (ret)
246                                                   : "r" (p), "r" (1), "r" (0)
# Line 204 | Line 249 | static inline int testandset(int *p)
249   }
250   #endif
251  
252 < #ifdef __i386__
252 > #if defined(__i386__) || defined(__x86_64__)
253   #define HAVE_TEST_AND_SET 1
254 < static inline int testandset(int *p)
254 > static inline int testandset(volatile int *p)
255   {
256 <        char ret;
257 <        long int readval;
258 <        
259 <        __asm__ __volatile__("lock; cmpxchgl %3, %1; sete %0"
260 <                                                 : "=q" (ret), "=m" (*p), "=a" (readval)
216 <                                                 : "r" (1), "m" (*p), "a" (0)
256 >        long int ret;
257 >        /* Note: the "xchg" instruction does not need a "lock" prefix */
258 >        __asm__ __volatile__("xchgl %k0, %1"
259 >                                                 : "=r" (ret), "=m" (*p)
260 >                                                 : "0" (1), "m" (*p)
261                                                   : "memory");
262          return ret;
263   }
# Line 221 | Line 265 | static inline int testandset(int *p)
265  
266   #ifdef __s390__
267   #define HAVE_TEST_AND_SET 1
268 < static inline int testandset(int *p)
268 > static inline int testandset(volatile int *p)
269   {
270          int ret;
271  
# Line 236 | Line 280 | static inline int testandset(int *p)
280  
281   #ifdef __alpha__
282   #define HAVE_TEST_AND_SET 1
283 < static inline int testandset(int *p)
283 > static inline int testandset(volatile int *p)
284   {
285          int ret;
286          unsigned long one;
# Line 256 | Line 300 | static inline int testandset(int *p)
300  
301   #ifdef __sparc__
302   #define HAVE_TEST_AND_SET 1
303 < static inline int testandset(int *p)
303 > static inline int testandset(volatile int *p)
304   {
305          int ret;
306  
# Line 271 | Line 315 | static inline int testandset(int *p)
315  
316   #ifdef __arm__
317   #define HAVE_TEST_AND_SET 1
318 < static inline int testandset(int *p)
318 > static inline int testandset(volatile int *p)
319   {
320          register unsigned int ret;
321          __asm__ __volatile__("swp %0, %1, [%2]"
# Line 286 | Line 330 | static inline int testandset(int *p)
330  
331   #if HAVE_TEST_AND_SET
332   #define HAVE_SPINLOCKS 1
333 < typedef int spinlock_t;
333 > typedef volatile int spinlock_t;
334  
335   static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
336  
# Line 313 | Line 357 | typedef struct timespec tm_time_t;
357   typedef struct timeval tm_time_t;
358   #endif
359  
360 + // Timing functions
361 + extern uint64 GetTicks_usec(void);
362 + extern void Delay_usec(uint32 usec);
363 +
364 + #if defined(HAVE_PTHREADS) || (defined(__linux__) && defined(__powerpc__))
365   // Setup pthread attributes
366   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
367 + #endif
368  
369   // Various definitions
370   typedef struct rgb_color {
# Line 324 | Line 374 | typedef struct rgb_color {
374          uint8           alpha;
375   } rgb_color;
376  
377 + // X11 display fast locks
378 + #ifdef HAVE_SPINLOCKS
379 + #define X11_LOCK_TYPE spinlock_t
380 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
381 + #define XDisplayLock() spin_lock(&x_display_lock)
382 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
383 + #elif defined(HAVE_PTHREADS)
384 + #define X11_LOCK_TYPE pthread_mutex_t
385 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
386 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
387 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
388 + #else
389 + #define XDisplayLock()
390 + #define XDisplayUnlock()
391 + #endif
392 + #ifdef X11_LOCK_TYPE
393 + extern X11_LOCK_TYPE x_display_lock;
394 + #endif
395 +
396   // Macro for calling MacOS routines
397   #define CallMacOS(type, tvect) call_macos((uint32)tvect)
398   #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines