ViewVC Help
View File | Revision Log | Show Annotations | Revision Graph | Root Listing
root/cebix/SheepShaver/src/Unix/sysdeps.h
(Generate patch)

Comparing SheepShaver/src/Unix/sysdeps.h (file contents):
Revision 1.4 by gbeauche, 2003-05-22T22:12:05Z vs.
Revision 1.38 by gbeauche, 2005-02-20T18:08:50Z

# Line 1 | Line 1
1   /*
2   *  sysdeps.h - System dependent definitions for Linux
3   *
4 < *  SheepShaver (C) 1997-2002 Christian Bauer and Marc Hellwig
4 > *  SheepShaver (C) 1997-2005 Christian Bauer and Marc Hellwig
5   *
6   *  This program is free software; you can redistribute it and/or modify
7   *  it under the terms of the GNU General Public License as published by
# Line 41 | Line 41
41   #include <assert.h>
42   #include <stdio.h>
43   #include <stdlib.h>
44 + #include <stddef.h>
45   #include <string.h>
46   #include <signal.h>
47  
48 + #ifdef HAVE_PTHREADS
49 + # include <pthread.h>
50 + #endif
51 +
52   #ifdef HAVE_FCNTL_H
53   # include <fcntl.h>
54   #endif
# Line 59 | Line 64
64   # endif
65   #endif
66  
67 < // Mac and host address space are the same
67 > // Fix offsetof() on FreeBSD and GCC >= 3.4
68 > #if defined(__FreeBSD__) && defined(__cplusplus)
69 > #undef offsetof
70 > /* The cast to "char &" below avoids problems with user-defined
71 >   "operator &", which can appear in a POD type.  */
72 > #define offsetof(TYPE, MEMBER)                          \
73 >  (__offsetof__ (reinterpret_cast <size_t>              \
74 >                 (&reinterpret_cast <char &>            \
75 >                  (static_cast<TYPE *> (0)->MEMBER))))
76 > #endif
77 >
78 > // Define for external components
79 > #define SHEEPSHAVER 1
80 >
81 > // Always use Real Addressing mode on native architectures
82 > // Otherwise, use Direct Addressing mode if NATMEM_OFFSET is set
83 > #if !defined(EMULATED_PPC)
84   #define REAL_ADDRESSING 1
85 <
86 < // Are we using a PPC emulator or the real thing?
87 < #ifdef __powerpc__
88 < #define EMULATED_PPC 0
85 > #elif defined(__CYGWIN__)
86 > #define DIRECT_ADDRESSING 1
87 > #define DIRECT_ADDRESSING_HACK 1
88 > /*
89 >  The following address translation functions were empirically
90 >  determined on a Windows XP system running Cygwin 1.5.12-1 so
91 >  that RAM size can be maximized (up to 960 MB) and avoiding
92 >  the use of a TLB. This also takes into account reduced address
93 >  space available when the Cygwin runtime is used.
94 > */
95 > #define DIRECT_ADDRESSING_VIRT2PHYS(ADDR) \
96 >        ((ADDR) + (((ADDR)  < 0x41000000) ? 0x39000000 : 0xcf800000))
97 > #define DIRECT_ADDRESSING_PHYS2VIRT(ADDR) \
98 >        ((ADDR) - (((ADDR) >= 0x39000000) ? 0x39000000 : 0xcf800000))
99 > #elif defined(NATMEM_OFFSET)
100 > #define DIRECT_ADDRESSING 1
101   #else
102 < #define EMULATED_PPC 1
102 > #define REAL_ADDRESSING 1
103   #endif
104  
105   #define POWERPC_ROM 1
106  
107 + #if EMULATED_PPC
108 + // Mac ROM is write protected when banked memory is used
109 + #if REAL_ADDRESSING || DIRECT_ADDRESSING
110 + # define ROM_IS_WRITE_PROTECTED 0
111 + # define USE_SCRATCHMEM_SUBTERFUGE 1
112 + #else
113 + # define ROM_IS_WRITE_PROTECTED 1
114 + #endif
115 + // Configure PowerPC emulator
116 + #define PPC_REENTRANT_JIT 1
117 + #define PPC_CHECK_INTERRUPTS 1
118 + #define PPC_DECODE_CACHE 1
119 + #define PPC_FLIGHT_RECORDER 1
120 + #define PPC_PROFILE_COMPILE_TIME 0
121 + #define PPC_PROFILE_GENERIC_CALLS 0
122 + #define KPX_MAX_CPUS 1
123 + #if ENABLE_DYNGEN
124 + // Don't bother with predecode cache when using JIT
125 + #define PPC_ENABLE_JIT 1
126 + #undef  PPC_DECODE_CACHE
127 + #endif
128 + #if defined(__i386__)
129 + #define DYNGEN_ASM_OPTS 1
130 + #endif
131 + #else
132 + // Mac ROM is write protected
133 + #define ROM_IS_WRITE_PROTECTED 1
134 + #define USE_SCRATCHMEM_SUBTERFUGE 0
135 + #endif
136 +
137   // Data types
138   typedef unsigned char uint8;
139   typedef signed char int8;
# Line 115 | Line 178 | typedef int64 intptr;
178   #error "Unsupported size of pointer"
179   #endif
180  
181 + /**
182 + *              Helper functions to byteswap data
183 + **/
184 +
185 + #if defined(__GNUC__)
186 + #if defined(__x86_64__) || defined(__i386__)
187 + // Linux/AMD64 currently has no asm optimized bswap_32() in <byteswap.h>
188 + #define opt_bswap_32 do_opt_bswap_32
189 + static inline uint32 do_opt_bswap_32(uint32 x)
190 + {
191 +  uint32 v;
192 +  __asm__ __volatile__ ("bswap %0" : "=r" (v) : "0" (x));
193 +  return v;
194 + }
195 + #endif
196 + #endif
197 +
198 + #ifdef HAVE_BYTESWAP_H
199 + #include <byteswap.h>
200 + #endif
201 +
202 + #ifdef  opt_bswap_16
203 + #undef  bswap_16
204 + #define bswap_16 opt_bswap_16
205 + #endif
206 + #ifndef bswap_16
207 + #define bswap_16 generic_bswap_16
208 + #endif
209 +
210 + static inline uint16 generic_bswap_16(uint16 x)
211 + {
212 +  return ((x & 0xff) << 8) | ((x >> 8) & 0xff);
213 + }
214 +
215 + #ifdef  opt_bswap_32
216 + #undef  bswap_32
217 + #define bswap_32 opt_bswap_32
218 + #endif
219 + #ifndef bswap_32
220 + #define bswap_32 generic_bswap_32
221 + #endif
222 +
223 + static inline uint32 generic_bswap_32(uint32 x)
224 + {
225 +  return (((x & 0xff000000) >> 24) |
226 +                  ((x & 0x00ff0000) >>  8) |
227 +                  ((x & 0x0000ff00) <<  8) |
228 +                  ((x & 0x000000ff) << 24) );
229 + }
230 +
231 + #if defined(__i386__)
232 + #define opt_bswap_64 do_opt_bswap_64
233 + static inline uint64 do_opt_bswap_64(uint64 x)
234 + {
235 +  return (bswap_32(x >> 32) | (((uint64)bswap_32((uint32)x)) << 32));
236 + }
237 + #endif
238 +
239 + #ifdef  opt_bswap_64
240 + #undef  bswap_64
241 + #define bswap_64 opt_bswap_64
242 + #endif
243 + #ifndef bswap_64
244 + #define bswap_64 generic_bswap_64
245 + #endif
246 +
247 + static inline uint64 generic_bswap_64(uint64 x)
248 + {
249 +  return (((x & UVAL64(0xff00000000000000)) >> 56) |
250 +                  ((x & UVAL64(0x00ff000000000000)) >> 40) |
251 +                  ((x & UVAL64(0x0000ff0000000000)) >> 24) |
252 +                  ((x & UVAL64(0x000000ff00000000)) >>  8) |
253 +                  ((x & UVAL64(0x00000000ff000000)) <<  8) |
254 +                  ((x & UVAL64(0x0000000000ff0000)) << 24) |
255 +                  ((x & UVAL64(0x000000000000ff00)) << 40) |
256 +                  ((x & UVAL64(0x00000000000000ff)) << 56) );
257 + }
258 +
259 + #ifdef WORDS_BIGENDIAN
260 + static inline uint16 tswap16(uint16 x) { return x; }
261 + static inline uint32 tswap32(uint32 x) { return x; }
262 + static inline uint64 tswap64(uint64 x) { return x; }
263 + #else
264 + static inline uint16 tswap16(uint16 x) { return bswap_16(x); }
265 + static inline uint32 tswap32(uint32 x) { return bswap_32(x); }
266 + static inline uint64 tswap64(uint64 x) { return bswap_64(x); }
267 + #endif
268 +
269 + // spin locks
270 + #ifdef __GNUC__
271 +
272 + #if defined(__powerpc__) || defined(__ppc__)
273 + #define HAVE_TEST_AND_SET 1
274 + static inline int testandset(volatile int *p)
275 + {
276 +        int ret;
277 +        __asm__ __volatile__("0:    lwarx       %0,0,%1\n"
278 +                                                 "      xor.    %0,%3,%0\n"
279 +                                                 "      bne             1f\n"
280 +                                                 "      stwcx.  %2,0,%1\n"
281 +                                                 "      bne-    0b\n"
282 +                                                 "1:    "
283 +                                                 : "=&r" (ret)
284 +                                                 : "r" (p), "r" (1), "r" (0)
285 +                                                 : "cr0", "memory");
286 +        return ret;
287 + }
288 + #endif
289 +
290 + /* FIXME: SheepShaver occasionnally hangs with those locks */
291 + #if 0 && (defined(__i386__) || defined(__x86_64__))
292 + #define HAVE_TEST_AND_SET 1
293 + static inline int testandset(volatile int *p)
294 + {
295 +        long int ret;
296 +        /* Note: the "xchg" instruction does not need a "lock" prefix */
297 +        __asm__ __volatile__("xchgl %k0, %1"
298 +                                                 : "=r" (ret), "=m" (*p)
299 +                                                 : "0" (1), "m" (*p)
300 +                                                 : "memory");
301 +        return ret;
302 + }
303 + #endif
304 +
305 + #ifdef __s390__
306 + #define HAVE_TEST_AND_SET 1
307 + static inline int testandset(volatile int *p)
308 + {
309 +        int ret;
310 +
311 +        __asm__ __volatile__("0: cs    %0,%1,0(%2)\n"
312 +                                                 "   jl    0b"
313 +                                                 : "=&d" (ret)
314 +                                                 : "r" (1), "a" (p), "0" (*p)
315 +                                                 : "cc", "memory" );
316 +        return ret;
317 + }
318 + #endif
319 +
320 + #ifdef __alpha__
321 + #define HAVE_TEST_AND_SET 1
322 + static inline int testandset(volatile int *p)
323 + {
324 +        int ret;
325 +        unsigned long one;
326 +
327 +        __asm__ __volatile__("0:        mov 1,%2\n"
328 +                                                 "      ldl_l %0,%1\n"
329 +                                                 "      stl_c %2,%1\n"
330 +                                                 "      beq %2,1f\n"
331 +                                                 ".subsection 2\n"
332 +                                                 "1:    br 0b\n"
333 +                                                 ".previous"
334 +                                                 : "=r" (ret), "=m" (*p), "=r" (one)
335 +                                                 : "m" (*p));
336 +        return ret;
337 + }
338 + #endif
339 +
340 + #ifdef __sparc__
341 + #define HAVE_TEST_AND_SET 1
342 + static inline int testandset(volatile int *p)
343 + {
344 +        int ret;
345 +
346 +        __asm__ __volatile__("ldstub    [%1], %0"
347 +                                                 : "=r" (ret)
348 +                                                 : "r" (p)
349 +                                                 : "memory");
350 +
351 +        return (ret ? 1 : 0);
352 + }
353 + #endif
354 +
355 + #ifdef __arm__
356 + #define HAVE_TEST_AND_SET 1
357 + static inline int testandset(volatile int *p)
358 + {
359 +        register unsigned int ret;
360 +        __asm__ __volatile__("swp %0, %1, [%2]"
361 +                                                 : "=r"(ret)
362 +                                                 : "0"(1), "r"(p));
363 +        
364 +        return ret;
365 + }
366 + #endif
367 +
368 + #endif /* __GNUC__ */
369 +
370 + typedef volatile int spinlock_t;
371 +
372 + static const spinlock_t SPIN_LOCK_UNLOCKED = 0;
373 +
374 + #if HAVE_TEST_AND_SET
375 + #define HAVE_SPINLOCKS 1
376 + static inline void spin_lock(spinlock_t *lock)
377 + {
378 +        while (testandset(lock));
379 + }
380 +
381 + static inline void spin_unlock(spinlock_t *lock)
382 + {
383 +        *lock = 0;
384 + }
385 +
386 + static inline int spin_trylock(spinlock_t *lock)
387 + {
388 +        return !testandset(lock);
389 + }
390 + #else
391 + static inline void spin_lock(spinlock_t *lock)
392 + {
393 + }
394 +
395 + static inline void spin_unlock(spinlock_t *lock)
396 + {
397 + }
398 +
399 + static inline int spin_trylock(spinlock_t *lock)
400 + {
401 +        return 1;
402 + }
403 + #endif
404 +
405   // Time data type for Time Manager emulation
406   #ifdef HAVE_CLOCK_GETTIME
407   typedef struct timespec tm_time_t;
# Line 122 | Line 409 | typedef struct timespec tm_time_t;
409   typedef struct timeval tm_time_t;
410   #endif
411  
412 + // Timing functions
413 + extern uint64 GetTicks_usec(void);
414 + extern void Delay_usec(uint32 usec);
415 +
416 + #if defined(HAVE_PTHREADS) || ((defined(__linux__) || defined(__NetBSD__)) && defined(__powerpc__))
417   // Setup pthread attributes
418   extern void Set_pthread_attr(pthread_attr_t *attr, int priority);
419 + #endif
420  
421   // Various definitions
422   typedef struct rgb_color {
# Line 133 | Line 426 | typedef struct rgb_color {
426          uint8           alpha;
427   } rgb_color;
428  
429 + // X11 display fast locks
430 + #ifdef HAVE_SPINLOCKS
431 + #define X11_LOCK_TYPE spinlock_t
432 + #define X11_LOCK_INIT SPIN_LOCK_UNLOCKED
433 + #define XDisplayLock() spin_lock(&x_display_lock)
434 + #define XDisplayUnlock() spin_unlock(&x_display_lock)
435 + #elif defined(HAVE_PTHREADS)
436 + #define X11_LOCK_TYPE pthread_mutex_t
437 + #define X11_LOCK_INIT PTHREAD_MUTEX_INITIALIZER
438 + #define XDisplayLock() pthread_mutex_lock(&x_display_lock);
439 + #define XDisplayUnlock() pthread_mutex_unlock(&x_display_lock);
440 + #else
441 + #define XDisplayLock()
442 + #define XDisplayUnlock()
443 + #endif
444 + #ifdef X11_LOCK_TYPE
445 + extern X11_LOCK_TYPE x_display_lock;
446 + #endif
447 +
448   // Macro for calling MacOS routines
449 < #define CallMacOS(type, tvect) call_macos((uint32)tvect)
450 < #define CallMacOS1(type, tvect, arg1) call_macos1((uint32)tvect, (uint32)arg1)
451 < #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uint32)tvect, (uint32)arg1, (uint32)arg2)
452 < #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3)
453 < #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4)
454 < #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5)
455 < #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6)
456 < #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uint32)tvect, (uint32)arg1, (uint32)arg2, (uint32)arg3, (uint32)arg4, (uint32)arg5, (uint32)arg6, (uint32)arg7)
449 > #define CallMacOS(type, tvect) call_macos((uintptr)tvect)
450 > #define CallMacOS1(type, tvect, arg1) call_macos1((uintptr)tvect, (uintptr)arg1)
451 > #define CallMacOS2(type, tvect, arg1, arg2) call_macos2((uintptr)tvect, (uintptr)arg1, (uintptr)arg2)
452 > #define CallMacOS3(type, tvect, arg1, arg2, arg3) call_macos3((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3)
453 > #define CallMacOS4(type, tvect, arg1, arg2, arg3, arg4) call_macos4((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4)
454 > #define CallMacOS5(type, tvect, arg1, arg2, arg3, arg4, arg5) call_macos5((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5)
455 > #define CallMacOS6(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6) call_macos6((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6)
456 > #define CallMacOS7(type, tvect, arg1, arg2, arg3, arg4, arg5, arg6, arg7) call_macos7((uintptr)tvect, (uintptr)arg1, (uintptr)arg2, (uintptr)arg3, (uintptr)arg4, (uintptr)arg5, (uintptr)arg6, (uintptr)arg7)
457  
458   #ifdef __cplusplus
459   extern "C" {

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines