diff --git a/boehm-gc/ChangeLog b/boehm-gc/ChangeLog index dd1d4a8888e910579334fa1e91ef05341a42998a..fc5cbe746a9050c9540f7939919bc04acf7acd5c 100644 --- a/boehm-gc/ChangeLog +++ b/boehm-gc/ChangeLog @@ -1,3 +1,11 @@ +2003-03-22 Richard Henderson <rth@redhat.com> + + * include/private/gc_locks.h [IA64]: Include ia64intrin.h. + (GC_test_and_set): Use __sync_lock_test_and_set. + (GC_clear): Use volatile assignment. + (GC_compare_and_exchange): Use __sync_bool_compare_and_swap. + (GC_memory_write_barrier): Use __sync_synchronize. + 2003-03-12 Andreas Schwab <schwab@suse.de> * configure.in: Avoid trailing /. in toolexeclibdir. diff --git a/boehm-gc/include/private/gc_locks.h b/boehm-gc/include/private/gc_locks.h index 33101f8c2d02726aa03cdc1c13eda08f61777b85..9b91ada917a5ca0f6e7cf487c0ff9932d7bfded1 100644 --- a/boehm-gc/include/private/gc_locks.h +++ b/boehm-gc/include/private/gc_locks.h @@ -100,17 +100,13 @@ # define GC_TEST_AND_SET_DEFINED # endif # if defined(IA64) +# include <ia64intrin.h> inline static int GC_test_and_set(volatile unsigned int *addr) { - long oldval, n = 1; - __asm__ __volatile__("xchg4 %0=%1,%2" - : "=r"(oldval), "=m"(*addr) - : "r"(n), "1"(*addr) : "memory"); - return oldval; + return __sync_lock_test_and_set(addr, 1); } # define GC_TEST_AND_SET_DEFINED - /* Should this handle post-increment addressing?? */ inline static void GC_clear(volatile unsigned int *addr) { - __asm__ __volatile__("st4.rel %0=r0" : "=m" (*addr) : : "memory"); + *addr = 0; } # define GC_CLEAR_DEFINED # endif @@ -326,20 +322,17 @@ # if defined(IA64) # if !defined(GENERIC_COMPARE_AND_SWAP) inline static GC_bool GC_compare_and_exchange(volatile GC_word *addr, - GC_word old, GC_word new_val) + GC_word old, + GC_word new_val) { - unsigned long oldval; - __asm__ __volatile__("mov ar.ccv=%4 ;; cmpxchg8.rel %0=%1,%2,ar.ccv" - : "=r"(oldval), "=m"(*addr) - : "r"(new_val), "1"(*addr), "r"(old) : "memory"); - return (oldval == old); + return __sync_bool_compare_and_swap (addr, old, new_val); } # endif /* !GENERIC_COMPARE_AND_SWAP */ # if 0 /* Shouldn't be needed; we use volatile stores instead. */ inline static void GC_memory_write_barrier() { - __asm__ __volatile__("mf" : : : "memory"); + __sync_synchronize (); } # endif /* 0 */ # endif /* IA64 */