1 /*
   2  * Copyright © 2007  Chris Wilson
   3  * Copyright © 2009,2010  Red Hat, Inc.
   4  * Copyright © 2011,2012  Google, Inc.
   5  *
   6  *  This is part of HarfBuzz, a text shaping library.
   7  *
   8  * Permission is hereby granted, without written agreement and without
   9  * license or royalty fees, to use, copy, modify, and distribute this
  10  * software and its documentation for any purpose, provided that the
  11  * above copyright notice and the following two paragraphs appear in
  12  * all copies of this software.
  13  *
  14  * IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR
  15  * DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES
  16  * ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN
  17  * IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH
  18  * DAMAGE.
  19  *
  20  * THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING,
  21  * BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
  22  * FITNESS FOR A PARTICULAR PURPOSE.  THE SOFTWARE PROVIDED HEREUNDER IS
  23  * ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO
  24  * PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS.
  25  *
  26  * Contributor(s):
  27  *      Chris Wilson <chris@chris-wilson.co.uk>
  28  * Red Hat Author(s): Behdad Esfahbod
  29  * Google Author(s): Behdad Esfahbod
  30  */
  31 
  32 #ifndef HB_ATOMIC_PRIVATE_HH
  33 #define HB_ATOMIC_PRIVATE_HH
  34 
  35 #include "hb-private.hh"
  36 
  37 
  38 /* atomic_int */
  39 
  40 /* We need external help for these */
  41 
  42 #if defined(hb_atomic_int_impl_add) \
  43  && defined(hb_atomic_ptr_impl_get) \
  44  && defined(hb_atomic_ptr_impl_cmpexch)
  45 
  46 /* Defined externally, i.e. in config.h; must have typedef'ed hb_atomic_int_impl_t as well. */
  47 
  48 
  49 #elif !defined(HB_NO_MT) && (defined(_WIN32) || defined(__CYGWIN__))
  50 
  51 #include <windows.h>
  52 
  53 /* MinGW has a convoluted history of supporting MemoryBarrier
  54  * properly.  As such, define a function to wrap the whole
  55  * thing. */
  56 static inline void _HBMemoryBarrier (void) {
  57 #if !defined(MemoryBarrier)
  58   long dummy = 0;
  59   InterlockedExchange (&dummy, 1);
  60 #else
  61   MemoryBarrier ();
  62 #endif
  63 }
  64 
  65 typedef LONG hb_atomic_int_impl_t;
  66 #define hb_atomic_int_impl_add(AI, V)           InterlockedExchangeAdd (&(AI), (V))
  67 
  68 #define hb_atomic_ptr_impl_get(P)               (_HBMemoryBarrier (), (void *) *(P))
  69 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       (InterlockedCompareExchangePointer ((void **) (P), (void *) (N), (void *) (O)) == (void *) (O))
  70 
  71 
  72 #elif !defined(HB_NO_MT) && defined(HAVE_INTEL_ATOMIC_PRIMITIVES)
  73 
  74 typedef int hb_atomic_int_impl_t;
  75 #define hb_atomic_int_impl_add(AI, V)           __sync_fetch_and_add (&(AI), (V))
  76 
  77 #define hb_atomic_ptr_impl_get(P)               (void *) (__sync_synchronize (), *(P))
  78 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       __sync_bool_compare_and_swap ((P), (O), (N))
  79 
  80 
  81 #elif !defined(HB_NO_MT) && defined(HAVE_SOLARIS_ATOMIC_OPS)
  82 
  83 #include <atomic.h>
  84 #include <mbarrier.h>
  85 
  86 typedef unsigned int hb_atomic_int_impl_t;
  87 #define hb_atomic_int_impl_add(AI, V)           ( ({__machine_rw_barrier ();}), atomic_add_int_nv (&(AI), (V)) - (V))
  88 
  89 #define hb_atomic_ptr_impl_get(P)               ( ({__machine_rw_barrier ();}), (void *) *(P))
  90 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       ( ({__machine_rw_barrier ();}), atomic_cas_ptr ((void **) (P), (void *) (O), (void *) (N)) == (void *) (O) ? true : false)
  91 
  92 
  93 #elif !defined(HB_NO_MT) && defined(__APPLE__)
  94 
  95 #include <libkern/OSAtomic.h>
  96 #ifdef __MAC_OS_X_MIN_REQUIRED
  97 #include <AvailabilityMacros.h>
  98 #elif defined(__IPHONE_OS_MIN_REQUIRED)
  99 #include <Availability.h>
 100 #endif
 101 
 102 
 103 typedef int32_t hb_atomic_int_impl_t;
 104 #define hb_atomic_int_impl_add(AI, V)           (OSAtomicAdd32Barrier ((V), &(AI)) - (V))
 105 
 106 #define hb_atomic_ptr_impl_get(P)               (OSMemoryBarrier (), (void *) *(P))
 107 #if (MAC_OS_X_VERSION_MIN_REQUIRED > MAC_OS_X_VERSION_10_4 || __IPHONE_VERSION_MIN_REQUIRED >= 20100)
 108 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       OSAtomicCompareAndSwapPtrBarrier ((void *) (O), (void *) (N), (void **) (P))
 109 #else
 110 #if __ppc64__ || __x86_64__ || __aarch64__
 111 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       OSAtomicCompareAndSwap64Barrier ((int64_t) (void *) (O), (int64_t) (void *) (N), (int64_t*) (P))
 112 #else
 113 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       OSAtomicCompareAndSwap32Barrier ((int32_t) (void *) (O), (int32_t) (void *) (N), (int32_t*) (P))
 114 #endif
 115 #endif
 116 
 117 
 118 #elif !defined(HB_NO_MT) && defined(_AIX) && defined(__IBMCPP__)
 119 
 120 #include <builtins.h>
 121 
 122 
 123 static inline int _hb_fetch_and_add(volatile int* AI, unsigned int V) {
 124   __lwsync();
 125   int result = __fetch_and_add(AI, V);
 126   __isync();
 127   return result;
 128 }
 129 static inline int _hb_compare_and_swaplp(volatile long* P, long O, long N) {
 130   __sync();
 131   int result = __compare_and_swaplp (P, &O, N);
 132   __sync();
 133   return result;
 134 }
 135 
 136 typedef int hb_atomic_int_impl_t;
 137 #define hb_atomic_int_impl_add(AI, V)           _hb_fetch_and_add (&(AI), (V))
 138 
 139 #define hb_atomic_ptr_impl_get(P)               (__sync(), (void *) *(P))
 140 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       _hb_compare_and_swaplp ((long*)(P), (long)(O), (long)(N))
 141 
 142 #elif !defined(HB_NO_MT)
 143 
 144 #define HB_ATOMIC_INT_NIL 1 /* Warn that fallback implementation is in use. */
 145 
 146 typedef volatile int hb_atomic_int_impl_t;
 147 #define hb_atomic_int_impl_add(AI, V)           (((AI) += (V)) - (V))
 148 
 149 #define hb_atomic_ptr_impl_get(P)               ((void *) *(P))
 150 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       (* (void * volatile *) (P) == (void *) (O) ? (* (void * volatile *) (P) = (void *) (N), true) : false)
 151 
 152 
 153 #else /* HB_NO_MT */
 154 
 155 typedef int hb_atomic_int_impl_t;
 156 #define hb_atomic_int_impl_add(AI, V)           (((AI) += (V)) - (V))
 157 
 158 #define hb_atomic_ptr_impl_get(P)               ((void *) *(P))
 159 #define hb_atomic_ptr_impl_cmpexch(P,O,N)       (* (void **) (P) == (void *) (O) ? (* (void **) (P) = (void *) (N), true) : false)
 160 
 161 
 162 #endif
 163 
 164 
 165 #define HB_ATOMIC_INT_INIT(V)          {V}
 166 
 167 struct hb_atomic_int_t
 168 {
 169   hb_atomic_int_impl_t v;
 170 
 171   inline void set_unsafe (int v_) { v = v_; }
 172   inline int get_unsafe (void) const { return v; }
 173   inline int inc (void) { return hb_atomic_int_impl_add (const_cast<hb_atomic_int_impl_t &> (v),  1); }
 174   inline int dec (void) { return hb_atomic_int_impl_add (const_cast<hb_atomic_int_impl_t &> (v), -1); }
 175 };
 176 
 177 
 178 #define hb_atomic_ptr_get(P) hb_atomic_ptr_impl_get(P)
 179 #define hb_atomic_ptr_cmpexch(P,O,N) hb_atomic_ptr_impl_cmpexch((P),(O),(N))
 180 
 181 
 182 #endif /* HB_ATOMIC_PRIVATE_HH */