2012-05-18 02:30:46 +02:00
|
|
|
/*
|
|
|
|
* Copyright © 2007 Chris Wilson
|
|
|
|
* Copyright © 2009,2010 Red Hat, Inc.
|
2012-05-18 02:50:38 +02:00
|
|
|
* Copyright © 2011,2012 Google, Inc.
|
2012-05-18 02:30:46 +02:00
|
|
|
*
|
|
|
|
* This is part of HarfBuzz, a text shaping library.
|
|
|
|
*
|
|
|
|
* Permission is hereby granted, without written agreement and without
|
|
|
|
* license or royalty fees, to use, copy, modify, and distribute this
|
|
|
|
* software and its documentation for any purpose, provided that the
|
|
|
|
* above copyright notice and the following two paragraphs appear in
|
|
|
|
* all copies of this software.
|
|
|
|
*
|
|
|
|
* IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR
|
|
|
|
* DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES
|
|
|
|
* ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN
|
|
|
|
* IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH
|
|
|
|
* DAMAGE.
|
|
|
|
*
|
|
|
|
* THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING,
|
|
|
|
* BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. THE SOFTWARE PROVIDED HEREUNDER IS
|
|
|
|
* ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO
|
|
|
|
* PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS.
|
|
|
|
*
|
|
|
|
* Contributor(s):
|
|
|
|
* Chris Wilson <chris@chris-wilson.co.uk>
|
|
|
|
* Red Hat Author(s): Behdad Esfahbod
|
|
|
|
* Google Author(s): Behdad Esfahbod
|
|
|
|
*/
|
|
|
|
|
2018-08-26 07:36:36 +02:00
|
|
|
#ifndef HB_ATOMIC_HH
|
|
|
|
#define HB_ATOMIC_HH
|
2012-05-18 02:30:46 +02:00
|
|
|
|
2018-08-26 07:36:36 +02:00
|
|
|
#include "hb.hh"
|
2018-12-31 00:49:34 +01:00
|
|
|
#include "hb-meta.hh"
|
2012-05-18 02:30:46 +02:00
|
|
|
|
|
|
|
|
2018-08-09 07:45:49 +02:00
|
|
|
/*
|
|
|
|
* Atomic integers and pointers.
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
2012-05-18 02:30:46 +02:00
|
|
|
/* We need external help for these */
|
|
|
|
|
2015-03-27 21:49:33 +01:00
|
|
|
#if defined(hb_atomic_int_impl_add) \
|
2018-08-01 04:29:49 +02:00
|
|
|
&& defined(hb_atomic_ptr_impl_get) \
|
2015-03-27 21:49:33 +01:00
|
|
|
&& defined(hb_atomic_ptr_impl_cmpexch)
|
2015-04-08 21:49:38 +02:00
|
|
|
|
2018-08-09 07:45:49 +02:00
|
|
|
/* Defined externally, i.e. in config.h. */
|
2012-05-18 02:30:46 +02:00
|
|
|
|
2012-05-18 03:53:24 +02:00
|
|
|
|
2018-09-26 22:37:18 +02:00
|
|
|
#elif !defined(HB_NO_MT) && defined(__ATOMIC_ACQUIRE)
|
2018-07-17 10:57:01 +02:00
|
|
|
|
2021-02-23 01:22:09 +01:00
|
|
|
/* C++11-style GCC primitives. We prefer these as they don't require linking to libstdc++ / libc++. */
|
2018-07-17 10:57:01 +02:00
|
|
|
|
2018-09-26 22:37:18 +02:00
|
|
|
#define _hb_memory_barrier() __sync_synchronize ()
|
|
|
|
|
2018-08-01 06:05:51 +02:00
|
|
|
#define hb_atomic_int_impl_add(AI, V) __atomic_fetch_add ((AI), (V), __ATOMIC_ACQ_REL)
|
|
|
|
#define hb_atomic_int_impl_set_relaxed(AI, V) __atomic_store_n ((AI), (V), __ATOMIC_RELAXED)
|
2018-09-26 22:37:18 +02:00
|
|
|
#define hb_atomic_int_impl_set(AI, V) __atomic_store_n ((AI), (V), __ATOMIC_RELEASE)
|
2018-08-01 06:05:51 +02:00
|
|
|
#define hb_atomic_int_impl_get_relaxed(AI) __atomic_load_n ((AI), __ATOMIC_RELAXED)
|
2018-09-26 22:37:18 +02:00
|
|
|
#define hb_atomic_int_impl_get(AI) __atomic_load_n ((AI), __ATOMIC_ACQUIRE)
|
2018-07-17 10:57:01 +02:00
|
|
|
|
2018-08-09 09:22:37 +02:00
|
|
|
#define hb_atomic_ptr_impl_set_relaxed(P, V) __atomic_store_n ((P), (V), __ATOMIC_RELAXED)
|
|
|
|
#define hb_atomic_ptr_impl_get_relaxed(P) __atomic_load_n ((P), __ATOMIC_RELAXED)
|
2018-09-26 22:37:18 +02:00
|
|
|
#define hb_atomic_ptr_impl_get(P) __atomic_load_n ((P), __ATOMIC_ACQUIRE)
|
2018-07-17 10:57:01 +02:00
|
|
|
static inline bool
|
|
|
|
_hb_atomic_ptr_impl_cmplexch (const void **P, const void *O_, const void *N)
|
|
|
|
{
|
|
|
|
const void *O = O_; // Need lvalue
|
|
|
|
return __atomic_compare_exchange_n ((void **) P, (void **) &O, (void *) N, true, __ATOMIC_ACQ_REL, __ATOMIC_RELAXED);
|
|
|
|
}
|
2018-08-01 23:13:59 +02:00
|
|
|
#define hb_atomic_ptr_impl_cmpexch(P,O,N) _hb_atomic_ptr_impl_cmplexch ((const void **) (P), (O), (N))
|
2018-07-17 10:57:01 +02:00
|
|
|
|
2021-02-23 01:23:53 +01:00
|
|
|
|
2021-02-23 01:22:09 +01:00
|
|
|
#elif !defined(HB_NO_MT)
|
2018-07-16 15:41:09 +02:00
|
|
|
|
2018-07-17 10:57:01 +02:00
|
|
|
/* C++11 atomics. */
|
2018-07-16 15:41:09 +02:00
|
|
|
|
|
|
|
#include <atomic>
|
|
|
|
|
2018-09-26 22:37:18 +02:00
|
|
|
#define _hb_memory_barrier() std::atomic_thread_fence(std::memory_order_ack_rel)
|
|
|
|
#define _hb_memory_r_barrier() std::atomic_thread_fence(std::memory_order_acquire)
|
|
|
|
#define _hb_memory_w_barrier() std::atomic_thread_fence(std::memory_order_release)
|
|
|
|
|
2018-08-01 06:05:51 +02:00
|
|
|
#define hb_atomic_int_impl_add(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->fetch_add ((V), std::memory_order_acq_rel))
|
|
|
|
#define hb_atomic_int_impl_set_relaxed(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_relaxed))
|
2018-09-27 23:20:26 +02:00
|
|
|
#define hb_atomic_int_impl_set(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std::memory_order_release))
|
2019-03-05 17:18:57 +01:00
|
|
|
#define hb_atomic_int_impl_get_relaxed(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_relaxed))
|
|
|
|
#define hb_atomic_int_impl_get(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::memory_order_acquire))
|
2018-07-16 15:41:09 +02:00
|
|
|
|
2018-08-09 09:22:37 +02:00
|
|
|
#define hb_atomic_ptr_impl_set_relaxed(P, V) (reinterpret_cast<std::atomic<void*> *> (P)->store ((V), std::memory_order_relaxed))
|
2019-03-05 17:18:57 +01:00
|
|
|
#define hb_atomic_ptr_impl_get_relaxed(P) (reinterpret_cast<std::atomic<void*> const *> (P)->load (std::memory_order_relaxed))
|
2018-09-26 22:37:18 +02:00
|
|
|
#define hb_atomic_ptr_impl_get(P) (reinterpret_cast<std::atomic<void*> *> (P)->load (std::memory_order_acquire))
|
2018-07-16 15:41:09 +02:00
|
|
|
static inline bool
|
|
|
|
_hb_atomic_ptr_impl_cmplexch (const void **P, const void *O_, const void *N)
|
|
|
|
{
|
|
|
|
const void *O = O_; // Need lvalue
|
2018-07-17 10:57:01 +02:00
|
|
|
return reinterpret_cast<std::atomic<const void*> *> (P)->compare_exchange_weak (O, N, std::memory_order_acq_rel, std::memory_order_relaxed);
|
2018-07-16 15:41:09 +02:00
|
|
|
}
|
2018-08-01 23:13:59 +02:00
|
|
|
#define hb_atomic_ptr_impl_cmpexch(P,O,N) _hb_atomic_ptr_impl_cmplexch ((const void **) (P), (O), (N))
|
2018-07-16 15:41:09 +02:00
|
|
|
|
|
|
|
|
2021-09-23 00:21:27 +02:00
|
|
|
#else /* defined(HB_NO_MT) */
|
2012-06-05 22:34:49 +02:00
|
|
|
|
2019-10-23 00:03:47 +02:00
|
|
|
#define hb_atomic_int_impl_add(AI, V) ((*(AI) += (V)) - (V))
|
2019-05-25 18:59:01 +02:00
|
|
|
#define _hb_memory_barrier() do {} while (0)
|
2018-08-09 07:45:49 +02:00
|
|
|
#define hb_atomic_ptr_impl_cmpexch(P,O,N) (* (void **) (P) == (void *) (O) ? (* (void **) (P) = (void *) (N), true) : false)
|
2012-06-05 23:27:20 +02:00
|
|
|
|
2012-05-18 02:30:46 +02:00
|
|
|
#endif
|
|
|
|
|
2015-03-27 21:49:33 +01:00
|
|
|
|
2022-07-18 00:47:16 +02:00
|
|
|
#ifndef _hb_compiler_memory_r_barrier
|
2022-07-18 01:03:40 +02:00
|
|
|
/* This we always use std::atomic for; and should never be disabled...
|
|
|
|
* except that MSVC gives me an internal compiler error on it. */
|
|
|
|
#if !defined(_MSC_VER)
|
|
|
|
#include <atomic>
|
2022-07-18 00:47:16 +02:00
|
|
|
#define _hb_compiler_memory_r_barrier() std::atomic_signal_fence (std::memory_order_acquire)
|
2022-07-18 01:03:40 +02:00
|
|
|
#else
|
|
|
|
#define _hb_compiler_memory_r_barrier() do {} while (0)
|
2022-07-18 00:47:16 +02:00
|
|
|
#endif
|
2022-07-18 01:03:40 +02:00
|
|
|
#endif
|
|
|
|
|
2022-07-18 00:47:16 +02:00
|
|
|
|
|
|
|
|
2018-08-01 22:59:31 +02:00
|
|
|
#ifndef _hb_memory_r_barrier
|
|
|
|
#define _hb_memory_r_barrier() _hb_memory_barrier ()
|
|
|
|
#endif
|
|
|
|
#ifndef _hb_memory_w_barrier
|
|
|
|
#define _hb_memory_w_barrier() _hb_memory_barrier ()
|
|
|
|
#endif
|
2018-08-01 06:05:51 +02:00
|
|
|
#ifndef hb_atomic_int_impl_set_relaxed
|
2018-08-01 07:21:21 +02:00
|
|
|
#define hb_atomic_int_impl_set_relaxed(AI, V) (*(AI) = (V))
|
2018-08-01 06:05:51 +02:00
|
|
|
#endif
|
|
|
|
#ifndef hb_atomic_int_impl_get_relaxed
|
2018-08-01 07:21:21 +02:00
|
|
|
#define hb_atomic_int_impl_get_relaxed(AI) (*(AI))
|
2018-08-01 06:05:51 +02:00
|
|
|
#endif
|
2018-08-09 09:22:37 +02:00
|
|
|
|
|
|
|
#ifndef hb_atomic_ptr_impl_set_relaxed
|
|
|
|
#define hb_atomic_ptr_impl_set_relaxed(P, V) (*(P) = (V))
|
|
|
|
#endif
|
|
|
|
#ifndef hb_atomic_ptr_impl_get_relaxed
|
|
|
|
#define hb_atomic_ptr_impl_get_relaxed(P) (*(P))
|
|
|
|
#endif
|
2018-09-27 23:20:26 +02:00
|
|
|
#ifndef hb_atomic_int_impl_set
|
2018-09-28 15:13:14 +02:00
|
|
|
inline void hb_atomic_int_impl_set (int *AI, int v) { _hb_memory_w_barrier (); *AI = v; }
|
2018-09-27 23:20:26 +02:00
|
|
|
#endif
|
2018-09-26 22:37:18 +02:00
|
|
|
#ifndef hb_atomic_int_impl_get
|
2018-11-13 19:01:13 +01:00
|
|
|
inline int hb_atomic_int_impl_get (const int *AI) { int v = *AI; _hb_memory_r_barrier (); return v; }
|
2018-09-26 22:37:18 +02:00
|
|
|
#endif
|
2018-08-01 07:51:38 +02:00
|
|
|
#ifndef hb_atomic_ptr_impl_get
|
2018-11-13 19:01:13 +01:00
|
|
|
inline void *hb_atomic_ptr_impl_get (void ** const P) { void *v = *P; _hb_memory_r_barrier (); return v; }
|
2018-08-01 07:51:38 +02:00
|
|
|
#endif
|
2018-08-01 06:05:51 +02:00
|
|
|
|
2015-03-27 21:49:33 +01:00
|
|
|
|
|
|
|
struct hb_atomic_int_t
|
|
|
|
{
|
2020-06-29 12:51:09 +02:00
|
|
|
hb_atomic_int_t () = default;
|
|
|
|
constexpr hb_atomic_int_t (int v) : v (v) {}
|
|
|
|
|
2022-08-03 20:54:03 +02:00
|
|
|
hb_atomic_int_t& operator = (int v_) { set_relaxed (v_); return *this; }
|
|
|
|
operator int () const { return get_relaxed (); }
|
|
|
|
|
2018-12-16 20:08:10 +01:00
|
|
|
void set_relaxed (int v_) { hb_atomic_int_impl_set_relaxed (&v, v_); }
|
2022-08-03 20:37:41 +02:00
|
|
|
void set_release (int v_) { hb_atomic_int_impl_set (&v, v_); }
|
2018-12-17 19:01:01 +01:00
|
|
|
int get_relaxed () const { return hb_atomic_int_impl_get_relaxed (&v); }
|
2022-08-03 20:37:41 +02:00
|
|
|
int get_acquire () const { return hb_atomic_int_impl_get (&v); }
|
2018-12-17 19:01:01 +01:00
|
|
|
int inc () { return hb_atomic_int_impl_add (&v, 1); }
|
|
|
|
int dec () { return hb_atomic_int_impl_add (&v, -1); }
|
2018-08-09 07:51:35 +02:00
|
|
|
|
2020-06-29 12:51:09 +02:00
|
|
|
int v = 0;
|
2015-03-27 21:49:33 +01:00
|
|
|
};
|
|
|
|
|
2018-08-09 09:22:37 +02:00
|
|
|
template <typename P>
|
|
|
|
struct hb_atomic_ptr_t
|
|
|
|
{
|
2019-04-16 22:45:53 +02:00
|
|
|
typedef hb_remove_pointer<P> T;
|
2018-08-09 09:22:37 +02:00
|
|
|
|
2020-06-29 12:51:09 +02:00
|
|
|
hb_atomic_ptr_t () = default;
|
|
|
|
constexpr hb_atomic_ptr_t (T* v) : v (v) {}
|
|
|
|
|
2018-12-16 20:08:10 +01:00
|
|
|
void init (T* v_ = nullptr) { set_relaxed (v_); }
|
|
|
|
void set_relaxed (T* v_) { hb_atomic_ptr_impl_set_relaxed (&v, v_); }
|
2018-12-17 19:01:01 +01:00
|
|
|
T *get_relaxed () const { return (T *) hb_atomic_ptr_impl_get_relaxed (&v); }
|
2022-08-03 20:43:28 +02:00
|
|
|
T *get_acquire () const { return (T *) hb_atomic_ptr_impl_get ((void **) &v); }
|
2018-12-16 20:08:10 +01:00
|
|
|
bool cmpexch (const T *old, T *new_) const { return hb_atomic_ptr_impl_cmpexch ((void **) &v, (void *) old, (void *) new_); }
|
2018-08-09 09:22:37 +02:00
|
|
|
|
2022-08-03 20:43:28 +02:00
|
|
|
T * operator -> () const { return get_acquire (); }
|
|
|
|
template <typename C> operator C * () const { return get_acquire (); }
|
2018-11-05 19:23:54 +01:00
|
|
|
|
2020-06-29 12:51:09 +02:00
|
|
|
T *v = nullptr;
|
2018-08-09 09:22:37 +02:00
|
|
|
};
|
2015-03-27 21:49:33 +01:00
|
|
|
|
2012-05-18 02:30:46 +02:00
|
|
|
|
2018-08-26 07:36:36 +02:00
|
|
|
#endif /* HB_ATOMIC_HH */
|