Rizin
unix-like reverse engineering framework and cli tools
atomic-ops.h
Go to the documentation of this file.
1 /* Copyright (c) 2013, Ben Noordhuis <info@bnoordhuis.nl>
2  *
3  * Permission to use, copy, modify, and/or distribute this software for any
4  * purpose with or without fee is hereby granted, provided that the above
5  * copyright notice and this permission notice appear in all copies.
6  *
7  * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
8  * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
9  * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
10  * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
11  * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
12  * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
13  * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
14  */
15 
16 #ifndef UV_ATOMIC_OPS_H_
17 #define UV_ATOMIC_OPS_H_
18 
19 #include "internal.h" /* UV_UNUSED */
20 
21 #if defined(__SUNPRO_C) || defined(__SUNPRO_CC)
22 #include <atomic.h>
23 #endif
24 
25 UV_UNUSED(static int cmpxchgi(int* ptr, int oldval, int newval));
26 UV_UNUSED(static void cpu_relax(void));
27 
28 /* Prefer hand-rolled assembly over the gcc builtins because the latter also
29  * issue full memory barriers.
30  */
31 UV_UNUSED(static int cmpxchgi(int* ptr, int oldval, int newval)) {
32 #if defined(__i386__) || defined(__x86_64__)
33  int out;
34  __asm__ __volatile__ ("lock; cmpxchg %2, %1;"
35  : "=a" (out), "+m" (*(volatile int*) ptr)
36  : "r" (newval), "0" (oldval)
37  : "memory");
38  return out;
39 #elif defined(__MVS__)
40  unsigned int op4;
41  if (__plo_CSST(ptr, (unsigned int*) &oldval, newval,
42  (unsigned int*) ptr, *ptr, &op4))
43  return oldval;
44  else
45  return op4;
46 #elif defined(__SUNPRO_C) || defined(__SUNPRO_CC)
47  return atomic_cas_uint((uint_t *)ptr, (uint_t)oldval, (uint_t)newval);
48 #else
49  return __sync_val_compare_and_swap(ptr, oldval, newval);
50 #endif
51 }
52 
53 UV_UNUSED(static void cpu_relax(void)) {
54 #if defined(__i386__) || defined(__x86_64__)
55  __asm__ __volatile__ ("rep; nop"); /* a.k.a. PAUSE */
56 #elif (defined(__arm__) && __ARM_ARCH >= 7) || defined(__aarch64__)
57  __asm__ volatile("yield");
58 #endif
59 }
60 
61 #endif /* UV_ATOMIC_OPS_H_ */
UV_UNUSED(static int cmpxchgi(int *ptr, int oldval, int newval))
Definition: atomic-ops.h:31
const lzma_allocator const uint8_t size_t uint8_t * out
Definition: block.h:528