|
typedef unsigned char | uchar |
|
typedef unsigned short | ushort |
|
#define | KMP_ATOMIC_VOLATILE volatile |
|
#define | KMP_CHECK_GTID |
|
#define | ATOMIC_BEGIN(TYPE_ID, OP_ID, TYPE, RET_TYPE) |
|
#define | ATOMIC_LOCK0 __kmp_atomic_lock |
|
#define | ATOMIC_LOCK1i __kmp_atomic_lock_1i |
|
#define | ATOMIC_LOCK2i __kmp_atomic_lock_2i |
|
#define | ATOMIC_LOCK4i __kmp_atomic_lock_4i |
|
#define | ATOMIC_LOCK4r __kmp_atomic_lock_4r |
|
#define | ATOMIC_LOCK8i __kmp_atomic_lock_8i |
|
#define | ATOMIC_LOCK8r __kmp_atomic_lock_8r |
|
#define | ATOMIC_LOCK8c __kmp_atomic_lock_8c |
|
#define | ATOMIC_LOCK10r __kmp_atomic_lock_10r |
|
#define | ATOMIC_LOCK16r __kmp_atomic_lock_16r |
|
#define | ATOMIC_LOCK16c __kmp_atomic_lock_16c |
|
#define | ATOMIC_LOCK20c __kmp_atomic_lock_20c |
|
#define | ATOMIC_LOCK32c __kmp_atomic_lock_32c |
|
#define | OP_CRITICAL(OP, LCK_ID) |
|
#define | OP_UPDATE_CRITICAL(TYPE, OP, LCK_ID) |
|
#define | OP_GOMP_CRITICAL(OP, FLAG) |
|
#define | OP_UPDATE_GOMP_CRITICAL(TYPE, OP, FLAG) |
|
#define | KMP_DO_PAUSE |
|
#define | OP_CMPXCHG(TYPE, BITS, OP) |
|
#define | ATOMIC_FIXED_ADD(TYPE_ID, OP_ID, TYPE, BITS, OP, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_CMPXCHG(TYPE_ID, OP_ID, TYPE, BITS, OP, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_CRIT_L(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | ATOMIC_CMPX_L(TYPE_ID, OP_ID, TYPE, BITS, OP, LCK_ID, MASK, GOMP_FLAG) |
|
#define | MIN_MAX_CRITSECT(OP, LCK_ID) |
|
#define | GOMP_MIN_MAX_CRITSECT(OP, FLAG) |
|
#define | MIN_MAX_CMPXCHG(TYPE, BITS, OP) |
|
#define | MIN_MAX_CRITICAL(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | MIN_MAX_COMPXCHG(TYPE_ID, OP_ID, TYPE, BITS, OP, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_CRIT_EQV(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | ATOMIC_CMPX_EQV(TYPE_ID, OP_ID, TYPE, BITS, OP, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_CRITICAL(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | ATOMIC_BEGIN_MIX(TYPE_ID, TYPE, OP_ID, RTYPE_ID, RTYPE) |
|
#define | ATOMIC_CRITICAL_FP(TYPE_ID, TYPE, OP_ID, OP, RTYPE_ID, RTYPE, LCK_ID, GOMP_FLAG) |
|
#define | ATOMIC_CMPXCHG_MIX(TYPE_ID, TYPE, OP_ID, BITS, OP, RTYPE_ID, RTYPE, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_CMPXCHG_CMPLX(TYPE_ID, TYPE, OP_ID, BITS, OP, RTYPE_ID, RTYPE, LCK_ID, MASK, GOMP_FLAG) |
|
#define | ATOMIC_BEGIN_READ(TYPE_ID, OP_ID, TYPE, RET_TYPE) |
|
#define | OP_CMPXCHG_READ(TYPE, BITS, OP) |
|
#define | OP_CRITICAL_READ(OP, LCK_ID) |
|
#define | OP_GOMP_CRITICAL_READ(OP, FLAG) |
|
#define | ATOMIC_FIXED_READ(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | ATOMIC_CMPXCHG_READ(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | ATOMIC_CRITICAL_READ(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | ATOMIC_BEGIN_CPT(TYPE_ID, OP_ID, TYPE, RET_TYPE) |
|
#define | OP_CRITICAL_CPT(OP, LCK_ID) |
|
#define | OP_UPDATE_CRITICAL_CPT(TYPE, OP, LCK_ID) |
|
#define | OP_GOMP_CRITICAL_CPT(TYPE, OP, FLAG) |
|
#define | OP_CMPXCHG_CPT(TYPE, BITS, OP) |
|
#define | ATOMIC_CMPXCHG_CPT(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | ATOMIC_FIXED_ADD_CPT(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | OP_CRITICAL_L_CPT(OP, LCK_ID) |
|
#define | OP_GOMP_CRITICAL_L_CPT(OP, FLAG) |
|
#define | ATOMIC_CMPX_L_CPT(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | MIN_MAX_CRITSECT_CPT(OP, LCK_ID) |
|
#define | GOMP_MIN_MAX_CRITSECT_CPT(OP, FLAG) |
|
#define | MIN_MAX_CMPXCHG_CPT(TYPE, BITS, OP) |
|
#define | MIN_MAX_CRITICAL_CPT(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | MIN_MAX_COMPXCHG_CPT(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | OP_GOMP_CRITICAL_EQV_CPT(OP, FLAG) |
|
#define | ATOMIC_CMPX_EQV_CPT(TYPE_ID, OP_ID, TYPE, BITS, OP, GOMP_FLAG) |
|
#define | ATOMIC_CRITICAL_CPT(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
#define | OP_CRITICAL_CPT_WRK(OP, LCK_ID) |
|
#define | OP_GOMP_CRITICAL_CPT_WRK(OP, FLAG) |
|
#define | ATOMIC_BEGIN_WRK(TYPE_ID, OP_ID, TYPE) |
|
#define | ATOMIC_CRITICAL_CPT_WRK(TYPE_ID, OP_ID, TYPE, OP, LCK_ID, GOMP_FLAG) |
|
int | __kmp_atomic_mode = 1 |
|
kmp_atomic_lock_t | __kmp_atomic_lock |
|
kmp_atomic_lock_t | __kmp_atomic_lock_1i |
|
kmp_atomic_lock_t | __kmp_atomic_lock_2i |
|
kmp_atomic_lock_t | __kmp_atomic_lock_4i |
|
kmp_atomic_lock_t | __kmp_atomic_lock_4r |
|
kmp_atomic_lock_t | __kmp_atomic_lock_8i |
|
kmp_atomic_lock_t | __kmp_atomic_lock_8r |
|
kmp_atomic_lock_t | __kmp_atomic_lock_8c |
|
kmp_atomic_lock_t | __kmp_atomic_lock_10r |
|
kmp_atomic_lock_t | __kmp_atomic_lock_16r |
|
kmp_atomic_lock_t | __kmp_atomic_lock_16c |
|
kmp_atomic_lock_t | __kmp_atomic_lock_20c |
|
kmp_atomic_lock_t | __kmp_atomic_lock_32c |
|
| sub |
|
| kmp_int32 |
|
| kmp_real32 |
|
KMP_ARCH_X86 | kmp_int64 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | kmp_real64 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | andb |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | kmp_int8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | div |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | kmp_uint8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | orb |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | shr |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | add |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | kmp_int16 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | mul |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | shl |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 | kmp_uint32 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | xor |
|
| orl |
|
| char |
|
KMP_ARCH_X86 | short |
|
| min |
|
| neqv |
|
KMP_ARCH_X86 KMP_ARCH_X86 | eqv |
|
| float8 |
|
| kmp_cmplx32 |
|
| cmplx8 |
|
| kmp_cmplx64 |
|
| rd |
|
KMP_ARCH_X86 KMP_ARCH_X86 long | double |
|
KMP_ARCH_X86 KMP_ARCH_X86 long | wr |
|
KMP_ARCH_X86 KMP_ARCH_X86 long KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 long | kmp_cmplx80 |
|
| sub_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | andb_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | div_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | orb_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | shr_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | add_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | mul_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | shl_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | xor_cpt |
|
| orl_cpt |
|
| min_cpt |
|
| neqv_cpt |
|
KMP_ARCH_X86 KMP_ARCH_X86 | eqv_cpt |
|
| ATOMIC_FIXED_ADD (fixed4, add, kmp_int32, 32,+, 4i, 3, 0) ATOMIC_FIXED_ADD(fixed4 |
|
| ATOMIC_CMPXCHG (float4, add, kmp_real32, 32,+, 4r, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(float4 |
|
KMP_ARCH_X86 | ATOMIC_FIXED_ADD (fixed8, add, kmp_int64, 64,+, 8i, 7, KMP_ARCH_X86) ATOMIC_FIXED_ADD(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (float8, add, kmp_real64, 64,+, 8r, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(float8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1, add, kmp_int8, 8,+, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1, div, kmp_int8, 8,/, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed1u |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1, mul, kmp_int8, 8, *, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1, shl, kmp_int8, 8,<<, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1u, shr, kmp_uint8, 8, > >, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed1, xor, kmp_int8, 8, ^, 1i, 0, 0) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed2, andb, kmp_int16, 16, &, 2i, 1, 0) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed2u, div, kmp_uint16, 16,/, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed2, orb, kmp_int16, 16,|, 2i, 1, 0) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed2u, shr, kmp_uint16, 16, > >, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed2, xor, kmp_int16, 16, ^, 2i, 1, 0) ATOMIC_CMPXCHG(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed4, div, kmp_int32, 32,/, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed4u |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed4, mul, kmp_int32, 32, *, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed4, shl, kmp_int32, 32,<<, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed4u, shr, kmp_uint32, 32, > >, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed8, andb, kmp_int64, 64, &, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed8u, div, kmp_uint64, 64,/, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed8, orb, kmp_int64, 64,|, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8, shr, kmp_int64, 64, > KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed8u, shr, kmp_uint64, 64, > >, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8, shr, kmp_int64, 64, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (float4, div, kmp_real32, 32,/, 4r, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2, shr, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8, shr, kmp_int64, 64, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG (float8, div, kmp_real64, 64,/, 8r, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG(float8 |
|
| ATOMIC_CMPX_L (fixed1, andl, char, 8, &&, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPX_L(fixed1 |
|
KMP_ARCH_X86 | ATOMIC_CMPX_L (fixed2, andl, short, 16, &&, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPX_L(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_L (fixed4, andl, kmp_int32, 32, &&, 4i, 3, 0) ATOMIC_CMPX_L(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_L (fixed8, andl, kmp_int64, 64, &&, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPX_L(fixed8 |
|
| MIN_MAX_COMPXCHG (fixed1, max, char, 8,<, 1i, 0, KMP_ARCH_X86) MIN_MAX_COMPXCHG(fixed1 |
|
KMP_ARCH_X86 | MIN_MAX_COMPXCHG (fixed2, max, short, 16,<, 2i, 1, KMP_ARCH_X86) MIN_MAX_COMPXCHG(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG (fixed4, max, kmp_int32, 32,<, 4i, 3, 0) MIN_MAX_COMPXCHG(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG (fixed8, max, kmp_int64, 64,<, 8i, 7, KMP_ARCH_X86) MIN_MAX_COMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG (float4, max, kmp_real32, 32,<, 4r, 3, KMP_ARCH_X86) MIN_MAX_COMPXCHG(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG (float8, max, kmp_real64, 64,<, 8r, 7, KMP_ARCH_X86) MIN_MAX_COMPXCHG(float8 |
|
| ATOMIC_CMPXCHG (fixed1, neqv, kmp_int8, 8, ^, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed2 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG (fixed4, neqv, kmp_int32, 32, ^, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_EQV (fixed1, eqv, kmp_int8, 8, ^~, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPX_EQV(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_EQV (fixed4, eqv, kmp_int32, 32, ^~, 4i, 3, KMP_ARCH_X86) ATOMIC_CMPX_EQV(fixed8 |
|
| ATOMIC_CMPXCHG_MIX (fixed1, char, mul, 8, *, float8, kmp_real64, 1i, 0, KMP_ARCH_X86) ATOMIC_CMPXCHG_MIX(fixed1 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG_MIX (fixed2, short, mul, 16, *, float8, kmp_real64, 2i, 1, KMP_ARCH_X86) ATOMIC_CMPXCHG_MIX(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_MIX (fixed4, kmp_int32, mul, 32, *, float8, kmp_real64, 4i, 3, 0) ATOMIC_CMPXCHG_MIX(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_MIX (fixed8, kmp_int64, mul, 64, *, float8, kmp_real64, 8i, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG_MIX(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_MIX (float4, kmp_real32, add, 32,+, float8, kmp_real64, 4r, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG_MIX(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_MIX (float4, kmp_real32, mul, 32, *, float8, kmp_real64, 4r, 3, KMP_ARCH_X86) ATOMIC_CMPXCHG_MIX(float4 |
|
| ATOMIC_CMPXCHG_CMPLX (cmplx4, kmp_cmplx32, add, 64,+, cmplx8, kmp_cmplx64, 8c, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG_CMPLX(cmplx4 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG_CMPLX (cmplx4, kmp_cmplx32, mul, 64, *, cmplx8, kmp_cmplx64, 8c, 7, KMP_ARCH_X86) ATOMIC_CMPXCHG_CMPLX(cmplx4 |
|
| ATOMIC_FIXED_READ (fixed8, rd, kmp_int64, 64,+, KMP_ARCH_X86) ATOMIC_CMPXCHG_READ(float4 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG_READ (float8, rd, kmp_real64, 64,+, KMP_ARCH_X86) ATOMIC_CMPXCHG_READ(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_READ (fixed2, rd, kmp_int16, 16,+, KMP_ARCH_X86) ATOMIC_CRITICAL_READ(float10 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long | ATOMIC_CRITICAL_READ (cmplx4, rd, kmp_cmplx32,+, 8c, 1) ATOMIC_CRITICAL_READ(cmplx8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long | ATOMIC_CRITICAL_READ (cmplx10, rd, kmp_cmplx80,+, 20c, 1) ATOMIC_XCHG_WR(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long KMP_ARCH_X86 | ATOMIC_XCHG_WR (fixed2, wr, kmp_int16, 16,=, KMP_ARCH_X86) ATOMIC_XCHG_WR(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_XCHG_WR (fixed8, wr, kmp_int64, 64,=, KMP_ARCH_X86) ATOMIC_XCHG_FLOAT_WR(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_XCHG_FLOAT_WR (float8, wr, kmp_real64, 64,=, KMP_ARCH_X86) ATOMIC_CRITICAL_WR(float10 |
|
KMP_ARCH_X86 KMP_ARCH_X86 long KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 long | ATOMIC_CRITICAL_WR (cmplx8, wr, kmp_cmplx64,=, 16c, 1) ATOMIC_CRITICAL_WR(cmplx10 |
|
| ATOMIC_FIXED_ADD_CPT (fixed4, add_cpt, kmp_int32, 32,+, 0) ATOMIC_FIXED_ADD_CPT(fixed4 |
|
| ATOMIC_FIXED_ADD_CPT (fixed8, add_cpt, kmp_int64, 64,+, KMP_ARCH_X86) ATOMIC_FIXED_ADD_CPT(fixed8 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (float4, add_cpt, kmp_real32, 32,+, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (float8, add_cpt, kmp_real64, 64,+, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(float8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1, add_cpt, kmp_int8, 8,+, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1, div_cpt, kmp_int8, 8,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed1u |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1, mul_cpt, kmp_int8, 8, *, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1, shl_cpt, kmp_int8, 8,<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1u, shr_cpt, kmp_uint8, 8, > >, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed1 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed1, xor_cpt, kmp_int8, 8, ^, 0) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed2, andb_cpt, kmp_int16, 16, &, 0) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed2u, div_cpt, kmp_uint16, 16,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed2, orb_cpt, kmp_int16, 16,|, 0) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed2u, shr_cpt, kmp_uint16, 16, > >, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed2, xor_cpt, kmp_int16, 16, ^, 0) ATOMIC_CMPXCHG_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed4, div_cpt, kmp_int32, 32,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed4u |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed4, mul_cpt, kmp_int32, 32, *, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed4, shl_cpt, kmp_int32, 32,<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed4u, shr_cpt, kmp_uint32, 32, > >, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed8, andb_cpt, kmp_int64, 64, &, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed8u, div_cpt, kmp_uint64, 64,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed8, orb_cpt, kmp_int64, 64,|, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8, shr_cpt, kmp_int64, 64, > KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed8u, shr_cpt, kmp_uint64, 64, > >, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8, shr_cpt, kmp_int64, 64, > KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (float4, div_cpt, kmp_real32, 32,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2, shr_cpt, kmp_int16, 16, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86<<, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8, shr_cpt, kmp_int64, 64, > KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (float8, div_cpt, kmp_real64, 64,/, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(float8 |
|
| ATOMIC_CMPX_L_CPT (fixed1, andl_cpt, char, 8, &&, KMP_ARCH_X86) ATOMIC_CMPX_L_CPT(fixed1 |
|
KMP_ARCH_X86 | ATOMIC_CMPX_L_CPT (fixed2, andl_cpt, short, 16, &&, KMP_ARCH_X86) ATOMIC_CMPX_L_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_L_CPT (fixed4, andl_cpt, kmp_int32, 32, &&, 0) ATOMIC_CMPX_L_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_L_CPT (fixed8, andl_cpt, kmp_int64, 64, &&, KMP_ARCH_X86) ATOMIC_CMPX_L_CPT(fixed8 |
|
| MIN_MAX_COMPXCHG_CPT (fixed1, max_cpt, char, 8,<, KMP_ARCH_X86) MIN_MAX_COMPXCHG_CPT(fixed1 |
|
KMP_ARCH_X86 | MIN_MAX_COMPXCHG_CPT (fixed2, max_cpt, short, 16,<, KMP_ARCH_X86) MIN_MAX_COMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG_CPT (fixed4, max_cpt, kmp_int32, 32,<, 0) MIN_MAX_COMPXCHG_CPT(fixed4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG_CPT (fixed8, max_cpt, kmp_int64, 64,<, KMP_ARCH_X86) MIN_MAX_COMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG_CPT (float4, max_cpt, kmp_real32, 32,<, KMP_ARCH_X86) MIN_MAX_COMPXCHG_CPT(float4 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_COMPXCHG_CPT (float8, max_cpt, kmp_real64, 64,<, KMP_ARCH_X86) MIN_MAX_COMPXCHG_CPT(float8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | MIN_MAX_CRITICAL_CPT (float10, max_cpt, long double,<, 10r, 1) MIN_MAX_CRITICAL_CPT(float10 |
|
| ATOMIC_CMPXCHG_CPT (fixed1, neqv_cpt, kmp_int8, 8, ^, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed2 |
|
KMP_ARCH_X86 | ATOMIC_CMPXCHG_CPT (fixed4, neqv_cpt, kmp_int32, 32, ^, KMP_ARCH_X86) ATOMIC_CMPXCHG_CPT(fixed8 |
|
KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_EQV_CPT (fixed1, eqv_cpt, kmp_int8, 8, ^~, KMP_ARCH_X86) ATOMIC_CMPX_EQV_CPT(fixed2 |
|
KMP_ARCH_X86 KMP_ARCH_X86 KMP_ARCH_X86 | ATOMIC_CMPX_EQV_CPT (fixed4, eqv_cpt, kmp_int32, 32, ^~, KMP_ARCH_X86) ATOMIC_CMPX_EQV_CPT(fixed8 |
|
| ATOMIC_CRITICAL_CPT (float10, add_cpt, long double,+, 10r, 1) ATOMIC_CRITICAL_CPT(float10 |
|
long | ATOMIC_CRITICAL_CPT (float10, mul_cpt, long double, *, 10r, 1) ATOMIC_CRITICAL_CPT(float10 |
|
long long | ATOMIC_CRITICAL_CPT_WRK (cmplx4, add_cpt, kmp_cmplx32,+, 8c, 1) ATOMIC_CRITICAL_CPT_WRK(cmplx4 |
|
long long | ATOMIC_CRITICAL_CPT_WRK (cmplx4, mul_cpt, kmp_cmplx32, *, 8c, 1) ATOMIC_CRITICAL_CPT_WRK(cmplx4 |
|
long long | ATOMIC_CRITICAL_CPT (cmplx8, add_cpt, kmp_cmplx64,+, 16c, 1) ATOMIC_CRITICAL_CPT(cmplx8 |
|
long long | ATOMIC_CRITICAL_CPT (cmplx8, mul_cpt, kmp_cmplx64, *, 16c, 1) ATOMIC_CRITICAL_CPT(cmplx8 |
|
long long | ATOMIC_CRITICAL_CPT (cmplx10, add_cpt, kmp_cmplx80,+, 20c, 1) ATOMIC_CRITICAL_CPT(cmplx10 |
|
long long | ATOMIC_CRITICAL_CPT (cmplx10, mul_cpt, kmp_cmplx80, *, 20c, 1) ATOMIC_CRITICAL_CPT(cmplx10 |
|
long long void | __kmpc_atomic_1 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_2 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_4 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_8 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_16 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_32 (ident_t *id_ref, int gtid, void *lhs, void *rhs, void(*f)(void *, void *, void *)) |
|
void | __kmpc_atomic_start (void) |
|
void | __kmpc_atomic_end (void) |
|