From 284058073d7856b778181eb37b06c7f7561d6ed1 Mon Sep 17 00:00:00 2001 From: Sebastian Lackner Date: Mon, 18 Jul 2016 16:05:41 +0200 Subject: [PATCH] vcomp: Implement 16-bit atomic instructions. Signed-off-by: Sebastian Lackner Signed-off-by: Alexandre Julliard --- dlls/vcomp/main.c | 115 ++++++++++++++++++++++++++++++++++++ dlls/vcomp/vcomp.spec | 22 +++---- dlls/vcomp100/vcomp100.spec | 22 +++---- dlls/vcomp110/vcomp110.spec | 22 +++---- dlls/vcomp120/vcomp120.spec | 22 +++---- dlls/vcomp90/vcomp90.spec | 22 +++---- 6 files changed, 170 insertions(+), 55 deletions(-) diff --git a/dlls/vcomp/main.c b/dlls/vcomp/main.c index 446b83da21a..32cfab4c980 100644 --- a/dlls/vcomp/main.c +++ b/dlls/vcomp/main.c @@ -238,6 +238,57 @@ static void CDECL _vcomp_fork_call_wrapper(void *wrapper, int nargs, __ms_va_lis #endif +#if defined(__GNUC__) && (defined(__i386__) || defined(__x86_64__)) + +static inline short interlocked_cmpxchg16(short *dest, short xchg, short compare) +{ + short ret; + __asm__ __volatile__( "lock; cmpxchgw %2,(%1)" + : "=a" (ret) : "r" (dest), "r" (xchg), "0" (compare) : "memory" ); + return ret; +} + +static inline short interlocked_xchg_add16(short *dest, short incr) +{ + short ret; + __asm__ __volatile__( "lock; xaddw %0,(%1)" + : "=r" (ret) : "r" (dest), "0" (incr) : "memory" ); + return ret; +} + +#else /* __GNUC__ */ + +#ifdef __GCC_HAVE_SYNC_COMPARE_AND_SWAP_2 +static inline short interlocked_cmpxchg16(short *dest, short xchg, short compare) +{ + return __sync_val_compare_and_swap(dest, compare, xchg); +} + +static inline short interlocked_xchg_add16(short *dest, short incr) +{ + return __sync_fetch_and_add(dest, incr); +} +#else +static short interlocked_cmpxchg16(short *dest, short xchg, short compare) +{ + EnterCriticalSection(&vcomp_section); + if (*dest == compare) *dest = xchg; else compare = *dest; + LeaveCriticalSection(&vcomp_section); + return compare; +} + +static short interlocked_xchg_add16(short *dest, short incr) +{ + short ret; + EnterCriticalSection(&vcomp_section); + ret = *dest; *dest += incr; + LeaveCriticalSection(&vcomp_section); + return ret; +} +#endif + +#endif /* __GNUC__ */ + static inline struct vcomp_thread_data *vcomp_get_thread_data(void) { return (struct vcomp_thread_data *)TlsGetValue(vcomp_context_tls); @@ -292,6 +343,70 @@ static void vcomp_free_thread_data(void) vcomp_set_thread_data(NULL); } +void CDECL _vcomp_atomic_add_i2(short *dest, short val) +{ + interlocked_xchg_add16(dest, val); +} + +void CDECL _vcomp_atomic_and_i2(short *dest, short val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old & val, old) != old); +} + +void CDECL _vcomp_atomic_div_i2(short *dest, short val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old / val, old) != old); +} + +void CDECL _vcomp_atomic_div_ui2(unsigned short *dest, unsigned short val) +{ + unsigned short old; + do old = *dest; while ((unsigned short)interlocked_cmpxchg16((short *)dest, old / val, old) != old); +} + +void CDECL _vcomp_atomic_mul_i2(short *dest, short val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old * val, old) != old); +} + +void CDECL _vcomp_atomic_or_i2(short *dest, short val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old | val, old) != old); +} + +void CDECL _vcomp_atomic_shl_i2(short *dest, unsigned int val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old << val, old) != old); +} + +void CDECL _vcomp_atomic_shr_i2(short *dest, unsigned int val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old >> val, old) != old); +} + +void CDECL _vcomp_atomic_shr_ui2(unsigned short *dest, unsigned int val) +{ + unsigned short old; + do old = *dest; while ((unsigned short)interlocked_cmpxchg16((short *)dest, old >> val, old) != old); +} + +void CDECL _vcomp_atomic_sub_i2(short *dest, short val) +{ + interlocked_xchg_add16(dest, -val); +} + +void CDECL _vcomp_atomic_xor_i2(short *dest, short val) +{ + short old; + do old = *dest; while (interlocked_cmpxchg16(dest, old ^ val, old) != old); +} + void CDECL _vcomp_atomic_add_i4(int *dest, int val) { interlocked_xchg_add(dest, val); diff --git a/dlls/vcomp/vcomp.spec b/dlls/vcomp/vcomp.spec index eff411e4d55..018ffffb98c 100644 --- a/dlls/vcomp/vcomp.spec +++ b/dlls/vcomp/vcomp.spec @@ -1,53 +1,53 @@ @ stub _vcomp_atomic_add_i1 -@ stub _vcomp_atomic_add_i2 +@ cdecl _vcomp_atomic_add_i2(ptr long) @ cdecl _vcomp_atomic_add_i4(ptr long) @ cdecl _vcomp_atomic_add_i8(ptr int64) @ cdecl _vcomp_atomic_add_r4(ptr float) @ cdecl _vcomp_atomic_add_r8(ptr double) @ stub _vcomp_atomic_and_i1 -@ stub _vcomp_atomic_and_i2 +@ cdecl _vcomp_atomic_and_i2(ptr long) @ cdecl _vcomp_atomic_and_i4(ptr long) @ cdecl _vcomp_atomic_and_i8(ptr int64) @ stub _vcomp_atomic_div_i1 -@ stub _vcomp_atomic_div_i2 +@ cdecl _vcomp_atomic_div_i2(ptr long) @ cdecl _vcomp_atomic_div_i4(ptr long) @ cdecl _vcomp_atomic_div_i8(ptr int64) @ cdecl _vcomp_atomic_div_r4(ptr float) @ cdecl _vcomp_atomic_div_r8(ptr double) @ stub _vcomp_atomic_div_ui1 -@ stub _vcomp_atomic_div_ui2 +@ cdecl _vcomp_atomic_div_ui2(ptr long) @ cdecl _vcomp_atomic_div_ui4(ptr long) @ cdecl _vcomp_atomic_div_ui8(ptr int64) @ stub _vcomp_atomic_mul_i1 -@ stub _vcomp_atomic_mul_i2 +@ cdecl _vcomp_atomic_mul_i2(ptr long) @ cdecl _vcomp_atomic_mul_i4(ptr long) @ cdecl _vcomp_atomic_mul_i8(ptr int64) @ cdecl _vcomp_atomic_mul_r4(ptr float) @ cdecl _vcomp_atomic_mul_r8(ptr double) @ stub _vcomp_atomic_or_i1 -@ stub _vcomp_atomic_or_i2 +@ cdecl _vcomp_atomic_or_i2(ptr long) @ cdecl _vcomp_atomic_or_i4(ptr long) @ cdecl _vcomp_atomic_or_i8(ptr int64) @ stub _vcomp_atomic_shl_i1 -@ stub _vcomp_atomic_shl_i2 +@ cdecl _vcomp_atomic_shl_i2(ptr long) @ cdecl _vcomp_atomic_shl_i4(ptr long) @ cdecl _vcomp_atomic_shl_i8(ptr long) @ stub _vcomp_atomic_shr_i1 -@ stub _vcomp_atomic_shr_i2 +@ cdecl _vcomp_atomic_shr_i2(ptr long) @ cdecl _vcomp_atomic_shr_i4(ptr long) @ cdecl _vcomp_atomic_shr_i8(ptr long) @ stub _vcomp_atomic_shr_ui1 -@ stub _vcomp_atomic_shr_ui2 +@ cdecl _vcomp_atomic_shr_ui2(ptr long) @ cdecl _vcomp_atomic_shr_ui4(ptr long) @ cdecl _vcomp_atomic_shr_ui8(ptr long) @ stub _vcomp_atomic_sub_i1 -@ stub _vcomp_atomic_sub_i2 +@ cdecl _vcomp_atomic_sub_i2(ptr long) @ cdecl _vcomp_atomic_sub_i4(ptr long) @ cdecl _vcomp_atomic_sub_i8(ptr int64) @ cdecl _vcomp_atomic_sub_r4(ptr float) @ cdecl _vcomp_atomic_sub_r8(ptr double) @ stub _vcomp_atomic_xor_i1 -@ stub _vcomp_atomic_xor_i2 +@ cdecl _vcomp_atomic_xor_i2(ptr long) @ cdecl _vcomp_atomic_xor_i4(ptr long) @ cdecl _vcomp_atomic_xor_i8(ptr int64) @ cdecl _vcomp_barrier() diff --git a/dlls/vcomp100/vcomp100.spec b/dlls/vcomp100/vcomp100.spec index ba1f414defd..41eb824d42e 100644 --- a/dlls/vcomp100/vcomp100.spec +++ b/dlls/vcomp100/vcomp100.spec @@ -1,53 +1,53 @@ @ stub _vcomp_atomic_add_i1 -@ stub _vcomp_atomic_add_i2 +@ cdecl _vcomp_atomic_add_i2(ptr long) vcomp._vcomp_atomic_add_i2 @ cdecl _vcomp_atomic_add_i4(ptr long) vcomp._vcomp_atomic_add_i4 @ cdecl _vcomp_atomic_add_i8(ptr int64) vcomp._vcomp_atomic_add_i8 @ cdecl _vcomp_atomic_add_r4(ptr float) vcomp._vcomp_atomic_add_r4 @ cdecl _vcomp_atomic_add_r8(ptr double) vcomp._vcomp_atomic_add_r8 @ stub _vcomp_atomic_and_i1 -@ stub _vcomp_atomic_and_i2 +@ cdecl _vcomp_atomic_and_i2(ptr long) vcomp._vcomp_atomic_and_i2 @ cdecl _vcomp_atomic_and_i4(ptr long) vcomp._vcomp_atomic_and_i4 @ cdecl _vcomp_atomic_and_i8(ptr int64) vcomp._vcomp_atomic_and_i8 @ stub _vcomp_atomic_div_i1 -@ stub _vcomp_atomic_div_i2 +@ cdecl _vcomp_atomic_div_i2(ptr long) vcomp._vcomp_atomic_div_i2 @ cdecl _vcomp_atomic_div_i4(ptr long) vcomp._vcomp_atomic_div_i4 @ cdecl _vcomp_atomic_div_i8(ptr int64) vcomp._vcomp_atomic_div_i8 @ cdecl _vcomp_atomic_div_r4(ptr float) vcomp._vcomp_atomic_div_r4 @ cdecl _vcomp_atomic_div_r8(ptr double) vcomp._vcomp_atomic_div_r8 @ stub _vcomp_atomic_div_ui1 -@ stub _vcomp_atomic_div_ui2 +@ cdecl _vcomp_atomic_div_ui2(ptr long) vcomp._vcomp_atomic_div_ui2 @ cdecl _vcomp_atomic_div_ui4(ptr long) vcomp._vcomp_atomic_div_ui4 @ cdecl _vcomp_atomic_div_ui8(ptr int64) vcomp._vcomp_atomic_div_ui8 @ stub _vcomp_atomic_mul_i1 -@ stub _vcomp_atomic_mul_i2 +@ cdecl _vcomp_atomic_mul_i2(ptr long) vcomp._vcomp_atomic_mul_i2 @ cdecl _vcomp_atomic_mul_i4(ptr long) vcomp._vcomp_atomic_mul_i4 @ cdecl _vcomp_atomic_mul_i8(ptr int64) vcomp._vcomp_atomic_mul_i8 @ cdecl _vcomp_atomic_mul_r4(ptr float) vcomp._vcomp_atomic_mul_r4 @ cdecl _vcomp_atomic_mul_r8(ptr double) vcomp._vcomp_atomic_mul_r8 @ stub _vcomp_atomic_or_i1 -@ stub _vcomp_atomic_or_i2 +@ cdecl _vcomp_atomic_or_i2(ptr long) vcomp._vcomp_atomic_or_i2 @ cdecl _vcomp_atomic_or_i4(ptr long) vcomp._vcomp_atomic_or_i4 @ cdecl _vcomp_atomic_or_i8(ptr int64) vcomp._vcomp_atomic_or_i8 @ stub _vcomp_atomic_shl_i1 -@ stub _vcomp_atomic_shl_i2 +@ cdecl _vcomp_atomic_shl_i2(ptr long) vcomp._vcomp_atomic_shl_i2 @ cdecl _vcomp_atomic_shl_i4(ptr long) vcomp._vcomp_atomic_shl_i4 @ cdecl _vcomp_atomic_shl_i8(ptr long) vcomp._vcomp_atomic_shl_i8 @ stub _vcomp_atomic_shr_i1 -@ stub _vcomp_atomic_shr_i2 +@ cdecl _vcomp_atomic_shr_i2(ptr long) vcomp._vcomp_atomic_shr_i2 @ cdecl _vcomp_atomic_shr_i4(ptr long) vcomp._vcomp_atomic_shr_i4 @ cdecl _vcomp_atomic_shr_i8(ptr long) vcomp._vcomp_atomic_shr_i8 @ stub _vcomp_atomic_shr_ui1 -@ stub _vcomp_atomic_shr_ui2 +@ cdecl _vcomp_atomic_shr_ui2(ptr long) vcomp._vcomp_atomic_shr_ui2 @ cdecl _vcomp_atomic_shr_ui4(ptr long) vcomp._vcomp_atomic_shr_ui4 @ cdecl _vcomp_atomic_shr_ui8(ptr long) vcomp._vcomp_atomic_shr_ui8 @ stub _vcomp_atomic_sub_i1 -@ stub _vcomp_atomic_sub_i2 +@ cdecl _vcomp_atomic_sub_i2(ptr long) vcomp._vcomp_atomic_sub_i2 @ cdecl _vcomp_atomic_sub_i4(ptr long) vcomp._vcomp_atomic_sub_i4 @ cdecl _vcomp_atomic_sub_i8(ptr int64) vcomp._vcomp_atomic_sub_i8 @ cdecl _vcomp_atomic_sub_r4(ptr float) vcomp._vcomp_atomic_sub_r4 @ cdecl _vcomp_atomic_sub_r8(ptr double) vcomp._vcomp_atomic_sub_r8 @ stub _vcomp_atomic_xor_i1 -@ stub _vcomp_atomic_xor_i2 +@ cdecl _vcomp_atomic_xor_i2(ptr long) vcomp._vcomp_atomic_xor_i2 @ cdecl _vcomp_atomic_xor_i4(ptr long) vcomp._vcomp_atomic_xor_i4 @ cdecl _vcomp_atomic_xor_i8(ptr int64) vcomp._vcomp_atomic_xor_i8 @ cdecl _vcomp_barrier() vcomp._vcomp_barrier diff --git a/dlls/vcomp110/vcomp110.spec b/dlls/vcomp110/vcomp110.spec index 8389d27fa45..c95310abb15 100644 --- a/dlls/vcomp110/vcomp110.spec +++ b/dlls/vcomp110/vcomp110.spec @@ -1,54 +1,54 @@ @ stub C2VectParallel @ stub _vcomp_atomic_add_i1 -@ stub _vcomp_atomic_add_i2 +@ cdecl _vcomp_atomic_add_i2(ptr long) vcomp._vcomp_atomic_add_i2 @ cdecl _vcomp_atomic_add_i4(ptr long) vcomp._vcomp_atomic_add_i4 @ cdecl _vcomp_atomic_add_i8(ptr int64) vcomp._vcomp_atomic_add_i8 @ cdecl _vcomp_atomic_add_r4(ptr float) vcomp._vcomp_atomic_add_r4 @ cdecl _vcomp_atomic_add_r8(ptr double) vcomp._vcomp_atomic_add_r8 @ stub _vcomp_atomic_and_i1 -@ stub _vcomp_atomic_and_i2 +@ cdecl _vcomp_atomic_and_i2(ptr long) vcomp._vcomp_atomic_and_i2 @ cdecl _vcomp_atomic_and_i4(ptr long) vcomp._vcomp_atomic_and_i4 @ cdecl _vcomp_atomic_and_i8(ptr int64) vcomp._vcomp_atomic_and_i8 @ stub _vcomp_atomic_div_i1 -@ stub _vcomp_atomic_div_i2 +@ cdecl _vcomp_atomic_div_i2(ptr long) vcomp._vcomp_atomic_div_i2 @ cdecl _vcomp_atomic_div_i4(ptr long) vcomp._vcomp_atomic_div_i4 @ cdecl _vcomp_atomic_div_i8(ptr int64) vcomp._vcomp_atomic_div_i8 @ cdecl _vcomp_atomic_div_r4(ptr float) vcomp._vcomp_atomic_div_r4 @ cdecl _vcomp_atomic_div_r8(ptr double) vcomp._vcomp_atomic_div_r8 @ stub _vcomp_atomic_div_ui1 -@ stub _vcomp_atomic_div_ui2 +@ cdecl _vcomp_atomic_div_ui2(ptr long) vcomp._vcomp_atomic_div_ui2 @ cdecl _vcomp_atomic_div_ui4(ptr long) vcomp._vcomp_atomic_div_ui4 @ cdecl _vcomp_atomic_div_ui8(ptr int64) vcomp._vcomp_atomic_div_ui8 @ stub _vcomp_atomic_mul_i1 -@ stub _vcomp_atomic_mul_i2 +@ cdecl _vcomp_atomic_mul_i2(ptr long) vcomp._vcomp_atomic_mul_i2 @ cdecl _vcomp_atomic_mul_i4(ptr long) vcomp._vcomp_atomic_mul_i4 @ cdecl _vcomp_atomic_mul_i8(ptr int64) vcomp._vcomp_atomic_mul_i8 @ cdecl _vcomp_atomic_mul_r4(ptr float) vcomp._vcomp_atomic_mul_r4 @ cdecl _vcomp_atomic_mul_r8(ptr double) vcomp._vcomp_atomic_mul_r8 @ stub _vcomp_atomic_or_i1 -@ stub _vcomp_atomic_or_i2 +@ cdecl _vcomp_atomic_or_i2(ptr long) vcomp._vcomp_atomic_or_i2 @ cdecl _vcomp_atomic_or_i4(ptr long) vcomp._vcomp_atomic_or_i4 @ cdecl _vcomp_atomic_or_i8(ptr int64) vcomp._vcomp_atomic_or_i8 @ stub _vcomp_atomic_shl_i1 -@ stub _vcomp_atomic_shl_i2 +@ cdecl _vcomp_atomic_shl_i2(ptr long) vcomp._vcomp_atomic_shl_i2 @ cdecl _vcomp_atomic_shl_i4(ptr long) vcomp._vcomp_atomic_shl_i4 @ cdecl _vcomp_atomic_shl_i8(ptr long) vcomp._vcomp_atomic_shl_i8 @ stub _vcomp_atomic_shr_i1 -@ stub _vcomp_atomic_shr_i2 +@ cdecl _vcomp_atomic_shr_i2(ptr long) vcomp._vcomp_atomic_shr_i2 @ cdecl _vcomp_atomic_shr_i4(ptr long) vcomp._vcomp_atomic_shr_i4 @ cdecl _vcomp_atomic_shr_i8(ptr long) vcomp._vcomp_atomic_shr_i8 @ stub _vcomp_atomic_shr_ui1 -@ stub _vcomp_atomic_shr_ui2 +@ cdecl _vcomp_atomic_shr_ui2(ptr long) vcomp._vcomp_atomic_shr_ui2 @ cdecl _vcomp_atomic_shr_ui4(ptr long) vcomp._vcomp_atomic_shr_ui4 @ cdecl _vcomp_atomic_shr_ui8(ptr long) vcomp._vcomp_atomic_shr_ui8 @ stub _vcomp_atomic_sub_i1 -@ stub _vcomp_atomic_sub_i2 +@ cdecl _vcomp_atomic_sub_i2(ptr long) vcomp._vcomp_atomic_sub_i2 @ cdecl _vcomp_atomic_sub_i4(ptr long) vcomp._vcomp_atomic_sub_i4 @ cdecl _vcomp_atomic_sub_i8(ptr int64) vcomp._vcomp_atomic_sub_i8 @ cdecl _vcomp_atomic_sub_r4(ptr float) vcomp._vcomp_atomic_sub_r4 @ cdecl _vcomp_atomic_sub_r8(ptr double) vcomp._vcomp_atomic_sub_r8 @ stub _vcomp_atomic_xor_i1 -@ stub _vcomp_atomic_xor_i2 +@ cdecl _vcomp_atomic_xor_i2(ptr long) vcomp._vcomp_atomic_xor_i2 @ cdecl _vcomp_atomic_xor_i4(ptr long) vcomp._vcomp_atomic_xor_i4 @ cdecl _vcomp_atomic_xor_i8(ptr int64) vcomp._vcomp_atomic_xor_i8 @ cdecl _vcomp_barrier() vcomp._vcomp_barrier diff --git a/dlls/vcomp120/vcomp120.spec b/dlls/vcomp120/vcomp120.spec index 8389d27fa45..c95310abb15 100644 --- a/dlls/vcomp120/vcomp120.spec +++ b/dlls/vcomp120/vcomp120.spec @@ -1,54 +1,54 @@ @ stub C2VectParallel @ stub _vcomp_atomic_add_i1 -@ stub _vcomp_atomic_add_i2 +@ cdecl _vcomp_atomic_add_i2(ptr long) vcomp._vcomp_atomic_add_i2 @ cdecl _vcomp_atomic_add_i4(ptr long) vcomp._vcomp_atomic_add_i4 @ cdecl _vcomp_atomic_add_i8(ptr int64) vcomp._vcomp_atomic_add_i8 @ cdecl _vcomp_atomic_add_r4(ptr float) vcomp._vcomp_atomic_add_r4 @ cdecl _vcomp_atomic_add_r8(ptr double) vcomp._vcomp_atomic_add_r8 @ stub _vcomp_atomic_and_i1 -@ stub _vcomp_atomic_and_i2 +@ cdecl _vcomp_atomic_and_i2(ptr long) vcomp._vcomp_atomic_and_i2 @ cdecl _vcomp_atomic_and_i4(ptr long) vcomp._vcomp_atomic_and_i4 @ cdecl _vcomp_atomic_and_i8(ptr int64) vcomp._vcomp_atomic_and_i8 @ stub _vcomp_atomic_div_i1 -@ stub _vcomp_atomic_div_i2 +@ cdecl _vcomp_atomic_div_i2(ptr long) vcomp._vcomp_atomic_div_i2 @ cdecl _vcomp_atomic_div_i4(ptr long) vcomp._vcomp_atomic_div_i4 @ cdecl _vcomp_atomic_div_i8(ptr int64) vcomp._vcomp_atomic_div_i8 @ cdecl _vcomp_atomic_div_r4(ptr float) vcomp._vcomp_atomic_div_r4 @ cdecl _vcomp_atomic_div_r8(ptr double) vcomp._vcomp_atomic_div_r8 @ stub _vcomp_atomic_div_ui1 -@ stub _vcomp_atomic_div_ui2 +@ cdecl _vcomp_atomic_div_ui2(ptr long) vcomp._vcomp_atomic_div_ui2 @ cdecl _vcomp_atomic_div_ui4(ptr long) vcomp._vcomp_atomic_div_ui4 @ cdecl _vcomp_atomic_div_ui8(ptr int64) vcomp._vcomp_atomic_div_ui8 @ stub _vcomp_atomic_mul_i1 -@ stub _vcomp_atomic_mul_i2 +@ cdecl _vcomp_atomic_mul_i2(ptr long) vcomp._vcomp_atomic_mul_i2 @ cdecl _vcomp_atomic_mul_i4(ptr long) vcomp._vcomp_atomic_mul_i4 @ cdecl _vcomp_atomic_mul_i8(ptr int64) vcomp._vcomp_atomic_mul_i8 @ cdecl _vcomp_atomic_mul_r4(ptr float) vcomp._vcomp_atomic_mul_r4 @ cdecl _vcomp_atomic_mul_r8(ptr double) vcomp._vcomp_atomic_mul_r8 @ stub _vcomp_atomic_or_i1 -@ stub _vcomp_atomic_or_i2 +@ cdecl _vcomp_atomic_or_i2(ptr long) vcomp._vcomp_atomic_or_i2 @ cdecl _vcomp_atomic_or_i4(ptr long) vcomp._vcomp_atomic_or_i4 @ cdecl _vcomp_atomic_or_i8(ptr int64) vcomp._vcomp_atomic_or_i8 @ stub _vcomp_atomic_shl_i1 -@ stub _vcomp_atomic_shl_i2 +@ cdecl _vcomp_atomic_shl_i2(ptr long) vcomp._vcomp_atomic_shl_i2 @ cdecl _vcomp_atomic_shl_i4(ptr long) vcomp._vcomp_atomic_shl_i4 @ cdecl _vcomp_atomic_shl_i8(ptr long) vcomp._vcomp_atomic_shl_i8 @ stub _vcomp_atomic_shr_i1 -@ stub _vcomp_atomic_shr_i2 +@ cdecl _vcomp_atomic_shr_i2(ptr long) vcomp._vcomp_atomic_shr_i2 @ cdecl _vcomp_atomic_shr_i4(ptr long) vcomp._vcomp_atomic_shr_i4 @ cdecl _vcomp_atomic_shr_i8(ptr long) vcomp._vcomp_atomic_shr_i8 @ stub _vcomp_atomic_shr_ui1 -@ stub _vcomp_atomic_shr_ui2 +@ cdecl _vcomp_atomic_shr_ui2(ptr long) vcomp._vcomp_atomic_shr_ui2 @ cdecl _vcomp_atomic_shr_ui4(ptr long) vcomp._vcomp_atomic_shr_ui4 @ cdecl _vcomp_atomic_shr_ui8(ptr long) vcomp._vcomp_atomic_shr_ui8 @ stub _vcomp_atomic_sub_i1 -@ stub _vcomp_atomic_sub_i2 +@ cdecl _vcomp_atomic_sub_i2(ptr long) vcomp._vcomp_atomic_sub_i2 @ cdecl _vcomp_atomic_sub_i4(ptr long) vcomp._vcomp_atomic_sub_i4 @ cdecl _vcomp_atomic_sub_i8(ptr int64) vcomp._vcomp_atomic_sub_i8 @ cdecl _vcomp_atomic_sub_r4(ptr float) vcomp._vcomp_atomic_sub_r4 @ cdecl _vcomp_atomic_sub_r8(ptr double) vcomp._vcomp_atomic_sub_r8 @ stub _vcomp_atomic_xor_i1 -@ stub _vcomp_atomic_xor_i2 +@ cdecl _vcomp_atomic_xor_i2(ptr long) vcomp._vcomp_atomic_xor_i2 @ cdecl _vcomp_atomic_xor_i4(ptr long) vcomp._vcomp_atomic_xor_i4 @ cdecl _vcomp_atomic_xor_i8(ptr int64) vcomp._vcomp_atomic_xor_i8 @ cdecl _vcomp_barrier() vcomp._vcomp_barrier diff --git a/dlls/vcomp90/vcomp90.spec b/dlls/vcomp90/vcomp90.spec index ba1f414defd..41eb824d42e 100644 --- a/dlls/vcomp90/vcomp90.spec +++ b/dlls/vcomp90/vcomp90.spec @@ -1,53 +1,53 @@ @ stub _vcomp_atomic_add_i1 -@ stub _vcomp_atomic_add_i2 +@ cdecl _vcomp_atomic_add_i2(ptr long) vcomp._vcomp_atomic_add_i2 @ cdecl _vcomp_atomic_add_i4(ptr long) vcomp._vcomp_atomic_add_i4 @ cdecl _vcomp_atomic_add_i8(ptr int64) vcomp._vcomp_atomic_add_i8 @ cdecl _vcomp_atomic_add_r4(ptr float) vcomp._vcomp_atomic_add_r4 @ cdecl _vcomp_atomic_add_r8(ptr double) vcomp._vcomp_atomic_add_r8 @ stub _vcomp_atomic_and_i1 -@ stub _vcomp_atomic_and_i2 +@ cdecl _vcomp_atomic_and_i2(ptr long) vcomp._vcomp_atomic_and_i2 @ cdecl _vcomp_atomic_and_i4(ptr long) vcomp._vcomp_atomic_and_i4 @ cdecl _vcomp_atomic_and_i8(ptr int64) vcomp._vcomp_atomic_and_i8 @ stub _vcomp_atomic_div_i1 -@ stub _vcomp_atomic_div_i2 +@ cdecl _vcomp_atomic_div_i2(ptr long) vcomp._vcomp_atomic_div_i2 @ cdecl _vcomp_atomic_div_i4(ptr long) vcomp._vcomp_atomic_div_i4 @ cdecl _vcomp_atomic_div_i8(ptr int64) vcomp._vcomp_atomic_div_i8 @ cdecl _vcomp_atomic_div_r4(ptr float) vcomp._vcomp_atomic_div_r4 @ cdecl _vcomp_atomic_div_r8(ptr double) vcomp._vcomp_atomic_div_r8 @ stub _vcomp_atomic_div_ui1 -@ stub _vcomp_atomic_div_ui2 +@ cdecl _vcomp_atomic_div_ui2(ptr long) vcomp._vcomp_atomic_div_ui2 @ cdecl _vcomp_atomic_div_ui4(ptr long) vcomp._vcomp_atomic_div_ui4 @ cdecl _vcomp_atomic_div_ui8(ptr int64) vcomp._vcomp_atomic_div_ui8 @ stub _vcomp_atomic_mul_i1 -@ stub _vcomp_atomic_mul_i2 +@ cdecl _vcomp_atomic_mul_i2(ptr long) vcomp._vcomp_atomic_mul_i2 @ cdecl _vcomp_atomic_mul_i4(ptr long) vcomp._vcomp_atomic_mul_i4 @ cdecl _vcomp_atomic_mul_i8(ptr int64) vcomp._vcomp_atomic_mul_i8 @ cdecl _vcomp_atomic_mul_r4(ptr float) vcomp._vcomp_atomic_mul_r4 @ cdecl _vcomp_atomic_mul_r8(ptr double) vcomp._vcomp_atomic_mul_r8 @ stub _vcomp_atomic_or_i1 -@ stub _vcomp_atomic_or_i2 +@ cdecl _vcomp_atomic_or_i2(ptr long) vcomp._vcomp_atomic_or_i2 @ cdecl _vcomp_atomic_or_i4(ptr long) vcomp._vcomp_atomic_or_i4 @ cdecl _vcomp_atomic_or_i8(ptr int64) vcomp._vcomp_atomic_or_i8 @ stub _vcomp_atomic_shl_i1 -@ stub _vcomp_atomic_shl_i2 +@ cdecl _vcomp_atomic_shl_i2(ptr long) vcomp._vcomp_atomic_shl_i2 @ cdecl _vcomp_atomic_shl_i4(ptr long) vcomp._vcomp_atomic_shl_i4 @ cdecl _vcomp_atomic_shl_i8(ptr long) vcomp._vcomp_atomic_shl_i8 @ stub _vcomp_atomic_shr_i1 -@ stub _vcomp_atomic_shr_i2 +@ cdecl _vcomp_atomic_shr_i2(ptr long) vcomp._vcomp_atomic_shr_i2 @ cdecl _vcomp_atomic_shr_i4(ptr long) vcomp._vcomp_atomic_shr_i4 @ cdecl _vcomp_atomic_shr_i8(ptr long) vcomp._vcomp_atomic_shr_i8 @ stub _vcomp_atomic_shr_ui1 -@ stub _vcomp_atomic_shr_ui2 +@ cdecl _vcomp_atomic_shr_ui2(ptr long) vcomp._vcomp_atomic_shr_ui2 @ cdecl _vcomp_atomic_shr_ui4(ptr long) vcomp._vcomp_atomic_shr_ui4 @ cdecl _vcomp_atomic_shr_ui8(ptr long) vcomp._vcomp_atomic_shr_ui8 @ stub _vcomp_atomic_sub_i1 -@ stub _vcomp_atomic_sub_i2 +@ cdecl _vcomp_atomic_sub_i2(ptr long) vcomp._vcomp_atomic_sub_i2 @ cdecl _vcomp_atomic_sub_i4(ptr long) vcomp._vcomp_atomic_sub_i4 @ cdecl _vcomp_atomic_sub_i8(ptr int64) vcomp._vcomp_atomic_sub_i8 @ cdecl _vcomp_atomic_sub_r4(ptr float) vcomp._vcomp_atomic_sub_r4 @ cdecl _vcomp_atomic_sub_r8(ptr double) vcomp._vcomp_atomic_sub_r8 @ stub _vcomp_atomic_xor_i1 -@ stub _vcomp_atomic_xor_i2 +@ cdecl _vcomp_atomic_xor_i2(ptr long) vcomp._vcomp_atomic_xor_i2 @ cdecl _vcomp_atomic_xor_i4(ptr long) vcomp._vcomp_atomic_xor_i4 @ cdecl _vcomp_atomic_xor_i8(ptr int64) vcomp._vcomp_atomic_xor_i8 @ cdecl _vcomp_barrier() vcomp._vcomp_barrier