Overview
Comment: | atomic_*.h: Add missing _Nullable |
---|---|
Downloads: | Tarball | ZIP archive | SQL archive |
Timelines: | family | ancestors | descendants | both | trunk |
Files: | files | file ages | folders |
SHA3-256: |
6f9c0e8eb2daa30341d1ac0d1d12ab1a |
User & Date: | js on 2017-04-14 05:13:09 |
Other Links: | manifest | tags |
Context
2017-04-14
| ||
05:14 | ObjFW.h: Add missing imports check-in: ddd6a56658 user: js tags: trunk | |
05:13 | atomic_*.h: Add missing _Nullable check-in: 6f9c0e8eb2 user: js tags: trunk | |
02:28 | OFSandbox: Revert using a bitfield check-in: 02ccf37477 user: js tags: trunk | |
Changes
Modified src/atomic_builtins.h from [07554058fd] to [2dc5b3a1a5].
︙ | ︙ | |||
22 23 24 25 26 27 28 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return __atomic_add_fetch(p, i, __ATOMIC_RELAXED); } | | | | 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return __atomic_add_fetch(p, i, __ATOMIC_RELAXED); } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { return __atomic_add_fetch(p, i, __ATOMIC_RELAXED); } static OF_INLINE int of_atomic_int_sub(volatile int *_Nonnull p, int i) { return __atomic_sub_fetch(p, i, __ATOMIC_RELAXED); } static OF_INLINE int32_t of_atomic_int32_sub(volatile int32_t *_Nonnull p, int32_t i) { return __atomic_sub_fetch(p, i, __ATOMIC_RELAXED); } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { return __atomic_sub_fetch(p, i, __ATOMIC_RELAXED); } static OF_INLINE int of_atomic_int_inc(volatile int *_Nonnull p) |
︙ | ︙ |
Modified src/atomic_no_threads.h from [d6831a1ea4] to [3924fe3a37].
︙ | ︙ | |||
22 23 24 25 26 27 28 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return (*p += i); } | | | | 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return (*p += i); } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { return (*(char* volatile*)p += i); } static OF_INLINE int of_atomic_int_sub(volatile int *_Nonnull p, int i) { return (*p -= i); } static OF_INLINE int32_t of_atomic_int32_sub(volatile int32_t *_Nonnull p, int32_t i) { return (*p -= i); } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { return (*(char* volatile*)p -= i); } static OF_INLINE int of_atomic_int_inc(volatile int *_Nonnull p) |
︙ | ︙ |
Modified src/atomic_osatomic.h from [eb1a6c5a77] to [7e5dffc102].
︙ | ︙ | |||
24 25 26 27 28 29 30 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return OSAtomicAdd32(i, p); } | | | 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return OSAtomicAdd32(i, p); } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { #ifdef __LP64__ return (void*)OSAtomicAdd64(i, (int64_t*)p); #else return (void*)OSAtomicAdd32(i, (int32_t*)p); #endif |
︙ | ︙ | |||
46 47 48 49 50 51 52 | static OF_INLINE int32_t of_atomic_int32_sub(volatile int32_t *_Nonnull p, int32_t i) { return OSAtomicAdd32(-i, p); } | | | 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 | static OF_INLINE int32_t of_atomic_int32_sub(volatile int32_t *_Nonnull p, int32_t i) { return OSAtomicAdd32(-i, p); } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { #ifdef __LP64__ return (void*)OSAtomicAdd64(-i, (int64_t*)p); #else return (void*)OSAtomicAdd32(-i, (int32_t*)p); #endif |
︙ | ︙ |
Modified src/atomic_powerpc.h from [3babc72639] to [56d99bc409].
︙ | ︙ | |||
42 43 44 45 46 47 48 | : "=&r"(i) : "r"(i), "r"(p) ); return i; } | | | 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 | : "=&r"(i) : "r"(i), "r"(p) ); return i; } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { __asm__ __volatile__ ( "0:\n\t" "lwarx %0, 0, %2\n\t" "add %0, %0, %1\n\t" "stwcx. %0, 0, %2\n\t" |
︙ | ︙ | |||
90 91 92 93 94 95 96 | : "=&r"(i) : "r"(i), "r"(p) ); return i; } | | | 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 | : "=&r"(i) : "r"(i), "r"(p) ); return i; } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { __asm__ __volatile__ ( "0:\n\t" "lwarx %0, 0, %2\n\t" "sub %0, %0, %1\n\t" "stwcx. %0, 0, %2\n\t" |
︙ | ︙ |
Modified src/atomic_sync_builtins.h from [cd9ea173f9] to [b3ae278d63].
︙ | ︙ | |||
22 23 24 25 26 27 28 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return __sync_add_and_fetch(p, i); } | | | | 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 | static OF_INLINE int32_t of_atomic_int32_add(volatile int32_t *_Nonnull p, int32_t i) { return __sync_add_and_fetch(p, i); } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { return __sync_add_and_fetch(p, (void*)i); } static OF_INLINE int of_atomic_int_sub(volatile int *_Nonnull p, int i) { return __sync_sub_and_fetch(p, i); } static OF_INLINE int32_t of_atomic_int32_sub(volatile int32_t *_Nonnull p, int32_t i) { return __sync_sub_and_fetch(p, i); } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { return __sync_sub_and_fetch(p, (void*)i); } static OF_INLINE int of_atomic_int_inc(volatile int *_Nonnull p) |
︙ | ︙ |
Modified src/atomic_x86.h from [3c92b1ba0d] to [7c455e67f3].
︙ | ︙ | |||
53 54 55 56 57 58 59 | : "+&r"(i) : "r"(i), "m"(*p) ); return i; } | | | 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 | : "+&r"(i) : "r"(i), "m"(*p) ); return i; } static OF_INLINE void *_Nullable of_atomic_ptr_add(void *volatile _Nullable *_Nonnull p, intptr_t i) { #if defined(OF_X86_64_ASM) __asm__ __volatile__ ( "lock\n\t" "xaddq %0, %2\n\t" "addq %1, %0" |
︙ | ︙ | |||
123 124 125 126 127 128 129 | : "+&r"(i) : "r"(i), "m"(*p) ); return i; } | | | 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 | : "+&r"(i) : "r"(i), "m"(*p) ); return i; } static OF_INLINE void *_Nullable of_atomic_ptr_sub(void *volatile _Nullable *_Nonnull p, intptr_t i) { #if defined(OF_X86_64_ASM) __asm__ __volatile__ ( "negq %0\n\t" "lock\n\t" "xaddq %0, %2\n\t" |
︙ | ︙ |