︙ | | |
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
|
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
|
-
+
|
#ifdef OF_HAVE_OSATOMIC
# include <libkern/OSAtomic.h>
#endif
OF_ASSUME_NONNULL_BEGIN
static OF_INLINE int
of_atomic_int_add(volatile __nonnull int *p, int i)
of_atomic_int_add(volatile int *OF_NONNULL p, int i)
{
#if !defined(OF_HAVE_THREADS)
return (*p += i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_add_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
if (sizeof(int) == 4)
|
︙ | | |
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
|
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
|
-
+
|
return OSAtomicAdd32Barrier(i, p);
#else
# error of_atomic_int_add not implemented!
#endif
}
static OF_INLINE int32_t
of_atomic_int32_add(volatile __nonnull int32_t *p, int32_t i)
of_atomic_int32_add(volatile int32_t *OF_NONNULL p, int32_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*p += i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_add_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
__asm__ __volatile__ (
|
︙ | | |
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
|
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
|
-
+
|
return OSAtomicAdd32Barrier(i, p);
#else
# error of_atomic_int32_add not implemented!
#endif
}
static OF_INLINE void*
of_atomic_ptr_add(__nullable void *volatile *__nonnull p, intptr_t i)
of_atomic_ptr_add(void *volatile OF_NULLABLE *OF_NONNULL p, intptr_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*(char* volatile*)p += i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_add_and_fetch(p, (void*)i);
#elif defined(OF_X86_64_ASM)
__asm__ __volatile__ (
|
︙ | | |
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
|
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
|
-
+
|
# endif
#else
# error of_atomic_ptr_add not implemented!
#endif
}
static OF_INLINE int
of_atomic_int_sub(volatile __nonnull int *p, int i)
of_atomic_int_sub(volatile int *OF_NONNULL p, int i)
{
#if !defined(OF_HAVE_THREADS)
return (*p -= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_sub_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
if (sizeof(int) == 4)
|
︙ | | |
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
|
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
|
-
+
|
return OSAtomicAdd32Barrier(-i, p);
#else
# error of_atomic_int_sub not implemented!
#endif
}
static OF_INLINE int32_t
of_atomic_int32_sub(volatile __nonnull int32_t *p, int32_t i)
of_atomic_int32_sub(volatile int32_t *OF_NONNULL p, int32_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*p -= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_sub_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
__asm__ __volatile__ (
|
︙ | | |
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
|
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
|
-
+
|
return OSAtomicAdd32Barrier(-i, p);
#else
# error of_atomic_int32_sub not implemented!
#endif
}
static OF_INLINE void*
of_atomic_ptr_sub(__nullable void *volatile *__nonnull p, intptr_t i)
of_atomic_ptr_sub(void *volatile OF_NULLABLE *OF_NONNULL p, intptr_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*(char* volatile*)p -= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_sub_and_fetch(p, (void*)i);
#elif defined(OF_X86_64_ASM)
__asm__ __volatile__ (
|
︙ | | |
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
|
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
|
-
+
|
# endif
#else
# error of_atomic_ptr_sub not implemented!
#endif
}
static OF_INLINE int
of_atomic_int_inc(volatile __nonnull int *p)
of_atomic_int_inc(volatile int *OF_NONNULL p)
{
#if !defined(OF_HAVE_THREADS)
return ++*p;
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_add_and_fetch(p, 1);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
int i;
|
︙ | | |
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
|
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
|
-
+
|
return OSAtomicIncrement32Barrier(p);
#else
# error of_atomic_int_inc not implemented!
#endif
}
static OF_INLINE int32_t
of_atomic_int32_inc(volatile __nonnull int32_t *p)
of_atomic_int32_inc(volatile int32_t *OF_NONNULL p)
{
#if !defined(OF_HAVE_THREADS)
return ++*p;
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_add_and_fetch(p, 1);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
int32_t i;
|
︙ | | |
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
|
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
|
-
+
|
return OSAtomicIncrement32Barrier(p);
#else
# error of_atomic_int32_inc not implemented!
#endif
}
static OF_INLINE int
of_atomic_int_dec(volatile __nonnull int *p)
of_atomic_int_dec(volatile int *OF_NONNULL p)
{
#if !defined(OF_HAVE_THREADS)
return --*p;
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_sub_and_fetch(p, 1);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
int i;
|
︙ | | |
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
|
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
|
-
+
|
return OSAtomicDecrement32Barrier(p);
#else
# error of_atomic_int_dec not implemented!
#endif
}
static OF_INLINE int32_t
of_atomic_int32_dec(volatile __nonnull int32_t *p)
of_atomic_int32_dec(volatile int32_t *OF_NONNULL p)
{
#if !defined(OF_HAVE_THREADS)
return --*p;
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_sub_and_fetch(p, 1);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
int32_t i;
|
︙ | | |
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
|
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
|
-
+
|
return OSAtomicDecrement32Barrier(p);
#else
# error of_atomic_int32_dec not implemented!
#endif
}
static OF_INLINE unsigned int
of_atomic_int_or(volatile __nonnull unsigned int *p, unsigned int i)
of_atomic_int_or(volatile unsigned int *OF_NONNULL p, unsigned int i)
{
#if !defined(OF_HAVE_THREADS)
return (*p |= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_or_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
if (sizeof(int) == 4)
|
︙ | | |
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
|
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
|
-
+
|
return OSAtomicOr32Barrier(i, p);
#else
# error of_atomic_int_or not implemented!
#endif
}
static OF_INLINE uint32_t
of_atomic_int32_or(volatile __nonnull uint32_t *p, uint32_t i)
of_atomic_int32_or(volatile uint32_t *OF_NONNULL p, uint32_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*p |= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_or_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
__asm__ __volatile__ (
|
︙ | | |
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
|
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
|
-
+
|
return OSAtomicOr32Barrier(i, p);
#else
# error of_atomic_int32_or not implemented!
#endif
}
static OF_INLINE unsigned int
of_atomic_int_and(volatile __nonnull unsigned int *p, unsigned int i)
of_atomic_int_and(volatile unsigned int *OF_NONNULL p, unsigned int i)
{
#if !defined(OF_HAVE_THREADS)
return (*p &= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_and_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
if (sizeof(int) == 4)
|
︙ | | |
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
|
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
|
-
+
|
return OSAtomicAnd32Barrier(i, p);
#else
# error of_atomic_int_and not implemented!
#endif
}
static OF_INLINE uint32_t
of_atomic_int32_and(volatile __nonnull uint32_t *p, uint32_t i)
of_atomic_int32_and(volatile uint32_t *OF_NONNULL p, uint32_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*p &= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_and_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
__asm__ __volatile__ (
|
︙ | | |
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
|
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
|
-
+
|
return OSAtomicAnd32Barrier(i, p);
#else
# error of_atomic_int32_and not implemented!
#endif
}
static OF_INLINE unsigned int
of_atomic_int_xor(volatile __nonnull unsigned int *p, unsigned int i)
of_atomic_int_xor(volatile unsigned int *OF_NONNULL p, unsigned int i)
{
#if !defined(OF_HAVE_THREADS)
return (*p ^= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_xor_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
if (sizeof(int) == 4)
|
︙ | | |
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
|
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
|
-
+
|
return OSAtomicXor32Barrier(i, p);
#else
# error of_atomic_int_xor not implemented!
#endif
}
static OF_INLINE uint32_t
of_atomic_int32_xor(volatile __nonnull uint32_t *p, uint32_t i)
of_atomic_int32_xor(volatile uint32_t *OF_NONNULL p, uint32_t i)
{
#if !defined(OF_HAVE_THREADS)
return (*p ^= i);
#elif defined(OF_HAVE_GCC_ATOMIC_OPS)
return __sync_xor_and_fetch(p, i);
#elif defined(OF_X86_64_ASM) || defined(OF_X86_ASM)
__asm__ __volatile__ (
|
︙ | | |
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
|
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
|
-
+
|
return OSAtomicXor32Barrier(i, p);
#else
# error of_atomic_int32_xor not implemented!
#endif
}
static OF_INLINE bool
of_atomic_int_cmpswap(volatile __nonnull int *p, int o, int n)
of_atomic_int_cmpswap(volatile int *OF_NONNULL p, int o, int n)
{
#if !defined(OF_HAVE_THREADS)
if (*p == o) {
*p = n;
return true;
}
|
︙ | | |
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
|
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
|
-
+
|
return OSAtomicCompareAndSwapIntBarrier(o, n, p);
#else
# error of_atomic_int_cmpswap not implemented!
#endif
}
static OF_INLINE bool
of_atomic_int32_cmpswap(volatile __nonnull int32_t *p, int32_t o, int32_t n)
of_atomic_int32_cmpswap(volatile int32_t *OF_NONNULL p, int32_t o, int32_t n)
{
#if !defined(OF_HAVE_THREADS)
if (*p == o) {
*p = n;
return true;
}
|
︙ | | |
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
|
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
|
-
-
+
+
|
return OSAtomicCompareAndSwap32Barrier(o, n, p);
#else
# error of_atomic_int32_cmpswap not implemented!
#endif
}
static OF_INLINE bool
of_atomic_ptr_cmpswap(__nullable void *volatile *__nonnull p,
__nullable void *o, __nullable void *n)
of_atomic_ptr_cmpswap(void *volatile OF_NULLABLE *OF_NONNULL p,
void *OF_NULLABLE o, void *OF_NULLABLE n)
{
#if !defined(OF_HAVE_THREADS)
if (*p == o) {
*p = n;
return true;
}
|
︙ | | |