@@ -12919,28 +12919,28 @@ void __ovld prefetch(const __global half16 *p, size_t num_elements);
12919
12919
* pointed by p. The function returns old.
12920
12920
*/
12921
12921
int __ovld atomic_add(volatile __global int *p, int val);
12922
- unsigned int __ovld atomic_add(volatile __global unsigned int *p, unsigned int val);
12922
+ uint __ovld atomic_add(volatile __global uint *p, uint val);
12923
12923
int __ovld atomic_add(volatile __local int *p, int val);
12924
- unsigned int __ovld atomic_add(volatile __local unsigned int *p, unsigned int val);
12924
+ uint __ovld atomic_add(volatile __local uint *p, uint val);
12925
12925
#ifdef __OPENCL_CPP_VERSION__
12926
12926
int __ovld atomic_add(volatile int *p, int val);
12927
- unsigned int __ovld atomic_add(volatile unsigned int *p, unsigned int val);
12927
+ uint __ovld atomic_add(volatile uint *p, uint val);
12928
12928
#endif
12929
12929
12930
12930
#if defined(cl_khr_global_int32_base_atomics)
12931
12931
int __ovld atom_add(volatile __global int *p, int val);
12932
- unsigned int __ovld atom_add(volatile __global unsigned int *p, unsigned int val);
12932
+ uint __ovld atom_add(volatile __global uint *p, uint val);
12933
12933
#endif
12934
12934
#if defined(cl_khr_local_int32_base_atomics)
12935
12935
int __ovld atom_add(volatile __local int *p, int val);
12936
- unsigned int __ovld atom_add(volatile __local unsigned int *p, unsigned int val);
12936
+ uint __ovld atom_add(volatile __local uint *p, uint val);
12937
12937
#endif
12938
12938
12939
12939
#if defined(cl_khr_int64_base_atomics)
12940
12940
long __ovld atom_add(volatile __global long *p, long val);
12941
- unsigned long __ovld atom_add(volatile __global unsigned long *p, unsigned long val);
12941
+ ulong __ovld atom_add(volatile __global ulong *p, ulong val);
12942
12942
long __ovld atom_add(volatile __local long *p, long val);
12943
- unsigned long __ovld atom_add(volatile __local unsigned long *p, unsigned long val);
12943
+ ulong __ovld atom_add(volatile __local ulong *p, ulong val);
12944
12944
#endif
12945
12945
12946
12946
/**
@@ -12949,28 +12949,28 @@ unsigned long __ovld atom_add(volatile __local unsigned long *p, unsigned long v
12949
12949
* returns old.
12950
12950
*/
12951
12951
int __ovld atomic_sub(volatile __global int *p, int val);
12952
- unsigned int __ovld atomic_sub(volatile __global unsigned int *p, unsigned int val);
12952
+ uint __ovld atomic_sub(volatile __global uint *p, uint val);
12953
12953
int __ovld atomic_sub(volatile __local int *p, int val);
12954
- unsigned int __ovld atomic_sub(volatile __local unsigned int *p, unsigned int val);
12954
+ uint __ovld atomic_sub(volatile __local uint *p, uint val);
12955
12955
#ifdef __OPENCL_CPP_VERSION__
12956
12956
int __ovld atomic_sub(volatile int *p, int val);
12957
- unsigned int __ovld atomic_sub(volatile unsigned int *p, unsigned int val);
12957
+ uint __ovld atomic_sub(volatile uint *p, uint val);
12958
12958
#endif
12959
12959
12960
12960
#if defined(cl_khr_global_int32_base_atomics)
12961
12961
int __ovld atom_sub(volatile __global int *p, int val);
12962
- unsigned int __ovld atom_sub(volatile __global unsigned int *p, unsigned int val);
12962
+ uint __ovld atom_sub(volatile __global uint *p, uint val);
12963
12963
#endif
12964
12964
#if defined(cl_khr_local_int32_base_atomics)
12965
12965
int __ovld atom_sub(volatile __local int *p, int val);
12966
- unsigned int __ovld atom_sub(volatile __local unsigned int *p, unsigned int val);
12966
+ uint __ovld atom_sub(volatile __local uint *p, uint val);
12967
12967
#endif
12968
12968
12969
12969
#if defined(cl_khr_int64_base_atomics)
12970
12970
long __ovld atom_sub(volatile __global long *p, long val);
12971
- unsigned long __ovld atom_sub(volatile __global unsigned long *p, unsigned long val);
12971
+ ulong __ovld atom_sub(volatile __global ulong *p, ulong val);
12972
12972
long __ovld atom_sub(volatile __local long *p, long val);
12973
- unsigned long __ovld atom_sub(volatile __local unsigned long *p, unsigned long val);
12973
+ ulong __ovld atom_sub(volatile __local ulong *p, ulong val);
12974
12974
#endif
12975
12975
12976
12976
/**
@@ -12979,31 +12979,31 @@ unsigned long __ovld atom_sub(volatile __local unsigned long *p, unsigned long v
12979
12979
* value.
12980
12980
*/
12981
12981
int __ovld atomic_xchg(volatile __global int *p, int val);
12982
- unsigned int __ovld atomic_xchg(volatile __global unsigned int *p, unsigned int val);
12982
+ uint __ovld atomic_xchg(volatile __global uint *p, uint val);
12983
12983
int __ovld atomic_xchg(volatile __local int *p, int val);
12984
- unsigned int __ovld atomic_xchg(volatile __local unsigned int *p, unsigned int val);
12984
+ uint __ovld atomic_xchg(volatile __local uint *p, uint val);
12985
12985
float __ovld atomic_xchg(volatile __global float *p, float val);
12986
12986
float __ovld atomic_xchg(volatile __local float *p, float val);
12987
12987
#ifdef __OPENCL_CPP_VERSION__
12988
12988
int __ovld atomic_xchg(volatile int *p, int val);
12989
- unsigned int __ovld atomic_xchg(volatile unsigned int *p, unsigned int val);
12989
+ uint __ovld atomic_xchg(volatile uint *p, uint val);
12990
12990
float __ovld atomic_xchg(volatile float *p, float val);
12991
12991
#endif
12992
12992
12993
12993
#if defined(cl_khr_global_int32_base_atomics)
12994
12994
int __ovld atom_xchg(volatile __global int *p, int val);
12995
- unsigned int __ovld atom_xchg(volatile __global unsigned int *p, unsigned int val);
12995
+ uint __ovld atom_xchg(volatile __global uint *p, uint val);
12996
12996
#endif
12997
12997
#if defined(cl_khr_local_int32_base_atomics)
12998
12998
int __ovld atom_xchg(volatile __local int *p, int val);
12999
- unsigned int __ovld atom_xchg(volatile __local unsigned int *p, unsigned int val);
12999
+ uint __ovld atom_xchg(volatile __local uint *p, uint val);
13000
13000
#endif
13001
13001
13002
13002
#if defined(cl_khr_int64_base_atomics)
13003
13003
long __ovld atom_xchg(volatile __global long *p, long val);
13004
13004
long __ovld atom_xchg(volatile __local long *p, long val);
13005
- unsigned long __ovld atom_xchg(volatile __global unsigned long *p, unsigned long val);
13006
- unsigned long __ovld atom_xchg(volatile __local unsigned long *p, unsigned long val);
13005
+ ulong __ovld atom_xchg(volatile __global ulong *p, ulong val);
13006
+ ulong __ovld atom_xchg(volatile __local ulong *p, ulong val);
13007
13007
#endif
13008
13008
13009
13009
/**
@@ -13013,28 +13013,28 @@ unsigned long __ovld atom_xchg(volatile __local unsigned long *p, unsigned long
13013
13013
* pointed by p. The function returns old.
13014
13014
*/
13015
13015
int __ovld atomic_inc(volatile __global int *p);
13016
- unsigned int __ovld atomic_inc(volatile __global unsigned int *p);
13016
+ uint __ovld atomic_inc(volatile __global uint *p);
13017
13017
int __ovld atomic_inc(volatile __local int *p);
13018
- unsigned int __ovld atomic_inc(volatile __local unsigned int *p);
13018
+ uint __ovld atomic_inc(volatile __local uint *p);
13019
13019
#ifdef __OPENCL_CPP_VERSION__
13020
13020
int __ovld atomic_inc(volatile int *p);
13021
- unsigned int __ovld atomic_inc(volatile unsigned int *p);
13021
+ uint __ovld atomic_inc(volatile uint *p);
13022
13022
#endif
13023
13023
13024
13024
#if defined(cl_khr_global_int32_base_atomics)
13025
13025
int __ovld atom_inc(volatile __global int *p);
13026
- unsigned int __ovld atom_inc(volatile __global unsigned int *p);
13026
+ uint __ovld atom_inc(volatile __global uint *p);
13027
13027
#endif
13028
13028
#if defined(cl_khr_local_int32_base_atomics)
13029
13029
int __ovld atom_inc(volatile __local int *p);
13030
- unsigned int __ovld atom_inc(volatile __local unsigned int *p);
13030
+ uint __ovld atom_inc(volatile __local uint *p);
13031
13031
#endif
13032
13032
13033
13033
#if defined(cl_khr_int64_base_atomics)
13034
13034
long __ovld atom_inc(volatile __global long *p);
13035
- unsigned long __ovld atom_inc(volatile __global unsigned long *p);
13035
+ ulong __ovld atom_inc(volatile __global ulong *p);
13036
13036
long __ovld atom_inc(volatile __local long *p);
13037
- unsigned long __ovld atom_inc(volatile __local unsigned long *p);
13037
+ ulong __ovld atom_inc(volatile __local ulong *p);
13038
13038
#endif
13039
13039
13040
13040
/**
@@ -13044,28 +13044,28 @@ unsigned long __ovld atom_inc(volatile __local unsigned long *p);
13044
13044
* pointed by p. The function returns old.
13045
13045
*/
13046
13046
int __ovld atomic_dec(volatile __global int *p);
13047
- unsigned int __ovld atomic_dec(volatile __global unsigned int *p);
13047
+ uint __ovld atomic_dec(volatile __global uint *p);
13048
13048
int __ovld atomic_dec(volatile __local int *p);
13049
- unsigned int __ovld atomic_dec(volatile __local unsigned int *p);
13049
+ uint __ovld atomic_dec(volatile __local uint *p);
13050
13050
#ifdef __OPENCL_CPP_VERSION__
13051
13051
int __ovld atomic_dec(volatile int *p);
13052
- unsigned int __ovld atomic_dec(volatile unsigned int *p);
13052
+ uint __ovld atomic_dec(volatile uint *p);
13053
13053
#endif
13054
13054
13055
13055
#if defined(cl_khr_global_int32_base_atomics)
13056
13056
int __ovld atom_dec(volatile __global int *p);
13057
- unsigned int __ovld atom_dec(volatile __global unsigned int *p);
13057
+ uint __ovld atom_dec(volatile __global uint *p);
13058
13058
#endif
13059
13059
#if defined(cl_khr_local_int32_base_atomics)
13060
13060
int __ovld atom_dec(volatile __local int *p);
13061
- unsigned int __ovld atom_dec(volatile __local unsigned int *p);
13061
+ uint __ovld atom_dec(volatile __local uint *p);
13062
13062
#endif
13063
13063
13064
13064
#if defined(cl_khr_int64_base_atomics)
13065
13065
long __ovld atom_dec(volatile __global long *p);
13066
- unsigned long __ovld atom_dec(volatile __global unsigned long *p);
13066
+ ulong __ovld atom_dec(volatile __global ulong *p);
13067
13067
long __ovld atom_dec(volatile __local long *p);
13068
- unsigned long __ovld atom_dec(volatile __local unsigned long *p);
13068
+ ulong __ovld atom_dec(volatile __local ulong *p);
13069
13069
#endif
13070
13070
13071
13071
/**
@@ -13076,28 +13076,28 @@ unsigned long __ovld atom_dec(volatile __local unsigned long *p);
13076
13076
* returns old.
13077
13077
*/
13078
13078
int __ovld atomic_cmpxchg(volatile __global int *p, int cmp, int val);
13079
- unsigned int __ovld atomic_cmpxchg(volatile __global unsigned int *p, unsigned int cmp, unsigned int val);
13079
+ uint __ovld atomic_cmpxchg(volatile __global uint *p, uint cmp, uint val);
13080
13080
int __ovld atomic_cmpxchg(volatile __local int *p, int cmp, int val);
13081
- unsigned int __ovld atomic_cmpxchg(volatile __local unsigned int *p, unsigned int cmp, unsigned int val);
13081
+ uint __ovld atomic_cmpxchg(volatile __local uint *p, uint cmp, uint val);
13082
13082
#ifdef __OPENCL_CPP_VERSION__
13083
13083
int __ovld atomic_cmpxchg(volatile int *p, int cmp, int val);
13084
- unsigned int __ovld atomic_cmpxchg(volatile unsigned int *p, unsigned int cmp, unsigned int val);
13084
+ uint __ovld atomic_cmpxchg(volatile uint *p, uint cmp, uint val);
13085
13085
#endif
13086
13086
13087
13087
#if defined(cl_khr_global_int32_base_atomics)
13088
13088
int __ovld atom_cmpxchg(volatile __global int *p, int cmp, int val);
13089
- unsigned int __ovld atom_cmpxchg(volatile __global unsigned int *p, unsigned int cmp, unsigned int val);
13089
+ uint __ovld atom_cmpxchg(volatile __global uint *p, uint cmp, uint val);
13090
13090
#endif
13091
13091
#if defined(cl_khr_local_int32_base_atomics)
13092
13092
int __ovld atom_cmpxchg(volatile __local int *p, int cmp, int val);
13093
- unsigned int __ovld atom_cmpxchg(volatile __local unsigned int *p, unsigned int cmp, unsigned int val);
13093
+ uint __ovld atom_cmpxchg(volatile __local uint *p, uint cmp, uint val);
13094
13094
#endif
13095
13095
13096
13096
#if defined(cl_khr_int64_base_atomics)
13097
13097
long __ovld atom_cmpxchg(volatile __global long *p, long cmp, long val);
13098
- unsigned long __ovld atom_cmpxchg(volatile __global unsigned long *p, unsigned long cmp, unsigned long val);
13098
+ ulong __ovld atom_cmpxchg(volatile __global ulong *p, ulong cmp, ulong val);
13099
13099
long __ovld atom_cmpxchg(volatile __local long *p, long cmp, long val);
13100
- unsigned long __ovld atom_cmpxchg(volatile __local unsigned long *p, unsigned long cmp, unsigned long val);
13100
+ ulong __ovld atom_cmpxchg(volatile __local ulong *p, ulong cmp, ulong val);
13101
13101
#endif
13102
13102
13103
13103
/**
@@ -13108,28 +13108,28 @@ unsigned long __ovld atom_cmpxchg(volatile __local unsigned long *p, unsigned lo
13108
13108
* returns old.
13109
13109
*/
13110
13110
int __ovld atomic_min(volatile __global int *p, int val);
13111
- unsigned int __ovld atomic_min(volatile __global unsigned int *p, unsigned int val);
13111
+ uint __ovld atomic_min(volatile __global uint *p, uint val);
13112
13112
int __ovld atomic_min(volatile __local int *p, int val);
13113
- unsigned int __ovld atomic_min(volatile __local unsigned int *p, unsigned int val);
13113
+ uint __ovld atomic_min(volatile __local uint *p, uint val);
13114
13114
#ifdef __OPENCL_CPP_VERSION__
13115
13115
int __ovld atomic_min(volatile int *p, int val);
13116
- unsigned int __ovld atomic_min(volatile unsigned int *p, unsigned int val);
13116
+ uint __ovld atomic_min(volatile uint *p, uint val);
13117
13117
#endif
13118
13118
13119
13119
#if defined(cl_khr_global_int32_extended_atomics)
13120
13120
int __ovld atom_min(volatile __global int *p, int val);
13121
- unsigned int __ovld atom_min(volatile __global unsigned int *p, unsigned int val);
13121
+ uint __ovld atom_min(volatile __global uint *p, uint val);
13122
13122
#endif
13123
13123
#if defined(cl_khr_local_int32_extended_atomics)
13124
13124
int __ovld atom_min(volatile __local int *p, int val);
13125
- unsigned int __ovld atom_min(volatile __local unsigned int *p, unsigned int val);
13125
+ uint __ovld atom_min(volatile __local uint *p, uint val);
13126
13126
#endif
13127
13127
13128
13128
#if defined(cl_khr_int64_extended_atomics)
13129
13129
long __ovld atom_min(volatile __global long *p, long val);
13130
- unsigned long __ovld atom_min(volatile __global unsigned long *p, unsigned long val);
13130
+ ulong __ovld atom_min(volatile __global ulong *p, ulong val);
13131
13131
long __ovld atom_min(volatile __local long *p, long val);
13132
- unsigned long __ovld atom_min(volatile __local unsigned long *p, unsigned long val);
13132
+ ulong __ovld atom_min(volatile __local ulong *p, ulong val);
13133
13133
#endif
13134
13134
13135
13135
/**
@@ -13140,28 +13140,28 @@ unsigned long __ovld atom_min(volatile __local unsigned long *p, unsigned long v
13140
13140
* returns old.
13141
13141
*/
13142
13142
int __ovld atomic_max(volatile __global int *p, int val);
13143
- unsigned int __ovld atomic_max(volatile __global unsigned int *p, unsigned int val);
13143
+ uint __ovld atomic_max(volatile __global uint *p, uint val);
13144
13144
int __ovld atomic_max(volatile __local int *p, int val);
13145
- unsigned int __ovld atomic_max(volatile __local unsigned int *p, unsigned int val);
13145
+ uint __ovld atomic_max(volatile __local uint *p, uint val);
13146
13146
#ifdef __OPENCL_CPP_VERSION__
13147
13147
int __ovld atomic_max(volatile int *p, int val);
13148
- unsigned int __ovld atomic_max(volatile unsigned int *p, unsigned int val);
13148
+ uint __ovld atomic_max(volatile uint *p, uint val);
13149
13149
#endif
13150
13150
13151
13151
#if defined(cl_khr_global_int32_extended_atomics)
13152
13152
int __ovld atom_max(volatile __global int *p, int val);
13153
- unsigned int __ovld atom_max(volatile __global unsigned int *p, unsigned int val);
13153
+ uint __ovld atom_max(volatile __global uint *p, uint val);
13154
13154
#endif
13155
13155
#if defined(cl_khr_local_int32_extended_atomics)
13156
13156
int __ovld atom_max(volatile __local int *p, int val);
13157
- unsigned int __ovld atom_max(volatile __local unsigned int *p, unsigned int val);
13157
+ uint __ovld atom_max(volatile __local uint *p, uint val);
13158
13158
#endif
13159
13159
13160
13160
#if defined(cl_khr_int64_extended_atomics)
13161
13161
long __ovld atom_max(volatile __global long *p, long val);
13162
- unsigned long __ovld atom_max(volatile __global unsigned long *p, unsigned long val);
13162
+ ulong __ovld atom_max(volatile __global ulong *p, ulong val);
13163
13163
long __ovld atom_max(volatile __local long *p, long val);
13164
- unsigned long __ovld atom_max(volatile __local unsigned long *p, unsigned long val);
13164
+ ulong __ovld atom_max(volatile __local ulong *p, ulong val);
13165
13165
#endif
13166
13166
13167
13167
/**
@@ -13171,28 +13171,28 @@ unsigned long __ovld atom_max(volatile __local unsigned long *p, unsigned long v
13171
13171
* pointed by p. The function returns old.
13172
13172
*/
13173
13173
int __ovld atomic_and(volatile __global int *p, int val);
13174
- unsigned int __ovld atomic_and(volatile __global unsigned int *p, unsigned int val);
13174
+ uint __ovld atomic_and(volatile __global uint *p, uint val);
13175
13175
int __ovld atomic_and(volatile __local int *p, int val);
13176
- unsigned int __ovld atomic_and(volatile __local unsigned int *p, unsigned int val);
13176
+ uint __ovld atomic_and(volatile __local uint *p, uint val);
13177
13177
#ifdef __OPENCL_CPP_VERSION__
13178
13178
int __ovld atomic_and(volatile int *p, int val);
13179
- unsigned int __ovld atomic_and(volatile unsigned int *p, unsigned int val);
13179
+ uint __ovld atomic_and(volatile uint *p, uint val);
13180
13180
#endif
13181
13181
13182
13182
#if defined(cl_khr_global_int32_extended_atomics)
13183
13183
int __ovld atom_and(volatile __global int *p, int val);
13184
- unsigned int __ovld atom_and(volatile __global unsigned int *p, unsigned int val);
13184
+ uint __ovld atom_and(volatile __global uint *p, uint val);
13185
13185
#endif
13186
13186
#if defined(cl_khr_local_int32_extended_atomics)
13187
13187
int __ovld atom_and(volatile __local int *p, int val);
13188
- unsigned int __ovld atom_and(volatile __local unsigned int *p, unsigned int val);
13188
+ uint __ovld atom_and(volatile __local uint *p, uint val);
13189
13189
#endif
13190
13190
13191
13191
#if defined(cl_khr_int64_extended_atomics)
13192
13192
long __ovld atom_and(volatile __global long *p, long val);
13193
- unsigned long __ovld atom_and(volatile __global unsigned long *p, unsigned long val);
13193
+ ulong __ovld atom_and(volatile __global ulong *p, ulong val);
13194
13194
long __ovld atom_and(volatile __local long *p, long val);
13195
- unsigned long __ovld atom_and(volatile __local unsigned long *p, unsigned long val);
13195
+ ulong __ovld atom_and(volatile __local ulong *p, ulong val);
13196
13196
#endif
13197
13197
13198
13198
/**
@@ -13202,28 +13202,28 @@ unsigned long __ovld atom_and(volatile __local unsigned long *p, unsigned long v
13202
13202
* pointed by p. The function returns old.
13203
13203
*/
13204
13204
int __ovld atomic_or(volatile __global int *p, int val);
13205
- unsigned int __ovld atomic_or(volatile __global unsigned int *p, unsigned int val);
13205
+ uint __ovld atomic_or(volatile __global uint *p, uint val);
13206
13206
int __ovld atomic_or(volatile __local int *p, int val);
13207
- unsigned int __ovld atomic_or(volatile __local unsigned int *p, unsigned int val);
13207
+ uint __ovld atomic_or(volatile __local uint *p, uint val);
13208
13208
#ifdef __OPENCL_CPP_VERSION__
13209
13209
int __ovld atomic_or(volatile int *p, int val);
13210
- unsigned int __ovld atomic_or(volatile unsigned int *p, unsigned int val);
13210
+ uint __ovld atomic_or(volatile uint *p, uint val);
13211
13211
#endif
13212
13212
13213
13213
#if defined(cl_khr_global_int32_extended_atomics)
13214
13214
int __ovld atom_or(volatile __global int *p, int val);
13215
- unsigned int __ovld atom_or(volatile __global unsigned int *p, unsigned int val);
13215
+ uint __ovld atom_or(volatile __global uint *p, uint val);
13216
13216
#endif
13217
13217
#if defined(cl_khr_local_int32_extended_atomics)
13218
13218
int __ovld atom_or(volatile __local int *p, int val);
13219
- unsigned int __ovld atom_or(volatile __local unsigned int *p, unsigned int val);
13219
+ uint __ovld atom_or(volatile __local uint *p, uint val);
13220
13220
#endif
13221
13221
13222
13222
#if defined(cl_khr_int64_extended_atomics)
13223
13223
long __ovld atom_or(volatile __global long *p, long val);
13224
- unsigned long __ovld atom_or(volatile __global unsigned long *p, unsigned long val);
13224
+ ulong __ovld atom_or(volatile __global ulong *p, ulong val);
13225
13225
long __ovld atom_or(volatile __local long *p, long val);
13226
- unsigned long __ovld atom_or(volatile __local unsigned long *p, unsigned long val);
13226
+ ulong __ovld atom_or(volatile __local ulong *p, ulong val);
13227
13227
#endif
13228
13228
13229
13229
/**
@@ -13233,28 +13233,28 @@ unsigned long __ovld atom_or(volatile __local unsigned long *p, unsigned long va
13233
13233
* pointed by p. The function returns old.
13234
13234
*/
13235
13235
int __ovld atomic_xor(volatile __global int *p, int val);
13236
- unsigned int __ovld atomic_xor(volatile __global unsigned int *p, unsigned int val);
13236
+ uint __ovld atomic_xor(volatile __global uint *p, uint val);
13237
13237
int __ovld atomic_xor(volatile __local int *p, int val);
13238
- unsigned int __ovld atomic_xor(volatile __local unsigned int *p, unsigned int val);
13238
+ uint __ovld atomic_xor(volatile __local uint *p, uint val);
13239
13239
#ifdef __OPENCL_CPP_VERSION__
13240
13240
int __ovld atomic_xor(volatile int *p, int val);
13241
- unsigned int __ovld atomic_xor(volatile unsigned int *p, unsigned int val);
13241
+ uint __ovld atomic_xor(volatile uint *p, uint val);
13242
13242
#endif
13243
13243
13244
13244
#if defined(cl_khr_global_int32_extended_atomics)
13245
13245
int __ovld atom_xor(volatile __global int *p, int val);
13246
- unsigned int __ovld atom_xor(volatile __global unsigned int *p, unsigned int val);
13246
+ uint __ovld atom_xor(volatile __global uint *p, uint val);
13247
13247
#endif
13248
13248
#if defined(cl_khr_local_int32_extended_atomics)
13249
13249
int __ovld atom_xor(volatile __local int *p, int val);
13250
- unsigned int __ovld atom_xor(volatile __local unsigned int *p, unsigned int val);
13250
+ uint __ovld atom_xor(volatile __local uint *p, uint val);
13251
13251
#endif
13252
13252
13253
13253
#if defined(cl_khr_int64_extended_atomics)
13254
13254
long __ovld atom_xor(volatile __global long *p, long val);
13255
- unsigned long __ovld atom_xor(volatile __global unsigned long *p, unsigned long val);
13255
+ ulong __ovld atom_xor(volatile __global ulong *p, ulong val);
13256
13256
long __ovld atom_xor(volatile __local long *p, long val);
13257
- unsigned long __ovld atom_xor(volatile __local unsigned long *p, unsigned long val);
13257
+ ulong __ovld atom_xor(volatile __local ulong *p, ulong val);
13258
13258
#endif
13259
13259
13260
13260
#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics)
0 commit comments