Skip to content

Commit a25f216

Browse files
committed
No need for atomicAdd for float2, conflicts with CUDA 12.1
1 parent 0fee1c1 commit a25f216

File tree

1 file changed

+1
-5
lines changed

1 file changed

+1
-5
lines changed

src/cudafeat/feature-online-batched-cmvn-cuda-kernels.cu

+1-5
Original file line numberDiff line numberDiff line change
@@ -24,18 +24,14 @@ __host__ __device__ inline float2 operator-(const float2 &a, const float2 &b) {
2424
retval.y = a.y - b.y;
2525
return retval;
2626
}
27+
2728
__host__ __device__ inline float2 operator+(const float2 &a, const float2 &b) {
2829
float2 retval;
2930
retval.x = a.x + b.x;
3031
retval.y = a.y + b.y;
3132
return retval;
3233
}
3334

34-
__device__ inline void atomicAdd(float2 *addr, float2 val) {
35-
atomicAdd(reinterpret_cast<float *>(addr), val.x);
36-
atomicAdd(reinterpret_cast<float *>(addr) + 1, val.y);
37-
}
38-
3935
__device__ inline void operator+=(float2 &a, float2 &b) {
4036
// overloading +=
4137
a.x += b.x;

0 commit comments

Comments
 (0)