@@ -183,8 +183,8 @@ struct __subgroup_radix_sort
183
183
oneapi::dpl::__ranges::__require_access (__cgh, __src);
184
184
185
185
// Get accessors to sycl::buffer (global memory) or sycl::local_accessor (local memory)
186
- auto __acc_buf_val = __buf_val.get_acc (__cgh);
187
- auto __acc_buf_count = __buf_count.get_acc (__cgh);
186
+ auto __exchange_lacc = __buf_val.get_acc (__cgh);
187
+ auto __counter_lacc = __buf_count.get_acc (__cgh);
188
188
189
189
const auto __fence_buf_val = __buf_val._space ;
190
190
const auto __fence_buf_count = __buf_count._space ;
@@ -215,7 +215,7 @@ struct __subgroup_radix_sort
215
215
216
216
// 1. "counting" phase
217
217
// counter initialization
218
- _CounterT* __pcounter = __dpl_sycl::__get_accessor_ptr (__acc_buf_count ) + __wi;
218
+ _CounterT* __pcounter = __dpl_sycl::__get_accessor_ptr (__counter_lacc ) + __wi;
219
219
220
220
_ONEDPL_PRAGMA_UNROLL
221
221
for (uint16_t __i = 0 ; __i < __bin_count; ++__i)
@@ -245,11 +245,11 @@ struct __subgroup_radix_sort
245
245
246
246
// scan contiguous numbers
247
247
uint16_t __bin_sum[__bin_count];
248
- __bin_sum[0 ] = __acc_buf_count [__wi * __bin_count];
248
+ __bin_sum[0 ] = __counter_lacc [__wi * __bin_count];
249
249
250
250
_ONEDPL_PRAGMA_UNROLL
251
251
for (uint16_t __i = 1 ; __i < __bin_count; ++__i)
252
- __bin_sum[__i] = __bin_sum[__i - 1 ] + __acc_buf_count [__wi * __bin_count + __i];
252
+ __bin_sum[__i] = __bin_sum[__i - 1 ] + __counter_lacc [__wi * __bin_count + __i];
253
253
254
254
__dpl_sycl::__group_barrier (__it /* , sycl::access::fence_space::local_space */ );
255
255
// exclusive scan local sum
@@ -258,10 +258,10 @@ struct __subgroup_radix_sort
258
258
// add to local sum, generate exclusive scan result
259
259
_ONEDPL_PRAGMA_UNROLL
260
260
for (uint16_t __i = 0 ; __i < __bin_count; ++__i)
261
- __acc_buf_count [__wi * __bin_count + __i + 1 ] = __sum_scan + __bin_sum[__i];
261
+ __counter_lacc [__wi * __bin_count + __i + 1 ] = __sum_scan + __bin_sum[__i];
262
262
263
263
if (__wi == 0 )
264
- __acc_buf_count [0 ] = 0 ;
264
+ __counter_lacc [0 ] = 0 ;
265
265
__dpl_sycl::__group_barrier (__it, __fence_buf_count);
266
266
}
267
267
@@ -298,7 +298,7 @@ struct __subgroup_radix_sort
298
298
{
299
299
const uint16_t __idx = __wi * __block_size + __i;
300
300
if (__idx < __n)
301
- __acc_buf_val [__idx].~_ValT ();
301
+ __exchange_lacc [__idx].~_ValT ();
302
302
}
303
303
304
304
return ;
@@ -312,7 +312,7 @@ struct __subgroup_radix_sort
312
312
{
313
313
const uint16_t __r = __indices[__i];
314
314
if (__r < __n)
315
- new (&__acc_buf_val [__r]) _ValT (::std::move (__values.__v [__i]));
315
+ new (&__exchange_lacc [__r]) _ValT (::std::move (__values.__v [__i]));
316
316
}
317
317
}
318
318
else
@@ -322,7 +322,7 @@ struct __subgroup_radix_sort
322
322
{
323
323
const uint16_t __r = __indices[__i];
324
324
if (__r < __n)
325
- __acc_buf_val [__r] = ::std::move (__values.__v [__i]);
325
+ __exchange_lacc [__r] = ::std::move (__values.__v [__i]);
326
326
}
327
327
}
328
328
__dpl_sycl::__group_barrier (__it, __fence_buf_val);
@@ -332,7 +332,7 @@ struct __subgroup_radix_sort
332
332
{
333
333
const uint16_t __idx = __wi * __block_size + __i;
334
334
if (__idx < __n)
335
- __values.__v [__i] = ::std::move (__acc_buf_val [__idx]);
335
+ __values.__v [__i] = ::std::move (__exchange_lacc [__idx]);
336
336
}
337
337
__dpl_sycl::__group_barrier (__it /* , sycl::access::fence_space::local_space */ );
338
338
}
0 commit comments