@@ -332,7 +332,7 @@ GPUdi() GPUTPCCompressionGatherKernels::Vec128* GPUTPCCompressionGatherKernels::
332332template <typename T, typename S>
333333GPUdi () bool GPUTPCCompressionGatherKernels::isAlignedTo(const S* ptr)
334334{
335- if CONSTEXPR (alignof (S) >= alignof (T)) {
335+ if constexpr (alignof (S) >= alignof (T)) {
336336 static_cast <void >(ptr);
337337 return true ;
338338 } else {
@@ -343,10 +343,10 @@ GPUdi() bool GPUTPCCompressionGatherKernels::isAlignedTo(const S* ptr)
343343template <>
344344GPUdi () void GPUTPCCompressionGatherKernels::compressorMemcpy<uint8_t>(uint8_t * GPUrestrict () dst, const uint8_t* GPUrestrict() src, uint32_t size, int32_t nThreads, int32_t iThread)
345345{
346- CONSTEXPR const int32_t vec128Elems = CpyVector<uint8_t , Vec128>::Size;
347- CONSTEXPR const int32_t vec64Elems = CpyVector<uint8_t , Vec64>::Size;
348- CONSTEXPR const int32_t vec32Elems = CpyVector<uint8_t , Vec32>::Size;
349- CONSTEXPR const int32_t vec16Elems = CpyVector<uint8_t , Vec16>::Size;
346+ constexpr const int32_t vec128Elems = CpyVector<uint8_t , Vec128>::Size;
347+ constexpr const int32_t vec64Elems = CpyVector<uint8_t , Vec64>::Size;
348+ constexpr const int32_t vec32Elems = CpyVector<uint8_t , Vec32>::Size;
349+ constexpr const int32_t vec16Elems = CpyVector<uint8_t , Vec16>::Size;
350350
351351 if (size >= uint32_t (nThreads * vec128Elems)) {
352352 compressorMemcpyVectorised<uint8_t , Vec128>(dst, src, size, nThreads, iThread);
@@ -364,9 +364,9 @@ GPUdi() void GPUTPCCompressionGatherKernels::compressorMemcpy<uint8_t>(uint8_t*
364364template <>
365365GPUdi () void GPUTPCCompressionGatherKernels::compressorMemcpy<uint16_t>(uint16_t * GPUrestrict () dst, const uint16_t* GPUrestrict() src, uint32_t size, int32_t nThreads, int32_t iThread)
366366{
367- CONSTEXPR const int32_t vec128Elems = CpyVector<uint16_t , Vec128>::Size;
368- CONSTEXPR const int32_t vec64Elems = CpyVector<uint16_t , Vec64>::Size;
369- CONSTEXPR const int32_t vec32Elems = CpyVector<uint16_t , Vec32>::Size;
367+ constexpr const int32_t vec128Elems = CpyVector<uint16_t , Vec128>::Size;
368+ constexpr const int32_t vec64Elems = CpyVector<uint16_t , Vec64>::Size;
369+ constexpr const int32_t vec32Elems = CpyVector<uint16_t , Vec32>::Size;
370370
371371 if (size >= uint32_t (nThreads * vec128Elems)) {
372372 compressorMemcpyVectorised<uint16_t , Vec128>(dst, src, size, nThreads, iThread);
@@ -382,8 +382,8 @@ GPUdi() void GPUTPCCompressionGatherKernels::compressorMemcpy<uint16_t>(uint16_t
382382template <>
383383GPUdi () void GPUTPCCompressionGatherKernels::compressorMemcpy<uint32_t>(uint32_t * GPUrestrict () dst, const uint32_t* GPUrestrict() src, uint32_t size, int32_t nThreads, int32_t iThread)
384384{
385- CONSTEXPR const int32_t vec128Elems = CpyVector<uint32_t , Vec128>::Size;
386- CONSTEXPR const int32_t vec64Elems = CpyVector<uint32_t , Vec64>::Size;
385+ constexpr const int32_t vec128Elems = CpyVector<uint32_t , Vec128>::Size;
386+ constexpr const int32_t vec64Elems = CpyVector<uint32_t , Vec64>::Size;
387387
388388 if (size >= uint32_t (nThreads * vec128Elems)) {
389389 compressorMemcpyVectorised<uint32_t , Vec128>(dst, src, size, nThreads, iThread);
@@ -446,8 +446,8 @@ GPUdi() void GPUTPCCompressionGatherKernels::compressorMemcpyBuffered(V* buf, T*
446446 V* GPUrestrict () dstAligned = nullptr ;
447447
448448 T* bufT = reinterpret_cast <T*>(buf);
449- CONSTEXPR const int32_t bufSize = GPUCA_WARP_SIZE;
450- CONSTEXPR const int32_t bufTSize = bufSize * sizeof (V) / sizeof (T);
449+ constexpr const int32_t bufSize = GPUCA_WARP_SIZE;
450+ constexpr const int32_t bufTSize = bufSize * sizeof (V) / sizeof (T);
451451
452452 for (uint32_t i = 0 ; i < nEntries; i++) {
453453 uint32_t srcPos = 0 ;
0 commit comments