summaryrefslogtreecommitdiff
path: root/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels
diff options
context:
space:
mode:
Diffstat (limited to 'thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels')
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/BoundSearchKernelsCL.h171
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/CopyKernelsCL.h261
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/FillKernelsCL.h179
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsCL.h255
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsFloat4CL.h255
-rw-r--r--thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/RadixSort32KernelsCL.h1817
6 files changed, 1466 insertions, 1472 deletions
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/BoundSearchKernelsCL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/BoundSearchKernelsCL.h
index 9c9e847138..1758dd41e3 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/BoundSearchKernelsCL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/BoundSearchKernelsCL.h
@@ -1,87 +1,86 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* boundSearchKernelsCL= \
-"/*\n"
-"Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Originally written by Takahiro Harada\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"typedef struct\n"
-"{\n"
-" u32 m_key; \n"
-" u32 m_value;\n"
-"}SortData;\n"
-"typedef struct\n"
-"{\n"
-" u32 m_nSrc;\n"
-" u32 m_nDst;\n"
-" u32 m_padding[2];\n"
-"} ConstBuffer;\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"__kernel\n"
-"void SearchSortDataLowerKernel(__global SortData* src, __global u32 *dst, \n"
-" unsigned int nSrc, unsigned int nDst)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < nSrc )\n"
-" {\n"
-" SortData first; first.m_key = (u32)(-1); first.m_value = (u32)(-1);\n"
-" SortData end; end.m_key = nDst; end.m_value = nDst;\n"
-" SortData iData = (gIdx==0)? first: src[gIdx-1];\n"
-" SortData jData = (gIdx==nSrc)? end: src[gIdx];\n"
-" if( iData.m_key != jData.m_key )\n"
-" {\n"
-"// for(u32 k=iData.m_key+1; k<=min(jData.m_key, nDst-1); k++)\n"
-" u32 k = jData.m_key;\n"
-" {\n"
-" dst[k] = gIdx;\n"
-" }\n"
-" }\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"__kernel\n"
-"void SearchSortDataUpperKernel(__global SortData* src, __global u32 *dst, \n"
-" unsigned int nSrc, unsigned int nDst)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX+1;\n"
-" if( gIdx < nSrc+1 )\n"
-" {\n"
-" SortData first; first.m_key = 0; first.m_value = 0;\n"
-" SortData end; end.m_key = nDst; end.m_value = nDst;\n"
-" SortData iData = src[gIdx-1];\n"
-" SortData jData = (gIdx==nSrc)? end: src[gIdx];\n"
-" if( iData.m_key != jData.m_key )\n"
-" {\n"
-" u32 k = iData.m_key;\n"
-" {\n"
-" dst[k] = gIdx;\n"
-" }\n"
-" }\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"__kernel\n"
-"void SubtractKernel(__global u32* A, __global u32 *B, __global u32 *C, \n"
-" unsigned int nSrc, unsigned int nDst)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" \n"
-" if( gIdx < nDst )\n"
-" {\n"
-" C[gIdx] = A[gIdx] - B[gIdx];\n"
-" }\n"
-"}\n"
-;
+static const char* boundSearchKernelsCL =
+ "/*\n"
+ "Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Originally written by Takahiro Harada\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "typedef struct\n"
+ "{\n"
+ " u32 m_key; \n"
+ " u32 m_value;\n"
+ "}SortData;\n"
+ "typedef struct\n"
+ "{\n"
+ " u32 m_nSrc;\n"
+ " u32 m_nDst;\n"
+ " u32 m_padding[2];\n"
+ "} ConstBuffer;\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "__kernel\n"
+ "void SearchSortDataLowerKernel(__global SortData* src, __global u32 *dst, \n"
+ " unsigned int nSrc, unsigned int nDst)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < nSrc )\n"
+ " {\n"
+ " SortData first; first.m_key = (u32)(-1); first.m_value = (u32)(-1);\n"
+ " SortData end; end.m_key = nDst; end.m_value = nDst;\n"
+ " SortData iData = (gIdx==0)? first: src[gIdx-1];\n"
+ " SortData jData = (gIdx==nSrc)? end: src[gIdx];\n"
+ " if( iData.m_key != jData.m_key )\n"
+ " {\n"
+ "// for(u32 k=iData.m_key+1; k<=min(jData.m_key, nDst-1); k++)\n"
+ " u32 k = jData.m_key;\n"
+ " {\n"
+ " dst[k] = gIdx;\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "__kernel\n"
+ "void SearchSortDataUpperKernel(__global SortData* src, __global u32 *dst, \n"
+ " unsigned int nSrc, unsigned int nDst)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX+1;\n"
+ " if( gIdx < nSrc+1 )\n"
+ " {\n"
+ " SortData first; first.m_key = 0; first.m_value = 0;\n"
+ " SortData end; end.m_key = nDst; end.m_value = nDst;\n"
+ " SortData iData = src[gIdx-1];\n"
+ " SortData jData = (gIdx==nSrc)? end: src[gIdx];\n"
+ " if( iData.m_key != jData.m_key )\n"
+ " {\n"
+ " u32 k = iData.m_key;\n"
+ " {\n"
+ " dst[k] = gIdx;\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "__kernel\n"
+ "void SubtractKernel(__global u32* A, __global u32 *B, __global u32 *C, \n"
+ " unsigned int nSrc, unsigned int nDst)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " \n"
+ " if( gIdx < nDst )\n"
+ " {\n"
+ " C[gIdx] = A[gIdx] - B[gIdx];\n"
+ " }\n"
+ "}\n";
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/CopyKernelsCL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/CopyKernelsCL.h
index e5670e3cd3..33c9279462 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/CopyKernelsCL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/CopyKernelsCL.h
@@ -1,132 +1,131 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* copyKernelsCL= \
-"/*\n"
-"Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
-"\n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Originally written by Takahiro Harada\n"
-"\n"
-"#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
-"#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
-"\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
-"#define AtomInc(x) atom_inc(&(x))\n"
-"#define AtomInc1(x, out) out = atom_inc(&(x))\n"
-"\n"
-"#define make_uint4 (uint4)\n"
-"#define make_uint2 (uint2)\n"
-"#define make_int2 (int2)\n"
-"\n"
-"typedef struct\n"
-"{\n"
-" int m_n;\n"
-" int m_padding[3];\n"
-"} ConstBuffer;\n"
-"\n"
-"\n"
-"\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void Copy1F4Kernel(__global float4* dst, __global float4* src, \n"
-" ConstBuffer cb)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-"\n"
-" if( gIdx < cb.m_n )\n"
-" {\n"
-" float4 a0 = src[gIdx];\n"
-"\n"
-" dst[ gIdx ] = a0;\n"
-" }\n"
-"}\n"
-"\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void Copy2F4Kernel(__global float4* dst, __global float4* src, \n"
-" ConstBuffer cb)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-"\n"
-" if( 2*gIdx <= cb.m_n )\n"
-" {\n"
-" float4 a0 = src[gIdx*2+0];\n"
-" float4 a1 = src[gIdx*2+1];\n"
-"\n"
-" dst[ gIdx*2+0 ] = a0;\n"
-" dst[ gIdx*2+1 ] = a1;\n"
-" }\n"
-"}\n"
-"\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void Copy4F4Kernel(__global float4* dst, __global float4* src, \n"
-" ConstBuffer cb)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-"\n"
-" if( 4*gIdx <= cb.m_n )\n"
-" {\n"
-" int idx0 = gIdx*4+0;\n"
-" int idx1 = gIdx*4+1;\n"
-" int idx2 = gIdx*4+2;\n"
-" int idx3 = gIdx*4+3;\n"
-"\n"
-" float4 a0 = src[idx0];\n"
-" float4 a1 = src[idx1];\n"
-" float4 a2 = src[idx2];\n"
-" float4 a3 = src[idx3];\n"
-"\n"
-" dst[ idx0 ] = a0;\n"
-" dst[ idx1 ] = a1;\n"
-" dst[ idx2 ] = a2;\n"
-" dst[ idx3 ] = a3;\n"
-" }\n"
-"}\n"
-"\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void CopyF1Kernel(__global float* dstF1, __global float* srcF1, \n"
-" ConstBuffer cb)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-"\n"
-" if( gIdx < cb.m_n )\n"
-" {\n"
-" float a0 = srcF1[gIdx];\n"
-"\n"
-" dstF1[ gIdx ] = a0;\n"
-" }\n"
-"}\n"
-"\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void CopyF2Kernel(__global float2* dstF2, __global float2* srcF2, \n"
-" ConstBuffer cb)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-"\n"
-" if( gIdx < cb.m_n )\n"
-" {\n"
-" float2 a0 = srcF2[gIdx];\n"
-"\n"
-" dstF2[ gIdx ] = a0;\n"
-" }\n"
-"}\n"
-"\n"
-"\n"
-;
+static const char* copyKernelsCL =
+ "/*\n"
+ "Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
+ "\n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Originally written by Takahiro Harada\n"
+ "\n"
+ "#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
+ "#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
+ "\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
+ "#define AtomInc(x) atom_inc(&(x))\n"
+ "#define AtomInc1(x, out) out = atom_inc(&(x))\n"
+ "\n"
+ "#define make_uint4 (uint4)\n"
+ "#define make_uint2 (uint2)\n"
+ "#define make_int2 (int2)\n"
+ "\n"
+ "typedef struct\n"
+ "{\n"
+ " int m_n;\n"
+ " int m_padding[3];\n"
+ "} ConstBuffer;\n"
+ "\n"
+ "\n"
+ "\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void Copy1F4Kernel(__global float4* dst, __global float4* src, \n"
+ " ConstBuffer cb)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ "\n"
+ " if( gIdx < cb.m_n )\n"
+ " {\n"
+ " float4 a0 = src[gIdx];\n"
+ "\n"
+ " dst[ gIdx ] = a0;\n"
+ " }\n"
+ "}\n"
+ "\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void Copy2F4Kernel(__global float4* dst, __global float4* src, \n"
+ " ConstBuffer cb)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ "\n"
+ " if( 2*gIdx <= cb.m_n )\n"
+ " {\n"
+ " float4 a0 = src[gIdx*2+0];\n"
+ " float4 a1 = src[gIdx*2+1];\n"
+ "\n"
+ " dst[ gIdx*2+0 ] = a0;\n"
+ " dst[ gIdx*2+1 ] = a1;\n"
+ " }\n"
+ "}\n"
+ "\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void Copy4F4Kernel(__global float4* dst, __global float4* src, \n"
+ " ConstBuffer cb)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ "\n"
+ " if( 4*gIdx <= cb.m_n )\n"
+ " {\n"
+ " int idx0 = gIdx*4+0;\n"
+ " int idx1 = gIdx*4+1;\n"
+ " int idx2 = gIdx*4+2;\n"
+ " int idx3 = gIdx*4+3;\n"
+ "\n"
+ " float4 a0 = src[idx0];\n"
+ " float4 a1 = src[idx1];\n"
+ " float4 a2 = src[idx2];\n"
+ " float4 a3 = src[idx3];\n"
+ "\n"
+ " dst[ idx0 ] = a0;\n"
+ " dst[ idx1 ] = a1;\n"
+ " dst[ idx2 ] = a2;\n"
+ " dst[ idx3 ] = a3;\n"
+ " }\n"
+ "}\n"
+ "\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void CopyF1Kernel(__global float* dstF1, __global float* srcF1, \n"
+ " ConstBuffer cb)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ "\n"
+ " if( gIdx < cb.m_n )\n"
+ " {\n"
+ " float a0 = srcF1[gIdx];\n"
+ "\n"
+ " dstF1[ gIdx ] = a0;\n"
+ " }\n"
+ "}\n"
+ "\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void CopyF2Kernel(__global float2* dstF2, __global float2* srcF2, \n"
+ " ConstBuffer cb)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ "\n"
+ " if( gIdx < cb.m_n )\n"
+ " {\n"
+ " float2 a0 = srcF2[gIdx];\n"
+ "\n"
+ " dstF2[ gIdx ] = a0;\n"
+ " }\n"
+ "}\n"
+ "\n"
+ "\n";
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/FillKernelsCL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/FillKernelsCL.h
index 4f8b96e489..983e652270 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/FillKernelsCL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/FillKernelsCL.h
@@ -1,91 +1,90 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* fillKernelsCL= \
-"/*\n"
-"Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Originally written by Takahiro Harada\n"
-"#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
-"#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
-"#define AtomInc(x) atom_inc(&(x))\n"
-"#define AtomInc1(x, out) out = atom_inc(&(x))\n"
-"#define make_uint4 (uint4)\n"
-"#define make_uint2 (uint2)\n"
-"#define make_int2 (int2)\n"
-"typedef struct\n"
-"{\n"
-" union\n"
-" {\n"
-" int4 m_data;\n"
-" uint4 m_unsignedData;\n"
-" float m_floatData;\n"
-" };\n"
-" int m_offset;\n"
-" int m_n;\n"
-" int m_padding[2];\n"
-"} ConstBuffer;\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void FillIntKernel(__global int* dstInt, int num_elements, int value, const int offset)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < num_elements )\n"
-" {\n"
-" dstInt[ offset+gIdx ] = value;\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void FillFloatKernel(__global float* dstFloat, int num_elements, float value, const int offset)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < num_elements )\n"
-" {\n"
-" dstFloat[ offset+gIdx ] = value;\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void FillUnsignedIntKernel(__global unsigned int* dstInt, const int num, const unsigned int value, const int offset)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < num )\n"
-" {\n"
-" dstInt[ offset+gIdx ] = value;\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void FillInt2Kernel(__global int2* dstInt2, const int num, const int2 value, const int offset)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < num )\n"
-" {\n"
-" dstInt2[ gIdx + offset] = make_int2( value.x, value.y );\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(64,1,1)))\n"
-"void FillInt4Kernel(__global int4* dstInt4, const int num, const int4 value, const int offset)\n"
-"{\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" if( gIdx < num )\n"
-" {\n"
-" dstInt4[ offset+gIdx ] = value;\n"
-" }\n"
-"}\n"
-;
+static const char* fillKernelsCL =
+ "/*\n"
+ "Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Originally written by Takahiro Harada\n"
+ "#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
+ "#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
+ "#define AtomInc(x) atom_inc(&(x))\n"
+ "#define AtomInc1(x, out) out = atom_inc(&(x))\n"
+ "#define make_uint4 (uint4)\n"
+ "#define make_uint2 (uint2)\n"
+ "#define make_int2 (int2)\n"
+ "typedef struct\n"
+ "{\n"
+ " union\n"
+ " {\n"
+ " int4 m_data;\n"
+ " uint4 m_unsignedData;\n"
+ " float m_floatData;\n"
+ " };\n"
+ " int m_offset;\n"
+ " int m_n;\n"
+ " int m_padding[2];\n"
+ "} ConstBuffer;\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void FillIntKernel(__global int* dstInt, int num_elements, int value, const int offset)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < num_elements )\n"
+ " {\n"
+ " dstInt[ offset+gIdx ] = value;\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void FillFloatKernel(__global float* dstFloat, int num_elements, float value, const int offset)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < num_elements )\n"
+ " {\n"
+ " dstFloat[ offset+gIdx ] = value;\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void FillUnsignedIntKernel(__global unsigned int* dstInt, const int num, const unsigned int value, const int offset)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < num )\n"
+ " {\n"
+ " dstInt[ offset+gIdx ] = value;\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void FillInt2Kernel(__global int2* dstInt2, const int num, const int2 value, const int offset)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < num )\n"
+ " {\n"
+ " dstInt2[ gIdx + offset] = make_int2( value.x, value.y );\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "void FillInt4Kernel(__global int4* dstInt4, const int num, const int4 value, const int offset)\n"
+ "{\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " if( gIdx < num )\n"
+ " {\n"
+ " dstInt4[ offset+gIdx ] = value;\n"
+ " }\n"
+ "}\n";
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsCL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsCL.h
index 27baab8331..fc5e7b865c 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsCL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsCL.h
@@ -1,129 +1,128 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* prefixScanKernelsCL= \
-"/*\n"
-"Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Originally written by Takahiro Harada\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"// takahiro end\n"
-"#define WG_SIZE 128 \n"
-"#define m_numElems x\n"
-"#define m_numBlocks y\n"
-"#define m_numScanBlocks z\n"
-"/*typedef struct\n"
-"{\n"
-" uint m_numElems;\n"
-" uint m_numBlocks;\n"
-" uint m_numScanBlocks;\n"
-" uint m_padding[1];\n"
-"} ConstBuffer;\n"
-"*/\n"
-"u32 ScanExclusive(__local u32* data, u32 n, int lIdx, int lSize)\n"
-"{\n"
-" u32 blocksum;\n"
-" int offset = 1;\n"
-" for(int nActive=n>>1; nActive>0; nActive>>=1, offset<<=1)\n"
-" {\n"
-" GROUP_LDS_BARRIER;\n"
-" for(int iIdx=lIdx; iIdx<nActive; iIdx+=lSize)\n"
-" {\n"
-" int ai = offset*(2*iIdx+1)-1;\n"
-" int bi = offset*(2*iIdx+2)-1;\n"
-" data[bi] += data[ai];\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" if( lIdx == 0 )\n"
-" {\n"
-" blocksum = data[ n-1 ];\n"
-" data[ n-1 ] = 0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" offset >>= 1;\n"
-" for(int nActive=1; nActive<n; nActive<<=1, offset>>=1 )\n"
-" {\n"
-" GROUP_LDS_BARRIER;\n"
-" for( int iIdx = lIdx; iIdx<nActive; iIdx += lSize )\n"
-" {\n"
-" int ai = offset*(2*iIdx+1)-1;\n"
-" int bi = offset*(2*iIdx+2)-1;\n"
-" u32 temp = data[ai];\n"
-" data[ai] = data[bi];\n"
-" data[bi] += temp;\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" return blocksum;\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void LocalScanKernel(__global u32* dst, __global u32 *src, __global u32 *sumBuffer,\n"
-" uint4 cb)\n"
-"{\n"
-" __local u32 ldsData[WG_SIZE*2];\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" ldsData[2*lIdx] = ( 2*gIdx < cb.m_numElems )? src[2*gIdx]: 0;\n"
-" ldsData[2*lIdx + 1] = ( 2*gIdx+1 < cb.m_numElems )? src[2*gIdx + 1]: 0;\n"
-" u32 sum = ScanExclusive(ldsData, WG_SIZE*2, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
-" if( lIdx == 0 ) sumBuffer[GET_GROUP_IDX] = sum;\n"
-" if( (2*gIdx) < cb.m_numElems )\n"
-" {\n"
-" dst[2*gIdx] = ldsData[2*lIdx];\n"
-" }\n"
-" if( (2*gIdx + 1) < cb.m_numElems )\n"
-" {\n"
-" dst[2*gIdx + 1] = ldsData[2*lIdx + 1];\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void AddOffsetKernel(__global u32 *dst, __global u32 *blockSum, uint4 cb)\n"
-"{\n"
-" const u32 blockSize = WG_SIZE*2;\n"
-" int myIdx = GET_GROUP_IDX+1;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" u32 iBlockSum = blockSum[myIdx];\n"
-" int endValue = min((myIdx+1)*(blockSize), cb.m_numElems);\n"
-" for(int i=myIdx*blockSize+lIdx; i<endValue; i+=GET_GROUP_SIZE)\n"
-" {\n"
-" dst[i] += iBlockSum;\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void TopLevelScanKernel(__global u32* dst, uint4 cb)\n"
-"{\n"
-" __local u32 ldsData[2048];\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" int lSize = GET_GROUP_SIZE;\n"
-" for(int i=lIdx; i<cb.m_numScanBlocks; i+=lSize )\n"
-" {\n"
-" ldsData[i] = (i<cb.m_numBlocks)? dst[i]:0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" u32 sum = ScanExclusive(ldsData, cb.m_numScanBlocks, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
-" for(int i=lIdx; i<cb.m_numBlocks; i+=lSize )\n"
-" {\n"
-" dst[i] = ldsData[i];\n"
-" }\n"
-" if( gIdx == 0 )\n"
-" {\n"
-" dst[cb.m_numBlocks] = sum;\n"
-" }\n"
-"}\n"
-;
+static const char* prefixScanKernelsCL =
+ "/*\n"
+ "Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Originally written by Takahiro Harada\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "// takahiro end\n"
+ "#define WG_SIZE 128 \n"
+ "#define m_numElems x\n"
+ "#define m_numBlocks y\n"
+ "#define m_numScanBlocks z\n"
+ "/*typedef struct\n"
+ "{\n"
+ " uint m_numElems;\n"
+ " uint m_numBlocks;\n"
+ " uint m_numScanBlocks;\n"
+ " uint m_padding[1];\n"
+ "} ConstBuffer;\n"
+ "*/\n"
+ "u32 ScanExclusive(__local u32* data, u32 n, int lIdx, int lSize)\n"
+ "{\n"
+ " u32 blocksum;\n"
+ " int offset = 1;\n"
+ " for(int nActive=n>>1; nActive>0; nActive>>=1, offset<<=1)\n"
+ " {\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for(int iIdx=lIdx; iIdx<nActive; iIdx+=lSize)\n"
+ " {\n"
+ " int ai = offset*(2*iIdx+1)-1;\n"
+ " int bi = offset*(2*iIdx+2)-1;\n"
+ " data[bi] += data[ai];\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " if( lIdx == 0 )\n"
+ " {\n"
+ " blocksum = data[ n-1 ];\n"
+ " data[ n-1 ] = 0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " offset >>= 1;\n"
+ " for(int nActive=1; nActive<n; nActive<<=1, offset>>=1 )\n"
+ " {\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for( int iIdx = lIdx; iIdx<nActive; iIdx += lSize )\n"
+ " {\n"
+ " int ai = offset*(2*iIdx+1)-1;\n"
+ " int bi = offset*(2*iIdx+2)-1;\n"
+ " u32 temp = data[ai];\n"
+ " data[ai] = data[bi];\n"
+ " data[bi] += temp;\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " return blocksum;\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void LocalScanKernel(__global u32* dst, __global u32 *src, __global u32 *sumBuffer,\n"
+ " uint4 cb)\n"
+ "{\n"
+ " __local u32 ldsData[WG_SIZE*2];\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " ldsData[2*lIdx] = ( 2*gIdx < cb.m_numElems )? src[2*gIdx]: 0;\n"
+ " ldsData[2*lIdx + 1] = ( 2*gIdx+1 < cb.m_numElems )? src[2*gIdx + 1]: 0;\n"
+ " u32 sum = ScanExclusive(ldsData, WG_SIZE*2, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
+ " if( lIdx == 0 ) sumBuffer[GET_GROUP_IDX] = sum;\n"
+ " if( (2*gIdx) < cb.m_numElems )\n"
+ " {\n"
+ " dst[2*gIdx] = ldsData[2*lIdx];\n"
+ " }\n"
+ " if( (2*gIdx + 1) < cb.m_numElems )\n"
+ " {\n"
+ " dst[2*gIdx + 1] = ldsData[2*lIdx + 1];\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void AddOffsetKernel(__global u32 *dst, __global u32 *blockSum, uint4 cb)\n"
+ "{\n"
+ " const u32 blockSize = WG_SIZE*2;\n"
+ " int myIdx = GET_GROUP_IDX+1;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " u32 iBlockSum = blockSum[myIdx];\n"
+ " int endValue = min((myIdx+1)*(blockSize), cb.m_numElems);\n"
+ " for(int i=myIdx*blockSize+lIdx; i<endValue; i+=GET_GROUP_SIZE)\n"
+ " {\n"
+ " dst[i] += iBlockSum;\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void TopLevelScanKernel(__global u32* dst, uint4 cb)\n"
+ "{\n"
+ " __local u32 ldsData[2048];\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " int lSize = GET_GROUP_SIZE;\n"
+ " for(int i=lIdx; i<cb.m_numScanBlocks; i+=lSize )\n"
+ " {\n"
+ " ldsData[i] = (i<cb.m_numBlocks)? dst[i]:0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " u32 sum = ScanExclusive(ldsData, cb.m_numScanBlocks, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
+ " for(int i=lIdx; i<cb.m_numBlocks; i+=lSize )\n"
+ " {\n"
+ " dst[i] = ldsData[i];\n"
+ " }\n"
+ " if( gIdx == 0 )\n"
+ " {\n"
+ " dst[cb.m_numBlocks] = sum;\n"
+ " }\n"
+ "}\n";
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsFloat4CL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsFloat4CL.h
index 5b13254796..15d1bc5195 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsFloat4CL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/PrefixScanKernelsFloat4CL.h
@@ -1,129 +1,128 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* prefixScanKernelsFloat4CL= \
-"/*\n"
-"Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Originally written by Takahiro Harada\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"// takahiro end\n"
-"#define WG_SIZE 128 \n"
-"#define m_numElems x\n"
-"#define m_numBlocks y\n"
-"#define m_numScanBlocks z\n"
-"/*typedef struct\n"
-"{\n"
-" uint m_numElems;\n"
-" uint m_numBlocks;\n"
-" uint m_numScanBlocks;\n"
-" uint m_padding[1];\n"
-"} ConstBuffer;\n"
-"*/\n"
-"float4 ScanExclusiveFloat4(__local float4* data, u32 n, int lIdx, int lSize)\n"
-"{\n"
-" float4 blocksum;\n"
-" int offset = 1;\n"
-" for(int nActive=n>>1; nActive>0; nActive>>=1, offset<<=1)\n"
-" {\n"
-" GROUP_LDS_BARRIER;\n"
-" for(int iIdx=lIdx; iIdx<nActive; iIdx+=lSize)\n"
-" {\n"
-" int ai = offset*(2*iIdx+1)-1;\n"
-" int bi = offset*(2*iIdx+2)-1;\n"
-" data[bi] += data[ai];\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" if( lIdx == 0 )\n"
-" {\n"
-" blocksum = data[ n-1 ];\n"
-" data[ n-1 ] = 0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" offset >>= 1;\n"
-" for(int nActive=1; nActive<n; nActive<<=1, offset>>=1 )\n"
-" {\n"
-" GROUP_LDS_BARRIER;\n"
-" for( int iIdx = lIdx; iIdx<nActive; iIdx += lSize )\n"
-" {\n"
-" int ai = offset*(2*iIdx+1)-1;\n"
-" int bi = offset*(2*iIdx+2)-1;\n"
-" float4 temp = data[ai];\n"
-" data[ai] = data[bi];\n"
-" data[bi] += temp;\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" return blocksum;\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void LocalScanKernel(__global float4* dst, __global float4* src, __global float4* sumBuffer, uint4 cb)\n"
-"{\n"
-" __local float4 ldsData[WG_SIZE*2];\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" ldsData[2*lIdx] = ( 2*gIdx < cb.m_numElems )? src[2*gIdx]: 0;\n"
-" ldsData[2*lIdx + 1] = ( 2*gIdx+1 < cb.m_numElems )? src[2*gIdx + 1]: 0;\n"
-" float4 sum = ScanExclusiveFloat4(ldsData, WG_SIZE*2, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
-" if( lIdx == 0 ) \n"
-" sumBuffer[GET_GROUP_IDX] = sum;\n"
-" if( (2*gIdx) < cb.m_numElems )\n"
-" {\n"
-" dst[2*gIdx] = ldsData[2*lIdx];\n"
-" }\n"
-" if( (2*gIdx + 1) < cb.m_numElems )\n"
-" {\n"
-" dst[2*gIdx + 1] = ldsData[2*lIdx + 1];\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void AddOffsetKernel(__global float4* dst, __global float4* blockSum, uint4 cb)\n"
-"{\n"
-" const u32 blockSize = WG_SIZE*2;\n"
-" int myIdx = GET_GROUP_IDX+1;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" float4 iBlockSum = blockSum[myIdx];\n"
-" int endValue = min((myIdx+1)*(blockSize), cb.m_numElems);\n"
-" for(int i=myIdx*blockSize+lIdx; i<endValue; i+=GET_GROUP_SIZE)\n"
-" {\n"
-" dst[i] += iBlockSum;\n"
-" }\n"
-"}\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"__kernel\n"
-"void TopLevelScanKernel(__global float4* dst, uint4 cb)\n"
-"{\n"
-" __local float4 ldsData[2048];\n"
-" int gIdx = GET_GLOBAL_IDX;\n"
-" int lIdx = GET_LOCAL_IDX;\n"
-" int lSize = GET_GROUP_SIZE;\n"
-" for(int i=lIdx; i<cb.m_numScanBlocks; i+=lSize )\n"
-" {\n"
-" ldsData[i] = (i<cb.m_numBlocks)? dst[i]:0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" float4 sum = ScanExclusiveFloat4(ldsData, cb.m_numScanBlocks, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
-" for(int i=lIdx; i<cb.m_numBlocks; i+=lSize )\n"
-" {\n"
-" dst[i] = ldsData[i];\n"
-" }\n"
-" if( gIdx == 0 )\n"
-" {\n"
-" dst[cb.m_numBlocks] = sum;\n"
-" }\n"
-"}\n"
-;
+static const char* prefixScanKernelsFloat4CL =
+ "/*\n"
+ "Copyright (c) 2012 Advanced Micro Devices, Inc. \n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Originally written by Takahiro Harada\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "// takahiro end\n"
+ "#define WG_SIZE 128 \n"
+ "#define m_numElems x\n"
+ "#define m_numBlocks y\n"
+ "#define m_numScanBlocks z\n"
+ "/*typedef struct\n"
+ "{\n"
+ " uint m_numElems;\n"
+ " uint m_numBlocks;\n"
+ " uint m_numScanBlocks;\n"
+ " uint m_padding[1];\n"
+ "} ConstBuffer;\n"
+ "*/\n"
+ "float4 ScanExclusiveFloat4(__local float4* data, u32 n, int lIdx, int lSize)\n"
+ "{\n"
+ " float4 blocksum;\n"
+ " int offset = 1;\n"
+ " for(int nActive=n>>1; nActive>0; nActive>>=1, offset<<=1)\n"
+ " {\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for(int iIdx=lIdx; iIdx<nActive; iIdx+=lSize)\n"
+ " {\n"
+ " int ai = offset*(2*iIdx+1)-1;\n"
+ " int bi = offset*(2*iIdx+2)-1;\n"
+ " data[bi] += data[ai];\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " if( lIdx == 0 )\n"
+ " {\n"
+ " blocksum = data[ n-1 ];\n"
+ " data[ n-1 ] = 0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " offset >>= 1;\n"
+ " for(int nActive=1; nActive<n; nActive<<=1, offset>>=1 )\n"
+ " {\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for( int iIdx = lIdx; iIdx<nActive; iIdx += lSize )\n"
+ " {\n"
+ " int ai = offset*(2*iIdx+1)-1;\n"
+ " int bi = offset*(2*iIdx+2)-1;\n"
+ " float4 temp = data[ai];\n"
+ " data[ai] = data[bi];\n"
+ " data[bi] += temp;\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " return blocksum;\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void LocalScanKernel(__global float4* dst, __global float4* src, __global float4* sumBuffer, uint4 cb)\n"
+ "{\n"
+ " __local float4 ldsData[WG_SIZE*2];\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " ldsData[2*lIdx] = ( 2*gIdx < cb.m_numElems )? src[2*gIdx]: 0;\n"
+ " ldsData[2*lIdx + 1] = ( 2*gIdx+1 < cb.m_numElems )? src[2*gIdx + 1]: 0;\n"
+ " float4 sum = ScanExclusiveFloat4(ldsData, WG_SIZE*2, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
+ " if( lIdx == 0 ) \n"
+ " sumBuffer[GET_GROUP_IDX] = sum;\n"
+ " if( (2*gIdx) < cb.m_numElems )\n"
+ " {\n"
+ " dst[2*gIdx] = ldsData[2*lIdx];\n"
+ " }\n"
+ " if( (2*gIdx + 1) < cb.m_numElems )\n"
+ " {\n"
+ " dst[2*gIdx + 1] = ldsData[2*lIdx + 1];\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void AddOffsetKernel(__global float4* dst, __global float4* blockSum, uint4 cb)\n"
+ "{\n"
+ " const u32 blockSize = WG_SIZE*2;\n"
+ " int myIdx = GET_GROUP_IDX+1;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " float4 iBlockSum = blockSum[myIdx];\n"
+ " int endValue = min((myIdx+1)*(blockSize), cb.m_numElems);\n"
+ " for(int i=myIdx*blockSize+lIdx; i<endValue; i+=GET_GROUP_SIZE)\n"
+ " {\n"
+ " dst[i] += iBlockSum;\n"
+ " }\n"
+ "}\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "__kernel\n"
+ "void TopLevelScanKernel(__global float4* dst, uint4 cb)\n"
+ "{\n"
+ " __local float4 ldsData[2048];\n"
+ " int gIdx = GET_GLOBAL_IDX;\n"
+ " int lIdx = GET_LOCAL_IDX;\n"
+ " int lSize = GET_GROUP_SIZE;\n"
+ " for(int i=lIdx; i<cb.m_numScanBlocks; i+=lSize )\n"
+ " {\n"
+ " ldsData[i] = (i<cb.m_numBlocks)? dst[i]:0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " float4 sum = ScanExclusiveFloat4(ldsData, cb.m_numScanBlocks, GET_LOCAL_IDX, GET_GROUP_SIZE);\n"
+ " for(int i=lIdx; i<cb.m_numBlocks; i+=lSize )\n"
+ " {\n"
+ " dst[i] = ldsData[i];\n"
+ " }\n"
+ " if( gIdx == 0 )\n"
+ " {\n"
+ " dst[cb.m_numBlocks] = sum;\n"
+ " }\n"
+ "}\n";
diff --git a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/RadixSort32KernelsCL.h b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/RadixSort32KernelsCL.h
index 8876c16aa6..fb4bdda303 100644
--- a/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/RadixSort32KernelsCL.h
+++ b/thirdparty/bullet/Bullet3OpenCL/ParallelPrimitives/kernels/RadixSort32KernelsCL.h
@@ -1,910 +1,909 @@
//this file is autogenerated using stringify.bat (premake --stringify) in the build folder of this project
-static const char* radixSort32KernelsCL= \
-"/*\n"
-"Bullet Continuous Collision Detection and Physics Library\n"
-"Copyright (c) 2011 Advanced Micro Devices, Inc. http://bulletphysics.org\n"
-"This software is provided 'as-is', without any express or implied warranty.\n"
-"In no event will the authors be held liable for any damages arising from the use of this software.\n"
-"Permission is granted to anyone to use this software for any purpose, \n"
-"including commercial applications, and to alter it and redistribute it freely, \n"
-"subject to the following restrictions:\n"
-"1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
-"2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
-"3. This notice may not be removed or altered from any source distribution.\n"
-"*/\n"
-"//Author Takahiro Harada\n"
-"//#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
-"#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
-"#pragma OPENCL EXTENSION cl_khr_global_int32_base_atomics : enable\n"
-"typedef unsigned int u32;\n"
-"#define GET_GROUP_IDX get_group_id(0)\n"
-"#define GET_LOCAL_IDX get_local_id(0)\n"
-"#define GET_GLOBAL_IDX get_global_id(0)\n"
-"#define GET_GROUP_SIZE get_local_size(0)\n"
-"#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
-"#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
-"#define AtomInc(x) atom_inc(&(x))\n"
-"#define AtomInc1(x, out) out = atom_inc(&(x))\n"
-"#define AtomAdd(x, value) atom_add(&(x), value)\n"
-"#define SELECT_UINT4( b, a, condition ) select( b,a,condition )\n"
-"#define make_uint4 (uint4)\n"
-"#define make_uint2 (uint2)\n"
-"#define make_int2 (int2)\n"
-"#define WG_SIZE 64\n"
-"#define ELEMENTS_PER_WORK_ITEM (256/WG_SIZE)\n"
-"#define BITS_PER_PASS 4\n"
-"#define NUM_BUCKET (1<<BITS_PER_PASS)\n"
-"typedef uchar u8;\n"
-"// this isn't optimization for VLIW. But just reducing writes. \n"
-"#define USE_2LEVEL_REDUCE 1\n"
-"//#define CHECK_BOUNDARY 1\n"
-"//#define NV_GPU 1\n"
-"// Cypress\n"
-"#define nPerWI 16\n"
-"// Cayman\n"
-"//#define nPerWI 20\n"
-"#define m_n x\n"
-"#define m_nWGs y\n"
-"#define m_startBit z\n"
-"#define m_nBlocksPerWG w\n"
-"/*\n"
-"typedef struct\n"
-"{\n"
-" int m_n;\n"
-" int m_nWGs;\n"
-" int m_startBit;\n"
-" int m_nBlocksPerWG;\n"
-"} ConstBuffer;\n"
-"*/\n"
-"typedef struct\n"
-"{\n"
-" unsigned int m_key;\n"
-" unsigned int m_value;\n"
-"} SortDataCL;\n"
-"uint prefixScanVectorEx( uint4* data )\n"
-"{\n"
-" u32 sum = 0;\n"
-" u32 tmp = data[0].x;\n"
-" data[0].x = sum;\n"
-" sum += tmp;\n"
-" tmp = data[0].y;\n"
-" data[0].y = sum;\n"
-" sum += tmp;\n"
-" tmp = data[0].z;\n"
-" data[0].z = sum;\n"
-" sum += tmp;\n"
-" tmp = data[0].w;\n"
-" data[0].w = sum;\n"
-" sum += tmp;\n"
-" return sum;\n"
-"}\n"
-"u32 localPrefixSum( u32 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory, int wgSize /*64 or 128*/ )\n"
-"{\n"
-" { // Set data\n"
-" sorterSharedMemory[lIdx] = 0;\n"
-" sorterSharedMemory[lIdx+wgSize] = pData;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" { // Prefix sum\n"
-" int idx = 2*lIdx + (wgSize+1);\n"
-"#if defined(USE_2LEVEL_REDUCE)\n"
-" if( lIdx < 64 )\n"
-" {\n"
-" u32 u0, u1, u2;\n"
-" u0 = sorterSharedMemory[idx-3];\n"
-" u1 = sorterSharedMemory[idx-2];\n"
-" u2 = sorterSharedMemory[idx-1];\n"
-" AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
-" GROUP_MEM_FENCE;\n"
-" u0 = sorterSharedMemory[idx-12];\n"
-" u1 = sorterSharedMemory[idx-8];\n"
-" u2 = sorterSharedMemory[idx-4];\n"
-" AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
-" GROUP_MEM_FENCE;\n"
-" u0 = sorterSharedMemory[idx-48];\n"
-" u1 = sorterSharedMemory[idx-32];\n"
-" u2 = sorterSharedMemory[idx-16];\n"
-" AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
-" GROUP_MEM_FENCE;\n"
-" if( wgSize > 64 )\n"
-" {\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-64];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-" sorterSharedMemory[idx-1] += sorterSharedMemory[idx-2];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#else\n"
-" if( lIdx < 64 )\n"
-" {\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-2]; \n"
-" GROUP_MEM_FENCE;\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-4];\n"
-" GROUP_MEM_FENCE;\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-8];\n"
-" GROUP_MEM_FENCE;\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-16];\n"
-" GROUP_MEM_FENCE;\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-32];\n"
-" GROUP_MEM_FENCE;\n"
-" if( wgSize > 64 )\n"
-" {\n"
-" sorterSharedMemory[idx] += sorterSharedMemory[idx-64];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-" sorterSharedMemory[idx-1] += sorterSharedMemory[idx-2];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#endif\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" *totalSum = sorterSharedMemory[wgSize*2-1];\n"
-" u32 addValue = sorterSharedMemory[lIdx+wgSize-1];\n"
-" return addValue;\n"
-"}\n"
-"//__attribute__((reqd_work_group_size(128,1,1)))\n"
-"uint4 localPrefixSum128V( uint4 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory )\n"
-"{\n"
-" u32 s4 = prefixScanVectorEx( &pData );\n"
-" u32 rank = localPrefixSum( s4, lIdx, totalSum, sorterSharedMemory, 128 );\n"
-" return pData + make_uint4( rank, rank, rank, rank );\n"
-"}\n"
-"//__attribute__((reqd_work_group_size(64,1,1)))\n"
-"uint4 localPrefixSum64V( uint4 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory )\n"
-"{\n"
-" u32 s4 = prefixScanVectorEx( &pData );\n"
-" u32 rank = localPrefixSum( s4, lIdx, totalSum, sorterSharedMemory, 64 );\n"
-" return pData + make_uint4( rank, rank, rank, rank );\n"
-"}\n"
-"u32 unpack4Key( u32 key, int keyIdx ){ return (key>>(keyIdx*8)) & 0xff;}\n"
-"u32 bit8Scan(u32 v)\n"
-"{\n"
-" return (v<<8) + (v<<16) + (v<<24);\n"
-"}\n"
-"//===\n"
-"#define MY_HISTOGRAM(idx) localHistogramMat[(idx)*WG_SIZE+lIdx]\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void StreamCountKernel( __global u32* gSrc, __global u32* histogramOut, int4 cb )\n"
-"{\n"
-" __local u32 localHistogramMat[NUM_BUCKET*WG_SIZE];\n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 lIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" for(int i=0; i<NUM_BUCKET; i++)\n"
-" {\n"
-" MY_HISTOGRAM(i) = 0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" u32 localKey;\n"
-" int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
-" int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
-" {\n"
-" // MY_HISTOGRAM( localKeys.x ) ++ is much expensive than atomic add as it requires read and write while atomics can just add on AMD\n"
-" // Using registers didn't perform well. It seems like use localKeys to address requires a lot of alu ops\n"
-" // AMD: AtomInc performs better while NV prefers ++\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-" {\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+i < n )\n"
-"#endif\n"
-" {\n"
-" localKey = (gSrc[addr+i]>>startBit) & 0xf;\n"
-"#if defined(NV_GPU)\n"
-" MY_HISTOGRAM( localKey )++;\n"
-"#else\n"
-" AtomInc( MY_HISTOGRAM( localKey ) );\n"
-"#endif\n"
-" }\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" \n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<GET_GROUP_SIZE; i++)\n"
-" {\n"
-" sum += localHistogramMat[lIdx*WG_SIZE+(i+lIdx)%GET_GROUP_SIZE];\n"
-" }\n"
-" histogramOut[lIdx*nWGs+wgIdx] = sum;\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void StreamCountSortDataKernel( __global SortDataCL* gSrc, __global u32* histogramOut, int4 cb )\n"
-"{\n"
-" __local u32 localHistogramMat[NUM_BUCKET*WG_SIZE];\n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 lIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" for(int i=0; i<NUM_BUCKET; i++)\n"
-" {\n"
-" MY_HISTOGRAM(i) = 0;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" u32 localKey;\n"
-" int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
-" int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
-" {\n"
-" // MY_HISTOGRAM( localKeys.x ) ++ is much expensive than atomic add as it requires read and write while atomics can just add on AMD\n"
-" // Using registers didn't perform well. It seems like use localKeys to address requires a lot of alu ops\n"
-" // AMD: AtomInc performs better while NV prefers ++\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-" {\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+i < n )\n"
-"#endif\n"
-" {\n"
-" localKey = (gSrc[addr+i].m_key>>startBit) & 0xf;\n"
-"#if defined(NV_GPU)\n"
-" MY_HISTOGRAM( localKey )++;\n"
-"#else\n"
-" AtomInc( MY_HISTOGRAM( localKey ) );\n"
-"#endif\n"
-" }\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" \n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<GET_GROUP_SIZE; i++)\n"
-" {\n"
-" sum += localHistogramMat[lIdx*WG_SIZE+(i+lIdx)%GET_GROUP_SIZE];\n"
-" }\n"
-" histogramOut[lIdx*nWGs+wgIdx] = sum;\n"
-" }\n"
-"}\n"
-"#define nPerLane (nPerWI/4)\n"
-"// NUM_BUCKET*nWGs < 128*nPerWI\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(128,1,1)))\n"
-"void PrefixScanKernel( __global u32* wHistogram1, int4 cb )\n"
-"{\n"
-" __local u32 ldsTopScanData[128*2];\n"
-" u32 lIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" u32 data[nPerWI];\n"
-" for(int i=0; i<nPerWI; i++)\n"
-" {\n"
-" data[i] = 0;\n"
-" if( (nPerWI*lIdx+i) < NUM_BUCKET*nWGs )\n"
-" data[i] = wHistogram1[nPerWI*lIdx+i];\n"
-" }\n"
-" uint4 myData = make_uint4(0,0,0,0);\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" myData.x += data[nPerLane*0+i];\n"
-" myData.y += data[nPerLane*1+i];\n"
-" myData.z += data[nPerLane*2+i];\n"
-" myData.w += data[nPerLane*3+i];\n"
-" }\n"
-" uint totalSum;\n"
-" uint4 scanned = localPrefixSum128V( myData, lIdx, &totalSum, ldsTopScanData );\n"
-"// for(int j=0; j<4; j++) // somehow it introduces a lot of branches\n"
-" { int j = 0;\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" u32 tmp = data[nPerLane*j+i];\n"
-" data[nPerLane*j+i] = sum;\n"
-" sum += tmp;\n"
-" }\n"
-" }\n"
-" { int j = 1;\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" u32 tmp = data[nPerLane*j+i];\n"
-" data[nPerLane*j+i] = sum;\n"
-" sum += tmp;\n"
-" }\n"
-" }\n"
-" { int j = 2;\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" u32 tmp = data[nPerLane*j+i];\n"
-" data[nPerLane*j+i] = sum;\n"
-" sum += tmp;\n"
-" }\n"
-" }\n"
-" { int j = 3;\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" u32 tmp = data[nPerLane*j+i];\n"
-" data[nPerLane*j+i] = sum;\n"
-" sum += tmp;\n"
-" }\n"
-" }\n"
-" for(int i=0; i<nPerLane; i++)\n"
-" {\n"
-" data[nPerLane*0+i] += scanned.x;\n"
-" data[nPerLane*1+i] += scanned.y;\n"
-" data[nPerLane*2+i] += scanned.z;\n"
-" data[nPerLane*3+i] += scanned.w;\n"
-" }\n"
-" for(int i=0; i<nPerWI; i++)\n"
-" {\n"
-" int index = nPerWI*lIdx+i;\n"
-" if (index < NUM_BUCKET*nWGs)\n"
-" wHistogram1[nPerWI*lIdx+i] = data[i];\n"
-" }\n"
-"}\n"
-"// 4 scan, 4 exchange\n"
-"void sort4Bits(u32 sortData[4], int startBit, int lIdx, __local u32* ldsSortData)\n"
-"{\n"
-" for(int bitIdx=0; bitIdx<BITS_PER_PASS; bitIdx++)\n"
-" {\n"
-" u32 mask = (1<<bitIdx);\n"
-" uint4 cmpResult = make_uint4( (sortData[0]>>startBit) & mask, (sortData[1]>>startBit) & mask, (sortData[2]>>startBit) & mask, (sortData[3]>>startBit) & mask );\n"
-" uint4 prefixSum = SELECT_UINT4( make_uint4(1,1,1,1), make_uint4(0,0,0,0), cmpResult != make_uint4(0,0,0,0) );\n"
-" u32 total;\n"
-" prefixSum = localPrefixSum64V( prefixSum, lIdx, &total, ldsSortData );\n"
-" {\n"
-" uint4 localAddr = make_uint4(lIdx*4+0,lIdx*4+1,lIdx*4+2,lIdx*4+3);\n"
-" uint4 dstAddr = localAddr - prefixSum + make_uint4( total, total, total, total );\n"
-" dstAddr = SELECT_UINT4( prefixSum, dstAddr, cmpResult != make_uint4(0, 0, 0, 0) );\n"
-" GROUP_LDS_BARRIER;\n"
-" ldsSortData[dstAddr.x] = sortData[0];\n"
-" ldsSortData[dstAddr.y] = sortData[1];\n"
-" ldsSortData[dstAddr.z] = sortData[2];\n"
-" ldsSortData[dstAddr.w] = sortData[3];\n"
-" GROUP_LDS_BARRIER;\n"
-" sortData[0] = ldsSortData[localAddr.x];\n"
-" sortData[1] = ldsSortData[localAddr.y];\n"
-" sortData[2] = ldsSortData[localAddr.z];\n"
-" sortData[3] = ldsSortData[localAddr.w];\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-" }\n"
-"}\n"
-"// 2 scan, 2 exchange\n"
-"void sort4Bits1(u32 sortData[4], int startBit, int lIdx, __local u32* ldsSortData)\n"
-"{\n"
-" for(uint ibit=0; ibit<BITS_PER_PASS; ibit+=2)\n"
-" {\n"
-" uint4 b = make_uint4((sortData[0]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[1]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[2]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[3]>>(startBit+ibit)) & 0x3);\n"
-" u32 key4;\n"
-" u32 sKeyPacked[4] = { 0, 0, 0, 0 };\n"
-" {\n"
-" sKeyPacked[0] |= 1<<(8*b.x);\n"
-" sKeyPacked[1] |= 1<<(8*b.y);\n"
-" sKeyPacked[2] |= 1<<(8*b.z);\n"
-" sKeyPacked[3] |= 1<<(8*b.w);\n"
-" key4 = sKeyPacked[0] + sKeyPacked[1] + sKeyPacked[2] + sKeyPacked[3];\n"
-" }\n"
-" u32 rankPacked;\n"
-" u32 sumPacked;\n"
-" {\n"
-" rankPacked = localPrefixSum( key4, lIdx, &sumPacked, ldsSortData, WG_SIZE );\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" u32 newOffset[4] = { 0,0,0,0 };\n"
-" {\n"
-" u32 sumScanned = bit8Scan( sumPacked );\n"
-" u32 scannedKeys[4];\n"
-" scannedKeys[0] = 1<<(8*b.x);\n"
-" scannedKeys[1] = 1<<(8*b.y);\n"
-" scannedKeys[2] = 1<<(8*b.z);\n"
-" scannedKeys[3] = 1<<(8*b.w);\n"
-" { // 4 scans at once\n"
-" u32 sum4 = 0;\n"
-" for(int ie=0; ie<4; ie++)\n"
-" {\n"
-" u32 tmp = scannedKeys[ie];\n"
-" scannedKeys[ie] = sum4;\n"
-" sum4 += tmp;\n"
-" }\n"
-" }\n"
-" {\n"
-" u32 sumPlusRank = sumScanned + rankPacked;\n"
-" { u32 ie = b.x;\n"
-" scannedKeys[0] += sumPlusRank;\n"
-" newOffset[0] = unpack4Key( scannedKeys[0], ie );\n"
-" }\n"
-" { u32 ie = b.y;\n"
-" scannedKeys[1] += sumPlusRank;\n"
-" newOffset[1] = unpack4Key( scannedKeys[1], ie );\n"
-" }\n"
-" { u32 ie = b.z;\n"
-" scannedKeys[2] += sumPlusRank;\n"
-" newOffset[2] = unpack4Key( scannedKeys[2], ie );\n"
-" }\n"
-" { u32 ie = b.w;\n"
-" scannedKeys[3] += sumPlusRank;\n"
-" newOffset[3] = unpack4Key( scannedKeys[3], ie );\n"
-" }\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" {\n"
-" ldsSortData[newOffset[0]] = sortData[0];\n"
-" ldsSortData[newOffset[1]] = sortData[1];\n"
-" ldsSortData[newOffset[2]] = sortData[2];\n"
-" ldsSortData[newOffset[3]] = sortData[3];\n"
-" GROUP_LDS_BARRIER;\n"
-" u32 dstAddr = 4*lIdx;\n"
-" sortData[0] = ldsSortData[dstAddr+0];\n"
-" sortData[1] = ldsSortData[dstAddr+1];\n"
-" sortData[2] = ldsSortData[dstAddr+2];\n"
-" sortData[3] = ldsSortData[dstAddr+3];\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-" }\n"
-"}\n"
-"#define SET_HISTOGRAM(setIdx, key) ldsSortData[(setIdx)*NUM_BUCKET+key]\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void SortAndScatterKernel( __global const u32* restrict gSrc, __global const u32* rHistogram, __global u32* restrict gDst, int4 cb )\n"
-"{\n"
-" __local u32 ldsSortData[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
-" __local u32 localHistogramToCarry[NUM_BUCKET];\n"
-" __local u32 localHistogram[NUM_BUCKET*2];\n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 lIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" if( lIdx < (NUM_BUCKET) )\n"
-" {\n"
-" localHistogramToCarry[lIdx] = rHistogram[lIdx*nWGs + wgIdx];\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" int nBlocks = n/blockSize - nBlocksPerWG*wgIdx;\n"
-" int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
-" {\n"
-" u32 myHistogram = 0;\n"
-" u32 sortData[ELEMENTS_PER_WORK_ITEM];\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" sortData[i] = ( addr+i < n )? gSrc[ addr+i ] : 0xffffffff;\n"
-"#else\n"
-" sortData[i] = gSrc[ addr+i ];\n"
-"#endif\n"
-" sort4Bits(sortData, startBit, lIdx, ldsSortData);\n"
-" u32 keys[ELEMENTS_PER_WORK_ITEM];\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-" keys[i] = (sortData[i]>>startBit) & 0xf;\n"
-" { // create histogram\n"
-" u32 setIdx = lIdx/16;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[lIdx] = 0;\n"
-" }\n"
-" ldsSortData[lIdx] = 0;\n"
-" GROUP_LDS_BARRIER;\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+i < n )\n"
-"#endif\n"
-"#if defined(NV_GPU)\n"
-" SET_HISTOGRAM( setIdx, keys[i] )++;\n"
-"#else\n"
-" AtomInc( SET_HISTOGRAM( setIdx, keys[i] ) );\n"
-"#endif\n"
-" \n"
-" GROUP_LDS_BARRIER;\n"
-" \n"
-" uint hIdx = NUM_BUCKET+lIdx;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<WG_SIZE/16; i++)\n"
-" {\n"
-" sum += SET_HISTOGRAM( i, lIdx );\n"
-" }\n"
-" myHistogram = sum;\n"
-" localHistogram[hIdx] = sum;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-"#if defined(USE_2LEVEL_REDUCE)\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[hIdx] = localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" u32 u0, u1, u2;\n"
-" u0 = localHistogram[hIdx-3];\n"
-" u1 = localHistogram[hIdx-2];\n"
-" u2 = localHistogram[hIdx-1];\n"
-" AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
-" GROUP_MEM_FENCE;\n"
-" u0 = localHistogram[hIdx-12];\n"
-" u1 = localHistogram[hIdx-8];\n"
-" u2 = localHistogram[hIdx-4];\n"
-" AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#else\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[hIdx] = localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-2];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-4];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-8];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#endif\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-" {\n"
-" for(int ie=0; ie<ELEMENTS_PER_WORK_ITEM; ie++)\n"
-" {\n"
-" int dataIdx = ELEMENTS_PER_WORK_ITEM*lIdx+ie;\n"
-" int binIdx = keys[ie];\n"
-" int groupOffset = localHistogramToCarry[binIdx];\n"
-" int myIdx = dataIdx - localHistogram[NUM_BUCKET+binIdx];\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+ie < n )\n"
-"#endif\n"
-" gDst[ groupOffset + myIdx ] = sortData[ie];\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogramToCarry[lIdx] += myHistogram;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-"}\n"
-"// 2 scan, 2 exchange\n"
-"void sort4Bits1KeyValue(u32 sortData[4], int sortVal[4], int startBit, int lIdx, __local u32* ldsSortData, __local int *ldsSortVal)\n"
-"{\n"
-" for(uint ibit=0; ibit<BITS_PER_PASS; ibit+=2)\n"
-" {\n"
-" uint4 b = make_uint4((sortData[0]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[1]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[2]>>(startBit+ibit)) & 0x3, \n"
-" (sortData[3]>>(startBit+ibit)) & 0x3);\n"
-" u32 key4;\n"
-" u32 sKeyPacked[4] = { 0, 0, 0, 0 };\n"
-" {\n"
-" sKeyPacked[0] |= 1<<(8*b.x);\n"
-" sKeyPacked[1] |= 1<<(8*b.y);\n"
-" sKeyPacked[2] |= 1<<(8*b.z);\n"
-" sKeyPacked[3] |= 1<<(8*b.w);\n"
-" key4 = sKeyPacked[0] + sKeyPacked[1] + sKeyPacked[2] + sKeyPacked[3];\n"
-" }\n"
-" u32 rankPacked;\n"
-" u32 sumPacked;\n"
-" {\n"
-" rankPacked = localPrefixSum( key4, lIdx, &sumPacked, ldsSortData, WG_SIZE );\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" u32 newOffset[4] = { 0,0,0,0 };\n"
-" {\n"
-" u32 sumScanned = bit8Scan( sumPacked );\n"
-" u32 scannedKeys[4];\n"
-" scannedKeys[0] = 1<<(8*b.x);\n"
-" scannedKeys[1] = 1<<(8*b.y);\n"
-" scannedKeys[2] = 1<<(8*b.z);\n"
-" scannedKeys[3] = 1<<(8*b.w);\n"
-" { // 4 scans at once\n"
-" u32 sum4 = 0;\n"
-" for(int ie=0; ie<4; ie++)\n"
-" {\n"
-" u32 tmp = scannedKeys[ie];\n"
-" scannedKeys[ie] = sum4;\n"
-" sum4 += tmp;\n"
-" }\n"
-" }\n"
-" {\n"
-" u32 sumPlusRank = sumScanned + rankPacked;\n"
-" { u32 ie = b.x;\n"
-" scannedKeys[0] += sumPlusRank;\n"
-" newOffset[0] = unpack4Key( scannedKeys[0], ie );\n"
-" }\n"
-" { u32 ie = b.y;\n"
-" scannedKeys[1] += sumPlusRank;\n"
-" newOffset[1] = unpack4Key( scannedKeys[1], ie );\n"
-" }\n"
-" { u32 ie = b.z;\n"
-" scannedKeys[2] += sumPlusRank;\n"
-" newOffset[2] = unpack4Key( scannedKeys[2], ie );\n"
-" }\n"
-" { u32 ie = b.w;\n"
-" scannedKeys[3] += sumPlusRank;\n"
-" newOffset[3] = unpack4Key( scannedKeys[3], ie );\n"
-" }\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" {\n"
-" ldsSortData[newOffset[0]] = sortData[0];\n"
-" ldsSortData[newOffset[1]] = sortData[1];\n"
-" ldsSortData[newOffset[2]] = sortData[2];\n"
-" ldsSortData[newOffset[3]] = sortData[3];\n"
-" ldsSortVal[newOffset[0]] = sortVal[0];\n"
-" ldsSortVal[newOffset[1]] = sortVal[1];\n"
-" ldsSortVal[newOffset[2]] = sortVal[2];\n"
-" ldsSortVal[newOffset[3]] = sortVal[3];\n"
-" GROUP_LDS_BARRIER;\n"
-" u32 dstAddr = 4*lIdx;\n"
-" sortData[0] = ldsSortData[dstAddr+0];\n"
-" sortData[1] = ldsSortData[dstAddr+1];\n"
-" sortData[2] = ldsSortData[dstAddr+2];\n"
-" sortData[3] = ldsSortData[dstAddr+3];\n"
-" sortVal[0] = ldsSortVal[dstAddr+0];\n"
-" sortVal[1] = ldsSortVal[dstAddr+1];\n"
-" sortVal[2] = ldsSortVal[dstAddr+2];\n"
-" sortVal[3] = ldsSortVal[dstAddr+3];\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void SortAndScatterSortDataKernel( __global const SortDataCL* restrict gSrc, __global const u32* rHistogram, __global SortDataCL* restrict gDst, int4 cb)\n"
-"{\n"
-" __local int ldsSortData[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
-" __local int ldsSortVal[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
-" __local u32 localHistogramToCarry[NUM_BUCKET];\n"
-" __local u32 localHistogram[NUM_BUCKET*2];\n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 lIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" if( lIdx < (NUM_BUCKET) )\n"
-" {\n"
-" localHistogramToCarry[lIdx] = rHistogram[lIdx*nWGs + wgIdx];\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" \n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" int nBlocks = n/blockSize - nBlocksPerWG*wgIdx;\n"
-" int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
-" {\n"
-" u32 myHistogram = 0;\n"
-" int sortData[ELEMENTS_PER_WORK_ITEM];\n"
-" int sortVal[ELEMENTS_PER_WORK_ITEM];\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" {\n"
-" sortData[i] = ( addr+i < n )? gSrc[ addr+i ].m_key : 0xffffffff;\n"
-" sortVal[i] = ( addr+i < n )? gSrc[ addr+i ].m_value : 0xffffffff;\n"
-" }\n"
-"#else\n"
-" {\n"
-" sortData[i] = gSrc[ addr+i ].m_key;\n"
-" sortVal[i] = gSrc[ addr+i ].m_value;\n"
-" }\n"
-"#endif\n"
-" sort4Bits1KeyValue(sortData, sortVal, startBit, lIdx, ldsSortData, ldsSortVal);\n"
-" u32 keys[ELEMENTS_PER_WORK_ITEM];\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-" keys[i] = (sortData[i]>>startBit) & 0xf;\n"
-" { // create histogram\n"
-" u32 setIdx = lIdx/16;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[lIdx] = 0;\n"
-" }\n"
-" ldsSortData[lIdx] = 0;\n"
-" GROUP_LDS_BARRIER;\n"
-" for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+i < n )\n"
-"#endif\n"
-"#if defined(NV_GPU)\n"
-" SET_HISTOGRAM( setIdx, keys[i] )++;\n"
-"#else\n"
-" AtomInc( SET_HISTOGRAM( setIdx, keys[i] ) );\n"
-"#endif\n"
-" \n"
-" GROUP_LDS_BARRIER;\n"
-" \n"
-" uint hIdx = NUM_BUCKET+lIdx;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" u32 sum = 0;\n"
-" for(int i=0; i<WG_SIZE/16; i++)\n"
-" {\n"
-" sum += SET_HISTOGRAM( i, lIdx );\n"
-" }\n"
-" myHistogram = sum;\n"
-" localHistogram[hIdx] = sum;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-"#if defined(USE_2LEVEL_REDUCE)\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[hIdx] = localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" u32 u0, u1, u2;\n"
-" u0 = localHistogram[hIdx-3];\n"
-" u1 = localHistogram[hIdx-2];\n"
-" u2 = localHistogram[hIdx-1];\n"
-" AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
-" GROUP_MEM_FENCE;\n"
-" u0 = localHistogram[hIdx-12];\n"
-" u1 = localHistogram[hIdx-8];\n"
-" u2 = localHistogram[hIdx-4];\n"
-" AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#else\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogram[hIdx] = localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-1];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-2];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-4];\n"
-" GROUP_MEM_FENCE;\n"
-" localHistogram[hIdx] += localHistogram[hIdx-8];\n"
-" GROUP_MEM_FENCE;\n"
-" }\n"
-"#endif\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-" {\n"
-" for(int ie=0; ie<ELEMENTS_PER_WORK_ITEM; ie++)\n"
-" {\n"
-" int dataIdx = ELEMENTS_PER_WORK_ITEM*lIdx+ie;\n"
-" int binIdx = keys[ie];\n"
-" int groupOffset = localHistogramToCarry[binIdx];\n"
-" int myIdx = dataIdx - localHistogram[NUM_BUCKET+binIdx];\n"
-"#if defined(CHECK_BOUNDARY)\n"
-" if( addr+ie < n )\n"
-" {\n"
-" if ((groupOffset + myIdx)<n)\n"
-" {\n"
-" if (sortData[ie]==sortVal[ie])\n"
-" {\n"
-" \n"
-" SortDataCL tmp;\n"
-" tmp.m_key = sortData[ie];\n"
-" tmp.m_value = sortVal[ie];\n"
-" if (tmp.m_key == tmp.m_value)\n"
-" gDst[groupOffset + myIdx ] = tmp;\n"
-" }\n"
-" \n"
-" }\n"
-" }\n"
-"#else\n"
-" if ((groupOffset + myIdx)<n)\n"
-" {\n"
-" gDst[ groupOffset + myIdx ].m_key = sortData[ie];\n"
-" gDst[ groupOffset + myIdx ].m_value = sortVal[ie];\n"
-" }\n"
-"#endif\n"
-" }\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" if( lIdx < NUM_BUCKET )\n"
-" {\n"
-" localHistogramToCarry[lIdx] += myHistogram;\n"
-" }\n"
-" GROUP_LDS_BARRIER;\n"
-" }\n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void SortAndScatterSortDataKernelSerial( __global const SortDataCL* restrict gSrc, __global const u32* rHistogram, __global SortDataCL* restrict gDst, int4 cb)\n"
-"{\n"
-" \n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 realLocalIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" int counter[NUM_BUCKET];\n"
-" \n"
-" if (realLocalIdx>0)\n"
-" return;\n"
-" \n"
-" for (int c=0;c<NUM_BUCKET;c++)\n"
-" counter[c]=0;\n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" \n"
-" int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++)\n"
-" {\n"
-" for (int lIdx=0;lIdx<WG_SIZE;lIdx++)\n"
-" {\n"
-" int addr2 = iblock*blockSize + blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" \n"
-" for(int j=0; j<ELEMENTS_PER_WORK_ITEM; j++)\n"
-" {\n"
-" int i = addr2+j;\n"
-" if( i < n )\n"
-" {\n"
-" int tableIdx;\n"
-" tableIdx = (gSrc[i].m_key>>startBit) & 0xf;//0xf = NUM_TABLES-1\n"
-" gDst[rHistogram[tableIdx*nWGs+wgIdx] + counter[tableIdx]] = gSrc[i];\n"
-" counter[tableIdx] ++;\n"
-" }\n"
-" }\n"
-" }\n"
-" }\n"
-" \n"
-"}\n"
-"__kernel\n"
-"__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
-"void SortAndScatterKernelSerial( __global const u32* restrict gSrc, __global const u32* rHistogram, __global u32* restrict gDst, int4 cb )\n"
-"{\n"
-" \n"
-" u32 gIdx = GET_GLOBAL_IDX;\n"
-" u32 realLocalIdx = GET_LOCAL_IDX;\n"
-" u32 wgIdx = GET_GROUP_IDX;\n"
-" u32 wgSize = GET_GROUP_SIZE;\n"
-" const int startBit = cb.m_startBit;\n"
-" const int n = cb.m_n;\n"
-" const int nWGs = cb.m_nWGs;\n"
-" const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
-" int counter[NUM_BUCKET];\n"
-" \n"
-" if (realLocalIdx>0)\n"
-" return;\n"
-" \n"
-" for (int c=0;c<NUM_BUCKET;c++)\n"
-" counter[c]=0;\n"
-" const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
-" \n"
-" int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
-" for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++)\n"
-" {\n"
-" for (int lIdx=0;lIdx<WG_SIZE;lIdx++)\n"
-" {\n"
-" int addr2 = iblock*blockSize + blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
-" \n"
-" for(int j=0; j<ELEMENTS_PER_WORK_ITEM; j++)\n"
-" {\n"
-" int i = addr2+j;\n"
-" if( i < n )\n"
-" {\n"
-" int tableIdx;\n"
-" tableIdx = (gSrc[i]>>startBit) & 0xf;//0xf = NUM_TABLES-1\n"
-" gDst[rHistogram[tableIdx*nWGs+wgIdx] + counter[tableIdx]] = gSrc[i];\n"
-" counter[tableIdx] ++;\n"
-" }\n"
-" }\n"
-" }\n"
-" }\n"
-" \n"
-"}\n"
-;
+static const char* radixSort32KernelsCL =
+ "/*\n"
+ "Bullet Continuous Collision Detection and Physics Library\n"
+ "Copyright (c) 2011 Advanced Micro Devices, Inc. http://bulletphysics.org\n"
+ "This software is provided 'as-is', without any express or implied warranty.\n"
+ "In no event will the authors be held liable for any damages arising from the use of this software.\n"
+ "Permission is granted to anyone to use this software for any purpose, \n"
+ "including commercial applications, and to alter it and redistribute it freely, \n"
+ "subject to the following restrictions:\n"
+ "1. The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.\n"
+ "2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.\n"
+ "3. This notice may not be removed or altered from any source distribution.\n"
+ "*/\n"
+ "//Author Takahiro Harada\n"
+ "//#pragma OPENCL EXTENSION cl_amd_printf : enable\n"
+ "#pragma OPENCL EXTENSION cl_khr_local_int32_base_atomics : enable\n"
+ "#pragma OPENCL EXTENSION cl_khr_global_int32_base_atomics : enable\n"
+ "typedef unsigned int u32;\n"
+ "#define GET_GROUP_IDX get_group_id(0)\n"
+ "#define GET_LOCAL_IDX get_local_id(0)\n"
+ "#define GET_GLOBAL_IDX get_global_id(0)\n"
+ "#define GET_GROUP_SIZE get_local_size(0)\n"
+ "#define GROUP_LDS_BARRIER barrier(CLK_LOCAL_MEM_FENCE)\n"
+ "#define GROUP_MEM_FENCE mem_fence(CLK_LOCAL_MEM_FENCE)\n"
+ "#define AtomInc(x) atom_inc(&(x))\n"
+ "#define AtomInc1(x, out) out = atom_inc(&(x))\n"
+ "#define AtomAdd(x, value) atom_add(&(x), value)\n"
+ "#define SELECT_UINT4( b, a, condition ) select( b,a,condition )\n"
+ "#define make_uint4 (uint4)\n"
+ "#define make_uint2 (uint2)\n"
+ "#define make_int2 (int2)\n"
+ "#define WG_SIZE 64\n"
+ "#define ELEMENTS_PER_WORK_ITEM (256/WG_SIZE)\n"
+ "#define BITS_PER_PASS 4\n"
+ "#define NUM_BUCKET (1<<BITS_PER_PASS)\n"
+ "typedef uchar u8;\n"
+ "// this isn't optimization for VLIW. But just reducing writes. \n"
+ "#define USE_2LEVEL_REDUCE 1\n"
+ "//#define CHECK_BOUNDARY 1\n"
+ "//#define NV_GPU 1\n"
+ "// Cypress\n"
+ "#define nPerWI 16\n"
+ "// Cayman\n"
+ "//#define nPerWI 20\n"
+ "#define m_n x\n"
+ "#define m_nWGs y\n"
+ "#define m_startBit z\n"
+ "#define m_nBlocksPerWG w\n"
+ "/*\n"
+ "typedef struct\n"
+ "{\n"
+ " int m_n;\n"
+ " int m_nWGs;\n"
+ " int m_startBit;\n"
+ " int m_nBlocksPerWG;\n"
+ "} ConstBuffer;\n"
+ "*/\n"
+ "typedef struct\n"
+ "{\n"
+ " unsigned int m_key;\n"
+ " unsigned int m_value;\n"
+ "} SortDataCL;\n"
+ "uint prefixScanVectorEx( uint4* data )\n"
+ "{\n"
+ " u32 sum = 0;\n"
+ " u32 tmp = data[0].x;\n"
+ " data[0].x = sum;\n"
+ " sum += tmp;\n"
+ " tmp = data[0].y;\n"
+ " data[0].y = sum;\n"
+ " sum += tmp;\n"
+ " tmp = data[0].z;\n"
+ " data[0].z = sum;\n"
+ " sum += tmp;\n"
+ " tmp = data[0].w;\n"
+ " data[0].w = sum;\n"
+ " sum += tmp;\n"
+ " return sum;\n"
+ "}\n"
+ "u32 localPrefixSum( u32 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory, int wgSize /*64 or 128*/ )\n"
+ "{\n"
+ " { // Set data\n"
+ " sorterSharedMemory[lIdx] = 0;\n"
+ " sorterSharedMemory[lIdx+wgSize] = pData;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " { // Prefix sum\n"
+ " int idx = 2*lIdx + (wgSize+1);\n"
+ "#if defined(USE_2LEVEL_REDUCE)\n"
+ " if( lIdx < 64 )\n"
+ " {\n"
+ " u32 u0, u1, u2;\n"
+ " u0 = sorterSharedMemory[idx-3];\n"
+ " u1 = sorterSharedMemory[idx-2];\n"
+ " u2 = sorterSharedMemory[idx-1];\n"
+ " AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
+ " GROUP_MEM_FENCE;\n"
+ " u0 = sorterSharedMemory[idx-12];\n"
+ " u1 = sorterSharedMemory[idx-8];\n"
+ " u2 = sorterSharedMemory[idx-4];\n"
+ " AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
+ " GROUP_MEM_FENCE;\n"
+ " u0 = sorterSharedMemory[idx-48];\n"
+ " u1 = sorterSharedMemory[idx-32];\n"
+ " u2 = sorterSharedMemory[idx-16];\n"
+ " AtomAdd( sorterSharedMemory[idx], u0+u1+u2 ); \n"
+ " GROUP_MEM_FENCE;\n"
+ " if( wgSize > 64 )\n"
+ " {\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-64];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ " sorterSharedMemory[idx-1] += sorterSharedMemory[idx-2];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#else\n"
+ " if( lIdx < 64 )\n"
+ " {\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-2]; \n"
+ " GROUP_MEM_FENCE;\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-4];\n"
+ " GROUP_MEM_FENCE;\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-8];\n"
+ " GROUP_MEM_FENCE;\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-16];\n"
+ " GROUP_MEM_FENCE;\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-32];\n"
+ " GROUP_MEM_FENCE;\n"
+ " if( wgSize > 64 )\n"
+ " {\n"
+ " sorterSharedMemory[idx] += sorterSharedMemory[idx-64];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ " sorterSharedMemory[idx-1] += sorterSharedMemory[idx-2];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#endif\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " *totalSum = sorterSharedMemory[wgSize*2-1];\n"
+ " u32 addValue = sorterSharedMemory[lIdx+wgSize-1];\n"
+ " return addValue;\n"
+ "}\n"
+ "//__attribute__((reqd_work_group_size(128,1,1)))\n"
+ "uint4 localPrefixSum128V( uint4 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory )\n"
+ "{\n"
+ " u32 s4 = prefixScanVectorEx( &pData );\n"
+ " u32 rank = localPrefixSum( s4, lIdx, totalSum, sorterSharedMemory, 128 );\n"
+ " return pData + make_uint4( rank, rank, rank, rank );\n"
+ "}\n"
+ "//__attribute__((reqd_work_group_size(64,1,1)))\n"
+ "uint4 localPrefixSum64V( uint4 pData, uint lIdx, uint* totalSum, __local u32* sorterSharedMemory )\n"
+ "{\n"
+ " u32 s4 = prefixScanVectorEx( &pData );\n"
+ " u32 rank = localPrefixSum( s4, lIdx, totalSum, sorterSharedMemory, 64 );\n"
+ " return pData + make_uint4( rank, rank, rank, rank );\n"
+ "}\n"
+ "u32 unpack4Key( u32 key, int keyIdx ){ return (key>>(keyIdx*8)) & 0xff;}\n"
+ "u32 bit8Scan(u32 v)\n"
+ "{\n"
+ " return (v<<8) + (v<<16) + (v<<24);\n"
+ "}\n"
+ "//===\n"
+ "#define MY_HISTOGRAM(idx) localHistogramMat[(idx)*WG_SIZE+lIdx]\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void StreamCountKernel( __global u32* gSrc, __global u32* histogramOut, int4 cb )\n"
+ "{\n"
+ " __local u32 localHistogramMat[NUM_BUCKET*WG_SIZE];\n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 lIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " for(int i=0; i<NUM_BUCKET; i++)\n"
+ " {\n"
+ " MY_HISTOGRAM(i) = 0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " u32 localKey;\n"
+ " int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
+ " int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
+ " {\n"
+ " // MY_HISTOGRAM( localKeys.x ) ++ is much expensive than atomic add as it requires read and write while atomics can just add on AMD\n"
+ " // Using registers didn't perform well. It seems like use localKeys to address requires a lot of alu ops\n"
+ " // AMD: AtomInc performs better while NV prefers ++\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ " {\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+i < n )\n"
+ "#endif\n"
+ " {\n"
+ " localKey = (gSrc[addr+i]>>startBit) & 0xf;\n"
+ "#if defined(NV_GPU)\n"
+ " MY_HISTOGRAM( localKey )++;\n"
+ "#else\n"
+ " AtomInc( MY_HISTOGRAM( localKey ) );\n"
+ "#endif\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " \n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<GET_GROUP_SIZE; i++)\n"
+ " {\n"
+ " sum += localHistogramMat[lIdx*WG_SIZE+(i+lIdx)%GET_GROUP_SIZE];\n"
+ " }\n"
+ " histogramOut[lIdx*nWGs+wgIdx] = sum;\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void StreamCountSortDataKernel( __global SortDataCL* gSrc, __global u32* histogramOut, int4 cb )\n"
+ "{\n"
+ " __local u32 localHistogramMat[NUM_BUCKET*WG_SIZE];\n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 lIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " for(int i=0; i<NUM_BUCKET; i++)\n"
+ " {\n"
+ " MY_HISTOGRAM(i) = 0;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " u32 localKey;\n"
+ " int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
+ " int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
+ " {\n"
+ " // MY_HISTOGRAM( localKeys.x ) ++ is much expensive than atomic add as it requires read and write while atomics can just add on AMD\n"
+ " // Using registers didn't perform well. It seems like use localKeys to address requires a lot of alu ops\n"
+ " // AMD: AtomInc performs better while NV prefers ++\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ " {\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+i < n )\n"
+ "#endif\n"
+ " {\n"
+ " localKey = (gSrc[addr+i].m_key>>startBit) & 0xf;\n"
+ "#if defined(NV_GPU)\n"
+ " MY_HISTOGRAM( localKey )++;\n"
+ "#else\n"
+ " AtomInc( MY_HISTOGRAM( localKey ) );\n"
+ "#endif\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " \n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<GET_GROUP_SIZE; i++)\n"
+ " {\n"
+ " sum += localHistogramMat[lIdx*WG_SIZE+(i+lIdx)%GET_GROUP_SIZE];\n"
+ " }\n"
+ " histogramOut[lIdx*nWGs+wgIdx] = sum;\n"
+ " }\n"
+ "}\n"
+ "#define nPerLane (nPerWI/4)\n"
+ "// NUM_BUCKET*nWGs < 128*nPerWI\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(128,1,1)))\n"
+ "void PrefixScanKernel( __global u32* wHistogram1, int4 cb )\n"
+ "{\n"
+ " __local u32 ldsTopScanData[128*2];\n"
+ " u32 lIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " u32 data[nPerWI];\n"
+ " for(int i=0; i<nPerWI; i++)\n"
+ " {\n"
+ " data[i] = 0;\n"
+ " if( (nPerWI*lIdx+i) < NUM_BUCKET*nWGs )\n"
+ " data[i] = wHistogram1[nPerWI*lIdx+i];\n"
+ " }\n"
+ " uint4 myData = make_uint4(0,0,0,0);\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " myData.x += data[nPerLane*0+i];\n"
+ " myData.y += data[nPerLane*1+i];\n"
+ " myData.z += data[nPerLane*2+i];\n"
+ " myData.w += data[nPerLane*3+i];\n"
+ " }\n"
+ " uint totalSum;\n"
+ " uint4 scanned = localPrefixSum128V( myData, lIdx, &totalSum, ldsTopScanData );\n"
+ "// for(int j=0; j<4; j++) // somehow it introduces a lot of branches\n"
+ " { int j = 0;\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " u32 tmp = data[nPerLane*j+i];\n"
+ " data[nPerLane*j+i] = sum;\n"
+ " sum += tmp;\n"
+ " }\n"
+ " }\n"
+ " { int j = 1;\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " u32 tmp = data[nPerLane*j+i];\n"
+ " data[nPerLane*j+i] = sum;\n"
+ " sum += tmp;\n"
+ " }\n"
+ " }\n"
+ " { int j = 2;\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " u32 tmp = data[nPerLane*j+i];\n"
+ " data[nPerLane*j+i] = sum;\n"
+ " sum += tmp;\n"
+ " }\n"
+ " }\n"
+ " { int j = 3;\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " u32 tmp = data[nPerLane*j+i];\n"
+ " data[nPerLane*j+i] = sum;\n"
+ " sum += tmp;\n"
+ " }\n"
+ " }\n"
+ " for(int i=0; i<nPerLane; i++)\n"
+ " {\n"
+ " data[nPerLane*0+i] += scanned.x;\n"
+ " data[nPerLane*1+i] += scanned.y;\n"
+ " data[nPerLane*2+i] += scanned.z;\n"
+ " data[nPerLane*3+i] += scanned.w;\n"
+ " }\n"
+ " for(int i=0; i<nPerWI; i++)\n"
+ " {\n"
+ " int index = nPerWI*lIdx+i;\n"
+ " if (index < NUM_BUCKET*nWGs)\n"
+ " wHistogram1[nPerWI*lIdx+i] = data[i];\n"
+ " }\n"
+ "}\n"
+ "// 4 scan, 4 exchange\n"
+ "void sort4Bits(u32 sortData[4], int startBit, int lIdx, __local u32* ldsSortData)\n"
+ "{\n"
+ " for(int bitIdx=0; bitIdx<BITS_PER_PASS; bitIdx++)\n"
+ " {\n"
+ " u32 mask = (1<<bitIdx);\n"
+ " uint4 cmpResult = make_uint4( (sortData[0]>>startBit) & mask, (sortData[1]>>startBit) & mask, (sortData[2]>>startBit) & mask, (sortData[3]>>startBit) & mask );\n"
+ " uint4 prefixSum = SELECT_UINT4( make_uint4(1,1,1,1), make_uint4(0,0,0,0), cmpResult != make_uint4(0,0,0,0) );\n"
+ " u32 total;\n"
+ " prefixSum = localPrefixSum64V( prefixSum, lIdx, &total, ldsSortData );\n"
+ " {\n"
+ " uint4 localAddr = make_uint4(lIdx*4+0,lIdx*4+1,lIdx*4+2,lIdx*4+3);\n"
+ " uint4 dstAddr = localAddr - prefixSum + make_uint4( total, total, total, total );\n"
+ " dstAddr = SELECT_UINT4( prefixSum, dstAddr, cmpResult != make_uint4(0, 0, 0, 0) );\n"
+ " GROUP_LDS_BARRIER;\n"
+ " ldsSortData[dstAddr.x] = sortData[0];\n"
+ " ldsSortData[dstAddr.y] = sortData[1];\n"
+ " ldsSortData[dstAddr.z] = sortData[2];\n"
+ " ldsSortData[dstAddr.w] = sortData[3];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " sortData[0] = ldsSortData[localAddr.x];\n"
+ " sortData[1] = ldsSortData[localAddr.y];\n"
+ " sortData[2] = ldsSortData[localAddr.z];\n"
+ " sortData[3] = ldsSortData[localAddr.w];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ " }\n"
+ "}\n"
+ "// 2 scan, 2 exchange\n"
+ "void sort4Bits1(u32 sortData[4], int startBit, int lIdx, __local u32* ldsSortData)\n"
+ "{\n"
+ " for(uint ibit=0; ibit<BITS_PER_PASS; ibit+=2)\n"
+ " {\n"
+ " uint4 b = make_uint4((sortData[0]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[1]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[2]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[3]>>(startBit+ibit)) & 0x3);\n"
+ " u32 key4;\n"
+ " u32 sKeyPacked[4] = { 0, 0, 0, 0 };\n"
+ " {\n"
+ " sKeyPacked[0] |= 1<<(8*b.x);\n"
+ " sKeyPacked[1] |= 1<<(8*b.y);\n"
+ " sKeyPacked[2] |= 1<<(8*b.z);\n"
+ " sKeyPacked[3] |= 1<<(8*b.w);\n"
+ " key4 = sKeyPacked[0] + sKeyPacked[1] + sKeyPacked[2] + sKeyPacked[3];\n"
+ " }\n"
+ " u32 rankPacked;\n"
+ " u32 sumPacked;\n"
+ " {\n"
+ " rankPacked = localPrefixSum( key4, lIdx, &sumPacked, ldsSortData, WG_SIZE );\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " u32 newOffset[4] = { 0,0,0,0 };\n"
+ " {\n"
+ " u32 sumScanned = bit8Scan( sumPacked );\n"
+ " u32 scannedKeys[4];\n"
+ " scannedKeys[0] = 1<<(8*b.x);\n"
+ " scannedKeys[1] = 1<<(8*b.y);\n"
+ " scannedKeys[2] = 1<<(8*b.z);\n"
+ " scannedKeys[3] = 1<<(8*b.w);\n"
+ " { // 4 scans at once\n"
+ " u32 sum4 = 0;\n"
+ " for(int ie=0; ie<4; ie++)\n"
+ " {\n"
+ " u32 tmp = scannedKeys[ie];\n"
+ " scannedKeys[ie] = sum4;\n"
+ " sum4 += tmp;\n"
+ " }\n"
+ " }\n"
+ " {\n"
+ " u32 sumPlusRank = sumScanned + rankPacked;\n"
+ " { u32 ie = b.x;\n"
+ " scannedKeys[0] += sumPlusRank;\n"
+ " newOffset[0] = unpack4Key( scannedKeys[0], ie );\n"
+ " }\n"
+ " { u32 ie = b.y;\n"
+ " scannedKeys[1] += sumPlusRank;\n"
+ " newOffset[1] = unpack4Key( scannedKeys[1], ie );\n"
+ " }\n"
+ " { u32 ie = b.z;\n"
+ " scannedKeys[2] += sumPlusRank;\n"
+ " newOffset[2] = unpack4Key( scannedKeys[2], ie );\n"
+ " }\n"
+ " { u32 ie = b.w;\n"
+ " scannedKeys[3] += sumPlusRank;\n"
+ " newOffset[3] = unpack4Key( scannedKeys[3], ie );\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " {\n"
+ " ldsSortData[newOffset[0]] = sortData[0];\n"
+ " ldsSortData[newOffset[1]] = sortData[1];\n"
+ " ldsSortData[newOffset[2]] = sortData[2];\n"
+ " ldsSortData[newOffset[3]] = sortData[3];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " u32 dstAddr = 4*lIdx;\n"
+ " sortData[0] = ldsSortData[dstAddr+0];\n"
+ " sortData[1] = ldsSortData[dstAddr+1];\n"
+ " sortData[2] = ldsSortData[dstAddr+2];\n"
+ " sortData[3] = ldsSortData[dstAddr+3];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ " }\n"
+ "}\n"
+ "#define SET_HISTOGRAM(setIdx, key) ldsSortData[(setIdx)*NUM_BUCKET+key]\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void SortAndScatterKernel( __global const u32* restrict gSrc, __global const u32* rHistogram, __global u32* restrict gDst, int4 cb )\n"
+ "{\n"
+ " __local u32 ldsSortData[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
+ " __local u32 localHistogramToCarry[NUM_BUCKET];\n"
+ " __local u32 localHistogram[NUM_BUCKET*2];\n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 lIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " if( lIdx < (NUM_BUCKET) )\n"
+ " {\n"
+ " localHistogramToCarry[lIdx] = rHistogram[lIdx*nWGs + wgIdx];\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " int nBlocks = n/blockSize - nBlocksPerWG*wgIdx;\n"
+ " int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
+ " {\n"
+ " u32 myHistogram = 0;\n"
+ " u32 sortData[ELEMENTS_PER_WORK_ITEM];\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " sortData[i] = ( addr+i < n )? gSrc[ addr+i ] : 0xffffffff;\n"
+ "#else\n"
+ " sortData[i] = gSrc[ addr+i ];\n"
+ "#endif\n"
+ " sort4Bits(sortData, startBit, lIdx, ldsSortData);\n"
+ " u32 keys[ELEMENTS_PER_WORK_ITEM];\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ " keys[i] = (sortData[i]>>startBit) & 0xf;\n"
+ " { // create histogram\n"
+ " u32 setIdx = lIdx/16;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[lIdx] = 0;\n"
+ " }\n"
+ " ldsSortData[lIdx] = 0;\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+i < n )\n"
+ "#endif\n"
+ "#if defined(NV_GPU)\n"
+ " SET_HISTOGRAM( setIdx, keys[i] )++;\n"
+ "#else\n"
+ " AtomInc( SET_HISTOGRAM( setIdx, keys[i] ) );\n"
+ "#endif\n"
+ " \n"
+ " GROUP_LDS_BARRIER;\n"
+ " \n"
+ " uint hIdx = NUM_BUCKET+lIdx;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<WG_SIZE/16; i++)\n"
+ " {\n"
+ " sum += SET_HISTOGRAM( i, lIdx );\n"
+ " }\n"
+ " myHistogram = sum;\n"
+ " localHistogram[hIdx] = sum;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ "#if defined(USE_2LEVEL_REDUCE)\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[hIdx] = localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " u32 u0, u1, u2;\n"
+ " u0 = localHistogram[hIdx-3];\n"
+ " u1 = localHistogram[hIdx-2];\n"
+ " u2 = localHistogram[hIdx-1];\n"
+ " AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
+ " GROUP_MEM_FENCE;\n"
+ " u0 = localHistogram[hIdx-12];\n"
+ " u1 = localHistogram[hIdx-8];\n"
+ " u2 = localHistogram[hIdx-4];\n"
+ " AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#else\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[hIdx] = localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-2];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-4];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-8];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#endif\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ " {\n"
+ " for(int ie=0; ie<ELEMENTS_PER_WORK_ITEM; ie++)\n"
+ " {\n"
+ " int dataIdx = ELEMENTS_PER_WORK_ITEM*lIdx+ie;\n"
+ " int binIdx = keys[ie];\n"
+ " int groupOffset = localHistogramToCarry[binIdx];\n"
+ " int myIdx = dataIdx - localHistogram[NUM_BUCKET+binIdx];\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+ie < n )\n"
+ "#endif\n"
+ " gDst[ groupOffset + myIdx ] = sortData[ie];\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogramToCarry[lIdx] += myHistogram;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ "}\n"
+ "// 2 scan, 2 exchange\n"
+ "void sort4Bits1KeyValue(u32 sortData[4], int sortVal[4], int startBit, int lIdx, __local u32* ldsSortData, __local int *ldsSortVal)\n"
+ "{\n"
+ " for(uint ibit=0; ibit<BITS_PER_PASS; ibit+=2)\n"
+ " {\n"
+ " uint4 b = make_uint4((sortData[0]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[1]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[2]>>(startBit+ibit)) & 0x3, \n"
+ " (sortData[3]>>(startBit+ibit)) & 0x3);\n"
+ " u32 key4;\n"
+ " u32 sKeyPacked[4] = { 0, 0, 0, 0 };\n"
+ " {\n"
+ " sKeyPacked[0] |= 1<<(8*b.x);\n"
+ " sKeyPacked[1] |= 1<<(8*b.y);\n"
+ " sKeyPacked[2] |= 1<<(8*b.z);\n"
+ " sKeyPacked[3] |= 1<<(8*b.w);\n"
+ " key4 = sKeyPacked[0] + sKeyPacked[1] + sKeyPacked[2] + sKeyPacked[3];\n"
+ " }\n"
+ " u32 rankPacked;\n"
+ " u32 sumPacked;\n"
+ " {\n"
+ " rankPacked = localPrefixSum( key4, lIdx, &sumPacked, ldsSortData, WG_SIZE );\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " u32 newOffset[4] = { 0,0,0,0 };\n"
+ " {\n"
+ " u32 sumScanned = bit8Scan( sumPacked );\n"
+ " u32 scannedKeys[4];\n"
+ " scannedKeys[0] = 1<<(8*b.x);\n"
+ " scannedKeys[1] = 1<<(8*b.y);\n"
+ " scannedKeys[2] = 1<<(8*b.z);\n"
+ " scannedKeys[3] = 1<<(8*b.w);\n"
+ " { // 4 scans at once\n"
+ " u32 sum4 = 0;\n"
+ " for(int ie=0; ie<4; ie++)\n"
+ " {\n"
+ " u32 tmp = scannedKeys[ie];\n"
+ " scannedKeys[ie] = sum4;\n"
+ " sum4 += tmp;\n"
+ " }\n"
+ " }\n"
+ " {\n"
+ " u32 sumPlusRank = sumScanned + rankPacked;\n"
+ " { u32 ie = b.x;\n"
+ " scannedKeys[0] += sumPlusRank;\n"
+ " newOffset[0] = unpack4Key( scannedKeys[0], ie );\n"
+ " }\n"
+ " { u32 ie = b.y;\n"
+ " scannedKeys[1] += sumPlusRank;\n"
+ " newOffset[1] = unpack4Key( scannedKeys[1], ie );\n"
+ " }\n"
+ " { u32 ie = b.z;\n"
+ " scannedKeys[2] += sumPlusRank;\n"
+ " newOffset[2] = unpack4Key( scannedKeys[2], ie );\n"
+ " }\n"
+ " { u32 ie = b.w;\n"
+ " scannedKeys[3] += sumPlusRank;\n"
+ " newOffset[3] = unpack4Key( scannedKeys[3], ie );\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " {\n"
+ " ldsSortData[newOffset[0]] = sortData[0];\n"
+ " ldsSortData[newOffset[1]] = sortData[1];\n"
+ " ldsSortData[newOffset[2]] = sortData[2];\n"
+ " ldsSortData[newOffset[3]] = sortData[3];\n"
+ " ldsSortVal[newOffset[0]] = sortVal[0];\n"
+ " ldsSortVal[newOffset[1]] = sortVal[1];\n"
+ " ldsSortVal[newOffset[2]] = sortVal[2];\n"
+ " ldsSortVal[newOffset[3]] = sortVal[3];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " u32 dstAddr = 4*lIdx;\n"
+ " sortData[0] = ldsSortData[dstAddr+0];\n"
+ " sortData[1] = ldsSortData[dstAddr+1];\n"
+ " sortData[2] = ldsSortData[dstAddr+2];\n"
+ " sortData[3] = ldsSortData[dstAddr+3];\n"
+ " sortVal[0] = ldsSortVal[dstAddr+0];\n"
+ " sortVal[1] = ldsSortVal[dstAddr+1];\n"
+ " sortVal[2] = ldsSortVal[dstAddr+2];\n"
+ " sortVal[3] = ldsSortVal[dstAddr+3];\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void SortAndScatterSortDataKernel( __global const SortDataCL* restrict gSrc, __global const u32* rHistogram, __global SortDataCL* restrict gDst, int4 cb)\n"
+ "{\n"
+ " __local int ldsSortData[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
+ " __local int ldsSortVal[WG_SIZE*ELEMENTS_PER_WORK_ITEM+16];\n"
+ " __local u32 localHistogramToCarry[NUM_BUCKET];\n"
+ " __local u32 localHistogram[NUM_BUCKET*2];\n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 lIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " if( lIdx < (NUM_BUCKET) )\n"
+ " {\n"
+ " localHistogramToCarry[lIdx] = rHistogram[lIdx*nWGs + wgIdx];\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " \n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " int nBlocks = n/blockSize - nBlocksPerWG*wgIdx;\n"
+ " int addr = blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++, addr+=blockSize)\n"
+ " {\n"
+ " u32 myHistogram = 0;\n"
+ " int sortData[ELEMENTS_PER_WORK_ITEM];\n"
+ " int sortVal[ELEMENTS_PER_WORK_ITEM];\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " {\n"
+ " sortData[i] = ( addr+i < n )? gSrc[ addr+i ].m_key : 0xffffffff;\n"
+ " sortVal[i] = ( addr+i < n )? gSrc[ addr+i ].m_value : 0xffffffff;\n"
+ " }\n"
+ "#else\n"
+ " {\n"
+ " sortData[i] = gSrc[ addr+i ].m_key;\n"
+ " sortVal[i] = gSrc[ addr+i ].m_value;\n"
+ " }\n"
+ "#endif\n"
+ " sort4Bits1KeyValue(sortData, sortVal, startBit, lIdx, ldsSortData, ldsSortVal);\n"
+ " u32 keys[ELEMENTS_PER_WORK_ITEM];\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ " keys[i] = (sortData[i]>>startBit) & 0xf;\n"
+ " { // create histogram\n"
+ " u32 setIdx = lIdx/16;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[lIdx] = 0;\n"
+ " }\n"
+ " ldsSortData[lIdx] = 0;\n"
+ " GROUP_LDS_BARRIER;\n"
+ " for(int i=0; i<ELEMENTS_PER_WORK_ITEM; i++)\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+i < n )\n"
+ "#endif\n"
+ "#if defined(NV_GPU)\n"
+ " SET_HISTOGRAM( setIdx, keys[i] )++;\n"
+ "#else\n"
+ " AtomInc( SET_HISTOGRAM( setIdx, keys[i] ) );\n"
+ "#endif\n"
+ " \n"
+ " GROUP_LDS_BARRIER;\n"
+ " \n"
+ " uint hIdx = NUM_BUCKET+lIdx;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " u32 sum = 0;\n"
+ " for(int i=0; i<WG_SIZE/16; i++)\n"
+ " {\n"
+ " sum += SET_HISTOGRAM( i, lIdx );\n"
+ " }\n"
+ " myHistogram = sum;\n"
+ " localHistogram[hIdx] = sum;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ "#if defined(USE_2LEVEL_REDUCE)\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[hIdx] = localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " u32 u0, u1, u2;\n"
+ " u0 = localHistogram[hIdx-3];\n"
+ " u1 = localHistogram[hIdx-2];\n"
+ " u2 = localHistogram[hIdx-1];\n"
+ " AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
+ " GROUP_MEM_FENCE;\n"
+ " u0 = localHistogram[hIdx-12];\n"
+ " u1 = localHistogram[hIdx-8];\n"
+ " u2 = localHistogram[hIdx-4];\n"
+ " AtomAdd( localHistogram[hIdx], u0 + u1 + u2 );\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#else\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogram[hIdx] = localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-1];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-2];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-4];\n"
+ " GROUP_MEM_FENCE;\n"
+ " localHistogram[hIdx] += localHistogram[hIdx-8];\n"
+ " GROUP_MEM_FENCE;\n"
+ " }\n"
+ "#endif\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ " {\n"
+ " for(int ie=0; ie<ELEMENTS_PER_WORK_ITEM; ie++)\n"
+ " {\n"
+ " int dataIdx = ELEMENTS_PER_WORK_ITEM*lIdx+ie;\n"
+ " int binIdx = keys[ie];\n"
+ " int groupOffset = localHistogramToCarry[binIdx];\n"
+ " int myIdx = dataIdx - localHistogram[NUM_BUCKET+binIdx];\n"
+ "#if defined(CHECK_BOUNDARY)\n"
+ " if( addr+ie < n )\n"
+ " {\n"
+ " if ((groupOffset + myIdx)<n)\n"
+ " {\n"
+ " if (sortData[ie]==sortVal[ie])\n"
+ " {\n"
+ " \n"
+ " SortDataCL tmp;\n"
+ " tmp.m_key = sortData[ie];\n"
+ " tmp.m_value = sortVal[ie];\n"
+ " if (tmp.m_key == tmp.m_value)\n"
+ " gDst[groupOffset + myIdx ] = tmp;\n"
+ " }\n"
+ " \n"
+ " }\n"
+ " }\n"
+ "#else\n"
+ " if ((groupOffset + myIdx)<n)\n"
+ " {\n"
+ " gDst[ groupOffset + myIdx ].m_key = sortData[ie];\n"
+ " gDst[ groupOffset + myIdx ].m_value = sortVal[ie];\n"
+ " }\n"
+ "#endif\n"
+ " }\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " if( lIdx < NUM_BUCKET )\n"
+ " {\n"
+ " localHistogramToCarry[lIdx] += myHistogram;\n"
+ " }\n"
+ " GROUP_LDS_BARRIER;\n"
+ " }\n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void SortAndScatterSortDataKernelSerial( __global const SortDataCL* restrict gSrc, __global const u32* rHistogram, __global SortDataCL* restrict gDst, int4 cb)\n"
+ "{\n"
+ " \n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 realLocalIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " int counter[NUM_BUCKET];\n"
+ " \n"
+ " if (realLocalIdx>0)\n"
+ " return;\n"
+ " \n"
+ " for (int c=0;c<NUM_BUCKET;c++)\n"
+ " counter[c]=0;\n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " \n"
+ " int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++)\n"
+ " {\n"
+ " for (int lIdx=0;lIdx<WG_SIZE;lIdx++)\n"
+ " {\n"
+ " int addr2 = iblock*blockSize + blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " \n"
+ " for(int j=0; j<ELEMENTS_PER_WORK_ITEM; j++)\n"
+ " {\n"
+ " int i = addr2+j;\n"
+ " if( i < n )\n"
+ " {\n"
+ " int tableIdx;\n"
+ " tableIdx = (gSrc[i].m_key>>startBit) & 0xf;//0xf = NUM_TABLES-1\n"
+ " gDst[rHistogram[tableIdx*nWGs+wgIdx] + counter[tableIdx]] = gSrc[i];\n"
+ " counter[tableIdx] ++;\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " \n"
+ "}\n"
+ "__kernel\n"
+ "__attribute__((reqd_work_group_size(WG_SIZE,1,1)))\n"
+ "void SortAndScatterKernelSerial( __global const u32* restrict gSrc, __global const u32* rHistogram, __global u32* restrict gDst, int4 cb )\n"
+ "{\n"
+ " \n"
+ " u32 gIdx = GET_GLOBAL_IDX;\n"
+ " u32 realLocalIdx = GET_LOCAL_IDX;\n"
+ " u32 wgIdx = GET_GROUP_IDX;\n"
+ " u32 wgSize = GET_GROUP_SIZE;\n"
+ " const int startBit = cb.m_startBit;\n"
+ " const int n = cb.m_n;\n"
+ " const int nWGs = cb.m_nWGs;\n"
+ " const int nBlocksPerWG = cb.m_nBlocksPerWG;\n"
+ " int counter[NUM_BUCKET];\n"
+ " \n"
+ " if (realLocalIdx>0)\n"
+ " return;\n"
+ " \n"
+ " for (int c=0;c<NUM_BUCKET;c++)\n"
+ " counter[c]=0;\n"
+ " const int blockSize = ELEMENTS_PER_WORK_ITEM*WG_SIZE;\n"
+ " \n"
+ " int nBlocks = (n)/blockSize - nBlocksPerWG*wgIdx;\n"
+ " for(int iblock=0; iblock<min(nBlocksPerWG, nBlocks); iblock++)\n"
+ " {\n"
+ " for (int lIdx=0;lIdx<WG_SIZE;lIdx++)\n"
+ " {\n"
+ " int addr2 = iblock*blockSize + blockSize*nBlocksPerWG*wgIdx + ELEMENTS_PER_WORK_ITEM*lIdx;\n"
+ " \n"
+ " for(int j=0; j<ELEMENTS_PER_WORK_ITEM; j++)\n"
+ " {\n"
+ " int i = addr2+j;\n"
+ " if( i < n )\n"
+ " {\n"
+ " int tableIdx;\n"
+ " tableIdx = (gSrc[i]>>startBit) & 0xf;//0xf = NUM_TABLES-1\n"
+ " gDst[rHistogram[tableIdx*nWGs+wgIdx] + counter[tableIdx]] = gSrc[i];\n"
+ " counter[tableIdx] ++;\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " }\n"
+ " \n"
+ "}\n";