mirror of
https://github.com/KhronosGroup/OpenCL-CTS.git
synced 2026-03-19 06:09:01 +00:00
New subgroups - full changes set (#1074)
* Extended subgroups - extended types types * Extended subgroups - non uniform vote tests * Extended subgroups - non uniform arithmetic tests * Extended subgroups - ballot tests * Extended subgroups - clustered reduce tests * Extended subgroups - shuffle tests * Extended subgroups - formating issues * Extended subgroups - review fixes * Extended subgroups - review fixes Fixed: removed additional brakes, kernel_sstr * Extended subgroups - fix macos build error * Extended subgroups - review fixes Fixed: mac os build error * Extended subgroups - data type verification example * Extended subgroups - error unification * Extended subgroups - fix header years * Extended subgroups - use is_half_nan * Extended subgroups - compare half as float * Review fixes mostly for ballot functions. - Modify kernels for better handling active/inactive workitems - Modify gen/chk functions for handling non uniform workgroup sizes - Introduce new variables naming convention - minor fixes * Extended subgroups - simplification data generation for ballot lsb/msb functions * Extended subgroups - minor fixes * Extended subgroups - move common code to function * Extended subgroups - formatting errors fix * Extended subgroups - fix build error * Extended subgroups - sub_group_elect more sophisticated Define mask which is 4bytes pattern where bit 1 means work item is active. If workitem in subgroup matches pattern then run sub_group_elect() * Extended subgroups - fix Ubuntu build error * Extended subgroups - voting function review fixes * adjust all function for using masks * remove calculate templates * merge code to one common template * check results only in active workitems * normalize values on host side * minor fixes * Extended subgroups - fix typos * Set of fixes and improvements after review * define WorkGroupParams to stop extended parameters list in function * better workitems mask handing (WorkGroupParams) * narrow values of data input generation to avoid overflows (arithmetic func) * implement work item masks for arithmetic functions * enable half type testing for reduction/scan/broadcast * minor fixes * Extended subgroups - fix Linux issues * Extended subgroups - fix sub_group_local_id data type * Extended subgroups - use vector instead of array. * Extended subgroups - change names to subgroup * Extended subgroups - uncomment code, fix build * Extended subgroups - build fix, use cl_half_from_float func * Extended subgroups - remove is_half_nan * Extended subgroups - do no use undef min/max * Extended subgroups - use parenthesis, fix formatting
This commit is contained in:
committed by
GitHub
parent
c5e4ca6c91
commit
71bef8563e
340
test_conformance/subgroups/test_subgroup_clustered_reduce.cpp
Normal file
340
test_conformance/subgroups/test_subgroup_clustered_reduce.cpp
Normal file
@@ -0,0 +1,340 @@
|
||||
//
|
||||
// Copyright (c) 2021 The Khronos Group Inc.
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
//
|
||||
#include "procs.h"
|
||||
#include "subhelpers.h"
|
||||
#include "subgroup_common_templates.h"
|
||||
#include "harness/typeWrappers.h"
|
||||
|
||||
#define CLUSTER_SIZE 4
|
||||
#define CLUSTER_SIZE_STR "4"
|
||||
|
||||
namespace {
|
||||
static const char *redadd_clustered_source =
|
||||
"__kernel void test_redadd_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_add(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_add(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redmax_clustered_source =
|
||||
"__kernel void test_redmax_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_max(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_max(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redmin_clustered_source =
|
||||
"__kernel void test_redmin_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_min(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_min(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redmul_clustered_source =
|
||||
"__kernel void test_redmul_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_mul(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_mul(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redand_clustered_source =
|
||||
"__kernel void test_redand_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_and(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_and(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redor_clustered_source =
|
||||
"__kernel void test_redor_clustered(const __global Type *in, __global int4 "
|
||||
"*xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_or(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_or(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redxor_clustered_source =
|
||||
"__kernel void test_redxor_clustered(const __global Type *in, __global "
|
||||
"int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_xor(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = sub_group_clustered_reduce_xor(in[gid], " CLUSTER_SIZE_STR
|
||||
");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redand_clustered_logical_source =
|
||||
"__kernel void test_redand_clustered_logical(const __global Type *in, "
|
||||
"__global int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_logical_and(in[gid], " CLUSTER_SIZE_STR
|
||||
")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = "
|
||||
"sub_group_clustered_reduce_logical_and(in[gid], " CLUSTER_SIZE_STR ");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redor_clustered_logical_source =
|
||||
"__kernel void test_redor_clustered_logical(const __global Type *in, "
|
||||
"__global int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if (sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_logical_or(in[gid], " CLUSTER_SIZE_STR
|
||||
")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = "
|
||||
"sub_group_clustered_reduce_logical_or(in[gid], " CLUSTER_SIZE_STR ");\n"
|
||||
"}\n";
|
||||
|
||||
static const char *redxor_clustered_logical_source =
|
||||
"__kernel void test_redxor_clustered_logical(const __global Type *in, "
|
||||
"__global int4 *xy, __global Type *out)\n"
|
||||
"{\n"
|
||||
" int gid = get_global_id(0);\n"
|
||||
" XY(xy,gid);\n"
|
||||
" xy[gid].w = 0;\n"
|
||||
" if ( sizeof(in[gid]) == "
|
||||
"sizeof(sub_group_clustered_reduce_logical_xor(in[gid], " CLUSTER_SIZE_STR
|
||||
")))\n"
|
||||
" {xy[gid].w = sizeof(in[gid]);}\n"
|
||||
" out[gid] = "
|
||||
"sub_group_clustered_reduce_logical_xor(in[gid], " CLUSTER_SIZE_STR ");\n"
|
||||
"}\n";
|
||||
|
||||
|
||||
// DESCRIPTION:
|
||||
// Test for reduce cluster functions
|
||||
template <typename Ty, ArithmeticOp operation> struct RED_CLU
|
||||
{
|
||||
static void gen(Ty *x, Ty *t, cl_int *m, const WorkGroupParams &test_params)
|
||||
{
|
||||
int nw = test_params.local_workgroup_size;
|
||||
int ns = test_params.subgroup_size;
|
||||
int ng = test_params.global_workgroup_size;
|
||||
ng = ng / nw;
|
||||
log_info(" sub_group_clustered_reduce_%s(%s, %d bytes) ...\n",
|
||||
operation_names(operation), TypeManager<Ty>::name(),
|
||||
sizeof(Ty));
|
||||
genrand<Ty, operation>(x, t, m, ns, nw, ng);
|
||||
}
|
||||
|
||||
static int chk(Ty *x, Ty *y, Ty *mx, Ty *my, cl_int *m,
|
||||
const WorkGroupParams &test_params)
|
||||
{
|
||||
int nw = test_params.local_workgroup_size;
|
||||
int ns = test_params.subgroup_size;
|
||||
int ng = test_params.global_workgroup_size;
|
||||
int nj = (nw + ns - 1) / ns;
|
||||
ng = ng / nw;
|
||||
|
||||
for (int k = 0; k < ng; ++k)
|
||||
{
|
||||
std::vector<cl_int> data_type_sizes;
|
||||
// Map to array indexed to array indexed by local ID and sub group
|
||||
for (int j = 0; j < nw; ++j)
|
||||
{
|
||||
mx[j] = x[j];
|
||||
my[j] = y[j];
|
||||
data_type_sizes.push_back(m[4 * j + 3]);
|
||||
}
|
||||
|
||||
for (cl_int dts : data_type_sizes)
|
||||
{
|
||||
if (dts != sizeof(Ty))
|
||||
{
|
||||
log_error("ERROR: sub_group_clustered_reduce_%s(%s) "
|
||||
"wrong data type size detected, expected: %d, "
|
||||
"used by device %d, in group %d\n",
|
||||
operation_names(operation),
|
||||
TypeManager<Ty>::name(), sizeof(Ty), dts, k);
|
||||
return TEST_FAIL;
|
||||
}
|
||||
}
|
||||
|
||||
for (int j = 0; j < nj; ++j)
|
||||
{
|
||||
int ii = j * ns;
|
||||
int n = ii + ns > nw ? nw - ii : ns;
|
||||
int midx = 4 * ii + 2;
|
||||
std::vector<Ty> clusters_results;
|
||||
int clusters_counter = ns / CLUSTER_SIZE;
|
||||
clusters_results.resize(clusters_counter);
|
||||
|
||||
// Compute target
|
||||
Ty tr = mx[ii];
|
||||
for (int i = 0; i < n; ++i)
|
||||
{
|
||||
if (i % CLUSTER_SIZE == 0)
|
||||
tr = mx[ii + i];
|
||||
else
|
||||
tr = calculate<Ty>(tr, mx[ii + i], operation);
|
||||
clusters_results[i / CLUSTER_SIZE] = tr;
|
||||
}
|
||||
|
||||
// Check result
|
||||
for (int i = 0; i < n; ++i)
|
||||
{
|
||||
Ty rr = my[ii + i];
|
||||
tr = clusters_results[i / CLUSTER_SIZE];
|
||||
if (!compare(rr, tr))
|
||||
{
|
||||
log_error(
|
||||
"ERROR: sub_group_clustered_reduce_%s(%s) mismatch "
|
||||
"for local id %d in sub group %d in group %d\n",
|
||||
operation_names(operation), TypeManager<Ty>::name(),
|
||||
i, j, k);
|
||||
return TEST_FAIL;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
x += nw;
|
||||
y += nw;
|
||||
m += 4 * nw;
|
||||
}
|
||||
log_info(" sub_group_clustered_reduce_%s(%s, %d bytes) ... passed\n",
|
||||
operation_names(operation), TypeManager<Ty>::name(),
|
||||
sizeof(Ty));
|
||||
return TEST_PASS;
|
||||
}
|
||||
};
|
||||
|
||||
template <typename T>
|
||||
int run_cluster_red_add_max_min_mul_for_type(RunTestForType rft)
|
||||
{
|
||||
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::add_>>(
|
||||
"test_redadd_clustered", redadd_clustered_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::max_>>(
|
||||
"test_redmax_clustered", redmax_clustered_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::min_>>(
|
||||
"test_redmin_clustered", redmin_clustered_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::mul_>>(
|
||||
"test_redmul_clustered", redmul_clustered_source);
|
||||
return error;
|
||||
}
|
||||
template <typename T> int run_cluster_and_or_xor_for_type(RunTestForType rft)
|
||||
{
|
||||
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::and_>>(
|
||||
"test_redand_clustered", redand_clustered_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::or_>>(
|
||||
"test_redor_clustered", redor_clustered_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::xor_>>(
|
||||
"test_redxor_clustered", redxor_clustered_source);
|
||||
return error;
|
||||
}
|
||||
template <typename T>
|
||||
int run_cluster_logical_and_or_xor_for_type(RunTestForType rft)
|
||||
{
|
||||
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_and>>(
|
||||
"test_redand_clustered_logical", redand_clustered_logical_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_or>>(
|
||||
"test_redor_clustered_logical", redor_clustered_logical_source);
|
||||
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_xor>>(
|
||||
"test_redxor_clustered_logical", redxor_clustered_logical_source);
|
||||
|
||||
return error;
|
||||
}
|
||||
}
|
||||
|
||||
int test_subgroup_functions_clustered_reduce(cl_device_id device,
|
||||
cl_context context,
|
||||
cl_command_queue queue,
|
||||
int num_elements)
|
||||
{
|
||||
std::vector<std::string> required_extensions = {
|
||||
"cl_khr_subgroup_clustered_reduce"
|
||||
};
|
||||
constexpr size_t global_work_size = 2000;
|
||||
constexpr size_t local_work_size = 200;
|
||||
WorkGroupParams test_params(global_work_size, local_work_size,
|
||||
required_extensions);
|
||||
RunTestForType rft(device, context, queue, num_elements, test_params);
|
||||
|
||||
int error = run_cluster_red_add_max_min_mul_for_type<cl_int>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_uint>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_long>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_ulong>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_short>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_ushort>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_char>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_uchar>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_float>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<cl_double>(rft);
|
||||
error |= run_cluster_red_add_max_min_mul_for_type<subgroups::cl_half>(rft);
|
||||
|
||||
error |= run_cluster_and_or_xor_for_type<cl_int>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_uint>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_long>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_ulong>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_short>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_ushort>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_char>(rft);
|
||||
error |= run_cluster_and_or_xor_for_type<cl_uchar>(rft);
|
||||
|
||||
error |= run_cluster_logical_and_or_xor_for_type<cl_int>(rft);
|
||||
return error;
|
||||
}
|
||||
Reference in New Issue
Block a user