mirror of
https://github.com/KhronosGroup/OpenCL-CTS.git
synced 2026-03-19 06:09:01 +00:00
* Extended subgroups - extended types types * Extended subgroups - non uniform vote tests * Extended subgroups - non uniform arithmetic tests * Extended subgroups - ballot tests * Extended subgroups - clustered reduce tests * Extended subgroups - shuffle tests * Extended subgroups - formating issues * Extended subgroups - review fixes * Extended subgroups - review fixes Fixed: removed additional brakes, kernel_sstr * Extended subgroups - fix macos build error * Extended subgroups - review fixes Fixed: mac os build error * Extended subgroups - data type verification example * Extended subgroups - error unification * Extended subgroups - fix header years * Extended subgroups - use is_half_nan * Extended subgroups - compare half as float * Review fixes mostly for ballot functions. - Modify kernels for better handling active/inactive workitems - Modify gen/chk functions for handling non uniform workgroup sizes - Introduce new variables naming convention - minor fixes * Extended subgroups - simplification data generation for ballot lsb/msb functions * Extended subgroups - minor fixes * Extended subgroups - move common code to function * Extended subgroups - formatting errors fix * Extended subgroups - fix build error * Extended subgroups - sub_group_elect more sophisticated Define mask which is 4bytes pattern where bit 1 means work item is active. If workitem in subgroup matches pattern then run sub_group_elect() * Extended subgroups - fix Ubuntu build error * Extended subgroups - voting function review fixes * adjust all function for using masks * remove calculate templates * merge code to one common template * check results only in active workitems * normalize values on host side * minor fixes * Extended subgroups - fix typos * Set of fixes and improvements after review * define WorkGroupParams to stop extended parameters list in function * better workitems mask handing (WorkGroupParams) * narrow values of data input generation to avoid overflows (arithmetic func) * implement work item masks for arithmetic functions * enable half type testing for reduction/scan/broadcast * minor fixes * Extended subgroups - fix Linux issues * Extended subgroups - fix sub_group_local_id data type * Extended subgroups - use vector instead of array. * Extended subgroups - change names to subgroup * Extended subgroups - uncomment code, fix build * Extended subgroups - build fix, use cl_half_from_float func * Extended subgroups - remove is_half_nan * Extended subgroups - do no use undef min/max * Extended subgroups - use parenthesis, fix formatting
341 lines
13 KiB
C++
341 lines
13 KiB
C++
//
|
|
// Copyright (c) 2021 The Khronos Group Inc.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
//
|
|
#include "procs.h"
|
|
#include "subhelpers.h"
|
|
#include "subgroup_common_templates.h"
|
|
#include "harness/typeWrappers.h"
|
|
|
|
#define CLUSTER_SIZE 4
|
|
#define CLUSTER_SIZE_STR "4"
|
|
|
|
namespace {
|
|
static const char *redadd_clustered_source =
|
|
"__kernel void test_redadd_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_add(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_add(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redmax_clustered_source =
|
|
"__kernel void test_redmax_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_max(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_max(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redmin_clustered_source =
|
|
"__kernel void test_redmin_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_min(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_min(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redmul_clustered_source =
|
|
"__kernel void test_redmul_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_mul(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_mul(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redand_clustered_source =
|
|
"__kernel void test_redand_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_and(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_and(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redor_clustered_source =
|
|
"__kernel void test_redor_clustered(const __global Type *in, __global int4 "
|
|
"*xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_or(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_or(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redxor_clustered_source =
|
|
"__kernel void test_redxor_clustered(const __global Type *in, __global "
|
|
"int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_xor(in[gid], " CLUSTER_SIZE_STR ")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = sub_group_clustered_reduce_xor(in[gid], " CLUSTER_SIZE_STR
|
|
");\n"
|
|
"}\n";
|
|
|
|
static const char *redand_clustered_logical_source =
|
|
"__kernel void test_redand_clustered_logical(const __global Type *in, "
|
|
"__global int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_logical_and(in[gid], " CLUSTER_SIZE_STR
|
|
")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = "
|
|
"sub_group_clustered_reduce_logical_and(in[gid], " CLUSTER_SIZE_STR ");\n"
|
|
"}\n";
|
|
|
|
static const char *redor_clustered_logical_source =
|
|
"__kernel void test_redor_clustered_logical(const __global Type *in, "
|
|
"__global int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if (sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_logical_or(in[gid], " CLUSTER_SIZE_STR
|
|
")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = "
|
|
"sub_group_clustered_reduce_logical_or(in[gid], " CLUSTER_SIZE_STR ");\n"
|
|
"}\n";
|
|
|
|
static const char *redxor_clustered_logical_source =
|
|
"__kernel void test_redxor_clustered_logical(const __global Type *in, "
|
|
"__global int4 *xy, __global Type *out)\n"
|
|
"{\n"
|
|
" int gid = get_global_id(0);\n"
|
|
" XY(xy,gid);\n"
|
|
" xy[gid].w = 0;\n"
|
|
" if ( sizeof(in[gid]) == "
|
|
"sizeof(sub_group_clustered_reduce_logical_xor(in[gid], " CLUSTER_SIZE_STR
|
|
")))\n"
|
|
" {xy[gid].w = sizeof(in[gid]);}\n"
|
|
" out[gid] = "
|
|
"sub_group_clustered_reduce_logical_xor(in[gid], " CLUSTER_SIZE_STR ");\n"
|
|
"}\n";
|
|
|
|
|
|
// DESCRIPTION:
|
|
// Test for reduce cluster functions
|
|
template <typename Ty, ArithmeticOp operation> struct RED_CLU
|
|
{
|
|
static void gen(Ty *x, Ty *t, cl_int *m, const WorkGroupParams &test_params)
|
|
{
|
|
int nw = test_params.local_workgroup_size;
|
|
int ns = test_params.subgroup_size;
|
|
int ng = test_params.global_workgroup_size;
|
|
ng = ng / nw;
|
|
log_info(" sub_group_clustered_reduce_%s(%s, %d bytes) ...\n",
|
|
operation_names(operation), TypeManager<Ty>::name(),
|
|
sizeof(Ty));
|
|
genrand<Ty, operation>(x, t, m, ns, nw, ng);
|
|
}
|
|
|
|
static int chk(Ty *x, Ty *y, Ty *mx, Ty *my, cl_int *m,
|
|
const WorkGroupParams &test_params)
|
|
{
|
|
int nw = test_params.local_workgroup_size;
|
|
int ns = test_params.subgroup_size;
|
|
int ng = test_params.global_workgroup_size;
|
|
int nj = (nw + ns - 1) / ns;
|
|
ng = ng / nw;
|
|
|
|
for (int k = 0; k < ng; ++k)
|
|
{
|
|
std::vector<cl_int> data_type_sizes;
|
|
// Map to array indexed to array indexed by local ID and sub group
|
|
for (int j = 0; j < nw; ++j)
|
|
{
|
|
mx[j] = x[j];
|
|
my[j] = y[j];
|
|
data_type_sizes.push_back(m[4 * j + 3]);
|
|
}
|
|
|
|
for (cl_int dts : data_type_sizes)
|
|
{
|
|
if (dts != sizeof(Ty))
|
|
{
|
|
log_error("ERROR: sub_group_clustered_reduce_%s(%s) "
|
|
"wrong data type size detected, expected: %d, "
|
|
"used by device %d, in group %d\n",
|
|
operation_names(operation),
|
|
TypeManager<Ty>::name(), sizeof(Ty), dts, k);
|
|
return TEST_FAIL;
|
|
}
|
|
}
|
|
|
|
for (int j = 0; j < nj; ++j)
|
|
{
|
|
int ii = j * ns;
|
|
int n = ii + ns > nw ? nw - ii : ns;
|
|
int midx = 4 * ii + 2;
|
|
std::vector<Ty> clusters_results;
|
|
int clusters_counter = ns / CLUSTER_SIZE;
|
|
clusters_results.resize(clusters_counter);
|
|
|
|
// Compute target
|
|
Ty tr = mx[ii];
|
|
for (int i = 0; i < n; ++i)
|
|
{
|
|
if (i % CLUSTER_SIZE == 0)
|
|
tr = mx[ii + i];
|
|
else
|
|
tr = calculate<Ty>(tr, mx[ii + i], operation);
|
|
clusters_results[i / CLUSTER_SIZE] = tr;
|
|
}
|
|
|
|
// Check result
|
|
for (int i = 0; i < n; ++i)
|
|
{
|
|
Ty rr = my[ii + i];
|
|
tr = clusters_results[i / CLUSTER_SIZE];
|
|
if (!compare(rr, tr))
|
|
{
|
|
log_error(
|
|
"ERROR: sub_group_clustered_reduce_%s(%s) mismatch "
|
|
"for local id %d in sub group %d in group %d\n",
|
|
operation_names(operation), TypeManager<Ty>::name(),
|
|
i, j, k);
|
|
return TEST_FAIL;
|
|
}
|
|
}
|
|
}
|
|
|
|
x += nw;
|
|
y += nw;
|
|
m += 4 * nw;
|
|
}
|
|
log_info(" sub_group_clustered_reduce_%s(%s, %d bytes) ... passed\n",
|
|
operation_names(operation), TypeManager<Ty>::name(),
|
|
sizeof(Ty));
|
|
return TEST_PASS;
|
|
}
|
|
};
|
|
|
|
template <typename T>
|
|
int run_cluster_red_add_max_min_mul_for_type(RunTestForType rft)
|
|
{
|
|
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::add_>>(
|
|
"test_redadd_clustered", redadd_clustered_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::max_>>(
|
|
"test_redmax_clustered", redmax_clustered_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::min_>>(
|
|
"test_redmin_clustered", redmin_clustered_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::mul_>>(
|
|
"test_redmul_clustered", redmul_clustered_source);
|
|
return error;
|
|
}
|
|
template <typename T> int run_cluster_and_or_xor_for_type(RunTestForType rft)
|
|
{
|
|
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::and_>>(
|
|
"test_redand_clustered", redand_clustered_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::or_>>(
|
|
"test_redor_clustered", redor_clustered_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::xor_>>(
|
|
"test_redxor_clustered", redxor_clustered_source);
|
|
return error;
|
|
}
|
|
template <typename T>
|
|
int run_cluster_logical_and_or_xor_for_type(RunTestForType rft)
|
|
{
|
|
int error = rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_and>>(
|
|
"test_redand_clustered_logical", redand_clustered_logical_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_or>>(
|
|
"test_redor_clustered_logical", redor_clustered_logical_source);
|
|
error |= rft.run_impl<T, RED_CLU<T, ArithmeticOp::logical_xor>>(
|
|
"test_redxor_clustered_logical", redxor_clustered_logical_source);
|
|
|
|
return error;
|
|
}
|
|
}
|
|
|
|
int test_subgroup_functions_clustered_reduce(cl_device_id device,
|
|
cl_context context,
|
|
cl_command_queue queue,
|
|
int num_elements)
|
|
{
|
|
std::vector<std::string> required_extensions = {
|
|
"cl_khr_subgroup_clustered_reduce"
|
|
};
|
|
constexpr size_t global_work_size = 2000;
|
|
constexpr size_t local_work_size = 200;
|
|
WorkGroupParams test_params(global_work_size, local_work_size,
|
|
required_extensions);
|
|
RunTestForType rft(device, context, queue, num_elements, test_params);
|
|
|
|
int error = run_cluster_red_add_max_min_mul_for_type<cl_int>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_uint>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_long>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_ulong>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_short>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_ushort>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_char>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_uchar>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_float>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<cl_double>(rft);
|
|
error |= run_cluster_red_add_max_min_mul_for_type<subgroups::cl_half>(rft);
|
|
|
|
error |= run_cluster_and_or_xor_for_type<cl_int>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_uint>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_long>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_ulong>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_short>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_ushort>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_char>(rft);
|
|
error |= run_cluster_and_or_xor_for_type<cl_uchar>(rft);
|
|
|
|
error |= run_cluster_logical_and_or_xor_for_type<cl_int>(rft);
|
|
return error;
|
|
}
|