diff --git a/help_function/help_function.xml b/help_function/help_function.xml index 8ffb210a6..a1ed854c3 100644 --- a/help_function/help_function.xml +++ b/help_function/help_function.xml @@ -83,6 +83,7 @@ + diff --git a/help_function/src/util_group_store_test.cpp b/help_function/src/util_group_store_test.cpp new file mode 100644 index 000000000..4a7feb2e4 --- /dev/null +++ b/help_function/src/util_group_store_test.cpp @@ -0,0 +1,263 @@ +// ====------ util_group_store_test.cpp------------ *- C++ -* ----===// + + +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +// +// ===----------------------------------------------------------------------===// +#include +#include +#include +#include +#include + + +template +bool helper_validation_function(const int *ptr, const char *func_name) { + for (int i = 0; i < 512; ++i) { + if (ptr[i] != i) { + if constexpr (S == dpct::group::store_algorithm::BLOCK_STORE_DIRECT) { + std::cout << func_name << "_blocked" + << " failed\n";} + else{ + std::cout << func_name << "_striped" + << " failed\n"; + } + std::ostream_iterator Iter(std::cout, ", "); + std::copy(ptr, ptr + 512, Iter); + std::cout << std::endl; + return false; + } + } + if constexpr (S == dpct::group::store_algorithm::BLOCK_STORE_DIRECT) { + std::cout << func_name << "_blocked" + << " passed\n";} + else{ + std::cout << func_name << "_striped" + << " passed\n"; + } + return true; +} + +bool subgroup_helper_validation_function(const int *ptr, const uint32_t *sg_sz, + const char *func_name) { + + int expected[512]; + int num_threads = 128; + int items_per_thread = 4; + uint32_t sg_sz_val = *sg_sz; + for (int i = 0; i < num_threads; ++i) { + for (int j = 0; j < items_per_thread; ++j) { + expected[items_per_thread * i + j] = + (i / sg_sz_val) * sg_sz_val * items_per_thread + sg_sz_val * j + + i % sg_sz_val; + } + } + + for (int i = 0; i < 512; ++i) { + if (ptr[i] != i) { + std::cout << " failed\n"; + std::ostream_iterator Iter(std::cout, ", "); + std::copy(ptr, ptr + 512, Iter); + std::cout << std::endl; + return false; + } + } + + std::cout << func_name << " pass\n"; + return true; +} + +template < dpct::group::store_algorithm S> bool test_group_store() { + // Tests dpct::group::workgroup_store using the specified store algorithm + sycl::queue q(dpct::get_default_queue()); + int data[512]; + int num_threads = 128; + int items_per_thread = 4; + if constexpr(S == dpct::group::store_algorithm::BLOCK_STORE_DIRECT) { + for (int i = 0; i < 512; i++) {data[i] = i;} + } + else{ + for (int i = 0; i < num_threads; ++i) { + for (int j = 0; j < items_per_thread; ++j) { + data[i * items_per_thread + j] = j * num_threads + i; + } + + } + } + + int data_out[512]; + sycl::buffer buffer(data, 512); + sycl::buffer buffer_out(data_out, 512); + + q.submit([&](sycl::handler &h) { + using group_store = + dpct::group::workgroup_store<4, S, int, int *, sycl::nd_item<3>>; + size_t temp_storage_size = group_store::get_local_memory_size(128); + sycl::local_accessor tacc(sycl::range<1>(temp_storage_size), h); + sycl::accessor dacc_read(buffer, h, sycl::read_only); + sycl::accessor dacc_write(buffer_out, h, sycl::read_write); + + h.parallel_for( + sycl::nd_range<3>(sycl::range<3>(1, 1, 128), sycl::range<3>(1, 1, 128)), + [=](sycl::nd_item<3> item) { + int thread_data[4]; + int global_index = + item.get_group(2) * item.get_local_range().get(2) + + item.get_local_id(2); + #pragma unroll + for (int i = 0; i < 4; ++i) { + thread_data[i] = dacc_read[global_index * 4 + i]; + } + + auto *d_w = + dacc_write.get_multi_ptr().get(); + auto *tmp = tacc.get_multi_ptr().get(); + + // Store thread_data of each work item from blocked arrangement + group_store(tmp).store(item, d_w, thread_data); + + }); + }); + q.wait_and_throw(); + + sycl::host_accessor data_accessor(buffer_out, sycl::read_write); + const int *ptr = data_accessor.get_multi_ptr(); + return helper_validation_function(ptr, "test_group_store"); +} + + +bool test_store_subgroup_striped_standalone() { + // Tests dpct::group::store_subgroup_striped as standalone method + + sycl::queue q(dpct::get_default_queue()); + int data[512]; + for(int i=0;i<512;i++){data[i]=i;} + sycl::buffer buffer(data, 512); + sycl::buffer sg_sz_buf{sycl::range<1>(1)}; + int data_out[512]; + sycl::buffer buffer_out(data_out, 512); + + q.submit([&](sycl::handler &h) { + sycl::accessor dacc_read(buffer, h, sycl::read_only); + sycl::accessor dacc_write(buffer_out, h, sycl::read_write); + sycl::accessor sg_sz_dacc(sg_sz_buf, h, sycl::read_write); + + h.parallel_for( + sycl::nd_range<3>(sycl::range<3>(1, 1, 128), sycl::range<3>(1, 1, 128)), + [=](sycl::nd_item<3> item) { + int thread_data[4]; + int global_index = + item.get_group(2) * item.get_local_range().get(2) + + item.get_local_id(2); + #pragma unroll + for (int i = 0; i < 4; ++i) { + thread_data[i] = dacc_read[global_index * 4 + i]; + } + auto *d_w = + dacc_write.get_multi_ptr().get(); + auto *sg_sz_acc = + sg_sz_dacc.get_multi_ptr().get(); + size_t gid = item.get_global_linear_id(); + if (gid == 0) { + sg_sz_acc[0] = item.get_sub_group().get_local_linear_range(); + } + dpct::group::store_subgroup_striped<4, int>(item, d_w, thread_data); + // reapply global mapping + global_index = + (item.get_group(2) * item.get_local_range().get(2)) + + item.get_local_id(2); // Each thread_data has 4 elements + #pragma unroll + for (int i = 0; i < 4; ++i) { + dacc_write[global_index * 4 + i] = thread_data[i]; + } + + }); + }); + q.wait_and_throw(); + + sycl::host_accessor data_accessor(buffer_out, sycl::read_only); + const int *ptr = data_accessor.get_multi_ptr(); + sycl::host_accessor data_accessor_sg(sg_sz_buf, sycl::read_only); + const uint32_t *ptr_sg = + data_accessor_sg.get_multi_ptr(); + return subgroup_helper_validation_function( + ptr, ptr_sg, "test_subgroup_striped_standalone"); +} + + + +template bool test_group_store_standalone() { + // Tests standalone methods for group store using the specified store algorithm + sycl::queue q(dpct::get_default_queue()); + int data[512]; + int num_threads = 128; + int items_per_thread = 4; + if constexpr(S == dpct::group::store_algorithm::BLOCK_STORE_DIRECT) { + for (int i = 0; i < 512; i++) {data[i] = i;} + } + else{ + for (int i = 0; i < num_threads; ++i) { + for (int j = 0; j < items_per_thread; ++j) { + data[i * items_per_thread + j] = j * num_threads + i; + } + + } + } + + std::cout< buffer(data, 512); + sycl::buffer buffer_out(data_out, 512); + + + q.submit([&](sycl::handler &h) { + sycl::accessor dacc_read(buffer, h, sycl::read_only); + sycl::accessor dacc_write(buffer_out, h, sycl::read_write); + h.parallel_for( + sycl::nd_range<3>(sycl::range<3>(1, 1, 128), sycl::range<3>(1, 1, 128)), + [=](sycl::nd_item<3> item) { + int thread_data[4]; + int global_index = + item.get_group(2) * item.get_local_range().get(2) + + item.get_local_id(2); + #pragma unroll + for (int i = 0; i < 4; ++i) { + thread_data[i] = dacc_read[global_index * 4 + i]; + } + + auto *d_w = + dacc_write.get_multi_ptr().get(); + // Store thread_data of each work item from blocked arrangement + if (S == dpct::group::store_algorithm::BLOCK_STORE_DIRECT) { + dpct::group::store_blocked<4, int>(item, d_w, thread_data); + } else { + dpct::group::store_striped<4, int>(item, d_w, thread_data); + } + + }); + }); + q.wait_and_throw(); + + sycl::host_accessor data_accessor(buffer_out, sycl::read_write); + const int *ptr = data_accessor.get_multi_ptr(); + return helper_validation_function(ptr, "test_group_load_store"); +} + +int main() { + + return !( + // Calls test_group_load with blocked and striped strategies , should pass + // both results. + test_group_store() && + test_group_store() && + // Calls test_load_subgroup_striped_standalone and should pass + test_store_subgroup_striped_standalone() && + // Calls test_group_load_standalone with blocked and striped strategies as + // free functions, should pass both results. + test_group_store_standalone() && + test_group_store_standalone()); +}