diff --git a/test/unittest/layers/unittest_layers_addition_cl.cpp b/test/unittest/layers/unittest_layers_addition_cl.cpp index e7feaaaa50..113e7fd690 100644 --- a/test/unittest/layers/unittest_layers_addition_cl.cpp +++ b/test/unittest/layers/unittest_layers_addition_cl.cpp @@ -7,6 +7,7 @@ * @brief Addition Layer Test * @see https://github.com/nnstreamer/nntrainer * @author Yash Singh + * @author Sungsik Kong * @bug No known bugs except for NYI items */ #include @@ -30,25 +31,26 @@ GTEST_PARAMETER_TEST(AdditionGPU, LayerSemanticsGpu, ::testing::Values(semantic_addition_gpu, semantic_addition_multi_gpu)); -auto addition_w32a32 = LayerGoldenTestParamType( +auto addition_w32a32_gpu = LayerGoldenTestParamType( nntrainer::createLayer, {}, "2:3:3:3,2:3:3:3", "added_w32a32.nnlayergolden", LayerGoldenTestParamOptions::DEFAULT, "nchw", "fp32", "fp32"); -auto addition_w32a32_2 = LayerGoldenTestParamType( +auto addition_w32a32_2_gpu = LayerGoldenTestParamType( nntrainer::createLayer, {}, "3:4:3:4,3:4:3:4", "added_w32a32_2.nnlayergolden", LayerGoldenTestParamOptions::DEFAULT, "nchw", "fp32", "fp32"); GTEST_PARAMETER_TEST(AdditionGPU, LayerGoldenTest, - ::testing::Values(addition_w32a32, addition_w32a32_2)); + ::testing::Values(addition_w32a32_gpu, + addition_w32a32_2_gpu)); #ifdef ENABLE_FP16 -auto addition_w16a16 = LayerGoldenTestParamType( +auto addition_w16a16_gpu = LayerGoldenTestParamType( nntrainer::createLayer, {}, "2:3:3:3,2:3:3:3", "added_w16a16.nnlayergolden", LayerGoldenTestParamOptions::DEFAULT, "nchw", "fp16", "fp16"); GTEST_PARAMETER_TEST(Addition16, LayerGoldenTest, - ::testing::Values(addition_w16a16)); + ::testing::Values(addition_w16a16_gpu)); #endif