From ef2581d5c623f78de5913f96fac3de8fd3a5b91f Mon Sep 17 00:00:00 2001 From: Nikita Kudriavtsev Date: Thu, 10 Sep 2020 12:56:21 +0300 Subject: [PATCH] [IE Myriad][IE Tests] Activation layer's constants parametrization. (#2071) CI passed: https://gitlab-icv.inn.intel.com/inference-engine/product-configs/merge_requests/870 --- .../single_layer_tests/activation.cpp | 70 +++++++++---------- .../single_layer_tests/comparison.cpp | 2 +- .../single_layer_tests/squeeze_unsqueeze.cpp | 2 +- .../single_layer_tests/activation.cpp | 20 +++--- .../single_layer_tests/activation.cpp | 62 ++++++++-------- .../single_layer_tests/comparison.cpp | 2 +- .../single_layer_tests/activation.cpp | 24 +++---- .../include/single_layer_tests/activation.hpp | 10 ++- .../src/single_layer_tests/activation.cpp | 40 +++++++---- .../common_test_utils/common_utils.hpp | 14 ++-- .../include/ngraph_functions/builders.hpp | 3 +- .../tests/ngraph_functions/src/activation.cpp | 21 +++--- 12 files changed, 144 insertions(+), 126 deletions(-) diff --git a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/activation.cpp b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/activation.cpp index 0088bb90e79..8450a3ed328 100644 --- a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/activation.cpp +++ b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/activation.cpp @@ -22,39 +22,39 @@ const std::vector netPrecisions = { InferenceEngine::Precision::FP16 }; -const std::vector activationTypes = { - Sigmoid, - Tanh, - Relu, - Exp, - Log, - Sign, - Abs, - Clamp, - Negative, - Acos, - Asin, - Atan, - Cos, - Cosh, - Floor, - Sin, - Sinh, - Sqrt, - Tan, - Elu, - Erf, - HardSigmoid, - Selu, - Ceiling, - Mish, - HSwish, - SoftPlus +const std::map>> activationTypes = { + {Sigmoid, {}}, + {Tanh, {}}, + {Relu, {}}, + {Exp, {}}, + {Log, {}}, + {Sign, {}}, + {Abs, {}}, + {Clamp, {{-2.0f, 2.0f}}}, + {Negative, {}}, + {Acos, {}}, + {Asin, {}}, + {Atan, {}}, + {Cos, {}}, + {Cosh, {}}, + {Floor, {}}, + {Sin, {}}, + {Sinh, {}}, + {Sqrt, {}}, + {Tan, {}}, + {Elu, {{0.1f}}}, + {Erf, {}}, + {HardSigmoid, {{0.2f, 0.5f}}}, + {Selu, {{1.6732f, 1.0507f}}}, + {Ceiling, {}}, + {Mish, {}}, + {HSwish, {}}, + {SoftPlus, {}} }; -const std::vector activationParamTypes = { - PReLu, - LeakyRelu, +const std::map>> activationParamTypes = { + {PReLu, {{-0.01f}}}, + {LeakyRelu, {{0.01f}}} }; std::map, std::vector>> basic = { @@ -68,16 +68,16 @@ std::map, std::vector>> preluBasic = { }; const auto basicCases = ::testing::Combine( - ::testing::ValuesIn(activationTypes), + ::testing::ValuesIn(CommonTestUtils::combineParams(activationTypes)), ::testing::ValuesIn(netPrecisions), - ::testing::ValuesIn(CommonTestUtils::combineShapes(basic)), + ::testing::ValuesIn(CommonTestUtils::combineParams(basic)), ::testing::Values(CommonTestUtils::DEVICE_CPU) ); const auto basicPreluCases = ::testing::Combine( - ::testing::ValuesIn(activationParamTypes), + ::testing::ValuesIn(CommonTestUtils::combineParams(activationParamTypes)), ::testing::ValuesIn(netPrecisions), - ::testing::ValuesIn(CommonTestUtils::combineShapes(preluBasic)), + ::testing::ValuesIn(CommonTestUtils::combineParams(preluBasic)), ::testing::Values(CommonTestUtils::DEVICE_CPU) ); diff --git a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/comparison.cpp b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/comparison.cpp index cc23a910c8f..5caf2dc2c2d 100644 --- a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/comparison.cpp +++ b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/comparison.cpp @@ -45,7 +45,7 @@ std::vector netPrecisions = { std::map additional_config = {}; const auto ComparisonTestParams = ::testing::Combine( - ::testing::ValuesIn(CommonTestUtils::combineShapes(inputShapes)), + ::testing::ValuesIn(CommonTestUtils::combineParams(inputShapes)), ::testing::ValuesIn(inputsPrecisions), ::testing::ValuesIn(comparisonOpTypes), ::testing::ValuesIn(secondInputTypes), diff --git a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/squeeze_unsqueeze.cpp b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/squeeze_unsqueeze.cpp index fae71fa4741..5562d658825 100644 --- a/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/squeeze_unsqueeze.cpp +++ b/inference-engine/tests/functional/plugin/cpu/shared_tests_instances/single_layer_tests/squeeze_unsqueeze.cpp @@ -31,7 +31,7 @@ const std::vector opTypes = { INSTANTIATE_TEST_CASE_P(Basic, SqueezeUnsqueezeLayerTest, ::testing::Combine( - ::testing::ValuesIn(CommonTestUtils::combineShapes(axesVectors)), + ::testing::ValuesIn(CommonTestUtils::combineParams(axesVectors)), ::testing::ValuesIn(opTypes), ::testing::ValuesIn(netPrecisions), ::testing::Values(CommonTestUtils::DEVICE_CPU)), diff --git a/inference-engine/tests/functional/plugin/gna/shared_tests_instances/single_layer_tests/activation.cpp b/inference-engine/tests/functional/plugin/gna/shared_tests_instances/single_layer_tests/activation.cpp index bedf839e4ed..c73ca47b915 100644 --- a/inference-engine/tests/functional/plugin/gna/shared_tests_instances/single_layer_tests/activation.cpp +++ b/inference-engine/tests/functional/plugin/gna/shared_tests_instances/single_layer_tests/activation.cpp @@ -25,14 +25,14 @@ const std::vector netPrecisions = { InferenceEngine::Precision::U8 }; -const std::vector activationTypes = { - Sigmoid, - Tanh, - Relu, - Exp, - Log, - Sign, - Abs +const std::map>> activationTypes = { + {Sigmoid, {}}, + {Tanh, {}}, + {Relu, {}}, + {Exp, {}}, + {Log, {}}, + {Sign, {}}, + {Abs, {}} }; std::map, std::vector>> basic = { @@ -42,9 +42,9 @@ std::map, std::vector>> basic = { }; const auto basicCases = ::testing::Combine( - ::testing::ValuesIn(activationTypes), + ::testing::ValuesIn(CommonTestUtils::combineParams(activationTypes)), ::testing::ValuesIn(netPrecisions), - ::testing::ValuesIn(CommonTestUtils::combineShapes(basic)), + ::testing::ValuesIn(CommonTestUtils::combineParams(basic)), ::testing::Values(CommonTestUtils::DEVICE_GNA) ); diff --git a/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/activation.cpp b/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/activation.cpp index 703ac25fae6..2b4644ce91a 100644 --- a/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/activation.cpp +++ b/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/activation.cpp @@ -16,35 +16,35 @@ const std::vector netPrecisions = { InferenceEngine::Precision::FP16 }; -const std::vector activationTypes = { - Sigmoid, - Tanh, - Relu, - Exp, - Log, - Sign, - Abs, - Gelu, - Clamp, - Negative, - Acos, - Asin, - Atan, - Cos, - Cosh, - Floor, - Sin, - Sinh, - Sqrt, - Tan, - Elu, - Erf, - HardSigmoid, - Selu, - Ceiling, - Mish, - HSwish, - SoftPlus +const std::map>> activationTypes = { + {Sigmoid, {}}, + {Tanh, {}}, + {Relu, {}}, + {Exp, {}}, + {Log, {}}, + {Sign, {}}, + {Abs, {}}, + {Gelu, {}}, + {Clamp, {{-2.0f, 2.0f}}}, + {Negative, {}}, + {Acos, {}}, + {Asin, {}}, + {Atan, {}}, + {Cos, {}}, + {Cosh, {}}, + {Floor, {}}, + {Sin, {}}, + {Sinh, {}}, + {Sqrt, {}}, + {Tan, {}}, + {Elu, {{0.1f}}}, + {Erf, {}}, + {HardSigmoid, {{0.2f, 0.5f}}}, + {Selu, {{1.6732f, 1.0507f}}}, + {Ceiling, {}}, + {Mish, {}}, + {HSwish, {}}, + {SoftPlus, {}} }; std::map, std::vector>> basic = { @@ -53,9 +53,9 @@ std::map, std::vector>> basic = { }; const auto basicCases = ::testing::Combine( - ::testing::ValuesIn(activationTypes), + ::testing::ValuesIn(CommonTestUtils::combineParams(activationTypes)), ::testing::ValuesIn(netPrecisions), - ::testing::ValuesIn(CommonTestUtils::combineShapes(basic)), + ::testing::ValuesIn(CommonTestUtils::combineParams(basic)), ::testing::Values(CommonTestUtils::DEVICE_GPU) ); diff --git a/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/comparison.cpp b/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/comparison.cpp index 2dbfcb1e5ff..7c907bfa1b0 100644 --- a/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/comparison.cpp +++ b/inference-engine/tests/functional/plugin/gpu/shared_tests_instances/single_layer_tests/comparison.cpp @@ -45,7 +45,7 @@ std::vector netPrecisions = { std::map additional_config = {}; const auto ComparisonTestParams = ::testing::Combine( - ::testing::ValuesIn(CommonTestUtils::combineShapes(inputShapes)), + ::testing::ValuesIn(CommonTestUtils::combineParams(inputShapes)), ::testing::ValuesIn(inputsPrecisions), ::testing::ValuesIn(comparisonOpTypes), ::testing::ValuesIn(secondInputTypes), diff --git a/inference-engine/tests/functional/plugin/myriad/shared_tests_instances/single_layer_tests/activation.cpp b/inference-engine/tests/functional/plugin/myriad/shared_tests_instances/single_layer_tests/activation.cpp index 714e50931c1..2b52438e261 100644 --- a/inference-engine/tests/functional/plugin/myriad/shared_tests_instances/single_layer_tests/activation.cpp +++ b/inference-engine/tests/functional/plugin/myriad/shared_tests_instances/single_layer_tests/activation.cpp @@ -15,16 +15,16 @@ const std::vector netPrecisions = { InferenceEngine::Precision::FP16 }; -const std::vector activationTypes = { - Sigmoid, - Tanh, - Relu, - Exp, - Log, - Gelu, - Mish, - SoftPlus, - Swish +const std::map>> activationTypes = { + {Sigmoid, {}}, + {Tanh, {}}, + {Relu, {}}, + {Exp, {}}, + {Log, {}}, + {Gelu, {}}, + {Mish, {}}, + {SoftPlus, {}}, + {Swish, {{0.05f}, {0.8f}, {1.0f}, {15.0f}}} }; std::map, std::vector>> basic = { @@ -33,9 +33,9 @@ std::map, std::vector>> basic = { }; const auto basicCases = ::testing::Combine( - ::testing::ValuesIn(activationTypes), + ::testing::ValuesIn(CommonTestUtils::combineParams(activationTypes)), ::testing::ValuesIn(netPrecisions), - ::testing::ValuesIn(CommonTestUtils::combineShapes(basic)), + ::testing::ValuesIn(CommonTestUtils::combineParams(basic)), ::testing::Values(CommonTestUtils::DEVICE_MYRIAD) ); diff --git a/inference-engine/tests/functional/plugin/shared/include/single_layer_tests/activation.hpp b/inference-engine/tests/functional/plugin/shared/include/single_layer_tests/activation.hpp index caa17fcbfdd..d4adc10568e 100644 --- a/inference-engine/tests/functional/plugin/shared/include/single_layer_tests/activation.hpp +++ b/inference-engine/tests/functional/plugin/shared/include/single_layer_tests/activation.hpp @@ -74,7 +74,7 @@ static std::map activationNames = }; typedef std::tuple< - ngraph::helpers::ActivationTypes, + std::pair>, // Activation type and constant value InferenceEngine::Precision, std::pair, std::vector>, std::string> activationParams; @@ -98,8 +98,12 @@ protected: void SetUp() override; private: - void generateActivationBlob(); - ngraph::ParameterVector createActivationParams(ngraph::element::Type ngPrc, std::vector inShape = {}); + void generateActivationBlob(std::vector constantsValue); + ngraph::ParameterVector createActivationParams( + ngraph::element::Type ngPrc, std::vector inShape = {}); + +private: + std::vector constantsValue; }; } // namespace LayerTestsDefinitions diff --git a/inference-engine/tests/functional/plugin/shared/src/single_layer_tests/activation.cpp b/inference-engine/tests/functional/plugin/shared/src/single_layer_tests/activation.cpp index 2801b1711e9..fc3f575d090 100644 --- a/inference-engine/tests/functional/plugin/shared/src/single_layer_tests/activation.cpp +++ b/inference-engine/tests/functional/plugin/shared/src/single_layer_tests/activation.cpp @@ -23,14 +23,15 @@ std::string ActivationLayerTest::getTestCaseName(const testing::TestParamInfo, std::vector> shapes; std::string targetDevice; - ngraph::helpers::ActivationTypes activationType; - std::tie(activationType, netPrecision, shapes, targetDevice) = obj.param; + std::pair> activationDecl; + std::tie(activationDecl, netPrecision, shapes, targetDevice) = obj.param; std::ostringstream result; const char separator = '_'; - result << activationNames[activationType] << separator; - result << "IS=" << CommonTestUtils::vec2str(shapes.first) << separator;; - result << "AS=" << CommonTestUtils::vec2str(shapes.second) << separator;; + result << activationNames[activationDecl.first] << separator; + result << "IS=" << CommonTestUtils::vec2str(shapes.first) << separator; + result << "AS=" << CommonTestUtils::vec2str(shapes.second) << separator; + result << "ConstantsValue=" << CommonTestUtils::vec2str(activationDecl.second) << separator; result << "netPRC=" << netPrecision.name() << separator; result << "targetDevice=" << targetDevice; return result.str(); @@ -39,10 +40,15 @@ std::string ActivationLayerTest::getTestCaseName(const testing::TestParamInfo, std::vector> shapes; - std::tie(activationType, netPrecision, shapes, targetDevice) = GetParam(); + std::pair> activationDecl; + std::tie(activationDecl, netPrecision, shapes, targetDevice) = GetParam(); + + activationType = activationDecl.first; + auto constantsValue = activationDecl.second; auto ngPrc = FuncTestUtils::PrecisionUtils::convertIE2nGraphPrc(netPrecision); auto params = ngraph::builder::makeParams(ngPrc, {shapes.first}); - auto activation = ngraph::builder::makeActivation(params[0], ngPrc, activationType, shapes.second); + auto activation = ngraph::builder::makeActivation(params[0], ngPrc, activationType, shapes.second, constantsValue); + function = std::make_shared(ngraph::NodeVector{activation}, params); } @@ -127,29 +133,29 @@ ngraph::ParameterVector ActivationParamLayerTest::createActivationParams(ngraph: } } -void ActivationParamLayerTest::generateActivationBlob() { +void ActivationParamLayerTest::generateActivationBlob(std::vector constantsValue) { switch (activationType) { case ngraph::helpers::ActivationTypes::PReLu: { auto blobNegativeSlope = inferRequest.GetBlob("negativeSlope"); - float negativeSlope = -0.01f; + float negativeSlope = constantsValue[0]; blobNegativeSlope = FuncTestUtils::createAndFillBlobWithFloatArray(blobNegativeSlope->getTensorDesc(), &negativeSlope, 1); } case ngraph::helpers::ActivationTypes::LeakyRelu: { auto blobLeakySlope = inferRequest.GetBlob("leakySlope"); - float leakySlope = 0.01f; + float leakySlope = constantsValue[0]; blobLeakySlope = FuncTestUtils::createAndFillBlobWithFloatArray(blobLeakySlope->getTensorDesc(), &leakySlope, 1); } case ngraph::helpers::ActivationTypes::HardSigmoid: { auto blobHardSigmoidAlpha = inferRequest.GetBlob("alpha"); auto blobHardSigmoidBeta = inferRequest.GetBlob("beta"); - float alpha = 0.2f, beta = 0.5f; + float alpha = constantsValue[0], beta = constantsValue[1]; blobHardSigmoidAlpha = FuncTestUtils::createAndFillBlobWithFloatArray(blobHardSigmoidAlpha->getTensorDesc(), &alpha, 1); blobHardSigmoidBeta = FuncTestUtils::createAndFillBlobWithFloatArray(blobHardSigmoidBeta->getTensorDesc(), &beta, 1); } case ngraph::helpers::ActivationTypes::Selu: { auto blobHardSigmoidAlpha = inferRequest.GetBlob("alpha"); auto blobHardSigmoidLambda = inferRequest.GetBlob("lambda"); - float alpha = 1.6732f, lambda = 1.0507f; + float alpha = constantsValue[0], lambda = constantsValue[1]; blobHardSigmoidAlpha = FuncTestUtils::createAndFillBlobWithFloatArray(blobHardSigmoidAlpha->getTensorDesc(), &alpha, 1); blobHardSigmoidLambda = FuncTestUtils::createAndFillBlobWithFloatArray(blobHardSigmoidLambda->getTensorDesc(), &lambda, 1); } @@ -164,7 +170,7 @@ void ActivationParamLayerTest::Infer() { auto blobInput = inferRequest.GetBlob("Input"); blobInput = FuncTestUtils::createAndFillBlobFloat(blobInput->getTensorDesc()); - generateActivationBlob(); + generateActivationBlob(constantsValue); inferRequest.Infer(); } @@ -173,12 +179,18 @@ void ActivationParamLayerTest::Infer() { void ActivationParamLayerTest::SetUp() { InferenceEngine::Precision netPrecision; std::pair, std::vector> shapes; - std::tie(activationType, netPrecision, shapes, targetDevice) = GetParam(); + std::pair> activationDecl; + std::tie(activationDecl, netPrecision, shapes, targetDevice) = GetParam(); + + activationType = activationDecl.first; + constantsValue = activationDecl.second; auto ngPrc = FuncTestUtils::PrecisionUtils::convertIE2nGraphPrc(netPrecision); auto params = ngraph::builder::makeParams(ngPrc, {shapes.first}); auto activationParams = createActivationParams(ngPrc); + params[0]->set_friendly_name("Input"); params.insert(params.end(), activationParams.begin(), activationParams.end()); + auto activation = ngraph::builder::makeActivation(params, ngPrc, activationType); function = std::make_shared(ngraph::NodeVector{activation}, params); } diff --git a/inference-engine/tests/ie_test_utils/common_test_utils/common_utils.hpp b/inference-engine/tests/ie_test_utils/common_test_utils/common_utils.hpp index 5627c617ebb..f5b2dcdfaf0 100644 --- a/inference-engine/tests/ie_test_utils/common_test_utils/common_utils.hpp +++ b/inference-engine/tests/ie_test_utils/common_test_utils/common_utils.hpp @@ -70,13 +70,13 @@ inline InferenceEngine::CNNLayerPtr getLayerByName(const InferenceEngine::CNNNet return getLayerByName(&icnnnetwork, layerName); } -template -std::vector, std::vector>> - combineShapes(const std::map, std::vector>>& inputShapes) { - std::vector, std::vector>> resVec; - for (auto& inputShape : inputShapes) { - for (auto& item : inputShape.second) { - resVec.push_back({inputShape.first, item}); +template +std::vector> combineParams( + const std::map>& keyValueSets) { + std::vector> resVec; + for (auto& keyValues : keyValueSets) { + for (auto& item : keyValues.second) { + resVec.push_back({keyValues.first, item}); } } return resVec; diff --git a/inference-engine/tests/ngraph_functions/include/ngraph_functions/builders.hpp b/inference-engine/tests/ngraph_functions/include/ngraph_functions/builders.hpp index bcddb14d905..22cf1da0077 100644 --- a/inference-engine/tests/ngraph_functions/include/ngraph_functions/builders.hpp +++ b/inference-engine/tests/ngraph_functions/include/ngraph_functions/builders.hpp @@ -191,7 +191,8 @@ std::shared_ptr makeVariadicSplit(const ngraph::Output &in, std::shared_ptr makeActivation(const ngraph::Output &in, const element::Type &type, ngraph::helpers::ActivationTypes activationType, - std::vector inShape = {}); + std::vector inShape = {}, + std::vector constantsValue = {}); std::shared_ptr makeActivation(const ngraph::ParameterVector ¶meters, const element::Type &type, diff --git a/inference-engine/tests/ngraph_functions/src/activation.cpp b/inference-engine/tests/ngraph_functions/src/activation.cpp index 34e3a7c864a..c09df345184 100644 --- a/inference-engine/tests/ngraph_functions/src/activation.cpp +++ b/inference-engine/tests/ngraph_functions/src/activation.cpp @@ -15,7 +15,8 @@ namespace builder { std::shared_ptr makeActivation(const ngraph::Output &in, const element::Type &type, ngraph::helpers::ActivationTypes activationType, - std::vector inShape) { + std::vector inShape, + std::vector constantsValue) { switch (activationType) { case ngraph::helpers::ActivationTypes::Sigmoid: return std::make_shared(in); @@ -27,7 +28,7 @@ std::shared_ptr makeActivation(const ngraph::Output &in, auto leaky_slope = std::make_shared( ngraph::element::f32, inShape, - std::vector{0.01f}); + constantsValue); return std::make_shared(in, leaky_slope); } case ngraph::helpers::ActivationTypes::Exp: @@ -41,7 +42,7 @@ std::shared_ptr makeActivation(const ngraph::Output &in, case ngraph::helpers::ActivationTypes::Gelu: return std::make_shared(in); case ngraph::helpers::ActivationTypes::Clamp: - return std::make_shared(in, -2.0, 2.0); + return std::make_shared(in, constantsValue[0], constantsValue[1]); case ngraph::helpers::ActivationTypes::Negative: return std::make_shared(in); case ngraph::helpers::ActivationTypes::Acos: @@ -65,21 +66,21 @@ std::shared_ptr makeActivation(const ngraph::Output &in, case ngraph::helpers::ActivationTypes::Tan: return std::make_shared(in); case ngraph::helpers::ActivationTypes::Elu: - return std::make_shared(in, 0.1); + return std::make_shared(in, constantsValue[0]); case ngraph::helpers::ActivationTypes::Erf: return std::make_shared(in); case ngraph::helpers::ActivationTypes::HardSigmoid: { auto hard_sigmoid_alpha = std::make_shared( - type, inShape, 0.2f); + type, inShape, constantsValue[0]); auto hard_sigmoid_beta = std::make_shared( - type, inShape, 0.5f); + type, inShape, constantsValue[1]); return std::make_shared(in, hard_sigmoid_alpha, hard_sigmoid_beta); } case ngraph::helpers::ActivationTypes::Selu: { auto selu_alpha = std::make_shared( - type, inShape, 1.6732f); + type, inShape, constantsValue[0]); auto selu_lambda = std::make_shared( - type, inShape, 1.0507f); + type, inShape, constantsValue[1]); return std::make_shared(in, selu_alpha, selu_lambda); } case ngraph::helpers::ActivationTypes::Ceiling: @@ -88,7 +89,7 @@ std::shared_ptr makeActivation(const ngraph::Output &in, auto negative_slope = std::make_shared( ngraph::element::f32, inShape, - std::vector{-0.01f}); + constantsValue); return std::make_shared(in, negative_slope); } case ngraph::helpers::ActivationTypes::Mish: @@ -98,7 +99,7 @@ std::shared_ptr makeActivation(const ngraph::Output &in, case ngraph::helpers::ActivationTypes::SoftPlus: return std::make_shared(in); case ngraph::helpers::ActivationTypes::Swish: { - auto beta = std::make_shared(type, inShape, 1.0f); + auto beta = std::make_shared(type, inShape, constantsValue[0]); return std::make_shared(in, beta); } default: