diff --git a/src/plugins/intel_gpu/include/intel_gpu/primitives/scale.hpp b/src/plugins/intel_gpu/include/intel_gpu/primitives/scale.hpp deleted file mode 100644 index 3d519970785..00000000000 --- a/src/plugins/intel_gpu/include/intel_gpu/primitives/scale.hpp +++ /dev/null @@ -1,74 +0,0 @@ -// Copyright (C) 2018-2022 Intel Corporation -// SPDX-License-Identifier: Apache-2.0 -// - -/////////////////////////////////////////////////////////////////////////////////////////////////// -#pragma once -#include "primitive.hpp" -#include - -namespace cldnn { -/// @addtogroup cpp_api C++ API -/// @{ -/// @addtogroup cpp_topology Network Topology -/// @{ -/// @addtogroup cpp_primitives Primitives -/// @{ - -/// @brief Performs elementwise product of input and scale_input. -/// @details Scale input dimension should be equal to input dimension or be 1 if it is not there.
-/// Input size : 2x3x4x5(BFYX)
-/// Possible scale inputs sizes :
-/// 2x3x4x5 - works the same as(axis == 0 == -4) in caffe
-/// 1x3x4x5 - works the same as(axis == 1 == -3) in caffe
-/// 1x1x4x5 - works the same as(axis == 2 == -2) in caffe
-/// 1x1x1x5 - works the same as(axis == 3 == -1) in caffe
-/// 1x1x1x1 - works the same as empty shape(scalar) in caffe
-/// When scale_input is the same as input, the behavior is the same as @ref eltwise with product operation.
-/// Performs scale over feature when the scale feature size is equal to input feature size.
-/// Performs scale over feature in batch when the scale feature and scale batch sizes are equal to input feature and input batch sizes.
-/// Optionally it can also add provided biases by providing bias data.
-struct scale : public primitive_base { - CLDNN_DECLARE_PRIMITIVE(scale) - - /// @brief Constructs scale primitive without adding bias. - /// @param id This primitive id. - /// @param input Input primitive id. - /// @param scale_input Scale input primitive id with values needed for product computation. - scale(const primitive_id& id, - const primitive_id& input, - const primitive_id& scale_input, // should be bfyx or yxfb, where each dimension can be 1, if all dimensions - // are 1 then this is scalar - const optional_data_type& output_dt = {}, - const padding& output_padding = padding()) - : primitive_base(id, {input, scale_input}, output_padding, output_dt), bias("") {} - - /// @brief Constructs scale primitive with optional adding bias. - /// @param id This primitive id. - /// @param input Input primitive id. - /// @param scale_input Scale input primitive id with values needed for product computation. - /// @param bias Primitive id containing bias data. - scale(const primitive_id& id, - const primitive_id& input, - const primitive_id& scale_input, // should be bfyx or yxfb, where each dimension can be 1, if all dimensions - // are 1 then this is scalar - const primitive_id& bias, // should be same size as scale_input - const optional_data_type& output_dt = {}, - const padding& output_padding = padding()) - : primitive_base(id, {input, scale_input}, output_padding, output_dt), bias(bias) {} - - /// @brief Primitive id containing bias data. - primitive_id bias; - -protected: - std::vector> get_dependencies() const override { - if (bias.empty()) - return {}; - else - return {bias}; - } -}; -/// @} -/// @} -/// @} -} // namespace cldnn diff --git a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_buffer_fusing.cpp b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_buffer_fusing.cpp index df44b776585..4102a41f291 100644 --- a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_buffer_fusing.cpp +++ b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_buffer_fusing.cpp @@ -11,7 +11,6 @@ #include "crop_inst.h" #include "eltwise_inst.h" #include "reshape_inst.h" -#include "scale_inst.h" #include "depth_to_space_inst.h" #include "resample_inst.h" #include "loop_inst.h" @@ -166,7 +165,7 @@ bool concat_in_place_optimization::match(concatenation_node& node) { // todo: we need add padding support for all optimized kernels to remove this condition if (!input->is_type() && !input->is_type() && !input->is_type() && !input->is_type() && !input->is_type() && - !input->is_type() && !input->is_type() && !input->is_type() && !input->is_type() && + !input->is_type() && !input->is_type() && !input->is_type() && !input->is_type()) return false; diff --git a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_primitive_fusing.cpp b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_primitive_fusing.cpp index 1a5ca190089..1fc3a938227 100644 --- a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_primitive_fusing.cpp +++ b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_primitive_fusing.cpp @@ -25,7 +25,6 @@ #include "permute_inst.h" #include "reshape_inst.h" #include "softmax_inst.h" -#include "scale_inst.h" #include "resample_inst.h" #include "depth_to_space_inst.h" #include "space_to_depth_inst.h" @@ -267,7 +266,7 @@ void prepare_primitive_fusing::fuse_activations(program &p) { !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && - !input.is_type() && !input.is_type() && !input.is_type() && + !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && !input.is_type() && @@ -817,8 +816,6 @@ void prepare_primitive_fusing::fuse_simple_primitives(program &p) { should_fuse |= input_data.is_type() && reduce_supports_fusings(input_data.as()); - should_fuse |= input_data.is_type(); - should_fuse |= input_data.is_type() && eltwise_supports_fusings(input_data.as()); if (!should_fuse) @@ -827,71 +824,6 @@ void prepare_primitive_fusing::fuse_simple_primitives(program &p) { p.fuse_nodes(input_data, activation_node, &fusing_history); }; - auto fuse_scale_f = [&](scale_node& scale_node) { - if (scale_node.get_dependencies().empty()) - CLDNN_ERROR_MESSAGE(scale_node.id(), "scale has invalid count of dependencies"); - - auto& input_data = scale_node.get_dependency(0); - if (input_data.get_users().size() != 1 || input_data.get_dependencies().empty()) - return; - - bool should_fuse = input_data.is_type() && - all_ones(input_data.as().get_primitive()->dilation); - - should_fuse |= input_data.is_type() && conv_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type() && fc_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type() && gemm_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type() && pooling_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type() && mvn_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type() && data_type_traits::is_i8_u8(input_data.get_dependency(0).get_output_layout().data_type); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type() && reduce_supports_fusings(input_data.as()); - - should_fuse |= input_data.is_type(); - - should_fuse |= input_data.is_type() && eltwise_supports_fusings(input_data.as()); - - if (!should_fuse) - return; - - p.fuse_nodes(input_data, scale_node, &fusing_history); - }; - auto fuse_quantize_f = [&](quantize_node& quantize_node) { auto& input_data = quantize_node.get_dependency(0); if (input_data.get_users().size() != 1 || input_data.get_dependencies().empty()) @@ -989,8 +921,6 @@ void prepare_primitive_fusing::fuse_simple_primitives(program &p) { should_fuse |= input_data.is_type() && eltwise_supports_fusings(input_data.as()) && quantize_node.get_scale_shift_opt(); - should_fuse |= input_data.is_type() && quantize_node.get_scale_shift_opt(); - should_fuse |= input_data.is_type() && input_data.as().get_primitive()->dimension == 1 && per_tensor_values; @@ -1034,7 +964,6 @@ void prepare_primitive_fusing::fuse_simple_primitives(program &p) { (parents[i]->is_type()) || (parents[i]->is_type()) || (parents[i]->is_type() && eltwise_supports_fusings(parents[i]->as())) || - (parents[i]->is_type()) || (parents[i]->is_type()) || (parents[i]->is_type()) || (parents[i]->is_type()) || @@ -1205,9 +1134,8 @@ void prepare_primitive_fusing::fuse_simple_primitives(program &p) { p.fuse_nodes(*fused_node, node, &fusing_history); }; - program_helpers::do_for_types(*node, + program_helpers::do_for_types(*node, fuse_activation_f, - fuse_scale_f, fuse_quantize_f, fuse_eltwise_f); } diff --git a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_quantization.cpp b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_quantization.cpp index b69ceb36b29..c6a191ee0c0 100644 --- a/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_quantization.cpp +++ b/src/plugins/intel_gpu/src/graph/graph_optimizer/prepare_quantization.cpp @@ -9,7 +9,6 @@ #include "reshape_inst.h" #include "reorder_inst.h" #include "binary_convolution_inst.h" -#include "scale_inst.h" #include "eltwise_inst.h" #include "data_inst.h" #include "pass_manager.h" diff --git a/src/plugins/intel_gpu/src/graph/impls/ocl/binary_convolution.cpp b/src/plugins/intel_gpu/src/graph/impls/ocl/binary_convolution.cpp index 8f0f019ad6f..2b00db68964 100644 --- a/src/plugins/intel_gpu/src/graph/impls/ocl/binary_convolution.cpp +++ b/src/plugins/intel_gpu/src/graph/impls/ocl/binary_convolution.cpp @@ -2,7 +2,6 @@ // SPDX-License-Identifier: Apache-2.0 // -#include "intel_gpu/primitives/scale.hpp" #include "intel_gpu/primitives/quantize.hpp" #include "binary_convolution_inst.h" #include "primitive_base.hpp" diff --git a/src/plugins/intel_gpu/src/graph/impls/ocl/register.cpp b/src/plugins/intel_gpu/src/graph/impls/ocl/register.cpp index 3194e3d3abf..0dc00265e2a 100644 --- a/src/plugins/intel_gpu/src/graph/impls/ocl/register.cpp +++ b/src/plugins/intel_gpu/src/graph/impls/ocl/register.cpp @@ -66,7 +66,6 @@ void register_implementations() { REGISTER_OCL(roi_align); REGISTER_OCL(roi_pooling); REGISTER_OCL(roll); - REGISTER_OCL(scale); REGISTER_OCL(scatter_update); REGISTER_OCL(scatter_nd_update); REGISTER_OCL(scatter_elements_update); diff --git a/src/plugins/intel_gpu/src/graph/impls/ocl/register.hpp b/src/plugins/intel_gpu/src/graph/impls/ocl/register.hpp index 1069f0f80f3..ec11c4b229d 100644 --- a/src/plugins/intel_gpu/src/graph/impls/ocl/register.hpp +++ b/src/plugins/intel_gpu/src/graph/impls/ocl/register.hpp @@ -61,7 +61,6 @@ #include "intel_gpu/primitives/roi_align.hpp" #include "intel_gpu/primitives/roi_pooling.hpp" #include "intel_gpu/primitives/roll.hpp" -#include "intel_gpu/primitives/scale.hpp" #include "intel_gpu/primitives/scatter_elements_update.hpp" #include "intel_gpu/primitives/scatter_nd_update.hpp" #include "intel_gpu/primitives/scatter_update.hpp" @@ -143,7 +142,6 @@ REGISTER_OCL(reverse_sequence); REGISTER_OCL(roi_align); REGISTER_OCL(roi_pooling); REGISTER_OCL(roll); -REGISTER_OCL(scale); REGISTER_OCL(scatter_update); REGISTER_OCL(scatter_elements_update); REGISTER_OCL(scatter_nd_update); diff --git a/src/plugins/intel_gpu/src/graph/impls/ocl/scale.cpp b/src/plugins/intel_gpu/src/graph/impls/ocl/scale.cpp deleted file mode 100644 index ed9a389c8f5..00000000000 --- a/src/plugins/intel_gpu/src/graph/impls/ocl/scale.cpp +++ /dev/null @@ -1,152 +0,0 @@ -// Copyright (C) 2018-2022 Intel Corporation -// SPDX-License-Identifier: Apache-2.0 -// - -#include "scale_inst.h" -#include "primitive_base.hpp" -#include "impls/implementation_map.hpp" -#include "kernel_selector_helper.h" -#include "eltwise/eltwise_kernel_selector.h" -#include "eltwise/eltwise_kernel_base.h" -#include "intel_gpu/runtime/error_handler.hpp" - -using namespace cldnn; - -namespace cldnn { -namespace ocl { - -struct scale_impl : typed_primitive_impl_ocl { - using parent = typed_primitive_impl_ocl; - using parent::parent; - - std::unique_ptr clone() const override { - return make_unique(*this); - } - - explicit scale_impl(const scale_impl& other) : parent(other), - _has_bias_term(other._has_bias_term) {} - - scale_impl(const scale_node& arg, const kernel_selector::kernel_data& kd) : parent(arg, kd) { - set_node_params(arg); - } - - void set_node_params(const program_node& arg) override { - IE_ASSERT(arg.is_type()); - const auto& node = arg.as(); - _has_bias_term = node.bias_term(); - } - -protected: - kernel_arguments_data get_arguments(typed_primitive_inst& instance, int32_t split) const override { - kernel_arguments_data args = parent::get_arguments(instance, split); - args.inputs = {instance.input_memory_ptr(), instance.scale_memory()}; - args.outputs = {instance.output_memory_ptr()}; - - if (_has_bias_term) { - args.inputs.push_back(instance.bias_memory()); - } - return args; - } - -public: - static primitive_impl* create(const scale_node& arg, const kernel_impl_params& impl_param) { - auto ew_params = get_default_params(impl_param); - auto ew_optional_params = - get_default_optional_params(arg.get_program()); - - ew_params.inputs.push_back(convert_data_tensor(impl_param.input_layouts[1])); - - ew_params.operations.push_back({{kernel_selector::eltwise_params::InputType::Buffer(0), - kernel_selector::eltwise_params::InputType::Buffer(1)}, - kernel_selector::eltwise_mode::MUL}); - - if (arg.bias_term()) { - ew_params.inputs.push_back(convert_data_tensor(impl_param.input_layouts[2])); - ew_params.operations.push_back({{kernel_selector::eltwise_params::InputType::Intermediate(0), - kernel_selector::eltwise_params::InputType::Buffer(2)}, - kernel_selector::eltwise_mode::ADD}); - } - - ew_params.layoutBased = true; - - auto& kernel_selector = kernel_selector::eltwise_kernel_selector::Instance(); - auto best_kernels = kernel_selector.GetBestKernels(ew_params, ew_optional_params); - - CLDNN_ERROR_BOOL(arg.id(), - "Best_kernel.empty()", - best_kernels.empty(), - "Cannot find a proper kernel with this arguments"); - - auto scale = new scale_impl(arg, best_kernels[0]); - - return scale; - } - -private: - bool _has_bias_term = false; -}; - -namespace detail { - -attach_scale_impl::attach_scale_impl() { - implementation_map::add(impl_types::ocl, scale_impl::create, { - std::make_tuple(data_types::f32, format::yxfb), - std::make_tuple(data_types::f16, format::yxfb), - std::make_tuple(data_types::i32, format::yxfb), - std::make_tuple(data_types::f32, format::byxf), - std::make_tuple(data_types::f16, format::byxf), - std::make_tuple(data_types::i32, format::byxf), - std::make_tuple(data_types::f32, format::bfyx), - std::make_tuple(data_types::f16, format::bfyx), - std::make_tuple(data_types::i32, format::bfyx), - std::make_tuple(data_types::u8, format::bfyx), - std::make_tuple(data_types::i8, format::bfyx), - std::make_tuple(data_types::f32, format::bfzyx), - std::make_tuple(data_types::f16, format::bfzyx), - std::make_tuple(data_types::i32, format::bfzyx), - std::make_tuple(data_types::u8, format::bfzyx), - std::make_tuple(data_types::i8, format::bfzyx), - std::make_tuple(data_types::f32, format::bfwzyx), - std::make_tuple(data_types::f16, format::bfwzyx), - std::make_tuple(data_types::i32, format::bfwzyx), - std::make_tuple(data_types::u8, format::bfwzyx), - std::make_tuple(data_types::i8, format::bfwzyx), - std::make_tuple(data_types::f32, format::b_fs_yx_fsv16), - std::make_tuple(data_types::f16, format::b_fs_yx_fsv16), - std::make_tuple(data_types::i32, format::b_fs_yx_fsv16), - std::make_tuple(data_types::u8, format::b_fs_yx_fsv16), - std::make_tuple(data_types::i8, format::b_fs_yx_fsv16), - std::make_tuple(data_types::f32, format::b_fs_zyx_fsv16), - std::make_tuple(data_types::f16, format::b_fs_zyx_fsv16), - std::make_tuple(data_types::i32, format::b_fs_zyx_fsv16), - std::make_tuple(data_types::i8, format::b_fs_zyx_fsv16), - std::make_tuple(data_types::u8, format::b_fs_zyx_fsv16), - std::make_tuple(data_types::f32, format::bs_fs_zyx_bsv16_fsv16), - std::make_tuple(data_types::f16, format::bs_fs_zyx_bsv16_fsv16), - std::make_tuple(data_types::i32, format::bs_fs_zyx_bsv16_fsv16), - std::make_tuple(data_types::f16, format::fs_b_yx_fsv32), - std::make_tuple(data_types::i32, format::fs_b_yx_fsv32), - std::make_tuple(data_types::f32, format::bs_fs_yx_bsv16_fsv16), - std::make_tuple(data_types::f16, format::bs_fs_yx_bsv16_fsv16), - std::make_tuple(data_types::i32, format::bs_fs_yx_bsv16_fsv16), - std::make_tuple(data_types::u8, format::b_fs_yx_fsv4), - std::make_tuple(data_types::i8, format::b_fs_yx_fsv4), - std::make_tuple(data_types::f16, format::b_fs_yx_fsv4), - std::make_tuple(data_types::f32, format::b_fs_yx_fsv4), - std::make_tuple(data_types::i32, format::b_fs_yx_fsv4), - std::make_tuple(data_types::u8, format::b_fs_yx_fsv32), - std::make_tuple(data_types::i8, format::b_fs_yx_fsv32), - std::make_tuple(data_types::f16, format::b_fs_yx_fsv32), - std::make_tuple(data_types::f32, format::b_fs_yx_fsv32), - std::make_tuple(data_types::i32, format::b_fs_yx_fsv32), - std::make_tuple(data_types::u8, format::b_fs_zyx_fsv32), - std::make_tuple(data_types::i8, format::b_fs_zyx_fsv32), - std::make_tuple(data_types::f16, format::b_fs_zyx_fsv32), - std::make_tuple(data_types::f32, format::b_fs_zyx_fsv32), - std::make_tuple(data_types::i32, format::b_fs_zyx_fsv32), - }); -} - -} // namespace detail -} // namespace ocl -} // namespace cldnn diff --git a/src/plugins/intel_gpu/src/graph/include/kernel_selector_helper.h b/src/plugins/intel_gpu/src/graph/include/kernel_selector_helper.h index 7feeefea247..e8c8a817b31 100644 --- a/src/plugins/intel_gpu/src/graph/include/kernel_selector_helper.h +++ b/src/plugins/intel_gpu/src/graph/include/kernel_selector_helper.h @@ -9,7 +9,6 @@ #include "intel_gpu/runtime/tensor.hpp" #include "intel_gpu/runtime/error_handler.hpp" #include "intel_gpu/primitives/eltwise.hpp" -#include "intel_gpu/primitives/scale.hpp" #include "intel_gpu/primitives/quantize.hpp" #include "intel_gpu/primitives/activation.hpp" #include "intel_gpu/primitives/primitive.hpp" diff --git a/src/plugins/intel_gpu/src/graph/include/scale_inst.h b/src/plugins/intel_gpu/src/graph/include/scale_inst.h deleted file mode 100644 index a7223684bea..00000000000 --- a/src/plugins/intel_gpu/src/graph/include/scale_inst.h +++ /dev/null @@ -1,59 +0,0 @@ -// Copyright (C) 2018-2022 Intel Corporation -// SPDX-License-Identifier: Apache-2.0 -// - -/////////////////////////////////////////////////////////////////////////////////////////////////// -#pragma once -#include "intel_gpu/primitives/scale.hpp" -#include "primitive_inst.h" -#include "kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_base.h" - -#include -#include - -namespace cldnn { - -template <> -struct typed_program_node : public typed_program_node_base { -private: - using parent = typed_program_node_base; - -public: - using parent::parent; - - typed_program_node(const std::shared_ptr prim, program& prog) : parent(prim, prog) { - support_padding_all(true); - } - program_node& input() const { return get_dependency(0); } - program_node& scale_in() const { return get_dependency(1); } - program_node& bias() const { return get_dependency(2); } - - bool bias_term() const { return get_primitive()->bias.length() != 0; } - - std::shared_ptr get_fuse_params() const override { - return std::make_shared(); - } -}; - -using scale_node = typed_program_node; - -template <> -class typed_primitive_inst : public typed_primitive_inst_base { - using parent = typed_primitive_inst_base; - -public: - static layout calc_output_layout(scale_node const& node, kernel_impl_params const& impl_param); - static std::string to_string(scale_node const& node); - -public: - typed_primitive_inst(network& network, scale_node const& desc); - - memory::ptr scale_memory() const { return dep_memory_ptr(1); } - memory::ptr bias_memory() const { return dep_memory_ptr(2); } - - bool bias_term() const { return _node.as().bias_term(); } -}; - -using scale_inst = typed_primitive_inst; - -} // namespace cldnn diff --git a/src/plugins/intel_gpu/src/graph/program.cpp b/src/plugins/intel_gpu/src/graph/program.cpp index 897f273f295..7a5e45a77e9 100644 --- a/src/plugins/intel_gpu/src/graph/program.cpp +++ b/src/plugins/intel_gpu/src/graph/program.cpp @@ -31,7 +31,6 @@ #include "reshape_inst.h" #include "quantize_inst.h" #include "activation_inst.h" -#include "scale_inst.h" #include "depth_to_space_inst.h" #include "convolution_inst.h" #include "concatenation_inst.h" @@ -1403,7 +1402,6 @@ void program::set_layout_optimizer_attributes(layout_optimizer& lo) { prim.type() != cldnn::border::type_id() && prim.type() != cldnn::resample::type_id() && prim.type() != cldnn::crop::type_id() && - prim.type() != cldnn::scale::type_id() && prim.type() != cldnn::depth_to_space::type_id() && prim.type() != cldnn::shuffle_channels::type_id() && (prim.type() != cldnn::mvn::type_id() @@ -1443,7 +1441,6 @@ void program::set_layout_optimizer_attributes(layout_optimizer& lo) { prim.type() != cldnn::reshape::type_id() && prim.type() != cldnn::input_layout::type_id() && prim.type() != cldnn::activation::type_id() && - prim.type() != cldnn::scale::type_id() && prim.type() != cldnn::softmax::type_id() && prim.type() != cldnn::fully_connected::type_id() && prim.type() != cldnn::generic_layer::type_id() && diff --git a/src/plugins/intel_gpu/src/graph/scale.cpp b/src/plugins/intel_gpu/src/graph/scale.cpp deleted file mode 100644 index 7691233ec89..00000000000 --- a/src/plugins/intel_gpu/src/graph/scale.cpp +++ /dev/null @@ -1,115 +0,0 @@ -// Copyright (C) 2018-2022 Intel Corporation -// SPDX-License-Identifier: Apache-2.0 -// - -#include "scale_inst.h" -#include "primitive_type_base.h" -#include "intel_gpu/runtime/error_handler.hpp" -#include "json_object.h" -#include - -namespace cldnn { -primitive_type_id scale::type_id() { - static primitive_type_base instance; - return &instance; -} - -layout scale_inst::calc_output_layout(scale_node const& node, kernel_impl_params const& impl_param) { - auto desc = impl_param.desc; - auto result = impl_param.get_non_padded_input_layout(); - auto scale_layout = impl_param.get_non_padded_input_layout(1); - - auto scale_x_size = scale_layout.spatial(0); - auto scale_y_size = scale_layout.spatial(1); - auto scale_z_size = scale_layout.spatial(2); - - auto input_x_size = result.spatial(0); - auto input_y_size = result.spatial(1); - auto input_z_size = result.spatial(2); - - if ((result.data_type == data_types::u8 || result.data_type == data_types::i8 || result.data_type == data_types::i32) && - (scale_layout.data_type == data_types::f32 || - scale_layout.data_type == data_types::f16)) - result.data_type = scale_layout.data_type; - - if (desc->output_data_type) - result.data_type = *desc->output_data_type; - - if (impl_param.has_fused_primitives()) { - result.data_type = impl_param.get_fused_output_layout().data_type; - } - - if (scale_x_size != 1) { - CLDNN_ERROR_NOT_EQUAL(desc->id, "Scale x size", scale_x_size, "input x size", input_x_size, ""); - } - if (scale_y_size != 1) { - CLDNN_ERROR_NOT_EQUAL(desc->id, "Scale y size", scale_y_size, "input y size", input_y_size, ""); - } - if (scale_z_size != 1) { - CLDNN_ERROR_NOT_EQUAL(desc->id, "Scale z size", scale_z_size, "input z size", input_z_size, ""); - } - - return result; -} - -std::string scale_inst::to_string(scale_node const& node) { - auto desc = node.get_primitive(); - auto node_info = node.desc_to_json(); - auto& input = node.input(); - auto& scale_input = node.scale_in(); - - std::stringstream primitive_description; - - json_composite scale_info; - scale_info.add("input", input.id()); - scale_info.add("scale input", scale_input.id()); - - node_info->add("scale info", scale_info); - node_info->dump(primitive_description); - - return primitive_description.str(); -} - -scale_inst::typed_primitive_inst(network& network, scale_node const& node) : parent(network, node) { - auto scale_layout = node.scale_in().get_output_layout(); - auto scale_format = scale_layout.format; - - auto scale_batch_size = scale_layout.batch(); - auto scale_feature_size = scale_layout.feature(); - - auto input_batch_size = node.input().get_output_layout().batch(); - auto input_feature_size = node.input().get_output_layout().feature(); - - if (scale_batch_size != 1) { - CLDNN_ERROR_NOT_EQUAL(node.id(), - "Scale batch size", - scale_batch_size, - "input batch size", - input_batch_size, - ""); - } - - if (scale_feature_size != 1) { - CLDNN_ERROR_NOT_EQUAL(node.id(), - "Scale feature size", - scale_feature_size, - "input feature size", - input_feature_size, - ""); - } - - if (!argument.bias.empty()) { - auto bias_layout = node.bias().get_output_layout(); - auto bias_format = bias_layout.format; - auto bias_raw_sizes = bias_layout.get_tensor().raw; - - CLDNN_ERROR_NOT_PROPER_FORMAT(node.id(), "Scale format", scale_format.value, "bias format", bias_format); - - for (size_t i = 0; i < bias_layout.get_tensor().raw.size(); ++i) { - if (scale_layout.get_tensor().raw[i] != bias_raw_sizes[i]) - CLDNN_ERROR_MESSAGE(node.id(), - "Scale input size do not match bias size! Size index:" + std::to_string(i)); - } - } -} -} // namespace cldnn diff --git a/src/plugins/intel_gpu/src/kernel_selector/common/common_types.h b/src/plugins/intel_gpu/src/kernel_selector/common/common_types.h index 72e19b77909..411fa17c505 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/common/common_types.h +++ b/src/plugins/intel_gpu/src/kernel_selector/common/common_types.h @@ -29,7 +29,6 @@ enum class KernelType { ACTIVATION, SOFT_MAX, ELTWISE, - SCALE, REORDER, RESHAPE, COUNT_NONZERO, diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_opt.h index 494c4f763a6..634f25a61bd 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_opt.h @@ -26,7 +26,6 @@ protected: JitConstants GetJitConstants(const activation_params& params, DispatchData dispatchData) const override; std::vector GetSupportedFusedOps() const override { return {FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION}; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_ref.h index 1f6c3075c19..10f16b1be32 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/activation/activation_kernel_ref.h @@ -23,7 +23,6 @@ public: JitConstants GetJitConstants(const activation_params& params, DispatchData dispatchData) const override; std::vector GetSupportedFusedOps() const override { return {FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION}; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/batch_to_space/batch_to_space_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/batch_to_space/batch_to_space_kernel_ref.h index 95ec6a210ce..c7b746683f1 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/batch_to_space/batch_to_space_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/batch_to_space/batch_to_space_kernel_ref.h @@ -23,7 +23,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1.cpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1.cpp index 12692da8de4..cd69e878543 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1.cpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1.cpp @@ -135,32 +135,6 @@ JitConstants BinaryConvolutionKernel1x1::GetFusedPrimitivesJitConstants(const bi std::string e_mul = "e_mul" + toCodeString(op_id); switch (fused_dep.GetType()) { - case KernelType::SCALE: { - std::string cast_type = (fused_dep.tensors[0].GetDType() == Datatype::F32) ? "as_float2" : "as_half2"; - if (fused_dep.tensors.size() == 1) { - std::string var_name = fused_dep_codegen.GetInputVarName(0); - prepare_data += "\\\n\t" + vec_data_type + " " + var_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += "\\\n\t" + data_type + " " + sc + " = (oc < 16) ? " + - get_shuffle(var_name + ".s0", "oc") + " : " + get_shuffle(var_name + ".s1", "oc") + ";"; - eltwise_fused_ops += "\\\n\tres = res*" + sc + ";"; - } else { - std::string var0_name = fused_dep_codegen.GetInputVarName(0); - std::string var1_name = fused_dep_codegen.GetInputVarName(1); - prepare_data += "\\\n\t" + vec_data_type + " " + var0_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - prepare_data += "\\\n\t" + vec_data_type + " " + var1_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(1), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += "\\\n\t" + data_type + " " + sc +" = (oc < 16) ? " + - get_shuffle(var0_name + ".s0", "oc") + " : " + get_shuffle(var0_name + ".s1", "oc") + ";"; - eltwise_fused_ops += "\\\n\t" + data_type + " " + sh + " = (oc < 16) ? " + - get_shuffle(var1_name + ".s0", "oc") + " : " + get_shuffle(var1_name + ".s1", "oc") + ";"; - eltwise_fused_ops += "\\\n\tres = res*" + sc + " + " + sh + ";"; - } - - break; - } - case KernelType::QUANTIZE: { std::string var_name_in = fused_dep_codegen.GetInputVarName(0); std::string var_name_out = fused_dep_codegen.GetInputVarName(3); diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1_b_fs_yx_fsv16.cpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1_b_fs_yx_fsv16.cpp index 58fb03ba39f..265d900be86 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1_b_fs_yx_fsv16.cpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_1x1_b_fs_yx_fsv16.cpp @@ -139,27 +139,6 @@ JitConstants BinaryConvolutionKernel1x1_b_fs_yx_fsv16::GetFusedPrimitivesJitCons std::string e_mul = "e_mul" + toCodeString(op_id); switch (fused_dep.GetType()) { - case KernelType::SCALE: { - std::string cast_type = (fused_dep.tensors[0].GetDType() == Datatype::F32) ? "as_float" : "as_half"; - if (fused_dep.tensors.size() == 1) { - std::string var_name = fused_dep_codegen.GetInputVarName(0); - prepare_data += "\\\n\t" + vec_data_type + " " + var_name + " = " + cast_type + - get_aligned_load(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += "\\\n\t" + data_type + " " + sc + " = " + get_shuffle(var_name, "oc") + ";"; - eltwise_fused_ops += "\\\n\tres = res*" + var_name + ";"; - } else { - std::string var0_name = fused_dep_codegen.GetInputVarName(0); - std::string var1_name = fused_dep_codegen.GetInputVarName(1); - prepare_data += "\\\n\t" + vec_data_type + " " + var0_name + " = " + cast_type + - get_aligned_load(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - prepare_data += "\\\n\t" + vec_data_type + " " + var1_name + " = " + cast_type + - get_aligned_load(fused_dep_codegen.GetInputPtrName(1), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += "\\\n\tres = res*" + var0_name + " + " + var1_name + ";"; - } - - break; - } - case KernelType::ACTIVATION: { auto p = fused_dep.GetOpParams(); base_activation_params activation = p->param; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_generic.cpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_generic.cpp index 892b0282497..66acbea44b0 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_generic.cpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/binary_convolution/binary_convolution_kernel_generic.cpp @@ -130,31 +130,6 @@ JitConstants BinaryConvolutionKernelGeneric::GetFusedPrimitivesJitConstants(cons std::string e_mul = "e_mul" + toCodeString(op_id); switch (fused_dep.GetType()) { - case KernelType::SCALE: { - std::string cast_type = (fused_dep.tensors[0].GetDType() == Datatype::F32) ? "as_float2" : "as_half2"; - if (fused_dep.tensors.size() == 1) { - std::string var_name = fused_dep_codegen.GetInputVarName(0); - prepare_data += vec_data_type + " " + var_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += data_type + " " + sc + " = (i < 16) ? " + var_name + ".s0" + " : " + var_name + ".s1;"; - eltwise_fused_ops += "res = res*" + sc +";"; - } else { - std::string var0_name = fused_dep_codegen.GetInputVarName(0); - std::string var1_name = fused_dep_codegen.GetInputVarName(1); - prepare_data += vec_data_type + " " + var0_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(0), "f_block*OC_BLOCK_SIZE") + ";"; - prepare_data += vec_data_type + " " + var1_name + " = " + cast_type + - get_aligned_load2(fused_dep_codegen.GetInputPtrName(1), "f_block*OC_BLOCK_SIZE") + ";"; - eltwise_fused_ops += - data_type + " " + sc + " = (i < 16) ? " + var0_name + ".s0" + " : " + var0_name + ".s1;"; - eltwise_fused_ops += - data_type + " " + sh + " = (i < 16) ? " + var1_name + ".s0" + " : " + var1_name + ".s1;"; - eltwise_fused_ops += "res = res*" + sc + " + " + sh + ";"; - } - - break; - } - case KernelType::QUANTIZE: { std::string var_name_in = fused_dep_codegen.GetInputVarName(0); std::string var_name_out = fused_dep_codegen.GetInputVarName(3); diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16.h index 41f603e5bf6..b63f6a3192a 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16.h @@ -36,7 +36,6 @@ protected: // so that it can decide whether to fuse eltwise along with reorder. return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::REORDER }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_1x1.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_1x1.h index 3a1cf5c57b8..97b14e30d37 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_1x1.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_1x1.h @@ -28,7 +28,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } bool Validate(const Params& p, const optional_params& o) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_depthwise.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_depthwise.h index 13a90562939..22f1ddcbdf5 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_depthwise.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_depthwise.h @@ -26,7 +26,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_imad_1x1.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_imad_1x1.h index 827d2be185c..1ec77fa3019 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_imad_1x1.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv16_imad_1x1.h @@ -32,7 +32,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv4_int8.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv4_int8.h index 7a29ce8cba5..fb08ef7c3ed 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv4_int8.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv4_int8.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv_16_32_imad_dw.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv_16_32_imad_dw.hpp index 39e8369f5c6..2b584b504f3 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv_16_32_imad_dw.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_yx_fsv_16_32_imad_dw.hpp @@ -27,7 +27,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16.h index 7271d92c04a..d50c8bc2aa1 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16.h @@ -49,7 +49,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16_imad.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16_imad.h index a827bf38756..24a23df4033 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16_imad.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_b_fs_zyx_fsv16_imad.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_os_iyx_osv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_os_iyx_osv16.h index d3816a737e5..d2d7fbfeeb4 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_os_iyx_osv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_os_iyx_osv16.h @@ -27,7 +27,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_b_fs_yx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_b_fs_yx_fsv16.h index a472f79acf9..442c9ec9ec4 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_b_fs_yx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_b_fs_yx_fsv16.h @@ -32,7 +32,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_bs_fs_yx_bsv16_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_bs_fs_yx_bsv16_fsv16.h index 2fd2cc93674..688925df005 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_bs_fs_yx_bsv16_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_bs_fs_yx_bsv16_fsv16.h @@ -26,7 +26,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_fs_byx_fsv32.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_fs_byx_fsv32.h index 9c10382f4af..664230f4ae0 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_fs_byx_fsv32.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_bfyx_to_fs_byx_fsv32.h @@ -30,7 +30,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32.h index f55c66de3db..0c9c993cfba 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32.h @@ -30,7 +30,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_1x1.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_1x1.h index 1423ba8a8d6..06aacfa9d70 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_1x1.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_1x1.h @@ -30,7 +30,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_depthwise.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_depthwise.h index 058383fa62a..024bf6d6156 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_depthwise.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_fs_byx_fsv32_depthwise.h @@ -30,7 +30,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad.h index f11fc2c74d1..06a68ae2823 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_1x1.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_1x1.h index b34d42bdaa7..611e12365d9 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_1x1.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_1x1.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_dw.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_dw.hpp index 2f2d6d52b2e..c49777c7cc5 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_dw.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_b_fs_yx_fsv4_dw.hpp @@ -32,7 +32,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_1x1.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_1x1.h index 33dcea915ac..0d137164506 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_1x1.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_1x1.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_3x3.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_3x3.h index 732bd3df582..95d4ae873b1 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_3x3.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_imad_bs_fs_yx_bsv16_fsv16_3x3.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32.h index 978ecd345da..19b81929d9b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32.h @@ -37,7 +37,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32_dw.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32_dw.h index 91bdec4b290..942ffd83d78 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32_dw.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_b_fs_yx_fsv32_dw.h @@ -30,7 +30,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv32.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv32.h index 7fbc335396b..4ac93a8098c 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv32.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv32.h @@ -40,7 +40,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv4.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv4.h index b3343c24456..294d54d34b4 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv4.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_mmad_bfyx_to_b_fs_yx_fsv4.h @@ -31,7 +31,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_ref.h index daa5a695e5e..c5697447ff2 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/convolution/convolution_kernel_ref.h @@ -32,7 +32,6 @@ protected: // so that it can decide whether to fuse eltwise along with reorder. return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::REORDER }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16.h index d5190b506c7..c18e15a43e2 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16.h @@ -34,7 +34,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16_dw.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16_dw.h index c3e55c27e61..f3612d1e262 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16_dw.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_b_fs_zyx_fsv16_dw.h @@ -51,7 +51,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_bfyx_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_bfyx_opt.h index 87c0bdcbc34..c54da79388e 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_bfyx_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_bfyx_opt.h @@ -25,7 +25,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_along_f_tile_bfx.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_along_f_tile_bfx.hpp index ecc1d8ea2e2..77b5e953f55 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_along_f_tile_bfx.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_along_f_tile_bfx.hpp @@ -28,7 +28,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_ref.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_ref.hpp index 0271a031af7..2a2ca58bc75 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_ref.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_imad_ref.hpp @@ -27,7 +27,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_ref.h index 96854907f75..0401f6d5e0a 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/deconvolution/deconvolution_kernel_ref.h @@ -24,7 +24,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/depth_to_space/depth_to_space_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/depth_to_space/depth_to_space_kernel_ref.h index ee46ad08942..49aefabf3d7 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/depth_to_space/depth_to_space_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/depth_to_space/depth_to_space_kernel_ref.h @@ -24,7 +24,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::REORDER, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv16.h index 455c3d56dbc..3cbfec6f17b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv16.h @@ -19,7 +19,6 @@ public: return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv4.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv4.h index e5ea1478c1a..a70a5bbbc40 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv4.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_b_fs_yx_fsv4.h @@ -29,7 +29,6 @@ public: return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } @@ -43,4 +42,4 @@ protected: const int vec_size = 4; }; -} // namespace kernel_selector \ No newline at end of file +} // namespace kernel_selector diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_base.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_base.h index 6d91c4ecf9b..0198e7a9ed6 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_base.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_base.h @@ -94,10 +94,6 @@ struct eltwise_fuse_params : fuse_params { eltwise_fuse_params(EltwiseMode mode) : fuse_params(KernelType::ELTWISE), mode(mode) {} }; -struct scale_fuse_params : fuse_params { - scale_fuse_params() : fuse_params(KernelType::SCALE) {} -}; - //////////////////////////////////////////////////////////////////////////////////////////////////////////////////////// // EltwiseKernelBase //////////////////////////////////////////////////////////////////////////////////////////////////////////////////////// diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_ref.h index ee45c4d1077..532ccd29be2 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/eltwise/eltwise_kernel_ref.h @@ -19,7 +19,6 @@ public: return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_io_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_io_ref.h index 033f66b56ea..2b8132ebb71 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_io_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_io_ref.h @@ -22,7 +22,6 @@ public: protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_tiled.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_tiled.h index ac03463292f..ffb4b6f72bb 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_tiled.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bf_tiled.h @@ -57,7 +57,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } JitConstants GetJitConstants(const fully_connected_params& params, const DispatchData& dispatchData) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bfyx_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bfyx_ref.h index e5bb4723b83..245de4f6c1b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bfyx_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_bfyx_ref.h @@ -24,7 +24,6 @@ protected: DispatchData SetDefault(const fully_connected_params& params, int autoTuneIndex = -1) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_io_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_io_ref.h index 4100e647d8f..cc7618be824 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_io_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_io_ref.h @@ -22,7 +22,6 @@ public: protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_oi_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_oi_ref.h index 31bc93dfc31..de2271b3d0e 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_oi_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_fb_oi_ref.h @@ -22,7 +22,6 @@ public: protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_imad.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_imad.h index d8561acbc9a..c48e53c9b77 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_imad.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_imad.h @@ -26,7 +26,6 @@ protected: JitConstants GetJitConstants(const fully_connected_params& params, const DispatchData& dispatchData) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_mmad.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_mmad.h index ac31cc54f32..8e84b3fec5b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_mmad.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_mmad.h @@ -34,7 +34,6 @@ protected: DispatchData SetDefault(const fully_connected_params& params, int autoTuneIndex = -1) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_yxfb_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_yxfb_ref.h index 5b757892445..d49f06cb79b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_yxfb_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/fully_connected/fully_connected_kernel_yxfb_ref.h @@ -23,7 +23,6 @@ public: protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_elements_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_elements_kernel_ref.h index fe44bda25a8..31fcb2ea98e 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_elements_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_elements_kernel_ref.h @@ -34,7 +34,6 @@ public: ParamsKey GetSupportedKey() const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_nd_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_nd_kernel_ref.h index d481fcbbcd1..0cbc4bf5596 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_nd_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gather/gather_nd_kernel_ref.h @@ -37,7 +37,6 @@ public: ParamsKey GetSupportedKey() const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8.h index ccf62ee039d..7be3cb2b66b 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8.h @@ -32,7 +32,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } bool Validate(const Params& params, const optional_params& options) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8_slm.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8_slm.h index ad059f435ab..68f8ecc3f94 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8_slm.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_mmad_int8_slm.h @@ -34,7 +34,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } bool Validate(const Params& params, const optional_params& options) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_ref.h index 2d47a7a6fe1..a55a31b0247 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_ref.h @@ -21,7 +21,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } bool Validate(const Params& params, const optional_params& options) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_tiled_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_tiled_opt.h index bb374dc8214..df3bc072b9c 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_tiled_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/gemm/gemm_kernel_tiled_opt.h @@ -29,7 +29,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, FusedOpType::ACTIVATION, - FusedOpType::SCALE, FusedOpType::ELTWISE }; } bool Validate(const Params& params, const optional_params& options) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_b_fs_yx_fsv16_imad.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_b_fs_yx_fsv16_imad.hpp index b95d50bf813..47d08bfc51f 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_b_fs_yx_fsv16_imad.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_b_fs_yx_fsv16_imad.hpp @@ -35,8 +35,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bfyx_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bfyx_opt.h index 14c09028cae..49dfd02a389 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bfyx_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bfyx_opt.h @@ -24,8 +24,7 @@ private: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } DispatchData SetDefault(const mvn_params& params) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bs_fs_yx_bsv32.hpp b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bs_fs_yx_bsv32.hpp index 9d52cb1dc62..7ca696d2ff8 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bs_fs_yx_bsv32.hpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_bs_fs_yx_bsv32.hpp @@ -34,8 +34,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_ref.h index 75288e65091..4c706227131 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/mvn/mvn_kernel_ref.h @@ -25,8 +25,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } std::string GetKernelName(const mvn_params&) const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/normalize/normalize_kernel_base.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/normalize/normalize_kernel_base.h index 7c7eae112c1..aae14230783 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/normalize/normalize_kernel_base.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/normalize/normalize_kernel_base.h @@ -50,8 +50,7 @@ protected: KernelsData GetCommonKernelsData(const Params& params, const optional_params&) const; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::ACTIVATION, - FusedOpType::SCALE }; + FusedOpType::ACTIVATION }; } bool Validate(const Params& params, const optional_params&) const override; Datatype GetActivationType(const normalize_params& params) const; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_ref.h index 007438849a3..ea78b649522 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_ref.h @@ -29,8 +29,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4.h index cf3e51eafd4..d2b8ebaab2a 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4.h @@ -27,8 +27,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4_fsv.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4_fsv.h index 43b18e1d1f7..e90eb2213b9 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4_fsv.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/permute/permute_kernel_tile_8x8_4x4_fsv.h @@ -27,8 +27,7 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::QUANTIZE, - FusedOpType::ELTWISE, - FusedOpType::SCALE + FusedOpType::ELTWISE }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv16.h index 7f44a69d5f3..47388d4b307 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv16.h @@ -23,7 +23,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv4.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv4.h index 6a50f9d9b07..82bec965125 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv4.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_yx_fsv4.h @@ -20,7 +20,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_zyx_fsv16_imad.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_zyx_fsv16_imad.h index fdf0ef84858..a0fdb398821 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_zyx_fsv16_imad.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_b_fs_zyx_fsv16_imad.h @@ -21,7 +21,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bfyx_block_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bfyx_block_opt.h index 68670c865a3..04e78d9511d 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bfyx_block_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bfyx_block_opt.h @@ -23,7 +23,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bs_fs_yx_bsv16_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bs_fs_yx_bsv16_fsv16.h index de9374fa665..7bfd420db5f 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bs_fs_yx_bsv16_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bs_fs_yx_bsv16_fsv16.h @@ -21,7 +21,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION}; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bsv16_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bsv16_fsv16.h index c9538cce459..f3903b58e73 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bsv16_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_bsv16_fsv16.h @@ -27,7 +27,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_opt.h index 020c75891ac..558d9571abb 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_opt.h @@ -23,7 +23,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_padding_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_padding_opt.h index 2a52e069313..f697684d971 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_padding_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_byxf_padding_opt.h @@ -23,7 +23,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_fs_b_yx_fsv32.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_fs_b_yx_fsv32.h index 65bd09afb52..03842e13da8 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_fs_b_yx_fsv32.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_fs_b_yx_fsv32.h @@ -23,7 +23,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_int8_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_int8_ref.h index b1a1f2df44c..b65bd14c0ee 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_int8_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_int8_ref.h @@ -21,7 +21,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } ParamsKey GetSupportedKey() const override; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_ref.h index 3776828f3fc..68116389cb0 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/pooling/pooling_kernel_gpu_ref.h @@ -18,7 +18,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_b_fs_yx_fsv16.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_b_fs_yx_fsv16.h index f67a6d93e5a..e06612a5101 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_b_fs_yx_fsv16.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_b_fs_yx_fsv16.h @@ -19,7 +19,6 @@ public: ParamsKey GetSupportedKey() const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ELTWISE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_ref.h index bce07293fde..4c33d02e135 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/reduce/reduce_kernel_ref.h @@ -19,7 +19,6 @@ public: JitConstants GetJitConstants(const reduce_params& params) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ELTWISE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_opt.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_opt.h index f79b275950e..e0ce20dbd9a 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_opt.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_opt.h @@ -24,7 +24,6 @@ protected: Datatype GetUnitType(const base_params& params) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ELTWISE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_ref.h index b6c82fb792c..422a070a6ca 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/resample/resample_kernel_ref.h @@ -19,7 +19,6 @@ public: JitConstants GetJitConstants(const resample_params& params) const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ELTWISE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_elements_update_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_elements_update_kernel_ref.h index b0eb85745cc..b8c1bffc806 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_elements_update_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_elements_update_kernel_ref.h @@ -33,7 +33,6 @@ public: ParamsKey GetSupportedKey() const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_nd_update_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_nd_update_kernel_ref.h index 11ecd4d3d54..c9e43704fae 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_nd_update_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_nd_update_kernel_ref.h @@ -37,7 +37,6 @@ public: ParamsKey GetSupportedKey() const override; std::vector GetSupportedFusedOps() const override { return { FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION, FusedOpType::ELTWISE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_update_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_update_kernel_ref.h index 64b07f5a2cb..cecbcfc562e 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_update_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/scatter_update/scatter_update_kernel_ref.h @@ -35,7 +35,6 @@ public: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_batch/space_to_batch_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_batch/space_to_batch_kernel_ref.h index 407eb41ae33..ba3cce63a94 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_batch/space_to_batch_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_batch/space_to_batch_kernel_ref.h @@ -23,7 +23,6 @@ protected: return { FusedOpType::ACTIVATION, FusedOpType::ELTWISE, - FusedOpType::SCALE, FusedOpType::QUANTIZE }; } diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_depth/space_to_depth_kernel_ref.h b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_depth/space_to_depth_kernel_ref.h index baa0638f653..2117fc799d7 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_depth/space_to_depth_kernel_ref.h +++ b/src/plugins/intel_gpu/src/kernel_selector/core/actual_kernels/space_to_depth/space_to_depth_kernel_ref.h @@ -41,7 +41,6 @@ protected: std::vector GetSupportedFusedOps() const override { return { FusedOpType::ELTWISE, FusedOpType::QUANTIZE, - FusedOpType::SCALE, FusedOpType::ACTIVATION }; } }; diff --git a/src/plugins/intel_gpu/src/kernel_selector/core/common/jitter.cpp b/src/plugins/intel_gpu/src/kernel_selector/core/common/jitter.cpp index 9225dbc780c..5b5b7b5f179 100644 --- a/src/plugins/intel_gpu/src/kernel_selector/core/common/jitter.cpp +++ b/src/plugins/intel_gpu/src/kernel_selector/core/common/jitter.cpp @@ -1485,7 +1485,6 @@ bool FusedOpsCodeGenerator::CanPreloadData(const FusedOpsConfiguration& conf) co std::string FusedOpsCodeGenerator::GetTypeStr() const { switch (desc.GetType()) { case KernelType::ELTWISE: return "eltwise"; - case KernelType::SCALE: return "scale"; case KernelType::QUANTIZE: return "quantize"; case KernelType::ACTIVATION: return "activation"; case KernelType::UNKNOWN: throw std::runtime_error("Invalid type of fused operation. Fused op can't have type UNKNOWN"); @@ -1578,7 +1577,7 @@ JitConstants FusedOpsCodeGenerator::MakeOpJitConstants(const FusedOpsConfigurati const auto& out_type = desc.output_tensor.GetDType(); if (conf.load_type == FusedOpsConfiguration::LoadType::FEATURE_SHUFFLE && - (desc.GetType() == KernelType::SCALE || desc.GetType() == KernelType::QUANTIZE)) { + desc.GetType() == KernelType::QUANTIZE) { is_shuffled = true; } @@ -1627,18 +1626,6 @@ JitConstants FusedOpsCodeGenerator::MakeOpJitConstants(const FusedOpsConfigurati } switch (desc.GetType()) { - case KernelType::SCALE: { - auto tmp_var = out_var + "_tmp"; - if (desc.tensors.size() > 1) { - op_decls += "\\\n\t" + GetType(get_acc_t(), vec_size) + " " + tmp_var + " = " - + input_vars[0] + " * " + input_vars[1] + " + " + input_vars[2] + ";"; - } else { - op_decls += "\\\n\t" + GetType(get_acc_t(), vec_size) + " " + tmp_var + " = " - + input_vars[0] + " * " + input_vars[1] + ";"; - } - op_decls += "\\\n\t" + GetOutputType(vec_size) + " " + out_var + " = " + ConvertToOutputType(tmp_var, vec_size) + ";"; - break; - } case KernelType::ELTWISE: { auto p = desc.GetOpParams(); if (!p) diff --git a/src/plugins/intel_gpu/tests/fusions/batch_to_space_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/batch_to_space_fusion_test.cpp index d515ff14885..2f3c86771c2 100644 --- a/src/plugins/intel_gpu/tests/fusions/batch_to_space_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/batch_to_space_fusion_test.cpp @@ -101,7 +101,7 @@ TEST_P(batch_to_space_scale_act_eltwise_quantize_u8, basic) { input_layout("input", get_input_layout(p)), batch_to_space("batch_to_space", "input", p.block_shape, p.crops_begin, p.crops_end, p.output_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "batch_to_space", "scale1_data"), + eltwise("scale1", { "batch_to_space", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), @@ -135,7 +135,7 @@ TEST_P(batch_to_space_scale_act_eltw, basic) { input_layout("input", get_input_layout(p)), batch_to_space("batch_to_space", "input", p.block_shape, p.crops_begin, p.crops_end, p.output_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "batch_to_space", "scale1_data"), + eltwise("scale1", { "batch_to_space", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), diff --git a/src/plugins/intel_gpu/tests/fusions/binary_convolution_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/binary_convolution_fusion_test.cpp index 524d2b21521..1c613365ce0 100644 --- a/src/plugins/intel_gpu/tests/fusions/binary_convolution_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/binary_convolution_fusion_test.cpp @@ -106,7 +106,7 @@ TEST_P(conv_bin_scale_activation, basic) { data("weights", get_mem(get_weights_layout(p), -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f/p.kernel.count())), binary_convolution("bin_conv_prim", "input", { "weights" }, p.stride, p.pad, p.dilation, p.out_shape, p.groups), - scale("scale", "bin_conv_prim", "scale_data"), + eltwise("scale", { "bin_conv_prim", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::relu), reorder("reorder_bfyx", "activation", p.default_format, data_types::f32) ); @@ -179,7 +179,7 @@ TEST_P(conv_bin_scale_conv_dw, dw_kernel_3x3_stride2) { data("weights_dw", get_mem(dw_weights_layout, -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1e-1f)), binary_convolution("bin_conv_prim", "input", { "weights" }, p.stride, p.pad, p.dilation, p.out_shape, p.groups), - scale("scale", "bin_conv_prim", "scale_data"), + eltwise("scale", { "bin_conv_prim", "scale_data" }, eltwise_mode::prod, p.default_type), convolution("conv_dw", "scale", { "weights_dw" }, p.out_shape.feature[0], dw_stride, dw_pad, dw_dilation), reorder("reorder_bfyx", "conv_dw", p.default_format, data_types::f32) ); @@ -202,7 +202,7 @@ TEST_P(conv_bin_scale_conv_dw, dw_kernel_3x3_stride1) { data("weights_dw", get_mem(dw_weights_layout, -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1e-1f)), binary_convolution("bin_conv_prim", "input", { "weights" }, p.stride, p.pad, p.dilation, p.out_shape, p.groups), - scale("scale", "bin_conv_prim", "scale_data"), + eltwise("scale", { "bin_conv_prim", "scale_data" }, eltwise_mode::prod, p.default_type), convolution("conv_dw", "scale", { "weights_dw" }, p.out_shape.feature[0], dw_stride, dw_pad, dw_dilation), reorder("reorder_bfyx", "conv_dw", p.default_format, data_types::f32) ); @@ -232,7 +232,7 @@ TEST_P(conv_bin_scale_conv_dw_prelu, dw_kernel_3x3_stride2) { data("weights_dw", get_mem(dw_weights_layout, -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1e-1f)), binary_convolution("bin_conv_prim", "input", { "weights" }, p.stride, p.pad, p.dilation, p.out_shape, p.groups), - scale("scale", "bin_conv_prim", "scale_data"), + eltwise("scale", { "bin_conv_prim", "scale_data" }, eltwise_mode::prod, p.default_type), convolution("conv_dw", "scale", { "weights_dw" }, p.out_shape.feature[0], dw_stride, dw_pad, dw_dilation), data("slope_data", get_mem(get_per_channel_layout(p))), activation("activation", "conv_dw", "slope_data", activation_func::relu_negative_slope), @@ -258,7 +258,7 @@ TEST_P(conv_bin_scale_conv_dw_prelu, dw_kernel_3x3_stride1) { data("weights_dw", get_mem(dw_weights_layout, -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1e-1f)), binary_convolution("bin_conv_prim", "input", { "weights" }, p.stride, p.pad, p.dilation, p.out_shape, p.groups), - scale("scale", "bin_conv_prim", "scale_data"), + eltwise("scale", { "bin_conv_prim", "scale_data" }, eltwise_mode::prod, p.default_type), convolution("conv_dw", "scale", { "weights_dw" }, p.out_shape.feature[0], dw_stride, dw_pad, dw_dilation), data("slope_data", get_mem(get_per_channel_layout(p))), activation("activation", "conv_dw", "slope_data", activation_func::relu_negative_slope), diff --git a/src/plugins/intel_gpu/tests/fusions/depth_to_space_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/depth_to_space_fusion_test.cpp index 07cedc44fa2..8c8fccdc128 100644 --- a/src/plugins/intel_gpu/tests/fusions/depth_to_space_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/depth_to_space_fusion_test.cpp @@ -104,7 +104,7 @@ TEST_P(depth_to_space_scale_act_eltwise_quantize_u8, basic) { input_layout("input", get_input_layout(p)), depth_to_space("depth_to_space", "input", p.block_size, p.mode), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "depth_to_space", "scale1_data"), + eltwise("scale1", { "depth_to_space", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), @@ -139,7 +139,7 @@ TEST_P(depth_to_space_scale_act_eltw, basic) { input_layout("input", get_input_layout(p)), depth_to_space("depth_to_space", "input", p.block_size, p.mode), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "depth_to_space", "scale1_data"), + eltwise("scale1", { "depth_to_space", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), diff --git a/src/plugins/intel_gpu/tests/fusions/eltwise_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/eltwise_fusion_test.cpp index af8972c19de..29241810ab6 100644 --- a/src/plugins/intel_gpu/tests/fusions/eltwise_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/eltwise_fusion_test.cpp @@ -365,7 +365,7 @@ TEST_P(eltwise_fp32_fused_prims, scale_activation) { input_layout("input2", get_input_layout2(p)), data("scale_data", get_mem(get_per_channel_layout(p), -10, 10)), eltwise("eltwise", { "input", "input2" }, p.mode, p.default_type), - scale("scale", "eltwise", "scale_data"), + eltwise("scale", { "eltwise", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::abs), reorder("out", "activation", p.default_format, data_types::f32) ); @@ -446,7 +446,7 @@ TEST_P(eltwise_fp32_scale, 6d) { input_layout("input2", get_input_layout2(p)), data("scale_data", get_mem(get_per_channel_layout(p), -10, 10)), eltwise("eltwise", { "input", "input2" }, p.mode, p.default_type), - scale("scale", "eltwise", "scale_data"), + eltwise("scale", { "eltwise", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("out", "scale", p.default_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/fusion_validity_test.cpp b/src/plugins/intel_gpu/tests/fusions/fusion_validity_test.cpp index 96585b7154a..3f0062043c1 100644 --- a/src/plugins/intel_gpu/tests/fusions/fusion_validity_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/fusion_validity_test.cpp @@ -106,7 +106,7 @@ TEST_P(format_mismatch_multiple_fusing, multiple_fused_node) { data("scale_data", get_mem(get_per_channel_layout(p), -10, 10)), data("eltwise_data", get_mem(get_default_layout(p), -10, 10)), resample("resample_prim", "input", p.out_shape, p.in_shape.feature[0], p.type), - scale("scale", "resample_prim", "scale_data"), + eltwise("scale", { "resample_prim", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::abs), eltwise("eltwise", { "activation", "eltwise_data" }, eltwise_mode::sum), reorder("reorder_bfyx", "eltwise", p.default_format, data_types::f32) diff --git a/src/plugins/intel_gpu/tests/fusions/gather_elements_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/gather_elements_fusion_test.cpp index a921365ff46..0c2d459cb93 100644 --- a/src/plugins/intel_gpu/tests/fusions/gather_elements_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/gather_elements_fusion_test.cpp @@ -153,7 +153,7 @@ TEST_P(gather_elements_scale_activation, basic) { data("scale_data", get_mem(get_per_channel_layout(p), -10, 10)), gather_elements("gather_elements_prim", "input", "gather_elements_indices", p.output_format, p.output_shape, p.axis), activation("activation", "gather_elements_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("reorder_bfyx", "scale", p.default_format, data_types::f32) ); @@ -197,7 +197,7 @@ TEST_P(gather_elements_activation_scale_eltwise, basic) { data("eltwise_data", get_mem(get_output_layout(p))), gather_elements("gather_elements_prim", "input", "gather_elements_indices", p.output_format, p.output_shape, p.axis), activation("activation", "gather_elements_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), eltwise("eltwise", { "scale", "eltwise_data" }, eltwise_mode::sum, p.data_type), reorder("reorder_bfyx", "eltwise", p.default_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/gather_nd_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/gather_nd_fusion_test.cpp index 8f2b4f2a6e3..dd533883a6a 100644 --- a/src/plugins/intel_gpu/tests/fusions/gather_nd_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/gather_nd_fusion_test.cpp @@ -187,7 +187,7 @@ TEST_P(gather_nd_activation_scale_eltwise, basic) { data("eltwise_data", get_mem(get_output_layout(p))), gather_nd("gather_nd_prim", "input", "gather_nd_indices", input_rank, p.indices_rank, p.batch_dims), activation("activation", "gather_nd_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), eltwise("eltwise", { "scale", "eltwise_data" }, eltwise_mode::sum, p.data_type), reorder("reorder_bfyx", "eltwise", p.default_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/gemm_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/gemm_fusion_test.cpp index c6c2b3bd2d4..cf6ef9a5706 100644 --- a/src/plugins/intel_gpu/tests/fusions/gemm_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/gemm_fusion_test.cpp @@ -222,7 +222,7 @@ TEST_P(gemm_2in_scale, basic) { input_layout("input1", get_input_layout(p, 1)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f/p.kernel.count())), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), - scale("scale", "gemm_prim", "scale_data"), + eltwise("scale", { "gemm_prim", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("reorder_bfyx", "scale", p.default_format, data_types::f32) ); @@ -237,7 +237,7 @@ TEST_P(gemm_2in_scale, fp16_scale_out) { input_layout("input1", get_input_layout(p, 1)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f/p.kernel.count())), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), - scale("scale", "gemm_prim", "scale_data", optional_data_type{ data_types::f16 }), + eltwise("scale", { "gemm_prim", "scale_data" }, eltwise_mode::prod, data_types::f16), reorder("reorder_bfyx", "scale", p.default_format, data_types::f32) ); @@ -272,7 +272,7 @@ TEST_P(gemm_2in_act_scale_quantize_i8, basic) { data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / p.kernel.count() / 255)), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), activation("activation", "gemm_prim", activation_func::exp), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), quantize("quantize", "scale", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::i8), reorder("reorder_bfyx", "quantize", p.default_format, data_types::f32) ); @@ -308,7 +308,7 @@ TEST_P(gemm_2in_act_scale_quantize_eltwise_i8, basic) { data("eltwise_data", get_mem(get_output_layout(p))), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), activation("activation", "gemm_prim", activation_func::exp), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), quantize("quantize", "scale", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::i8), eltwise("sum", { "quantize", "eltwise_data" }, eltwise_mode::sum, data_types::f32), reorder("reorder_bfyx", "sum", p.default_format, data_types::f32) @@ -334,7 +334,7 @@ TEST_P(gemm_2in_act_scale_eltwise, basic) { data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / p.kernel.count() / 255)), data("eltwise_data", get_mem(get_output_layout(p))), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), - scale("scale", "gemm_prim", "scale_data"), + eltwise("scale", { "gemm_prim", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::negative), eltwise("sum", { "activation", "eltwise_data" }, eltwise_mode::sum, data_types::f32), reorder("reorder_bfyx", "sum", p.default_format, data_types::f32) @@ -355,7 +355,7 @@ TEST_P(gemm_2in_act_scale_eltwise, broadcast_eltwise) { data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / p.kernel.count() / 255)), data("eltwise_data", get_mem(get_single_element_layout(p))), gemm("gemm_prim", { "input0", "input1" }, data_types::f32), - scale("scale", "gemm_prim", "scale_data"), + eltwise("scale", { "gemm_prim", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::negative), eltwise("sum", { "activation", "eltwise_data" }, eltwise_mode::sum, data_types::f32), reorder("reorder_bfyx", "sum", p.default_format, data_types::f32) diff --git a/src/plugins/intel_gpu/tests/fusions/mvn_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/mvn_fusion_test.cpp index 5c143102d2d..25429fe790b 100644 --- a/src/plugins/intel_gpu/tests/fusions/mvn_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/mvn_fusion_test.cpp @@ -138,7 +138,7 @@ TEST_P(mvn_scale_quantize_i8, basic) { input_layout("input", get_input_layout(p)), mvn("mvn", "input", p.normalize_variance, 1e-10f, false, false), data("scale_data", get_mem(get_per_channel_layout(p))), - scale("scale", "mvn", "scale_data"), + eltwise("scale", { "mvn", "scale_data" }, eltwise_mode::prod, p.default_type), data("in_low", get_mem(get_per_channel_layout(p), min_random, 0)), data("in_high", get_mem(get_per_channel_layout(p), 1, max_random)), data("out_low", get_mem(get_single_element_layout(p), -127, 127)), @@ -182,7 +182,7 @@ TEST_P(mvn_scale_activation_eltwise_fp32_quantize_i8, basic) { input_layout("input", get_input_layout(p)), mvn("mvn", "input", p.normalize_variance, 1e-10f, false, false), data("scale_data", get_mem(get_per_channel_layout(p))), - scale("scale", "mvn", "scale_data"), + eltwise("scale", { "mvn", "scale_data" }, eltwise_mode::prod, p.default_type), activation("act", "scale", activation_func::hyperbolic_tan), data("eltw_data", get_mem(layout{ p.input_type, p.default_format, p.elwise_size })), eltwise("eltw", { "act", "eltw_data" }, eltwise_mode::sum, data_types::f32), diff --git a/src/plugins/intel_gpu/tests/fusions/normalize_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/normalize_fusion_test.cpp index b529a74b43a..2fce59694c1 100644 --- a/src/plugins/intel_gpu/tests/fusions/normalize_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/normalize_fusion_test.cpp @@ -105,7 +105,7 @@ TEST_P(normalize_i8_float, basic) { data("weights", get_mem(get_weights_layout(p))), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f/255)), normalize("normalizel2", "input", "weights", p.across_spatial), - scale("scale", "normalizel2", "scale_data"), + eltwise("scale", { "normalizel2", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::abs), reorder("output_reorder", "activation", p.default_format, data_types::f32) ); @@ -115,6 +115,6 @@ TEST_P(normalize_i8_float, basic) { } INSTANTIATE_TEST_SUITE_P(fusings_gpu, normalize_i8_float, ::testing::ValuesIn(std::vector{ - normalize_test_params{ CASE_NORMALIZE_I8_1, false, 2, 4 }, - normalize_test_params{ CASE_NORMALIZE_I8_1, true, 2, 4 }, + normalize_test_params{ CASE_NORMALIZE_I8_1, false, 3, 4 }, + normalize_test_params{ CASE_NORMALIZE_I8_1, true, 3, 4 }, })); diff --git a/src/plugins/intel_gpu/tests/fusions/permute_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/permute_fusion_test.cpp index f38f9ee318d..75844e05f3e 100644 --- a/src/plugins/intel_gpu/tests/fusions/permute_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/permute_fusion_test.cpp @@ -166,7 +166,7 @@ TEST_P(permute_activation_scale_eltwise, basic) { data("eltwise_data", get_mem(layout{ p.data_type, p.input_format, p.out_shape })), data("scale_data", get_mem(get_per_channel_layout(p), 5e-1f)), permute("permute", "input", p.permute_order), - scale("scale", "permute", "scale_data"), + eltwise("scale", { "permute", "scale_data" }, eltwise_mode::prod, p.default_type), activation("actv", "scale", activation_func::relu), eltwise("eltwise", { "actv", "eltwise_data" }, eltwise_mode::sum, p.data_type), reorder("reorder_bfyx", "eltwise", p.default_format, p.default_type) @@ -290,10 +290,10 @@ TEST_P(permute_scale_actv_eltw_scale_actv_quant_i8, basic) { data("eltw_data", get_mem(layout(p.data_type, p.input_format, p.out_shape))), data("scale2_data", get_mem(get_per_channel_layout(p), 1e-1f)), permute("permute", "input", p.permute_order), - scale("scale1", "permute", "scale1_data"), + eltwise("scale1", { "permute", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.data_type), - scale("scale2", "eltw", "scale2_data"), + eltwise("scale2", { "eltw", "scale2_data" }, eltwise_mode::prod, p.default_type), activation("actv2", "scale2", activation_func::relu), quantize("quant", "actv2", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::i8), reorder("out", "quant", p.default_format, p.default_type) @@ -364,10 +364,10 @@ TEST_P(permute_scale_eltwise_actv_scale_actv, basic) { data("scale_data1", get_mem(get_per_channel_layout(p), 1e-1f)), data("scale_data2", get_mem(get_per_channel_layout(p), 1e-1f)), permute("permute", "input", p.permute_order), - scale("scale1", "permute", "scale_data1"), + eltwise("scale1", { "permute", "scale_data1" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), eltwise("eltwise", { "actv1", "eltwise_data" }, eltwise_mode::sum, p.default_type), - scale("scale2", "eltwise", "scale_data2"), + eltwise("scale2", { "eltwise", "scale_data2" }, eltwise_mode::prod, p.default_type), activation("actv2", "scale2", activation_func::relu), reorder("reorder_bfyx", "actv2", p.default_format, p.default_type) ); diff --git a/src/plugins/intel_gpu/tests/fusions/pooling_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/pooling_fusion_test.cpp index 00e86e44b60..87bd7c3be68 100644 --- a/src/plugins/intel_gpu/tests/fusions/pooling_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/pooling_fusion_test.cpp @@ -191,7 +191,7 @@ TEST_P(pooling_f32_scale, basic) { input_layout("input", get_input_layout(p)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 9.0f)), pooling("pooling", "input", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data"), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("output_reorder", "scale", format::bfyx, data_types::f32) ); @@ -211,7 +211,7 @@ TEST_P(pooling_f32_scale, fp16_scale_out) { input_layout("input", get_input_layout(p)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 9.0f)), pooling("pooling", "input", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data", optional_data_type{ data_types::f16 }), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, data_types::f16), reorder("output_reorder", "scale", format::bfyx, data_types::f32) ); @@ -251,7 +251,7 @@ TEST_P(pooling_scale_activation_quantize, basic) { data("out_hi", get_mem(get_single_element_layout(p), 255)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 16.0f)), pooling("pooling", "input", "", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data"), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::relu), quantize("quantize", "activation", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::u8), reorder("output_reorder", "quantize", p.default_format, data_types::f32) @@ -277,7 +277,7 @@ TEST_P(pooling_scale_activation_quantize, i8_output_data_type) { data("out_hi", get_mem(get_single_element_layout(p), -127, 127)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 16.0f)), pooling("pooling", "input", "", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data"), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::relu), quantize("quantize", "activation", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::i8), reorder("output_reorder", "quantize", p.default_format, data_types::f32) @@ -303,7 +303,7 @@ TEST_P(pooling_scale_activation_quantize, per_channel) { data("out_hi", get_mem(get_single_element_layout(p), 255)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 16.0f)), pooling("pooling", "input", "", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data"), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::atan), quantize("quantize", "activation", "in_lo", "in_hi", "out_lo", "out_hi", 255, data_types::u8), reorder("output_reorder", "quantize", p.default_format, data_types::f32) @@ -371,7 +371,7 @@ TEST_P(pooling_scale_activation, basic) { input_layout("input", get_input_layout(p)), data("scale_data", get_mem(get_per_channel_layout(p), 1.0f / 16.0f)), pooling("pooling", "input", "", p.pool_mode, kernel, stride, pad), - scale("scale", "pooling", "scale_data"), + eltwise("scale", { "pooling", "scale_data" }, eltwise_mode::prod, p.default_type), activation("activation", "scale", activation_func::relu), reorder("output_reorder", "activation", p.default_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/scatter_elements_update_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/scatter_elements_update_fusion_test.cpp index 191e6837e3a..850c7757954 100644 --- a/src/plugins/intel_gpu/tests/fusions/scatter_elements_update_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/scatter_elements_update_fusion_test.cpp @@ -134,7 +134,7 @@ TEST_P(scatter_elements_update_scale_activation_eltwise, basic) { data("eltwise_data", get_mem(layout{ p.data_type, p.input_format, p.input_shape })), scatter_elements_update("scatter_elements_update_prim", "input", "scatter_elements_update_indices", "scatter_elements_update_updates", p.axis), activation("activation", "scatter_elements_update_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), eltwise("eltwise", { "scale", "eltwise_data" }, eltwise_mode::sum, p.data_type), reorder("reorder_bfyx", "eltwise", p.default_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/scatter_nd_update_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/scatter_nd_update_fusion_test.cpp index 9b9896645d6..70fc060ab5e 100644 --- a/src/plugins/intel_gpu/tests/fusions/scatter_nd_update_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/scatter_nd_update_fusion_test.cpp @@ -348,7 +348,7 @@ TEST_P(scatter_nd_update_scale_activation_eltwise, basic) { data("eltwise_data", get_mem(layout{ p.data_type, p.input_format, p.input_shape })), scatter_nd_update("scatter_nd_update_prim", "input", "scatter_nd_update_indices", "scatter_nd_update_updates", p.indices_rank), activation("activation", "scatter_nd_update_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), eltwise("eltwise", { "scale", "eltwise_data" }, eltwise_mode::sum, p.data_type), reorder("reorder_bfyx", "eltwise", p.input_format, data_types::f32) ); diff --git a/src/plugins/intel_gpu/tests/fusions/scatter_update_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/scatter_update_fusion_test.cpp index 0a7191fb723..dc577e44f76 100644 --- a/src/plugins/intel_gpu/tests/fusions/scatter_update_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/scatter_update_fusion_test.cpp @@ -148,7 +148,7 @@ TEST_P(scatter_update_scale_activation, basic) { data("scale_data", get_mem(get_per_channel_layout(p), -10, 10)), scatter_update("scatter_update_prim", "input", "scatter_update_indices", "scatter_update_updates", p.axis), activation("activation", "scatter_update_prim", activation_func::abs), - scale("scale", "activation", "scale_data"), + eltwise("scale", { "activation", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("reorder_bfyx", "scale", p.default_format, data_types::f32) ); @@ -157,29 +157,29 @@ TEST_P(scatter_update_scale_activation, basic) { } INSTANTIATE_TEST_SUITE_P(fusings_gpu, scatter_update_scale_activation, ::testing::ValuesIn(std::vector{ - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_1, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_2, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_3, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_4, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_5, 2, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_1, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_2, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_3, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_4, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP32_5, 3, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_1, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_2, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_3, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_4, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_5, 2, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_1, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_2, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_3, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_4, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_FP16_5, 3, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_1, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_2, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_3, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_4, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_5, 2, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_1, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_2, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_3, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_4, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP32_5, 3, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_1, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_2, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_3, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_4, 2, 4 }, - scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_5, 2, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_1, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_2, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_3, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_4, 3, 4 }, + scatter_update_test_params{ CASE_SCATTER_UPDATE_5D_FP16_5, 3, 4 }, })); class scatter_update_scale_activation_eltwise : public ScatterUpdatePrimitiveFusingTest {}; @@ -194,7 +194,7 @@ TEST_P(scatter_update_scale_activation_eltwise, basic) { scatter_update("scatter_update_prim", "input", "scatter_update_indices", "scatter_update_updates", p.axis), activation("activation", "scatter_update_prim", activation_func::abs), eltwise("eltw", { "activation", "eltw_data" }, eltwise_mode::sum, p.default_type), - scale("scale", "eltw", "scale_data"), + eltwise("scale", { "eltw", "scale_data" }, eltwise_mode::prod, p.default_type), reorder("reorder_bfyx", "scale", p.default_format, data_types::f32) ); tolerance = 1e-5f; diff --git a/src/plugins/intel_gpu/tests/fusions/space_to_batch_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/space_to_batch_fusion_test.cpp index 9b997b82734..85fd2283688 100644 --- a/src/plugins/intel_gpu/tests/fusions/space_to_batch_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/space_to_batch_fusion_test.cpp @@ -99,7 +99,7 @@ TEST_P(space_to_batch_scale_act_eltwise_quantize_u8, basic) { input_layout("input", get_input_layout(p)), space_to_batch("space_to_batch", "input", p.block_shape, p.pads_begin, p.pads_end, p.output_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "space_to_batch", "scale1_data"), + eltwise("scale1", { "space_to_batch", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), @@ -134,7 +134,7 @@ TEST_P(space_to_batch_scale_act_eltw, basic) { input_layout("input", get_input_layout(p)), space_to_batch("space_to_batch", "input", p.block_shape, p.pads_begin, p.pads_end, p.output_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "space_to_batch", "scale1_data"), + eltwise("scale1", { "space_to_batch", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), diff --git a/src/plugins/intel_gpu/tests/fusions/space_to_depth_fusion_test.cpp b/src/plugins/intel_gpu/tests/fusions/space_to_depth_fusion_test.cpp index 98d827f4f4d..7d7b3983cfd 100644 --- a/src/plugins/intel_gpu/tests/fusions/space_to_depth_fusion_test.cpp +++ b/src/plugins/intel_gpu/tests/fusions/space_to_depth_fusion_test.cpp @@ -104,7 +104,7 @@ TEST_P(space_to_depth_scale_act_eltwise_quantize_u8, basic) { input_layout("input", get_input_layout(p)), space_to_depth("space_to_depth", "input", p.mode, p.block_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "space_to_depth", "scale1_data"), + eltwise("scale1", { "space_to_depth", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), @@ -139,7 +139,7 @@ TEST_P(space_to_depth_scale_act_eltw, basic) { input_layout("input", get_input_layout(p)), space_to_depth("space_to_depth", "input", p.mode, p.block_size), data("scale1_data", get_mem(get_per_channel_layout(p), -0.125f)), - scale("scale1", "space_to_depth", "scale1_data"), + eltwise("scale1", { "space_to_depth", "scale1_data" }, eltwise_mode::prod, p.default_type), activation("actv1", "scale1", activation_func::relu), data("eltw_data", get_mem(layout(p.default_type, p.input_format, p.output_size))), eltwise("eltw", { "actv1", "eltw_data" }, eltwise_mode::sum, p.default_type), diff --git a/src/plugins/intel_gpu/tests/test_cases/condition_gpu_test.cpp b/src/plugins/intel_gpu/tests/test_cases/condition_gpu_test.cpp index e803159eef4..d5502b88c91 100644 --- a/src/plugins/intel_gpu/tests/test_cases/condition_gpu_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/condition_gpu_test.cpp @@ -11,7 +11,6 @@ #include #include #include -#include #include #include @@ -68,7 +67,7 @@ TEST(DISABLED_condition_gpu, basic_equal_comp) { condition("condi", "input", branch_true, branch_false, "compare", cond_functions::EQUAL) ); topology.add( - scale("output", "condi", "scale_data") + eltwise("output", { "condi", "scale_data" }, eltwise_mode::prod) ); network net(engine, topology, bs); @@ -402,12 +401,12 @@ TEST(DISABLED_condition_gpu, basic_nested_ifs) { topology nested_true; { - nested_true.add(scale("scale_5", "condi_nested", "scale_5_data"), + nested_true.add(eltwise("scale_5", { "condi_nested", "scale_5_data" }, eltwise_mode::prod), data("scale_5_data", scale_5_mem)); } topology nested_false; { - nested_false.add(scale("scale_10", "condi_nested", "scale_10_data"), + nested_false.add(eltwise("scale_10", { "condi_nested", "scale_10_data" }, eltwise_mode::prod), data("scale_10_data", scale_10_mem)); } diff --git a/src/plugins/intel_gpu/tests/test_cases/convolution_gpu_test.cpp b/src/plugins/intel_gpu/tests/test_cases/convolution_gpu_test.cpp index 09959f65d67..5ae84311680 100644 --- a/src/plugins/intel_gpu/tests/test_cases/convolution_gpu_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/convolution_gpu_test.cpp @@ -6078,7 +6078,7 @@ TEST_P(convolution_gpu_block_layout3D, bfzyx_bsv16_fsv16_fp32_fused_ops) set_values(scale_mem, { scalar }); topology.add(data("scalar", scale_mem)); - topology.add(scale("scale", "conv_bsv16_fsv16", "scalar")); + topology.add(eltwise("scale", { "conv_bsv16_fsv16", "scalar" }, eltwise_mode::prod)); topology.add(reorder("reorder_bfzyx", "scale", format::bfzyx, data_types::f32)); @@ -6517,7 +6517,7 @@ TEST_P(convolution_gpu_block_layout, bfyx_bsv16_fsv16_fp32_fused_ops) set_values(scale_mem, { scalar }); topology.add(data("scalar", scale_mem)); - topology.add(scale("scale", "conv_bsv16_fsv16", "scalar")); + topology.add(eltwise("scale", { "conv_bsv16_fsv16", "scalar" }, eltwise_mode::prod)); topology.add(reorder("reorder_bfyx", "scale", format::bfyx, data_types::f32)); @@ -8255,9 +8255,10 @@ public: topo.add(cldnn::data("scale_scale", scale_mem)); topo.add(cldnn::data("scale_shift", shift_mem)); - topo.add(cldnn::scale("scale", "conv", "scale_scale", "scale_shift")); + topo.add(cldnn::eltwise("scale", { "conv", "scale_scale" }, eltwise_mode::prod)); + topo.add(cldnn::eltwise("shift", { "scale", "scale_shift" }, eltwise_mode::sum)); // Work-around since if scale is output it will not be fused - topo.add(cldnn::reorder("scale_wa_reorder", "scale", format::bfyx, this->output_type())); + topo.add(cldnn::reorder("scale_wa_reorder", "shift", format::bfyx, this->output_type())); return topo; } diff --git a/src/plugins/intel_gpu/tests/test_cases/memory_test.cpp b/src/plugins/intel_gpu/tests/test_cases/memory_test.cpp index d81b4b8eda2..ea29747f083 100644 --- a/src/plugins/intel_gpu/tests/test_cases/memory_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/memory_test.cpp @@ -13,7 +13,6 @@ #include #include #include -#include using namespace cldnn; using namespace ::tests; @@ -475,8 +474,8 @@ TEST(memory_pool, non_opt_intermidate_opt_after) { auto reshape = cldnn::reshape("reshape", "concat", reshape_tensor); auto crop1 = cldnn::crop("crop1", "reshape", { 1, 1, 1, 1 }, { 0, 0, 0, 0 }); auto crop2 = cldnn::crop("crop2", "reshape", { 1, 1, 1, 1 }, { 1, 0, 0, 0 }); - auto eltwise1 = cldnn::scale("elt1", "crop1", "scale_mem"); - auto eltwise2 = cldnn::scale("elt2", "crop2", "scale_mem"); + auto eltwise1 = cldnn::eltwise("elt1", { "crop1", "scale_mem" }, eltwise_mode::prod); + auto eltwise2 = cldnn::eltwise("elt2", { "crop2", "scale_mem" }, eltwise_mode::prod); auto topology = cldnn::topology( input, input2, @@ -522,8 +521,8 @@ TEST(memory_pool, add_mem_dep_test) { auto actv2 = cldnn::activation("input_activ2", "input1", activation_func::abs); auto crop1 = cldnn::crop("crop1", "input_activ1", { 1, 1, 2, 2 }, { 0, 0, 0, 0 }); auto crop2 = cldnn::crop("crop2", "input_activ2", { 1, 1, 2, 2 }, { 0, 1, 0, 0 }); - auto eltwise1 = cldnn::scale("elt1", "crop1", "scale_mem"); - auto eltwise2 = cldnn::scale("elt2", "crop2", "scale_mem"); + auto eltwise1 = cldnn::eltwise("elt1", { "crop1", "scale_mem" }, eltwise_mode::prod); + auto eltwise2 = cldnn::eltwise("elt2", { "crop2", "scale_mem" }, eltwise_mode::prod); auto actv3 = cldnn::activation("out3", "elt1", activation_func::abs); auto actv4 = cldnn::activation("out4", "elt2", activation_func::abs); diff --git a/src/plugins/intel_gpu/tests/test_cases/reorder_gpu_test.cpp b/src/plugins/intel_gpu/tests/test_cases/reorder_gpu_test.cpp index f54205eccb6..ffc8be9eba7 100644 --- a/src/plugins/intel_gpu/tests/test_cases/reorder_gpu_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/reorder_gpu_test.cpp @@ -1406,7 +1406,7 @@ TEST(reorder_gpu_opt, remove_redundant_activation_fuse) reorder("r1", "in", format::bfyx, data_types::f32), activation("relu", "r1", activation_func::relu_negative_slope, { 0.01f, 0.0f }), data("scale_data", scale_mem), - scale("output", "relu", "scale_data") + eltwise("output", { "relu", "scale_data" }, eltwise_mode::prod) }; build_options opts; diff --git a/src/plugins/intel_gpu/tests/test_cases/reshape_gpu_test.cpp b/src/plugins/intel_gpu/tests/test_cases/reshape_gpu_test.cpp index ea888654537..eb4a53e82fa 100644 --- a/src/plugins/intel_gpu/tests/test_cases/reshape_gpu_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/reshape_gpu_test.cpp @@ -603,8 +603,9 @@ TEST(reshape_gpu_f32, shrink_chain_partial) { topology.add(reshape("reshape", "relu", tensor(spatial(2, 2)))); topology.add(reorder("reorder", "reshape", format::bfyx, data_types::f32)); topology.add(reshape("reshape1", "reorder", tensor(feature(4)))); - topology.add(scale("scale", "reshape1", "scale_in", "shift_in")); - topology.add(reorder("out_reorder", "scale", format::yxfb, data_types::f32)); + topology.add(eltwise("scale", { "reshape1", "scale_in" }, eltwise_mode::prod)); + topology.add(eltwise("shift", { "scale", "shift_in" }, eltwise_mode::sum)); + topology.add(reorder("out_reorder", "shift", format::yxfb, data_types::f32)); std::vector input_vec = {-1.f, 2.f, -3.f, 4.f}; std::vector out = {5.f, 12.f, 15.f, 32.0f}; @@ -642,8 +643,9 @@ TEST(reshape_gpu_f32, shrink_chain_full) { topology.add(reshape("reshape", "relu", tensor(spatial(2, 2)))); topology.add(reorder("reorder", "reshape", format::bfyx, data_types::f32)); topology.add(reshape("reshape1", "reorder", tensor(feature(4)))); - topology.add(scale("scale", "reshape1", "scale_in", "shift_in")); - topology.add(reorder("out_reorder", "scale", format::yxfb, data_types::f32)); + topology.add(eltwise("scale", { "reshape1", "scale_in" }, eltwise_mode::prod)); + topology.add(eltwise("shift", { "scale", "shift_in" }, eltwise_mode::sum)); + topology.add(reorder("out_reorder", "shift", format::yxfb, data_types::f32)); std::vector input_vec = {-1.f, 2.f, -3.f, 4.f}; std::vector out = {5.f, 12.f, 15.f, 32.0f}; diff --git a/src/plugins/intel_gpu/tests/test_cases/split_gpu_test.cpp b/src/plugins/intel_gpu/tests/test_cases/split_gpu_test.cpp index 69ddf71c894..c964fc71196 100644 --- a/src/plugins/intel_gpu/tests/test_cases/split_gpu_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/split_gpu_test.cpp @@ -6,7 +6,6 @@ #include #include -#include #include #include @@ -634,9 +633,9 @@ TEST(split_gpu_f32, basic_in2x3x2x2_split_scale_feature_bfyx) { { "out1",{ 0, 1, 0, 0 } }, { "out2",{ 0, 2, 0, 0 } } })); - topology.add(scale("scale0", "split:out0", "scale_input0")); - topology.add(scale("scale1", "split:out1", "scale_input1")); - topology.add(scale("scale2", "split:out2", "scale_input2")); + topology.add(eltwise("scale0", { "split:out0", "scale_input0" }, eltwise_mode::prod)); + topology.add(eltwise("scale1", { "split:out1", "scale_input1" }, eltwise_mode::prod)); + topology.add(eltwise("scale2", { "split:out2", "scale_input2" }, eltwise_mode::prod)); std::vector scale_input_vec0 = { 1.f }; set_values(scale_input0, scale_input_vec0); diff --git a/src/plugins/intel_gpu/tests/test_cases/topology_test.cpp b/src/plugins/intel_gpu/tests/test_cases/topology_test.cpp index 13c523b66e0..199845668e4 100644 --- a/src/plugins/intel_gpu/tests/test_cases/topology_test.cpp +++ b/src/plugins/intel_gpu/tests/test_cases/topology_test.cpp @@ -10,7 +10,6 @@ #include #include #include -#include #include #include #include @@ -299,20 +298,6 @@ protected: return true; } }; - class scale_layer_type : public topology_layer_type - { - virtual bool AddPrimitive(cldnn::topology& topology, cldnn::primitive_id id, cldnn::layout output_layout, std::deque& input_layouts) - { - // for now using just one set of params - // todo: randomize params - cldnn::primitive_id input_id = topology_generator::CreateLayerId(); - input_layouts.push_back({ input_id, output_layout }); - cldnn::primitive_id scale_params_id = id + "_scale_params"; - AddRandomMemory(topology, scale_params_id, output_layout); - topology.add(cldnn::scale(id, input_id, scale_params_id, "")); - return true; - } - }; class softmax_layer_type : public topology_layer_type { virtual bool AddPrimitive(cldnn::topology& topology, cldnn::primitive_id id, cldnn::layout output_layout, std::deque& input_layouts) @@ -456,7 +441,6 @@ std::vector(new topology_test::topology_generator::activation_layer_type()), std::shared_ptr(new topology_test::topology_generator::depth_concatenate_layer_type()), std::shared_ptr(new topology_test::topology_generator::eltwise_layer_type()), - std::shared_ptr(new topology_test::topology_generator::scale_layer_type()), std::shared_ptr(new topology_test::topology_generator::softmax_layer_type()), // Only add new types at the end }; diff --git a/src/plugins/intel_gpu/tests/test_utils/test_utils.h b/src/plugins/intel_gpu/tests/test_utils/test_utils.h index ad78e33e7cb..be8e606691f 100644 --- a/src/plugins/intel_gpu/tests/test_utils/test_utils.h +++ b/src/plugins/intel_gpu/tests/test_utils/test_utils.h @@ -17,7 +17,6 @@ #include #include #include -#include #include #include #include @@ -528,11 +527,8 @@ inline void PrintTupleTo(const std::tuple, std::sha << " Spatial bins x: " << p->spatial_bins_x << " Spatial bins y: " << p->spatial_bins_y << " Output dim: " << p->output_dim; - } else if(primitive->type == cldnn::scale::type_id()) { - auto s = std::static_pointer_cast(primitive); - (void)s; } else if(primitive->type == cldnn::softmax::type_id()) { - auto sm = std::static_pointer_cast(primitive); + auto sm = std::static_pointer_cast(primitive); (void)sm; } else if (primitive->type == cldnn::reorder::type_id()) { auto reorder = std::static_pointer_cast(primitive);