From 7eb53586f553aa7fa32d5d0532a91b1122c27cca Mon Sep 17 00:00:00 2001 From: Roman Lyamin Date: Mon, 30 Oct 2023 21:58:34 +0400 Subject: [PATCH] [GPU] Added WA for MarkDequantizationSubgraph to work correctly in the plugin (#20767) --- .../intel_gpu/src/plugin/transformations_pipeline.cpp | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp b/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp index f96122a6531..68268790c62 100644 --- a/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp +++ b/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp @@ -109,6 +109,7 @@ #include "transformations/convert_precision.hpp" #include "transformations/init_node_info.hpp" #include "transformations/rt_info/fused_names_attribute.hpp" +#include "transformations/rt_info/keep_const_precision.hpp" #include "transformations/smart_reshape/matmul_sr.hpp" #include "plugin/transformations/convert_matmul_to_fc.hpp" @@ -159,6 +160,12 @@ void TransformationsPipeline::apply(std::shared_ptr func) { auto pass_config = manager.get_pass_config(); manager.set_per_pass_validation(false); + // Temporary solution, global rt info cleanup is needed + for (auto& node : func->get_ops()) { + ov::enable_constant_folding(node); + ov::disable_keep_const_precision(node); + } + enableInt8 = config.get_property(ov::intel_gpu::enable_lp_transformations) && ov::pass::low_precision::LowPrecision::isFunctionQuantized(func); if (enableInt8) { manager.register_pass(