From ee256e801c69386787c41ee4b34629f3e3edf14e Mon Sep 17 00:00:00 2001 From: Vladimir Paramuzov Date: Mon, 19 Dec 2022 11:19:07 +0400 Subject: [PATCH] [GPU] Removed legacy public gpu config and related processing (#14707) --- .../include/ie/cldnn/cldnn_config.hpp | 130 ------------------ src/inference/include/ie/ie_plugin_config.hpp | 2 +- .../intel_gpu/plugin/device_config.hpp | 10 -- .../intel_gpu/src/plugin/device_config.cpp | 106 +------------- src/plugins/intel_gpu/src/plugin/graph.cpp | 7 +- src/plugins/intel_gpu/src/plugin/plugin.cpp | 3 +- src/plugins/intel_gpu/src/plugin/program.cpp | 4 - .../intel_gpu/src/plugin/remote_context.cpp | 10 +- .../src/plugin/transformations_pipeline.cpp | 8 -- .../behavior/plugin/configuration_tests.cpp | 25 ---- 10 files changed, 11 insertions(+), 294 deletions(-) delete mode 100644 src/inference/include/ie/cldnn/cldnn_config.hpp diff --git a/src/inference/include/ie/cldnn/cldnn_config.hpp b/src/inference/include/ie/cldnn/cldnn_config.hpp deleted file mode 100644 index ef91b1333dc..00000000000 --- a/src/inference/include/ie/cldnn/cldnn_config.hpp +++ /dev/null @@ -1,130 +0,0 @@ -// Copyright (C) 2018-2022 Intel Corporation -// SPDX-License-Identifier: Apache-2.0 -// - -/** - * @brief A header for advanced hardware related properties for clDNN plugin - * To use in SetConfig() method of plugins - * - * @file cldnn_config.hpp - */ -#pragma once - -#include "gpu/gpu_config.hpp" -#include "ie_api.h" -#include "ie_plugin_config.hpp" - -namespace InferenceEngine { - -/** - * @brief GPU plugin configuration - */ -namespace CLDNNConfigParams { - -/** - * @brief shortcut for defining configuration keys - */ -#define CLDNN_CONFIG_KEY(name) InferenceEngine::CLDNNConfigParams::_CONFIG_KEY(CLDNN_##name) -#define DECLARE_CLDNN_CONFIG_KEY(name) DECLARE_CONFIG_KEY(CLDNN_##name) -#define DECLARE_CLDNN_CONFIG_VALUE(name) DECLARE_CONFIG_VALUE(CLDNN_##name) - -/** - * @brief This key instructs the clDNN plugin to use the OpenCL queue priority hint - * as defined in https://www.khronos.org/registry/OpenCL/specs/opencl-2.1-extensions.pdf - * this option should be used with an unsigned integer value (1 is lowest priority) - * 0 means no priority hint is set and default queue is created. - */ -INFERENCE_ENGINE_DEPRECATED("Use InferenceEngine::GPUConfigParams::GPU_PLUGIN_PRIORITY instead") -DECLARE_CLDNN_CONFIG_KEY(PLUGIN_PRIORITY); - -/** - * @brief This key instructs the clDNN plugin to use throttle hints the OpenCL queue throttle hint - * as defined in https://www.khronos.org/registry/OpenCL/specs/opencl-2.1-extensions.pdf, - * chapter 9.19. This option should be used with an unsigned integer value (1 is lowest energy consumption) - * 0 means no throttle hint is set and default queue created. - */ -INFERENCE_ENGINE_DEPRECATED("Use InferenceEngine::GPUConfigParams::GPU_PLUGIN_THROTTLE instead") -DECLARE_CLDNN_CONFIG_KEY(PLUGIN_THROTTLE); - -/** - * @brief This key controls clDNN memory pool optimization. - * Turned off by default. - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CLDNN_CONFIG_KEY(MEM_POOL); - -/** - * @brief This key defines the directory name to which clDNN graph visualization will be dumped. - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CLDNN_CONFIG_KEY(GRAPH_DUMPS_DIR); - -/** - * @brief This key defines the directory name to which full program sources will be dumped. - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CLDNN_CONFIG_KEY(SOURCES_DUMPS_DIR); - -/** - * @brief This key enables FP16 precision for quantized models. - * By default the model is converted to FP32 precision before running LPT. If this key is enabled (default), then - * non-quantized layers will be converted back to FP16 after LPT, which might imrpove the performance if a model has a - * lot of compute operations in non-quantized path. This key has no effect if current device doesn't have INT8 - * optimization capabilities. - */ -DECLARE_CLDNN_CONFIG_KEY(ENABLE_FP16_FOR_QUANTIZED_MODELS); - -/** - * @brief This key should be set to correctly handle NV12 input without pre-processing. - * Turned off by default. - */ -INFERENCE_ENGINE_DEPRECATED("Use InferenceEngine::GPUConfigParams::GPU_NV12_TWO_INPUTS instead") -DECLARE_CLDNN_CONFIG_KEY(NV12_TWO_INPUTS); - -} // namespace CLDNNConfigParams - -namespace PluginConfigParams { - -/** - * @brief This key enables dumping of the kernels used by the plugin for custom layers. - * - * This option should be used with values: PluginConfigParams::YES or PluginConfigParams::NO (default) - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CONFIG_KEY(DUMP_KERNELS); - -/** - * @brief This key controls performance tuning done or used by the plugin. - * - * This option should be used with values: - * PluginConfigParams::TUNING_DISABLED (default) - * PluginConfigParams::TUNING_USE_EXISTING - use existing data from tuning file - * PluginConfigParams::TUNING_CREATE - create tuning data for parameters not present in tuning file - * PluginConfigParams::TUNING_UPDATE - perform non-tuning updates like removal of invalid/deprecated data - * PluginConfigParams::TUNING_RETUNE - create tuning data for all parameters, even if already present - * - * For values TUNING_CREATE and TUNING_RETUNE the file will be created if it does not exist. - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CONFIG_KEY(TUNING_MODE); - -INFERENCE_ENGINE_DEPRECATED("The config value will be removed") -DECLARE_CONFIG_VALUE(TUNING_CREATE); -INFERENCE_ENGINE_DEPRECATED("The config value will be removed") -DECLARE_CONFIG_VALUE(TUNING_USE_EXISTING); -INFERENCE_ENGINE_DEPRECATED("The config value will be removed") -DECLARE_CONFIG_VALUE(TUNING_DISABLED); -INFERENCE_ENGINE_DEPRECATED("The config value will be removed") -DECLARE_CONFIG_VALUE(TUNING_UPDATE); -INFERENCE_ENGINE_DEPRECATED("The config value will be removed") -DECLARE_CONFIG_VALUE(TUNING_RETUNE); - -/** - * @brief This key defines the tuning data filename to be created/used - */ -INFERENCE_ENGINE_DEPRECATED("The config key will be removed") -DECLARE_CONFIG_KEY(TUNING_FILE); - -} // namespace PluginConfigParams - -} // namespace InferenceEngine diff --git a/src/inference/include/ie/ie_plugin_config.hpp b/src/inference/include/ie/ie_plugin_config.hpp index 2962b4802af..0624066d9bd 100644 --- a/src/inference/include/ie/ie_plugin_config.hpp +++ b/src/inference/include/ie/ie_plugin_config.hpp @@ -475,9 +475,9 @@ DECLARE_CONFIG_KEY(FORCE_TBB_TERMINATE); } // namespace InferenceEngine +#include "gpu/gpu_config.hpp" #include "hetero/hetero_plugin_config.hpp" #include "multi-device/multi_device_config.hpp" // remove in 2022.1 major release -#include "cldnn/cldnn_config.hpp" #include "gna/gna_config.hpp" diff --git a/src/plugins/intel_gpu/include/intel_gpu/plugin/device_config.hpp b/src/plugins/intel_gpu/include/intel_gpu/plugin/device_config.hpp index 1b46d891a35..227390ee2ef 100644 --- a/src/plugins/intel_gpu/include/intel_gpu/plugin/device_config.hpp +++ b/src/plugins/intel_gpu/include/intel_gpu/plugin/device_config.hpp @@ -24,18 +24,13 @@ struct Config { useProfiling(false), dumpCustomKernels(false), exclusiveAsyncRequests(false), - memory_pool_on(true), enableDynamicBatch(false), enableInt8(true), nv12_two_inputs(false), - enable_fp16_for_quantized_models(true), queuePriority(cldnn::priority_mode_types::med), queueThrottle(cldnn::throttle_mode_types::med), max_dynamic_batch(1), customLayers({}), - tuningConfig(), - graph_dumps_dir(""), - sources_dumps_dir(""), kernels_cache_dir(""), inference_precision(ov::element::undefined), task_exec_config({"GPU plugin internal task executor", // name @@ -70,18 +65,13 @@ struct Config { bool useProfiling; bool dumpCustomKernels; bool exclusiveAsyncRequests; - bool memory_pool_on; bool enableDynamicBatch; bool enableInt8; bool nv12_two_inputs; - bool enable_fp16_for_quantized_models; cldnn::priority_mode_types queuePriority; cldnn::throttle_mode_types queueThrottle; int max_dynamic_batch; CustomLayerMap customLayers; - cldnn::tuning_config_options tuningConfig; - std::string graph_dumps_dir; - std::string sources_dumps_dir; std::string kernels_cache_dir; ov::element::Type inference_precision; InferenceEngine::IStreamsExecutor::Config task_exec_config; diff --git a/src/plugins/intel_gpu/src/plugin/device_config.cpp b/src/plugins/intel_gpu/src/plugin/device_config.cpp index d70f2480f85..a71e7483b4f 100644 --- a/src/plugins/intel_gpu/src/plugin/device_config.cpp +++ b/src/plugins/intel_gpu/src/plugin/device_config.cpp @@ -7,7 +7,6 @@ #include #include -#include #include #include @@ -92,16 +91,7 @@ void Config::UpdateFromMap(const std::map& configMap, } else { IE_THROW(NotFound) << "Unsupported property value by plugin: " << val; } - } else if (key.compare(PluginConfigParams::KEY_DUMP_KERNELS) == 0) { - if (val.compare(PluginConfigParams::YES) == 0) { - dumpCustomKernels = true; - } else if (val.compare(PluginConfigParams::NO) == 0) { - dumpCustomKernels = false; - } else { - IE_THROW(NotFound) << "Unsupported property value by plugin: " << val; - } - } else if (key.compare(GPUConfigParams::KEY_GPU_PLUGIN_PRIORITY) == 0 || - key.compare(CLDNNConfigParams::KEY_CLDNN_PLUGIN_PRIORITY) == 0) { + } else if (key.compare(GPUConfigParams::KEY_GPU_PLUGIN_PRIORITY) == 0) { std::stringstream ss(val); uint32_t uVal(0); ss >> uVal; @@ -160,8 +150,7 @@ void Config::UpdateFromMap(const std::map& configMap, task_exec_config._streams = std::min(task_exec_config._streams, static_cast(std::thread::hardware_concurrency())); } - } else if (key.compare(GPUConfigParams::KEY_GPU_PLUGIN_THROTTLE) == 0 || - key.compare(CLDNNConfigParams::KEY_CLDNN_PLUGIN_THROTTLE) == 0) { + } else if (key.compare(GPUConfigParams::KEY_GPU_PLUGIN_THROTTLE) == 0) { std::stringstream ss(val); uint32_t uVal(0); ss >> uVal; @@ -200,45 +189,11 @@ void Config::UpdateFromMap(const std::map& configMap, for (auto& file : configFiles) { CustomLayer::LoadFromFile(file, customLayers); } - } else if (key.compare(PluginConfigParams::KEY_TUNING_MODE) == 0) { - if (val.compare(PluginConfigParams::TUNING_DISABLED) == 0) { - tuningConfig.mode = cldnn::tuning_mode::tuning_disabled; - } else if (val.compare(PluginConfigParams::TUNING_CREATE) == 0) { - tuningConfig.mode = cldnn::tuning_mode::tuning_tune_and_cache; - } else if (val.compare(PluginConfigParams::TUNING_USE_EXISTING) == 0) { - tuningConfig.mode = cldnn::tuning_mode::tuning_use_cache; - } else if (val.compare(PluginConfigParams::TUNING_UPDATE) == 0) { - tuningConfig.mode = cldnn::tuning_mode::tuning_use_and_update; - } else if (val.compare(PluginConfigParams::TUNING_RETUNE) == 0) { - tuningConfig.mode = cldnn::tuning_mode::tuning_retune_and_cache; - } else { - IE_THROW(NotFound) << "Unsupported tuning mode value by plugin: " << val; - } - } else if (key.compare(PluginConfigParams::KEY_TUNING_FILE) == 0) { - tuningConfig.cache_file_path = val; - } else if (key.compare(CLDNNConfigParams::KEY_CLDNN_MEM_POOL) == 0) { - if (val.compare(PluginConfigParams::YES) == 0) { - memory_pool_on = true; - } else if (val.compare(PluginConfigParams::NO) == 0) { - memory_pool_on = false; - } else { - IE_THROW(NotFound) << "Unsupported memory pool flag value: " << val; - } - } else if (key.compare(CLDNNConfigParams::KEY_CLDNN_GRAPH_DUMPS_DIR) == 0) { - if (!val.empty()) { - graph_dumps_dir = val; - createDirectory(graph_dumps_dir); - } } else if (key.compare(PluginConfigParams::KEY_CACHE_DIR) == 0 || key == ov::cache_dir) { if (!val.empty()) { kernels_cache_dir = val; createDirectory(kernels_cache_dir); } - } else if (key.compare(CLDNNConfigParams::KEY_CLDNN_SOURCES_DUMPS_DIR) == 0) { - if (!val.empty()) { - sources_dumps_dir = val; - createDirectory(sources_dumps_dir); - } } else if (key.compare(PluginConfigParams::KEY_EXCLUSIVE_ASYNC_REQUESTS) == 0) { if (val.compare(PluginConfigParams::YES) == 0) { exclusiveAsyncRequests = true; @@ -282,8 +237,7 @@ void Config::UpdateFromMap(const std::map& configMap, } else { IE_THROW(NotFound) << "Unsupported property value by plugin: " << val; } - } else if (key.compare(GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS) == 0 || - key.compare(CLDNNConfigParams::KEY_CLDNN_NV12_TWO_INPUTS) == 0) { + } else if (key.compare(GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS) == 0) { if (val.compare(PluginConfigParams::YES) == 0) { nv12_two_inputs = true; } else if (val.compare(PluginConfigParams::NO) == 0) { @@ -291,14 +245,6 @@ void Config::UpdateFromMap(const std::map& configMap, } else { IE_THROW(NotFound) << "Unsupported NV12 flag value: " << val; } - } else if (key.compare(CLDNNConfigParams::KEY_CLDNN_ENABLE_FP16_FOR_QUANTIZED_MODELS) == 0) { - if (val.compare(PluginConfigParams::YES) == 0) { - enable_fp16_for_quantized_models = true; - } else if (val.compare(PluginConfigParams::NO) == 0) { - enable_fp16_for_quantized_models = false; - } else { - IE_THROW(NotFound) << "Unsupported KEY_CLDNN_ENABLE_FP16_FOR_QUANTIZED_MODELS flag value: " << val; - } } else if (key.compare(GPUConfigParams::KEY_GPU_MAX_NUM_THREADS) == 0 || key == ov::compilation_num_threads) { int max_threads = std::max(1, static_cast(std::thread::hardware_concurrency())); try { @@ -353,39 +299,22 @@ void Config::adjustKeyMapValues() { key_config_map[ov::enable_profiling.name()] = PluginConfigParams::NO; } - if (dumpCustomKernels) - key_config_map[PluginConfigParams::KEY_DUMP_KERNELS] = PluginConfigParams::YES; - else - key_config_map[PluginConfigParams::KEY_DUMP_KERNELS] = PluginConfigParams::NO; - if (exclusiveAsyncRequests) key_config_map[PluginConfigParams::KEY_EXCLUSIVE_ASYNC_REQUESTS] = PluginConfigParams::YES; else key_config_map[PluginConfigParams::KEY_EXCLUSIVE_ASYNC_REQUESTS] = PluginConfigParams::NO; - if (memory_pool_on) - key_config_map[CLDNNConfigParams::KEY_CLDNN_MEM_POOL] = PluginConfigParams::YES; - else - key_config_map[CLDNNConfigParams::KEY_CLDNN_MEM_POOL] = PluginConfigParams::NO; - if (enableDynamicBatch) key_config_map[PluginConfigParams::KEY_DYN_BATCH_ENABLED] = PluginConfigParams::YES; else key_config_map[PluginConfigParams::KEY_DYN_BATCH_ENABLED] = PluginConfigParams::NO; if (nv12_two_inputs) { - key_config_map[CLDNNConfigParams::KEY_CLDNN_NV12_TWO_INPUTS] = PluginConfigParams::YES; key_config_map[GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS] = PluginConfigParams::YES; } else { - key_config_map[CLDNNConfigParams::KEY_CLDNN_NV12_TWO_INPUTS] = PluginConfigParams::NO; key_config_map[GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS] = PluginConfigParams::NO; } - if (enable_fp16_for_quantized_models) - key_config_map[CLDNNConfigParams::KEY_CLDNN_ENABLE_FP16_FOR_QUANTIZED_MODELS] = PluginConfigParams::YES; - else - key_config_map[CLDNNConfigParams::KEY_CLDNN_ENABLE_FP16_FOR_QUANTIZED_MODELS] = PluginConfigParams::NO; - key_config_map[ov::hint::inference_precision.name()] = inference_precision.get_type_name(); { @@ -420,7 +349,6 @@ void Config::adjustKeyMapValues() { default: break; } - key_config_map[CLDNNConfigParams::KEY_CLDNN_PLUGIN_PRIORITY] = qp; key_config_map[GPUConfigParams::KEY_GPU_PLUGIN_PRIORITY] = qp; } { @@ -448,7 +376,6 @@ void Config::adjustKeyMapValues() { default: break; } - key_config_map[CLDNNConfigParams::KEY_CLDNN_PLUGIN_THROTTLE] = qt; key_config_map[GPUConfigParams::KEY_GPU_PLUGIN_THROTTLE] = qt; } { @@ -471,30 +398,7 @@ void Config::adjustKeyMapValues() { hostTaskPriority = ov::util::to_string(ov::hint::Priority::MEDIUM); key_config_map[ov::intel_gpu::hint::host_task_priority.name()] = hostTaskPriority; } - { - std::string tm = PluginConfigParams::TUNING_DISABLED; - switch (tuningConfig.mode) { - case cldnn::tuning_mode::tuning_tune_and_cache: - tm = PluginConfigParams::TUNING_CREATE; - break; - case cldnn::tuning_mode::tuning_use_cache: - tm = PluginConfigParams::TUNING_USE_EXISTING; - break; - case cldnn::tuning_mode::tuning_use_and_update: - tm = PluginConfigParams::TUNING_UPDATE; - break; - case cldnn::tuning_mode::tuning_retune_and_cache: - tm = PluginConfigParams::TUNING_RETUNE; - break; - default: - break; - } - key_config_map[PluginConfigParams::KEY_TUNING_MODE] = tm; - key_config_map[PluginConfigParams::KEY_TUNING_FILE] = tuningConfig.cache_file_path; - } - key_config_map[CLDNNConfigParams::KEY_CLDNN_GRAPH_DUMPS_DIR] = graph_dumps_dir; - key_config_map[CLDNNConfigParams::KEY_CLDNN_SOURCES_DUMPS_DIR] = sources_dumps_dir; key_config_map[PluginConfigParams::KEY_CACHE_DIR] = kernels_cache_dir; key_config_map[ov::cache_dir.name()] = kernels_cache_dir; @@ -560,12 +464,8 @@ bool Config::CanShareContextWith(const Config& other) const { return this->throughput_streams == other.throughput_streams && this->useProfiling == other.useProfiling && this->dumpCustomKernels == other.dumpCustomKernels && - this->memory_pool_on == other.memory_pool_on && this->queueThrottle == other.queueThrottle && this->queuePriority == other.queuePriority && - this->sources_dumps_dir == other.sources_dumps_dir && - this->tuningConfig.mode == other.tuningConfig.mode && - this->tuningConfig.cache_file_path == other.tuningConfig.cache_file_path && this->kernels_cache_dir == other.kernels_cache_dir && this->device_id == other.device_id && this->task_exec_config._streams == other.task_exec_config._streams && diff --git a/src/plugins/intel_gpu/src/plugin/graph.cpp b/src/plugins/intel_gpu/src/plugin/graph.cpp index 1ae2bc88c46..c108dbf5e00 100644 --- a/src/plugins/intel_gpu/src/plugin/graph.cpp +++ b/src/plugins/intel_gpu/src/plugin/graph.cpp @@ -13,7 +13,6 @@ #include "intel_gpu/plugin/graph.hpp" #include "intel_gpu/plugin/simple_math.hpp" -#include #include "intel_gpu/plugin/infer_request.hpp" #include "intel_gpu/plugin/itt.hpp" @@ -136,14 +135,14 @@ std::shared_ptr Graph::BuildNetwork(std::shared_ptr(program, m_stream_id); } - - if (!m_config.graph_dumps_dir.empty() && m_stream_id == 0) { + GPU_DEBUG_GET_INSTANCE(debug_config); + GPU_DEBUG_IF(!debug_config->dump_graphs.empty() && m_stream_id == 0) { static int net_id = 0; auto steps_info = network->get_optimizer_passes_info(); size_t step_idx = 0; for (auto& step : steps_info) { CNNNetwork net(GetExecGraphInfoByPrimitivesInfo(step.second, true)); - net.serialize(m_config.graph_dumps_dir + std::to_string(net_id) + "_" + + net.serialize(debug_config->dump_graphs + std::to_string(net_id) + "_" + std::to_string(step_idx) + "_" + step.first + "_graph.xml"); step_idx++; } diff --git a/src/plugins/intel_gpu/src/plugin/plugin.cpp b/src/plugins/intel_gpu/src/plugin/plugin.cpp index 77692f61f90..84b842e99a9 100644 --- a/src/plugins/intel_gpu/src/plugin/plugin.cpp +++ b/src/plugins/intel_gpu/src/plugin/plugin.cpp @@ -353,7 +353,6 @@ void Plugin::SetConfig(const std::map &config) { streamsSet = config.find(PluginConfigParams::KEY_GPU_THROUGHPUT_STREAMS) != config.end() || config.find(ov::num_streams.name()) != config.end(); throttlingSet = config.find(GPUConfigParams::KEY_GPU_PLUGIN_THROTTLE) != config.end() || - config.find(CLDNNConfigParams::KEY_CLDNN_PLUGIN_THROTTLE) != config.end() || config.find(ov::intel_gpu::hint::queue_throttle.name()) != config.end(); std::string device_id; cldnn::device_info device_info = device_map.begin()->second->get_info(); @@ -882,7 +881,7 @@ Parameter Plugin::GetMetric(const std::string& name, const std::map Program::BuildProgram(const std::vectoris_dynamic()) { allow_new_shape_infer = true; @@ -326,7 +323,6 @@ std::shared_ptr Program::BuildProgram(const std::vectorsecond; - bool enable_profiling = (m_config.useProfiling || - (m_config.tuningConfig.mode == cldnn::tuning_mode::tuning_tune_and_cache) || - (m_config.tuningConfig.mode == cldnn::tuning_mode::tuning_retune_and_cache)); - auto engine_params = Plugin::GetParams(m_config, dev, m_external_queue); m_engine = cldnn::engine::create(engine_params.engine_type, engine_params.runtime_type, dev, - cldnn::engine_configuration(enable_profiling, + cldnn::engine_configuration(m_config.useProfiling, engine_params.queue_type, - m_config.sources_dumps_dir, + std::string(), m_config.queuePriority, m_config.queueThrottle, - m_config.memory_pool_on, + true, engine_params.use_unified_shared_memory, m_config.kernels_cache_dir, m_config.throughput_streams), diff --git a/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp b/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp index 3d07b1e417c..f9ccc9b1d05 100644 --- a/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp +++ b/src/plugins/intel_gpu/src/plugin/transformations_pipeline.cpp @@ -424,14 +424,6 @@ void TransformationsPipeline::apply(std::shared_ptr func) { OV_ITT_SCOPED_TASK(itt::domains::intel_gpu_plugin, "TransformationsPipeline::apply::lpt"); using namespace ngraph::pass::low_precision; - // Conversion to FP32 might be needed for quantized models that face any fp16 related issues (e.g. overflow) for non-quantized layers - // With this key users can work-around such issues - if (!config.enable_fp16_for_quantized_models) { - ngraph::pass::Manager manager; - manager.register_pass(precisions_array {{ ngraph::element::f16, ngraph::element::f32 }}); - manager.run_passes(func); - } - auto supportedPrecisions = std::vector({ PrecisionsRestriction::create({ {{0}, {ngraph::element::u8, ngraph::element::i8}}, diff --git a/src/tests/functional/plugin/gpu/shared_tests_instances/behavior/plugin/configuration_tests.cpp b/src/tests/functional/plugin/gpu/shared_tests_instances/behavior/plugin/configuration_tests.cpp index ee05707a3b2..6c98f7f912f 100644 --- a/src/tests/functional/plugin/gpu/shared_tests_instances/behavior/plugin/configuration_tests.cpp +++ b/src/tests/functional/plugin/gpu/shared_tests_instances/behavior/plugin/configuration_tests.cpp @@ -3,7 +3,6 @@ // #include "behavior/plugin/configuration_tests.hpp" -#include "cldnn/cldnn_config.hpp" #include "gpu/gpu_config.hpp" using namespace BehaviorTestsDefinitions; @@ -28,8 +27,6 @@ namespace { {{InferenceEngine::PluginConfigParams::KEY_GPU_THROUGHPUT_STREAMS, "OFF"}}, {{InferenceEngine::PluginConfigParams::KEY_PERF_COUNT, "ON"}}, {{InferenceEngine::PluginConfigParams::KEY_CONFIG_FILE, "unknown_file"}}, - {{InferenceEngine::PluginConfigParams::KEY_DUMP_KERNELS, "ON"}}, - {{InferenceEngine::PluginConfigParams::KEY_TUNING_MODE, "TUNING_UNKNOWN_MODE"}}, {{InferenceEngine::PluginConfigParams::KEY_DEVICE_ID, "DEVICE_UNKNOWN"}}}; }; @@ -55,10 +52,6 @@ namespace { {InferenceEngine::PluginConfigParams::KEY_PERF_COUNT, "ON"}}, {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU}, {InferenceEngine::PluginConfigParams::KEY_CONFIG_FILE, "unknown_file"}}, - {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU}, - {InferenceEngine::PluginConfigParams::KEY_DUMP_KERNELS, "ON"}}, - {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU}, - {InferenceEngine::PluginConfigParams::KEY_TUNING_MODE, "TUNING_UNKNOWN_MODE"}}, {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU}, {InferenceEngine::PluginConfigParams::KEY_DEVICE_ID, "DEVICE_UNKNOWN"}}, {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_CPU}, @@ -80,12 +73,6 @@ namespace { {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU + std::string(",") + CommonTestUtils::DEVICE_CPU}, {InferenceEngine::PluginConfigParams::KEY_CONFIG_FILE, "unknown_file"}}, - {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, - CommonTestUtils::DEVICE_GPU + std::string(",") + CommonTestUtils::DEVICE_CPU}, - {InferenceEngine::PluginConfigParams::KEY_DUMP_KERNELS, "ON"}}, - {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, - CommonTestUtils::DEVICE_GPU + std::string(",") + CommonTestUtils::DEVICE_CPU}, - {InferenceEngine::PluginConfigParams::KEY_TUNING_MODE, "TUNING_UNKNOWN_MODE"}}, {{InferenceEngine::MultiDeviceConfigParams::KEY_MULTI_DEVICE_PRIORITIES, CommonTestUtils::DEVICE_GPU + std::string(",") + CommonTestUtils::DEVICE_CPU}, {InferenceEngine::PluginConfigParams::KEY_DEVICE_ID, "DEVICE_UNKNOWN"}}, @@ -113,10 +100,6 @@ namespace { {InferenceEngine::PluginConfigParams::KEY_PERF_COUNT, "ON"}}, {{CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG), CommonTestUtils::DEVICE_GPU}, {InferenceEngine::PluginConfigParams::KEY_CONFIG_FILE, "unknown_file"}}, - {{CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG), CommonTestUtils::DEVICE_GPU}, - {InferenceEngine::PluginConfigParams::KEY_DUMP_KERNELS, "ON"}}, - {{CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG), CommonTestUtils::DEVICE_GPU}, - {InferenceEngine::PluginConfigParams::KEY_TUNING_MODE, "TUNING_UNKNOWN_MODE"}}, {{CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG), CommonTestUtils::DEVICE_GPU}, {InferenceEngine::PluginConfigParams::KEY_DEVICE_ID, "DEVICE_UNKNOWN"}}}; }; @@ -155,14 +138,6 @@ namespace { IE_SUPPRESS_DEPRECATED_START auto conf_gpu = []() { return std::vector>{ - // Deprecated - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_NV12_TWO_INPUTS, InferenceEngine::PluginConfigParams::YES}}, - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_NV12_TWO_INPUTS, InferenceEngine::PluginConfigParams::NO}}, - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_PLUGIN_THROTTLE, "0"}}, - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_PLUGIN_THROTTLE, "1"}}, - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_PLUGIN_PRIORITY, "0"}}, - {{InferenceEngine::CLDNNConfigParams::KEY_CLDNN_PLUGIN_PRIORITY, "1"}}, - {{InferenceEngine::GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS, InferenceEngine::PluginConfigParams::YES}}, {{InferenceEngine::GPUConfigParams::KEY_GPU_NV12_TWO_INPUTS, InferenceEngine::PluginConfigParams::NO}}, {{InferenceEngine::GPUConfigParams::KEY_GPU_PLUGIN_THROTTLE, "0"}},