From 84fc6fb6262e22114381e6ecb637d4ce219ad144 Mon Sep 17 00:00:00 2001 From: Aleksandr Voron Date: Tue, 29 Aug 2023 15:08:46 +0200 Subject: [PATCH] [CPU][ARM] Disable default fp16 inference precision (#19445) --- src/plugins/intel_cpu/src/config.cpp | 10 ++-------- 1 file changed, 2 insertions(+), 8 deletions(-) diff --git a/src/plugins/intel_cpu/src/config.cpp b/src/plugins/intel_cpu/src/config.cpp index 719764188da..c616df1ecf8 100644 --- a/src/plugins/intel_cpu/src/config.cpp +++ b/src/plugins/intel_cpu/src/config.cpp @@ -245,16 +245,10 @@ void Config::readProperties(const std::map &prop, Mode // when both execution_mode and inference_precision are specified if (!inferencePrecisionSetExplicitly) { if (executionMode == ov::hint::ExecutionMode::PERFORMANCE) { - inferencePrecision = ov::element::f32; -#if defined(OV_CPU_ARM_ENABLE_FP16) - //fp16 precision is used as default precision on ARM for non-convolution networks - //fp16 ACL convolution is slower than fp32 - if (modelType != ModelType::CNN) - inferencePrecision = ov::element::f16; -#else if (mayiuse(avx512_core_bf16)) inferencePrecision = ov::element::bf16; -#endif + else + inferencePrecision = ov::element::f32; } else { inferencePrecision = ov::element::f32; }