[CPU][ARM] Disable default fp16 inference precision (#19445)

This commit is contained in:
Aleksandr Voron 2023-08-29 15:08:46 +02:00 committed by GitHub
parent f6ab1e4833
commit 84fc6fb626
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -245,16 +245,10 @@ void Config::readProperties(const std::map<std::string, std::string> &prop, Mode
// when both execution_mode and inference_precision are specified
if (!inferencePrecisionSetExplicitly) {
if (executionMode == ov::hint::ExecutionMode::PERFORMANCE) {
inferencePrecision = ov::element::f32;
#if defined(OV_CPU_ARM_ENABLE_FP16)
//fp16 precision is used as default precision on ARM for non-convolution networks
//fp16 ACL convolution is slower than fp32
if (modelType != ModelType::CNN)
inferencePrecision = ov::element::f16;
#else
if (mayiuse(avx512_core_bf16))
inferencePrecision = ov::element::bf16;
#endif
else
inferencePrecision = ov::element::f32;
} else {
inferencePrecision = ov::element::f32;
}