CVS-108963 Coverity fixes (#17161)

This commit is contained in:
Evgenya Stepyreva 2023-04-25 01:03:56 +04:00 committed by GitHub
parent 64b5a4595a
commit 758ec32001
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -452,7 +452,7 @@ AutoBatchExecutableNetwork::AutoBatchExecutableNetwork(
AutoBatchExecutableNetwork::~AutoBatchExecutableNetwork() { AutoBatchExecutableNetwork::~AutoBatchExecutableNetwork() {
_terminate = true; _terminate = true;
for (auto w : _workerRequests) { for (const auto& w : _workerRequests) {
w->_thread.join(); w->_thread.join();
} }
_workerRequests.clear(); _workerRequests.clear();
@ -698,7 +698,7 @@ DeviceInformation AutoBatchInferencePlugin::ParseMetaDevice(
metaDevice.config = GetCore()->GetSupportedConfig(metaDevice.deviceName, user_config); metaDevice.config = GetCore()->GetSupportedConfig(metaDevice.deviceName, user_config);
// check that no irrelevant config-keys left // check that no irrelevant config-keys left
for (auto k : user_config) { for (const auto& k : user_config) {
const auto& name = k.first; const auto& name = k.first;
if (metaDevice.config.find(name) == metaDevice.config.end() && if (metaDevice.config.find(name) == metaDevice.config.end() &&
!ov::util::contains(supported_configKeys, name)) { !ov::util::contains(supported_configKeys, name)) {
@ -916,7 +916,7 @@ InferenceEngine::IExecutableNetworkInternal::Ptr AutoBatchInferencePlugin::LoadN
// TODO: use the per-network metric (22.2) rather than plugin-level // TODO: use the per-network metric (22.2) rather than plugin-level
auto stats = auto stats =
pCore->GetMetric(device, ov::intel_gpu::memory_statistics.name()).as<std::map<std::string, uint64_t>>(); pCore->GetMetric(device, ov::intel_gpu::memory_statistics.name()).as<std::map<std::string, uint64_t>>();
for (auto s : stats) for (const auto& s : stats)
footprint += s.second; footprint += s.second;
return footprint; return footprint;
}; };
@ -939,7 +939,7 @@ InferenceEngine::IExecutableNetworkInternal::Ptr AutoBatchInferencePlugin::LoadN
} }
// auto-batch settings // auto-batch settings
std::unordered_map<std::string, InferenceEngine::Parameter> networkConfig; std::unordered_map<std::string, InferenceEngine::Parameter> networkConfig;
for (auto c : fullConfig) { for (const auto& c : fullConfig) {
if (supported_configKeys.end() != std::find(supported_configKeys.begin(), supported_configKeys.end(), c.first)) if (supported_configKeys.end() != std::find(supported_configKeys.begin(), supported_configKeys.end(), c.first))
networkConfig.insert(c); networkConfig.insert(c);
} }
@ -981,7 +981,7 @@ InferenceEngine::QueryNetworkResult AutoBatchInferencePlugin::QueryNetwork(
if (!core) if (!core)
return InferenceEngine::QueryNetworkResult(); return InferenceEngine::QueryNetworkResult();
auto cfg = user_config; auto cfg = user_config;
for (auto c : cfg) { for (const auto& c : cfg) {
if (c.first == CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG) || c.first == ov::device::priorities.name()) { if (c.first == CONFIG_KEY(AUTO_BATCH_DEVICE_CONFIG) || c.first == ov::device::priorities.name()) {
auto val = c.second; auto val = c.second;
cfg.erase(c.first); cfg.erase(c.first);