[GNA] Disable AddFakeQuantizeFusion (#5037)
* issue 52034 * as per-channel quantization not supported in FQ layer in GNAPlugin
This commit is contained in:
parent
320009d3b2
commit
77d4fbc286
@ -55,6 +55,7 @@
|
||||
#include <transformations/common_optimizations/fq_reshape_fusion.hpp>
|
||||
#include <transformations/common_optimizations/pull_transpose_through_fq.hpp>
|
||||
#include <transformations/common_optimizations/relu_fake_quantize_fusion.hpp>
|
||||
#include <transformations/common_optimizations/add_fake_quantize_fusion.hpp>
|
||||
|
||||
#if GNA_LIB_VER == 2
|
||||
#include <gna2-model-api.h>
|
||||
@ -635,6 +636,8 @@ void GNAPlugin::LoadNetwork(CNNNetwork & _network) {
|
||||
pass_config->disable<ngraph::pass::FakeQuantizeReshapeFusion>();
|
||||
pass_config->disable<ngraph::pass::PullTransposeThroughFQUp>();
|
||||
pass_config->disable<ngraph::pass::ReluFakeQuantizeFusion>();
|
||||
// Consider to enable after per-channel quantization on FakeQuantize layer is supported in GNAPlugin, see issue 52034
|
||||
pass_config->disable<ngraph::pass::AddFakeQuantizeFusion>();
|
||||
manager.run_passes(graph);
|
||||
convertedNetwork = InferenceEngine::details::convertFunctionToICNNNetwork(graph, clonedNetwork);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user