[POT] Removed Multiply from ignored ops for transformers (#15600)

* [POT] Removed Multiply from ignored ops for transformers

* Add VPU to change_configurations_by_model_type
This commit is contained in:
Liubov Talamanova
2023-03-07 14:37:29 +00:00
committed by GitHub
parent 15e43e0cc2
commit e6ad0a5154
2 changed files with 2 additions and 2 deletions

View File

@@ -528,7 +528,7 @@ def _fake_quantize_to_types(model, hardware_config):
def change_configurations_by_model_type(model, config, fq_configuration, hardware_config):
if config['model_type'] == 'transformer' and config['target_device'] in ['ANY', 'CPU', 'GPU']:
if config['model_type'] == 'transformer' and config['target_device'] in ['ANY', 'CPU', 'CPU_SPR', 'GPU', 'VPU']:
change_configurations_by_model_type_transformer(model, fq_configuration, hardware_config)

View File

@@ -323,7 +323,7 @@ def get_input_shape_for_bias(activations_statistics, input_node_name):
def get_ignored_operations(model):
operation = {"transformer": [{"type": "Add"}, {"type": "Power"},
{"type": "Squeeze"}, {"type": "Multiply"},
{"type": "Squeeze"},
{"type": "Subtract"}, {"type": "ReduceMean"},
{"type": "SquaredDifference"}, {"type": "MVN"}]}
return operation[model]