* Mark all failed ONNX layer tests as XFail
* Add additional xfailed marks
* Add one more failed tests into XFail
* Add conditions for CPU/GPU failures
* Revert "Add conditions for CPU/GPU failures"
This reverts commit 790524c59c.
* Add failures separation for CPU/GPU
* Replace all xfail with skip
257 lines
13 KiB
Python
257 lines
13 KiB
Python
# Copyright (C) 2018-2023 Intel Corporation
|
|
# SPDX-License-Identifier: Apache-2.0
|
|
|
|
import numpy as np
|
|
import pytest
|
|
from common.layer_test_class import check_ir_version
|
|
from common.onnx_layer_test_class import OnnxRuntimeLayerTest
|
|
|
|
from openvino.tools.mo.front.common.partial_infer.utils import int64_array
|
|
from unit_tests.utils.graph import build_graph
|
|
|
|
|
|
def second_input_data_of_reshape(src_shape, axis):
|
|
if axis == 0:
|
|
return [1, -1]
|
|
if axis == 1:
|
|
return [0, -1]
|
|
if axis > 1:
|
|
return [int(np.prod(int64_array(src_shape[: axis]))), -1]
|
|
return [-1, int(np.prod(int64_array(src_shape[len(src_shape) + axis:])))]
|
|
|
|
|
|
def get_flatten_shape(src_shape, axis):
|
|
flatten_axis = axis if axis >= 0 else len(src_shape) + axis
|
|
if flatten_axis == 0:
|
|
fst_dim = 1
|
|
snd_dim = int(np.prod(int64_array(src_shape)))
|
|
elif flatten_axis == 1:
|
|
fst_dim = src_shape[0]
|
|
snd_dim = int(np.prod(int64_array(src_shape[1:])))
|
|
else:
|
|
fst_dim = int(np.prod(int64_array(src_shape[: flatten_axis])))
|
|
snd_dim = int(np.prod(int64_array(src_shape[flatten_axis:])))
|
|
return [fst_dim, snd_dim]
|
|
|
|
|
|
class TestLog(OnnxRuntimeLayerTest):
|
|
def _prepare_input(self, inputs_dict):
|
|
for input in inputs_dict.keys():
|
|
inputs_dict[input] = np.random.rand(*(inputs_dict[input])).astype(
|
|
np.float32) * 255 + 0.5
|
|
return inputs_dict
|
|
|
|
def create_net(self, shape, logsoftmax_axis, ir_version):
|
|
"""
|
|
ONNX net IR net
|
|
|
|
Input->LogSoftmax->Output => Input->Softmax->Log->Output
|
|
|
|
"""
|
|
|
|
#
|
|
# Create ONNX model
|
|
#
|
|
import onnx
|
|
from onnx import helper
|
|
from onnx import TensorProto
|
|
|
|
input = helper.make_tensor_value_info('input', TensorProto.FLOAT, shape)
|
|
output = helper.make_tensor_value_info('output', TensorProto.FLOAT, shape)
|
|
|
|
node_def = onnx.helper.make_node(
|
|
'LogSoftmax',
|
|
inputs=['input'],
|
|
outputs=['output'],
|
|
axis=logsoftmax_axis
|
|
)
|
|
|
|
# Create the graph (GraphProto)
|
|
graph_def = helper.make_graph(
|
|
[node_def],
|
|
'test_model',
|
|
[input],
|
|
[output],
|
|
)
|
|
|
|
# Create the model (ModelProto)
|
|
onnx_net = helper.make_model(graph_def, producer_name='test_model')
|
|
|
|
#
|
|
# Create reference IR net
|
|
#
|
|
ref_net = None
|
|
if check_ir_version(10, None, ir_version):
|
|
converted_shape = shape if len(shape) != 1 else shape[0]
|
|
flatten_shape = get_flatten_shape(shape, logsoftmax_axis)
|
|
reshape_data_val = second_input_data_of_reshape(shape, logsoftmax_axis)
|
|
reduce_sum_shape = np.copy(flatten_shape)
|
|
reduce_sum_shape[1] = 1
|
|
|
|
if len(shape) == 2 and shape == flatten_shape:
|
|
ref_nodes_attributes = {
|
|
'input': {'kind': 'op', 'type': 'Parameter', 'shape': converted_shape},
|
|
'input_data': {'shape': shape, 'kind': 'data', 'value': None},
|
|
'flatten_shape_val': {'shape': int64_array(reshape_data_val).shape,
|
|
'kind': 'data',
|
|
'value': int64_array(reshape_data_val)},
|
|
'flatten_shape': {'type': 'Const', 'kind': 'op', 'shape': 2},
|
|
'flatten_shape_data': {'shape': int64_array([2]), 'kind': 'data',
|
|
'value': None},
|
|
'reshape': {'kind': 'op', 'type': 'Reshape'},
|
|
'reshape_data': {'kind': 'data', 'shape': flatten_shape, 'value': None},
|
|
'reduce_max_axis_val': {'shape': int64_array([1]).shape, 'kind': 'data',
|
|
'value': int64_array([1])},
|
|
'reduce_max_axis': {'type': 'Const', 'kind': 'op', 'shape': 1},
|
|
'reduce_max_axis_data': {'shape': int64_array([1]), 'kind': 'data',
|
|
'value': None},
|
|
'reduce_max': {'type': 'ReduceMax', 'kind': 'op', 'keep_dims': True},
|
|
'reduce_max_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'sub_first': {'type': 'Subtract', 'kind': 'op'},
|
|
'sub_first_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'reduce_sum_axis_val': {'shape': int64_array([1]).shape, 'kind': 'data',
|
|
'value': int64_array([1])},
|
|
'reduce_sum_axis': {'type': 'Const', 'kind': 'op', 'shape': 1},
|
|
'reduce_sum_axis_data': {'shape': int64_array([1]), 'kind': 'data',
|
|
'value': None},
|
|
'reduce_sum': {'type': 'ReduceSum', 'kind': 'op', 'keep_dims': True},
|
|
'reduce_sum_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'exp': {'type': 'Exp', 'kind': 'op'},
|
|
'exp_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'log': {'type': 'Log', 'kind': 'op'},
|
|
'log_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'sub_second': {'type': 'Subtract', 'kind': 'op'},
|
|
'sub_second_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'result': {'kind': 'op', 'type': 'Result'},
|
|
}
|
|
|
|
ref_edges = [
|
|
('input', 'input_data'),
|
|
('flatten_shape_val', 'flatten_shape'),
|
|
('flatten_shape', 'flatten_shape_data'),
|
|
('flatten_shape_data', 'reshape', {'in': 1}),
|
|
('input_data', 'reshape', {'in': 0}),
|
|
('reshape', 'reshape_data'),
|
|
('reduce_max_axis_val', 'reduce_max_axis'),
|
|
('reduce_max_axis', 'reduce_max_axis_data'),
|
|
('reduce_max_axis_data', 'reduce_max', {'in': 1}),
|
|
('reduce_max', 'reduce_max_data'),
|
|
('reshape_data', 'reduce_max', {'out': 0, 'in': 0}),
|
|
('reshape_data', 'sub_first', {'out': 0, 'in': 0}),
|
|
('reduce_max_data', 'sub_first', {'in': 1}),
|
|
('sub_first', 'sub_first_data'),
|
|
('reduce_sum_axis_val', 'reduce_sum_axis'),
|
|
('reduce_sum_axis', 'reduce_sum_axis_data'),
|
|
('reduce_sum_axis_data', 'reduce_sum', {'in': 1}),
|
|
('reduce_sum', 'reduce_sum_data'),
|
|
('sub_first_data', 'exp'),
|
|
('exp', 'exp_data'),
|
|
('exp_data', 'reduce_sum', {'in': 0}),
|
|
('reduce_sum_data', 'log'),
|
|
('log', 'log_data'),
|
|
('log_data', 'sub_second', {'in': 1}),
|
|
('sub_second', 'sub_second_data'),
|
|
('sub_first_data', 'sub_second', {'out': 0, 'in': 0}),
|
|
('sub_second_data', 'result'),
|
|
]
|
|
else:
|
|
ref_nodes_attributes = {
|
|
'input': {'kind': 'op', 'type': 'Parameter', 'shape': converted_shape},
|
|
'input_data': {'shape': shape, 'kind': 'data', 'value': None},
|
|
'flatten_shape_val': {'shape': int64_array(reshape_data_val).shape,
|
|
'kind': 'data',
|
|
'value': int64_array(reshape_data_val)},
|
|
'flatten_shape': {'type': 'Const', 'kind': 'op', 'shape': 2},
|
|
'flatten_shape_data': {'shape': int64_array([2]), 'kind': 'data',
|
|
'value': None},
|
|
'reshape': {'kind': 'op', 'type': 'Reshape'},
|
|
'reshape_data': {'kind': 'data', 'shape': flatten_shape, 'value': None},
|
|
'reduce_max_axis_val': {'shape': int64_array([1]).shape, 'kind': 'data',
|
|
'value': int64_array([1])},
|
|
'reduce_max_axis': {'type': 'Const', 'kind': 'op', 'shape': 1},
|
|
'reduce_max_axis_data': {'shape': int64_array([1]), 'kind': 'data',
|
|
'value': None},
|
|
'reduce_max': {'type': 'ReduceMax', 'kind': 'op', 'keep_dims': True},
|
|
'reduce_max_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'sub_first': {'type': 'Subtract', 'kind': 'op'},
|
|
'sub_first_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'reduce_sum_axis_val': {'shape': int64_array([1]).shape, 'kind': 'data',
|
|
'value': int64_array([1])},
|
|
'reduce_sum_axis': {'type': 'Const', 'kind': 'op', 'shape': 1},
|
|
'reduce_sum_axis_data': {'shape': int64_array([1]), 'kind': 'data',
|
|
'value': None},
|
|
'reduce_sum': {'type': 'ReduceSum', 'kind': 'op', 'keep_dims': True},
|
|
'reduce_sum_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'exp': {'type': 'Exp', 'kind': 'op'},
|
|
'exp_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'log': {'type': 'Log', 'kind': 'op'},
|
|
'log_data': {'shape': reduce_sum_shape, 'kind': 'data', 'value': None},
|
|
'sub_second': {'type': 'Subtract', 'kind': 'op'},
|
|
'sub_second_data': {'shape': flatten_shape, 'kind': 'data', 'value': None},
|
|
'last_shape_val': {'shape': int64_array(shape).shape, 'kind': 'data',
|
|
'value': int64_array(shape)},
|
|
'last_shape': {'type': 'Const', 'kind': 'op', 'shape': len(shape)},
|
|
'last_shape_data': {'shape': int64_array([len(shape)]), 'kind': 'data',
|
|
'value': None},
|
|
'last_reshape': {'kind': 'op', 'type': 'Reshape'},
|
|
'last_reshape_data': {'kind': 'data', 'shape': shape, 'value': None},
|
|
'result': {'kind': 'op', 'type': 'Result'},
|
|
}
|
|
|
|
ref_edges = [
|
|
('input', 'input_data'),
|
|
('flatten_shape_val', 'flatten_shape'),
|
|
('flatten_shape', 'flatten_shape_data'),
|
|
('flatten_shape_data', 'reshape', {'in': 1}),
|
|
('input_data', 'reshape', {'in': 0}),
|
|
('reshape', 'reshape_data'),
|
|
('reduce_max_axis_val', 'reduce_max_axis'),
|
|
('reduce_max_axis', 'reduce_max_axis_data'),
|
|
('reduce_max_axis_data', 'reduce_max', {'in': 1}),
|
|
('reduce_max', 'reduce_max_data'),
|
|
('reshape_data', 'reduce_max', {'out': 0, 'in': 0}),
|
|
('reshape_data', 'sub_first', {'out': 0, 'in': 0}),
|
|
('reduce_max_data', 'sub_first', {'in': 1}),
|
|
('sub_first', 'sub_first_data'),
|
|
('reduce_sum_axis_val', 'reduce_sum_axis'),
|
|
('reduce_sum_axis', 'reduce_sum_axis_data'),
|
|
('reduce_sum_axis_data', 'reduce_sum', {'in': 1}),
|
|
('reduce_sum', 'reduce_sum_data'),
|
|
('sub_first_data', 'exp'),
|
|
('exp', 'exp_data'),
|
|
('exp_data', 'reduce_sum', {'in': 0}),
|
|
('reduce_sum_data', 'log'),
|
|
('log', 'log_data'),
|
|
('log_data', 'sub_second', {'in': 1}),
|
|
('sub_second', 'sub_second_data'),
|
|
('sub_first_data', 'sub_second', {'out': 0, 'in': 0}),
|
|
('last_shape_val', 'last_shape'),
|
|
('last_shape', 'last_shape_data'),
|
|
('last_shape_data', 'last_reshape', {'in': 1}),
|
|
('sub_second_data', 'last_reshape', {'in': 0}),
|
|
('last_reshape', 'last_reshape_data'),
|
|
('last_reshape_data', 'result'),
|
|
]
|
|
|
|
ref_net = build_graph(ref_nodes_attributes, ref_edges)
|
|
return onnx_net, ref_net
|
|
|
|
test_data_precommit = [
|
|
dict(shape=[2, 4], logsoftmax_axis=-1),
|
|
dict(shape=[2, 3, 2, 5, 6], logsoftmax_axis=-2)]
|
|
|
|
test_data = [
|
|
dict(shape=[10, 12], logsoftmax_axis=-1),
|
|
dict(shape=[4, 5, 3], logsoftmax_axis=1),
|
|
dict(shape=[6, 8, 5, 7], logsoftmax_axis=2),
|
|
dict(shape=[2, 3, 2, 5, 6], logsoftmax_axis=-2)]
|
|
|
|
@pytest.mark.parametrize("params", test_data)
|
|
@pytest.mark.nightly
|
|
def test_log(self, params, ie_device, precision, ir_version, temp_dir, use_old_api):
|
|
if ie_device == 'GPU':
|
|
pytest.skip('GREEN_SUITE')
|
|
self._test(*self.create_net(**params, ir_version=ir_version), ie_device, precision,
|
|
ir_version,
|
|
temp_dir=temp_dir, use_old_api=use_old_api)
|