Add exceptions

This commit is contained in:
Ilya Churaev 2020-12-24 18:48:43 +03:00
parent 300a88a644
commit 0489ba376f
162 changed files with 627 additions and 418 deletions

View File

@ -41,7 +41,7 @@ shared_ptr<Node> op::Abs::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Abs>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Abs_clone_with_new_inputs is disabled!");
}
namespace absop
@ -77,10 +77,9 @@ namespace absop
bool op::Abs::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v0_Abs_evaluate)
{
rc = absop::evaluate_abs(inputs[0], outputs[0], shape_size(get_output_shape(0)));
return absop::evaluate_abs(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return rc;
throw ngraph_error("v0_Abs_evaluate is disabled!");
}

View File

@ -50,7 +50,7 @@ shared_ptr<Node> op::Acos::clone_with_new_inputs(const OutputVector& new_args) c
check_new_args_count(this, new_args);
return make_shared<Acos>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Acos_clone_with_new_inputs is disabled!");
}
namespace acosop
@ -85,10 +85,9 @@ namespace acosop
bool op::Acos::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v0_Acos_evaluate)
{
rc = acosop::evaluate_acos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
return acosop::evaluate_acos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return rc;
throw ngraph_error("v0_Acos_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Acosh::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<Acosh>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v3_Acosh_clone_with_new_inputs is disabled!");
}
namespace acoshop
@ -74,7 +74,6 @@ namespace acoshop
bool op::v3::Acosh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v3_Acosh_evaluate) { rc = acoshop::evaluate_acosh(inputs[0], outputs[0]); }
return rc;
NGRAPH_OP_SCOPE(v3_Acosh_evaluate) { return acoshop::evaluate_acosh(inputs[0], outputs[0]); }
throw ngraph_error("v3_Acosh_evaluate is disabled!");
}

View File

@ -86,7 +86,7 @@ bool op::v1::Add::visit_attributes(AttributeVisitor& visitor)
BinaryElementwiseArithmetic::visit_attributes(visitor);
return true;
}
return false;
throw ngraph_error("v1_Add_visit_attributes is disabled!");
}
shared_ptr<Node> op::v1::Add::clone_with_new_inputs(const OutputVector& new_args) const
@ -96,15 +96,14 @@ shared_ptr<Node> op::v1::Add::clone_with_new_inputs(const OutputVector& new_args
check_new_args_count(this, new_args);
return make_shared<op::v1::Add>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Add_clone_with_new_inputs is disabled!");
}
bool op::v1::Add::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v1_Add_evaluate)
{
rc = add::evaluate_add(inputs[0], inputs[1], outputs[0], get_autob());
return add::evaluate_add(inputs[0], inputs[1], outputs[0], get_autob());
}
return rc;
throw ngraph_error("v1_Add_evaluate is disabled!");
}

View File

@ -39,7 +39,7 @@ bool op::v1::LogicalAnd::visit_attributes(AttributeVisitor& visitor)
BinaryElementwiseLogical::visit_attributes(visitor);
return true;
}
return false;
throw ngraph_error("v1_LogicalAnd_visit_attributes is disabled!");
}
shared_ptr<Node> op::v1::LogicalAnd::clone_with_new_inputs(const OutputVector& new_args) const
@ -49,7 +49,7 @@ shared_ptr<Node> op::v1::LogicalAnd::clone_with_new_inputs(const OutputVector& n
check_new_args_count(this, new_args);
return make_shared<v1::LogicalAnd>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_LogicalAnd_clone_with_new_inputs is disabled!");
}
namespace logand
@ -94,10 +94,9 @@ namespace logand
bool op::v1::LogicalAnd::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v1_LogicalAnd_evaluate)
{
rc = logand::evaluate_logand(inputs[0], inputs[1], outputs[0], get_autob());
return logand::evaluate_logand(inputs[0], inputs[1], outputs[0], get_autob());
}
return rc;
throw ngraph_error("v1_LogicalAnd_evaluate is disabled!");
}

View File

@ -51,7 +51,7 @@ shared_ptr<Node> op::Asin::clone_with_new_inputs(const OutputVector& new_args) c
check_new_args_count(this, new_args);
return make_shared<Asin>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Asin_clone_with_new_inputs is disabled!");
}
namespace asinop
@ -86,10 +86,9 @@ namespace asinop
bool op::Asin::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v0_Asin_evaluate)
{
rc = asinop::evaluate_asin(inputs[0], outputs[0], shape_size(get_output_shape(0)));
return asinop::evaluate_asin(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return rc;
throw ngraph_error("v0_Asin_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Asinh::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<Asinh>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v3_Asinh_clone_with_new_inputs is disabled!");
}
namespace asinhop
@ -74,7 +74,6 @@ namespace asinhop
bool op::v3::Asinh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v3_Asinh_evaluate) { rc = asinhop::evaluate_asinh(inputs[0], outputs[0]); }
return rc;
NGRAPH_OP_SCOPE(v3_Asinh_evaluate) { return asinhop::evaluate_asinh(inputs[0], outputs[0]); }
throw ngraph_error("v3_Asinh_evaluate is disabled!");
}

View File

@ -77,7 +77,9 @@ void op::v3::Assign::validate_and_infer_types()
{
set_output_type(0, arg_t, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v3_Assign_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v3::Assign::clone_with_new_inputs(const OutputVector& new_args) const
@ -87,7 +89,7 @@ shared_ptr<Node> op::v3::Assign::clone_with_new_inputs(const OutputVector& new_a
check_new_args_count(this, new_args);
return make_shared<Assign>(new_args.at(0), m_variable_id);
}
return nullptr;
throw ngraph_error("v3_Assign_clone_with_new_inputs is disabled!");
}
bool op::v3::Assign::visit_attributes(AttributeVisitor& visitor)
@ -97,5 +99,5 @@ bool op::v3::Assign::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("variable_id", m_variable_id);
return true;
}
return false;
throw ngraph_error("v3_Assign_visit_attributes is disabled!");
}

View File

@ -50,7 +50,7 @@ shared_ptr<Node> op::Atan::clone_with_new_inputs(const OutputVector& new_args) c
check_new_args_count(this, new_args);
return make_shared<Atan>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Atan_clone_with_new_inputs is disabled!");
}
namespace atanop
@ -85,10 +85,9 @@ namespace atanop
bool op::Atan::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v0_Atan_evaluate)
{
rc = atanop::evaluate_atan(inputs[0], outputs[0], shape_size(get_output_shape(0)));
return atanop::evaluate_atan(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return rc;
throw ngraph_error("v0_Atan_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Atanh::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<Atanh>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v3_Atanh_clone_with_new_inputs is disabled!");
}
namespace atanhop
@ -74,7 +74,6 @@ namespace atanhop
bool op::v3::Atanh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
bool rc = false;
NGRAPH_OP_SCOPE(v3_Atanh_evaluate) { rc = atanhop::evaluate_atanh(inputs[0], outputs[0]); }
return rc;
NGRAPH_OP_SCOPE(v3_Atanh_evaluate) { return atanhop::evaluate_atanh(inputs[0], outputs[0]); }
throw ngraph_error("v3_Atanh_evaluate is disabled!");
}

View File

@ -77,7 +77,7 @@ bool op::v1::AvgPool::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("rounding_type", m_rounding_type);
return true;
}
return false;
throw ngraph_error("v1_AvgPool_visit_attributes is disabled!");
}
void op::v1::AvgPool::validate_and_infer_types()
@ -148,7 +148,9 @@ void op::v1::AvgPool::validate_and_infer_types()
!m_exclude_pad,
m_rounding_type == op::RoundingType::CEIL)
: output_shape);
return;
}
throw ngraph_error("v1_AvgPool_validate_and_infer_types is disabled!");
}
const Shape& op::v1::AvgPool::get_kernel() const
@ -235,7 +237,7 @@ shared_ptr<Node> op::v1::AvgPool::clone_with_new_inputs(const OutputVector& new_
m_rounding_type,
m_auto_pad);
}
return nullptr;
throw ngraph_error("v1_AvgPool_clone_with_new_inputs is disabled!");
}
shared_ptr<Node> op::v1::AvgPool::get_default_value() const

View File

@ -45,7 +45,7 @@ bool op::v0::BatchNormInference::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("epsilon", m_epsilon);
return true;
}
return false;
throw ngraph_error("v0_BatchNormInference_visit_attributes is disabled!");
}
void op::v0::BatchNormInference::validate_and_infer_types()
@ -71,7 +71,9 @@ void op::v0::BatchNormInference::validate_and_infer_types()
get_input_partial_shape(INPUT_VARIANCE));
set_output_type(0, result_et, result_batch_shape);
return;
}
throw ngraph_error("v0_BatchNormInference_validate_and_infer_types is disabled!");
}
std::shared_ptr<Node>
@ -87,7 +89,7 @@ std::shared_ptr<Node>
new_args.at(4),
m_epsilon);
}
return nullptr;
throw ngraph_error("v0_BatchNormInference_clone_with_new_inputs is disabled!");
}
NGRAPH_RTTI_DEFINITION(op::v5::BatchNormInference, "BatchNormInference", 5);
@ -111,7 +113,7 @@ bool op::v5::BatchNormInference::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("epsilon", m_epsilon);
return true;
}
return false;
throw ngraph_error("v5_BatchNormInference_visit_attributes is disabled!");
}
void op::v5::BatchNormInference::validate_and_infer_types()
@ -137,7 +139,9 @@ void op::v5::BatchNormInference::validate_and_infer_types()
get_input_partial_shape(INPUT_VARIANCE));
set_output_type(0, result_et, result_batch_shape);
return;
}
throw ngraph_error("v5_BatchNormInference_validate_and_infer_types is disabled!");
}
std::shared_ptr<Node>
@ -153,5 +157,5 @@ std::shared_ptr<Node>
new_args.at(4),
m_epsilon);
}
return nullptr;
throw ngraph_error("v5_BatchNormInference_clone_with_new_inputs is disabled!");
}

View File

@ -129,7 +129,9 @@ void op::v1::BatchToSpace::validate_and_infer_types()
{
set_output_type(0, data_type, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v1_BatchToSpace_validate_and_infer_types is disabled!");
}
std::shared_ptr<ngraph::Node>
@ -141,13 +143,13 @@ std::shared_ptr<ngraph::Node>
return make_shared<BatchToSpace>(
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3));
}
return nullptr;
throw ngraph_error("v1_BatchToSpace_clone_with_new_inputs is disabled!");
}
bool ngraph::op::v1::BatchToSpace::visit_attributes(ngraph::AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_BatchToSpace_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_BatchToSpace_visit_attributes is disabled!");
}
namespace
@ -268,5 +270,5 @@ bool ngraph::op::v1::BatchToSpace::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v1_BatchToSpace) { return batch_to_space_evaluate(outputs, inputs); }
return false;
throw ngraph_error("v1_BatchToSpace is disabled!");
}

View File

@ -151,7 +151,9 @@ void op::v1::BinaryConvolution::validate_and_infer_types()
m_dilations);
set_output_type(0, data_batch_et, result_shape);
return;
}
throw ngraph_error("v1_BinaryConvolution_validate_and_infer_types is disabled!");
}
shared_ptr<Node>
@ -170,7 +172,7 @@ shared_ptr<Node>
m_pad_value,
m_auto_pad);
}
return nullptr;
throw ngraph_error("v1_BinaryConvolution_clone_with_new_inputs is disabled!");
}
bool op::v1::BinaryConvolution::visit_attributes(AttributeVisitor& visitor)
@ -186,7 +188,7 @@ bool op::v1::BinaryConvolution::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_pad", m_auto_pad);
return true;
}
return false;
throw ngraph_error("v1_BinaryConvolution_visit_attributes is disabled!");
}
namespace ngraph

View File

@ -203,7 +203,9 @@ void op::v3::Broadcast::validate_and_infer_types()
set_input_is_relevant_to_shape(2); // axes_mapping - Broadcast type
}
set_output_type(0, get_input_element_type(0), result_shape);
return;
}
throw ngraph_error("v3_Broadcast_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v3::Broadcast::clone_with_new_inputs(const OutputVector& new_args) const
@ -225,7 +227,7 @@ shared_ptr<Node> op::v3::Broadcast::clone_with_new_inputs(const OutputVector& ne
throw ngraph_error("Not supported number of Broadcast:v3 args");
}
}
return nullptr;
throw ngraph_error("v3_Broadcast_clone_with_new_inputs is disabled!");
}
bool op::v3::Broadcast::visit_attributes(AttributeVisitor& visitor)
@ -235,14 +237,14 @@ bool op::v3::Broadcast::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("mode", m_mode);
return true;
}
return false;
throw ngraph_error("v3_Broadcast_visit_attributes is disabled!");
}
bool op::v3::Broadcast::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v3_Broadcast_evaluate) { return broadcast_evaluate(outputs, inputs); }
return false;
throw ngraph_error("v3_Broadcast_evaluate is disabled!");
}
namespace
@ -315,7 +317,9 @@ void op::v1::Broadcast::validate_and_infer_types()
set_input_is_relevant_to_shape(0); // arg - Result element type
set_input_is_relevant_to_shape(1); // target_shape - Result shape
set_input_is_relevant_to_shape(2); // axes_mapping - Broadcast type
return;
}
throw ngraph_error("v1_Broadcast_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::Broadcast::clone_with_new_inputs(const OutputVector& new_args) const
@ -326,7 +330,7 @@ shared_ptr<Node> op::v1::Broadcast::clone_with_new_inputs(const OutputVector& ne
return make_shared<v1::Broadcast>(
new_args.at(0), new_args.at(1), new_args.at(2), m_broadcast_spec);
}
return nullptr;
throw ngraph_error("v1_Broadcast_clone_with_new_inputs is disabled!");
}
bool op::v1::Broadcast::visit_attributes(AttributeVisitor& visitor)
@ -336,7 +340,7 @@ bool op::v1::Broadcast::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("mode", m_broadcast_spec);
return true;
}
return false;
throw ngraph_error("v1_Broadcast_visit_attributes is disabled!");
}
bool op::v1::Broadcast::evaluate(const HostTensorVector& outputs,
@ -346,5 +350,5 @@ bool op::v1::Broadcast::evaluate(const HostTensorVector& outputs,
{
return op::util::BroadcastBase::evaluate(outputs, inputs);
}
return false;
throw ngraph_error("v1_Broadcast_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ bool op::v3::Bucketize::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("with_right_bound", m_with_right_bound);
return true;
}
return false;
throw ngraph_error("v3_Bucketize_visit_attributes is disabled!");
}
void op::v3::Bucketize::validate_and_infer_types()
@ -68,7 +68,9 @@ void op::v3::Bucketize::validate_and_infer_types()
set_output_size(1);
set_output_type(0, m_output_type, data_pshape);
return;
}
throw ngraph_error("v3_Bucketize_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v3::Bucketize::clone_with_new_inputs(const OutputVector& inputs) const
@ -80,5 +82,5 @@ shared_ptr<Node> op::v3::Bucketize::clone_with_new_inputs(const OutputVector& in
return make_shared<v3::Bucketize>(
inputs.at(0), inputs.at(1), m_output_type, m_with_right_bound);
}
return nullptr;
throw ngraph_error("v3_Bucketize_clone_with_new_inputs is disabled!");
}

View File

@ -39,7 +39,7 @@ shared_ptr<Node> op::Ceiling::clone_with_new_inputs(const OutputVector& new_args
check_new_args_count(this, new_args);
return make_shared<Ceiling>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Ceiling_clone_with_new_inputs is disabled!");
}
namespace ceiling
@ -91,5 +91,5 @@ bool op::Ceiling::evaluate(const HostTensorVector& outputs, const HostTensorVect
{
return ceiling::evaluate_ceiling(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Ceiling_evaluate is disabled!");
}

View File

@ -91,7 +91,7 @@ bool op::v0::Clamp::evaluate(const HostTensorVector& outputs, const HostTensorVe
return clamp::evaluate_clamp(
inputs[0], outputs[0], get_min(), get_max(), shape_size(get_input_shape(0)));
}
return false;
throw ngraph_error("v0_Clamp_evaluate is disabled!");
}
NGRAPH_RTTI_DEFINITION(op::v0::Clamp, "Clamp", 0);
@ -239,7 +239,7 @@ shared_ptr<Node> op::Clamp::clone_with_new_inputs(const OutputVector& new_args)
return make_shared<Clamp>(new_args.at(0), m_min, m_max);
}
return nullptr;
throw ngraph_error("v0_Clamp_clone_with_new_inputs is disabled!");
}
bool op::Clamp::visit_attributes(AttributeVisitor& visitor)
@ -250,5 +250,5 @@ bool op::Clamp::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("max", m_max);
return true;
}
return false;
throw ngraph_error("v0_Clamp_visit_attributes is disabled!");
}

View File

@ -46,7 +46,7 @@ bool op::Concat::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("axis", m_axis);
return true;
}
return false;
throw ngraph_error("v0_Concat_visit_attributes is disabled!");
}
void op::Concat::validate_and_infer_types()
@ -114,7 +114,9 @@ void op::Concat::validate_and_infer_types()
{
set_output_type(0, inputs_et, PartialShape::dynamic(concatenation_axis_output_dim));
}
return;
}
throw ngraph_error("v0_Concat_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::Concat::clone_with_new_inputs(const OutputVector& new_args) const
@ -124,7 +126,7 @@ shared_ptr<Node> op::Concat::clone_with_new_inputs(const OutputVector& new_args)
// TODO(amprocte): Should we check the new_args count here?
return make_shared<Concat>(new_args, m_axis);
}
return nullptr;
throw ngraph_error("v0_Concat_clone_with_new_inputs is disabled!");
}
namespace
@ -162,5 +164,5 @@ bool op::Concat::evaluate(const HostTensorVector& outputs, const HostTensorVecto
auto concat_axis = get_axis() < 0 ? get_axis() + inputs[0]->get_shape().size() : get_axis();
return evaluate_concat(inputs, outputs[0], concat_axis);
}
return false;
throw ngraph_error("v0_Concat_evaluate is disabled!");
}

View File

@ -554,7 +554,7 @@ shared_ptr<Node> op::Constant::clone_with_new_inputs(const OutputVector& new_arg
check_new_args_count(this, new_args);
return make_shared<Constant>(*this);
}
return nullptr;
throw ngraph_error("v0_Constant_clone_with_new_inputs is disabled!");
}
template <typename T>
@ -641,7 +641,7 @@ bool op::v0::Constant::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("value", m_data);
return true;
}
return false;
throw ngraph_error("v0_Constant_visit_attributes is disabled!");
}
bool op::v0::Constant::evaluate(const HostTensorVector& outputs,
@ -653,7 +653,7 @@ bool op::v0::Constant::evaluate(const HostTensorVector& outputs,
output->write(get_data_ptr(), output->get_size_in_bytes());
return true;
}
return false;
throw ngraph_error("v0_Constant_evaluate is disabled!");
}
//

View File

@ -37,7 +37,9 @@ void op::Convert::validate_and_infer_types()
NGRAPH_OP_SCOPE(v0_Convert_validate_and_infer_types)
{
set_output_type(0, m_destination_type, get_input_partial_shape(0));
return;
}
throw ngraph_error("v0_Convert_validate_and_infer_types is disabled!");
}
bool op::Convert::visit_attributes(AttributeVisitor& visitor)
@ -47,7 +49,7 @@ bool op::Convert::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("destination_type", m_destination_type);
return true;
}
return false;
throw ngraph_error("v0_Convert_visit_attributes is disabled!");
}
shared_ptr<Node> op::Convert::clone_with_new_inputs(const OutputVector& new_args) const
@ -57,7 +59,7 @@ shared_ptr<Node> op::Convert::clone_with_new_inputs(const OutputVector& new_args
check_new_args_count(this, new_args);
return make_shared<Convert>(new_args.at(0), m_destination_type);
}
return nullptr;
throw ngraph_error("v0_Convert_clone_with_new_inputs is disabled!");
}
namespace convert
@ -134,5 +136,5 @@ bool op::v0::Convert::evaluate(const HostTensorVector& output_values,
{
return convert::evaluate_convert(input_values[0], output_values[0]);
}
return false;
throw ngraph_error("v0_Convert_evaluate is disabled!");
}

View File

@ -35,13 +35,15 @@ void op::v1::ConvertLike::validate_and_infer_types()
NGRAPH_OP_SCOPE(v1_ConvertLike_validate_and_infer_types)
{
set_output_type(0, get_input_element_type(1), get_input_partial_shape(0));
return;
}
throw ngraph_error("v1_ConvertLike_validate_and_infer_types is disabled!");
}
bool op::v1::ConvertLike::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_ConvertLike_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_ConvertLike_visit_attributes is disabled!");
}
shared_ptr<Node> op::v1::ConvertLike::clone_with_new_inputs(const OutputVector& new_args) const
@ -51,5 +53,5 @@ shared_ptr<Node> op::v1::ConvertLike::clone_with_new_inputs(const OutputVector&
check_new_args_count(this, new_args);
return make_shared<ConvertLike>(new_args.at(0), new_args.at(1));
}
return nullptr;
throw ngraph_error("v1_ConvertLike_clone_with_new_inputs is disabled!");
}

View File

@ -56,7 +56,7 @@ bool op::v1::Convolution::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_pad", m_auto_pad);
return true;
}
return false;
throw ngraph_error("v1_Convolution_visit_attributes is disabled!");
}
void op::v1::Convolution::validate_and_infer_types()
@ -149,7 +149,9 @@ void op::v1::Convolution::validate_and_infer_types()
m_dilations);
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v1_Convolution_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::Convolution::clone_with_new_inputs(const OutputVector& new_args) const
@ -165,7 +167,7 @@ shared_ptr<Node> op::v1::Convolution::clone_with_new_inputs(const OutputVector&
m_dilations,
m_auto_pad);
}
return nullptr;
throw ngraph_error("v1_Convolution_clone_with_new_inputs is disabled!");
}
constexpr NodeTypeInfo op::v1::ConvolutionBackpropData::type_info;
@ -206,7 +208,7 @@ bool op::v1::ConvolutionBackpropData::visit_attributes(AttributeVisitor& visitor
visitor.on_attribute("output_padding", m_output_padding);
return true;
}
return false;
throw ngraph_error("v1_ConvolutionBackpropData_visit_attributes is disabled!");
}
op::v1::ConvolutionBackpropData::ConvolutionBackpropData(const Output<Node>& data,
@ -447,7 +449,9 @@ void op::v1::ConvolutionBackpropData::validate_and_infer_types()
set_input_is_relevant_to_shape(0);
set_input_is_relevant_to_shape(1);
set_output_type(0, result_et, output_pshape);
return;
}
throw ngraph_error("v1_ConvolutionBackpropData_validate_and_infer_types is disabled!");
}
shared_ptr<Node>
@ -480,5 +484,5 @@ shared_ptr<Node>
m_output_padding);
}
}
return nullptr;
throw ngraph_error("v1_ConvolutionBackpropData_clone_with_new_inputs is disabled!");
}

View File

@ -38,7 +38,7 @@ op::Cos::Cos(const Output<Node>& arg)
bool op::Cos::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Cos_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Cos_visit_attributes is disabled!");
}
shared_ptr<Node> op::Cos::clone_with_new_inputs(const OutputVector& new_args) const
@ -48,7 +48,7 @@ shared_ptr<Node> op::Cos::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Cos>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Cos_clone_with_new_inputs is disabled!");
}
namespace cosop
@ -87,5 +87,5 @@ bool op::Cos::evaluate(const HostTensorVector& outputs, const HostTensorVector&
{
return cosop::evaluate_cos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Cos_evaluate is disabled!");
}

View File

@ -37,7 +37,7 @@ op::Cosh::Cosh(const Output<Node>& arg)
bool op::Cosh::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Cosh_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Cosh_visit_attributes is disabled!");
}
shared_ptr<Node> op::Cosh::clone_with_new_inputs(const OutputVector& new_args) const
@ -47,7 +47,7 @@ shared_ptr<Node> op::Cosh::clone_with_new_inputs(const OutputVector& new_args) c
check_new_args_count(this, new_args);
return make_shared<Cosh>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Cosh_clone_with_new_inputs is disabled!");
}
namespace coshop
@ -86,5 +86,5 @@ bool op::Cosh::evaluate(const HostTensorVector& outputs, const HostTensorVector&
{
return coshop::evaluate_cosh(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Cosh_evaluate is disabled!");
}

View File

@ -98,7 +98,9 @@ void op::CTCGreedyDecoder::validate_and_infer_types()
}
}
set_output_type(0, input_et, PartialShape{batch_size, time_size, 1, 1});
return;
}
throw ngraph_error("v0_CTCGreedyDecoder_validate_and_infer_types is disabled!");
}
bool op::CTCGreedyDecoder::visit_attributes(AttributeVisitor& visitor)
@ -108,7 +110,7 @@ bool op::CTCGreedyDecoder::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("ctc_merge_repeated", m_ctc_merge_repeated);
return true;
}
return false;
throw ngraph_error("v0_CTCGreedyDecoder_visit_attributes is disabled!");
}
shared_ptr<Node> op::CTCGreedyDecoder::clone_with_new_inputs(const OutputVector& new_args) const
@ -118,5 +120,5 @@ shared_ptr<Node> op::CTCGreedyDecoder::clone_with_new_inputs(const OutputVector&
check_new_args_count(this, new_args);
return make_shared<CTCGreedyDecoder>(new_args.at(0), new_args.at(1), m_ctc_merge_repeated);
}
return nullptr;
throw ngraph_error("v0_CTCGreedyDecoder_clone_with_new_inputs is disabled!");
}

View File

@ -229,7 +229,9 @@ void op::v4::CTCLoss::validate_and_infer_types()
{
set_output_type(0, logits_type, PartialShape{Dimension::dynamic()});
}
return;
}
throw ngraph_error("v4_CTCLoss_validate_and_infer_types is disabled!");
}
bool op::v4::CTCLoss::visit_attributes(AttributeVisitor& visitor)
@ -241,7 +243,7 @@ bool op::v4::CTCLoss::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("unique", unique_);
return true;
}
return false;
throw ngraph_error("v4_CTCLoss_visit_attributes is disabled!");
}
shared_ptr<Node> op::v4::CTCLoss::clone_with_new_inputs(const OutputVector& new_args) const
@ -275,5 +277,5 @@ shared_ptr<Node> op::v4::CTCLoss::clone_with_new_inputs(const OutputVector& new_
throw ngraph_error("Incorrect number of arguments");
}
}
return nullptr;
throw ngraph_error("v4_CTCLoss_clone_with_new_inputs is disabled!");
}

View File

@ -53,7 +53,7 @@ bool op::v0::CumSum::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("reverse", m_reverse);
return true;
}
return false;
throw ngraph_error("v0_CumSum_visit_attributes is disabled!");
}
void op::v0::CumSum::validate_and_infer_types()
@ -76,7 +76,9 @@ void op::v0::CumSum::validate_and_infer_types()
"axis element type must be either int64_t or int32_t but got (",
axis_type,
").");
return;
}
throw ngraph_error("v0_CumSum_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v0::CumSum::clone_with_new_inputs(const OutputVector& new_args) const
@ -86,7 +88,7 @@ shared_ptr<Node> op::v0::CumSum::clone_with_new_inputs(const OutputVector& new_a
check_new_args_count(this, new_args);
return make_shared<op::CumSum>(new_args.at(0), new_args.at(1), m_exclusive, m_reverse);
}
return nullptr;
throw ngraph_error("v0_CumSum_clone_with_new_inputs is disabled!");
}
shared_ptr<Node> op::v0::CumSum::get_default_value() const

View File

@ -62,7 +62,7 @@ bool op::v1::DeformableConvolution::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("deformable_group", m_deformable_group);
return true;
}
return false;
throw ngraph_error("v1_DeformableConvolution_visit_attributes is disabled!");
}
void op::v1::DeformableConvolution::validate_and_infer_types()
@ -201,7 +201,9 @@ void op::v1::DeformableConvolution::validate_and_infer_types()
m_dilations);
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v1_DeformableConvolution_validate_and_infer_types is disabled!");
}
shared_ptr<Node>
@ -221,5 +223,5 @@ shared_ptr<Node>
m_group,
m_deformable_group);
}
return nullptr;
throw ngraph_error("v1_DeformableConvolution_clone_with_new_inputs is disabled!");
}

View File

@ -83,7 +83,7 @@ bool op::v1::DeformablePSROIPooling::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("part_size", m_part_size);
return true;
}
return false;
throw ngraph_error("v1_DeformablePSROIPooling_visit_attributes is disabled!");
}
void op::v1::DeformablePSROIPooling::validate_and_infer_types()
@ -131,7 +131,9 @@ void op::v1::DeformablePSROIPooling::validate_and_infer_types()
}
set_output_type(0, input_et, PartialShape(output_dim_vec));
return;
}
throw ngraph_error("v1_DeformablePSROIPooling_validate_and_infer_types is disabled!");
}
shared_ptr<Node>
@ -172,5 +174,5 @@ shared_ptr<Node>
throw ngraph_error("Not supported number of DeformablePSROIPooling args");
}
}
return nullptr;
throw ngraph_error("v1_DeformablePSROIPooling_clone_with_new_inputs is disabled!");
}

View File

@ -60,7 +60,7 @@ bool op::DepthToSpace::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("mode", m_mode);
return true;
}
return false;
throw ngraph_error("v0_DepthToSpace_visit_attributes is disabled!");
}
shared_ptr<Node> op::DepthToSpace::clone_with_new_inputs(const OutputVector& new_args) const
@ -73,7 +73,7 @@ shared_ptr<Node> op::DepthToSpace::clone_with_new_inputs(const OutputVector& new
}
return make_shared<DepthToSpace>(new_args.at(0), m_mode, m_blocksize);
}
return nullptr;
throw ngraph_error("v0_DepthToSpace_clone_with_new_inputs is disabled!");
}
void op::DepthToSpace::validate_and_infer_types()
@ -121,7 +121,9 @@ void op::DepthToSpace::validate_and_infer_types()
{
set_output_type(0, data_type, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_DepthToSpace_validate_and_infer_types is disabled!");
}
bool op::DepthToSpace::evaluate_depth_to_space(const HostTensorVector& outputs,
@ -255,7 +257,7 @@ bool op::DepthToSpace::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v0_DepthToSpace_evaluate) { return evaluate_depth_to_space(outputs, inputs); }
return false;
throw ngraph_error("v0_DepthToSpace_evaluate is disabled!");
}
namespace ngraph
{

View File

@ -268,7 +268,9 @@ void op::DetectionOutput::validate_and_infer_types()
output_shape.push_back(7);
set_output_type(0, box_logits_et, output_shape);
return;
}
throw ngraph_error("v0_DetectionOutput_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::DetectionOutput::clone_with_new_inputs(const OutputVector& new_args) const
@ -297,7 +299,7 @@ shared_ptr<Node> op::DetectionOutput::clone_with_new_inputs(const OutputVector&
m_attrs);
}
}
return nullptr;
throw ngraph_error("v0_DetectionOutput_clone_with_new_inputs is disabled!");
}
bool op::DetectionOutput::visit_attributes(AttributeVisitor& visitor)
@ -322,5 +324,5 @@ bool op::DetectionOutput::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("objectness_score", m_attrs.objectness_score);
return true;
}
return false;
throw ngraph_error("v0_DetectionOutput_visit_attributes is disabled!");
}

View File

@ -98,7 +98,7 @@ bool op::v1::Divide::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("m_pythondiv", m_pythondiv);
return true;
}
return false;
throw ngraph_error("v1_Divide_visit_attributes is disabled!");
}
shared_ptr<Node> op::v1::Divide::clone_with_new_inputs(const OutputVector& new_args) const
@ -109,7 +109,7 @@ shared_ptr<Node> op::v1::Divide::clone_with_new_inputs(const OutputVector& new_a
return make_shared<op::v1::Divide>(
new_args.at(0), new_args.at(1), this->is_pythondiv(), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Divide_clone_with_new_inputs is disabled!");
}
bool op::v1::Divide::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
@ -119,5 +119,5 @@ bool op::v1::Divide::evaluate(const HostTensorVector& outputs, const HostTensorV
return divide::evaluate_divide(
inputs[0], inputs[1], outputs[0], get_autob(), is_pythondiv());
}
return false;
throw ngraph_error("v1_Divide_evaluate is disabled!");
}

View File

@ -39,7 +39,7 @@ bool ngraph::op::v0::Elu::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("alpha", m_alpha);
return true;
}
return false;
throw ngraph_error("v0_Elu_visit_attributes is disabled!");
}
void op::v0::Elu::validate_and_infer_types()
@ -48,7 +48,9 @@ void op::v0::Elu::validate_and_infer_types()
{
set_output_size(1);
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
return;
}
throw ngraph_error("v0_Elu_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::Elu::clone_with_new_inputs(const OutputVector& new_args) const
@ -58,5 +60,5 @@ shared_ptr<Node> op::Elu::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Elu>(new_args.at(0), m_alpha);
}
return nullptr;
throw ngraph_error("v0_Elu_clone_with_new_inputs is disabled!");
}

View File

@ -180,7 +180,9 @@ void op::v3::EmbeddingSegmentsSum::validate_and_infer_types()
}
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v3_EmbeddingSegmentsSum_validate_and_infer_types is disabled!");
}
shared_ptr<Node>
@ -213,5 +215,5 @@ shared_ptr<Node>
throw ngraph_error("Incorrect number of arguments");
}
}
return nullptr;
throw ngraph_error("v3_EmbeddingSegmentsSum_clone_with_new_inputs is disabled!");
}

View File

@ -73,5 +73,5 @@ shared_ptr<Node>
throw ngraph_error("Incorrect number of arguments");
}
}
return nullptr;
throw ngraph_error("v3_EmbeddingBagOffsetsSum_clone_with_new_inputs is disabled!");
}

View File

@ -56,5 +56,5 @@ shared_ptr<Node>
throw ngraph_error("Incorrect number of arguments");
}
}
return nullptr;
throw ngraph_error("v3_EmbeddingBagPackedSum_clone_with_new_inputs is disabled!");
}

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Equal::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<op::v1::Equal>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Equal_clone_with_new_inputs is disabled!");
}
bool op::v1::Equal::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
@ -91,5 +91,5 @@ bool op::v1::Equal::evaluate(const HostTensorVector& outputs, const HostTensorVe
{
return equal::evaluate_equal(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Equal_evaluate is disabled!");
}

View File

@ -30,7 +30,7 @@ constexpr NodeTypeInfo op::Erf::type_info;
bool ngraph::op::v0::Erf::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Erf_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Erf_visit_attributes is disabled!");
}
shared_ptr<Node> op::Erf::clone_with_new_inputs(const OutputVector& new_args) const
@ -40,7 +40,7 @@ shared_ptr<Node> op::Erf::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Erf>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Erf_clone_with_new_inputs is disabled!");
}
op::Erf::Erf(const Output<Node>& arg)
@ -85,5 +85,5 @@ bool op::Erf::evaluate(const HostTensorVector& outputs, const HostTensorVector&
{
return erfop::evaluate_erf(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Erf_evaluate is disabled!");
}

View File

@ -36,7 +36,7 @@ op::Exp::Exp(const Output<Node>& arg)
bool ngraph::op::v0::Exp::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Exp_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Exp_visit_attributes is disabled!");
}
shared_ptr<Node> op::Exp::clone_with_new_inputs(const OutputVector& new_args) const
@ -46,7 +46,7 @@ shared_ptr<Node> op::Exp::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Exp>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Exp_clone_with_new_inputs is disabled!");
}
namespace expop
@ -85,5 +85,5 @@ bool op::Exp::evaluate(const HostTensorVector& outputs, const HostTensorVector&
{
return expop::evaluate_exp(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Exp_evaluate is disabled!");
}

View File

@ -148,7 +148,9 @@ void op::v3::ExtractImagePatches::validate_and_infer_types()
set_output_type(0, get_input_element_type(0), output_Pshape);
}
return;
}
throw ngraph_error("v3_ExtractImagePatches_validate_and_infer_types is disabled!");
}
bool op::v3::ExtractImagePatches::visit_attributes(AttributeVisitor& visitor)
@ -161,7 +163,7 @@ bool op::v3::ExtractImagePatches::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_pad", m_padding);
return true;
}
return false;
throw ngraph_error("v3_ExtractImagePatches_visit_attributes is disabled!");
}
shared_ptr<Node>
@ -176,5 +178,5 @@ shared_ptr<Node>
m_patch_selection_rates,
m_padding);
}
return nullptr;
throw ngraph_error("v3_ExtractImagePatches_clone_with_new_inputs is disabled!");
}

View File

@ -85,7 +85,9 @@ void op::FakeQuantize::validate_and_infer_types()
}
}
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
return;
}
throw ngraph_error("v0_FakeQuantize_validate_and_infer_types is disabled!");
}
bool ngraph::op::v0::FakeQuantize::visit_attributes(AttributeVisitor& visitor)
@ -96,7 +98,7 @@ bool ngraph::op::v0::FakeQuantize::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_broadcast", m_auto_broadcast);
return true;
}
return false;
throw ngraph_error("v0_FakeQuantize_visit_attributes is disabled!");
}
OutputVector op::FakeQuantize::decompose_op() const
@ -186,5 +188,5 @@ shared_ptr<Node> op::FakeQuantize::clone_with_new_inputs(const OutputVector& new
m_levels,
m_auto_broadcast);
}
return nullptr;
throw ngraph_error("v0_FakeQuantize_clone_with_new_inputs is disabled!");
}

View File

@ -35,7 +35,7 @@ op::Floor::Floor(const Output<Node>& arg)
bool ngraph::op::v0::Floor::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Floor_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Floor_visit_attributes is disabled!");
}
shared_ptr<Node> op::Floor::clone_with_new_inputs(const OutputVector& new_args) const
@ -45,7 +45,7 @@ shared_ptr<Node> op::Floor::clone_with_new_inputs(const OutputVector& new_args)
check_new_args_count(this, new_args);
return make_shared<Floor>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Floor_clone_with_new_inputs is disabled!");
}
namespace floorop
@ -97,5 +97,5 @@ bool op::Floor::evaluate(const HostTensorVector& outputs, const HostTensorVector
{
return floorop::evaluate_floor(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Floor_evaluate is disabled!");
}

View File

@ -39,7 +39,7 @@ shared_ptr<Node> op::v1::FloorMod::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<FloorMod>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_FloorMod_clone_with_new_inputs is disabled!");
}
namespace floor_mod
@ -90,11 +90,11 @@ bool op::v1::FloorMod::evaluate(const HostTensorVector& outputs,
{
return floor_mod::evaluate_floor_mod(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_FloorMod_evaluate is disabled!");
}
bool op::v1::FloorMod::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_FloorMod_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_FloorMod_visit_attributes is disabled!");
}

View File

@ -48,7 +48,7 @@ op::v1::Gather::Gather(const Output<Node>& params,
bool ngraph::op::v1::Gather::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_Gather_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_Gather_visit_attributes is disabled!");
}
void op::v1::Gather::validate_and_infer_types()
@ -115,7 +115,9 @@ void op::v1::Gather::validate_and_infer_types()
}
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v1_Gather_validate_and_infer_types is disabled!");
}
int64_t op::v1::Gather::get_axis() const
@ -145,7 +147,7 @@ shared_ptr<Node> op::v1::Gather::clone_with_new_inputs(const OutputVector& new_a
return make_shared<v1::Gather>(
new_args.at(PARAMS), new_args.at(INDICES), new_args.at(AXIS));
}
return nullptr;
throw ngraph_error("v1_Gather_clone_with_new_inputs is disabled!");
}
namespace gather
@ -323,7 +325,7 @@ bool op::v1::Gather::evaluate_gather(const HostTensorVector& outputs,
bool op::v1::Gather::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v1_Gather_evaluate) { return evaluate_gather(outputs, inputs); }
return false;
throw ngraph_error("v1_Gather_evaluate is disabled!");
}
bool op::v1::Gather::constant_fold(OutputVector& output_values, const OutputVector& input_values)

View File

@ -119,7 +119,9 @@ void op::v6::GatherElements::validate_and_infer_types()
}
}
set_output_type(0, data_type, output_pshape);
return;
}
throw ngraph_error("v6_GatherElements_validate_and_infer_types is disabled!");
}
bool op::v6::GatherElements::visit_attributes(AttributeVisitor& visitor)
@ -129,7 +131,7 @@ bool op::v6::GatherElements::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("axis", m_axis);
return true;
}
return false;
throw ngraph_error("v6_GatherElements_visit_attributes is disabled!");
}
shared_ptr<Node> op::v6::GatherElements::clone_with_new_inputs(const OutputVector& new_args) const
@ -139,5 +141,5 @@ shared_ptr<Node> op::v6::GatherElements::clone_with_new_inputs(const OutputVecto
check_new_args_count(this, new_args);
return make_shared<op::v6::GatherElements>(new_args.at(0), new_args.at(1), m_axis);
}
return nullptr;
throw ngraph_error("v6_GatherElements_clone_with_new_inputs is disabled!");
}

View File

@ -149,7 +149,9 @@ void op::v5::GatherND::validate_and_infer_types()
{
set_output_type(0, data_type, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v5_GatherND_validate_and_infer_types is disabled!");
}
bool op::v5::GatherND::visit_attributes(AttributeVisitor& visitor)
@ -159,7 +161,7 @@ bool op::v5::GatherND::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("batch_dims", m_batch_dims);
return true;
}
return false;
throw ngraph_error("v5_GatherND_visit_attributes is disabled!");
}
shared_ptr<Node> op::v5::GatherND::clone_with_new_inputs(const OutputVector& new_args) const
@ -169,5 +171,5 @@ shared_ptr<Node> op::v5::GatherND::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v5::GatherND>(new_args.at(0), new_args.at(1), m_batch_dims);
}
return nullptr;
throw ngraph_error("v5_GatherND_clone_with_new_inputs is disabled!");
}

View File

@ -40,13 +40,13 @@ shared_ptr<Node> op::v1::GatherTree::clone_with_new_inputs(const OutputVector& n
return make_shared<v1::GatherTree>(
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3));
}
return nullptr;
throw ngraph_error("v1_GatherTree_clone_with_new_inputs is disabled!");
}
bool ngraph::op::v1::GatherTree::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_GatherTree_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_GatherTree_visit_attributes is disabled!");
}
void op::v1::GatherTree::validate_and_infer_types()
@ -88,5 +88,7 @@ void op::v1::GatherTree::validate_and_infer_types()
const auto& step_ids_et = get_input_element_type(0);
set_output_type(0, step_ids_et, step_ids_rank);
return;
}
throw ngraph_error("v1_GatherTree_validate_and_infer_types is disabled!");
}

View File

@ -43,7 +43,7 @@ op::Gelu::Gelu(const Output<Node>& data)
bool ngraph::op::v0::Gelu::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Gelu_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Gelu_visit_attributes is disabled!");
}
// f(x) = 0.5 * x * (1.0 + erf( x / sqrt(2.0) )
@ -77,7 +77,7 @@ shared_ptr<Node> op::Gelu::clone_with_new_inputs(const OutputVector& new_args) c
}
return make_shared<Gelu>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Gelu_clone_with_new_inputs is disabled!");
}
void op::Gelu::pre_validate_and_infer_types()

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Greater::clone_with_new_inputs(const OutputVector& new_
check_new_args_count(this, new_args);
return make_shared<op::v1::Greater>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Greater_clone_with_new_inputs is disabled!");
}
bool op::v1::Greater::evaluate(const HostTensorVector& outputs,
@ -92,5 +92,5 @@ bool op::v1::Greater::evaluate(const HostTensorVector& outputs,
{
return greaterop::evaluate_greater(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Greater_evaluate is disabled!");
}

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::GreaterEqual::clone_with_new_inputs(const OutputVector&
check_new_args_count(this, new_args);
return make_shared<op::v1::GreaterEqual>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_GreaterEqual_clone_with_new_inputs is disabled!");
}
bool op::v1::GreaterEqual::evaluate(const HostTensorVector& outputs,
@ -93,5 +93,5 @@ bool op::v1::GreaterEqual::evaluate(const HostTensorVector& outputs,
return greater_equalop::evaluate_greater_equal(
inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_GreaterEqual_evaluate is disabled!");
}

View File

@ -49,7 +49,7 @@ bool ngraph::op::v0::GRN::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("bias", m_bias);
return true;
}
return false;
throw ngraph_error("v0_GRN_visit_attributes is disabled!");
}
void op::GRN::pre_validate_and_infer_types()
@ -109,5 +109,5 @@ shared_ptr<Node> op::GRN::clone_with_new_inputs(const OutputVector& new_args) co
}
return make_shared<GRN>(new_args.at(0), m_bias);
}
return nullptr;
throw ngraph_error("v0_GRN_clone_with_new_inputs is disabled!");
}

View File

@ -70,7 +70,7 @@ bool ngraph::op::v1::GroupConvolution::visit_attributes(AttributeVisitor& visito
visitor.on_attribute("auto_pad", m_auto_pad);
return true;
}
return false;
throw ngraph_error("v1_GroupConvolution_visit_attributes is disabled!");
}
void op::v1::GroupConvolution::validate_and_infer_types()
@ -179,7 +179,9 @@ void op::v1::GroupConvolution::validate_and_infer_types()
m_dilations);
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v1_GroupConvolution_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::GroupConvolution::clone_with_new_inputs(const OutputVector& new_args) const
@ -195,7 +197,7 @@ shared_ptr<Node> op::v1::GroupConvolution::clone_with_new_inputs(const OutputVec
m_dilations,
m_auto_pad);
}
return nullptr;
throw ngraph_error("v1_GroupConvolution_clone_with_new_inputs is disabled!");
}
//------------------------------------------------------------------------------
@ -277,7 +279,7 @@ bool ngraph::op::v1::GroupConvolutionBackpropData::visit_attributes(AttributeVis
visitor.on_attribute("output_padding", m_output_padding);
return true;
}
return false;
throw ngraph_error("v1_GroupConvolutionBackpropData_visit_attributes is disabled!");
}
bool op::v1::GroupConvolutionBackpropData::is_dynamic() const
@ -596,5 +598,5 @@ shared_ptr<Node>
m_output_padding);
}
}
return nullptr;
throw ngraph_error("v1_GroupConvolutionBackpropData_clone_with_new_inputs is disabled!");
}

View File

@ -108,7 +108,7 @@ bool op::v3::GRUCell::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("linear_before_reset", m_linear_before_reset);
return op::util::RNNCellBase::visit_attributes(visitor);
}
return false;
throw ngraph_error("v3_GRUCell_visit_attributes is disabled!");
}
void op::v3::GRUCell::validate_and_infer_types()
@ -209,7 +209,9 @@ void op::v3::GRUCell::validate_and_infer_types()
// Set output size, type and shape
set_output_size(1);
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
return;
}
throw ngraph_error("v3_GRUCell_validate_and_infer_types is disabled!");
}
void op::v3::GRUCell::add_default_bias_input()
@ -258,5 +260,5 @@ shared_ptr<Node> op::v3::GRUCell::clone_with_new_inputs(const OutputVector& new_
throw ngraph_error("Incorrect number of new arguments");
}
}
return nullptr;
throw ngraph_error("v3_GRUCell_clone_with_new_inputs is disabled!");
}

View File

@ -179,7 +179,9 @@ void op::v5::GRUSequence::validate_and_infer_types()
{merged_batch_size, merged_num_directions, x_pshape[1], merged_hidden_size});
set_output_type(
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
return;
}
throw ngraph_error("v5_GRUSequence_validate_and_infer_types is disabled!");
}
bool op::v5::GRUSequence::visit_attributes(AttributeVisitor& visitor)
@ -190,7 +192,7 @@ bool op::v5::GRUSequence::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("linear_before_reset", m_linear_before_reset);
return op::util::RNNCellBase::visit_attributes(visitor);
}
return false;
throw ngraph_error("v5_GRUSequence_visit_attributes is disabled!");
}
shared_ptr<Node> op::v5::GRUSequence::clone_with_new_inputs(const OutputVector& new_args) const
@ -212,5 +214,5 @@ shared_ptr<Node> op::v5::GRUSequence::clone_with_new_inputs(const OutputVector&
m_clip,
m_linear_before_reset);
}
return nullptr;
throw ngraph_error("v5_GRUSequence_clone_with_new_inputs is disabled!");
}

View File

@ -43,7 +43,7 @@ op::HardSigmoid::HardSigmoid(const Output<Node>& data,
bool ngraph::op::v0::HardSigmoid::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_HardSigmoid_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_HardSigmoid_visit_attributes is disabled!");
}
void op::HardSigmoid::pre_validate_and_infer_types()
@ -110,5 +110,5 @@ shared_ptr<Node> op::HardSigmoid::clone_with_new_inputs(const OutputVector& new_
return make_shared<HardSigmoid>(new_args.at(0), new_args.at(1), new_args.at(2));
}
return nullptr;
throw ngraph_error("v0_HardSigmoid_clone_with_new_inputs is disabled!");
}

View File

@ -36,7 +36,7 @@ op::v5::HSigmoid::HSigmoid(const Output<Node>& arg)
bool op::v5::HSigmoid::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v5_HSigmoid_visit_attributes) { return true; }
return false;
throw ngraph_error("v5_HSigmoid_visit_attributes is disabled!");
}
shared_ptr<Node> op::v5::HSigmoid::clone_with_new_inputs(const OutputVector& new_args) const
@ -45,7 +45,7 @@ shared_ptr<Node> op::v5::HSigmoid::clone_with_new_inputs(const OutputVector& new
{
return make_shared<op::v5::HSigmoid>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v5_HSigmoid_clone_with_new_inputs is disabled!");
}
namespace
@ -82,5 +82,5 @@ bool op::v5::HSigmoid::evaluate(const HostTensorVector& outputs,
{
return evaluate_hsigmoid(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v5_HSigmoid_evaluate is disabled!");
}

View File

@ -36,7 +36,7 @@ op::v4::HSwish::HSwish(const Output<Node>& arg)
bool op::v4::HSwish::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v4_HSwish_visit_attributes) { return true; }
return false;
throw ngraph_error("v4_HSwish_visit_attributes is disabled!");
}
shared_ptr<Node> op::v4::HSwish::clone_with_new_inputs(const OutputVector& new_args) const
@ -45,7 +45,7 @@ shared_ptr<Node> op::v4::HSwish::clone_with_new_inputs(const OutputVector& new_a
{
return make_shared<op::v4::HSwish>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v4_HSwish_clone_with_new_inputs is disabled!");
}
namespace hswish
@ -81,5 +81,5 @@ bool op::v4::HSwish::evaluate(const HostTensorVector& outputs, const HostTensorV
{
return hswish::evaluate_hswish(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v4_HSwish_evaluate is disabled!");
}

View File

@ -49,7 +49,7 @@ bool op::v0::Interpolate::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("pads_end", m_attrs.pads_end);
return true;
}
return false;
throw ngraph_error("v0_Interpolate_visit_attributes is disabled!");
}
void op::v0::Interpolate::validate_and_infer_types()
@ -81,7 +81,9 @@ void op::v0::Interpolate::validate_and_infer_types()
}
}
set_output_type(0, get_input_element_type(0), output_shape);
return;
}
throw ngraph_error("v0_Interpolate_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v0::Interpolate::clone_with_new_inputs(const OutputVector& new_args) const
@ -91,7 +93,7 @@ shared_ptr<Node> op::v0::Interpolate::clone_with_new_inputs(const OutputVector&
check_new_args_count(this, new_args);
return make_shared<op::v0::Interpolate>(new_args.at(0), new_args.at(1), m_attrs);
}
return nullptr;
throw ngraph_error("v0_Interpolate_clone_with_new_inputs is disabled!");
}
namespace ngraph
@ -157,7 +159,7 @@ bool op::v4::Interpolate::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("cube_coeff", m_attrs.cube_coeff);
return true;
}
return false;
throw ngraph_error("v4_Interpolate_visit_attributes is disabled!");
}
std::vector<int64_t> op::v4::Interpolate::get_axes() const
@ -290,7 +292,9 @@ void op::v4::Interpolate::validate_and_infer_types()
}
set_output_type(0, get_input_element_type(0), output_shape);
return;
}
throw ngraph_error("v4_Interpolate_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v4::Interpolate::clone_with_new_inputs(const OutputVector& new_args) const
@ -306,7 +310,7 @@ shared_ptr<Node> op::v4::Interpolate::clone_with_new_inputs(const OutputVector&
return make_shared<op::v4::Interpolate>(
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3), m_attrs);
}
return nullptr;
throw ngraph_error("v4_Interpolate_clone_with_new_inputs is disabled!");
}
namespace
@ -523,7 +527,7 @@ bool op::v4::Interpolate::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v4_Interpolate_evaluate) { return evaluate_interpolate(outputs, inputs); }
return false;
throw ngraph_error("v4_Interpolate_evaluate is disabled!");
}
namespace ngraph

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Less::clone_with_new_inputs(const OutputVector& new_arg
check_new_args_count(this, new_args);
return make_shared<op::v1::Less>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Less_clone_with_new_inputs is disabled!");
}
bool op::v1::Less::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
@ -91,5 +91,5 @@ bool op::v1::Less::evaluate(const HostTensorVector& outputs, const HostTensorVec
{
return lessop::evaluate_less(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Less_evaluate is disabled!");
}

View File

@ -43,7 +43,7 @@ shared_ptr<Node> op::v1::LessEqual::clone_with_new_inputs(const OutputVector& ne
check_new_args_count(this, new_args);
return make_shared<v1::LessEqual>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_LessEqual_clone_with_new_inputs is disabled!");
}
namespace less_equalop
@ -92,5 +92,5 @@ bool op::v1::LessEqual::evaluate(const HostTensorVector& outputs,
{
return less_equalop::evaluate_less_equal(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_LessEqual_evaluate is disabled!");
}

View File

@ -36,7 +36,7 @@ op::Log::Log(const Output<Node>& arg)
bool ngraph::op::v0::Log::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Log_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Log_visit_attributes is disabled!");
}
shared_ptr<Node> op::Log::clone_with_new_inputs(const OutputVector& new_args) const
@ -46,7 +46,7 @@ shared_ptr<Node> op::Log::clone_with_new_inputs(const OutputVector& new_args) co
check_new_args_count(this, new_args);
return make_shared<Log>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Log_clone_with_new_inputs is disabled!");
}
namespace logop
@ -85,5 +85,5 @@ bool op::Log::evaluate(const HostTensorVector& outputs, const HostTensorVector&
{
return logop::evaluate_log(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Log_evaluate is disabled!");
}

View File

@ -38,7 +38,7 @@ bool op::v5::LogSoftmax::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("axis", m_axis);
return true;
}
return false;
throw ngraph_error("v5_LogSoftmax_visit_attributes is disabled!");
}
void op::v5::LogSoftmax::validate_and_infer_types()
@ -57,7 +57,9 @@ void op::v5::LogSoftmax::validate_and_infer_types()
").");
set_output_type(0, get_input_element_type(0), input_shape);
return;
}
throw ngraph_error("v5_LogSoftmax_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v5::LogSoftmax::clone_with_new_inputs(const OutputVector& new_args) const
@ -67,5 +69,5 @@ shared_ptr<Node> op::v5::LogSoftmax::clone_with_new_inputs(const OutputVector& n
check_new_args_count(this, new_args);
return make_shared<op::v5::LogSoftmax>(new_args.at(0), m_axis);
}
return nullptr;
throw ngraph_error("v5_LogSoftmax_clone_with_new_inputs is disabled!");
}

View File

@ -45,7 +45,7 @@ bool op::v5::Loop::visit_attributes(AttributeVisitor& visitor)
return false;
}
return false;
throw ngraph_error("v5_Loop_visit_attributes is disabled!");
}
void op::v5::Loop::validate_and_infer_types()
@ -310,7 +310,9 @@ void op::v5::Loop::validate_and_infer_types()
}
}
}
return;
}
throw ngraph_error("v5_Loop_validate_and_infer_types is disabled!");
}
std::shared_ptr<Node> op::v5::Loop::clone_with_new_inputs(const OutputVector& new_args) const
@ -392,7 +394,7 @@ std::shared_ptr<Node> op::v5::Loop::clone_with_new_inputs(const OutputVector& ne
}
return move(op);
}
return nullptr;
throw ngraph_error("v5_Loop_clone_with_new_inputs is disabled!");
}
Output<Node> op::v5::Loop::get_concatenated_slices(const Output<Node>& value,
@ -421,5 +423,5 @@ bool op::v5::Loop::evaluate(const HostTensorVector& outputs, const HostTensorVec
inputs);
return true;
}
return false;
throw ngraph_error("v5_Loop_evaluate is disabled!");
}

View File

@ -113,7 +113,9 @@ void op::LRN::validate_and_infer_types()
"Axes input must be integral numbers, but are: ",
axes_type,
").");
return;
}
throw ngraph_error("v0_LRN_validate_and_infer_types is disabled!");
}
bool ngraph::op::v0::LRN::visit_attributes(AttributeVisitor& visitor)
@ -126,7 +128,7 @@ bool ngraph::op::v0::LRN::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("size", m_size);
return true;
}
return false;
throw ngraph_error("v0_LRN_visit_attributes is disabled!");
}
shared_ptr<Node> op::LRN::clone_with_new_inputs(const OutputVector& new_args) const
@ -137,5 +139,5 @@ shared_ptr<Node> op::LRN::clone_with_new_inputs(const OutputVector& new_args) co
return make_shared<op::LRN>(
new_args.at(0), new_args.at(1), m_alpha, m_beta, m_bias, m_size);
}
return nullptr;
throw ngraph_error("v0_LRN_clone_with_new_inputs is disabled!");
}

View File

@ -142,7 +142,7 @@ bool ngraph::op::v0::LSTMCell::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("weights_format", m_weights_format);
return true;
}
return false;
throw ngraph_error("v0_LSTMCell_visit_attributes is disabled!");
}
void op::v0::LSTMCell::validate_and_infer_types()
@ -297,7 +297,9 @@ void op::v0::LSTMCell::validate_and_infer_types()
set_output_size(2);
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
set_output_type(1, result_et, {merged_batch_size, merged_hidden_size});
return;
}
throw ngraph_error("v0_LSTMCell_validate_and_infer_types is disabled!");
}
Output<Node> op::v0::LSTMCell::get_default_bias_input() const
@ -371,7 +373,7 @@ shared_ptr<Node> op::v0::LSTMCell::clone_with_new_inputs(const OutputVector& new
throw ngraph_error("Incorrect number of new arguments");
}
}
return nullptr;
throw ngraph_error("v0_LSTMCell_clone_with_new_inputs is disabled!");
}
namespace ngraph
@ -459,7 +461,7 @@ bool ngraph::op::v4::LSTMCell::visit_attributes(AttributeVisitor& visitor)
{
return op::util::RNNCellBase::visit_attributes(visitor);
}
return false;
throw ngraph_error("v4_LSTMCell_visit_attributes is disabled!");
}
void op::v4::LSTMCell::validate_and_infer_types()
@ -573,7 +575,9 @@ void op::v4::LSTMCell::validate_and_infer_types()
set_output_size(2);
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
set_output_type(1, result_et, {merged_batch_size, merged_hidden_size});
return;
}
throw ngraph_error("v4_LSTMCell_validate_and_infer_types is disabled!");
}
Output<Node> op::v4::LSTMCell::get_default_bias_input() const
@ -619,5 +623,5 @@ shared_ptr<Node> op::v4::LSTMCell::clone_with_new_inputs(const OutputVector& new
throw ngraph_error("Incorrect number of new arguments");
}
}
return nullptr;
throw ngraph_error("v4_LSTMCell_clone_with_new_inputs is disabled!");
}

View File

@ -48,7 +48,7 @@ bool ngraph::op::v0::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("weights_format", m_weights_format);
return true;
}
return false;
throw ngraph_error("v0_LSTMSequence_visit_attributes is disabled!");
}
OutputVector op::v0::LSTMSequence::decompose_op() const
{
@ -121,7 +121,7 @@ shared_ptr<Node> op::v0::LSTMSequence::clone_with_new_inputs(const OutputVector&
throw ngraph_error("Incorrect number of new arguments");
}
}
return nullptr;
throw ngraph_error("v0_LSTMSequence_clone_with_new_inputs is disabled!");
}
shared_ptr<Node> op::v0::LSTMSequence::get_masked_node(const Output<Node>& data,
@ -437,7 +437,9 @@ void op::v0::LSTMSequence::validate_and_infer_types()
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
set_output_type(
2, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
return;
}
throw ngraph_error("v0_LSTMSequence_validate_and_infer_types is disabled!");
}
bool ngraph::op::v5::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
@ -447,7 +449,7 @@ bool ngraph::op::v5::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("direction", m_direction);
return op::util::RNNCellBase::visit_attributes(visitor);
}
return false;
throw ngraph_error("v5_LSTMSequence_visit_attributes is disabled!");
}
shared_ptr<Node> op::v5::LSTMSequence::clone_with_new_inputs(const OutputVector& new_args) const
@ -476,7 +478,7 @@ shared_ptr<Node> op::v5::LSTMSequence::clone_with_new_inputs(const OutputVector&
throw ngraph_error("Incorrect number of new arguments");
}
}
return nullptr;
throw ngraph_error("v5_LSTMSequence_clone_with_new_inputs is disabled!");
}
void op::v5::LSTMSequence::validate_and_infer_types()
@ -622,5 +624,7 @@ void op::v5::LSTMSequence::validate_and_infer_types()
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
set_output_type(
2, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
return;
}
throw ngraph_error("v5_LSTMSequence_validate_and_infer_types is disabled!");
}

View File

@ -46,7 +46,7 @@ bool ngraph::op::v0::MatMul::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("transpose_b", m_transpose_b);
return true;
}
return false;
throw ngraph_error("v0_MatMul_visit_attributes is disabled!");
}
shared_ptr<Node> op::MatMul::clone_with_new_inputs(const OutputVector& new_args) const
@ -56,7 +56,7 @@ shared_ptr<Node> op::MatMul::clone_with_new_inputs(const OutputVector& new_args)
check_new_args_count(this, new_args);
return make_shared<MatMul>(new_args.at(0), new_args.at(1), m_transpose_a, m_transpose_b);
}
return nullptr;
throw ngraph_error("v0_MatMul_clone_with_new_inputs is disabled!");
}
namespace matmul
@ -272,7 +272,7 @@ bool op::MatMul::evaluate(const HostTensorVector& outputs, const HostTensorVecto
return matmul::evaluate_matmul(
inputs[0], inputs[1], outputs[0], get_transpose_a(), get_transpose_b());
}
return false;
throw ngraph_error("v0_MatMul_evaluate is disabled!");
}
void ngraph::op::v0::MatMul::validate_and_infer_types()
@ -309,5 +309,7 @@ void ngraph::op::v0::MatMul::validate_and_infer_types()
{
set_output_type(0, result_et, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_MatMul_validate_and_infer_types is disabled!");
}

View File

@ -75,7 +75,7 @@ shared_ptr<Node> op::v1::ReduceMax::clone_with_new_inputs(const OutputVector& ne
check_new_args_count(this, new_args);
return make_shared<op::v1::ReduceMax>(new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v1_ReduceMax_clone_with_new_inputs is disabled!");
}
bool op::v1::ReduceMax::evaluate(const HostTensorVector& outputs,
@ -85,5 +85,5 @@ bool op::v1::ReduceMax::evaluate(const HostTensorVector& outputs,
{
return maxop::evaluate_max(inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
}
return false;
throw ngraph_error("v1_ReduceMax_evaluate is disabled!");
}

View File

@ -80,7 +80,7 @@ bool ngraph::op::v1::MaxPool::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_pad", m_auto_pad);
return true;
}
return false;
throw ngraph_error("v1_MaxPool_visit_attributes is disabled!");
}
void op::v1::MaxPool::validate_and_infer_types()
@ -138,7 +138,9 @@ void op::v1::MaxPool::validate_and_infer_types()
true,
m_rounding_type == op::RoundingType::CEIL)
: output_shape);
return;
}
throw ngraph_error("v1_MaxPool_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::MaxPool::clone_with_new_inputs(const OutputVector& new_args) const
@ -154,7 +156,7 @@ shared_ptr<Node> op::v1::MaxPool::clone_with_new_inputs(const OutputVector& new_
m_rounding_type,
m_auto_pad);
}
return nullptr;
throw ngraph_error("v1_MaxPool_clone_with_new_inputs is disabled!");
}
shared_ptr<Node> op::v1::MaxPool::get_default_value() const
@ -247,5 +249,5 @@ bool op::v1::MaxPool::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v1_MaxPool_evaluate) { return evaluate_maxpool(outputs, inputs); }
return false;
throw ngraph_error("v1_MaxPool_evaluate is disabled!");
}

View File

@ -89,7 +89,7 @@ shared_ptr<Node> op::v1::Maximum::clone_with_new_inputs(const OutputVector& new_
check_new_args_count(this, new_args);
return make_shared<op::v1::Maximum>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Maximum_clone_with_new_inputs is disabled!");
}
bool op::v1::Maximum::evaluate(const HostTensorVector& outputs,
@ -99,5 +99,5 @@ bool op::v1::Maximum::evaluate(const HostTensorVector& outputs,
{
return maximumop::evaluate_maximum(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Maximum_evaluate is disabled!");
}

View File

@ -77,7 +77,7 @@ shared_ptr<Node> op::v1::ReduceMin::clone_with_new_inputs(const OutputVector& ne
check_new_args_count(this, new_args);
return make_shared<op::v1::ReduceMin>(new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v1_ReduceMin_clone_with_new_inputs is disabled!");
}
bool op::v1::ReduceMin::evaluate(const HostTensorVector& outputs,
@ -87,5 +87,5 @@ bool op::v1::ReduceMin::evaluate(const HostTensorVector& outputs,
{
return minop::evaluate_min(inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
}
return false;
throw ngraph_error("v1_ReduceMin_evaluate is disabled!");
}

View File

@ -87,7 +87,7 @@ shared_ptr<Node> op::v1::Minimum::clone_with_new_inputs(const OutputVector& new_
check_new_args_count(this, new_args);
return make_shared<op::v1::Minimum>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Minimum_clone_with_new_inputs is disabled!");
}
bool op::v1::Minimum::evaluate(const HostTensorVector& outputs,
@ -97,5 +97,5 @@ bool op::v1::Minimum::evaluate(const HostTensorVector& outputs,
{
return minimumop::evaluate_minimum(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Minimum_evaluate is disabled!");
}

View File

@ -36,7 +36,7 @@ op::v4::Mish::Mish(const Output<Node>& arg)
bool op::v4::Mish::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v4_Mish_visit_attributes) { return true; }
return false;
throw ngraph_error("v4_Mish_visit_attributes is disabled!");
}
void op::v4::Mish::validate_and_infer_types()
@ -45,7 +45,9 @@ void op::v4::Mish::validate_and_infer_types()
{
set_output_size(1);
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
return;
}
throw ngraph_error("v4_Mish_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v4::Mish::clone_with_new_inputs(const OutputVector& new_args) const
@ -55,7 +57,7 @@ shared_ptr<Node> op::v4::Mish::clone_with_new_inputs(const OutputVector& new_arg
check_new_args_count(this, new_args);
return make_shared<Mish>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v4_Mish_clone_with_new_inputs is disabled!");
}
namespace mish
@ -89,5 +91,5 @@ bool op::v4::Mish::evaluate(const HostTensorVector& outputs, const HostTensorVec
{
return mish::evaluate_mish(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v4_Mish_evaluate is disabled!");
}

View File

@ -46,7 +46,7 @@ bool ngraph::op::v1::Mod::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("auto_broadcast", m_auto_broadcast);
return true;
}
return false;
throw ngraph_error("v1_Mod_visit_attributes is disabled!");
}
OutputVector op::v1::Mod::decompose_op() const
@ -75,5 +75,5 @@ shared_ptr<Node> op::v1::Mod::clone_with_new_inputs(const OutputVector& new_args
{
return make_shared<Mod>(new_args.at(0), new_args.at(1), m_auto_broadcast);
}
return nullptr;
throw ngraph_error("v1_Mod_clone_with_new_inputs is disabled!");
}

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v0::Multiply::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v0::Multiply>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v0_Multiply_clone_with_new_inputs is disabled!");
}
bool op::v0::Multiply::evaluate(const HostTensorVector& outputs,
@ -92,7 +92,7 @@ bool op::v0::Multiply::evaluate(const HostTensorVector& outputs,
{
return multiplyop::evaluate_multiply(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v0_Multiply_evaluate is disabled!");
}
// ------------------------------------ v1 -------------------------------------
@ -114,7 +114,7 @@ shared_ptr<Node> op::v1::Multiply::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v1::Multiply>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Multiply_clone_with_new_inputs is disabled!");
}
bool op::v1::Multiply::evaluate(const HostTensorVector& outputs,
@ -124,5 +124,5 @@ bool op::v1::Multiply::evaluate(const HostTensorVector& outputs,
{
return multiplyop::evaluate_multiply(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Multiply_evaluate is disabled!");
}

View File

@ -78,7 +78,9 @@ void op::MVN::validate_and_infer_types()
}
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
return;
}
throw ngraph_error("v0_MVN_validate_and_infer_types is disabled!");
}
OutputVector op::MVN::decompose_op() const
@ -119,7 +121,7 @@ shared_ptr<Node> op::MVN::clone_with_new_inputs(const OutputVector& new_args) co
new_args.size());
return make_shared<MVN>(new_args.at(0), m_reduction_axes, m_normalize_variance, m_eps);
}
return nullptr;
throw ngraph_error("v0_MVN_clone_with_new_inputs is disabled!");
}
bool op::MVN::visit_attributes(AttributeVisitor& visitor)
@ -132,7 +134,7 @@ bool op::MVN::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("reduction_axes", m_reduction_axes);
return true;
}
return false;
throw ngraph_error("v0_MVN_visit_attributes is disabled!");
}
// ------------------------------ V6 ------------------------------
@ -194,7 +196,9 @@ void op::v6::MVN::validate_and_infer_types()
}
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
return;
}
throw ngraph_error("v6_MVN_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v6::MVN::clone_with_new_inputs(const OutputVector& new_args) const
@ -208,7 +212,7 @@ shared_ptr<Node> op::v6::MVN::clone_with_new_inputs(const OutputVector& new_args
return make_shared<op::v6::MVN>(
new_args.at(0), new_args.at(1), m_normalize_variance, m_eps, m_eps_mode);
}
return nullptr;
throw ngraph_error("v6_MVN_clone_with_new_inputs is disabled!");
}
bool op::v6::MVN::visit_attributes(AttributeVisitor& visitor)
@ -220,5 +224,5 @@ bool op::v6::MVN::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("eps_mode", m_eps_mode);
return true;
}
return false;
throw ngraph_error("v6_MVN_visit_attributes is disabled!");
}

View File

@ -33,7 +33,7 @@ op::Negative::Negative(const Output<Node>& arg)
bool ngraph::op::v0::Negative::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Negative_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Negative_visit_attributes is disabled!");
}
shared_ptr<Node> op::Negative::clone_with_new_inputs(const OutputVector& new_args) const
@ -43,7 +43,7 @@ shared_ptr<Node> op::Negative::clone_with_new_inputs(const OutputVector& new_arg
check_new_args_count(this, new_args);
return make_shared<Negative>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v0_Negative_clone_with_new_inputs is disabled!");
}
namespace negativeop
@ -83,7 +83,7 @@ bool op::Negative::evaluate(const HostTensorVector& outputs, const HostTensorVec
return negativeop::evaluate_negative(
inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v0_Negative_evaluate is disabled!");
}
shared_ptr<Node> ngraph::operator-(const Output<Node>& arg0)

View File

@ -90,7 +90,7 @@ std::shared_ptr<Node>
m_box_encoding,
m_sort_result_descending);
}
return nullptr;
throw ngraph_error("v1_NonMaxSuppression_clone_with_new_inputs is disabled!");
}
bool ngraph::op::v1::NonMaxSuppression::visit_attributes(AttributeVisitor& visitor)
@ -101,7 +101,7 @@ bool ngraph::op::v1::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
visitor.on_attribute("sort_result_descending", m_sort_result_descending);
return true;
}
return false;
throw ngraph_error("v1_NonMaxSuppression_visit_attributes is disabled!");
}
void op::v1::NonMaxSuppression::validate_and_infer_types()
@ -204,7 +204,9 @@ void op::v1::NonMaxSuppression::validate_and_infer_types()
out_shape[0] = std::min(num_boxes, max_output_boxes_per_class * num_classes);
}
set_output_type(0, output_element_type, out_shape);
return;
}
throw ngraph_error("v1_NonMaxSuppression_validate_and_infer_types is disabled!");
}
int64_t op::v1::NonMaxSuppression::max_boxes_output_from_input() const
@ -309,7 +311,7 @@ std::shared_ptr<Node>
m_sort_result_descending,
m_output_type);
}
return nullptr;
throw ngraph_error("v3_NonMaxSuppression_clone_with_new_inputs is disabled!");
}
bool ngraph::op::v3::NonMaxSuppression::visit_attributes(AttributeVisitor& visitor)
@ -321,7 +323,7 @@ bool ngraph::op::v3::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
visitor.on_attribute("output_type", m_output_type);
return true;
}
return false;
throw ngraph_error("v3_NonMaxSuppression_visit_attributes is disabled!");
}
void op::v3::NonMaxSuppression::validate()
@ -430,7 +432,9 @@ void op::v3::NonMaxSuppression::validate_and_infer_types()
}
}
set_output_type(0, m_output_type, out_shape);
return;
}
throw ngraph_error("v3_NonMaxSuppression_validate_and_infer_types is disabled!");
}
int64_t op::v3::NonMaxSuppression::max_boxes_output_from_input() const
@ -539,7 +543,7 @@ std::shared_ptr<Node>
m_sort_result_descending,
m_output_type);
}
return nullptr;
throw ngraph_error("v4_NonMaxSuppression_clone_with_new_inputs is disabled!");
}
void op::v4::NonMaxSuppression::validate_and_infer_types()
@ -571,7 +575,9 @@ void op::v4::NonMaxSuppression::validate_and_infer_types()
}
}
set_output_type(0, m_output_type, out_shape);
return;
}
throw ngraph_error("v4_NonMaxSuppression_validate_and_infer_types is disabled!");
}
// ------------------------------ V5 ------------------------------
@ -722,7 +728,7 @@ std::shared_ptr<Node>
break;
}
}
return nullptr;
throw ngraph_error("v5_NonMaxSuppression_clone_with_new_inputs is disabled!");
}
namespace
@ -935,7 +941,7 @@ bool ngraph::op::v5::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
visitor.on_attribute("output_type", m_output_type);
return true;
}
return false;
throw ngraph_error("v5_NonMaxSuppression_visit_attributes is disabled!");
}
void op::v5::NonMaxSuppression::validate_and_infer_types()
@ -971,7 +977,9 @@ void op::v5::NonMaxSuppression::validate_and_infer_types()
set_output_type(0, m_output_type, out_shape);
set_output_type(1, element::f32, out_shape);
set_output_type(2, m_output_type, Shape{1});
return;
}
throw ngraph_error("v5_NonMaxSuppression_validate_and_infer_types is disabled!");
}
namespace ngraph

View File

@ -53,7 +53,7 @@ bool ngraph::op::v3::NonZero::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("output_type", m_output_type);
return true;
}
return false;
throw ngraph_error("v3_NonZero_visit_attributes is disabled!");
}
void op::v3::NonZero::validate_and_infer_types()
@ -84,7 +84,9 @@ void op::v3::NonZero::validate_and_infer_types()
}
set_input_is_relevant_to_shape(0);
return;
}
throw ngraph_error("v3_NonZero_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v3::NonZero::clone_with_new_inputs(const OutputVector& new_args) const
@ -94,7 +96,7 @@ shared_ptr<Node> op::v3::NonZero::clone_with_new_inputs(const OutputVector& new_
check_new_args_count(this, new_args);
return make_shared<v3::NonZero>(new_args.at(0), m_output_type);
}
return nullptr;
throw ngraph_error("v3_NonZero_clone_with_new_inputs is disabled!");
}
namespace nonzero
@ -178,5 +180,5 @@ bool op::v3::NonZero::evaluate(const HostTensorVector& outputs,
{
return nonzero::evaluate_nonzero(inputs[0], outputs[0]);
}
return false;
throw ngraph_error("v3_NonZero_evaluate is disabled!");
}

View File

@ -52,7 +52,7 @@ bool ngraph::op::v0::NormalizeL2::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("eps_mode", m_eps_mode);
return true;
}
return false;
throw ngraph_error("v0_NormalizeL2_visit_attributes is disabled!");
}
void op::NormalizeL2::pre_validate_and_infer_types()
@ -129,5 +129,5 @@ shared_ptr<Node> op::NormalizeL2::clone_with_new_inputs(const OutputVector& new_
}
return make_shared<NormalizeL2>(new_args.at(0), new_args.at(1), m_eps, m_eps_mode);
}
return nullptr;
throw ngraph_error("v0_NormalizeL2_clone_with_new_inputs is disabled!");
}

View File

@ -39,7 +39,7 @@ op::v1::LogicalNot::LogicalNot(const Output<Node>& arg)
bool ngraph::op::v1::LogicalNot::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_LogicalNot_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_LogicalNot_visit_attributes is disabled!");
}
// TODO(amprocte): Update this to allow only boolean, for consistency with logical binops.
@ -52,7 +52,9 @@ void op::v1::LogicalNot::validate_and_infer_types()
PartialShape& args_pshape = std::get<1>(args_et_pshape);
set_output_type(0, args_et, args_pshape);
return;
}
throw ngraph_error("v1_LogicalNot_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::LogicalNot::clone_with_new_inputs(const OutputVector& new_args) const
@ -62,7 +64,7 @@ shared_ptr<Node> op::v1::LogicalNot::clone_with_new_inputs(const OutputVector& n
check_new_args_count(this, new_args);
return make_shared<v1::LogicalNot>(new_args.at(0));
}
return nullptr;
throw ngraph_error("v1_LogicalNot_clone_with_new_inputs is disabled!");
}
namespace notop
@ -103,5 +105,5 @@ bool op::v1::LogicalNot::evaluate(const HostTensorVector& outputs,
{
return notop::evaluate_not(inputs[0], outputs[0], shape_size(get_output_shape(0)));
}
return false;
throw ngraph_error("v1_LogicalNot_evaluate is disabled!");
}

View File

@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::NotEqual::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v1::NotEqual>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_NotEqual_clone_with_new_inputs is disabled!");
}
bool op::v1::NotEqual::evaluate(const HostTensorVector& outputs,
@ -92,11 +92,11 @@ bool op::v1::NotEqual::evaluate(const HostTensorVector& outputs,
{
return not_equalop::evaluate_not_equal(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_NotEqual_evaluate is disabled!");
}
bool op::v1::NotEqual::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v1_NotEqual_visit_attributes) { return true; }
return false;
throw ngraph_error("v1_NotEqual_visit_attributes is disabled!");
}

View File

@ -120,7 +120,9 @@ void op::v1::OneHot::validate_and_infer_types()
}
set_output_type(0, on_value_et, result_shape);
return;
}
throw ngraph_error("v1_OneHot_validate_and_infer_types is disabled!");
}
bool ngraph::op::v1::OneHot::visit_attributes(AttributeVisitor& visitor)
@ -130,7 +132,7 @@ bool ngraph::op::v1::OneHot::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("axis", m_axis);
return true;
}
return false;
throw ngraph_error("v1_OneHot_visit_attributes is disabled!");
}
shared_ptr<Node> op::v1::OneHot::clone_with_new_inputs(const OutputVector& new_args) const
@ -141,7 +143,7 @@ shared_ptr<Node> op::v1::OneHot::clone_with_new_inputs(const OutputVector& new_a
return make_shared<v1::OneHot>(
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3), m_axis);
}
return nullptr;
throw ngraph_error("v1_OneHot_clone_with_new_inputs is disabled!");
}
namespace detail
@ -224,5 +226,5 @@ bool op::v1::OneHot::evaluate(const HostTensorVector& output_values,
{
return detail::evaluate_onehot(output_values, input_values, get_axis());
}
return false;
throw ngraph_error("v1_OneHot_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ shared_ptr<Node> op::v1::LogicalOr::clone_with_new_inputs(const OutputVector& ne
check_new_args_count(this, new_args);
return make_shared<v1::LogicalOr>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_LogicalOr_clone_with_new_inputs is disabled!");
}
namespace logor
@ -90,5 +90,5 @@ bool op::v1::LogicalOr::evaluate(const HostTensorVector& outputs,
{
return logor::evaluate_logor(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_LogicalOr_evaluate is disabled!");
}

View File

@ -80,7 +80,7 @@ bool ngraph::op::v1::Pad::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("pad_mode", m_pad_mode);
return true;
}
return false;
throw ngraph_error("v1_Pad_visit_attributes is disabled!");
}
void op::v1::Pad::validate_and_infer_types()
@ -202,7 +202,9 @@ void op::v1::Pad::validate_and_infer_types()
{
set_output_type(0, get_input_element_type(0), PartialShape::dynamic());
}
return;
}
throw ngraph_error("v1_Pad_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v1::Pad::clone_with_new_inputs(const OutputVector& new_args) const
@ -220,7 +222,7 @@ shared_ptr<Node> op::v1::Pad::clone_with_new_inputs(const OutputVector& new_args
return make_shared<v1::Pad>(new_args.at(0), new_args.at(1), new_args.at(2), m_pad_mode);
}
}
return nullptr;
throw ngraph_error("v1_Pad_clone_with_new_inputs is disabled!");
}
bool op::v1::Pad::evaluate_pad(const HostTensorVector& outputs,
@ -257,5 +259,5 @@ bool op::v1::Pad::evaluate_pad(const HostTensorVector& outputs,
bool op::v1::Pad::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(v1_Pad_evaluate) { return evaluate_pad(outputs, inputs); }
return false;
throw ngraph_error("v1_Pad_evaluate is disabled!");
}

View File

@ -41,7 +41,7 @@ bool op::Parameter::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("element_type", m_element_type);
return true;
}
return false;
throw ngraph_error("v0_Parameter_visit_attributes is disabled!");
}
void op::Parameter::validate_and_infer_types()
@ -50,7 +50,9 @@ void op::Parameter::validate_and_infer_types()
{
Op::validate_and_infer_types();
set_output_type(0, m_element_type, m_partial_shape);
return;
}
throw ngraph_error("v0_Parameter_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::Parameter::clone_with_new_inputs(const OutputVector& new_args) const
@ -60,7 +62,7 @@ shared_ptr<Node> op::Parameter::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<Parameter>(m_element_type, m_partial_shape);
}
return nullptr;
throw ngraph_error("v0_Parameter_clone_with_new_inputs is disabled!");
}
bool op::Parameter::is_relevant_to_shapes() const

View File

@ -85,7 +85,7 @@ shared_ptr<Node> op::v1::Power::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<op::v1::Power>(new_args.at(0), new_args.at(1), this->get_autob());
}
return nullptr;
throw ngraph_error("v1_Power_clone_with_new_inputs is disabled!");
}
bool op::v1::Power::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
@ -94,5 +94,5 @@ bool op::v1::Power::evaluate(const HostTensorVector& outputs, const HostTensorVe
{
return power::evaluate_power(inputs[0], inputs[1], outputs[0], get_autob());
}
return false;
throw ngraph_error("v1_Power_evaluate is disabled!");
}

View File

@ -43,7 +43,7 @@ op::PRelu::PRelu(const Output<Node>& data, const Output<Node>& slope)
bool ngraph::op::v0::PRelu::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_PRelu_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_PRelu_visit_attributes is disabled!");
}
void ngraph::op::v0::PRelu::pre_validate_and_infer_types()
@ -97,7 +97,7 @@ shared_ptr<Node> op::PRelu::clone_with_new_inputs(const OutputVector& new_args)
}
return make_shared<PRelu>(new_args.at(0), new_args.at(1));
}
return nullptr;
throw ngraph_error("v0_PRelu_clone_with_new_inputs is disabled!");
}
namespace prelu
@ -136,5 +136,5 @@ bool op::PRelu::evaluate(const HostTensorVector& outputs, const HostTensorVector
{
return prelu::evaluate_prelu(inputs[0], inputs[1], outputs[0]);
}
return false;
throw ngraph_error("v0_PRelu_evaluate is disabled!");
}

View File

@ -83,7 +83,9 @@ void op::PriorBox::validate_and_infer_types()
{
set_output_type(0, element::f32, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_PriorBox_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::PriorBox::clone_with_new_inputs(const OutputVector& new_args) const
@ -93,7 +95,7 @@ shared_ptr<Node> op::PriorBox::clone_with_new_inputs(const OutputVector& new_arg
check_new_args_count(this, new_args);
return make_shared<PriorBox>(new_args.at(0), new_args.at(1), m_attrs);
}
return nullptr;
throw ngraph_error("v0_PriorBox_clone_with_new_inputs is disabled!");
}
int64_t op::PriorBox::number_of_priors(const PriorBoxAttrs& attrs)
@ -159,7 +161,7 @@ bool op::PriorBox::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("scale_all_sizes", m_attrs.scale_all_sizes);
return true;
}
return false;
throw ngraph_error("v0_PriorBox_visit_attributes is disabled!");
}
namespace prior_box
@ -207,5 +209,5 @@ bool op::v0::PriorBox::evaluate(const HostTensorVector& outputs,
{
return prior_box::evaluate_prior_box(inputs[0], inputs[1], outputs[0], get_attrs());
}
return false;
throw ngraph_error("v0_PriorBox_evaluate is disabled!");
}

View File

@ -88,7 +88,9 @@ void op::PriorBoxClustered::validate_and_infer_types()
{
set_output_type(0, element::f32, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_PriorBoxClustered_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::PriorBoxClustered::clone_with_new_inputs(const OutputVector& new_args) const
@ -98,7 +100,7 @@ shared_ptr<Node> op::PriorBoxClustered::clone_with_new_inputs(const OutputVector
check_new_args_count(this, new_args);
return make_shared<PriorBoxClustered>(new_args.at(0), new_args.at(1), m_attrs);
}
return nullptr;
throw ngraph_error("v0_PriorBoxClustered_clone_with_new_inputs is disabled!");
}
bool op::PriorBoxClustered::visit_attributes(AttributeVisitor& visitor)
@ -132,7 +134,7 @@ bool op::PriorBoxClustered::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("variance", m_attrs.variances);
return true;
}
return false;
throw ngraph_error("v0_PriorBoxClustered_visit_attributes is disabled!");
}
namespace prior_box_clustered
@ -181,5 +183,5 @@ bool op::v0::PriorBoxClustered::evaluate(const HostTensorVector& outputs,
return prior_box_clustered::evaluate_prior_box(
inputs[0], inputs[1], outputs[0], get_attrs());
}
return false;
throw ngraph_error("v0_PriorBoxClustered_evaluate is disabled!");
}

View File

@ -85,7 +85,9 @@ void op::v0::Proposal::validate_and_infer_types()
{
set_output_type(0, get_input_element_type(0), PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_Proposal_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v0::Proposal::clone_with_new_inputs(const OutputVector& new_args) const
@ -96,7 +98,7 @@ shared_ptr<Node> op::v0::Proposal::clone_with_new_inputs(const OutputVector& new
return make_shared<op::v0::Proposal>(
new_args.at(0), new_args.at(1), new_args.at(2), m_attrs);
}
return nullptr;
throw ngraph_error("v0_Proposal_clone_with_new_inputs is disabled!");
}
bool op::v0::Proposal::visit_attributes(AttributeVisitor& visitor)
@ -119,7 +121,7 @@ bool op::v0::Proposal::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("framework", m_attrs.framework);
return true;
}
return false;
throw ngraph_error("v0_Proposal_visit_attributes is disabled!");
}
NGRAPH_RTTI_DEFINITION(op::v4::Proposal, "Proposal", 4);
@ -149,7 +151,9 @@ void op::v4::Proposal::validate_and_infer_types()
1, get_input_element_type(0), PartialShape{batch_size * m_attrs.post_nms_topn});
else
set_output_type(1, get_input_element_type(0), PartialShape::dynamic());
return;
}
throw ngraph_error("v4_Proposal_validate_and_infer_types is disabled!");
}
std::shared_ptr<Node> op::v4::Proposal::clone_with_new_inputs(const OutputVector& new_args) const
@ -160,5 +164,5 @@ std::shared_ptr<Node> op::v4::Proposal::clone_with_new_inputs(const OutputVector
return make_shared<op::v4::Proposal>(
new_args.at(0), new_args.at(1), new_args.at(2), m_attrs);
}
return nullptr;
throw ngraph_error("v4_Proposal_clone_with_new_inputs is disabled!");
}

View File

@ -54,7 +54,7 @@ bool ngraph::op::v0::PSROIPooling::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("spatial_bins_y", m_spatial_bins_y);
return true;
}
return false;
throw ngraph_error("v0_PSROIPooling_visit_attributes is disabled!");
}
void op::PSROIPooling::validate_and_infer_types()
@ -137,7 +137,9 @@ void op::PSROIPooling::validate_and_infer_types()
}
set_output_type(0, feat_maps_et, output_shape);
}
return;
}
throw ngraph_error("v0_PSROIPooling_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::PSROIPooling::clone_with_new_inputs(const OutputVector& new_args) const
@ -154,5 +156,5 @@ shared_ptr<Node> op::PSROIPooling::clone_with_new_inputs(const OutputVector& new
m_spatial_bins_y,
m_mode);
}
return nullptr;
throw ngraph_error("v0_PSROIPooling_clone_with_new_inputs is disabled!");
}

View File

@ -161,7 +161,9 @@ void op::Quantize::validate_and_infer_types()
{
set_output_type(0, quantized_type, PartialShape::dynamic());
}
return;
}
throw ngraph_error("v0_Quantize_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::Quantize::clone_with_new_inputs(const OutputVector& new_args) const
@ -172,5 +174,5 @@ shared_ptr<Node> op::Quantize::clone_with_new_inputs(const OutputVector& new_arg
return make_shared<Quantize>(
new_args.at(0), new_args.at(1), new_args.at(2), m_type, m_axes, m_round_mode);
}
return nullptr;
throw ngraph_error("v0_Quantize_clone_with_new_inputs is disabled!");
}

View File

@ -75,7 +75,7 @@ bool ngraph::op::v4::Range::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("output_type", m_output_type);
return true;
}
return false;
throw ngraph_error("v4_Range_visit_attributes is disabled!");
}
void op::v4::Range::validate_and_infer_types()
@ -187,7 +187,9 @@ void op::v4::Range::validate_and_infer_types()
result = PartialShape{Dimension(static_cast<int64_t>(strided))};
}
set_output_type(0, m_output_type, result);
return;
}
throw ngraph_error("v4_Range_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v4::Range::clone_with_new_inputs(const OutputVector& new_args) const
@ -198,7 +200,7 @@ shared_ptr<Node> op::v4::Range::clone_with_new_inputs(const OutputVector& new_ar
return make_shared<v4::Range>(
new_args.at(0), new_args.at(1), new_args.at(2), m_output_type);
}
return nullptr;
throw ngraph_error("v4_Range_clone_with_new_inputs is disabled!");
}
template <typename T>
@ -323,7 +325,7 @@ bool op::v4::Range::evaluate(const HostTensorVector& outputs, const HostTensorVe
HostTensorPtr step = inputs[2];
return rangeop::evaluate_power(out, start, stop, step, m_output_type, 4);
}
return false;
throw ngraph_error("v4_Range_evaluate is disabled!");
}
constexpr NodeTypeInfo op::v0::Range::type_info;
@ -437,7 +439,7 @@ static PartialShape infer_output_shape(const op::v0::Range* node, const element:
bool ngraph::op::v0::Range::visit_attributes(AttributeVisitor& visitor)
{
NGRAPH_OP_SCOPE(v0_Range_visit_attributes) { return true; }
return false;
throw ngraph_error("v0_Range_visit_attributes is disabled!");
}
void op::v0::Range::validate_and_infer_types()
@ -522,7 +524,9 @@ void op::v0::Range::validate_and_infer_types()
#endif
set_output_type(0, result_et, result_shape);
return;
}
throw ngraph_error("v0_Range_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::v0::Range::clone_with_new_inputs(const OutputVector& new_args) const
@ -532,7 +536,7 @@ shared_ptr<Node> op::v0::Range::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<Range>(new_args.at(0), new_args.at(1), new_args.at(2));
}
return nullptr;
throw ngraph_error("v0_Range_clone_with_new_inputs is disabled!");
}
template <element::Type_t ET, typename T>
@ -550,5 +554,5 @@ bool op::v0::Range::evaluate(const HostTensorVector& outputs, const HostTensorVe
HostTensorPtr step = inputs[2];
return rangeop::evaluate_power(out, start, stop, step, start->get_element_type(), 0);
}
return false;
throw ngraph_error("v0_Range_evaluate is disabled!");
}

View File

@ -42,7 +42,9 @@ void op::ReadValue::validate_and_infer_types()
else
m_variable->update(info);
set_output_type(0, arg_t, output_shape);
return;
}
throw ngraph_error("v3_ReadValue_validate_and_infer_types is disabled!");
}
shared_ptr<Node> op::ReadValue::clone_with_new_inputs(const OutputVector& new_args) const
@ -52,7 +54,7 @@ shared_ptr<Node> op::ReadValue::clone_with_new_inputs(const OutputVector& new_ar
check_new_args_count(this, new_args);
return make_shared<ReadValue>(new_args.at(0), m_variable_id);
}
return nullptr;
throw ngraph_error("v3_ReadValue_clone_with_new_inputs is disabled!");
}
bool op::v3::ReadValue::visit_attributes(AttributeVisitor& visitor)
@ -62,5 +64,5 @@ bool op::v3::ReadValue::visit_attributes(AttributeVisitor& visitor)
visitor.on_attribute("variable_id", m_variable_id);
return true;
}
return false;
throw ngraph_error("v3_ReadValue_visit_attributes is disabled!");
}

View File

@ -46,7 +46,7 @@ shared_ptr<Node> op::v4::ReduceL1::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v4::ReduceL1>(new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v4_ReduceL1_clone_with_new_inputs is disabled!");
}
namespace reduce_l1
@ -90,5 +90,5 @@ bool op::v4::ReduceL1::evaluate(const HostTensorVector& outputs,
return reduce_l1::evaluate_sum(
inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
}
return false;
throw ngraph_error("v4_ReduceL1_evaluate is disabled!");
}

View File

@ -46,7 +46,7 @@ shared_ptr<Node> op::v4::ReduceL2::clone_with_new_inputs(const OutputVector& new
check_new_args_count(this, new_args);
return make_shared<op::v4::ReduceL2>(new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v4_ReduceL2_clone_with_new_inputs is disabled!");
}
namespace reduce_l2
@ -88,5 +88,5 @@ bool op::v4::ReduceL2::evaluate(const HostTensorVector& outputs,
return reduce_l2::evaluate_reduce_l2(
inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
}
return false;
throw ngraph_error("v4_ReduceL2_evaluate is disabled!");
}

View File

@ -42,7 +42,7 @@ shared_ptr<Node> op::v1::ReduceLogicalAnd::clone_with_new_inputs(const OutputVec
return make_shared<op::v1::ReduceLogicalAnd>(
new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v1_ReduceLogicalAnd_clone_with_new_inputs is disabled!");
}
namespace
@ -87,5 +87,5 @@ bool op::v1::ReduceLogicalAnd::evaluate(const HostTensorVector& outputs,
const auto& out = outputs[0];
return evaluate_reduce_logical_and(data, axes, out, get_keep_dims());
}
return false;
throw ngraph_error("v1_ReduceLogicalAnd_evaluate is disabled!");
}

View File

@ -42,7 +42,7 @@ shared_ptr<Node> op::v1::ReduceLogicalOr::clone_with_new_inputs(const OutputVect
return make_shared<op::v1::ReduceLogicalOr>(
new_args.at(0), new_args.at(1), get_keep_dims());
}
return nullptr;
throw ngraph_error("v1_ReduceLogicalOr_clone_with_new_inputs is disabled!");
}
namespace
@ -87,5 +87,5 @@ bool op::v1::ReduceLogicalOr::evaluate(const HostTensorVector& outputs,
const auto& out = outputs[0];
return evaluate_reduce_logical_or(data, axes, out, get_keep_dims());
}
return false;
throw ngraph_error("v1_ReduceLogicalOr_evaluate is disabled!");
}

Some files were not shown because too many files have changed in this diff Show More