Add exceptions
This commit is contained in:
parent
300a88a644
commit
0489ba376f
@ -41,7 +41,7 @@ shared_ptr<Node> op::Abs::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Abs>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Abs_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace absop
|
||||
@ -77,10 +77,9 @@ namespace absop
|
||||
|
||||
bool op::Abs::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v0_Abs_evaluate)
|
||||
{
|
||||
rc = absop::evaluate_abs(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
return absop::evaluate_abs(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v0_Abs_evaluate is disabled!");
|
||||
}
|
||||
|
@ -50,7 +50,7 @@ shared_ptr<Node> op::Acos::clone_with_new_inputs(const OutputVector& new_args) c
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Acos>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Acos_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace acosop
|
||||
@ -85,10 +85,9 @@ namespace acosop
|
||||
|
||||
bool op::Acos::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v0_Acos_evaluate)
|
||||
{
|
||||
rc = acosop::evaluate_acos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
return acosop::evaluate_acos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v0_Acos_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Acosh::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Acosh>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Acosh_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace acoshop
|
||||
@ -74,7 +74,6 @@ namespace acoshop
|
||||
|
||||
bool op::v3::Acosh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v3_Acosh_evaluate) { rc = acoshop::evaluate_acosh(inputs[0], outputs[0]); }
|
||||
return rc;
|
||||
NGRAPH_OP_SCOPE(v3_Acosh_evaluate) { return acoshop::evaluate_acosh(inputs[0], outputs[0]); }
|
||||
throw ngraph_error("v3_Acosh_evaluate is disabled!");
|
||||
}
|
||||
|
@ -86,7 +86,7 @@ bool op::v1::Add::visit_attributes(AttributeVisitor& visitor)
|
||||
BinaryElementwiseArithmetic::visit_attributes(visitor);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Add_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::Add::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -96,15 +96,14 @@ shared_ptr<Node> op::v1::Add::clone_with_new_inputs(const OutputVector& new_args
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Add>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Add_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Add::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v1_Add_evaluate)
|
||||
{
|
||||
rc = add::evaluate_add(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
return add::evaluate_add(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v1_Add_evaluate is disabled!");
|
||||
}
|
||||
|
@ -39,7 +39,7 @@ bool op::v1::LogicalAnd::visit_attributes(AttributeVisitor& visitor)
|
||||
BinaryElementwiseLogical::visit_attributes(visitor);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_LogicalAnd_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::LogicalAnd::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -49,7 +49,7 @@ shared_ptr<Node> op::v1::LogicalAnd::clone_with_new_inputs(const OutputVector& n
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<v1::LogicalAnd>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_LogicalAnd_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace logand
|
||||
@ -94,10 +94,9 @@ namespace logand
|
||||
bool op::v1::LogicalAnd::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v1_LogicalAnd_evaluate)
|
||||
{
|
||||
rc = logand::evaluate_logand(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
return logand::evaluate_logand(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v1_LogicalAnd_evaluate is disabled!");
|
||||
}
|
||||
|
@ -51,7 +51,7 @@ shared_ptr<Node> op::Asin::clone_with_new_inputs(const OutputVector& new_args) c
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Asin>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Asin_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace asinop
|
||||
@ -86,10 +86,9 @@ namespace asinop
|
||||
|
||||
bool op::Asin::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v0_Asin_evaluate)
|
||||
{
|
||||
rc = asinop::evaluate_asin(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
return asinop::evaluate_asin(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v0_Asin_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Asinh::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Asinh>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Asinh_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace asinhop
|
||||
@ -74,7 +74,6 @@ namespace asinhop
|
||||
|
||||
bool op::v3::Asinh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v3_Asinh_evaluate) { rc = asinhop::evaluate_asinh(inputs[0], outputs[0]); }
|
||||
return rc;
|
||||
NGRAPH_OP_SCOPE(v3_Asinh_evaluate) { return asinhop::evaluate_asinh(inputs[0], outputs[0]); }
|
||||
throw ngraph_error("v3_Asinh_evaluate is disabled!");
|
||||
}
|
||||
|
@ -77,7 +77,9 @@ void op::v3::Assign::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, arg_t, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_Assign_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v3::Assign::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -87,7 +89,7 @@ shared_ptr<Node> op::v3::Assign::clone_with_new_inputs(const OutputVector& new_a
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Assign>(new_args.at(0), m_variable_id);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Assign_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v3::Assign::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -97,5 +99,5 @@ bool op::v3::Assign::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("variable_id", m_variable_id);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_Assign_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -50,7 +50,7 @@ shared_ptr<Node> op::Atan::clone_with_new_inputs(const OutputVector& new_args) c
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Atan>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Atan_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace atanop
|
||||
@ -85,10 +85,9 @@ namespace atanop
|
||||
|
||||
bool op::Atan::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v0_Atan_evaluate)
|
||||
{
|
||||
rc = atanop::evaluate_atan(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
return atanop::evaluate_atan(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return rc;
|
||||
throw ngraph_error("v0_Atan_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ shared_ptr<Node> op::v3::Atanh::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Atanh>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Atanh_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace atanhop
|
||||
@ -74,7 +74,6 @@ namespace atanhop
|
||||
|
||||
bool op::v3::Atanh::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
bool rc = false;
|
||||
NGRAPH_OP_SCOPE(v3_Atanh_evaluate) { rc = atanhop::evaluate_atanh(inputs[0], outputs[0]); }
|
||||
return rc;
|
||||
NGRAPH_OP_SCOPE(v3_Atanh_evaluate) { return atanhop::evaluate_atanh(inputs[0], outputs[0]); }
|
||||
throw ngraph_error("v3_Atanh_evaluate is disabled!");
|
||||
}
|
||||
|
@ -77,7 +77,7 @@ bool op::v1::AvgPool::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("rounding_type", m_rounding_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_AvgPool_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::AvgPool::validate_and_infer_types()
|
||||
@ -148,7 +148,9 @@ void op::v1::AvgPool::validate_and_infer_types()
|
||||
!m_exclude_pad,
|
||||
m_rounding_type == op::RoundingType::CEIL)
|
||||
: output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_AvgPool_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
const Shape& op::v1::AvgPool::get_kernel() const
|
||||
@ -235,7 +237,7 @@ shared_ptr<Node> op::v1::AvgPool::clone_with_new_inputs(const OutputVector& new_
|
||||
m_rounding_type,
|
||||
m_auto_pad);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_AvgPool_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::AvgPool::get_default_value() const
|
||||
|
@ -45,7 +45,7 @@ bool op::v0::BatchNormInference::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("epsilon", m_epsilon);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_BatchNormInference_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::BatchNormInference::validate_and_infer_types()
|
||||
@ -71,7 +71,9 @@ void op::v0::BatchNormInference::validate_and_infer_types()
|
||||
get_input_partial_shape(INPUT_VARIANCE));
|
||||
|
||||
set_output_type(0, result_et, result_batch_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_BatchNormInference_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
std::shared_ptr<Node>
|
||||
@ -87,7 +89,7 @@ std::shared_ptr<Node>
|
||||
new_args.at(4),
|
||||
m_epsilon);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_BatchNormInference_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
NGRAPH_RTTI_DEFINITION(op::v5::BatchNormInference, "BatchNormInference", 5);
|
||||
@ -111,7 +113,7 @@ bool op::v5::BatchNormInference::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("epsilon", m_epsilon);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_BatchNormInference_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v5::BatchNormInference::validate_and_infer_types()
|
||||
@ -137,7 +139,9 @@ void op::v5::BatchNormInference::validate_and_infer_types()
|
||||
get_input_partial_shape(INPUT_VARIANCE));
|
||||
|
||||
set_output_type(0, result_et, result_batch_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_BatchNormInference_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
std::shared_ptr<Node>
|
||||
@ -153,5 +157,5 @@ std::shared_ptr<Node>
|
||||
new_args.at(4),
|
||||
m_epsilon);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_BatchNormInference_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -129,7 +129,9 @@ void op::v1::BatchToSpace::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, data_type, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_BatchToSpace_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
std::shared_ptr<ngraph::Node>
|
||||
@ -141,13 +143,13 @@ std::shared_ptr<ngraph::Node>
|
||||
return make_shared<BatchToSpace>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_BatchToSpace_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v1::BatchToSpace::visit_attributes(ngraph::AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_BatchToSpace_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_BatchToSpace_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -268,5 +270,5 @@ bool ngraph::op::v1::BatchToSpace::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_BatchToSpace) { return batch_to_space_evaluate(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v1_BatchToSpace is disabled!");
|
||||
}
|
||||
|
@ -151,7 +151,9 @@ void op::v1::BinaryConvolution::validate_and_infer_types()
|
||||
m_dilations);
|
||||
|
||||
set_output_type(0, data_batch_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_BinaryConvolution_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -170,7 +172,7 @@ shared_ptr<Node>
|
||||
m_pad_value,
|
||||
m_auto_pad);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_BinaryConvolution_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::BinaryConvolution::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -186,7 +188,7 @@ bool op::v1::BinaryConvolution::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_pad", m_auto_pad);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_BinaryConvolution_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
namespace ngraph
|
||||
|
@ -203,7 +203,9 @@ void op::v3::Broadcast::validate_and_infer_types()
|
||||
set_input_is_relevant_to_shape(2); // axes_mapping - Broadcast type
|
||||
}
|
||||
set_output_type(0, get_input_element_type(0), result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_Broadcast_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v3::Broadcast::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -225,7 +227,7 @@ shared_ptr<Node> op::v3::Broadcast::clone_with_new_inputs(const OutputVector& ne
|
||||
throw ngraph_error("Not supported number of Broadcast:v3 args");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Broadcast_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v3::Broadcast::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -235,14 +237,14 @@ bool op::v3::Broadcast::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("mode", m_mode);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_Broadcast_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
bool op::v3::Broadcast::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v3_Broadcast_evaluate) { return broadcast_evaluate(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v3_Broadcast_evaluate is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -315,7 +317,9 @@ void op::v1::Broadcast::validate_and_infer_types()
|
||||
set_input_is_relevant_to_shape(0); // arg - Result element type
|
||||
set_input_is_relevant_to_shape(1); // target_shape - Result shape
|
||||
set_input_is_relevant_to_shape(2); // axes_mapping - Broadcast type
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_Broadcast_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::Broadcast::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -326,7 +330,7 @@ shared_ptr<Node> op::v1::Broadcast::clone_with_new_inputs(const OutputVector& ne
|
||||
return make_shared<v1::Broadcast>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), m_broadcast_spec);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Broadcast_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Broadcast::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -336,7 +340,7 @@ bool op::v1::Broadcast::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("mode", m_broadcast_spec);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Broadcast_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Broadcast::evaluate(const HostTensorVector& outputs,
|
||||
@ -346,5 +350,5 @@ bool op::v1::Broadcast::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return op::util::BroadcastBase::evaluate(outputs, inputs);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Broadcast_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ bool op::v3::Bucketize::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("with_right_bound", m_with_right_bound);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_Bucketize_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v3::Bucketize::validate_and_infer_types()
|
||||
@ -68,7 +68,9 @@ void op::v3::Bucketize::validate_and_infer_types()
|
||||
|
||||
set_output_size(1);
|
||||
set_output_type(0, m_output_type, data_pshape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_Bucketize_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v3::Bucketize::clone_with_new_inputs(const OutputVector& inputs) const
|
||||
@ -80,5 +82,5 @@ shared_ptr<Node> op::v3::Bucketize::clone_with_new_inputs(const OutputVector& in
|
||||
return make_shared<v3::Bucketize>(
|
||||
inputs.at(0), inputs.at(1), m_output_type, m_with_right_bound);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_Bucketize_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -39,7 +39,7 @@ shared_ptr<Node> op::Ceiling::clone_with_new_inputs(const OutputVector& new_args
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Ceiling>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Ceiling_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace ceiling
|
||||
@ -91,5 +91,5 @@ bool op::Ceiling::evaluate(const HostTensorVector& outputs, const HostTensorVect
|
||||
{
|
||||
return ceiling::evaluate_ceiling(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Ceiling_evaluate is disabled!");
|
||||
}
|
||||
|
@ -91,7 +91,7 @@ bool op::v0::Clamp::evaluate(const HostTensorVector& outputs, const HostTensorVe
|
||||
return clamp::evaluate_clamp(
|
||||
inputs[0], outputs[0], get_min(), get_max(), shape_size(get_input_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Clamp_evaluate is disabled!");
|
||||
}
|
||||
|
||||
NGRAPH_RTTI_DEFINITION(op::v0::Clamp, "Clamp", 0);
|
||||
@ -239,7 +239,7 @@ shared_ptr<Node> op::Clamp::clone_with_new_inputs(const OutputVector& new_args)
|
||||
|
||||
return make_shared<Clamp>(new_args.at(0), m_min, m_max);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Clamp_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::Clamp::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -250,5 +250,5 @@ bool op::Clamp::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("max", m_max);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Clamp_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ bool op::Concat::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("axis", m_axis);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Concat_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::Concat::validate_and_infer_types()
|
||||
@ -114,7 +114,9 @@ void op::Concat::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, inputs_et, PartialShape::dynamic(concatenation_axis_output_dim));
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Concat_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Concat::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -124,7 +126,7 @@ shared_ptr<Node> op::Concat::clone_with_new_inputs(const OutputVector& new_args)
|
||||
// TODO(amprocte): Should we check the new_args count here?
|
||||
return make_shared<Concat>(new_args, m_axis);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Concat_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -162,5 +164,5 @@ bool op::Concat::evaluate(const HostTensorVector& outputs, const HostTensorVecto
|
||||
auto concat_axis = get_axis() < 0 ? get_axis() + inputs[0]->get_shape().size() : get_axis();
|
||||
return evaluate_concat(inputs, outputs[0], concat_axis);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Concat_evaluate is disabled!");
|
||||
}
|
||||
|
@ -554,7 +554,7 @@ shared_ptr<Node> op::Constant::clone_with_new_inputs(const OutputVector& new_arg
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Constant>(*this);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Constant_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
template <typename T>
|
||||
@ -641,7 +641,7 @@ bool op::v0::Constant::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("value", m_data);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Constant_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
bool op::v0::Constant::evaluate(const HostTensorVector& outputs,
|
||||
@ -653,7 +653,7 @@ bool op::v0::Constant::evaluate(const HostTensorVector& outputs,
|
||||
output->write(get_data_ptr(), output->get_size_in_bytes());
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Constant_evaluate is disabled!");
|
||||
}
|
||||
|
||||
//
|
||||
|
@ -37,7 +37,9 @@ void op::Convert::validate_and_infer_types()
|
||||
NGRAPH_OP_SCOPE(v0_Convert_validate_and_infer_types)
|
||||
{
|
||||
set_output_type(0, m_destination_type, get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Convert_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::Convert::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -47,7 +49,7 @@ bool op::Convert::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("destination_type", m_destination_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Convert_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Convert::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -57,7 +59,7 @@ shared_ptr<Node> op::Convert::clone_with_new_inputs(const OutputVector& new_args
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Convert>(new_args.at(0), m_destination_type);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Convert_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace convert
|
||||
@ -134,5 +136,5 @@ bool op::v0::Convert::evaluate(const HostTensorVector& output_values,
|
||||
{
|
||||
return convert::evaluate_convert(input_values[0], output_values[0]);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Convert_evaluate is disabled!");
|
||||
}
|
||||
|
@ -35,13 +35,15 @@ void op::v1::ConvertLike::validate_and_infer_types()
|
||||
NGRAPH_OP_SCOPE(v1_ConvertLike_validate_and_infer_types)
|
||||
{
|
||||
set_output_type(0, get_input_element_type(1), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_ConvertLike_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::ConvertLike::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_ConvertLike_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_ConvertLike_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::ConvertLike::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -51,5 +53,5 @@ shared_ptr<Node> op::v1::ConvertLike::clone_with_new_inputs(const OutputVector&
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<ConvertLike>(new_args.at(0), new_args.at(1));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ConvertLike_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -56,7 +56,7 @@ bool op::v1::Convolution::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_pad", m_auto_pad);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Convolution_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::Convolution::validate_and_infer_types()
|
||||
@ -149,7 +149,9 @@ void op::v1::Convolution::validate_and_infer_types()
|
||||
m_dilations);
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_Convolution_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::Convolution::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -165,7 +167,7 @@ shared_ptr<Node> op::v1::Convolution::clone_with_new_inputs(const OutputVector&
|
||||
m_dilations,
|
||||
m_auto_pad);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Convolution_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
constexpr NodeTypeInfo op::v1::ConvolutionBackpropData::type_info;
|
||||
@ -206,7 +208,7 @@ bool op::v1::ConvolutionBackpropData::visit_attributes(AttributeVisitor& visitor
|
||||
visitor.on_attribute("output_padding", m_output_padding);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_ConvolutionBackpropData_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
op::v1::ConvolutionBackpropData::ConvolutionBackpropData(const Output<Node>& data,
|
||||
@ -447,7 +449,9 @@ void op::v1::ConvolutionBackpropData::validate_and_infer_types()
|
||||
set_input_is_relevant_to_shape(0);
|
||||
set_input_is_relevant_to_shape(1);
|
||||
set_output_type(0, result_et, output_pshape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_ConvolutionBackpropData_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -480,5 +484,5 @@ shared_ptr<Node>
|
||||
m_output_padding);
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ConvolutionBackpropData_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -38,7 +38,7 @@ op::Cos::Cos(const Output<Node>& arg)
|
||||
bool op::Cos::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Cos_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Cos_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Cos::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -48,7 +48,7 @@ shared_ptr<Node> op::Cos::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Cos>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Cos_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace cosop
|
||||
@ -87,5 +87,5 @@ bool op::Cos::evaluate(const HostTensorVector& outputs, const HostTensorVector&
|
||||
{
|
||||
return cosop::evaluate_cos(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Cos_evaluate is disabled!");
|
||||
}
|
||||
|
@ -37,7 +37,7 @@ op::Cosh::Cosh(const Output<Node>& arg)
|
||||
bool op::Cosh::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Cosh_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Cosh_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Cosh::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -47,7 +47,7 @@ shared_ptr<Node> op::Cosh::clone_with_new_inputs(const OutputVector& new_args) c
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Cosh>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Cosh_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace coshop
|
||||
@ -86,5 +86,5 @@ bool op::Cosh::evaluate(const HostTensorVector& outputs, const HostTensorVector&
|
||||
{
|
||||
return coshop::evaluate_cosh(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Cosh_evaluate is disabled!");
|
||||
}
|
||||
|
@ -98,7 +98,9 @@ void op::CTCGreedyDecoder::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, input_et, PartialShape{batch_size, time_size, 1, 1});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_CTCGreedyDecoder_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::CTCGreedyDecoder::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -108,7 +110,7 @@ bool op::CTCGreedyDecoder::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("ctc_merge_repeated", m_ctc_merge_repeated);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_CTCGreedyDecoder_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::CTCGreedyDecoder::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -118,5 +120,5 @@ shared_ptr<Node> op::CTCGreedyDecoder::clone_with_new_inputs(const OutputVector&
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<CTCGreedyDecoder>(new_args.at(0), new_args.at(1), m_ctc_merge_repeated);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_CTCGreedyDecoder_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -229,7 +229,9 @@ void op::v4::CTCLoss::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, logits_type, PartialShape{Dimension::dynamic()});
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_CTCLoss_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v4::CTCLoss::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -241,7 +243,7 @@ bool op::v4::CTCLoss::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("unique", unique_);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_CTCLoss_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v4::CTCLoss::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -275,5 +277,5 @@ shared_ptr<Node> op::v4::CTCLoss::clone_with_new_inputs(const OutputVector& new_
|
||||
throw ngraph_error("Incorrect number of arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_CTCLoss_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -53,7 +53,7 @@ bool op::v0::CumSum::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("reverse", m_reverse);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_CumSum_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::CumSum::validate_and_infer_types()
|
||||
@ -76,7 +76,9 @@ void op::v0::CumSum::validate_and_infer_types()
|
||||
"axis element type must be either int64_t or int32_t but got (",
|
||||
axis_type,
|
||||
").");
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_CumSum_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::CumSum::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -86,7 +88,7 @@ shared_ptr<Node> op::v0::CumSum::clone_with_new_inputs(const OutputVector& new_a
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::CumSum>(new_args.at(0), new_args.at(1), m_exclusive, m_reverse);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_CumSum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::CumSum::get_default_value() const
|
||||
|
@ -62,7 +62,7 @@ bool op::v1::DeformableConvolution::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("deformable_group", m_deformable_group);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_DeformableConvolution_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::DeformableConvolution::validate_and_infer_types()
|
||||
@ -201,7 +201,9 @@ void op::v1::DeformableConvolution::validate_and_infer_types()
|
||||
m_dilations);
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_DeformableConvolution_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -221,5 +223,5 @@ shared_ptr<Node>
|
||||
m_group,
|
||||
m_deformable_group);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_DeformableConvolution_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -83,7 +83,7 @@ bool op::v1::DeformablePSROIPooling::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("part_size", m_part_size);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_DeformablePSROIPooling_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::DeformablePSROIPooling::validate_and_infer_types()
|
||||
@ -131,7 +131,9 @@ void op::v1::DeformablePSROIPooling::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, input_et, PartialShape(output_dim_vec));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_DeformablePSROIPooling_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -172,5 +174,5 @@ shared_ptr<Node>
|
||||
throw ngraph_error("Not supported number of DeformablePSROIPooling args");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_DeformablePSROIPooling_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -60,7 +60,7 @@ bool op::DepthToSpace::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("mode", m_mode);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_DepthToSpace_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::DepthToSpace::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -73,7 +73,7 @@ shared_ptr<Node> op::DepthToSpace::clone_with_new_inputs(const OutputVector& new
|
||||
}
|
||||
return make_shared<DepthToSpace>(new_args.at(0), m_mode, m_blocksize);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_DepthToSpace_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
void op::DepthToSpace::validate_and_infer_types()
|
||||
@ -121,7 +121,9 @@ void op::DepthToSpace::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, data_type, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_DepthToSpace_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::DepthToSpace::evaluate_depth_to_space(const HostTensorVector& outputs,
|
||||
@ -255,7 +257,7 @@ bool op::DepthToSpace::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_DepthToSpace_evaluate) { return evaluate_depth_to_space(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v0_DepthToSpace_evaluate is disabled!");
|
||||
}
|
||||
namespace ngraph
|
||||
{
|
||||
|
@ -268,7 +268,9 @@ void op::DetectionOutput::validate_and_infer_types()
|
||||
output_shape.push_back(7);
|
||||
|
||||
set_output_type(0, box_logits_et, output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_DetectionOutput_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::DetectionOutput::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -297,7 +299,7 @@ shared_ptr<Node> op::DetectionOutput::clone_with_new_inputs(const OutputVector&
|
||||
m_attrs);
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_DetectionOutput_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::DetectionOutput::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -322,5 +324,5 @@ bool op::DetectionOutput::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("objectness_score", m_attrs.objectness_score);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_DetectionOutput_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -98,7 +98,7 @@ bool op::v1::Divide::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("m_pythondiv", m_pythondiv);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Divide_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::Divide::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -109,7 +109,7 @@ shared_ptr<Node> op::v1::Divide::clone_with_new_inputs(const OutputVector& new_a
|
||||
return make_shared<op::v1::Divide>(
|
||||
new_args.at(0), new_args.at(1), this->is_pythondiv(), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Divide_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Divide::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
@ -119,5 +119,5 @@ bool op::v1::Divide::evaluate(const HostTensorVector& outputs, const HostTensorV
|
||||
return divide::evaluate_divide(
|
||||
inputs[0], inputs[1], outputs[0], get_autob(), is_pythondiv());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Divide_evaluate is disabled!");
|
||||
}
|
||||
|
@ -39,7 +39,7 @@ bool ngraph::op::v0::Elu::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("alpha", m_alpha);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Elu_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::Elu::validate_and_infer_types()
|
||||
@ -48,7 +48,9 @@ void op::v0::Elu::validate_and_infer_types()
|
||||
{
|
||||
set_output_size(1);
|
||||
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Elu_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Elu::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -58,5 +60,5 @@ shared_ptr<Node> op::Elu::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Elu>(new_args.at(0), m_alpha);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Elu_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -180,7 +180,9 @@ void op::v3::EmbeddingSegmentsSum::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_EmbeddingSegmentsSum_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -213,5 +215,5 @@ shared_ptr<Node>
|
||||
throw ngraph_error("Incorrect number of arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_EmbeddingSegmentsSum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -73,5 +73,5 @@ shared_ptr<Node>
|
||||
throw ngraph_error("Incorrect number of arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_EmbeddingBagOffsetsSum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -56,5 +56,5 @@ shared_ptr<Node>
|
||||
throw ngraph_error("Incorrect number of arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_EmbeddingBagPackedSum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Equal::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Equal>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Equal_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Equal::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
@ -91,5 +91,5 @@ bool op::v1::Equal::evaluate(const HostTensorVector& outputs, const HostTensorVe
|
||||
{
|
||||
return equal::evaluate_equal(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Equal_evaluate is disabled!");
|
||||
}
|
||||
|
@ -30,7 +30,7 @@ constexpr NodeTypeInfo op::Erf::type_info;
|
||||
bool ngraph::op::v0::Erf::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Erf_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Erf_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Erf::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -40,7 +40,7 @@ shared_ptr<Node> op::Erf::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Erf>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Erf_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
op::Erf::Erf(const Output<Node>& arg)
|
||||
@ -85,5 +85,5 @@ bool op::Erf::evaluate(const HostTensorVector& outputs, const HostTensorVector&
|
||||
{
|
||||
return erfop::evaluate_erf(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Erf_evaluate is disabled!");
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ op::Exp::Exp(const Output<Node>& arg)
|
||||
bool ngraph::op::v0::Exp::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Exp_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Exp_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Exp::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -46,7 +46,7 @@ shared_ptr<Node> op::Exp::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Exp>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Exp_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace expop
|
||||
@ -85,5 +85,5 @@ bool op::Exp::evaluate(const HostTensorVector& outputs, const HostTensorVector&
|
||||
{
|
||||
return expop::evaluate_exp(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Exp_evaluate is disabled!");
|
||||
}
|
||||
|
@ -148,7 +148,9 @@ void op::v3::ExtractImagePatches::validate_and_infer_types()
|
||||
|
||||
set_output_type(0, get_input_element_type(0), output_Pshape);
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_ExtractImagePatches_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v3::ExtractImagePatches::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -161,7 +163,7 @@ bool op::v3::ExtractImagePatches::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_pad", m_padding);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_ExtractImagePatches_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node>
|
||||
@ -176,5 +178,5 @@ shared_ptr<Node>
|
||||
m_patch_selection_rates,
|
||||
m_padding);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_ExtractImagePatches_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -85,7 +85,9 @@ void op::FakeQuantize::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_FakeQuantize_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v0::FakeQuantize::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -96,7 +98,7 @@ bool ngraph::op::v0::FakeQuantize::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_broadcast", m_auto_broadcast);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_FakeQuantize_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
OutputVector op::FakeQuantize::decompose_op() const
|
||||
@ -186,5 +188,5 @@ shared_ptr<Node> op::FakeQuantize::clone_with_new_inputs(const OutputVector& new
|
||||
m_levels,
|
||||
m_auto_broadcast);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_FakeQuantize_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -35,7 +35,7 @@ op::Floor::Floor(const Output<Node>& arg)
|
||||
bool ngraph::op::v0::Floor::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Floor_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Floor_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Floor::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -45,7 +45,7 @@ shared_ptr<Node> op::Floor::clone_with_new_inputs(const OutputVector& new_args)
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Floor>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Floor_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace floorop
|
||||
@ -97,5 +97,5 @@ bool op::Floor::evaluate(const HostTensorVector& outputs, const HostTensorVector
|
||||
{
|
||||
return floorop::evaluate_floor(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Floor_evaluate is disabled!");
|
||||
}
|
||||
|
@ -39,7 +39,7 @@ shared_ptr<Node> op::v1::FloorMod::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<FloorMod>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_FloorMod_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace floor_mod
|
||||
@ -90,11 +90,11 @@ bool op::v1::FloorMod::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return floor_mod::evaluate_floor_mod(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_FloorMod_evaluate is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::FloorMod::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_FloorMod_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_FloorMod_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -48,7 +48,7 @@ op::v1::Gather::Gather(const Output<Node>& params,
|
||||
bool ngraph::op::v1::Gather::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_Gather_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_Gather_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::Gather::validate_and_infer_types()
|
||||
@ -115,7 +115,9 @@ void op::v1::Gather::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_Gather_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
int64_t op::v1::Gather::get_axis() const
|
||||
@ -145,7 +147,7 @@ shared_ptr<Node> op::v1::Gather::clone_with_new_inputs(const OutputVector& new_a
|
||||
return make_shared<v1::Gather>(
|
||||
new_args.at(PARAMS), new_args.at(INDICES), new_args.at(AXIS));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Gather_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace gather
|
||||
@ -323,7 +325,7 @@ bool op::v1::Gather::evaluate_gather(const HostTensorVector& outputs,
|
||||
bool op::v1::Gather::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_Gather_evaluate) { return evaluate_gather(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v1_Gather_evaluate is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Gather::constant_fold(OutputVector& output_values, const OutputVector& input_values)
|
||||
|
@ -119,7 +119,9 @@ void op::v6::GatherElements::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, data_type, output_pshape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v6_GatherElements_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v6::GatherElements::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -129,7 +131,7 @@ bool op::v6::GatherElements::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("axis", m_axis);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v6_GatherElements_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v6::GatherElements::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -139,5 +141,5 @@ shared_ptr<Node> op::v6::GatherElements::clone_with_new_inputs(const OutputVecto
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v6::GatherElements>(new_args.at(0), new_args.at(1), m_axis);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v6_GatherElements_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -149,7 +149,9 @@ void op::v5::GatherND::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, data_type, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_GatherND_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v5::GatherND::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -159,7 +161,7 @@ bool op::v5::GatherND::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("batch_dims", m_batch_dims);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_GatherND_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v5::GatherND::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -169,5 +171,5 @@ shared_ptr<Node> op::v5::GatherND::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v5::GatherND>(new_args.at(0), new_args.at(1), m_batch_dims);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_GatherND_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -40,13 +40,13 @@ shared_ptr<Node> op::v1::GatherTree::clone_with_new_inputs(const OutputVector& n
|
||||
return make_shared<v1::GatherTree>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_GatherTree_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v1::GatherTree::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_GatherTree_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_GatherTree_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::GatherTree::validate_and_infer_types()
|
||||
@ -88,5 +88,7 @@ void op::v1::GatherTree::validate_and_infer_types()
|
||||
|
||||
const auto& step_ids_et = get_input_element_type(0);
|
||||
set_output_type(0, step_ids_et, step_ids_rank);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_GatherTree_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
@ -43,7 +43,7 @@ op::Gelu::Gelu(const Output<Node>& data)
|
||||
bool ngraph::op::v0::Gelu::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Gelu_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Gelu_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
// f(x) = 0.5 * x * (1.0 + erf( x / sqrt(2.0) )
|
||||
@ -77,7 +77,7 @@ shared_ptr<Node> op::Gelu::clone_with_new_inputs(const OutputVector& new_args) c
|
||||
}
|
||||
return make_shared<Gelu>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Gelu_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
void op::Gelu::pre_validate_and_infer_types()
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Greater::clone_with_new_inputs(const OutputVector& new_
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Greater>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Greater_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Greater::evaluate(const HostTensorVector& outputs,
|
||||
@ -92,5 +92,5 @@ bool op::v1::Greater::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return greaterop::evaluate_greater(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Greater_evaluate is disabled!");
|
||||
}
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::GreaterEqual::clone_with_new_inputs(const OutputVector&
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::GreaterEqual>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_GreaterEqual_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::GreaterEqual::evaluate(const HostTensorVector& outputs,
|
||||
@ -93,5 +93,5 @@ bool op::v1::GreaterEqual::evaluate(const HostTensorVector& outputs,
|
||||
return greater_equalop::evaluate_greater_equal(
|
||||
inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_GreaterEqual_evaluate is disabled!");
|
||||
}
|
||||
|
@ -49,7 +49,7 @@ bool ngraph::op::v0::GRN::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("bias", m_bias);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_GRN_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::GRN::pre_validate_and_infer_types()
|
||||
@ -109,5 +109,5 @@ shared_ptr<Node> op::GRN::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
}
|
||||
return make_shared<GRN>(new_args.at(0), m_bias);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_GRN_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -70,7 +70,7 @@ bool ngraph::op::v1::GroupConvolution::visit_attributes(AttributeVisitor& visito
|
||||
visitor.on_attribute("auto_pad", m_auto_pad);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_GroupConvolution_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::GroupConvolution::validate_and_infer_types()
|
||||
@ -179,7 +179,9 @@ void op::v1::GroupConvolution::validate_and_infer_types()
|
||||
m_dilations);
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_GroupConvolution_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::GroupConvolution::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -195,7 +197,7 @@ shared_ptr<Node> op::v1::GroupConvolution::clone_with_new_inputs(const OutputVec
|
||||
m_dilations,
|
||||
m_auto_pad);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_GroupConvolution_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
//------------------------------------------------------------------------------
|
||||
@ -277,7 +279,7 @@ bool ngraph::op::v1::GroupConvolutionBackpropData::visit_attributes(AttributeVis
|
||||
visitor.on_attribute("output_padding", m_output_padding);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_GroupConvolutionBackpropData_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::GroupConvolutionBackpropData::is_dynamic() const
|
||||
@ -596,5 +598,5 @@ shared_ptr<Node>
|
||||
m_output_padding);
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_GroupConvolutionBackpropData_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -108,7 +108,7 @@ bool op::v3::GRUCell::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("linear_before_reset", m_linear_before_reset);
|
||||
return op::util::RNNCellBase::visit_attributes(visitor);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_GRUCell_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v3::GRUCell::validate_and_infer_types()
|
||||
@ -209,7 +209,9 @@ void op::v3::GRUCell::validate_and_infer_types()
|
||||
// Set output size, type and shape
|
||||
set_output_size(1);
|
||||
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_GRUCell_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
void op::v3::GRUCell::add_default_bias_input()
|
||||
@ -258,5 +260,5 @@ shared_ptr<Node> op::v3::GRUCell::clone_with_new_inputs(const OutputVector& new_
|
||||
throw ngraph_error("Incorrect number of new arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_GRUCell_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -179,7 +179,9 @@ void op::v5::GRUSequence::validate_and_infer_types()
|
||||
{merged_batch_size, merged_num_directions, x_pshape[1], merged_hidden_size});
|
||||
set_output_type(
|
||||
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_GRUSequence_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool op::v5::GRUSequence::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -190,7 +192,7 @@ bool op::v5::GRUSequence::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("linear_before_reset", m_linear_before_reset);
|
||||
return op::util::RNNCellBase::visit_attributes(visitor);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_GRUSequence_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v5::GRUSequence::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -212,5 +214,5 @@ shared_ptr<Node> op::v5::GRUSequence::clone_with_new_inputs(const OutputVector&
|
||||
m_clip,
|
||||
m_linear_before_reset);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_GRUSequence_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -43,7 +43,7 @@ op::HardSigmoid::HardSigmoid(const Output<Node>& data,
|
||||
bool ngraph::op::v0::HardSigmoid::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_HardSigmoid_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_HardSigmoid_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::HardSigmoid::pre_validate_and_infer_types()
|
||||
@ -110,5 +110,5 @@ shared_ptr<Node> op::HardSigmoid::clone_with_new_inputs(const OutputVector& new_
|
||||
|
||||
return make_shared<HardSigmoid>(new_args.at(0), new_args.at(1), new_args.at(2));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_HardSigmoid_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ op::v5::HSigmoid::HSigmoid(const Output<Node>& arg)
|
||||
bool op::v5::HSigmoid::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v5_HSigmoid_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v5_HSigmoid_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v5::HSigmoid::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -45,7 +45,7 @@ shared_ptr<Node> op::v5::HSigmoid::clone_with_new_inputs(const OutputVector& new
|
||||
{
|
||||
return make_shared<op::v5::HSigmoid>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_HSigmoid_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -82,5 +82,5 @@ bool op::v5::HSigmoid::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return evaluate_hsigmoid(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_HSigmoid_evaluate is disabled!");
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ op::v4::HSwish::HSwish(const Output<Node>& arg)
|
||||
bool op::v4::HSwish::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v4_HSwish_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v4_HSwish_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v4::HSwish::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -45,7 +45,7 @@ shared_ptr<Node> op::v4::HSwish::clone_with_new_inputs(const OutputVector& new_a
|
||||
{
|
||||
return make_shared<op::v4::HSwish>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_HSwish_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace hswish
|
||||
@ -81,5 +81,5 @@ bool op::v4::HSwish::evaluate(const HostTensorVector& outputs, const HostTensorV
|
||||
{
|
||||
return hswish::evaluate_hswish(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_HSwish_evaluate is disabled!");
|
||||
}
|
||||
|
@ -49,7 +49,7 @@ bool op::v0::Interpolate::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("pads_end", m_attrs.pads_end);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Interpolate_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::Interpolate::validate_and_infer_types()
|
||||
@ -81,7 +81,9 @@ void op::v0::Interpolate::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, get_input_element_type(0), output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Interpolate_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::Interpolate::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -91,7 +93,7 @@ shared_ptr<Node> op::v0::Interpolate::clone_with_new_inputs(const OutputVector&
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v0::Interpolate>(new_args.at(0), new_args.at(1), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Interpolate_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace ngraph
|
||||
@ -157,7 +159,7 @@ bool op::v4::Interpolate::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("cube_coeff", m_attrs.cube_coeff);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_Interpolate_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
std::vector<int64_t> op::v4::Interpolate::get_axes() const
|
||||
@ -290,7 +292,9 @@ void op::v4::Interpolate::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, get_input_element_type(0), output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_Interpolate_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v4::Interpolate::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -306,7 +310,7 @@ shared_ptr<Node> op::v4::Interpolate::clone_with_new_inputs(const OutputVector&
|
||||
return make_shared<op::v4::Interpolate>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_Interpolate_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -523,7 +527,7 @@ bool op::v4::Interpolate::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v4_Interpolate_evaluate) { return evaluate_interpolate(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v4_Interpolate_evaluate is disabled!");
|
||||
}
|
||||
|
||||
namespace ngraph
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::Less::clone_with_new_inputs(const OutputVector& new_arg
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Less>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Less_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Less::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
@ -91,5 +91,5 @@ bool op::v1::Less::evaluate(const HostTensorVector& outputs, const HostTensorVec
|
||||
{
|
||||
return lessop::evaluate_less(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Less_evaluate is disabled!");
|
||||
}
|
||||
|
@ -43,7 +43,7 @@ shared_ptr<Node> op::v1::LessEqual::clone_with_new_inputs(const OutputVector& ne
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<v1::LessEqual>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_LessEqual_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace less_equalop
|
||||
@ -92,5 +92,5 @@ bool op::v1::LessEqual::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return less_equalop::evaluate_less_equal(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_LessEqual_evaluate is disabled!");
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ op::Log::Log(const Output<Node>& arg)
|
||||
bool ngraph::op::v0::Log::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Log_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Log_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Log::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -46,7 +46,7 @@ shared_ptr<Node> op::Log::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Log>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Log_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace logop
|
||||
@ -85,5 +85,5 @@ bool op::Log::evaluate(const HostTensorVector& outputs, const HostTensorVector&
|
||||
{
|
||||
return logop::evaluate_log(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Log_evaluate is disabled!");
|
||||
}
|
||||
|
@ -38,7 +38,7 @@ bool op::v5::LogSoftmax::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("axis", m_axis);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_LogSoftmax_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v5::LogSoftmax::validate_and_infer_types()
|
||||
@ -57,7 +57,9 @@ void op::v5::LogSoftmax::validate_and_infer_types()
|
||||
").");
|
||||
|
||||
set_output_type(0, get_input_element_type(0), input_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_LogSoftmax_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v5::LogSoftmax::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -67,5 +69,5 @@ shared_ptr<Node> op::v5::LogSoftmax::clone_with_new_inputs(const OutputVector& n
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v5::LogSoftmax>(new_args.at(0), m_axis);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_LogSoftmax_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -45,7 +45,7 @@ bool op::v5::Loop::visit_attributes(AttributeVisitor& visitor)
|
||||
|
||||
return false;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_Loop_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v5::Loop::validate_and_infer_types()
|
||||
@ -310,7 +310,9 @@ void op::v5::Loop::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_Loop_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
std::shared_ptr<Node> op::v5::Loop::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -392,7 +394,7 @@ std::shared_ptr<Node> op::v5::Loop::clone_with_new_inputs(const OutputVector& ne
|
||||
}
|
||||
return move(op);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_Loop_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
Output<Node> op::v5::Loop::get_concatenated_slices(const Output<Node>& value,
|
||||
@ -421,5 +423,5 @@ bool op::v5::Loop::evaluate(const HostTensorVector& outputs, const HostTensorVec
|
||||
inputs);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_Loop_evaluate is disabled!");
|
||||
}
|
||||
|
@ -113,7 +113,9 @@ void op::LRN::validate_and_infer_types()
|
||||
"Axes input must be integral numbers, but are: ",
|
||||
axes_type,
|
||||
").");
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_LRN_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v0::LRN::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -126,7 +128,7 @@ bool ngraph::op::v0::LRN::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("size", m_size);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_LRN_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::LRN::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -137,5 +139,5 @@ shared_ptr<Node> op::LRN::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
return make_shared<op::LRN>(
|
||||
new_args.at(0), new_args.at(1), m_alpha, m_beta, m_bias, m_size);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_LRN_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -142,7 +142,7 @@ bool ngraph::op::v0::LSTMCell::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("weights_format", m_weights_format);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_LSTMCell_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::LSTMCell::validate_and_infer_types()
|
||||
@ -297,7 +297,9 @@ void op::v0::LSTMCell::validate_and_infer_types()
|
||||
set_output_size(2);
|
||||
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
|
||||
set_output_type(1, result_et, {merged_batch_size, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_LSTMCell_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
Output<Node> op::v0::LSTMCell::get_default_bias_input() const
|
||||
@ -371,7 +373,7 @@ shared_ptr<Node> op::v0::LSTMCell::clone_with_new_inputs(const OutputVector& new
|
||||
throw ngraph_error("Incorrect number of new arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_LSTMCell_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace ngraph
|
||||
@ -459,7 +461,7 @@ bool ngraph::op::v4::LSTMCell::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
return op::util::RNNCellBase::visit_attributes(visitor);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_LSTMCell_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v4::LSTMCell::validate_and_infer_types()
|
||||
@ -573,7 +575,9 @@ void op::v4::LSTMCell::validate_and_infer_types()
|
||||
set_output_size(2);
|
||||
set_output_type(0, result_et, {merged_batch_size, merged_hidden_size});
|
||||
set_output_type(1, result_et, {merged_batch_size, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_LSTMCell_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
Output<Node> op::v4::LSTMCell::get_default_bias_input() const
|
||||
@ -619,5 +623,5 @@ shared_ptr<Node> op::v4::LSTMCell::clone_with_new_inputs(const OutputVector& new
|
||||
throw ngraph_error("Incorrect number of new arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_LSTMCell_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -48,7 +48,7 @@ bool ngraph::op::v0::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("weights_format", m_weights_format);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_LSTMSequence_visit_attributes is disabled!");
|
||||
}
|
||||
OutputVector op::v0::LSTMSequence::decompose_op() const
|
||||
{
|
||||
@ -121,7 +121,7 @@ shared_ptr<Node> op::v0::LSTMSequence::clone_with_new_inputs(const OutputVector&
|
||||
throw ngraph_error("Incorrect number of new arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_LSTMSequence_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::LSTMSequence::get_masked_node(const Output<Node>& data,
|
||||
@ -437,7 +437,9 @@ void op::v0::LSTMSequence::validate_and_infer_types()
|
||||
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
|
||||
set_output_type(
|
||||
2, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_LSTMSequence_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v5::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -447,7 +449,7 @@ bool ngraph::op::v5::LSTMSequence::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("direction", m_direction);
|
||||
return op::util::RNNCellBase::visit_attributes(visitor);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_LSTMSequence_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v5::LSTMSequence::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -476,7 +478,7 @@ shared_ptr<Node> op::v5::LSTMSequence::clone_with_new_inputs(const OutputVector&
|
||||
throw ngraph_error("Incorrect number of new arguments");
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_LSTMSequence_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
void op::v5::LSTMSequence::validate_and_infer_types()
|
||||
@ -622,5 +624,7 @@ void op::v5::LSTMSequence::validate_and_infer_types()
|
||||
1, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
|
||||
set_output_type(
|
||||
2, result_et, {merged_batch_size, merged_num_directions, merged_hidden_size});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_LSTMSequence_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ bool ngraph::op::v0::MatMul::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("transpose_b", m_transpose_b);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_MatMul_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::MatMul::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -56,7 +56,7 @@ shared_ptr<Node> op::MatMul::clone_with_new_inputs(const OutputVector& new_args)
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<MatMul>(new_args.at(0), new_args.at(1), m_transpose_a, m_transpose_b);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_MatMul_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace matmul
|
||||
@ -272,7 +272,7 @@ bool op::MatMul::evaluate(const HostTensorVector& outputs, const HostTensorVecto
|
||||
return matmul::evaluate_matmul(
|
||||
inputs[0], inputs[1], outputs[0], get_transpose_a(), get_transpose_b());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_MatMul_evaluate is disabled!");
|
||||
}
|
||||
|
||||
void ngraph::op::v0::MatMul::validate_and_infer_types()
|
||||
@ -309,5 +309,7 @@ void ngraph::op::v0::MatMul::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, result_et, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_MatMul_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
@ -75,7 +75,7 @@ shared_ptr<Node> op::v1::ReduceMax::clone_with_new_inputs(const OutputVector& ne
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::ReduceMax>(new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ReduceMax_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::ReduceMax::evaluate(const HostTensorVector& outputs,
|
||||
@ -85,5 +85,5 @@ bool op::v1::ReduceMax::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return maxop::evaluate_max(inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_ReduceMax_evaluate is disabled!");
|
||||
}
|
||||
|
@ -80,7 +80,7 @@ bool ngraph::op::v1::MaxPool::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_pad", m_auto_pad);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_MaxPool_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::MaxPool::validate_and_infer_types()
|
||||
@ -138,7 +138,9 @@ void op::v1::MaxPool::validate_and_infer_types()
|
||||
true,
|
||||
m_rounding_type == op::RoundingType::CEIL)
|
||||
: output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_MaxPool_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::MaxPool::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -154,7 +156,7 @@ shared_ptr<Node> op::v1::MaxPool::clone_with_new_inputs(const OutputVector& new_
|
||||
m_rounding_type,
|
||||
m_auto_pad);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_MaxPool_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::MaxPool::get_default_value() const
|
||||
@ -247,5 +249,5 @@ bool op::v1::MaxPool::evaluate(const HostTensorVector& outputs,
|
||||
const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_MaxPool_evaluate) { return evaluate_maxpool(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v1_MaxPool_evaluate is disabled!");
|
||||
}
|
||||
|
@ -89,7 +89,7 @@ shared_ptr<Node> op::v1::Maximum::clone_with_new_inputs(const OutputVector& new_
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Maximum>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Maximum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Maximum::evaluate(const HostTensorVector& outputs,
|
||||
@ -99,5 +99,5 @@ bool op::v1::Maximum::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return maximumop::evaluate_maximum(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Maximum_evaluate is disabled!");
|
||||
}
|
||||
|
@ -77,7 +77,7 @@ shared_ptr<Node> op::v1::ReduceMin::clone_with_new_inputs(const OutputVector& ne
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::ReduceMin>(new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ReduceMin_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::ReduceMin::evaluate(const HostTensorVector& outputs,
|
||||
@ -87,5 +87,5 @@ bool op::v1::ReduceMin::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return minop::evaluate_min(inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_ReduceMin_evaluate is disabled!");
|
||||
}
|
||||
|
@ -87,7 +87,7 @@ shared_ptr<Node> op::v1::Minimum::clone_with_new_inputs(const OutputVector& new_
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Minimum>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Minimum_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Minimum::evaluate(const HostTensorVector& outputs,
|
||||
@ -97,5 +97,5 @@ bool op::v1::Minimum::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return minimumop::evaluate_minimum(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Minimum_evaluate is disabled!");
|
||||
}
|
||||
|
@ -36,7 +36,7 @@ op::v4::Mish::Mish(const Output<Node>& arg)
|
||||
bool op::v4::Mish::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v4_Mish_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v4_Mish_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v4::Mish::validate_and_infer_types()
|
||||
@ -45,7 +45,9 @@ void op::v4::Mish::validate_and_infer_types()
|
||||
{
|
||||
set_output_size(1);
|
||||
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_Mish_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v4::Mish::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -55,7 +57,7 @@ shared_ptr<Node> op::v4::Mish::clone_with_new_inputs(const OutputVector& new_arg
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Mish>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_Mish_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace mish
|
||||
@ -89,5 +91,5 @@ bool op::v4::Mish::evaluate(const HostTensorVector& outputs, const HostTensorVec
|
||||
{
|
||||
return mish::evaluate_mish(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_Mish_evaluate is disabled!");
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ bool ngraph::op::v1::Mod::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("auto_broadcast", m_auto_broadcast);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Mod_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
OutputVector op::v1::Mod::decompose_op() const
|
||||
@ -75,5 +75,5 @@ shared_ptr<Node> op::v1::Mod::clone_with_new_inputs(const OutputVector& new_args
|
||||
{
|
||||
return make_shared<Mod>(new_args.at(0), new_args.at(1), m_auto_broadcast);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Mod_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v0::Multiply::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v0::Multiply>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Multiply_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v0::Multiply::evaluate(const HostTensorVector& outputs,
|
||||
@ -92,7 +92,7 @@ bool op::v0::Multiply::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return multiplyop::evaluate_multiply(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Multiply_evaluate is disabled!");
|
||||
}
|
||||
|
||||
// ------------------------------------ v1 -------------------------------------
|
||||
@ -114,7 +114,7 @@ shared_ptr<Node> op::v1::Multiply::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Multiply>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Multiply_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Multiply::evaluate(const HostTensorVector& outputs,
|
||||
@ -124,5 +124,5 @@ bool op::v1::Multiply::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return multiplyop::evaluate_multiply(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Multiply_evaluate is disabled!");
|
||||
}
|
||||
|
@ -78,7 +78,9 @@ void op::MVN::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_MVN_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
OutputVector op::MVN::decompose_op() const
|
||||
@ -119,7 +121,7 @@ shared_ptr<Node> op::MVN::clone_with_new_inputs(const OutputVector& new_args) co
|
||||
new_args.size());
|
||||
return make_shared<MVN>(new_args.at(0), m_reduction_axes, m_normalize_variance, m_eps);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_MVN_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::MVN::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -132,7 +134,7 @@ bool op::MVN::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("reduction_axes", m_reduction_axes);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_MVN_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
// ------------------------------ V6 ------------------------------
|
||||
@ -194,7 +196,9 @@ void op::v6::MVN::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, get_input_element_type(0), get_input_partial_shape(0));
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v6_MVN_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v6::MVN::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -208,7 +212,7 @@ shared_ptr<Node> op::v6::MVN::clone_with_new_inputs(const OutputVector& new_args
|
||||
return make_shared<op::v6::MVN>(
|
||||
new_args.at(0), new_args.at(1), m_normalize_variance, m_eps, m_eps_mode);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v6_MVN_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v6::MVN::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -220,5 +224,5 @@ bool op::v6::MVN::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("eps_mode", m_eps_mode);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v6_MVN_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -33,7 +33,7 @@ op::Negative::Negative(const Output<Node>& arg)
|
||||
bool ngraph::op::v0::Negative::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Negative_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Negative_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Negative::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -43,7 +43,7 @@ shared_ptr<Node> op::Negative::clone_with_new_inputs(const OutputVector& new_arg
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Negative>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Negative_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace negativeop
|
||||
@ -83,7 +83,7 @@ bool op::Negative::evaluate(const HostTensorVector& outputs, const HostTensorVec
|
||||
return negativeop::evaluate_negative(
|
||||
inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Negative_evaluate is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> ngraph::operator-(const Output<Node>& arg0)
|
||||
|
@ -90,7 +90,7 @@ std::shared_ptr<Node>
|
||||
m_box_encoding,
|
||||
m_sort_result_descending);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_NonMaxSuppression_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v1::NonMaxSuppression::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -101,7 +101,7 @@ bool ngraph::op::v1::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
|
||||
visitor.on_attribute("sort_result_descending", m_sort_result_descending);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_NonMaxSuppression_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::NonMaxSuppression::validate_and_infer_types()
|
||||
@ -204,7 +204,9 @@ void op::v1::NonMaxSuppression::validate_and_infer_types()
|
||||
out_shape[0] = std::min(num_boxes, max_output_boxes_per_class * num_classes);
|
||||
}
|
||||
set_output_type(0, output_element_type, out_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_NonMaxSuppression_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
int64_t op::v1::NonMaxSuppression::max_boxes_output_from_input() const
|
||||
@ -309,7 +311,7 @@ std::shared_ptr<Node>
|
||||
m_sort_result_descending,
|
||||
m_output_type);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_NonMaxSuppression_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v3::NonMaxSuppression::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -321,7 +323,7 @@ bool ngraph::op::v3::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
|
||||
visitor.on_attribute("output_type", m_output_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_NonMaxSuppression_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v3::NonMaxSuppression::validate()
|
||||
@ -430,7 +432,9 @@ void op::v3::NonMaxSuppression::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, m_output_type, out_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_NonMaxSuppression_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
int64_t op::v3::NonMaxSuppression::max_boxes_output_from_input() const
|
||||
@ -539,7 +543,7 @@ std::shared_ptr<Node>
|
||||
m_sort_result_descending,
|
||||
m_output_type);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_NonMaxSuppression_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
void op::v4::NonMaxSuppression::validate_and_infer_types()
|
||||
@ -571,7 +575,9 @@ void op::v4::NonMaxSuppression::validate_and_infer_types()
|
||||
}
|
||||
}
|
||||
set_output_type(0, m_output_type, out_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_NonMaxSuppression_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
// ------------------------------ V5 ------------------------------
|
||||
@ -722,7 +728,7 @@ std::shared_ptr<Node>
|
||||
break;
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v5_NonMaxSuppression_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -935,7 +941,7 @@ bool ngraph::op::v5::NonMaxSuppression::visit_attributes(AttributeVisitor& visit
|
||||
visitor.on_attribute("output_type", m_output_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v5_NonMaxSuppression_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v5::NonMaxSuppression::validate_and_infer_types()
|
||||
@ -971,7 +977,9 @@ void op::v5::NonMaxSuppression::validate_and_infer_types()
|
||||
set_output_type(0, m_output_type, out_shape);
|
||||
set_output_type(1, element::f32, out_shape);
|
||||
set_output_type(2, m_output_type, Shape{1});
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v5_NonMaxSuppression_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
namespace ngraph
|
||||
|
@ -53,7 +53,7 @@ bool ngraph::op::v3::NonZero::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("output_type", m_output_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_NonZero_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v3::NonZero::validate_and_infer_types()
|
||||
@ -84,7 +84,9 @@ void op::v3::NonZero::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_input_is_relevant_to_shape(0);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_NonZero_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v3::NonZero::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -94,7 +96,7 @@ shared_ptr<Node> op::v3::NonZero::clone_with_new_inputs(const OutputVector& new_
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<v3::NonZero>(new_args.at(0), m_output_type);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_NonZero_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace nonzero
|
||||
@ -178,5 +180,5 @@ bool op::v3::NonZero::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return nonzero::evaluate_nonzero(inputs[0], outputs[0]);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_NonZero_evaluate is disabled!");
|
||||
}
|
||||
|
@ -52,7 +52,7 @@ bool ngraph::op::v0::NormalizeL2::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("eps_mode", m_eps_mode);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_NormalizeL2_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::NormalizeL2::pre_validate_and_infer_types()
|
||||
@ -129,5 +129,5 @@ shared_ptr<Node> op::NormalizeL2::clone_with_new_inputs(const OutputVector& new_
|
||||
}
|
||||
return make_shared<NormalizeL2>(new_args.at(0), new_args.at(1), m_eps, m_eps_mode);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_NormalizeL2_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -39,7 +39,7 @@ op::v1::LogicalNot::LogicalNot(const Output<Node>& arg)
|
||||
bool ngraph::op::v1::LogicalNot::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_LogicalNot_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_LogicalNot_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
// TODO(amprocte): Update this to allow only boolean, for consistency with logical binops.
|
||||
@ -52,7 +52,9 @@ void op::v1::LogicalNot::validate_and_infer_types()
|
||||
PartialShape& args_pshape = std::get<1>(args_et_pshape);
|
||||
|
||||
set_output_type(0, args_et, args_pshape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_LogicalNot_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::LogicalNot::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -62,7 +64,7 @@ shared_ptr<Node> op::v1::LogicalNot::clone_with_new_inputs(const OutputVector& n
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<v1::LogicalNot>(new_args.at(0));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_LogicalNot_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace notop
|
||||
@ -103,5 +105,5 @@ bool op::v1::LogicalNot::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return notop::evaluate_not(inputs[0], outputs[0], shape_size(get_output_shape(0)));
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_LogicalNot_evaluate is disabled!");
|
||||
}
|
||||
|
@ -82,7 +82,7 @@ shared_ptr<Node> op::v1::NotEqual::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::NotEqual>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_NotEqual_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::NotEqual::evaluate(const HostTensorVector& outputs,
|
||||
@ -92,11 +92,11 @@ bool op::v1::NotEqual::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return not_equalop::evaluate_not_equal(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_NotEqual_evaluate is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::NotEqual::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_NotEqual_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v1_NotEqual_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -120,7 +120,9 @@ void op::v1::OneHot::validate_and_infer_types()
|
||||
}
|
||||
|
||||
set_output_type(0, on_value_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_OneHot_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
bool ngraph::op::v1::OneHot::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -130,7 +132,7 @@ bool ngraph::op::v1::OneHot::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("axis", m_axis);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_OneHot_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::OneHot::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -141,7 +143,7 @@ shared_ptr<Node> op::v1::OneHot::clone_with_new_inputs(const OutputVector& new_a
|
||||
return make_shared<v1::OneHot>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), new_args.at(3), m_axis);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_OneHot_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace detail
|
||||
@ -224,5 +226,5 @@ bool op::v1::OneHot::evaluate(const HostTensorVector& output_values,
|
||||
{
|
||||
return detail::evaluate_onehot(output_values, input_values, get_axis());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_OneHot_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ shared_ptr<Node> op::v1::LogicalOr::clone_with_new_inputs(const OutputVector& ne
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<v1::LogicalOr>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_LogicalOr_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace logor
|
||||
@ -90,5 +90,5 @@ bool op::v1::LogicalOr::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return logor::evaluate_logor(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_LogicalOr_evaluate is disabled!");
|
||||
}
|
||||
|
@ -80,7 +80,7 @@ bool ngraph::op::v1::Pad::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("pad_mode", m_pad_mode);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Pad_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v1::Pad::validate_and_infer_types()
|
||||
@ -202,7 +202,9 @@ void op::v1::Pad::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, get_input_element_type(0), PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v1_Pad_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v1::Pad::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -220,7 +222,7 @@ shared_ptr<Node> op::v1::Pad::clone_with_new_inputs(const OutputVector& new_args
|
||||
return make_shared<v1::Pad>(new_args.at(0), new_args.at(1), new_args.at(2), m_pad_mode);
|
||||
}
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Pad_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Pad::evaluate_pad(const HostTensorVector& outputs,
|
||||
@ -257,5 +259,5 @@ bool op::v1::Pad::evaluate_pad(const HostTensorVector& outputs,
|
||||
bool op::v1::Pad::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v1_Pad_evaluate) { return evaluate_pad(outputs, inputs); }
|
||||
return false;
|
||||
throw ngraph_error("v1_Pad_evaluate is disabled!");
|
||||
}
|
||||
|
@ -41,7 +41,7 @@ bool op::Parameter::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("element_type", m_element_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Parameter_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::Parameter::validate_and_infer_types()
|
||||
@ -50,7 +50,9 @@ void op::Parameter::validate_and_infer_types()
|
||||
{
|
||||
Op::validate_and_infer_types();
|
||||
set_output_type(0, m_element_type, m_partial_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Parameter_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Parameter::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -60,7 +62,7 @@ shared_ptr<Node> op::Parameter::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Parameter>(m_element_type, m_partial_shape);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Parameter_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::Parameter::is_relevant_to_shapes() const
|
||||
|
@ -85,7 +85,7 @@ shared_ptr<Node> op::v1::Power::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v1::Power>(new_args.at(0), new_args.at(1), this->get_autob());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_Power_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v1::Power::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const
|
||||
@ -94,5 +94,5 @@ bool op::v1::Power::evaluate(const HostTensorVector& outputs, const HostTensorVe
|
||||
{
|
||||
return power::evaluate_power(inputs[0], inputs[1], outputs[0], get_autob());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_Power_evaluate is disabled!");
|
||||
}
|
||||
|
@ -43,7 +43,7 @@ op::PRelu::PRelu(const Output<Node>& data, const Output<Node>& slope)
|
||||
bool ngraph::op::v0::PRelu::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_PRelu_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_PRelu_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void ngraph::op::v0::PRelu::pre_validate_and_infer_types()
|
||||
@ -97,7 +97,7 @@ shared_ptr<Node> op::PRelu::clone_with_new_inputs(const OutputVector& new_args)
|
||||
}
|
||||
return make_shared<PRelu>(new_args.at(0), new_args.at(1));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_PRelu_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace prelu
|
||||
@ -136,5 +136,5 @@ bool op::PRelu::evaluate(const HostTensorVector& outputs, const HostTensorVector
|
||||
{
|
||||
return prelu::evaluate_prelu(inputs[0], inputs[1], outputs[0]);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PRelu_evaluate is disabled!");
|
||||
}
|
||||
|
@ -83,7 +83,9 @@ void op::PriorBox::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, element::f32, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_PriorBox_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::PriorBox::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -93,7 +95,7 @@ shared_ptr<Node> op::PriorBox::clone_with_new_inputs(const OutputVector& new_arg
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<PriorBox>(new_args.at(0), new_args.at(1), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_PriorBox_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
int64_t op::PriorBox::number_of_priors(const PriorBoxAttrs& attrs)
|
||||
@ -159,7 +161,7 @@ bool op::PriorBox::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("scale_all_sizes", m_attrs.scale_all_sizes);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PriorBox_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
namespace prior_box
|
||||
@ -207,5 +209,5 @@ bool op::v0::PriorBox::evaluate(const HostTensorVector& outputs,
|
||||
{
|
||||
return prior_box::evaluate_prior_box(inputs[0], inputs[1], outputs[0], get_attrs());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PriorBox_evaluate is disabled!");
|
||||
}
|
||||
|
@ -88,7 +88,9 @@ void op::PriorBoxClustered::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, element::f32, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_PriorBoxClustered_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::PriorBoxClustered::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -98,7 +100,7 @@ shared_ptr<Node> op::PriorBoxClustered::clone_with_new_inputs(const OutputVector
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<PriorBoxClustered>(new_args.at(0), new_args.at(1), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_PriorBoxClustered_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::PriorBoxClustered::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -132,7 +134,7 @@ bool op::PriorBoxClustered::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("variance", m_attrs.variances);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PriorBoxClustered_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
namespace prior_box_clustered
|
||||
@ -181,5 +183,5 @@ bool op::v0::PriorBoxClustered::evaluate(const HostTensorVector& outputs,
|
||||
return prior_box_clustered::evaluate_prior_box(
|
||||
inputs[0], inputs[1], outputs[0], get_attrs());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PriorBoxClustered_evaluate is disabled!");
|
||||
}
|
||||
|
@ -85,7 +85,9 @@ void op::v0::Proposal::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, get_input_element_type(0), PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Proposal_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::Proposal::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -96,7 +98,7 @@ shared_ptr<Node> op::v0::Proposal::clone_with_new_inputs(const OutputVector& new
|
||||
return make_shared<op::v0::Proposal>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Proposal_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v0::Proposal::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -119,7 +121,7 @@ bool op::v0::Proposal::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("framework", m_attrs.framework);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Proposal_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
NGRAPH_RTTI_DEFINITION(op::v4::Proposal, "Proposal", 4);
|
||||
@ -149,7 +151,9 @@ void op::v4::Proposal::validate_and_infer_types()
|
||||
1, get_input_element_type(0), PartialShape{batch_size * m_attrs.post_nms_topn});
|
||||
else
|
||||
set_output_type(1, get_input_element_type(0), PartialShape::dynamic());
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_Proposal_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
std::shared_ptr<Node> op::v4::Proposal::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -160,5 +164,5 @@ std::shared_ptr<Node> op::v4::Proposal::clone_with_new_inputs(const OutputVector
|
||||
return make_shared<op::v4::Proposal>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), m_attrs);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_Proposal_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -54,7 +54,7 @@ bool ngraph::op::v0::PSROIPooling::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("spatial_bins_y", m_spatial_bins_y);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_PSROIPooling_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::PSROIPooling::validate_and_infer_types()
|
||||
@ -137,7 +137,9 @@ void op::PSROIPooling::validate_and_infer_types()
|
||||
}
|
||||
set_output_type(0, feat_maps_et, output_shape);
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_PSROIPooling_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::PSROIPooling::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -154,5 +156,5 @@ shared_ptr<Node> op::PSROIPooling::clone_with_new_inputs(const OutputVector& new
|
||||
m_spatial_bins_y,
|
||||
m_mode);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_PSROIPooling_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -161,7 +161,9 @@ void op::Quantize::validate_and_infer_types()
|
||||
{
|
||||
set_output_type(0, quantized_type, PartialShape::dynamic());
|
||||
}
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Quantize_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::Quantize::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -172,5 +174,5 @@ shared_ptr<Node> op::Quantize::clone_with_new_inputs(const OutputVector& new_arg
|
||||
return make_shared<Quantize>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), m_type, m_axes, m_round_mode);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Quantize_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
@ -75,7 +75,7 @@ bool ngraph::op::v4::Range::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("output_type", m_output_type);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_Range_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v4::Range::validate_and_infer_types()
|
||||
@ -187,7 +187,9 @@ void op::v4::Range::validate_and_infer_types()
|
||||
result = PartialShape{Dimension(static_cast<int64_t>(strided))};
|
||||
}
|
||||
set_output_type(0, m_output_type, result);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v4_Range_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v4::Range::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -198,7 +200,7 @@ shared_ptr<Node> op::v4::Range::clone_with_new_inputs(const OutputVector& new_ar
|
||||
return make_shared<v4::Range>(
|
||||
new_args.at(0), new_args.at(1), new_args.at(2), m_output_type);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_Range_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
template <typename T>
|
||||
@ -323,7 +325,7 @@ bool op::v4::Range::evaluate(const HostTensorVector& outputs, const HostTensorVe
|
||||
HostTensorPtr step = inputs[2];
|
||||
return rangeop::evaluate_power(out, start, stop, step, m_output_type, 4);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_Range_evaluate is disabled!");
|
||||
}
|
||||
|
||||
constexpr NodeTypeInfo op::v0::Range::type_info;
|
||||
@ -437,7 +439,7 @@ static PartialShape infer_output_shape(const op::v0::Range* node, const element:
|
||||
bool ngraph::op::v0::Range::visit_attributes(AttributeVisitor& visitor)
|
||||
{
|
||||
NGRAPH_OP_SCOPE(v0_Range_visit_attributes) { return true; }
|
||||
return false;
|
||||
throw ngraph_error("v0_Range_visit_attributes is disabled!");
|
||||
}
|
||||
|
||||
void op::v0::Range::validate_and_infer_types()
|
||||
@ -522,7 +524,9 @@ void op::v0::Range::validate_and_infer_types()
|
||||
#endif
|
||||
|
||||
set_output_type(0, result_et, result_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v0_Range_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::v0::Range::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -532,7 +536,7 @@ shared_ptr<Node> op::v0::Range::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<Range>(new_args.at(0), new_args.at(1), new_args.at(2));
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v0_Range_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
template <element::Type_t ET, typename T>
|
||||
@ -550,5 +554,5 @@ bool op::v0::Range::evaluate(const HostTensorVector& outputs, const HostTensorVe
|
||||
HostTensorPtr step = inputs[2];
|
||||
return rangeop::evaluate_power(out, start, stop, step, start->get_element_type(), 0);
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v0_Range_evaluate is disabled!");
|
||||
}
|
||||
|
@ -42,7 +42,9 @@ void op::ReadValue::validate_and_infer_types()
|
||||
else
|
||||
m_variable->update(info);
|
||||
set_output_type(0, arg_t, output_shape);
|
||||
return;
|
||||
}
|
||||
throw ngraph_error("v3_ReadValue_validate_and_infer_types is disabled!");
|
||||
}
|
||||
|
||||
shared_ptr<Node> op::ReadValue::clone_with_new_inputs(const OutputVector& new_args) const
|
||||
@ -52,7 +54,7 @@ shared_ptr<Node> op::ReadValue::clone_with_new_inputs(const OutputVector& new_ar
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<ReadValue>(new_args.at(0), m_variable_id);
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v3_ReadValue_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
bool op::v3::ReadValue::visit_attributes(AttributeVisitor& visitor)
|
||||
@ -62,5 +64,5 @@ bool op::v3::ReadValue::visit_attributes(AttributeVisitor& visitor)
|
||||
visitor.on_attribute("variable_id", m_variable_id);
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v3_ReadValue_visit_attributes is disabled!");
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ shared_ptr<Node> op::v4::ReduceL1::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v4::ReduceL1>(new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_ReduceL1_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace reduce_l1
|
||||
@ -90,5 +90,5 @@ bool op::v4::ReduceL1::evaluate(const HostTensorVector& outputs,
|
||||
return reduce_l1::evaluate_sum(
|
||||
inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_ReduceL1_evaluate is disabled!");
|
||||
}
|
||||
|
@ -46,7 +46,7 @@ shared_ptr<Node> op::v4::ReduceL2::clone_with_new_inputs(const OutputVector& new
|
||||
check_new_args_count(this, new_args);
|
||||
return make_shared<op::v4::ReduceL2>(new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v4_ReduceL2_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace reduce_l2
|
||||
@ -88,5 +88,5 @@ bool op::v4::ReduceL2::evaluate(const HostTensorVector& outputs,
|
||||
return reduce_l2::evaluate_reduce_l2(
|
||||
inputs[0], outputs[0], get_reduction_axes(), get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v4_ReduceL2_evaluate is disabled!");
|
||||
}
|
||||
|
@ -42,7 +42,7 @@ shared_ptr<Node> op::v1::ReduceLogicalAnd::clone_with_new_inputs(const OutputVec
|
||||
return make_shared<op::v1::ReduceLogicalAnd>(
|
||||
new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ReduceLogicalAnd_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -87,5 +87,5 @@ bool op::v1::ReduceLogicalAnd::evaluate(const HostTensorVector& outputs,
|
||||
const auto& out = outputs[0];
|
||||
return evaluate_reduce_logical_and(data, axes, out, get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_ReduceLogicalAnd_evaluate is disabled!");
|
||||
}
|
||||
|
@ -42,7 +42,7 @@ shared_ptr<Node> op::v1::ReduceLogicalOr::clone_with_new_inputs(const OutputVect
|
||||
return make_shared<op::v1::ReduceLogicalOr>(
|
||||
new_args.at(0), new_args.at(1), get_keep_dims());
|
||||
}
|
||||
return nullptr;
|
||||
throw ngraph_error("v1_ReduceLogicalOr_clone_with_new_inputs is disabled!");
|
||||
}
|
||||
|
||||
namespace
|
||||
@ -87,5 +87,5 @@ bool op::v1::ReduceLogicalOr::evaluate(const HostTensorVector& outputs,
|
||||
const auto& out = outputs[0];
|
||||
return evaluate_reduce_logical_or(data, axes, out, get_keep_dims());
|
||||
}
|
||||
return false;
|
||||
throw ngraph_error("v1_ReduceLogicalOr_evaluate is disabled!");
|
||||
}
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user