From 5a111bfb27a0a9f0555d325cea5b26b7e151f1fd Mon Sep 17 00:00:00 2001 From: Ilya Churaev Date: Wed, 14 Apr 2021 10:35:20 +0300 Subject: [PATCH] Enable nonnull-compare flag for nGraph (#5213) * Enable nonnull-compare flag for nGraph * Fixed checks * Fixed MSVC * Fixed build --- ngraph/CMakeLists.txt | 7 ++++--- ngraph/core/src/op/clamp.cpp | 3 +-- ngraph/core/src/op/concat.cpp | 8 ++++---- ngraph/core/src/op/convert.cpp | 4 ++-- ngraph/core/src/op/exp.cpp | 3 +-- ngraph/core/src/op/gather.cpp | 8 ++++---- ngraph/core/src/op/gelu.cpp | 3 +-- ngraph/core/src/op/hsigmoid.cpp | 3 +-- ngraph/core/src/op/hswish.cpp | 3 +-- ngraph/core/src/op/min.cpp | 6 +++--- ngraph/core/src/op/mish.cpp | 3 +-- ngraph/core/src/op/prelu.cpp | 3 +-- ngraph/core/src/op/reduce_prod.cpp | 4 ++-- ngraph/core/src/op/relu.cpp | 3 +-- ngraph/core/src/op/reshape.cpp | 4 ++-- ngraph/core/src/op/scatter_nd_update.cpp | 6 +++--- ngraph/core/src/op/select.cpp | 4 ++-- ngraph/core/src/op/shape_of.cpp | 10 +++++----- ngraph/core/src/op/sigmoid.cpp | 3 +-- ngraph/core/src/op/softmax.cpp | 3 +-- ngraph/core/src/op/softplus.cpp | 3 +-- ngraph/core/src/op/squeeze.cpp | 8 ++++---- ngraph/core/src/op/strided_slice.cpp | 6 +++--- ngraph/core/src/op/swish.cpp | 3 +-- ngraph/core/src/op/unsqueeze.cpp | 4 ++-- .../core/src/op/util/binary_elementwise_arithmetic.cpp | 6 +++--- ngraph/core/src/op/util/broadcast_base.cpp | 5 ++--- .../onnx_import/src/op/global_average_pool.cpp | 2 +- ngraph/frontend/onnx_import/src/op/global_max_pool.cpp | 2 +- 29 files changed, 59 insertions(+), 71 deletions(-) diff --git a/ngraph/CMakeLists.txt b/ngraph/CMakeLists.txt index fb7b461b205..7077d05cc28 100644 --- a/ngraph/CMakeLists.txt +++ b/ngraph/CMakeLists.txt @@ -83,7 +83,6 @@ option(NGRAPH_LIB_VERSIONING_ENABLE "Enable shared library versioning" OFF) option(NGRAPH_PYTHON_BUILD_ENABLE "Enable build nGraph python package wheel" OFF) option(NGRAPH_DYNAMIC_COMPONENTS_ENABLE "Enable dynamic loading of components" ON) option(NGRAPH_EXPORT_TARGETS_ENABLE "Enable exporting nGraph cmake export targets" ON) -option(NGRAPH_WARNINGS_AS_ERRORS "Make all nGraph compile-time warnings into errors" OFF) option(NGRAPH_ADDRESS_SANITIZER_ENABLE "Compiles and links with Address Sanitizer" OFF) option(NGRAPH_THREAD_SANITIZER_ENABLE "Compiles and links with Thread Sanitizer" OFF) option(NGRAPH_UB_SANITIZER_ENABLE "Compiles and links with Undefined Behavior Sanitizer" OFF) @@ -109,7 +108,6 @@ message(STATUS "NGRAPH_THREAD_SANITIZER_ENABLE: ${NGRAPH_THREAD_SANITIZER_ message(STATUS "NGRAPH_UB_SANITIZER_ENABLE: ${NGRAPH_UB_SANITIZER_ENABLE}") message(STATUS "NGRAPH_USE_PROTOBUF_LITE: ${NGRAPH_USE_PROTOBUF_LITE}") message(STATUS "NGRAPH_UNIT_TEST_ENABLE: ${NGRAPH_UNIT_TEST_ENABLE}") -message(STATUS "NGRAPH_WARNINGS_AS_ERRORS: ${NGRAPH_WARNINGS_AS_ERRORS}") # Setup CMAKE_ARGS to be forwarded to External Projects set(NGRAPH_FORWARD_CMAKE_ARGS @@ -199,7 +197,10 @@ if (WIN32) string(REPLACE "/W3" "/W0" CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS}") endif() -if (NOT WIN32 AND NGRAPH_WARNINGS_AS_ERRORS) +if (NOT CMAKE_CXX_COMPILER_ID STREQUAL "MSVC") + if (CMAKE_CXX_COMPILER_ID STREQUAL "GNU" AND (NOT CMAKE_CXX_COMPILER_VERSION VERSION_LESS 6.0)) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -Wnonnull-compare") + endif() set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -Werror") endif() diff --git a/ngraph/core/src/op/clamp.cpp b/ngraph/core/src/op/clamp.cpp index d2f9aa920e5..809694242fe 100644 --- a/ngraph/core/src/op/clamp.cpp +++ b/ngraph/core/src/op/clamp.cpp @@ -103,8 +103,7 @@ namespace clamp bool op::v0::Clamp::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Clamp_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return clamp::evaluate_clamp(inputs[0], outputs[0], get_min(), get_max()); } diff --git a/ngraph/core/src/op/concat.cpp b/ngraph/core/src/op/concat.cpp index e7d06e830c4..77437fffe50 100644 --- a/ngraph/core/src/op/concat.cpp +++ b/ngraph/core/src/op/concat.cpp @@ -141,9 +141,9 @@ namespace bool op::Concat::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Concat_evaluate); - NGRAPH_CHECK(this, !inputs.empty()); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, inputs.size())); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(!inputs.empty()); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, inputs.size())); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); auto concat_axis = get_axis() < 0 ? get_axis() + inputs[0]->get_shape().size() : get_axis(); return evaluate_concat(inputs, outputs[0], concat_axis); } @@ -155,4 +155,4 @@ bool op::Concat::evaluate_lower(const HostTensorVector& output_values) const bool op::Concat::evaluate_upper(const HostTensorVector& output_values) const { return default_upper_bound_evaluator(this, output_values); -} \ No newline at end of file +} diff --git a/ngraph/core/src/op/convert.cpp b/ngraph/core/src/op/convert.cpp index 3a0a6307b28..80fe8de11c8 100644 --- a/ngraph/core/src/op/convert.cpp +++ b/ngraph/core/src/op/convert.cpp @@ -150,8 +150,8 @@ bool op::v0::Convert::evaluate(const HostTensorVector& output_values, const HostTensorVector& input_values) const { NGRAPH_OP_SCOPE(v0_Convert_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(input_values, 1)); - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(input_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); return convert::evaluate_convert(input_values[0], output_values[0]); } diff --git a/ngraph/core/src/op/exp.cpp b/ngraph/core/src/op/exp.cpp index 81378380750..27d9a2beba2 100644 --- a/ngraph/core/src/op/exp.cpp +++ b/ngraph/core/src/op/exp.cpp @@ -68,7 +68,6 @@ namespace expop bool op::Exp::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Exp_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return expop::evaluate_exp(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/gather.cpp b/ngraph/core/src/op/gather.cpp index df1be923b2e..6475cb8ae55 100644 --- a/ngraph/core/src/op/gather.cpp +++ b/ngraph/core/src/op/gather.cpp @@ -488,8 +488,8 @@ bool op::v1::Gather::evaluate_gather(const HostTensorVector& outputs, bool op::v1::Gather::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v1_Gather_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 3)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 3)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return evaluate_gather(outputs, inputs); } @@ -548,8 +548,8 @@ bool op::v7::Gather::evaluate_gather(const HostTensorVector& outputs, bool op::v7::Gather::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v7_Gather_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 3)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 3)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return evaluate_gather(outputs, inputs); } diff --git a/ngraph/core/src/op/gelu.cpp b/ngraph/core/src/op/gelu.cpp index ffd342ee491..9bd27023875 100644 --- a/ngraph/core/src/op/gelu.cpp +++ b/ngraph/core/src/op/gelu.cpp @@ -187,7 +187,6 @@ namespace gelu bool op::v7::Gelu::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v7_Gelu_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return gelu::evaluate_gelu(inputs[0], outputs[0], m_approximation_mode); } diff --git a/ngraph/core/src/op/hsigmoid.cpp b/ngraph/core/src/op/hsigmoid.cpp index 7965aae84bf..6ce9477dea7 100644 --- a/ngraph/core/src/op/hsigmoid.cpp +++ b/ngraph/core/src/op/hsigmoid.cpp @@ -66,7 +66,6 @@ bool op::v5::HSigmoid::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v5_HSigmoid_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return evaluate_hsigmoid(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/hswish.cpp b/ngraph/core/src/op/hswish.cpp index 69d1967a397..d4401b21c1d 100644 --- a/ngraph/core/src/op/hswish.cpp +++ b/ngraph/core/src/op/hswish.cpp @@ -64,7 +64,6 @@ namespace hswish bool op::v4::HSwish::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v4_HSwish_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return hswish::evaluate_hswish(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/min.cpp b/ngraph/core/src/op/min.cpp index 300bd8add3c..39439ca762c 100644 --- a/ngraph/core/src/op/min.cpp +++ b/ngraph/core/src/op/min.cpp @@ -67,8 +67,8 @@ bool op::v1::ReduceMin::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v1_ReduceMin_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 2)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 2)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return minop::evaluate_min(inputs[0], outputs[0], get_reduction_axes(), get_keep_dims()); } @@ -84,4 +84,4 @@ bool op::v1::ReduceMin::evaluate_upper(const HostTensorVector& output_values) co if (!input_value(1).get_tensor().has_and_set_bound()) return false; return default_upper_bound_evaluator(this, output_values); -} \ No newline at end of file +} diff --git a/ngraph/core/src/op/mish.cpp b/ngraph/core/src/op/mish.cpp index d4eb1aae6b2..ccd059b0c73 100644 --- a/ngraph/core/src/op/mish.cpp +++ b/ngraph/core/src/op/mish.cpp @@ -71,7 +71,6 @@ namespace mish bool op::v4::Mish::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v4_Mish_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return mish::evaluate_mish(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/prelu.cpp b/ngraph/core/src/op/prelu.cpp index 8e4abd68c27..a7e3ac0aee0 100644 --- a/ngraph/core/src/op/prelu.cpp +++ b/ngraph/core/src/op/prelu.cpp @@ -78,7 +78,6 @@ namespace prelu bool op::PRelu::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_PRelu_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 2)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 2)); return prelu::evaluate_prelu(inputs[0], inputs[1], outputs[0]); } diff --git a/ngraph/core/src/op/reduce_prod.cpp b/ngraph/core/src/op/reduce_prod.cpp index d24ac763f65..c0e8836ebf2 100644 --- a/ngraph/core/src/op/reduce_prod.cpp +++ b/ngraph/core/src/op/reduce_prod.cpp @@ -73,8 +73,8 @@ bool op::v1::ReduceProd::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v1_ReduceProd_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 2)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 2)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return reduce_prod::evaluate_product( inputs[0], outputs[0], get_reduction_axes(), get_keep_dims()); } diff --git a/ngraph/core/src/op/relu.cpp b/ngraph/core/src/op/relu.cpp index 29eef5bec67..1b015b9e70f 100644 --- a/ngraph/core/src/op/relu.cpp +++ b/ngraph/core/src/op/relu.cpp @@ -63,8 +63,7 @@ namespace relu bool op::Relu::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Relu_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return relu::evaluate_relu(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/reshape.cpp b/ngraph/core/src/op/reshape.cpp index af62a16fb6a..b74fbd1a25a 100644 --- a/ngraph/core/src/op/reshape.cpp +++ b/ngraph/core/src/op/reshape.cpp @@ -323,8 +323,8 @@ bool op::v1::Reshape::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v1_Reshape_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 2)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 2)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return evaluate_reshape(outputs, inputs); } diff --git a/ngraph/core/src/op/scatter_nd_update.cpp b/ngraph/core/src/op/scatter_nd_update.cpp index 2c4716c9c01..a887daf2cb1 100644 --- a/ngraph/core/src/op/scatter_nd_update.cpp +++ b/ngraph/core/src/op/scatter_nd_update.cpp @@ -91,9 +91,9 @@ bool op::v3::ScatterNDUpdate::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v3_ScatterNDUpdate_evaluate); - NGRAPH_CHECK(this, !inputs.empty()); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 3)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(!inputs.empty()); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 3)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return scatter::evaluate_scatter(inputs[0], inputs[1], inputs[2], outputs[0]); } diff --git a/ngraph/core/src/op/select.cpp b/ngraph/core/src/op/select.cpp index ba31f8fb866..d5fc2d7ef33 100644 --- a/ngraph/core/src/op/select.cpp +++ b/ngraph/core/src/op/select.cpp @@ -146,8 +146,8 @@ bool op::v1::Select::evaluate(const HostTensorVector& output_values, const HostTensorVector& input_values) const { NGRAPH_OP_SCOPE(v1_Select_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(input_values, 3)); - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(input_values, 3)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); const auto autob = get_auto_broadcast(); return detail::evaluate_select( output_values, input_values, autob, output_values[0]->get_element_type()); diff --git a/ngraph/core/src/op/shape_of.cpp b/ngraph/core/src/op/shape_of.cpp index 67ff07c157e..b17924c8809 100644 --- a/ngraph/core/src/op/shape_of.cpp +++ b/ngraph/core/src/op/shape_of.cpp @@ -216,8 +216,8 @@ bool op::v3::ShapeOf::evaluate(const HostTensorVector& output_values, const HostTensorVector& input_values) const { NGRAPH_OP_SCOPE(v3_ShapeOf_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(input_values, 1)); - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(input_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); return shape_of::evaluate_shape_of(output_values[0], input_values[0]); } @@ -280,8 +280,8 @@ bool op::v0::ShapeOf::evaluate(const HostTensorVector& output_values, const HostTensorVector& input_values) const { NGRAPH_OP_SCOPE(v0_ShapeOf_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(input_values, 1)); - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(input_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); return shape_of::evaluate_shape_of(output_values[0], input_values[0]); } @@ -301,4 +301,4 @@ bool op::v0::ShapeOf::evaluate_lower(const HostTensorVector& output_values) cons bool op::v0::ShapeOf::evaluate_upper(const HostTensorVector& output_values) const { return shape_of::evaluate_bound_shape(this, output_values, true); -} \ No newline at end of file +} diff --git a/ngraph/core/src/op/sigmoid.cpp b/ngraph/core/src/op/sigmoid.cpp index f365c118002..8a0353b24c0 100644 --- a/ngraph/core/src/op/sigmoid.cpp +++ b/ngraph/core/src/op/sigmoid.cpp @@ -64,7 +64,6 @@ namespace sigmoid bool op::Sigmoid::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Sigmoid_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return sigmoid::evaluate_sigmoid(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/softmax.cpp b/ngraph/core/src/op/softmax.cpp index 272fb33f63c..6dfe5d05422 100644 --- a/ngraph/core/src/op/softmax.cpp +++ b/ngraph/core/src/op/softmax.cpp @@ -94,8 +94,7 @@ bool op::v1::Softmax::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v1_Softmax_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); outputs[0]->set_unary(inputs[0]); return evaluate_softmax(inputs[0], outputs[0], AxisSet{m_axis}); } diff --git a/ngraph/core/src/op/softplus.cpp b/ngraph/core/src/op/softplus.cpp index 4d56ed71ab9..aefaca0d07f 100644 --- a/ngraph/core/src/op/softplus.cpp +++ b/ngraph/core/src/op/softplus.cpp @@ -71,7 +71,6 @@ bool op::v4::SoftPlus::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v4_SoftPlus_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1) && validate_host_tensor_vector(inputs, 1)); return softplus::evaluate_softplus(inputs[0], outputs[0]); } diff --git a/ngraph/core/src/op/squeeze.cpp b/ngraph/core/src/op/squeeze.cpp index f21a048b49f..0b2095f257c 100644 --- a/ngraph/core/src/op/squeeze.cpp +++ b/ngraph/core/src/op/squeeze.cpp @@ -202,14 +202,14 @@ bool op::v0::Squeeze::evaluate(const HostTensorVector& outputs, { NGRAPH_OP_SCOPE(v0_Squeeze_evaluate); // TODO: change the behaviour after the support of Squeeze with one input - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, inputs.size())); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, inputs.size())); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return squeeze::evaluate_squeeze(inputs[0], inputs[1], outputs[0]); } bool op::v0::Squeeze::evaluate_lower(const HostTensorVector& output_values) const { - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); if (inputs().size() > 1 && !input_value(1).get_tensor().has_and_set_bound()) return false; return default_lower_bound_evaluator(this, output_values); @@ -217,7 +217,7 @@ bool op::v0::Squeeze::evaluate_lower(const HostTensorVector& output_values) cons bool op::v0::Squeeze::evaluate_upper(const HostTensorVector& output_values) const { - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); if (inputs().size() > 1 && !input_value(1).get_tensor().has_and_set_bound()) return false; return default_upper_bound_evaluator(this, output_values); diff --git a/ngraph/core/src/op/strided_slice.cpp b/ngraph/core/src/op/strided_slice.cpp index e4dab49dd04..1519c7489b0 100644 --- a/ngraph/core/src/op/strided_slice.cpp +++ b/ngraph/core/src/op/strided_slice.cpp @@ -274,8 +274,8 @@ bool op::v1::StridedSlice::evaluate(const HostTensorVector& output_values, { NGRAPH_OP_SCOPE(v1_StridedSlice_evaluate); // FIXME: 4th input is optional, but it is required by the following code - NGRAPH_CHECK(this, validate_host_tensor_vector(input_values, 4)); - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(input_values, 4)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); return strided_slice::evaluate_strided_slice(input_values[0], input_values[1], input_values[2], @@ -304,4 +304,4 @@ bool op::v1::StridedSlice::evaluate_upper(const HostTensorVector& output_values) !input_value(3).get_tensor().has_and_set_bound()) return false; return default_upper_bound_evaluator(this, output_values); -} \ No newline at end of file +} diff --git a/ngraph/core/src/op/swish.cpp b/ngraph/core/src/op/swish.cpp index fa963888b2b..72320269e8c 100644 --- a/ngraph/core/src/op/swish.cpp +++ b/ngraph/core/src/op/swish.cpp @@ -123,8 +123,7 @@ bool op::v4::Swish::evaluate(const HostTensorVector& outputs, const HostTensorVe { NGRAPH_OP_SCOPE(v4_Swish_evaluate); NGRAPH_CHECK( - this, validate_host_tensor_vector(outputs, 1) && - (validate_host_tensor_vector(inputs, 2) || validate_host_tensor_vector(inputs, 1))); + (validate_host_tensor_vector(inputs, 2) || validate_host_tensor_vector(inputs, 1))); return swish::evaluate_swish(inputs, outputs[0]); } diff --git a/ngraph/core/src/op/unsqueeze.cpp b/ngraph/core/src/op/unsqueeze.cpp index b6f849ad276..daac35a7a97 100644 --- a/ngraph/core/src/op/unsqueeze.cpp +++ b/ngraph/core/src/op/unsqueeze.cpp @@ -138,8 +138,8 @@ bool op::v0::Unsqueeze::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(v0_Unsqueeze_evaluate); - NGRAPH_CHECK(this, validate_host_tensor_vector(inputs, 2)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 2)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); return unsqueeze::evaluate_unsqueeze(inputs[0], inputs[1], outputs[0]); } diff --git a/ngraph/core/src/op/util/binary_elementwise_arithmetic.cpp b/ngraph/core/src/op/util/binary_elementwise_arithmetic.cpp index 22edda297b0..5c8205c23da 100644 --- a/ngraph/core/src/op/util/binary_elementwise_arithmetic.cpp +++ b/ngraph/core/src/op/util/binary_elementwise_arithmetic.cpp @@ -58,7 +58,7 @@ bool op::util::BinaryElementwiseArithmetic::visit_attributes(AttributeVisitor& v bool op::util::BinaryElementwiseArithmetic::evaluate_upper( const HostTensorVector& output_values) const { - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); HostTensorVector lower_output_tensors; for (const auto& output : output_values) lower_output_tensors.push_back( @@ -71,7 +71,7 @@ bool op::util::BinaryElementwiseArithmetic::evaluate_upper( bool op::util::BinaryElementwiseArithmetic::evaluate_lower( const HostTensorVector& output_values) const { - NGRAPH_CHECK(this, validate_host_tensor_vector(output_values, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(output_values, 1)); HostTensorVector upper_output_tensors; for (const auto& output : output_values) upper_output_tensors.push_back( @@ -79,4 +79,4 @@ bool op::util::BinaryElementwiseArithmetic::evaluate_lower( if (!interval_bound_evaluator(this, output_values, upper_output_tensors)) return false; return true; -} \ No newline at end of file +} diff --git a/ngraph/core/src/op/util/broadcast_base.cpp b/ngraph/core/src/op/util/broadcast_base.cpp index 38519c9df6e..ed49fc70322 100644 --- a/ngraph/core/src/op/util/broadcast_base.cpp +++ b/ngraph/core/src/op/util/broadcast_base.cpp @@ -513,9 +513,8 @@ bool op::util::BroadcastBase::evaluate(const HostTensorVector& outputs, const HostTensorVector& inputs) const { NGRAPH_OP_SCOPE(util_BroadcastBase_evaluate); - NGRAPH_CHECK(this, - validate_host_tensor_vector(inputs, 2) || validate_host_tensor_vector(inputs, 3)); - NGRAPH_CHECK(this, validate_host_tensor_vector(outputs, 1)); + NGRAPH_CHECK(validate_host_tensor_vector(inputs, 2) || validate_host_tensor_vector(inputs, 3)); + NGRAPH_CHECK(validate_host_tensor_vector(outputs, 1)); Shape target_shape = get_target_shape(inputs[1]); diff --git a/ngraph/frontend/onnx_import/src/op/global_average_pool.cpp b/ngraph/frontend/onnx_import/src/op/global_average_pool.cpp index 1db380722c7..2d99e69da5d 100644 --- a/ngraph/frontend/onnx_import/src/op/global_average_pool.cpp +++ b/ngraph/frontend/onnx_import/src/op/global_average_pool.cpp @@ -41,7 +41,7 @@ namespace ngraph // Input shape: [N, C, H, W, D] // Input spatial dimensions are H, W and D // Expected spatial dims indexes: [2, 3, 4] - uint64_t data_spatial_rank = data_rank_value - 2; + size_t data_spatial_rank = data_rank_value - 2; auto reduce_axes_vector = std::vector(data_spatial_rank); std::iota(reduce_axes_vector.begin(), reduce_axes_vector.end(), 2); auto reduce_axes = default_opset::Constant::create( diff --git a/ngraph/frontend/onnx_import/src/op/global_max_pool.cpp b/ngraph/frontend/onnx_import/src/op/global_max_pool.cpp index a3afb4cf00f..866b7ac0201 100644 --- a/ngraph/frontend/onnx_import/src/op/global_max_pool.cpp +++ b/ngraph/frontend/onnx_import/src/op/global_max_pool.cpp @@ -41,7 +41,7 @@ namespace ngraph // Input shape: [N, C, H, W, D] // Input spatial dimensions are H, W and D // Expected spatial dims indexes: [2, 3, 4] - uint64_t data_spatial_rank = data_rank_value - 2; + size_t data_spatial_rank = data_rank_value - 2; auto reduce_axes_vector = std::vector(data_spatial_rank); std::iota(reduce_axes_vector.begin(), reduce_axes_vector.end(), 2); auto reduce_axes = default_opset::Constant::create(