API Conformance report: C++ & Merge XML Part (#11245)
* Separate Summary, OpSummary, ApiSummary * final separatin * ChangeNamespaces * git status * filename * link error * Fix linking * Fix compilation * Report - ieplugin * Next step * Fix build * Changing inheritance + fix build * Fix run * Summary * Fix comments * Fix the run * fix build * #Extend report * Fix build * fix template * api_report flag * rebase to master branch * fix * fix build * myriad * fix problem with crash * Fix some mistakes * python merge * fix tests * tmp * Update Merge_xml script * Fix op * fix build * Fix bug with --report_unique_name * build * remove extra * gg * gpu build * c * Fix issue with win * infer_req * compiled + exec net * ov_plugin * ie_plugin * Fix comments * ff * fix last comment * fix build * fix template func * Apply commnets * Apply commnets * fix ci * build * build * build * inl * Remove extra * fix merge_xml * fix build * remarks * skip one test Co-authored-by: Alexander Zhogov <alexander.zhogov@intel.com>
This commit is contained in:
@@ -17,7 +17,7 @@ TEST_P(OVInferRequestBatchedTests, SetInputTensors_Batch_Non_0) {
|
||||
auto batch_shape = Shape{batch, 3, 3, 3};
|
||||
auto model = OVInferRequestBatchedTests::create_n_inputs(1, element::f32, batch_shape, "CNHW");
|
||||
const std::string tensor_name = "tensor_input0";
|
||||
auto execNet = ie->compile_model(model, targetDevice);
|
||||
auto execNet = ie->compile_model(model, target_device);
|
||||
ov::InferRequest req;
|
||||
req = execNet.create_infer_request();
|
||||
std::vector<ov::Tensor> tensors(batch, ov::Tensor(element::f32, one_shape));
|
||||
@@ -31,7 +31,7 @@ TEST_P(OVInferRequestBatchedTests, SetInputTensors_remote_tensor_default) {
|
||||
auto batch_shape = Shape{batch, 4, 4, 4};
|
||||
auto model = OVInferRequestBatchedTests::create_n_inputs(1, element::f32, batch_shape, "NCHW");
|
||||
const std::string tensor_name = "tensor_input0";
|
||||
auto execNet = ie->compile_model(model, targetDevice);
|
||||
auto execNet = ie->compile_model(model, target_device);
|
||||
ov::InferRequest req;
|
||||
req = execNet.create_infer_request();
|
||||
std::vector<ov::Tensor> tensors(batch - 1, ov::Tensor(element::f32, one_shape));
|
||||
@@ -49,7 +49,7 @@ TEST_P(OVInferRequestBatchedTests, SetInputTensors_Strides) {
|
||||
auto model = OVInferRequestBatchedTests::create_n_inputs(2, element::f32, batch_shape, "NCHW");
|
||||
std::vector<float> buffer1(one_shape_size_stride, 10);
|
||||
std::vector<float> buffer2(one_shape_size_stride, 20);
|
||||
auto execNet = ie->compile_model(model, targetDevice);
|
||||
auto execNet = ie->compile_model(model, target_device);
|
||||
// Create InferRequest
|
||||
ov::InferRequest req;
|
||||
req = execNet.create_infer_request();
|
||||
|
||||
@@ -69,7 +69,13 @@ INSTANTIATE_TEST_SUITE_P(
|
||||
// IE Class SetConfig
|
||||
//
|
||||
|
||||
using IEClassSetConfigTestHETERO = BehaviorTestsUtils::IEClassNetworkTest;
|
||||
class IEClassSetConfigTestHETERO : public BehaviorTestsUtils::IEClassNetworkTest,
|
||||
public BehaviorTestsUtils::IEPluginTestBase {
|
||||
void SetUp() override {
|
||||
IEClassNetworkTest::SetUp();
|
||||
IEPluginTestBase::SetUp();
|
||||
}
|
||||
};
|
||||
|
||||
TEST_F(IEClassSetConfigTestHETERO, smoke_SetConfigNoThrow) {
|
||||
{
|
||||
@@ -115,7 +121,13 @@ INSTANTIATE_TEST_SUITE_P(
|
||||
smoke_IEClassGetConfigTest, IEClassGetConfigTest,
|
||||
::testing::Values(CommonTestUtils::DEVICE_TEMPLATE));
|
||||
|
||||
using IEClassGetConfigTestTEMPLATE = BehaviorTestsUtils::IEClassNetworkTest;
|
||||
class IEClassGetConfigTestTEMPLATE : public BehaviorTestsUtils::IEClassNetworkTest,
|
||||
public BehaviorTestsUtils::IEPluginTestBase {
|
||||
void SetUp() override {
|
||||
IEClassNetworkTest::SetUp();
|
||||
IEPluginTestBase::SetUp();
|
||||
}
|
||||
};
|
||||
|
||||
TEST_F(IEClassGetConfigTestTEMPLATE, smoke_GetConfigNoThrow) {
|
||||
InferenceEngine::Core ie = BehaviorTestsUtils::createIECoreWithTemplate();
|
||||
|
||||
Reference in New Issue
Block a user