86 lines
2.8 KiB
C++
86 lines
2.8 KiB
C++
//*****************************************************************************
|
|
// Copyright 2017-2020 Intel Corporation
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
//*****************************************************************************
|
|
|
|
#include "int_backend_visibility.hpp"
|
|
|
|
#include "backend_manager.hpp"
|
|
#include "int_backend.hpp"
|
|
#include "int_executable.hpp"
|
|
#include "ngraph/except.hpp"
|
|
#include "ngraph/runtime/host_tensor.hpp"
|
|
#include "ngraph/util.hpp"
|
|
|
|
using namespace std;
|
|
using namespace ngraph;
|
|
|
|
extern "C" INTERPRETER_BACKEND_API void ngraph_register_interpreter_backend()
|
|
{
|
|
runtime::BackendManager::register_backend("INTERPRETER", [](const std::string& /* config */) {
|
|
return std::make_shared<runtime::interpreter::INTBackend>();
|
|
});
|
|
}
|
|
|
|
runtime::interpreter::INTBackend::INTBackend()
|
|
{
|
|
}
|
|
|
|
runtime::interpreter::INTBackend::INTBackend(const vector<string>& unsupported_op_name_list)
|
|
: m_unsupported_op_name_list{unsupported_op_name_list.begin(), unsupported_op_name_list.end()}
|
|
{
|
|
}
|
|
|
|
shared_ptr<runtime::Tensor> runtime::interpreter::INTBackend::create_tensor()
|
|
{
|
|
return make_shared<runtime::HostTensor>();
|
|
}
|
|
|
|
shared_ptr<runtime::Tensor>
|
|
runtime::interpreter::INTBackend::create_tensor(const element::Type& type, const Shape& shape)
|
|
{
|
|
return make_shared<runtime::HostTensor>(type, shape);
|
|
}
|
|
|
|
shared_ptr<runtime::Tensor> runtime::interpreter::INTBackend::create_tensor(
|
|
const element::Type& type, const Shape& shape, void* memory_pointer)
|
|
{
|
|
return make_shared<runtime::HostTensor>(type, shape, memory_pointer);
|
|
}
|
|
|
|
shared_ptr<runtime::Executable>
|
|
runtime::interpreter::INTBackend::compile(shared_ptr<Function> function,
|
|
bool enable_performance_collection)
|
|
{
|
|
return make_shared<INTExecutable>(function, enable_performance_collection);
|
|
}
|
|
|
|
bool runtime::interpreter::INTBackend::is_supported(const Node& node) const
|
|
{
|
|
return m_unsupported_op_name_list.find(node.description()) == m_unsupported_op_name_list.end();
|
|
}
|
|
|
|
bool runtime::interpreter::INTBackend::set_config(const map<string, string>& config, string& error)
|
|
{
|
|
bool rc = false;
|
|
auto it = config.find("test_echo");
|
|
error = "";
|
|
if (it != config.end())
|
|
{
|
|
error = it->second;
|
|
rc = true;
|
|
}
|
|
return rc;
|
|
}
|