Skip to content

Commit

Permalink
Fix build
Browse files Browse the repository at this point in the history
  • Loading branch information
vurusovs committed Dec 6, 2023
1 parent 368db6d commit d235581
Show file tree
Hide file tree
Showing 11 changed files with 17 additions and 14 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -200,7 +200,7 @@ void FusedConvBackpropData::validate_and_infer_types() {
// If auto_pad has one of following mode we infer paddings. Otherwise in
// EXPLICIT auto_pad mode we use what is provided.
if (auto_pad_ == ov::op::PadType::SAME_UPPER || auto_pad_ == ov::op::PadType::SAME_LOWER) {
ov::opset1::infer_conv_backprop_auto_padding(
ngraph::opset1::infer_conv_backprop_auto_padding(
ov::Shape{std::next(data_shape.begin(), 2), std::end(data_shape)},
ov::Shape{std::next(filters_shape.begin(), 2), std::end(filters_shape)},
output_shape,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@

#include <ie_precision.hpp>
#include <ngraph/node.hpp>
#include "openvino/opsets/opset1.hpp"
#include <vector>

#include "cuda_test_constants.hpp"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@

#include <ie_precision.hpp>
#include <ngraph/node.hpp>
#include "openvino/opsets/opset1.hpp"
#include <vector>

#include "cuda_test_constants.hpp"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,11 +14,11 @@
#include <memory>
#include <ngraph/function.hpp>
#include <ngraph/node.hpp>
#include <ngraph/opsets/opset1.hpp>
#include <ngraph/shape.hpp>
#include <ngraph/type/element_type.hpp>
#include <ov_models/builders.hpp>
#include <ov_models/utils/ov_helpers.hpp>
#include "openvino/opsets/opset1.hpp"
#include <openvino/op/util/attr_types.hpp>
#include <shared_test_classes/base/layer_test_utils.hpp>
#include <shared_test_classes/single_layer/activation.hpp>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -98,7 +98,7 @@ class FullyConnectedLayerTest : public testing::WithParamInterface<FullyConnecte
auto MatMul = std::make_shared<ov::op::v0::MatMul>(
params[0], secondaryInput, shapeRelatedParams.input1.second, shapeRelatedParams.input2.second);
auto Add = std::make_shared<ov::op::v1::Add>(MatMul, thirdInput);
ov::ResultVector results{std::make_shared<ov::opset1::Result>(Add)};
ov::ResultVector results{std::make_shared<ov::op::v0::Result>(Add)};
function = std::make_shared<ngraph::Function>(results, params, "FullyConnected");
}
};
Expand Down Expand Up @@ -224,7 +224,7 @@ class FullyConnectedLayer2MatMulTest : public testing::WithParamInterface<FullyC
shapeRelatedParams.matmul2_input1.second,
shapeRelatedParams.matmul2_input2.second);
auto Add = std::make_shared<ov::op::v1::Add>(matMul0, matMul1);
ov::ResultVector results{std::make_shared<ov::opset1::Result>(Add)};
ov::ResultVector results{std::make_shared<ov::op::v0::Result>(Add)};
function = std::make_shared<ngraph::Function>(results, params, "FullyConnected");
}
};
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ class CUDNNGRUCellTest : public UnsymmetricalComparer<GRUCellTest> {
const auto& ops = function->get_ordered_ops();
int seed = SEED_FIRST;
for (const auto& op : ops) {
if (std::dynamic_pointer_cast<ov::opset1::Constant>(op)) {
if (std::dynamic_pointer_cast<ov::op::v0::Constant>(op)) {
ov::Tensor random_tensor(op->get_element_type(), op->get_shape());
ov::test::utils::fill_tensor_random(random_tensor, up_to - start_from, start_from, 1, seed++);
function->replace_node(op, std::make_shared<ov::op::v0::Constant>(random_tensor));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ class CUDNNGRUSequenceTest : public UnsymmetricalComparer<GRUSequenceTest> {
const auto& ops = function->get_ordered_ops();
int seed = 1;
for (const auto& op : ops) {
if (std::dynamic_pointer_cast<ov::opset1::Constant>(op)) {
if (std::dynamic_pointer_cast<ov::op::v0::Constant>(op)) {
if (op->get_element_type() == ov::element::Type_t::f32) {
ov::Tensor random_tensor(op->get_element_type(), op->get_shape());
ov::test::utils::fill_tensor_random(random_tensor, up_to - start_from, start_from, 1, seed++);
Expand Down Expand Up @@ -56,7 +56,7 @@ class LPCNetCUDNNGRUSequenceTest : public UnsymmetricalComparer<GRUSequenceTest>
const auto& ops = function->get_ordered_ops();
int seed = 1;
for (const auto& op : ops) {
if (std::dynamic_pointer_cast<ov::opset1::Constant>(op)) {
if (std::dynamic_pointer_cast<ov::op::v0::Constant>(op)) {
if (op->get_element_type() == ov::element::Type_t::f32) {
ov::Tensor random_tensor(op->get_element_type(), op->get_shape());
ov::test::utils::fill_tensor_random(random_tensor, up_to - start_from, start_from, 1, seed++);
Expand Down Expand Up @@ -118,8 +118,8 @@ class LPCNetCUDNNGRUSequenceTest : public UnsymmetricalComparer<GRUSequenceTest>
true,
direction,
mode);
ov::ResultVector results{std::make_shared<ov::opset1::Result>(gru_sequence->output(0)),
std::make_shared<ov::opset1::Result>(gru_sequence->output(1))};
ov::ResultVector results{std::make_shared<ov::op::v0::Result>(gru_sequence->output(0)),
std::make_shared<ov::op::v0::Result>(gru_sequence->output(1))};
function = std::make_shared<ngraph::Function>(results, params, "gru_sequence");
bool ti_found = ngraph::helpers::is_tensor_iterator_exist(function);
EXPECT_EQ(ti_found, false);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ class CUDNNLSTMCellTest : public LSTMCellTest {
const auto& ops = function->get_ordered_ops();
int seed = SEED_FIRST;
for (const auto& op : ops) {
if (std::dynamic_pointer_cast<ov::opset1::Constant>(op)) {
if (std::dynamic_pointer_cast<ov::op::v0::Constant>(op)) {
ov::Tensor random_tensor(op->get_element_type(), op->get_shape());
ov::test::utils::fill_tensor_random(random_tensor, up_to - start_from, start_from, 1, seed);
auto constant = std::make_shared<ov::op::v0::Constant>(random_tensor);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ class CUDALSTMSequenceTest : public UnsymmetricalComparer<LSTMSequenceTest> {
int counter = 1;
const auto& ops = function->get_ordered_ops();
for (const auto& op : ops) {
if (std::dynamic_pointer_cast<ov::opset1::Constant>(op)) {
if (std::dynamic_pointer_cast<ov::op::v0::Constant>(op)) {
if (op->get_element_type() == ov::element::Type_t::f32) {
ov::Tensor random_tensor(op->get_element_type(), op->get_shape());
ov::test::utils::fill_tensor_random(random_tensor, up_to - start_from, start_from, 1, counter++);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@
//

#include "single_layer_tests/topk.hpp"
#include "openvino/opsets/opset4.hpp"

#include <vector>

Expand Down
6 changes: 3 additions & 3 deletions modules/nvidia_plugin/tests/unit/cuda_multi_graph_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@ class AddMul {
const auto add1 = ngraph::builder::makeEltwise(params[2], params[3], EltwiseTypes::ADD);

const auto mul = ngraph::builder::makeEltwise(add0, add1, EltwiseTypes::MULTIPLY);
const auto result = std::make_shared<ov::opset1::Result>(mul);
const auto result = std::make_shared<ov::op::v0::Result>(mul);
return std::make_shared<ov::Model>(result, params, "AddMul");
}

Expand Down Expand Up @@ -114,8 +114,8 @@ class AddConcat {

constexpr int64_t axis = CONCAT_AXIS;
const auto concat =
std::make_shared<ov::opset1::Concat>(ov::OutputVector{add0, add1}, axis);
const auto result = std::make_shared<ov::opset1::Result>(concat);
std::make_shared<ov::op::v0::Concat>(ov::OutputVector{add0, add1}, axis);
const auto result = std::make_shared<ov::op::v0::Result>(concat);
return std::make_shared<ov::Model>(result, params, "AddConcat");
}

Expand Down

0 comments on commit d235581

Please sign in to comment.