[GPU] Fix warnings (#16516)

* fix a few warnings

* cast size_t to uint32_t
This commit is contained in:
Andrei Gorbachev 2023-03-24 09:26:24 +00:00 committed by GitHub
parent 86c4489aca
commit cabb917b1f
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 6 additions and 6 deletions

View File

@ -84,11 +84,11 @@ std::vector<layout> broadcast_inst::calc_output_layouts(broadcast_node const& /*
ov::op::v3::shape_infer(&op, input_shapes, output_shapes, const_data); ov::op::v3::shape_infer(&op, input_shapes, output_shapes, const_data);
} else if (impl_param.input_layouts.size() >= 2) { } else if (impl_param.input_layouts.size() >= 2) {
auto input1 = impl_param.get_input_layout(1); auto input1 = impl_param.get_input_layout(1);
int output_rank = input1.get<ShapeType>().size(); auto output_rank = input1.get<ShapeType>().size();
if (input1.is_static()) { if (input1.is_static()) {
output_rank = input1.get_dim(0); // target shape rank is set as second input. output_rank = input1.get_dim(0); // target shape rank is set as second input.
} }
output_shapes[0] = ShapeType::dynamic(std::max(output_rank, 1)); output_shapes[0] = ShapeType::dynamic(std::max(static_cast<int>(output_rank), 1));
} }
format output_format = format::adjust_to_rank(input0_layout.format, output_shapes[0].size()); format output_format = format::adjust_to_rank(input0_layout.format, output_shapes[0].size());

View File

@ -66,7 +66,7 @@ public:
const auto& mode = primitive->mode; const auto& mode = primitive->mode;
const auto& sort_type = primitive->sort; const auto& sort_type = primitive->sort;
const auto& values_first = primitive->values_first; const auto& values_first = primitive->values_first;
const auto& outputs_num = primitive->input_size() == 3 ? 2 : primitive->output_size(); const auto& outputs_num = primitive->input_size() == 3 ? 2 : static_cast<uint32_t>(primitive->output_size());
auto argm_params = get_default_params<kernel_selector::arg_max_min_params>(impl_param); auto argm_params = get_default_params<kernel_selector::arg_max_min_params>(impl_param);
auto argm_optional_params = auto argm_optional_params =

View File

@ -279,7 +279,7 @@ protected:
void update_kernels_list_to_skip() { void update_kernels_list_to_skip() {
for (size_t i = 0; i < _kernel_data.kernels.size(); ++i) { for (size_t i = 0; i < _kernel_data.kernels.size(); ++i) {
auto gws = _kernel_data.kernels[i].params.workGroups.global; auto gws = _kernel_data.kernels[i].params.workGroups.global;
_kernel_data.kernels[i].skip_execution = (std::accumulate(gws.begin(), gws.end(), 1, std::multiplies<size_t>()) == 0); _kernel_data.kernels[i].skip_execution = (std::accumulate(gws.begin(), gws.end(), static_cast<size_t>(1), std::multiplies<size_t>()) == 0);
} }
} }

View File

@ -385,7 +385,7 @@ network::network(cldnn::BinaryInputBuffer& ib, const ExecutionConfig& config, st
_primitives[_primitive_id] = new_primitive_inst; _primitives[_primitive_id] = new_primitive_inst;
} }
int exec_order_size; size_t exec_order_size;
ib >> exec_order_size; ib >> exec_order_size;
_exec_order.clear(); _exec_order.clear();
@ -540,7 +540,7 @@ void network::save(cldnn::BinaryOutputBuffer& ob) {
} }
} }
int exec_order_size = _exec_order.size(); size_t exec_order_size = _exec_order.size();
ob << exec_order_size; ob << exec_order_size;
for (const auto& p_inst : _exec_order) { for (const auto& p_inst : _exec_order) {