Commit 589a1c0e authored by Robert Kimball's avatar Robert Kimball Committed by Scott Cyphers

fix centos gpu build (#1950)

parent f9fd9d4c
......@@ -1520,11 +1520,9 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::MaxPool* node)
padded_size * args[0].get_element_type().size());
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}},
input_shape,
input_shape_padded,
padding_below,
padding_interior);
std::vector<std::string> dtypes = {input_type, output_type};
pad_index = cuda_emitter->build_pad(
dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
// asymetric padding has been applied, zero out padding vectors to
// ensure cuDNN does not assume padding during pooling
......@@ -2573,8 +2571,8 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::ReplaceSlice* node,
Shape input_strides = row_major_strides(input_shape);
Shape replace_strides = row_major_strides(replace_shape);
size_t pad_index = build_pad(
{{input_type, output_type}}, replace_shape, input_shape, lower_bounds, slice_strides);
std::vector<std::string> dtypes = {input_type, output_type};
size_t pad_index = build_pad(dtypes, replace_shape, input_shape, lower_bounds, slice_strides);
if (in_place_op)
{
......
......@@ -529,12 +529,10 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::Convolution* node)
idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{args[0].get_element_type().c_type_string(),
out[0].get_element_type().c_type_string()}},
input_shape,
input_shape_padded,
padding_below,
padding_interior);
std::vector<std::string> dtypes = {args[0].get_element_type().c_type_string(),
out[0].get_element_type().c_type_string()};
pad_index = cuda_emitter->build_pad(
dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
// asymetric padding has been applied, zero out padding vectors to
// ensure cudnn does not assume padding
......@@ -653,11 +651,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop
idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}},
output_shape,
output_shape_padded,
padding_below,
padding_interior);
std::vector<std::string> dtypes = {input_type, output_type};
pad_index = cuda_emitter->build_pad(
dtypes, output_shape, output_shape_padded, padding_below, padding_interior);
slice_index = cuda_emitter->build_slice({{input_type, output_type}},
output_shape_padded,
......@@ -772,11 +768,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop
idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}},
input_shape_0,
input_shape_padded,
padding_below,
padding_interior);
std::vector<std::string> dtypes = {input_type, output_type};
pad_index = cuda_emitter->build_pad(
dtypes, input_shape_0, input_shape_padded, padding_below, padding_interior);
// asymetric padding has been applied, zero out padding vectors to
// ensure cudnn does not assume padding
......@@ -865,11 +859,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::MaxPool* node)
padded_size * args[0].get_element_type().size());
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}},
input_shape,
input_shape_padded,
padding_below,
padding_interior);
std::vector<std::string> dtypes = {input_type, output_type};
pad_index = cuda_emitter->build_pad(
dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
// asymetric padding has been applied, zero out padding vectors to
// ensure cuDNN does not assume padding during pooling
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment