Commit 589a1c0e authored by Robert Kimball's avatar Robert Kimball Committed by Scott Cyphers

fix centos gpu build (#1950)

parent f9fd9d4c
...@@ -1520,11 +1520,9 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::MaxPool* node) ...@@ -1520,11 +1520,9 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::MaxPool* node)
padded_size * args[0].get_element_type().size()); padded_size * args[0].get_element_type().size());
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter(); auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}}, std::vector<std::string> dtypes = {input_type, output_type};
input_shape, pad_index = cuda_emitter->build_pad(
input_shape_padded, dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
padding_below,
padding_interior);
// asymetric padding has been applied, zero out padding vectors to // asymetric padding has been applied, zero out padding vectors to
// ensure cuDNN does not assume padding during pooling // ensure cuDNN does not assume padding during pooling
...@@ -2573,8 +2571,8 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::ReplaceSlice* node, ...@@ -2573,8 +2571,8 @@ size_t runtime::gpu::CUDAEmitter::build_primitive(const op::ReplaceSlice* node,
Shape input_strides = row_major_strides(input_shape); Shape input_strides = row_major_strides(input_shape);
Shape replace_strides = row_major_strides(replace_shape); Shape replace_strides = row_major_strides(replace_shape);
size_t pad_index = build_pad( std::vector<std::string> dtypes = {input_type, output_type};
{{input_type, output_type}}, replace_shape, input_shape, lower_bounds, slice_strides); size_t pad_index = build_pad(dtypes, replace_shape, input_shape, lower_bounds, slice_strides);
if (in_place_op) if (in_place_op)
{ {
......
...@@ -529,12 +529,10 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::Convolution* node) ...@@ -529,12 +529,10 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::Convolution* node)
idx_workspace = allocator.reserve_workspace(temp_size, true); idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter(); auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{args[0].get_element_type().c_type_string(), std::vector<std::string> dtypes = {args[0].get_element_type().c_type_string(),
out[0].get_element_type().c_type_string()}}, out[0].get_element_type().c_type_string()};
input_shape, pad_index = cuda_emitter->build_pad(
input_shape_padded, dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
padding_below,
padding_interior);
// asymetric padding has been applied, zero out padding vectors to // asymetric padding has been applied, zero out padding vectors to
// ensure cudnn does not assume padding // ensure cudnn does not assume padding
...@@ -653,11 +651,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop ...@@ -653,11 +651,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop
idx_workspace = allocator.reserve_workspace(temp_size, true); idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter(); auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}}, std::vector<std::string> dtypes = {input_type, output_type};
output_shape, pad_index = cuda_emitter->build_pad(
output_shape_padded, dtypes, output_shape, output_shape_padded, padding_below, padding_interior);
padding_below,
padding_interior);
slice_index = cuda_emitter->build_slice({{input_type, output_type}}, slice_index = cuda_emitter->build_slice({{input_type, output_type}},
output_shape_padded, output_shape_padded,
...@@ -772,11 +768,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop ...@@ -772,11 +768,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::ConvolutionBackprop
idx_workspace = allocator.reserve_workspace(temp_size, true); idx_workspace = allocator.reserve_workspace(temp_size, true);
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter(); auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}}, std::vector<std::string> dtypes = {input_type, output_type};
input_shape_0, pad_index = cuda_emitter->build_pad(
input_shape_padded, dtypes, input_shape_0, input_shape_padded, padding_below, padding_interior);
padding_below,
padding_interior);
// asymetric padding has been applied, zero out padding vectors to // asymetric padding has been applied, zero out padding vectors to
// ensure cudnn does not assume padding // ensure cudnn does not assume padding
...@@ -865,11 +859,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::MaxPool* node) ...@@ -865,11 +859,9 @@ size_t runtime::gpu::CUDNNEmitter::build_primitive(const op::MaxPool* node)
padded_size * args[0].get_element_type().size()); padded_size * args[0].get_element_type().size());
auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter(); auto& cuda_emitter = m_primitive_emitter->get_cuda_emitter();
pad_index = cuda_emitter->build_pad({{input_type, output_type}}, std::vector<std::string> dtypes = {input_type, output_type};
input_shape, pad_index = cuda_emitter->build_pad(
input_shape_padded, dtypes, input_shape, input_shape_padded, padding_below, padding_interior);
padding_below,
padding_interior);
// asymetric padding has been applied, zero out padding vectors to // asymetric padding has been applied, zero out padding vectors to
// ensure cuDNN does not assume padding during pooling // ensure cuDNN does not assume padding during pooling
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment