1. 15 Oct, 2018 6 commits
  2. 14 Oct, 2018 1 commit
  3. 13 Oct, 2018 6 commits
  4. 12 Oct, 2018 10 commits
  5. 11 Oct, 2018 3 commits
  6. 10 Oct, 2018 3 commits
    • Fenglei's avatar
      add back missing part (#1785) · a41c1baa
      Fenglei authored
      a41c1baa
    • Fenglei's avatar
      nvgpu one hot update (#1773) · 6cd35432
      Fenglei authored
      * update onehot
      
      * clang
      
      * fix bugs
      
      * format
      
      * add output_datatype_size to hash
      
      * typo
      
      * hash
      6cd35432
    • Nick Korovaiko's avatar
      Reshape Sinking (#1701) · f642bc4c
      Nick Korovaiko authored
      * reshape sinking working on mnist_conv
      
      * forgot to add reshape_sinking files
      
      * refactoring of binary case
      
      * Quantize/Dequantize case, fix add case, add assert
      
      * address bob and scott's feedback
      
      * debug
      
      * fix a bug where reshapes are removed too early
      f642bc4c
  7. 09 Oct, 2018 4 commits
  8. 08 Oct, 2018 4 commits
    • Robert Kimball's avatar
      optimize operator== (#1765) · c5f0bd9d
      Robert Kimball authored
      c5f0bd9d
    • Chris Sullivan's avatar
      Update pad on nvpgu (#1759) · 40ff77bd
      Chris Sullivan authored
      * Add pad with fill operator using the outward-in index pattern.
      
      * Remove static pad and rename build_pad_dynamic -> build_pad. Update maxpool 1d padding.
      
      * Formatting.
      
      * Split build_pad_dynamic into build_pad and build_pad_fill.
      
      * Add test coverage for fixed bug in op::Pad for gpu.
      40ff77bd
    • Jayaram Bobba's avatar
      IAT: Skip reshapes that are removing or adding size-1 dimensions (#1684) · 519b18ac
      Jayaram Bobba authored
      * Reshape optimizations for when unit-sized dimensions are added/removed from tensors
      
      * Added unit tests for eliminating squeeze and expand_dims operations
      
      * Bug fix to expand dims layout
      
      * Style fix
      519b18ac
    • Jayaram Bobba's avatar
      IAT: More convolution folding optimizations (#1712) · 00b4453d
      Jayaram Bobba authored
      * Check output shape when setting memory layout for slice op.
      
      * Miscellaneous fusion and other optimizations for inception-resnetv2
      - ConvBias Batchnorm folding
      - ConvBias Affine folding
      - Check if MKLDNN can slice a given layout and select layouts
        appropriately
      
      * Fixed unit test and bug in conv bias pattern
      
      * Addressed PR feedback
      
      * Addressed PR feedback
      00b4453d
  9. 06 Oct, 2018 2 commits
  10. 05 Oct, 2018 1 commit
    • gcwenger's avatar
      Support LRN for NVGPU Backend (#1740) · fe06f325
      gcwenger authored
      * LRN WIP
      
      * Explicit lambda captures.
      
      * Switched to Ayan's new caching routine.
      
      * Remove commented out lrn from manifest.
      
      * Fixed clang 3.9 error.
      
      * Corrected lrn hash. Only call cudnnSetLRNDescriptor once.
      
      * Simplified lrn hash. Removed redundant parameters. No longer passing CUDNN_LRN_CROSS_CHANNEL_DIM1 as parameter because it's the only choice for cudnnLRNCrossChannelForward.
      fe06f325