1. 30 Jun, 2018 1 commit
    • Nick Korovaiko's avatar
      LoopKernel Collector (#1128) · 784735d6
      Nick Korovaiko authored
      * collector
      
      * keeping track of inputs; simplifying a merging stratey; adding LKGraph
      
      * LoopKernel Collector
      
      * address feedback
      
      * address feedback 2
      
      * address feedback 3
      784735d6
  2. 29 Jun, 2018 4 commits
  3. 28 Jun, 2018 8 commits
    • Nishant Patel's avatar
      Reshape bias to 1D for cpufusion of conv+bias bprop (#1151) · 1574031c
      Nishant Patel authored
      * Reshape bias to 1D for conv + bias bprop fusion
      
      * Reshape goe2 back to 2D before replacing
      1574031c
    • Fenglei's avatar
      check cudnn version (#1175) · cf3e2992
      Fenglei authored
      cf3e2992
    • Nishant Patel's avatar
      Support dimshuffle/transpose with MKLDNN (#1129) · 846f6bfe
      Nishant Patel authored
      * Reshape 4d
      
      * Support dimshuffles/transpose with MKLDNN
      
      * Addressing PR Feedback
      
      * Use Eigen for 3D dimshuffles
      846f6bfe
    • Pruthvi's avatar
      - Added workspace for rnn fprop kernel (#1153) · d861ba32
      Pruthvi authored
      - fixes segfault issue for GNMT model execution through ngraph-mxnet
      d861ba32
    • Matthew Brookhart's avatar
      working generate_adjoints (#1173) · aa36865c
      Matthew Brookhart authored
      aa36865c
    • Fenglei's avatar
      enable cudnn datatype support (#1122) · eef2b19d
      Fenglei authored
      * enable multi datatpye support for Cudnn. refactor binary ops using cudnn
      
      * fix bugs
      
      * add tests to skip list that CUDNN does not support
      
      * not int support on cudnn for backward pooling
      
      * no GPU.dot_4d_5d_multi_axis_big_fp64_VERY_SLOW test anymore
      
      * clang format
      
      * throw if datatype is int8 or int32 for backward pooling
      
      * comments
      
      * fix list in unit_test.manifest
      
      * add type support for alpha, beta
      
      * fix bugs
      
      * datatype support for alpha, beta
      
      * missing ()
      
      * clang format
      
      * batchnorm backward bug fix
      
      * remove debug info
      
      * change member function name to snake case. remove comments
      
      * use nullptr instead of NULL
      
      * code style, use cuDNN everywhere in comments
      
      * add cudnn host parameters memory manager.
      
      * change name to allocate_by_datatype
      
      * compiled
      
      * debug
      
      * fix bug: using list instead of vector, vector address will change each time it resize
      
      * add CUDNN_DATA_UINT8 and CUDNN_DATA_UINT8x4
      eef2b19d
    • Adam Straw's avatar
      constant broadcast folding (#1139) · 35b04e6a
      Adam Straw authored
      * constant broadcast folding
      
      * code review feedback
      35b04e6a
    • Chris Sullivan's avatar
      Add extra hash parameters to broadcast and max pool (#1163) · 13f00048
      Chris Sullivan authored
      * Move maxpool and avgpool into CudaKernelBuilder and add cache parameters to kernel name for broadcast which are required for correct lookup.
      
      * Styling.
      
      * Add space before avg_pool.
      13f00048
  4. 27 Jun, 2018 5 commits
  5. 26 Jun, 2018 10 commits
  6. 25 Jun, 2018 4 commits
  7. 23 Jun, 2018 1 commit
  8. 22 Jun, 2018 2 commits
  9. 21 Jun, 2018 2 commits
  10. 20 Jun, 2018 3 commits