1. 09 Mar, 2018 1 commit
    • Pruthvi's avatar
      Pruthvi/sigmoid (#614) · 5885c09a
      Pruthvi authored
      * - Added sigmoid fusion pass
      - added mkldnn emitter code for sigmoid
      
      * - corrected sigmoid expected values
      - add layout assignment for sigmoid op
      
      * - added assert's in cpu fusion for sigmoid
      - style fix
      
      * remove debug prints
      
      * NGMX-371 #comment addressed PR comments - Added sigmoid unit test case with 3D input ii) support in cpu_emmiter for sigmoid to handle all input shapes
      
      * NGMX-371 #comment use shape_size() to calculate the 1d input size
      5885c09a
  2. 08 Mar, 2018 11 commits
    • Jayaram Bobba's avatar
      Merge pull request #613 from NervanaSystems/jbobba/batchnorm-layouts · e46184a1
      Jayaram Bobba authored
      Jbobba/batchnorm layouts
      e46184a1
    • Jayaram Bobba's avatar
    • Jayaram Bobba's avatar
      d6000754
    • Jayaram Bobba's avatar
      9af9031e
    • Nick Korovaiko's avatar
      Optimize Broadcast in MatMulBias (#604) · 9cca4073
      Nick Korovaiko authored
      * remove broadcast from matmulbias
      
      * fix comments
      
      * working gemm-based broadcast
      
      * fix clang warning
      9cca4073
    • Jayaram Bobba's avatar
      Optimize MKLDNN filter conversions · bb06a619
      Jayaram Bobba authored
      bb06a619
    • Jayaram Bobba's avatar
    • Chris Sullivan's avatar
      Abstraction for GPU unary elementwise ops (#587) · 529362b5
      Chris Sullivan authored
      * straightforward gpu.cos implementation following previous patterns prior to refactor
      
      * Generalized unary elementwise gpu op impl.. New unary elementwise ops can
      be added to the type annotations in gpu_cuda_kernel_ops.hpp. Next step
      is to refactor the llvm interface in gpu_emitters.hpp for similar generality.
      
      * Added gpu_emitter.hpp:EmitUnaryElementwise.
      
      Function adds cuda kernel based on ngraph::op::op_type::description.
      This can service all unary elementwise ops run on the gpu.
      
      * The following elementwise unary ops now use the EmitUnaryElementwise emitter:
      * GPU.abs
      * GPU.acos
      * GPU.asin
      * GPU.atan
      * GPU.ceiling
      * GPU.cos
      * GPU.cosh
      * GPU.exp
      * GPU.floor
      * GPU.log
      * GPU.not
      * GPU.sign
      * GPU.sin
      * GPU.sinh
      * GPU.tan
      * GPU.tanh
      Unary elementwise ops Sign and Not need extra consideration.
      
      * tanh test changed to test::all_close for fp comparison (also done for tan in commit 65fa7c6de34c8277fe2a4801644f6bb64574f4ff).
      
      * GPU backend skips added for recent softmax test and updated aliased output test that uses op::Constant.
      
      * code format update
      
      * changed cuda builder interface names to unary/binary/arbitrary, added impl. note to gpu_cuda_kernel_ops, cleaned code format
      
      * updated ngraph-cpp reference
      
      * Fixing incorrect github conflict resolution.
      
      * Added GPU emitter for op::Result.
      For now it simply copies the output tensor.
      
      All but 3 tests now pass. The remaining
      failing tests are:
      * GPU.dot_0_0
      * GPU.dot_matrix_2x0_0x2
      * GPU.dot_2x0_0
      
      * Removed call to handle memory aliasing in gpu_external_function.
      
      * fix gpu emitter bug that will return in the middle of function
      
      * Merge pull request #609 from NervanaSystems/tfl/fix_return_bug
      
      fix gpu emitter bug that will return in the middle of function
      
      * GPU backend skips added for recent softmax test and updated aliased output test that uses op::Constant.
      529362b5
    • Fenglei's avatar
      Merge pull request #599 from NervanaSystems/tfl/gpu_fix_constant_bug · a02aab01
      Fenglei authored
      Fix constant bug on GPU
      a02aab01
    • Robert Kimball's avatar
      61fa9d55
    • Chris Sullivan's avatar
      GPU op::Result implementation (#611) · 905cafd2
      Chris Sullivan authored
      * Added GPU emitter for op::Result.
      For now it simply copies the output tensor.
      
      All but 3 tests now pass. The remaining
      failing tests are:
      * GPU.dot_0_0
      * GPU.dot_matrix_2x0_0x2
      * GPU.dot_2x0_0
      
      * Removed call to handle memory aliasing in gpu_external_function.
      
      * fix gpu emitter bug that will return in the middle of function
      
      * Merge pull request #609 from NervanaSystems/tfl/fix_return_bug
      
      fix gpu emitter bug that will return in the middle of function
      
      * GPU backend skips added for recent softmax test and updated aliased output test that uses op::Constant.
      905cafd2
  3. 07 Mar, 2018 5 commits
  4. 06 Mar, 2018 13 commits
  5. 05 Mar, 2018 9 commits
  6. 04 Mar, 2018 1 commit