1. 29 Aug, 2018 2 commits
  2. 27 Aug, 2018 1 commit
  3. 15 Aug, 2018 1 commit
  4. 13 Aug, 2018 1 commit
  5. 10 Aug, 2018 1 commit
  6. 07 Aug, 2018 1 commit
    • Jayaram Bobba's avatar
      Switch to using more expressive layout descriptors instead of numeric layout names (#1278) · 69c51c27
      Jayaram Bobba authored
      * Switch to using mkldnn memory descriptors for layout
      
      * More changes for using mkldnn descriptor instead of format
      
      * Removed mkldnn format from cpu layout descriptor. TODO - shuffle folding
      
      * Rotate mkldnn layouts on transpose
      
      * Modifications to builder reshape to skip rotated layouts
      
      * More fixes to layouts and removes axis order from cpu layout descriptor
      
      * Code cleanup
      
      * Removed shuffle folding pass since the functionality is subsumed by the layout pass
      
      * Canonicalize a few more formats to keep MKLDNN happy.
      
      * Style fixes
      
      * Style fixes
      
      * Style fixes
      
      * Addressed PR feedback and added reshape passthrough for non-transpose cases
      
      * Adjust named formats for weights tensors to keep MKLDNN happy
      
      * Style fixes
      
      * resolved merge issues
      69c51c27
  7. 18 Jul, 2018 1 commit
  8. 17 Jul, 2018 1 commit
    • Jayaram Bobba's avatar
      Added more convolution variants to DEX (#1223) · 9bb0b653
      Jayaram Bobba authored
      * CPU Direct Execution: Implement ConvertLayout and refactor
      
      * CPU Direct Execution: Implement Convolution
      
      * 1) Adds computation reuse to direct execution
      2) Add avg_pool, broadcast and convolution_bias to direct execution
      3) Moved some computation reuse utility functions to graph_utils
      
      * Use lists instead of vectors to avoid reallocation overheads
      
      * - Added convolution variants to direct execution
      - Removed ConvolutionBiasRelu, use ConvolutionBias instead
      - Reduced code duplication by moving functionality to mkldnn_emitter
        from cpu_emitter
      
      * Style fix
      
      * Moved mkldnn build_convolution to a templated method
      
      * Style fix
      
      * refactored mkldnn conv bprop builders
      
      * Style fix
      9bb0b653
  9. 11 Jul, 2018 1 commit
  10. 03 Jul, 2018 1 commit
  11. 02 Jul, 2018 3 commits
    • Sandeep's avatar
      move sigmoid to core fusion (#1132) · d05b5e39
      Sandeep authored
      * declare sigmoid for core fusion
      
      * add simple test for sigmoid
      
      * info fusion status
      
      * cp op as main op
      
      * builds as expected
      
      * move sigmoid fusion code
      
      * add reference kernel
      
      * sigmoid bprop reference kernel and clang-format
      
      * add delta to bprop
      
      * fprop called
      
      * compiles bprop
      
      * move tests
      
      * serializer support
      
      * address comments in code
      
      * add doc
      
      * naming similar to core ops
      
      * fix failing test
      
      * fix failing test
      
      * address clang issue
      
      * more changes
      
      * change test macro
      d05b5e39
    • Pruthvi's avatar
      MKLDNN BoundedRelu implementation for Relu6 (#1179) · eaa6091c
      Pruthvi authored
      * 1. Added MKLDNNN BoundedRelu op support for Relu6
      2. CpuLayout && CPU assignment pass for BoundedRelu Op
      3. Unit test inter v/s CPU for BoundedReluOp
      4. MKLDNN and default emitter code for BoundedReluOp
      
      * Removed Debug prints
      
      * 1. Added support for boundedrelu to work on any constant literal
      2. unit test case for rank2, rank3, rank4 for bounded relu without serialized graph
      
      * Removed is_six() method
      eaa6091c
    • Louis Feng's avatar
      Conv+bias shape check for better error detection (#1176) · e42e5815
      Louis Feng authored
      * Reshape bias to 1D for conv + bias bprop fusion
      
      * Reshape goe2 back to 2D before replacing
      
      * added shape checks to validate conv+bias op.
      
      * removed conv+bias backprop merge for separate PR review.
      
      * fixed conv_bias_bprop test.
      
      * minor changes to error messages.
      e42e5815
  12. 30 Jun, 2018 1 commit
    • Nick Korovaiko's avatar
      LoopKernel Collector (#1128) · 784735d6
      Nick Korovaiko authored
      * collector
      
      * keeping track of inputs; simplifying a merging stratey; adding LKGraph
      
      * LoopKernel Collector
      
      * address feedback
      
      * address feedback 2
      
      * address feedback 3
      784735d6
  13. 26 Jun, 2018 1 commit
    • Jayaram Bobba's avatar
      Convolution sum fusion (#1146) · 82ee0a77
      Jayaram Bobba authored
      * inplace compute
      
      * fix warnings
      
      * Initial support for convolution sum fusion
      
      * Added in-place support for conv sum fusion and test cases
      
      * reverting spurious changes
      
      * Bug fix to account for inplace input in conv sum fusion
      
      * fix compilation error
      
      * Addressed PR feedback
      82ee0a77
  14. 22 Jun, 2018 1 commit
  15. 19 Jun, 2018 1 commit
    • Nick Korovaiko's avatar
      Loop Kernel Op + Tests (#1028) · 96295aaa
      Nick Korovaiko authored
      * loop kernel + tests
      
      * remove commented out code
      
      * remove commented code; add comments
      
      * copy_with_new_args +test
      
      * add comment
      
      * fix comp errors
      96295aaa
  16. 15 Jun, 2018 1 commit
    • Pruthvi's avatar
      RNN fusion across layers (#1085) · f75b8006
      Pruthvi authored
      * - Added graph pass for fusing RNN op across layer
      - Added test case for inter v/s cpu for verifying layer fused RNN
      - more sanity checks in the RNN fusion graph pass
      - added support to replace the recurrent cell state correctly in the fused RNN op
      
      * Fixed multi layer rnn fusion unit test failure
      
      * Addressed PR comments
      f75b8006
  17. 13 Jun, 2018 1 commit
    • Nick Korovaiko's avatar
      Group Convolution (#1041) · 4a2c3c9c
      Nick Korovaiko authored
      *  group conv init
      
      * add GroupConvolution op; refine checks in fusion logic
      
      * add an emitter, cpu assigment
      
      * cpu_layout
      
      * add checks to algebraic simplification
      
      * updating emitter logic for groupconvolution
      
      * working before refactoring
      
      * moving primitive creation logic to mkldnn_emitter
      
      * group convolution graph test
      
      * rename an opt
      
      * address jbobba's feedback
      4a2c3c9c
  18. 07 Jun, 2018 1 commit
    • Louis Feng's avatar
      ngraph-1676 batch dot fusion (#1071) · 6f5e3ac7
      Louis Feng authored
      * batch dot pattern wip.
      
      * batch dot pattern wip.
      
      * added batch dot op.
      
      * batch dot compute testing.
      
      * correct gemm parameters.
      
      * renaming matrix fusions passes and update tests.
      
      * clean up.
      
      * clang format.
      
      * more clean ups.
      
      * clang format.
      
      * added CPUBatchDotFusion to default cpu passes.
      
      * added missing header.
      
      * added element type check.
      6f5e3ac7
  19. 06 Jun, 2018 1 commit
  20. 31 May, 2018 1 commit
  21. 30 May, 2018 2 commits
  22. 23 May, 2018 1 commit
    • Pruthvi's avatar
      LSTM fusion + RNN fusion across time slice's for single layer (#826) · 1d08f073
      Pruthvi authored
      * - Added pattren matcher for LSTM cell
      
      * WIP added support to replace lstm cell instead of subgraph
      
      * WIP LSTM pattern matcher, fuses recurrent cells
      
      * WIP added RNN CPU op
      
      * WIP mkldnn emmiter code for fprop RNN
      
      * WIP RNN mkldnn integration
      - Added mkldnn kernel for uni directional LSTM in the CPU emitter
      
      * add a getter for root node
      
      * recurrent graph rewrite
      
      * fix perms, rename match_root -> get_match_root
      
      * fix comp errors
      
      * make match_root return the topmost match; fix tests
      
      * - WIP GetOutputElement for handling multiple LSTM o/ps
      - use RecurrentGraphRewrite for replacing node after matching LSTM cells
      
      * WIP LSTM multi Output + debug prints
      
      * moved LSTM fusion to cpu_fusion
      
      * WIP added RNN superfused OP
      
      * WIP towards RNN layer fusion
      
      * WIP multiple output slicing RNN
      
      * WIP RNN mulitple o/ps fusion across layer
      
      * WIP corrected input params for fused RNN OP
      
      * concat corrosponding param's across differnt LSTM to form inputs to RNN fused op
      
      * i) Added  test case for RNN kernel ii) runs without error's
      
      * refactored and moved LSTM class to standalone file
      
      * Rename RNN -> Rnn , LSTM -> Lstm
      
      * WIP replace lstm slices to the consumer op
      
      * Slicing works on multiple RNN layers
      
      * fixed all bugs
      
      * - Added CPU RNN Recurrent Fusion
      - Added CPU LSTM fusion
      - removed debug code
      - style fix
      
      * - Added support to compute src_iter and dst_iter instead of taking zero_memory_desc
      - Added unit test to compute one LSTM cell
      
      *  changed RNN op signature to accept number of states in basic unit of RNN(GRU/LSTM/ vanilla RNN) cell
      
      * added sanity checks for RNN op
      
      * Fixed issue related to patching the graph while replacing the RNN sliced outputs
      
      * Fixed issue to feed the input symbols in the order X0, X1, ...Xt to the RNN op
      
      * Added unit test for multi layer RNN fusion
      
      * Removed debug statements
      
      * Added mulitlayered serialized graph ii) fixed compilation issue
      
      * Addressed PR comments
      
      * i) WIP MKLDNN layout for RNN Op ii) added test case for INTERPRETER v/s CPU Rnn results
      
      * - Fixed bug w.r.to src_layer feature size in rnn mkldnn emitter code
      - Refactored cpu_fusion rnn test case
      
      * merge origin/master with branch pruthvi/lstm_fusion
      
      * style fix
      
      * Added test case for multiple RNN layers
      
      * i) make rnn as mkldnn op if it meets the constraints ii) assert if rnn is not mkldnn op
      
      * fix unit test failure
      
      * - Added support to reliabily identify the hiddent state and input symbols from the nodes collected by Pattern matcher
      - Fixed failing unit tests
      
      * style fix
      
      * - removed "node type" dependency to replace the intermediate LSTM outputs
      
      * Addressed PR comments
      
      * Fix unit test
      
      * - added MKLDNN emitter for LSTM op
      - graph pass to concat LSTM input recurrent state tensors
      - CPU layout assignment for LSTM Op
      - Fixed bug in rnn/lstm unit test's
      - made changes to use replace_output instead of replace_node for replacing matched graph nodes in LSTM/RNN fusion pass
      
      (cherry picked from commit d16fc709265cc0a73e60c6d5f6d2878e7b908aca)
      
      * style fix
      
      * Renamed passes and style fixes
      1d08f073
  23. 21 May, 2018 1 commit
  24. 16 May, 2018 1 commit
  25. 08 May, 2018 1 commit
    • Nick Korovaiko's avatar
      MaxPoolWithIndices (#900) · a174c8c9
      Nick Korovaiko authored
      * MaxPoolWithIndices CPU Fusion
      
      * fix test to pass checks in cpu_fusion
      
      * pass test
      
      * clean up
      
      * add a new pass, add layouts
      
      * remove the opt from cpu_fusion
      
      * refactor cpu_layout logic for maxpool, clean up comments
      
      * add comment w.r.t. indices tensor
      
      * rename to cpu_workspace_insertion
      
      * add CPUWorkspaceInsertion pass for TF
      a174c8c9
  26. 04 May, 2018 1 commit
  27. 23 Apr, 2018 1 commit
  28. 18 Apr, 2018 1 commit
    • Nick Korovaiko's avatar
      Weight Fusion (#853) · 8cb48d37
      Nick Korovaiko authored
      * CPU weight fusion initial version
      
      * add tests for weight_fusion
      
      * address @jbobba's feedback
      
      * before cleaning up convolution_weight_optimization.cpp
      
      * clean up, rename, fix perms, fix format
      8cb48d37
  29. 16 Apr, 2018 1 commit
  30. 13 Apr, 2018 1 commit
    • Robert Kimball's avatar
      Remove legacy Backend API (#848) · ec501913
      Robert Kimball authored
      * remove deprecated
      
      * remove all legacy Backend API usage
      
      remove deprecated files
      
      * pull in changes from master
      
      * fix GPU calls
      
      * disable tests in convolution generator
      
      * update per PR comments. Enable performance counter feature.
      
      * update per PR comments
      
      * fix build error
      
      * fix conditionally compiled test :(
      ec501913
  31. 09 Apr, 2018 1 commit
  32. 04 Apr, 2018 1 commit
    • Nick Korovaiko's avatar
      Support multi-output ops in Adjoints (#796) · 5f0e8dc3
      Nick Korovaiko authored
      * refactor Adjoints to support multi-output ops
      
      * passing tests
      
      * switch to generate_adjoints(deltas) and backprop_node
      
      * remove debugging code
      
      * fix error msg
      
      * fix typo adjoitns
      
      * fix comp errors in mnist_mlp
      5f0e8dc3
  33. 03 Apr, 2018 1 commit
  34. 02 Apr, 2018 1 commit
    • Pruthvi's avatar
      Pruthvi/bn to support globalstats (#783) · 1d80cabe
      Pruthvi authored
      * WIP support bn training for global_stats
      
      (cherry picked from commit eb81a37328ea177b1d58c9eebdbb345e0fa25f0d)
      
      * - Style fix
      - Fix test case
      
      * Addressed PR comments
      - added support for bn training/inference with a same ctor
      - added more verbose comments in bn header
      
      * Fixed bn serializer and default value in bn ctor for bwd compatibility
      
      * proposed docs change
      
      * - Addressed PR comments
        - added support to compute bn inference/training using same mkldnn kernel with global stats
      
      * fix unit bn relu unit test
      1d80cabe
  35. 30 Mar, 2018 1 commit
  36. 29 Mar, 2018 1 commit