1. 21 Sep, 2019 1 commit
  2. 19 Sep, 2019 1 commit
  3. 05 Sep, 2019 1 commit
    • Amy Zhuang's avatar
      Use mkl-dnn v1.0 or v0.x depending on compilation flag. (#3227) · e26d602a
      Amy Zhuang authored
      * Use mkl-dnn v1.0 or v0.x depending on compilation flag.
      
      * Change cpu builder files.
      
      * Modify cmake files.
      
      Use mkldnn-v1.0 for DEX if NGRAPH_USE_MKLDNN_V1 is set to true, otherwise use mkldnn-v0.x.
      
      CODEGEN only builds with mkldnn-v1.0.
      
      * Implement mkldnn utility functions for mkldnn-v1.0.
      
      User mode scratchpad management for mkldnn-v1.0.
      
      * Query scratchpad size and allocate a buffer of max scratchpad size.
      
      * Do not create mkldnn::memory when query scratchpad size of Reorder.
      
      Modify mkldnn utility functions.
      
      Fix convolution_forward_init and inner_product_forward_init.
      
      Modify CPURuntimeContextCG.
      
      * Add user mode scratchpad to CODEGEN.
      
      * mkldnn-v1.0 splits LSTM states. Update Rnn/Lstm Op accordingly.
      
      * Address PR feedback: use MKLDNN_MAJOR_VERSION.
      
      * Modify cpu rnn fusion pass and related unit tests.
      
      * Change Rnn/Lstm arg types to Output.
      
      * Fix Lstm for CODEGEN.
      
      * Set native layout for Slice when input format is blocked.
      
      * Do not print scratchpad size.
      
      * Change external_mkldnn_v1.cmake.
      
      Fix a typo.
      
      * Add mkldnn_v1.patch for mkldnn-v1.0.
      
      * Address PR feedback.
      
      * Define MKLDNN_ERROR_MESSAGE.
      
      * Address PR feedback: change to NGRAPH_USE_LEGACY_MKLDNN.
      
      * Fix a bug.
      
      * Remove unused variable.
      
      * Fix compiler warnings.
      
      * Fix a bug for CODEGEN.
      
      * Move variable only needed for mkldnn-v0.20 inside #if.
      
      * Remove unused variables.
      
      * No in place Reshape rotation for blocked data layout with mkldnn-v1.0.
      
      * Modify mkldnn_v1.patch to force mkldnn to link to libiomp.
      
      * Fix style.
      
      * Change path for find_library and find_file.
      
      * Do not insert ConvertLayout before/after Quantize/DeQuantize for blocked data layout.
      
      * Write strides information to visualized graph.
      
      * Move variables only needed for mkldnn-v0 under #if.
      
      * Move more variables in rnn fusion.
      
      * Fix ConvertLayout constant folding for mkldnn-v1.0.
      e26d602a
  4. 29 Aug, 2019 1 commit
    • Nagy Mostafa's avatar
      [MLIR] Fixes for cpu_fusion.validate_fuse_gru_inputs (#3511) · ef58667f
      Nagy Mostafa authored
      * WIP
      
      * Fix incorrect CK output adjustment
      
      * Bug fix and enroce sanity check
      
      * Change cycle search depth, and fix sanity check
      
      * cpu_fusion.validate_fuse_gru_inputs passes.
      
      * Fix as_single_output to be able to always create a GOE
      
      * minor fix. style-apply
      
      * Clean up debug msgs
      
      * Switch to backward cycle check
      
      * Enable failing test
      
      * PR fixes
      
      * Address feedback: Add fwd cycle checks. Make cycle checking depth configurable
      ef58667f
  5. 26 Aug, 2019 1 commit
    • Diego Caballero's avatar
      [MLIR] Disable three more tests with MLIR enabled. (#3496) · 1683e200
      Diego Caballero authored
      * [MLIR] Bump MLIR repo 8/20/2019
      
      MLIR_
      commit 0cdb20a6add19bc96c20dad28589a1e54e4d8469
      Author: Lei Zhang <antiagainst@google.com>
      Date:   Tue Aug 20 13:33:41 2019 -0700
      
          Add spv.specConstant and spv._reference_of
      
      LLVM:
      commit 3b9a27b6908040881dad394022f8c472c15c0784
      Author: Simon Pilgrim <llvm-dev@redking.me.uk>
      Date:   Tue Aug 20 17:54:37 2019 +0000
      
          Fix typo in comment. NFCI.
      
      * [MLIR] Disable three more tests with MLIR enabled.
      
      This PR disables validate_fuse_gru_inputs, reshape_layout_optimizations4
      and reshape_layout_optimizations5:
        1. trivial_in_place_relu_fail: It checks tensors pool offset. There is
           no memory pool in MLIR atm.
        2. validate_fuse_gru_inputs: It creates an infinite cycle in
           MLIR subgraph extraction pass (under investigation).
        3. reshape_layout_optimizations4/5: They fail due to CompiledKernel
           being not expected by CPULayout pass.
      
      * Disable cpu_quant_fusion.qconcat
      1683e200
  6. 21 Aug, 2019 1 commit
  7. 15 Aug, 2019 2 commits
    • Pruthvi's avatar
      LSTM MKLDNN integration for ONNX LSTM op (#3327) · e5d606b8
      Pruthvi authored
      * - Add graph pass method for onnx lstmcell rewrite with lstm cpu op
      - insert reshapes to keep the weights in ldigo format
      - test case for onnx LstmCell to CPU Lstm
      
      * fix typo
      
      * - check LSTMCell for the fused op decomposistion in the backend
      
      * - fix bug in onnx_lstm graph pass
      - passes unit test
      
      * style-fix
      
      * - fix compilation error
      - use IFCO gate ordering for bias
      
      *  - Skip LSTMCell to LSTM CPU fusion for peephole
      
      * - add comment && remove duplicate function
      
      * -use dynamic_pointer_cast to check for constant
      
      * - onnx bias will be of shape (2 * gates_count * hidden_size) bias of Wb and Rb are concatenated, we will split the bias, add and rearrange in order IFCO
      
      * - Use most derived LSTM ctor for pattern matching
      
      * - Style Fix
      
      * style fix
      
      * Address PR comments
      
      * - add support for graph pass (MKLDNN version > 1) for mapping LSTMCell -> LSTM CPU op
      
      * fix unit test failure for MKLDNN V1.0
      e5d606b8
    • Diego Caballero's avatar
      [MLIR] Disable CPU fusion + degug tracer tests in MLIR (#3442) · 5bbd199b
      Diego Caballero authored
      CPU fusion is disabled in MLIR since fused ops are not supported in
      nGraph dialect. CPU debug tracer test doesn't expect CompiledKernel ops
      generated for MLIR.
      5bbd199b
  8. 31 Jul, 2019 1 commit
  9. 23 Jul, 2019 2 commits
  10. 08 Jul, 2019 1 commit
  11. 03 Jul, 2019 1 commit
  12. 01 Jul, 2019 1 commit
  13. 19 Jun, 2019 1 commit
  14. 18 Jun, 2019 1 commit
  15. 14 Jun, 2019 1 commit
    • gaurides's avatar
      Fuse Dropout (#3006) · 8c38db04
      gaurides authored
      * Initial implementation
      
      * Added test case
      
      * Bug fix; Dropout with 2 outputs, WIP
      
      * Fixed in unit-testl; WIP for model
      
      * Nothing is working
      
      * Revert "Nothing is working"
      
      This reverts commit d3ff09bb7a0d0519ab70ac85f2e7f30721afea96.
      
      * Fixed unit-test; fusion with 2 outputs
      
      * Fix style check, file permissions
      
      * Changed input arg to Node
      
      * Fix order of declaration
      
      * Improved performance
      
      * some cleanup
      
      * Fixed CI error
      
      * Fixed review comments
      
      * Fix CI error
      
      * Remove unused variable
      
      * Fix other CI errors
      
      * Changed type
      
      * Fix style check
      
      * Add codegen code for Dropout
      
      * addressed PR feedback; will add codegen support later
      
      * Cleanup; change variable name
      
      * Support for use_seed
      
      * Add setter for use_seed
      
      * Add setter for use_seed
      
      * Fix CI error
      
      * Make use_seed as arg
      
      * Fix CI error
      
      * Fix CI error
      8c38db04
  16. 12 Jun, 2019 1 commit
  17. 08 Jun, 2019 1 commit
  18. 05 Jun, 2019 1 commit
  19. 02 Jun, 2019 2 commits
  20. 24 May, 2019 1 commit
    • Michał Karzyński's avatar
      [Fused] LeakyRelu op (#2919) · 5650e913
      Michał Karzyński authored
      * [Fused] LeakyRelu op
      
      * Add LeakyRelu to serializer
      
      * Add unit tests
      
      * Fix merge branch 'master' into mkarzyns/fused_leaky_relu
      
      * Change broadcasting rules to NumPy style
      
      * Remove std:: and ngraph:: prefixes
      
      * Rename CPU Runtime LeakyRelu to CPULeakyRelu
      
      * Style apply
      
      * Fix cpu_fusion.fuse_leaky_relu test
      
      * Use eigen's tanh in the fused sigmoid multiply kernel (#2946)
      
      * Merge branch 'master' into mkarzyns/fused_leaky_relu
      
      * Add LeakyRelu to Intel GPU backend op list
      
      * Add LeakyRelu to Intel GPU backend op list
      5650e913
  21. 23 May, 2019 2 commits
  22. 22 May, 2019 3 commits
  23. 13 May, 2019 1 commit
    • Scott Cyphers's avatar
      Fix clang compiler warnings (#2898) · b94a042d
      Scott Cyphers authored
      * Fix clang compiler warnings
      
      * Remove unintended file.
      
      * style
      
      * Not part of PR
      
      * Another extra closure ref
      
      * More warnings from merges
      
      * Lambda arg was used
      b94a042d
  24. 29 Apr, 2019 1 commit
  25. 26 Apr, 2019 1 commit
  26. 17 Apr, 2019 1 commit
    • gaurides's avatar
      DeconvBias (#2716) · 03f13e4b
      gaurides authored
      * deconv optimizations for dcgan
      
      * Added test cases
      
      * modified some tests, not working at this point
      
      * Removed temp code
      
      * fixes to get unit test to pass
      
      * Added node validation checks
      
      * Update mkldnn emitter to memory reuse design
      
      * Code cleanup
      
      * Fix to enable deconv select the right kernel
      
      * Fix file permissions
      
      * Disabled unit test cases
      
      * Remove unused variable
      
      * Address PR feedback
      
      * Removed dead code
      
      * Style check
      
      * removed dead code
      03f13e4b
  27. 16 Apr, 2019 1 commit
    • Jayaram Bobba's avatar
      Moves some fused convolution ops to core FusedOps (#2733) · 6b5016e5
      Jayaram Bobba authored
      * - Moves some fused convolution ops to core FusedOps
      - Adds support for decomposing and replacing multi-output FusedOps
      - Adds query callbacks to FusedOpDecomposition to check if a FusedOp is
        supported by a backend
      - Adds core fusion patterns for FusedOps
      -
      
      * style fix
      
      * Added comments on FOP_FUSIONS
      
      * gpu convolution 1d bug fix (#2741)
      
      * Fix bug with dex-only compilation and addressed PR comments
      6b5016e5
  28. 12 Apr, 2019 1 commit
    • Adam Procter's avatar
      Deprecate direct access to descriptor::Input and descriptor::Output (#2724) · 5490bae5
      Adam Procter authored
      * Add NodeInput and NodeOutput classes
      
      * Deprecate Node::get_inputs, Node::get_outputs, Node::get_output_inputs. Remove Node::get_input_from and Node::get_output_from
      
      * Privatize most fields of Node
      
      * Make deprecation of descriptor-munching classes optional
      
      * Review comments
      
      * Adapt ReshapeSinking to use raw pointers for NodeInput
      
      * Fix ZDTE (thought I had already done in this branch, weird); style
      
      * wip
      
      * Change get_node_outputs() and get_node_inputs() to return vectors
      
      * Updates after merge
      
      * Whoops, forgot to define these functions
      
      * {NodeInput,NodeOutput} -> {Input,Output}
      
      * Kill shared_ptr in Output
      
      * Move Input and Output into node.hpp
      
      * Templatize the underlying node (sub)type in Input and Output
      
      * Eliminate some get_input_* and get_output_* functions
      
      * Change get_outputs and get_inputs back to their original names; rename NGRAPH_DEPRECATE_IO_DESCRIPTORS to NGRAPH_DEPRECATE_OLD_NODE_APIS
      
      * Miscellaneous cleanup
      
      * More cleanup
      
      * Unbreak CPU build
      
      * Simplify unit tests
      
      * Make Node less friendly
      
      * Deprecate more get_output_* and get_input_* functions
      
      * A couple of PR comments
      
      * Make the deprecation stuff more generally available
      
      * Better comment
      
      * Be more consistent about [] vs. at
      5490bae5
  29. 11 Apr, 2019 1 commit
    • Louis Feng's avatar
      [Dynamic Shape] Moving BatchDot to Core Op (#2691) · cc8dd452
      Louis Feng authored
      * batch dot WIP.
      
      * cpu backend refactor and unit tests pass.
      
      * WIP.
      
      * batch dot interpreter impelementation.
      
      * minor clean up.
      
      * more clean up.
      
      * patching the gpu backends.
      
      * added more tests, fixes, etc.
      
      * fixed compile error.
      
      * renamed batch dot to batch matmul.
      
      * refactor WIP.
      
      * fixes some tests and formating.
      
      * more fixes.
      cc8dd452
  30. 09 Apr, 2019 1 commit
  31. 26 Mar, 2019 1 commit
  32. 21 Mar, 2019 1 commit
    • tsocha's avatar
      [ONNX] Enable Pad modes for ONNX pad operator (#2590) · f8146495
      tsocha authored
      * Add support for negative padding
      
      * Use std::bind in pad builder check
      
      * Add support for negative padding in CPU backend
      
      * Updated kernel to do pad+slice
      
      * Remove type conversion warnings
      
      * Fix review comments
      
      * Remove interior padding from core op and interpreter stuff
      
      * Update backends other than GPU for retirement of padding_interior
      
      * Skeleton of support for edge/reflect padding
      
      * Post-merge cleanup
      
      * Attempt reference implementation for EDGE.
      
      * Fix the edge-padding reference, and add some unit tests
      
      * Implement REFLECT padding ref; add tests
      
      * Fixes to the CPU stuff so it compiles now
      
      * Fix test
      
      * Add support for different pad modes
      
      * Restore a stub get_padding_interior function, and tweak some stale comments
      
      * Update ONNX importer to not supply interior padding value; add checks for padding-too-small for EDGE and REFLECT
      
      * Typo
      
      * Bop a warning
      
      * Attempt fix to INTELGPU backend
      
      * Attempt another fix to INTELGPU backend
      
      * Fix pyapi
      
      * Style apply
      
      * Add support for padding modes
      
      * Remove unnecesary node validation checks
      
      * Remove tests for minimal reflect and edge pad
      
      * Remove commented tests
      
      * Remove unnecesary Asserts
      
      * Little update of pad documentation
      
      * Monospace for pad_mode options
      
      * Revert "Remove tests for minimal reflect and edge pad"
      
      This reverts commit 81e4787ea47195b832cab1452dde698bc05776fe.
      
      * Revert "Remove unnecesary node validation checks"
      
      This reverts commit 7e68db7564f3c9b1fd40e7db1d1bda4e0677cad9.
      
      * Test only spatial dims
      
      * axis -> spatial axis
      
      * Fix typo
      
      * Style check
      
      * Update test
      
      * Add CoordinateDiff include
      
      * Remove pad_mode from tree visualization
      
      * Convert padding into NVShape
      
      * Skip failing tests on GPU
      
      * Revert mode change
      
      * Remove merge artifact
      
      * Rename pad kernel into pad_ref
      f8146495
  33. 18 Mar, 2019 1 commit
    • Robert Kimball's avatar
      Change floating point comparisons from == to all_close_f (#2620) · 56e160ba
      Robert Kimball authored
      * change float comparisons from == to all_close_f
      
      * style
      
      * address a few more direct float comparisons
      
      * add missing include
      
      * specify tightest tolerance for Broadcast and Reshape tests
      
      * Increased tightness of float testing
      
      Increased tightness of float testing via MIN_FLOAT_TOLERANCE_BITS parameter
      
      * style
      56e160ba
  34. 26 Feb, 2019 1 commit
    • Jayaram Bobba's avatar
      More quantized fusion patterns (#2480) · b8106133
      Jayaram Bobba authored
      * Add QuantizedConcat
      
      * Remove unused variables and add check for size of mins and maxes vector
      
      * Resolve conflicts
      
      * Merged with master and addressed some PR feedback
      
      * Maxpool and Avgpool fusions. Exclude Q from conv+relu fusion
      
      * Remove single-user check from fusions
      
      * Quantized concat fusion
      
      * workaround: do reshape sinking by default
      
      * style fix
      
      * check scales for QuantizedConcat
      
      * use compare_constants
      
      * remove stale comment
      
      * Handle all concat cases from arg size 2 to 6
      
      * addressed feedback
      b8106133