- 06 Dec, 2018 1 commit
-
-
Pruthvi authored
* - Added reorder support for rnn weights_layer/iter * i) fixed compilation issues ii) working but still observing precision error * i) fixed failing rnn unit test for DEX ii) refactored workspace in RNN mkldnn emitter * i) added support for src reorder to TNC from NTC * reorder support for rnn output fron NTC to TNC * - added support for rnn weight reorder ldgoi -> ldigo - code refactor for lstm/rnn kernel in mkldnn emitter * - refactor rnn mkldnnn kernel, change variable names * fix RNN codegen kernel * disbale layer rnn fusion pass, to test CI * method to validate recurrent rnn inputs * add correlated macthes for Recurrent RNN PM * - simplify reorder logic for rnn_weights - fix graph pattern for fusing rnn cell across time steps * do weights reorders in rnn timesteps fusion * refactored LSTM graph pass * - Bug fix for finding the lstm inputs determenstically - Refactored LSTM graph pass to single pass - made changes to LSTM RNN time step fusion graph pass * - use replace_node instead of replace_output in Lstm_step_wise fusion graph pass * fix compilation error * Fix GNMT rnn fusion * check if the node is in use before replacing in RNN graph passes * i) fix style ii) fix topo sort issue in RNN graph pass * style fix * fix bug in simplify_concat pass * replaces Lstm1 -> {GOE1, GOE2} -> {Slice1, Slice2} -> Concat -> Lstm2 with Lstm1 -> Lstm2 * cse for convert layout * addressed PR comments * - optimization pass to remove Lstm1 -> {GOE1, GOE2} -> {Slice1, Slice2} -> Lstm2 - conditional fusing of LSTM cells only for the decoder * made changes to multi layer RNN fusion callback * fix asserts in RNN op * - added support to fuse layers when slc=dlc for RNN cells - bug fix on the sanity checks for RNN Op * - support RNN layer fusion till slc = dlc - bug fixes in multi layer rnn fusion call back * capture reshape in the RNN weights * Addressed PR comments * - added comments in multi layer PM call back - fuse only if slc == DLC across layers * restore deleted 3_lstm_cell_forward.json file * fix typo * fix failing unit tets * When processing in place slice, do not change the offset of the slice node if the argument pointer comes from function input. * Address PR feedback: process in place slice after propagating in place input. * Set INTERMEDIATE role before propagating in place input. * Do not add temporaries to the variable name map before propagating in place input in codegen. * Fix a bug in codegen. * Fix a bug in codegen slice. * reenable disabled rnn unit test * fix compiler error * - bug fix in the slicing logic for the layer fused rnn cell - fix failing rnn unit test * - Addressed PR comments - removed redundant checks from the rnn graph pass - simplified rnn call back replace node logic * - added new multilayer rnn *.json file - fix test case * [PRIVATE BRANCH] Style fixes (#2080) * Style fixes * change order of lstm gates * [PRIVATE BRANCH] Jbobba/rnn fusion review (#2113) * Style fixes for single-layer RNN fusion * Style fixes to multi-layer RNN * style fix * disable GPU test
-
- 14 Nov, 2018 1 commit
-
-
Adam Rogowiec authored
* Unit tests for conv2d causing errors. * UT for conv3D_bias * Fix padding order. `padding below` in nGraph terminology means padding added at the beginning of the axis. Whereas `padding above` means padding added at the end of the axis. * Rename test to sth more descriptive. * Apply clang-format. * Fix handling of `SAME_UPPER/LOWER` auto_pads mode for convolution/pooling ops. * Fix order of padding_below/above. Signed-off-by:
Adam Rogowiec <adam.rogowiec@intel.com> * Fix error in calculating output data shape.
-
- 13 Nov, 2018 1 commit
-
-
Adam Rogowiec authored
* Add static keyword for helper function. * Fix MatMul for cases where left hand side is 1D vector. - Add unit-test for this case. * Add new line at the end of file. * Log warning when dealing with scalars * Apply clang-format * Review: fix spelling, rename test model.
-
- 30 Oct, 2018 1 commit
-
-
Michał Karzyński authored
* Add ArgMin operator * Add ArgMax and a basic test case * Rename variables * Apply workaround for problems with Reshape on i64 * Review comments * Review comments
-
- 23 Oct, 2018 1 commit
-
-
Artur Wojcik authored
* onnx: add information about a domain to operators set Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: updates after review Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: update comments in the code Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix bug in node's description method Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix CentOS compilation Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: more after review changes Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com>
-
- 22 Oct, 2018 1 commit
-
-
Nick Korovaiko authored
* split bn into bn_inference bn_training * fix warnings * Add GPU support for the new BN ops (#1569) * Add GPU support and change batchnorm_globalstats test to use BNInference. * Changed test back to using BNTraining for global stats and updated cudnn backend to account for it. * Fix issues in merge with master. * Formatting. * CPU fixes * remove 5-arg training BN for now * more fixes * python batchnorm changes * fix onnx_import * fix a call BatchNormInference c-tor * yet another fix to BatchNormInference c-tor * AND yet another fix to batchnorm_inference c-tor * ops.py * address adam's feedback * Remove unnecessary parameter/argument. * remove batch_norm_training_relu_with_global_stats * remove bn_relu (training)
-
- 15 Oct, 2018 2 commits
-
-
Michał Karzyński authored
* [ONNX] Assert all op types supported * Apply clang-format * Address code review comments * Fix #include statements
-
Adam Rogowiec authored
* Update ONNX Squeeze Op implementation to conform with doc. Add unit test. * Apply code-format. * Correct attribute value type. * Change used loop structure. * Modified version of loops. - Without erase and with minimal computation time complexity. * Run CI
-
- 10 Oct, 2018 1 commit
-
-
Nick Korovaiko authored
* reshape sinking working on mnist_conv * forgot to add reshape_sinking files * refactoring of binary case * Quantize/Dequantize case, fix add case, add assert * address bob and scott's feedback * debug * fix a bug where reshapes are removed too early
-
- 26 Sep, 2018 1 commit
-
-
Adam Straw authored
* adding nGraph Quantize op * unit test failing for floating point exception * unit test working in float * unit test working in uint8 * improved type checking and polished unit test - passing * quantized axes working * inclusive project method * add round mode * TODO cleanup * code format * adding serializer support - fails build * add serializer support * make CPU quantize op work; new tests for int8, clamp) * fix build failure * fix GPU build issue * fix GPU unit test manifest * use quantized offset * add is_quantized field to element::Type * add reduce function to coordinate.hpp
-
- 14 Sep, 2018 1 commit
-
-
tsocha authored
* [ONNX] Non-linear operators * Review fix pt. 1 * Review fix pt. 2 * Non-linear tests * style check * Exception fix * Test fix
-
- 12 Sep, 2018 2 commits
-
-
Adam Rogowiec authored
* Add missing header. * Test for ReduceSum * Simple tests for reductions - L1/L2/LogSum/LogSumExp/Max/Mean/Min/Prod/SumSquare. * Add floating point literal suffix * Fix typo
-
tsocha authored
* [ONNX] Shape operator * Review fix pt. 1 * Style check
-
- 04 Sep, 2018 1 commit
-
-
tsocha authored
-
- 03 Sep, 2018 1 commit
-
-
Adam Rogowiec authored
* Move reshape utils down to reshape namespace. * Reshape operation. * Reshape operator binding. * Error fixes. * Reshape unit tests. * Move flatten utility function to reshape namespace. * Fix unused catched exception object * Add Constant support for int64 * Review fix. * clang-format * Review fix part 2. * Enable output shape as a second node input (only Constant). * Unit test for "dynamic" output shape (from Constant node). * Review fixes. * Make sure second Reshape op input is Constant node.
-
- 31 Aug, 2018 2 commits
- 30 Aug, 2018 1 commit
-
-
tsocha authored
-
- 29 Aug, 2018 2 commits
-
-
Michał Karzyński authored
* [ONNX] Sum op * [ONNX] Generic variadic op template * Add support for Min op * clang-format * Add support for Max op * Add support for Mean op * Docs, code cleanup * Docs, code cleanup
-
tsocha authored
-
- 28 Aug, 2018 2 commits
-
-
tsocha authored
* [ONNX] Softmax operator * Review fix pt. 1 * Review fix pt. 2 * Add softmax test * Update onnx_import.cpp
-
Michał Karzyński authored
-
- 27 Aug, 2018 1 commit
-
-
tsocha authored
* [ONNX] MatMul operator * Add NL on EOF * Review fix pt. 1
-
- 24 Aug, 2018 2 commits
-
-
tsocha authored
* Enable Mul OP * Reshape, broadcasting utils and Gemm op * Style check * Review fix pt. 1 * Review fix pt. 2 * Reuse documentation
-
Michał Karzyński authored
* Move batch_norm implementation to a .cpp file * Move split implementation to a .cpp file
-
- 23 Aug, 2018 1 commit
-
-
Michał Karzyński authored
* [ONNX] Refactor exceptions * [ONNX] Attribute helper functions * [ONNX] Convolution operation
-
- 21 Aug, 2018 1 commit
-
-
Michał Karzyński authored
* [ONNX] Add Relu op
-
- 14 Aug, 2018 1 commit
-
-
Adam Rogowiec authored
* onnx: add 'constant' operator Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: getting attribute value by name Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix code style Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix clang compilation warnings Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: exception Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add 'split' operator Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add public interface Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add initial unit test for importer Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: initial implementetion of operator' set Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * [WIP] Unit test for split operation. * Fix Split Op bounds calculation + UT * clang format * Split Op with variable parts unit test. * Remove unused headers * General purpose exceptions. * Change not_supported_error message template. * Add new general purpose errors. * ONNX BatchNormalization operation. * Clang-format * Update CMake * Add fixed test data. * Add missing ngraph install prefix for cmake in travis Dockerfile. * Remove -Wno-zero-as-null-pointer-constant * Code review * Apply clang-format-3.9 * Add missing onnx_import interface files to CMakeList * Clean code. * Fix test. * Apply clang-format-3.9 * Copyright notice format * Remove inputs in separate files * use all_close to compare floating point values * missed changing one CPU to INTERPRETER for unit test
-
- 10 Aug, 2018 1 commit
-
-
Artur Wojcik authored
* onnx: add 'constant' operator Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: getting attribute value by name Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix code style Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: fix clang compilation warnings Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: exception Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add 'split' operator Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add public interface Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: add initial unit test for importer Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * onnx: initial implementetion of operator' set Signed-off-by:
Artur Wojcik <artur.wojcik@intel.com> * [WIP] Unit test for split operation. * Fix Split Op bounds calculation + UT * clang format * Split Op with variable parts unit test. * Remove unused headers * Add missing ngraph install prefix for cmake in travis Dockerfile. * Remove -Wno-zero-as-null-pointer-constant * Code review * Apply clang-format-3.9 * Add missing onnx_import interface files to CMakeList * Copyright notice format
-
- 30 Jun, 2018 1 commit
-
-
Pruthvi authored
* - Fixed replace output for the multi layer recurrent cell state tensor output - Modified rnn add_output to consider direction and n_layer while calculating the output size for mkldnn dst_layer and dst_iter * fix unit test failure
-
- 15 Jun, 2018 1 commit
-
-
Pruthvi authored
* - Added graph pass for fusing RNN op across layer - Added test case for inter v/s cpu for verifying layer fused RNN - more sanity checks in the RNN fusion graph pass - added support to replace the recurrent cell state correctly in the fused RNN op * Fixed multi layer rnn fusion unit test failure * Addressed PR comments
-
- 07 Jun, 2018 1 commit
-
-
Louis Feng authored
* batch dot pattern wip. * batch dot pattern wip. * added batch dot op. * batch dot compute testing. * correct gemm parameters. * renaming matrix fusions passes and update tests. * clean up. * clang format. * more clean ups. * clang format. * added CPUBatchDotFusion to default cpu passes. * added missing header. * added element type check.
-
- 31 May, 2018 1 commit
-
-
Louis Feng authored
-
- 23 May, 2018 1 commit
-
-
Pruthvi authored
* - Added pattren matcher for LSTM cell * WIP added support to replace lstm cell instead of subgraph * WIP LSTM pattern matcher, fuses recurrent cells * WIP added RNN CPU op * WIP mkldnn emmiter code for fprop RNN * WIP RNN mkldnn integration - Added mkldnn kernel for uni directional LSTM in the CPU emitter * add a getter for root node * recurrent graph rewrite * fix perms, rename match_root -> get_match_root * fix comp errors * make match_root return the topmost match; fix tests * - WIP GetOutputElement for handling multiple LSTM o/ps - use RecurrentGraphRewrite for replacing node after matching LSTM cells * WIP LSTM multi Output + debug prints * moved LSTM fusion to cpu_fusion * WIP added RNN superfused OP * WIP towards RNN layer fusion * WIP multiple output slicing RNN * WIP RNN mulitple o/ps fusion across layer * WIP corrected input params for fused RNN OP * concat corrosponding param's across differnt LSTM to form inputs to RNN fused op * i) Added test case for RNN kernel ii) runs without error's * refactored and moved LSTM class to standalone file * Rename RNN -> Rnn , LSTM -> Lstm * WIP replace lstm slices to the consumer op * Slicing works on multiple RNN layers * fixed all bugs * - Added CPU RNN Recurrent Fusion - Added CPU LSTM fusion - removed debug code - style fix * - Added support to compute src_iter and dst_iter instead of taking zero_memory_desc - Added unit test to compute one LSTM cell * changed RNN op signature to accept number of states in basic unit of RNN(GRU/LSTM/ vanilla RNN) cell * added sanity checks for RNN op * Fixed issue related to patching the graph while replacing the RNN sliced outputs * Fixed issue to feed the input symbols in the order X0, X1, ...Xt to the RNN op * Added unit test for multi layer RNN fusion * Removed debug statements * Added mulitlayered serialized graph ii) fixed compilation issue * Addressed PR comments * i) WIP MKLDNN layout for RNN Op ii) added test case for INTERPRETER v/s CPU Rnn results * - Fixed bug w.r.to src_layer feature size in rnn mkldnn emitter code - Refactored cpu_fusion rnn test case * merge origin/master with branch pruthvi/lstm_fusion * style fix * Added test case for multiple RNN layers * i) make rnn as mkldnn op if it meets the constraints ii) assert if rnn is not mkldnn op * fix unit test failure * - Added support to reliabily identify the hiddent state and input symbols from the nodes collected by Pattern matcher - Fixed failing unit tests * style fix * - removed "node type" dependency to replace the intermediate LSTM outputs * Addressed PR comments * Fix unit test * - added MKLDNN emitter for LSTM op - graph pass to concat LSTM input recurrent state tensors - CPU layout assignment for LSTM Op - Fixed bug in rnn/lstm unit test's - made changes to use replace_output instead of replace_node for replacing matched graph nodes in LSTM/RNN fusion pass (cherry picked from commit d16fc709265cc0a73e60c6d5f6d2878e7b908aca) * style fix * Renamed passes and style fixes
-
- 30 Mar, 2018 1 commit
-
-
Nick Korovaiko authored
* initial refactoring using PM * unit test pass * cosmetic changes * add another rnn test * address louis' feedback * lower-case labels
-
- 09 Mar, 2018 1 commit
-
-
Pruthvi authored
* - Added sigmoid fusion pass - added mkldnn emitter code for sigmoid * - corrected sigmoid expected values - add layout assignment for sigmoid op * - added assert's in cpu fusion for sigmoid - style fix * remove debug prints * NGMX-371 #comment addressed PR comments - Added sigmoid unit test case with 3D input ii) support in cpu_emmiter for sigmoid to handle all input shapes * NGMX-371 #comment use shape_size() to calculate the 1d input size
-
- 27 Feb, 2018 1 commit
-
-
sharathns93 authored
* add TF serialzed graphs
-
- 22 Feb, 2018 1 commit
-
-
nikolay.korovaiko authored
-
- 20 Feb, 2018 1 commit
-
-
Ashok Emani authored
* add mxnet sockeye Seq2Seq model * update test with sockeye model
-
- 14 Feb, 2018 1 commit
-
-
Pruthvi authored
* fuse dot(a,b) + c cblas_gemm working on mlp rebase & small fixes enable debug output support replacing function's outputs * WIP pattern matching for variance * - Added pattern matcher graph to look up variance(sub graph) in bn - Added test case to verify the variance graph pattern * added batch norm mean pattern matcher. * remove reshapes (cherry picked from commit ecad321fb1b1bc3f7facda229beb940118ca0701) * fixed mean test to use Matcher. * resolve merge conflict in test/pattern.cpp * WIP bn fprop pattern * fprop bn fusion working * - Added unit test case to read the bn serializeed *.json file and run bn fprop fusion pass - Added batchnorm header file and defined the bn class to emit the mkldnn kernel - Added pattern matcher for fprop bn in CPU graph_rewrite pass * WIP MKLDNN fprop bn emitter code * completed fprop batchnorm kernel in CPU emitter * fixed bug in the emitter code for fprop bn * - Fixed copilation issues - unit tests are passing for bn emitter fprop code * Added support to compute fprop bn with mean annd variance as input * resolved compilation issues * refactored bn fprop code * - added batchnorm src file to the CMakeFilelist - moved bn fusion under CPU runtime/pass/cpu_fusion - fixed compilation issue * Resolved compilation issues in bn emitted code * Addded debug statements in fprop bn emitted code * added batchnorm.cpp src file * - Added test case to test fprop batchnorm with known tensor values - fixed bug related to defining weights in fprop bn * - Added test case for fprop batchnorm Op - Added test case for mean and variance pattern matcher - Added fprop bn *.json file with input having 4dmis mb2c3h2w2 - refactored fprop bn op class * Style fix * - Removed Debug symbols * - Fixed header template with correct year - appended mkldnn.hpp in the CPU generated code * Addressed PR review comments - added support for batchnorm op in serializer and de-serializer - added more sanity in bn constructor - renamed "BatchnormFprop" -> BatchNorm * - Addressed PR review comments - replaced auto with speicfic mkldnn::type in emitted bn kernel - modified function signature to take 'eps' as double instead of <Node> type * added missing header files, resolved compilation issue * style fix * Addressed PR comments 1. initilized member variables for bn in the same order as they are defined 2. renamed bn member variables to start with m_* as per coding convention 3. moved bn fusion test to test/cpu_fusion.cpp 4. style fix 5. added more checks to evaluate type and shape of inputs to bn * Added support for EMITDECL macro for batchnorm * - made correction to batchnorm src file name batchnorm -> batch_norm as per coding guidelines - corrected bn copy_with_new_args() method * Removed redundant SqrtOp support in serializer
-