- 11 Dec, 2018 3 commits
-
-
gaurides authored
* fix crash in ngraph-tf test conv_ops_test.Conv2DTest.testConv2DKernelSmallerThanStrideSame * fix file perms * correct checks
-
Amy Zhuang authored
* Bug fixes to unordered map checks * No in-place slice for non-native MKLDNN layouts * is_op
-
Nick Korovaiko authored
-
- 10 Dec, 2018 1 commit
-
-
harryskim authored
* Removed winml from stack diagram * Removed winml from full stack diagram * Update README.md * update the diagram without winml * Changed sentence about WinML * Removed duplication
-
- 08 Dec, 2018 4 commits
-
-
Robert Kimball authored
* change all_close tests to return gtest AssertionResult instead of bool to allow for better error messages * change throw to return error * address PR comments and fix compile error
-
Jayaram Bobba authored
* reenable mkldnn convolution for large padding * specify precision tolerance to unit test * pass tolerance values to all_close
-
Robert Kimball authored
* move GPU specific test to GPU only * fix unit test invocation * fix compile error * fix compile error * style * fix runtime error
-
Nick Korovaiko authored
* make GOE extend from util::Op * fix build breaks
-
- 07 Dec, 2018 6 commits
-
-
Jayaram Bobba authored
* initial commit for update slice op * Finished up update_slice fusion and added codegen support * style fixes * Added unit test for in-place update-slice strided * change pattern name
-
Robert Kimball authored
* change compile call to return Handle * make CPU require compile() before call() * fix unit tests to call compile() before call() * fix failing ops * update unit test * revert some changes * more fixups * more diff cleanup * a few more issues addressed * more fixes * update API * more updates * fix test_ops.py * fix * another attempt to fix * fix unit test * fix test error
-
Sergey Shalnov authored
-
gcwenger authored
* Double support for all_close_f * all_close_f uses fixed number of mantissa bits now. Simplified testing code. * Initialize test data members in constructor to values which will cause test failure. Setup then sets them correctly. * Reduce info printed out during all_close_f unit tests.
-
Sang Ik Lee authored
-
Adam Straw authored
* re-enable quantize_clamp_int32 test on CPU * MLKDNN typo
-
- 06 Dec, 2018 14 commits
-
-
Nishant Patel authored
* Quantize the bias to int32 * Bias scale fix * mnist works * Quantize Bias * Introduce Quantize op in the graph to quantize bias & feedback * Add QuantizedConvBiasAdd * Comments and some refactoring * Add test case with float bias and enable int32 as quantized type in ngraph * Change shape of scale from Shape{} to Shape{1} in the backend * Add QuantizedConvBiasSignedAdd * Fix Layouts, clean up and a test case for QCBA * Test case for QCBSA * cleanup mkldnn_emitter.hpp * fix build error * Constant fold
-
Sergey Shalnov authored
-
Nick Korovaiko authored
* one output passing tests clean up fix build breaks * move generators into a separate file
-
Nick Korovaiko authored
-
Nick Korovaiko authored
* an env var to disable individual fusions * fix env var name
-
Nick Korovaiko authored
* give fusions names * fix build breaks * fix perms
-
Nick Korovaiko authored
* abort messages in matcher.cpp * style fixes
-
gcwenger authored
* Isolated per op testing when comparing graphs for better determination of source of accuracy divergence. * Improve clarity of comment
-
Michał Karzyński authored
* Update README for PyPI * Update README for PyPI * Remove redundant newlines * Fix links
-
Adam Rogowiec authored
* Uniform quotes style . * Fix comment style. * Check setup.py with flake8. - Fix flake8 errors. * Move function out of class scope. * Fix function paramter list * Fix formatting.
-
Fenglei authored
* add some helper function * update with new helper function * update reduce to nd with new helper function * update float sum to stable sum * fix bug * update all reduce to stable sum for float * fix bug and pass the sum stable test * remove debug info * style * update with shape * fix bug * add host parameters to cuda_emitter * clang format * fix bugs * add element::type support * format * add a cached value with datatype name * add init_reduce_value * unroll loop * optimization * remove the need for init_value * add memset kernel * add memcpy * working version * remove debug info * add comments, clean up code. * change in_idx to input_idx * fix bug * change args name for memset in emitter * pass element::Type instead of string * the op::reduce come with init value, add support * resolve codacy-bot comment * fix bug * resove codacy-bot comment * remove unused comments, resolve comments * cuda reduce for max, min, mul, reduce op init value, format * use type::info * use type info for numeric_limits * remove code from gpu_host_parameters * header * remvoe outdated comments * add helper to check if stable sum is needed * add stable sum test for double * remove extra line * consolidate helper functions * no need list now. * remove extra ; * clang format * style * add skip test for cpu and intelGPU side * add line between groups of headers * add two simple stable sum test for float and double * skip test for intelGPU
-
Fabian Boemer authored
-
Pruthvi authored
* - Added reorder support for rnn weights_layer/iter * i) fixed compilation issues ii) working but still observing precision error * i) fixed failing rnn unit test for DEX ii) refactored workspace in RNN mkldnn emitter * i) added support for src reorder to TNC from NTC * reorder support for rnn output fron NTC to TNC * - added support for rnn weight reorder ldgoi -> ldigo - code refactor for lstm/rnn kernel in mkldnn emitter * - refactor rnn mkldnnn kernel, change variable names * fix RNN codegen kernel * disbale layer rnn fusion pass, to test CI * method to validate recurrent rnn inputs * add correlated macthes for Recurrent RNN PM * - simplify reorder logic for rnn_weights - fix graph pattern for fusing rnn cell across time steps * do weights reorders in rnn timesteps fusion * refactored LSTM graph pass * - Bug fix for finding the lstm inputs determenstically - Refactored LSTM graph pass to single pass - made changes to LSTM RNN time step fusion graph pass * - use replace_node instead of replace_output in Lstm_step_wise fusion graph pass * fix compilation error * Fix GNMT rnn fusion * check if the node is in use before replacing in RNN graph passes * i) fix style ii) fix topo sort issue in RNN graph pass * style fix * fix bug in simplify_concat pass * replaces Lstm1 -> {GOE1, GOE2} -> {Slice1, Slice2} -> Concat -> Lstm2 with Lstm1 -> Lstm2 * cse for convert layout * addressed PR comments * - optimization pass to remove Lstm1 -> {GOE1, GOE2} -> {Slice1, Slice2} -> Lstm2 - conditional fusing of LSTM cells only for the decoder * made changes to multi layer RNN fusion callback * fix asserts in RNN op * - added support to fuse layers when slc=dlc for RNN cells - bug fix on the sanity checks for RNN Op * - support RNN layer fusion till slc = dlc - bug fixes in multi layer rnn fusion call back * capture reshape in the RNN weights * Addressed PR comments * - added comments in multi layer PM call back - fuse only if slc == DLC across layers * restore deleted 3_lstm_cell_forward.json file * fix typo * fix failing unit tets * When processing in place slice, do not change the offset of the slice node if the argument pointer comes from function input. * Address PR feedback: process in place slice after propagating in place input. * Set INTERMEDIATE role before propagating in place input. * Do not add temporaries to the variable name map before propagating in place input in codegen. * Fix a bug in codegen. * Fix a bug in codegen slice. * reenable disabled rnn unit test * fix compiler error * - bug fix in the slicing logic for the layer fused rnn cell - fix failing rnn unit test * - Addressed PR comments - removed redundant checks from the rnn graph pass - simplified rnn call back replace node logic * - added new multilayer rnn *.json file - fix test case * [PRIVATE BRANCH] Style fixes (#2080) * Style fixes * change order of lstm gates * [PRIVATE BRANCH] Jbobba/rnn fusion review (#2113) * Style fixes for single-layer RNN fusion * Style fixes to multi-layer RNN * style fix * disable GPU test
-
Pruthvi authored
* fix fialing bn test * fix style
-
- 05 Dec, 2018 9 commits
-
-
Jayaram Bobba authored
* fix expand layout for padded dimensions * enable squeeze padded layouts
-
Nick Korovaiko authored
-
Pruthvi authored
-
Pruthvi authored
* - modified cpu_assignment pass to support bn with input 5D - added test cases for 5D bn and 5D bn+relu * - Address PR comments - used mkldnn_utils to validate bn for mkldnn * fix compilation error * Addressed PR comments - added helpers in mkldnn_utils for assigning ngraph Op as MKLDNN op - helper funnction for bn mkldnn assignment * fix clang error
-
Amy Zhuang authored
-
Robert Kimball authored
Bob/ext2
-
Chris Sullivan authored
-
Robert Kimball authored
This reverts commit 1c4aa225.
-
Robert Kimball authored
Fix the incorrect way to query the size of Constant tensor, which lead invalid memory access
-
- 04 Dec, 2018 3 commits
-
-
Scott Cyphers authored
-
Sergey Shalnov authored
* IntelGPU backend: Use clDNN matrix operations for nGraph::Dot * Update unit_test.manifest
-
Scott Cyphers authored
-