- 15 Oct, 2018 1 commit
-
-
tsocha authored
Add support for an environment variable `NGRAPH_ONNX_IMPORT_ENABLE` to setup.py This variable controls whether the Python API is build with or without support for ONNX import functionality built into core nGraph.
-
- 14 Oct, 2018 1 commit
-
-
gcwenger authored
* Improved AvgPool unit test coverage. Fixed small bug that was revealed. * Renamed disabled unit tests to reflect new names. * Ran clang-format on backend_test.in.cpp to fix format. * Renamed cpu_results->backend_results in two unit tests.
-
- 13 Oct, 2018 6 commits
-
-
mchrusci authored
* Added Jenkins Job No and URL * Update prepare_environment.sh - clone onnx models - remove wheel build artifacts * Link onnx models * Update Jenkinsfile - Only send email on fail, - Change ONNX branch warning to info field - Disable cloning after tox tests * Update prepare_environment.sh - Added comments - link ONNX models directory * Update cleanup
-
Robert Kimball authored
-
Nick Korovaiko authored
* dynamic to static casts * ScalarConstantLikeBase needs a dynamic cast * revert change
-
Robert Kimball authored
-
Nick Korovaiko authored
* unary, binary folding * fix divide wrong template args * add tests * fix merge breaks
-
gcwenger authored
-
- 12 Oct, 2018 10 commits
-
-
Robert Kimball authored
* Why am I still needing to fix license headers? * fix a few more in test
-
Jayaram Bobba authored
* Handle zero sized buffer requests * Removed aligned buffer initialize() and moved allocation to constructor instead
-
Artur Wojcik authored
* onnx: add register operator macro Signed-off-by: Artur Wojcik <artur.wojcik@intel.com> * onnx: add set information to 'op' namespace Signed-off-by: Artur Wojcik <artur.wojcik@intel.com>
-
L.S. Cook authored
* initial rough draft * import changes md to release notes for readability * import changes md to release notes for readability * remove draft * update README with link to Release Notes * release notes link * fix graphic placement * Wording change on ONNX supported frameworks
-
Ayan Moitra authored
* return nullptr when workspace size is zero+modify insert method to accept const lvalue ref * Unit test added
-
Robert Kimball authored
* move backend specific test files to test/backend directory * remove unit_test_control * move tests back to test root * fix comment * wip * fix manifest
-
Robert Kimball authored
* update test to verify all header files are complete, meaning they include what they use. * disable
-
Ayan Moitra authored
* Project initialization commit * Added unit tests for 3D tensors for argmax * Refactored reduce to be used by argmax argmin. argmax argmin still has some issues. WIP * [WIP]First working version of ArgMax ArgMin * added reduce buffer for the cudnn api calls * added reduce buffer for the cudnn api calls * Further modifications. Using rvalues to pass enums to build reduce method * more unit tests added * Incorporate Fenglei's comments * Incorporating Chris's first set of comments * small change to test file * Resolving clang issue that was causing argmin test to fail * Incorporate Chris's comments * clang format issue
-
Artur Wojcik authored
Signed-off-by: Artur Wojcik <artur.wojcik@intel.com>
-
Amy Zhuang authored
-
- 11 Oct, 2018 3 commits
-
-
Nick Korovaiko authored
-
Nick Korovaiko authored
-
Robert Kimball authored
* updated unit tests * remove debug comments
-
- 10 Oct, 2018 3 commits
-
-
Fenglei authored
-
Fenglei authored
* update onehot * clang * fix bugs * format * add output_datatype_size to hash * typo * hash
-
Nick Korovaiko authored
* reshape sinking working on mnist_conv * forgot to add reshape_sinking files * refactoring of binary case * Quantize/Dequantize case, fix add case, add assert * address bob and scott's feedback * debug * fix a bug where reshapes are removed too early
-
- 09 Oct, 2018 4 commits
-
-
Adam Procter authored
-
Robert Kimball authored
-
Chris Sullivan authored
* add find algorithm for convolution without extra padding * Use cudnnFind* or cudnnGet* depending on tuning param boolean. Add select function to search the perf results of the cudnn queries. * Formatting. * Algo search no longer binary, now it is either off, a heuristic search (cudnnGet*) or an explicit search (cudnnFind*). * Formatting. * switch to explicit. * Throw if no suitable cudnn algo found. * Formatting * Remove comment.
-
Jayaram Bobba authored
* Added a bounds check for mkldnn layout descriptor creation * Added dims check
-
- 08 Oct, 2018 4 commits
-
-
Robert Kimball authored
-
Chris Sullivan authored
* Add pad with fill operator using the outward-in index pattern. * Remove static pad and rename build_pad_dynamic -> build_pad. Update maxpool 1d padding. * Formatting. * Split build_pad_dynamic into build_pad and build_pad_fill. * Add test coverage for fixed bug in op::Pad for gpu.
-
Jayaram Bobba authored
* Reshape optimizations for when unit-sized dimensions are added/removed from tensors * Added unit tests for eliminating squeeze and expand_dims operations * Bug fix to expand dims layout * Style fix
-
Jayaram Bobba authored
* Check output shape when setting memory layout for slice op. * Miscellaneous fusion and other optimizations for inception-resnetv2 - ConvBias Batchnorm folding - ConvBias Affine folding - Check if MKLDNN can slice a given layout and select layouts appropriately * Fixed unit test and bug in conv bias pattern * Addressed PR feedback * Addressed PR feedback
-
- 06 Oct, 2018 2 commits
-
-
gcwenger authored
* Eliminated two warnings introduced in #1459 * Removed unnecessary call to reserve_workspace.
-
VINOD KUMAR DEVARAMPATI authored
* added constant folding for dequantize * modified as per review comments
-
- 05 Oct, 2018 6 commits
-
-
gcwenger authored
* LRN WIP * Explicit lambda captures. * Switched to Ayan's new caching routine. * Remove commented out lrn from manifest. * Fixed clang 3.9 error. * Corrected lrn hash. Only call cudnnSetLRNDescriptor once. * Simplified lrn hash. Removed redundant parameters. No longer passing CUDNN_LRN_CROSS_CHANNEL_DIM1 as parameter because it's the only choice for cudnnLRNCrossChannelForward.
-
Jaikrishnan Menon authored
-
Scott Cyphers authored
* More op doc, fix formatting * sqrt, tan * Formatting.
-
Robert Kimball authored
-
Robert Kimball authored
* address klocwork issue * move class init * more klocwork * more klocwork * more klocwork * comment on where the magic number is from * address review comments * address review comments
-
Chris Sullivan authored
* Add op::Sigmoid to nvgpu. * Bring rnn fusion and concat passes over into GPU from IA. This is a temporary move until generalization and gpu specification can occur. * Add LSTM fusion and cudnn inference kernel. Next need recurrent fusion and layer fusion. * Formatting * Removed unecessary extra output from LSTM op (rnn with seq. length = 1, so y = hy). * Add RNN fusion of LSTM cells within a recurrent layer. * Formatting. * Add fusion across RNN layers. * Formatting. * Add algebraic simplification. * Added rnn fusion tests. * Updated conditional on LSTM fusion to better distinguish bound nodes as ht vs xt. * Formatting. * Removed print statements. * Formatting. * Committing missing file. * Remove concat inputs pass and mkldnn references. * fix cmake paths * conflict resolution with merge from master. * remove explicit lstm op support. bare LSTM ops are converted to RNN ops for emission. * Formatting. * Use NGRAPH_ASSERT. Formatting of intel copyright. * Add check on the feature size (shape) of the recurrent (hidden) input and cell state, to ensure they are the same size. * fix wrong rnn header * Formatting. * Add back lstm op to dispatch table. * Added RNN test which shows cudnn rnn kernel is not producing correct results. * With update to AlgSimpl. to simplify concat-reshape-slice, the check modifed in this commit needed to be relaxed. * Bug fix in parameter tensor packing. * Alias third output element of RNN for cell state (bug fix). * Resolve numerical correctness issue with negative values in RNN (bug fix). Add minimal test to evaluate LSTM and compare with values calculated by hand. * Add tensor parameter sizes to kernel hash as they are kernel-specific. * Add 2 layer lstm fusion test against by-hand solution. * Export param concatenation to graph for cudnn kernel at both the single rnn layer and multi-layer. * Formatting. * Finishing touches after merge: add support for macro expansed dispatch via op_tbl. * Simplify macro support for gpu ops. * Add CUDNN_VERSION >= 7200 defguards for RNN fusion. Need to decide how to notify user of increased performance with >= 7200. * Revert lstm_analytic test to explicitly copy data to tensor params. * Removed namespace arg from NGRAPH_GPU_OP. * Refactored macros to different header so op_tbl only contains op list. * Defguard on cudnn_descriptor<cudnnRNNDataDescriptor_t>. * doubles -> floats * Reorg. pass asserts, prepare to replace with non-throwing pass failures. * Remove Lstm op and replace it with Rnn. * Format * Utilize RETURN_IF_FALSE in rnn pass to avoid any RT asserts. Note that falling back to raw (no passes) graph for 2rnn_3lstm json from mxnet models results in a double free inside of the memory layout pass. Appears to be a bug in Reshape pass through. * Removed print statements. Add check on input data and recurrent data. * Don't reuse memory for non-destructive ops. * Add back Rnn test. * Formatting. * Clean up comments. * Update test per review comments.
-