- 19 Jun, 2019 7 commits
-
-
Tomasz Dołbniak authored
* Fix of the failing tests: test_unary_op_array & test_unary_op_scalar * Parametrize the logical_not test with input_data * Code formatting * Whitespace removal because flake8
-
Denise Kutnick authored
* make check_inputs check for input_count >= expected_input_count * scrub unit test manifest after check_inputs change
-
Daiki AMINAKA authored
-
Amy Zhuang authored
* Enable Gather and ScatterAdd to use Eigen kernel for int8 type. * Reduce number of supported ranks. * Fix a bug.
-
Scott Cyphers authored
* Serialize nodes by reference * Most od deserialization * deserialize node * review comments
-
Jayaram Bobba authored
-
Robert Kimball authored
This gets all compile-time flag use into cpp files rather than headers. This does not work if it is in a header. (#3096)
-
- 18 Jun, 2019 4 commits
-
-
-
Adam Rogowiec authored
* Move split utility functions into core builder. * Move activation functions to nGraph core. * RNN cell base class. * LSTM cell fused operator. * Update LSTM ONNX operator to use LSTMCell fused op. * Use Constant::create instead of make_constant. * Remove ngraph:: prefixes and include standard headers. * Store member shared_ptrs as object. * Formatting. * Run validation at the end of constructor. * Add more doc to ActivationFunction. * Run FusedOpDecomposition pass two times in interpreter backend. * Remove unnecesary class member. * Add node validation. * Disambiguate constructors. * Add type property test. * Formatting and add comment with equations. * Update IGPU backend with LSTMCell fused op. * Fix: clip activation function input. * Unit tests. * Workaround for nested fused op: run FusedOpDecomposition twice. * Fix compilation on CentOS and on GPU. * PR feedback. * Fix CentOS bugs. * Address review comments. Remove stored inputs as class members. Use node inputs directly in decomposition. * Fix errors. * Review feedback: don't use decompose_op while generating Function in UTs. * Fix merge artifacts. * Move RNNCellBase to op/util directory. * Fix typo for avg_pool setter method. * Set default values for optional inputs. * Fix typo in comment.
-
Tomasz Dołbniak authored
-
Adam Procter authored
Fix corner case where op::Constant ctor is called with a zero-element shape and a vector of one string (#3082)
-
- 17 Jun, 2019 3 commits
-
-
Scott Cyphers authored
-
Robert Kimball authored
-
Michał Karzyński authored
-
- 16 Jun, 2019 1 commit
-
-
Adam Rogowiec authored
* Adding GroupConvTranspose fused operator. * Add missing header and remove commented code. * Remove unused variable. * Add a few more convieniece constructors. * Add more type prop UTs. * Remove unused post validation functions. * Style apply. * Fix conversion of vector to CoordinateDiff * Add GroupConvolutionTranspose to intel gpu backend. * Add documentation. * Use default (python-like) divide.
-
- 15 Jun, 2019 2 commits
-
-
Denise Kutnick authored
-
Rob Earhart authored
-
- 14 Jun, 2019 15 commits
-
-
gaurides authored
-
Jayaram Bobba authored
* Added missing attribute to Result serialization * Fix default layout attribute * use helper routine for optional attrs
-
Tomasz Dołbniak authored
* Initial implementation of the Shrink op * Multiply the values by the correct masks * Basic test case for Shrink with floats * Shrink test on integers * Code formatting * Shrink documentation and typo fix * Rephrase the Shrink docs * Out of <memory> ;)
-
Adam Procter authored
* Add DynElimination for Broadcast * Change silent bailouts for invalid shape/ETs to NGRAPH_CHECKs
-
Amy Zhuang authored
-
Tomasz Dołbniak authored
* Correct the dockerfile name and the way to pass multiple env vars to a container * Docs fix * Typo fix
-
Adam Procter authored
* Add execution tests for dynamic reduction ops; fix validation logic * Add dynamic_GPU.all to manifest * Be explicit about the dynamic_GPU prefix in manifest
-
Dmitry Yershov authored
-
gaurides authored
* Initial implementation * Added test case * Bug fix; Dropout with 2 outputs, WIP * Fixed in unit-testl; WIP for model * Nothing is working * Revert "Nothing is working" This reverts commit d3ff09bb7a0d0519ab70ac85f2e7f30721afea96. * Fixed unit-test; fusion with 2 outputs * Fix style check, file permissions * Changed input arg to Node * Fix order of declaration * Improved performance * some cleanup * Fixed CI error * Fixed review comments * Fix CI error * Remove unused variable * Fix other CI errors * Changed type * Fix style check * Add codegen code for Dropout * addressed PR feedback; will add codegen support later * Cleanup; change variable name * Support for use_seed * Add setter for use_seed * Add setter for use_seed * Fix CI error * Make use_seed as arg * Fix CI error * Fix CI error
-
Scott Cyphers authored
-
Scott Cyphers authored
-
Adam Procter authored
* Simple DynElimination test (not passing yet) * Implement DynElimination for DynSlice; simple test passing, but more needed * Add test generator for DynSlice * Add more tests (one not passing) * Rename update_reference.sh to update_convolution_reference.sh, to (hopefully) reduce confusion * Comment edits * Fix a couple more bugs, add a bunch of unit tests * A few more tests for the negative-stride slicing issue that's worrying me. * Refactor dyn slice test to improve compile time * Update comment * Further test refactoring: generate separate tests rather than one big one * More element type coverage because hey why not * Add more tests, per review comments
-
Daiki AMINAKA authored
-
Robert Kimball authored
* Copy friendly name when copying node * add unit test * style
-
Robert Kimball authored
* wip * hybrid as a static backend and not part of ngraph * only for linux * fix link problem * style * remove hybrid * fix compile error
-
- 13 Jun, 2019 5 commits
-
-
Jayaram Bobba authored
* Change reduction operations to 2-input dynamic variants with convenience constructors for cases where reduction AxisSet is known at op construction time * Modify rest of arithmetic and logical reduction ops to 2-input dynamic variants. Some fixes to existing passes to keep constant reduction axes inputs intact * add new All tests to GPU manifest
-
Michal Chruscinski authored
* Disable incremental building * Incremental building disablement as parameter
-
gaurides authored
* Backward comptability for GenerateMask nbench * PR feedback - use get_or_default
-
Jayaram Bobba authored
* Added support for ceil mode in AvgPool * Added ceil mode to MaxPool * remove extra semicolon * Add more constructor variants to support pybind which seems to have issues with multiple optional arguments * More constructor variants for AvgPool * More constructor variants for MaxPool * Style fix * Avoid constructor delegation * Revert "Avoid constructor delegation" This reverts commit 8efd59127bc9a16bae93b3c6b67dbcccfa95648f.
-
Michał Karzyński authored
* Unit tests for ConvInteger * Add ONNX ConvInteger op * Add QuantizedConvInteger builder * Add unit tests * Exclude tests on nVidia GPU backend * Fix merge artifact * Add const-correctness and allow RVO
-
- 12 Jun, 2019 3 commits
-
-
Dmitry Yershov authored
IntelGPU backend: Switch to new clDNN: Fixed _GLIBCXX_USE_CXX11_ABI redifinition error during ngrap-bridge build (#3049)
-
Sang Ik Lee authored
* Change behavior of elementwise divide for integral type to match Python. * Fix CPU codegen. * Temp fix: Disable failing UT for IntelGPU * Divide: Add constructor option to specify rounding mode for Integral types. * Update serializer to support legacy Divide dump. * Restore modified UT.
-
Rob Earhart authored
* Rename PlaidML_Executable::save -> save_as_format * Repair regression in PlaidML tensor impl This was caused by the recent removal of the offset parameter for tensor read/write operations -- we missed a spot where read/write were being called for synchronization purposes. * Disable a few more PlaidML tests pending triage * Skip elision of reshape->reshape It turns out this doesn't work, because the downstream reshape's input_order axis vector is incorrect if the upstream reshape is removed. * Add element type to PlaidML tensor debug output * Use nGraph booleans for PlaidML boolean data We'd previously been using i8; that's been deprecated for boolean data now that we have an explicit boolean element type. * Set PlaidML convolution output shapes correctly We weren't transposing the output shape; we were computing the right data, but the incorrect shape metadata causes validation to fail. * Add a PlaidML implicit broadcast op Better nGraph shape validation was tripping up PlaidML's use of a reshape to replace explicit broadcasts with implicit NumPy-style broadcasts (since the reshape's output shape would be incorrect for the downstream elementwise operation). Adding this implicit broadcast operation lets PlaidML tell nGraph something useful about the shapes, making validation pass (when it's otherwise correct).
-