1. 19 Jun, 2019 7 commits
  2. 18 Jun, 2019 4 commits
    • Sang Ik Lee's avatar
    • Adam Rogowiec's avatar
      [Fused Op] LSTMCell (#2966) · 6b528fb8
      Adam Rogowiec authored
      * Move split utility functions into core builder.
      
      * Move activation functions to nGraph core.
      
      * RNN cell base class.
      
      * LSTM cell fused operator.
      
      * Update LSTM ONNX operator to use LSTMCell fused op.
      
      * Use Constant::create instead of make_constant.
      
      * Remove ngraph:: prefixes and include standard headers.
      
      * Store member shared_ptrs as object.
      
      * Formatting.
      
      * Run validation at the end of constructor.
      
      * Add more doc to ActivationFunction.
      
      * Run FusedOpDecomposition pass two times in interpreter backend.
      
      * Remove unnecesary class member.
      
      * Add node validation.
      
      * Disambiguate constructors.
      
      * Add type property test.
      
      * Formatting and add comment with equations.
      
      * Update IGPU backend with LSTMCell fused op.
      
      * Fix: clip activation function input.
      
      * Unit tests.
      
      * Workaround for nested fused op: run FusedOpDecomposition twice.
      
      * Fix compilation on CentOS and on GPU.
      
      * PR feedback.
      
      * Fix CentOS bugs.
      
      * Address review comments.
      
      Remove stored inputs as class members. Use node inputs directly in
      decomposition.
      
      * Fix errors.
      
      * Review feedback: don't use decompose_op while generating Function in UTs.
      
      * Fix merge artifacts.
      
      * Move RNNCellBase to op/util directory.
      
      * Fix typo for avg_pool setter method.
      
      * Set default values for optional inputs.
      
      * Fix typo in comment.
      6b528fb8
    • Tomasz Dołbniak's avatar
    • Adam Procter's avatar
      Fix corner case where op::Constant ctor is called with a zero-element shape and… · b28ed620
      Adam Procter authored
      Fix corner case where op::Constant ctor is called with a zero-element shape and a vector of one string (#3082)
      
      b28ed620
  3. 17 Jun, 2019 3 commits
  4. 16 Jun, 2019 1 commit
    • Adam Rogowiec's avatar
      [FUSED] Group Transpose Convolution (#3040) · e21db881
      Adam Rogowiec authored
      * Adding GroupConvTranspose fused operator.
      
      * Add missing header and remove commented code.
      
      * Remove unused variable.
      
      * Add a few more convieniece constructors.
      
      * Add more type prop UTs.
      
      * Remove unused post validation functions.
      
      * Style apply.
      
      * Fix conversion of vector to CoordinateDiff
      
      * Add GroupConvolutionTranspose to intel gpu backend.
      
      * Add documentation.
      
      * Use default (python-like) divide.
      e21db881
  5. 15 Jun, 2019 2 commits
  6. 14 Jun, 2019 15 commits
  7. 13 Jun, 2019 5 commits
    • Jayaram Bobba's avatar
      Change reduction operations to 2-input dynamic variants (#2972) · 9ba4a78a
      Jayaram Bobba authored
      * Change reduction operations to 2-input dynamic variants with convenience constructors for cases where reduction AxisSet is known at op construction time
      
      * Modify rest of arithmetic and logical reduction ops to 2-input dynamic variants. Some fixes to existing passes to keep constant reduction axes inputs intact
      
      * add new All tests to GPU manifest
      9ba4a78a
    • Michal Chruscinski's avatar
      [ONNX CI] Disable incremental building (#3059) · ff5d79ca
      Michal Chruscinski authored
      * Disable incremental building
      
      * Incremental building disablement as parameter
      ff5d79ca
    • gaurides's avatar
      Backward compatibility for GenerateMask deserialize (#3057) · 68e626b1
      gaurides authored
      * Backward comptability for GenerateMask nbench
      
      * PR feedback - use get_or_default
      68e626b1
    • Jayaram Bobba's avatar
      Added support for ceil mode in AvgPool (#3027) · a809ed7f
      Jayaram Bobba authored
      * Added support for ceil mode in AvgPool
      
      * Added ceil mode to MaxPool
      
      * remove extra semicolon
      
      * Add more constructor variants to support pybind which seems to have issues with multiple optional arguments
      
      * More constructor variants for AvgPool
      
      * More constructor variants for MaxPool
      
      * Style fix
      
      * Avoid constructor delegation
      
      * Revert "Avoid constructor delegation"
      
      This reverts commit 8efd59127bc9a16bae93b3c6b67dbcccfa95648f.
      a809ed7f
    • Michał Karzyński's avatar
      [ONNX] Add ConvInteger op (#3012) · fa300fae
      Michał Karzyński authored
      * Unit tests for ConvInteger
      
      * Add ONNX ConvInteger op
      
      * Add QuantizedConvInteger builder
      
      * Add unit tests
      
      * Exclude tests on nVidia GPU backend
      
      * Fix merge artifact
      
      * Add const-correctness and allow RVO
      fa300fae
  8. 12 Jun, 2019 3 commits
    • Dmitry Yershov's avatar
      IntelGPU backend: Switch to new clDNN: Fixed _GLIBCXX_USE_CXX11_ABI redifinition… · 073aedcd
      Dmitry Yershov authored
      IntelGPU backend: Switch to new clDNN: Fixed _GLIBCXX_USE_CXX11_ABI redifinition error during ngrap-bridge build (#3049)
      
      073aedcd
    • Sang Ik Lee's avatar
      Change behavior of elementwise divide for integral type to match Python. (#3034) · 745c4001
      Sang Ik Lee authored
      * Change behavior of elementwise divide for integral type to match Python.
      
      * Fix CPU codegen.
      
      * Temp fix: Disable failing UT for IntelGPU
      
      * Divide: Add constructor option to specify rounding mode for Integral types.
      
      * Update serializer to support legacy Divide dump.
      
      * Restore modified UT.
      745c4001
    • Rob Earhart's avatar
      Update PlaidML backend for current nGraph (#3030) · 397740fe
      Rob Earhart authored
      * Rename PlaidML_Executable::save -> save_as_format
      
      * Repair regression in PlaidML tensor impl
      
      This was caused by the recent removal of the offset parameter for tensor read/write operations -- we missed a
      spot where read/write were being called for synchronization purposes.
      
      * Disable a few more PlaidML tests pending triage
      
      * Skip elision of reshape->reshape
      
      It turns out this doesn't work, because the downstream reshape's input_order axis vector is incorrect if the
      upstream reshape is removed.
      
      * Add element type to PlaidML tensor debug output
      
      * Use nGraph booleans for PlaidML boolean data
      
      We'd previously been using i8; that's been deprecated for boolean data now that we have an explicit boolean
      element type.
      
      * Set PlaidML convolution output shapes correctly
      
      We weren't transposing the output shape; we were computing the right data, but the incorrect shape metadata
      causes validation to fail.
      
      * Add a PlaidML implicit broadcast op
      
      Better nGraph shape validation was tripping up PlaidML's use of a reshape to replace explicit broadcasts with
      implicit NumPy-style broadcasts (since the reshape's output shape would be incorrect for the downstream
      elementwise operation).  Adding this implicit broadcast operation lets PlaidML tell nGraph something useful
      about the shapes, making validation pass (when it's otherwise correct).
      397740fe