1. 13 Jun, 2019 1 commit
  2. 12 Jun, 2019 7 commits
    • Scott Cyphers's avatar
      Merge branch 'master' into mlir · 4d24d157
      Scott Cyphers authored
      4d24d157
    • Dmitry Yershov's avatar
      IntelGPU backend: Switch to new clDNN: Fixed _GLIBCXX_USE_CXX11_ABI redifinition… · 073aedcd
      Dmitry Yershov authored
      IntelGPU backend: Switch to new clDNN: Fixed _GLIBCXX_USE_CXX11_ABI redifinition error during ngrap-bridge build (#3049)
      
      073aedcd
    • Sang Ik Lee's avatar
      Change behavior of elementwise divide for integral type to match Python. (#3034) · 745c4001
      Sang Ik Lee authored
      * Change behavior of elementwise divide for integral type to match Python.
      
      * Fix CPU codegen.
      
      * Temp fix: Disable failing UT for IntelGPU
      
      * Divide: Add constructor option to specify rounding mode for Integral types.
      
      * Update serializer to support legacy Divide dump.
      
      * Restore modified UT.
      745c4001
    • Rob Earhart's avatar
      Update PlaidML backend for current nGraph (#3030) · 397740fe
      Rob Earhart authored
      * Rename PlaidML_Executable::save -> save_as_format
      
      * Repair regression in PlaidML tensor impl
      
      This was caused by the recent removal of the offset parameter for tensor read/write operations -- we missed a
      spot where read/write were being called for synchronization purposes.
      
      * Disable a few more PlaidML tests pending triage
      
      * Skip elision of reshape->reshape
      
      It turns out this doesn't work, because the downstream reshape's input_order axis vector is incorrect if the
      upstream reshape is removed.
      
      * Add element type to PlaidML tensor debug output
      
      * Use nGraph booleans for PlaidML boolean data
      
      We'd previously been using i8; that's been deprecated for boolean data now that we have an explicit boolean
      element type.
      
      * Set PlaidML convolution output shapes correctly
      
      We weren't transposing the output shape; we were computing the right data, but the incorrect shape metadata
      causes validation to fail.
      
      * Add a PlaidML implicit broadcast op
      
      Better nGraph shape validation was tripping up PlaidML's use of a reshape to replace explicit broadcasts with
      implicit NumPy-style broadcasts (since the reshape's output shape would be incorrect for the downstream
      elementwise operation).  Adding this implicit broadcast operation lets PlaidML tell nGraph something useful
      about the shapes, making validation pass (when it's otherwise correct).
      397740fe
    • Scott Cyphers's avatar
      GenerateMask correction (#3029) · f0552cc8
      Scott Cyphers authored
      * GenerateMask correction
      Add an attribute that controls if the seed should be set on each use
      Convert to new virtual method for description implementatin
      
      * Support for switching to dynamic attributes.
      
      * GenerateMask changes in CPU backend (#3042)
      
      * Add CPU builder and kernel for new GenerateMask API
      
      * Remove dead code
      
      * Fix unit-test, PR feedback, file permissions
      
      * Disable new test for non-supporting backends
      
      * Fix CI error
      
      * Codegen support
      
      * Style check
      
      * Fix CI error
      f0552cc8
    • Michał Karzyński's avatar
      [ONNX] Add MatMulInteger op (#3011) · e51c5824
      Michał Karzyński authored
      * Unit tests for MatMulInteger
      
      * Add ONNX MatMulInteger op
      
      * Add QuantizedLinearMatmulInteger builder
      
      * Additional unit test
      
      * Exclude tests on nVidia GPU backend
      
      * Add 4D test case
      
      * Enable >2D MatMulInteger
      
      * Refactoring to MatMulFactory - step 1
      
      * Refactoring to MatMulFactory - step 2
      
      * Remove `using namespace ngraph` to make `Node` unambiguous.
      
      * Disable quantized ops tests on GPU backend
      
      * Remove unused `includes`
      
      * Remove redundant dynamic_pointer_cast
      
      * Remove redundant `move`
      
      * Add const correctness
      
      * Code review comments
      
      * Style apply
      
      * Add documentation
      
      * Use more complex shapes in tests
      e51c5824
    • Pruthvi's avatar
  3. 11 Jun, 2019 5 commits
  4. 10 Jun, 2019 5 commits
  5. 08 Jun, 2019 1 commit
  6. 07 Jun, 2019 3 commits
    • Pruthvi's avatar
      layout aware "copy_from" for cpu tensor (#2991) · 27bc5f39
      Pruthvi authored
      * layout-aware cpu tensor copy_from
      
      * - add unit test case for tensor "copy_from"
      - set default layout after copy_from src to destination
      
      * - added test case to copy between two different layouts
      
      * - fix clang error
       - added test case to check copy_from if the two tensor layouts are different
      - fix the LayoutDesc comparision
      
      * fix windows build error
      
      * - move "cop_from" unit tests from tensor to cpu_test
      
      * fix compilation error
      
      * Added more unit tests
      
      * fix read and write API in cpu_tensor_view
      27bc5f39
    • Adam Procter's avatar
    • Nishant Patel's avatar
      Use convolution_direct algo (#3028) · 3762538c
      Nishant Patel authored
      * Failing conv test case
      
      * Opt for mkldnn::algorithm::convolution_direct if input channel is less than 8
      
      * Comment
      3762538c
  7. 06 Jun, 2019 3 commits
  8. 05 Jun, 2019 9 commits
  9. 04 Jun, 2019 2 commits
  10. 03 Jun, 2019 4 commits