Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in / Register
Toggle navigation
N
ngraph
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Packages
Packages
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
submodule
ngraph
Commits
ee5d46c1
Unverified
Commit
ee5d46c1
authored
Jul 02, 2019
by
Scott Cyphers
Committed by
GitHub
Jul 02, 2019
Browse files
Options
Browse Files
Download
Plain Diff
Merge branch 'master' into louisfeng/NGCORE-456_fusion-type
parents
7a2f9603
7c16ee6c
Hide whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
141 additions
and
75 deletions
+141
-75
external_mlir.cmake
cmake/external_mlir.cmake
+2
-2
CMakeLists.txt
src/contrib/mlir/CMakeLists.txt
+1
-0
compiler.cpp
src/contrib/mlir/compiler.cpp
+13
-7
compiler.hpp
src/contrib/mlir/compiler.hpp
+1
-1
dialect.cpp
src/contrib/mlir/dialect/dialect.cpp
+3
-3
dialect.hpp
src/contrib/mlir/dialect/dialect.hpp
+4
-2
lowerer.cpp
src/contrib/mlir/lowerer.cpp
+69
-33
lowerer.hpp
src/contrib/mlir/lowerer.hpp
+2
-0
op_lowerers.inc
src/contrib/mlir/op_lowerers.inc
+13
-11
cpu_external_function.cpp
src/ngraph/runtime/cpu/cpu_external_function.cpp
+3
-0
gpu_compiled_function.cpp
src/ngraph/runtime/gpu/gpu_compiled_function.cpp
+0
-3
intelgpu_backend.cpp
src/ngraph/runtime/intelgpu/intelgpu_backend.cpp
+0
-4
int_executable.cpp
src/ngraph/runtime/interpreter/int_executable.cpp
+0
-3
plaidml_compiler.cpp
src/ngraph/runtime/plaidml/plaidml_compiler.cpp
+2
-0
plaidml_config.cpp
src/ngraph/runtime/plaidml/plaidml_config.cpp
+2
-0
unit_test.manifest
src/ngraph/runtime/plaidml/unit_test.manifest
+26
-6
No files found.
cmake/external_mlir.cmake
View file @
ee5d46c1
...
...
@@ -20,8 +20,8 @@ set(MLIR_LLVM_REPO_URL https://github.com/llvm/llvm-project.git)
set
(
MLIR_REPO_URL https://github.com/tensorflow/mlir.git
)
# Change these commit IDs to move to latest stable versions
set
(
MLIR_LLVM_COMMIT_ID
bb2b527
)
set
(
MLIR_COMMIT_ID
49f7efc
)
set
(
MLIR_LLVM_COMMIT_ID
c0cad98
)
set
(
MLIR_COMMIT_ID
82d5084
)
set
(
MLIR_PROJECT_ROOT
${
CMAKE_CURRENT_BINARY_DIR
}
/mlir_project
)
set
(
MLIR_LLVM_ROOT
${
MLIR_PROJECT_ROOT
}
/llvm-projects
)
set
(
MLIR_SOURCE_DIR
${
MLIR_LLVM_ROOT
}
/llvm/projects/mlir
)
...
...
src/contrib/mlir/CMakeLists.txt
View file @
ee5d46c1
...
...
@@ -56,6 +56,7 @@ if (NGRAPH_MLIR_ENABLE)
MLIRExecutionEngine
MLIRIR
MLIRLLVMIR
MLIRStandardToLLVM
MLIRParser
MLIRPass
MLIRTargetLLVMIR
...
...
src/contrib/mlir/compiler.cpp
View file @
ee5d46c1
...
...
@@ -34,11 +34,12 @@
#include <llvm/Support/MemoryBuffer.h>
#include <llvm/Support/SourceMgr.h>
#include <llvm/Support/TargetSelect.h>
#include <mlir/Conversion/StandardToLLVM/ConvertStandardToLLVM.h>
#include <mlir/Conversion/StandardToLLVM/ConvertStandardToLLVMPass.h>
#include <mlir/ExecutionEngine/ExecutionEngine.h>
#include <mlir/ExecutionEngine/MemRefUtils.h>
#include <mlir/ExecutionEngine/OptUtils.h>
#include <mlir/LLVMIR/LLVMDialect.h>
#include <mlir/LLVMIR/Transforms.h>
#include <mlir/Pass/PassManager.h>
#include <mlir/Target/LLVMIR.h>
#include <mlir/Transforms/DialectConversion.h>
...
...
@@ -50,6 +51,7 @@
using
llvm
::
SmallVector
;
using
llvm
::
StringRef
;
using
llvm
::
make_unique
;
using
namespace
ngraph
::
runtime
::
ngmlir
;
#define COMPILE_OP_DECL(op_name) \
...
...
@@ -75,7 +77,7 @@ void MLIRCompiler::init_mlir()
if
(
!
initialized
)
{
mlir
::
registerDialect
<
mlir
::
NGDialect
>
();
mlir
::
registerDialect
<
mlir
::
NG
raphOps
Dialect
>
();
// Register any LLVM command line options
llvm
::
cl
::
ParseEnvironmentOptions
(
"ngraph"
,
"MLIR_LLVM_OPTIONS"
,
""
);
initialized
=
true
;
...
...
@@ -133,7 +135,7 @@ void MLIRCompiler::build_ng_dialect_module()
}
// create builder
m_builder
=
llvm
::
make_unique
<
mlir
::
FuncBuilder
>
(
function
.
get
());
m_builder
=
llvm
::
make_unique
<
mlir
::
OpBuilder
>
(
function
->
getBody
());
build_ng_dialect
();
m_module
->
getFunctions
().
push_back
(
function
.
release
());
if
(
failed
(
m_module
->
verify
()))
...
...
@@ -359,10 +361,14 @@ void MLIRCompiler::execute()
NGRAPH_CHECK
(
m_module
,
"MLIR module is not ready."
);
// Lower Standard dialect to LLVM dialect.
auto
converter
=
mlir
::
createStdToLLVMConverter
();
auto
r
=
converter
->
convert
(
m_module
.
get
());
(
void
)
r
;
NGRAPH_CHECK
(
succeeded
(
r
),
"second conversion failed"
);
mlir
::
LLVMTypeConverter
llvm_converter
(
&
m_context
);
OwningRewritePatternList
patterns
;
mlir
::
populateStdToLLVMConversionPatterns
(
llvm_converter
,
patterns
);
mlir
::
ConversionTarget
target
(
m_context
);
target
.
addLegalDialect
<
mlir
::
LLVM
::
LLVMDialect
>
();
auto
result
=
applyConversionPatterns
(
*
m_module
,
target
,
llvm_converter
,
std
::
move
(
patterns
));
NGRAPH_CHECK
(
succeeded
(
result
),
"Standard to LLVM dialect conversion failed"
);
dump_mlir_module
(
"LLVM-IR Dialect Dump:"
);
...
...
src/contrib/mlir/compiler.hpp
View file @
ee5d46c1
...
...
@@ -132,7 +132,7 @@ namespace ngraph
mlir
::
MLIRContext
m_context
;
std
::
unique_ptr
<
mlir
::
Module
>
m_module
;
std
::
unique_ptr
<
mlir
::
Func
Builder
>
m_builder
;
std
::
unique_ptr
<
mlir
::
Op
Builder
>
m_builder
;
std
::
unique_ptr
<
mlir
::
ExecutionEngine
>
m_engine
;
using
TensorToInfo
=
std
::
pair
<
descriptor
::
Tensor
*
,
TensorInfo
>
;
...
...
src/contrib/mlir/dialect/dialect.cpp
View file @
ee5d46c1
...
...
@@ -21,8 +21,8 @@
using
namespace
mlir
;
NG
Dialect
::
NG
Dialect
(
mlir
::
MLIRContext
*
ctx
)
:
mlir
::
Dialect
(
"ng"
,
ctx
)
NG
raphOpsDialect
::
NGraphOps
Dialect
(
mlir
::
MLIRContext
*
ctx
)
:
mlir
::
Dialect
(
getDialectNamespace
()
,
ctx
)
{
addTypes
<
NGTensorType
>
();
addTypes
<
NGIntegerType
>
();
...
...
@@ -34,7 +34,7 @@ NGDialect::NGDialect(mlir::MLIRContext* ctx)
>
();
}
void
NGDialect
::
printType
(
mlir
::
Type
type
,
raw_ostream
&
os
)
const
void
NG
raphOps
Dialect
::
printType
(
mlir
::
Type
type
,
raw_ostream
&
os
)
const
{
switch
(
type
.
getKind
())
{
...
...
src/contrib/mlir/dialect/dialect.hpp
View file @
ee5d46c1
...
...
@@ -25,15 +25,17 @@
#include "ngraph/check.hpp"
namespace
mlir
{
class
NGDialect
:
public
mlir
::
Dialect
class
NG
raphOps
Dialect
:
public
mlir
::
Dialect
{
public
:
explicit
NGDialect
(
mlir
::
MLIRContext
*
ctx
);
explicit
NG
raphOps
Dialect
(
mlir
::
MLIRContext
*
ctx
);
mlir
::
Type
parseType
(
llvm
::
StringRef
tyData
,
mlir
::
Location
loc
)
const
override
{
NGRAPH_CHECK
(
false
,
"Unsupported type parsing."
);
return
mlir
::
Type
();
}
void
printType
(
mlir
::
Type
type
,
llvm
::
raw_ostream
&
os
)
const
override
;
static
StringRef
getDialectNamespace
()
{
return
"ng"
;
}
};
}
src/contrib/mlir/lowerer.cpp
View file @
ee5d46c1
...
...
@@ -41,31 +41,34 @@ namespace
class
DialectLoweringPass
;
/// Base class for nGraph operation conversions to affine/standard dialect. Provides
/// conversion patterns with an access to the DialectLoweringPass which holds the state of the
/// conversion.
class
NGraphOpLowering
:
public
ConversionPattern
{
public
:
NGraphOpLowering
(
StringRef
rootOpName
,
MLIRContext
*
context
,
DialectLoweringPass
&
pass
)
:
ConversionPattern
(
rootOpName
,
/*benefit=*/
1
,
context
)
,
m_pass
(
pass
){};
protected
:
// Back-reference to the lowering pass which contains the lowering state, including the
// nGraph type converter.
DialectLoweringPass
&
m_pass
;
};
#include "op_lowerers.inc"
///
Use Dialect Converson Framework
class
DialectLowerer
:
public
DialectConversion
///
Conversion from types in the nGraph dialect to the Standard dialect.
class
NGraphTypeConverter
:
public
TypeConverter
{
public
:
DialectLowerer
(
DialectLoweringPass
&
pass
)
:
DialectConversion
()
,
m_pass
(
pass
)
NGraphTypeConverter
()
:
TypeConverter
()
{
}
Type
convertType
(
Type
t
)
override
;
protected
:
// Initialize the list of converters.
void
initConverters
(
OwningRewritePatternList
&
patterns
,
MLIRContext
*
mlirContext
)
override
{
RewriteListBuilder
<
NGAddOpConversion
,
NGDotOpConversion
,
NGReturnOpConversion
>::
build
(
patterns
,
mlirContext
,
m_pass
);
}
private
:
DialectLoweringPass
&
m_pass
;
llvm
::
BumpPtrAllocator
allocator
;
};
/// Dialect Lowering Pass to affine ops
...
...
@@ -73,14 +76,17 @@ namespace
{
public
:
DialectLoweringPass
(
ngmlir
::
MLIRCompiler
&
compiler
)
:
m_dialectLowerer
(
*
this
)
,
m_compiler
(
compiler
)
:
m_compiler
(
compiler
)
{
}
void
runOnModule
()
override
;
SmallVector
<
Value
*
,
4
>
buildOutputDefs
(
Operation
*
op
,
PatternRewriter
&
rewriter
);
private
:
/// Collect a set of patterns to convert from the nGraph dialect to Affine dialect.
void
populateNGraphToAffineConversionPatterns
(
OwningRewritePatternList
&
patterns
);
mlir
::
Function
*
getCallDecl
(
StringRef
name
,
ArrayRef
<
Type
>
args
,
ArrayRef
<
Type
>
output
,
...
...
@@ -90,7 +96,7 @@ namespace
Value
*
insertMemMgrDef
(
PatternRewriter
*
rewriter
=
nullptr
);
private
:
DialectLowerer
m_dialectLower
er
;
NGraphTypeConverter
m_typeConvert
er
;
// Value holding mem manager passed pointer
SmallVector
<
Value
*
,
4
>
m_memMgrDefs
;
...
...
@@ -101,21 +107,39 @@ namespace
void
DialectLoweringPass
::
runOnModule
()
{
// Create type converter and initialize conversion patterns.
NGraphTypeConverter
converter
;
OwningRewritePatternList
patterns
;
populateNGraphToAffineConversionPatterns
(
patterns
);
// Create target that defines legal ops for nGraph dialect to be lowered to.
ConversionTarget
target
(
getContext
());
// TODO: Remove NGFakeInputOp. We need to set NGFakeInputOp as legal op because we generate
// it as part of the lowering to affine/standard.
target
.
addLegalDialect
<
AffineOpsDialect
,
StandardOpsDialect
>
();
target
.
addLegalOp
<
NGFakeInputOp
>
();
// capture output values by looking for the Return and grabbing the values
// the order of the returned values matches the order of the lowered func signature for
// results. This is used to find the arg_id that a defined value maps to if it is an output
findOutputValues
();
if
(
failed
(
m_dialectLowerer
.
convert
(
&
getModule
(
))))
if
(
failed
(
applyConversionPatterns
(
getModule
(),
target
,
converter
,
std
::
move
(
patterns
))))
{
getModule
().
getContext
()
->
emitError
(
mlir
::
UnknownLoc
::
get
(
getModule
().
getContext
()),
"Error lowering dialect
\n
"
);
emitError
(
mlir
::
UnknownLoc
::
get
(
&
getContext
()),
"Error lowering nGraph dialect
\n
"
);
signalPassFailure
();
}
processFakeInstrs
();
}
void
DialectLoweringPass
::
populateNGraphToAffineConversionPatterns
(
OwningRewritePatternList
&
patterns
)
{
RewriteListBuilder
<
NGAddOpConversion
,
NGDotOpConversion
,
NGReturnOpConversion
>::
build
(
patterns
,
&
getContext
(),
*
this
);
}
void
DialectLoweringPass
::
findOutputValues
()
{
// get original function
...
...
@@ -138,6 +162,9 @@ namespace
outputCount
=
ret
.
getNumOperands
();
});
// will be populated with lowered output values later
// TODO: This resize is making debugging obscure. When the container is not populated due
// to a bug, null pointers are used by the consumer leading to a crash more difficult to
// root-cause. We should try to change the current approach or introduce verification code.
m_loweredOutputValues
.
resize
(
outputCount
,
nullptr
);
}
...
...
@@ -146,10 +173,11 @@ namespace
{
// it would be nice to insert one fake def at the start of the new func
// however, due to how DialectConversion framework works, new func is only
// materialized after conversion is done (rewriter->getFunction, or even rewriter->getInsertionBlock()->getFunction()
// will give you the original func). This makes it very convoluted to insert instructions at entry block.
// materialized after conversion is done (rewriter->getFunction, or even
// rewriter->getInsertionBlock()->getFunction() will give you the original func). This
// makes it very convoluted to insert instructions at entry block.
auto
op
=
rewriter
->
create
<
NGFakeInputOp
>
(
rewriter
->
getUnknownLoc
(),
IndexType
::
get
(
getModule
().
getContext
()));
IndexType
::
get
(
&
getContext
()));
// will be fixed later to read passed arg instead.
m_memMgrDefs
.
push_back
(
op
.
getResult
());
return
op
.
getResult
();
...
...
@@ -167,8 +195,7 @@ namespace
unsigned
argId
=
(
int
)
attr
.
getInt
();
auto
fakeOp
=
rewriter
.
create
<
NGFakeInputOp
>
(
op
->
getLoc
(),
m_dialectLowerer
.
convertType
(
origResult
->
getType
())
/* convert to lowered type */
m_typeConverter
.
convertType
(
origResult
->
getType
())
/* convert to lowered type */
);
// Fake instrution is short-lived. Verify here.
fakeOp
.
verify
();
...
...
@@ -181,7 +208,7 @@ namespace
auto
tensorType
=
origResult
->
getType
().
cast
<
NGTensorType
>
();
auto
callBackFunc
=
getCallDecl
(
"__mlir_allocate"
,
{
rewriter
.
getIndexType
(),
rewriter
.
getIndexType
()},
{
m_
dialectLower
er
.
convertType
(
tensorType
)},
{
m_
typeConvert
er
.
convertType
(
tensorType
)},
rewriter
);
auto
size
=
tensorType
.
getSizeInBytes
();
...
...
@@ -261,10 +288,10 @@ namespace
return
callBackFuncPtr
;
}
// NGDialect converters
Type
DialectLower
er
::
convertType
(
Type
type
)
Type
NGraphTypeConvert
er
::
convertType
(
Type
type
)
{
// We may need to refactor this code to a external utility if type conversion is needed
// outside of the lowering context since
DialectLower
er is private.
// outside of the lowering context since
NGraphTypeConvert
er is private.
if
(
auto
tensor_type
=
type
.
dyn_cast
<
NGTensorType
>
())
{
...
...
@@ -294,7 +321,7 @@ namespace
}
#define REWRITER(OP) \
void OP##Conversion::rewrite(
\
PatternMatchResult OP##Conversion::matchAndRewrite(
\
Operation* op, ArrayRef<Value*> operands, PatternRewriter& rewriter) const
// ADD
...
...
@@ -334,6 +361,8 @@ namespace
});
// clang-format on
rewriter
.
replaceOp
(
op
,
{
result
});
return
matchSuccess
();
}
REWRITER
(
NGDotOp
)
...
...
@@ -396,9 +425,16 @@ namespace
});
rewriter
.
replaceOp
(
op
,
{
result
});
return
matchSuccess
();
}
REWRITER
(
NGReturnOp
)
{
rewriter
.
replaceOpWithNewOp
<
ReturnOp
>
(
op
);
return
matchSuccess
();
}
REWRITER
(
NGReturnOp
)
{
rewriter
.
replaceOpWithNewOp
<
ReturnOp
>
(
op
);
}
#undef REWRITER
}
...
...
src/contrib/mlir/lowerer.hpp
View file @
ee5d46c1
...
...
@@ -27,6 +27,8 @@ namespace ngraph
namespace
ngmlir
{
class
MLIRCompiler
;
using
OwningRewritePatternList
=
std
::
vector
<
std
::
unique_ptr
<
mlir
::
RewritePattern
>>
;
}
}
}
...
...
src/contrib/mlir/op_lowerers.inc
View file @
ee5d46c1
...
...
@@ -17,17 +17,19 @@
// Add new dialect ops lowerers to this file
#define DECL_OP_CONV(OP) \
class
OP
##Conversion : public mlir::DialectConversionPattern \
{
\
public
:
\
explicit
OP
##Conversion(mlir::MLIRContext *context, DialectLoweringPass& pass)\
:
mlir
::
DialectConversionPattern
(
mlir
::
OP
::
getOperationName
(),
1
,
context
),
\
m_pass
(
pass
)
\
{}
\
void
rewrite
(
Operation
*
op
,
ArrayRef
<
Value
*>
operands
,
PatternRewriter
&
rewriter
)
const
override
;
\
DialectLoweringPass
&
m_pass
;
\
};
#define DECL_OP_CONV(OP) \
class
OP
##Conversion : public NGraphOpLowering \
{
\
public
:
\
explicit
OP
##Conversion(mlir::MLIRContext* context, DialectLoweringPass& pass) \
:
NGraphOpLowering
(
mlir
::
OP
::
getOperationName
(),
context
,
pass
)
\
{
\
}
\
\
PatternMatchResult
matchAndRewrite
(
Operation
*
op
,
\
ArrayRef
<
Value
*>
operands
,
\
PatternRewriter
&
rewriter
)
const
override
;
\
};
DECL_OP_CONV
(
NGAddOp
)
DECL_OP_CONV
(
NGDotOp
)
...
...
src/ngraph/runtime/cpu/cpu_external_function.cpp
View file @
ee5d46c1
...
...
@@ -33,7 +33,10 @@
#include "ngraph/codegen/execution_engine.hpp"
#endif
#ifdef NGRAPH_MLIR_ENABLE
#include "contrib/mlir/pass/mlir_subgraph_extraction.hpp"
#endif
#include "ngraph/descriptor/input.hpp"
#include "ngraph/descriptor/output.hpp"
#include "ngraph/file_util.hpp"
...
...
src/ngraph/runtime/gpu/gpu_compiled_function.cpp
View file @
ee5d46c1
...
...
@@ -173,9 +173,6 @@ void runtime::gpu::GPUCompiledFunction::compile()
pass_manager
.
register_pass
<
runtime
::
gpu
::
pass
::
BatchNormCache
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
LikeReplacement
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
FusedOpDecomposition
>
();
// Run this pass for the second time since, some fused operators like LSTMCell may use
// other fused operators inside.
pass_manager
.
register_pass
<
ngraph
::
pass
::
FusedOpDecomposition
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
ImplicitBroadcastElimination
>
();
pass_manager
.
register_pass
<
runtime
::
gpu
::
pass
::
GPULayout
>
(
this
);
pass_manager
.
register_pass
<
ngraph
::
pass
::
AssignLayout
<
descriptor
::
layout
::
DenseTensorLayout
>>
();
...
...
src/ngraph/runtime/intelgpu/intelgpu_backend.cpp
View file @
ee5d46c1
...
...
@@ -430,10 +430,6 @@ shared_ptr<runtime::Executable>
if
(
m_disable_backend_optimizations
<
2
)
{
pass_manager
.
register_pass
<
ngraph
::
pass
::
FusedOpDecomposition
>
(
IntelGPUBackend
::
is_supported_impl
);
// Run this pass for the second time since, some fused operators like LSTMCell may use
// other fused operators inside.
pass_manager
.
register_pass
<
ngraph
::
pass
::
FusedOpDecomposition
>
(
IntelGPUBackend
::
is_supported_impl
);
pass_manager
.
register_pass
<
ngraph
::
pass
::
ImplicitBroadcastElimination
>
();
...
...
src/ngraph/runtime/interpreter/int_executable.cpp
View file @
ee5d46c1
...
...
@@ -47,9 +47,6 @@ runtime::interpreter::INTExecutable::INTExecutable(const shared_ptr<Function>& f
pass
::
Manager
pass_manager
;
pass_manager
.
register_pass
<
pass
::
LikeReplacement
>
();
pass_manager
.
register_pass
<
pass
::
FusedOpDecomposition
>
();
// Run this pass for the second time since, some fused operators like LSTMCell may use
// other fused operators inside.
pass_manager
.
register_pass
<
pass
::
FusedOpDecomposition
>
();
pass_manager
.
register_pass
<
pass
::
ImplicitBroadcastElimination
>
();
pass_manager
.
register_pass
<
pass
::
AssignLayout
<
DenseTensorLayout
>>
();
pass_manager
.
register_pass
<
pass
::
Liveness
>
();
...
...
src/ngraph/runtime/plaidml/plaidml_compiler.cpp
View file @
ee5d46c1
...
...
@@ -20,6 +20,7 @@
#include "ngraph/pass/algebraic_simplification.hpp"
#include "ngraph/pass/core_fusion.hpp"
#include "ngraph/pass/cse.hpp"
#include "ngraph/pass/fused_op_decomposition.hpp"
#include "ngraph/pass/get_output_element_elimination.hpp"
#include "ngraph/pass/like_replacement.hpp"
#include "ngraph/pass/liveness.hpp"
...
...
@@ -87,6 +88,7 @@ std::shared_ptr<ngraph::runtime::plaidml::PlaidML_Executable>
ngraph
::
pass
::
Manager
pass_manager
;
// We apply the same general-purposes passes as the CPU backend.
pass_manager
.
register_pass
<
ngraph
::
pass
::
FusedOpDecomposition
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
LikeReplacement
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
NopElimination
>
();
pass_manager
.
register_pass
<
ngraph
::
pass
::
ZeroDimTensorElimination
>
();
...
...
src/ngraph/runtime/plaidml/plaidml_config.cpp
View file @
ee5d46c1
...
...
@@ -242,6 +242,8 @@ ngraph::runtime::plaidml::Config
}
// Reject unknown options
NGRAPH_ERR
<<
"Unrecognized PlaidML backend option: "
<<
std
::
string
{
oname_begin
,
static_cast
<
std
::
size_t
>
(
oname_end
-
oname_begin
)};
err
=
true
;
}
...
...
src/ngraph/runtime/plaidml/unit_test.manifest
View file @
ee5d46c1
...
...
@@ -229,12 +229,6 @@ backwards_softmax_underflow
backwards_softmax_3d
batch_mat_mul_forward
dot_matrix_2x0_0x2
rnn_cell_no_bias
rnn_cell_bias_clip
rnn_cell_activation_function
gru_cell_bias_clip
gru_cell_linear_before_reset
gru_cell_activation_function
# dgkutnic ww24.5: these tests are to be triaged by the PlaidML team
# ww25.2: re-scrubbed this list of tests after fixing check_inputs
...
...
@@ -259,3 +253,29 @@ group_conv_transpose
group_conv_transpose_output_shape
divide_python_rounding_int32
backwards_batchmatmul_tensor2_tensor2
# unsupported ops: 'QuantizedConvolution', 'QuantizedDot', 'TopK', 'Erf', 'EmbeddingLookup'
model_quant_conv_linear
model_conv_integer_no_zero_point
model_matmul_integer_no_zero_point
model_matmul_integer_4d_no_zero_point
model_top_k
model_erf
model_erf_int32
model_hardmax
# node validation error: "Argument shapes are inconsistent."
model_lstm_fwd_with_clip
model_lstm_fwd_mixed_seq
model_lstm_fwd_hardsigmoid_activation
model_reduce_log_sum
model_reduce_log_sum_exp
model_reduce_mean
# result mismatch
model_dequantize_linear_scalar_zero_scale_int8
model_softmax
avg_pool_3d_uneven_strided_padded
rnn_cell_activation_function
gru_cell_bias_clip
gru_cell_linear_before_reset
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment