Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in / Register
Toggle navigation
O
opencv
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Packages
Packages
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
submodule
opencv
Commits
2a6637af
Commit
2a6637af
authored
Feb 11, 2020
by
Alexander Alekhin
Browse files
Options
Browse Files
Download
Plain Diff
Merge pull request #16516 from alalek:dnn_dump_network
parents
d4422e47
6eba1a4d
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
214 additions
and
104 deletions
+214
-104
dnn.cpp
modules/dnn/src/dnn.cpp
+214
-104
No files found.
modules/dnn/src/dnn.cpp
View file @
2a6637af
...
@@ -62,6 +62,8 @@ namespace cv {
...
@@ -62,6 +62,8 @@ namespace cv {
namespace
dnn
{
namespace
dnn
{
CV__DNN_EXPERIMENTAL_NS_BEGIN
CV__DNN_EXPERIMENTAL_NS_BEGIN
static
size_t
DNN_NETWORK_DUMP
=
utils
::
getConfigurationParameterSizeT
(
"OPENCV_DNN_NETWORK_DUMP"
,
0
);
// this option is useful to run valgrind memory errors detection
// this option is useful to run valgrind memory errors detection
static
bool
DNN_DISABLE_MEMORY_OPTIMIZATIONS
=
utils
::
getConfigurationParameterBool
(
"OPENCV_DNN_DISABLE_MEMORY_OPTIMIZATIONS"
,
false
);
static
bool
DNN_DISABLE_MEMORY_OPTIMIZATIONS
=
utils
::
getConfigurationParameterBool
(
"OPENCV_DNN_DISABLE_MEMORY_OPTIMIZATIONS"
,
false
);
...
@@ -86,6 +88,7 @@ using std::vector;
...
@@ -86,6 +88,7 @@ using std::vector;
using
std
::
map
;
using
std
::
map
;
using
std
::
make_pair
;
using
std
::
make_pair
;
using
std
::
set
;
using
std
::
set
;
using
std
::
string
;
//==================================================================================================
//==================================================================================================
...
@@ -1054,12 +1057,19 @@ static Ptr<BackendWrapper> wrapMat(int backendId, int targetId, cv::Mat& m)
...
@@ -1054,12 +1057,19 @@ static Ptr<BackendWrapper> wrapMat(int backendId, int targetId, cv::Mat& m)
return
Ptr
<
BackendWrapper
>
();
// TODO Error?
return
Ptr
<
BackendWrapper
>
();
// TODO Error?
}
}
static
int
g_networkId
=
0
;
struct
Net
::
Impl
struct
Net
::
Impl
{
{
typedef
std
::
map
<
int
,
LayerShapes
>
LayersShapesMap
;
typedef
std
::
map
<
int
,
LayerShapes
>
LayersShapesMap
;
typedef
std
::
map
<
int
,
LayerData
>
MapIdToLayerData
;
typedef
std
::
map
<
int
,
LayerData
>
MapIdToLayerData
;
const
int
networkId
;
// network global identifier
int
networkDumpCounter
;
// dump counter
Impl
()
Impl
()
:
networkId
(
CV_XADD
(
&
g_networkId
,
1
))
,
networkDumpCounter
(
0
)
{
{
//allocate fake net input layer
//allocate fake net input layer
netInputLayer
=
Ptr
<
DataLayer
>
(
new
DataLayer
());
netInputLayer
=
Ptr
<
DataLayer
>
(
new
DataLayer
());
...
@@ -1223,6 +1233,11 @@ struct Net::Impl
...
@@ -1223,6 +1233,11 @@ struct Net::Impl
{
{
CV_TRACE_FUNCTION
();
CV_TRACE_FUNCTION
();
if
(
DNN_NETWORK_DUMP
>
0
&&
networkDumpCounter
==
0
)
{
dumpNetworkToFile
();
}
if
(
preferableBackend
==
DNN_BACKEND_DEFAULT
)
if
(
preferableBackend
==
DNN_BACKEND_DEFAULT
)
preferableBackend
=
(
Backend
)
PARAM_DNN_BACKEND_DEFAULT
;
preferableBackend
=
(
Backend
)
PARAM_DNN_BACKEND_DEFAULT
;
#ifdef HAVE_INF_ENGINE
#ifdef HAVE_INF_ENGINE
...
@@ -1299,6 +1314,11 @@ struct Net::Impl
...
@@ -1299,6 +1314,11 @@ struct Net::Impl
netWasAllocated
=
true
;
netWasAllocated
=
true
;
this
->
blobsToKeep
=
blobsToKeep_
;
this
->
blobsToKeep
=
blobsToKeep_
;
if
(
DNN_NETWORK_DUMP
>
0
)
{
dumpNetworkToFile
();
}
}
}
}
}
...
@@ -2979,6 +2999,31 @@ struct Net::Impl
...
@@ -2979,6 +2999,31 @@ struct Net::Impl
static
static
Net
createNetworkFromModelOptimizer
(
InferenceEngine
::
CNNNetwork
&
ieNet
);
Net
createNetworkFromModelOptimizer
(
InferenceEngine
::
CNNNetwork
&
ieNet
);
#endif
#endif
string
dump
();
void
dumpNetworkToFile
()
{
#ifndef OPENCV_DNN_DISABLE_NETWORK_AUTO_DUMP
String
dumpFileName
=
cv
::
format
(
"ocv_dnn_net_%05d_%02d.dot"
,
networkId
,
networkDumpCounter
++
);
try
{
string
dumpStr
=
dump
();
std
::
ofstream
out
(
dumpFileName
.
c_str
(),
std
::
ios
::
out
|
std
::
ios
::
binary
);
out
<<
dumpStr
;
}
catch
(
const
std
::
exception
&
e
)
{
std
::
ofstream
out
((
dumpFileName
+
".error"
).
c_str
(),
std
::
ios
::
out
);
out
<<
"Exception: "
<<
e
.
what
()
<<
std
::
endl
;
}
catch
(...)
{
std
::
ofstream
out
((
dumpFileName
+
".error"
).
c_str
(),
std
::
ios
::
out
);
out
<<
"Can't dump: unknown exception"
<<
std
::
endl
;
}
#endif
}
};
};
Net
::
Net
()
:
impl
(
new
Net
::
Impl
)
Net
::
Net
()
:
impl
(
new
Net
::
Impl
)
...
@@ -3496,20 +3541,26 @@ int Net::getLayerId(const String &layer)
...
@@ -3496,20 +3541,26 @@ int Net::getLayerId(const String &layer)
return
impl
->
getLayerId
(
layer
);
return
impl
->
getLayerId
(
layer
);
}
}
String
parseLayerParams
(
const
String
&
name
,
const
LayerParams
&
lp
)
{
static
string
dumpLayerParameterSize
(
const
string
&
name
,
const
LayerParams
&
lp
)
{
std
::
ostringstream
out
(
name
,
std
::
ios
::
ate
);
DictValue
param
=
lp
.
get
(
name
);
DictValue
param
=
lp
.
get
(
name
);
std
::
ostringstream
out
;
switch
(
param
.
size
())
out
<<
name
<<
" "
;
{
switch
(
param
.
size
())
{
case
1
:
out
<<
" : "
;
break
;
case
1
:
out
<<
": "
;
break
;
case
2
:
out
<<
" (HxW): "
;
break
;
case
2
:
out
<<
"(HxW): "
;
break
;
case
3
:
out
<<
" (DxHxW): "
;
break
;
case
3
:
out
<<
"(DxHxW): "
;
break
;
default
:
default
:
CV_Error
(
Error
::
StsNotImplemented
,
format
(
"Unsupported %s size = %d"
,
name
.
c_str
(),
param
.
size
()));
CV_LOG_INFO
(
NULL
,
format
(
"DNN/dumpLayerParameterSize(): Unsupported '%s' size = %d"
,
name
.
c_str
(),
param
.
size
()));
out
<<
": "
;
}
}
for
(
size_t
i
=
0
;
i
<
param
.
size
()
-
1
;
i
++
)
{
for
(
size_t
i
=
0
;
i
<
param
.
size
();
i
++
)
out
<<
param
.
get
<
int
>
(
i
)
<<
" x "
;
{
if
(
i
>
0
)
out
<<
" x "
;
out
<<
param
.
get
<
int
>
(
i
);
}
}
out
<<
param
.
get
<
int
>
(
param
.
size
()
-
1
)
<<
"
\\
l"
;
return
out
.
str
();
return
out
.
str
();
}
}
...
@@ -3517,23 +3568,33 @@ String Net::dump()
...
@@ -3517,23 +3568,33 @@ String Net::dump()
{
{
CV_Assert
(
!
empty
());
CV_Assert
(
!
empty
());
if
(
impl
->
netInputLayer
->
inputsData
.
empty
())
bool
hasInput
=
!
impl
->
netInputLayer
->
inputsData
.
empty
();
CV_Error
(
Error
::
StsError
,
"Requested set input"
);
if
(
!
impl
->
netWasAllocated
)
if
(
hasInput
)
impl
->
setUpNet
();
{
if
(
!
impl
->
netWasAllocated
)
impl
->
setUpNet
();
}
return
impl
->
dump
();
}
string
Net
::
Impl
::
dump
()
{
bool
hasInput
=
!
netInputLayer
->
inputsData
.
empty
();
std
::
ostringstream
out
;
std
::
ostringstream
out
;
std
::
map
<
int
,
LayerData
>&
map
=
impl
->
layers
;
const
std
::
map
<
int
,
LayerData
>&
map
=
layers
;
int
prefBackend
=
impl
->
preferableBackend
;
Backend
prefBackend
=
(
Backend
)
preferableBackend
;
std
::
vector
<
std
::
vector
<
int
>
>
skippedLayers
;
std
::
vector
<
std
::
vector
<
int
>
>
skippedLayers
;
std
::
vector
<
int
>
skipId
;
std
::
vector
<
int
>
skipId
;
std
::
vector
<
int
>
allLayers
(
map
.
size
(),
-
1
);
std
::
vector
<
int
>
allLayers
(
map
.
size
(),
-
1
);
int
idPrev
=
-
1
;
int
idPrev
=
-
1
;
Ptr
<
BackendNode
>
prevNode
;
Ptr
<
BackendNode
>
prevNode
;
for
(
std
::
map
<
int
,
LayerData
>::
reverse_iterator
rit
=
map
.
rbegin
();
rit
!=
map
.
rend
();
++
rit
)
for
(
std
::
map
<
int
,
LayerData
>::
const_
reverse_iterator
rit
=
map
.
rbegin
();
rit
!=
map
.
rend
();
++
rit
)
{
{
std
::
map
<
int
,
Ptr
<
BackendNode
>
>::
iterator
itBackend
=
rit
->
second
.
backendNodes
.
find
(
prefBackend
);
std
::
map
<
int
,
Ptr
<
BackendNode
>
>::
const_
iterator
itBackend
=
rit
->
second
.
backendNodes
.
find
(
prefBackend
);
if
(
prefBackend
==
DNN_BACKEND_OPENCV
||
itBackend
==
rit
->
second
.
backendNodes
.
end
()
||
if
(
prefBackend
==
DNN_BACKEND_OPENCV
||
itBackend
==
rit
->
second
.
backendNodes
.
end
()
||
itBackend
->
second
.
empty
())
itBackend
->
second
.
empty
())
{
{
...
@@ -3572,154 +3633,203 @@ String Net::dump()
...
@@ -3572,154 +3633,203 @@ String Net::dump()
prevNode
=
itBackend
->
second
;
prevNode
=
itBackend
->
second
;
}
}
}
}
String
colors
[]
=
{
"#ffffb3"
,
"#fccde5"
,
"#8dd3c7"
,
"#bebada"
,
"#80b1d3"
,
"#fdb462"
};
string
colors
[]
=
{
"#ffffb3"
,
"#fccde5"
,
"#8dd3c7"
,
"#bebada"
,
"#80b1d3"
,
"#fdb462"
};
String
backend
;
string
backend
;
switch
(
prefBackend
)
{
switch
(
prefBackend
)
{
case
DNN_BACKEND_DEFAULT
:
backend
=
"DEFAULT/"
;
break
;
case
DNN_BACKEND_DEFAULT
:
backend
=
"DEFAULT/"
;
break
;
case
DNN_BACKEND_HALIDE
:
backend
=
"HALIDE/"
;
break
;
case
DNN_BACKEND_HALIDE
:
backend
=
"HALIDE/"
;
break
;
case
DNN_BACKEND_INFERENCE_ENGINE
:
// fallthru
case
DNN_BACKEND_INFERENCE_ENGINE
:
// fallthru
case
DNN_BACKEND_INFERENCE_ENGINE_NN_BUILDER_2019
:
backend
=
"DLIE/"
;
break
;
case
DNN_BACKEND_INFERENCE_ENGINE_NN_BUILDER_2019
:
backend
=
"DLIE/"
;
break
;
case
DNN_BACKEND_INFERENCE_ENGINE_NGRAPH
:
backend
=
"NGRAPH/"
;
break
;
case
DNN_BACKEND_INFERENCE_ENGINE_NGRAPH
:
backend
=
"NGRAPH/"
;
break
;
case
DNN_BACKEND_OPENCV
:
backend
=
"OCV/"
;
break
;
case
DNN_BACKEND_OPENCV
:
backend
=
"OCV/"
;
break
;
// don't use default:
}
}
out
<<
"digraph G {
"
<<
'\n'
;
out
<<
"digraph G {
\n
"
;
// Add nodes
// Add nodes
for
(
std
::
map
<
int
,
LayerData
>::
iterator
it
=
map
.
begin
();
it
!=
map
.
end
();
++
it
)
for
(
std
::
map
<
int
,
LayerData
>::
const_
iterator
it
=
map
.
begin
();
it
!=
map
.
end
();
++
it
)
{
{
String
name
=
it
->
second
.
params
.
name
;
const
LayerData
&
ld
=
it
->
second
;
if
(
allLayers
[
it
->
first
]
==
-
1
&&
!
name
.
empty
())
{
string
name
=
ld
.
params
.
name
;
out
<<
" "
<<
"
\"
"
<<
name
<<
"
\"
"
<<
" [label=
\"
"
;
std
::
vector
<
int
>
clusterIds
(
1
,
it
->
first
);
skipId
.
clear
();
if
(
allLayers
[
it
->
first
]
==
-
1
&&
!
name
.
empty
())
skipId
.
push_back
(
it
->
first
);
{
out
<<
"
\t\"
"
<<
name
<<
"
\"
[label=
\"
"
;
}
}
else
if
(
name
.
empty
()
||
it
->
first
!=
skippedLayers
[
allLayers
[
it
->
first
]][
0
])
else
if
(
name
.
empty
()
||
it
->
first
!=
skippedLayers
[
allLayers
[
it
->
first
]][
0
])
{
continue
;
continue
;
else
{
// first node in cluster : it->first == skippedLayers[allLayers[it->first]][0]
}
else
// first node in cluster : it->first == skippedLayers[allLayers[it->first]][0]
{
int
cluster
=
allLayers
[
it
->
first
];
int
cluster
=
allLayers
[
it
->
first
];
out
<<
"
"
<<
"
\"
"
<<
"cluster_"
<<
cluster
<<
"
\"
"
<<
" [label=
\"
{"
;
out
<<
"
\t\"
"
<<
"cluster_"
<<
cluster
<<
"
\
"
[label=
\"
{"
;
skipId
=
skippedLayers
[
allLayers
[
it
->
first
]];
// vertices in current cluster
clusterIds
=
skippedLayers
[
allLayers
[
it
->
first
]];
// vertices in current cluster
}
}
for
(
int
i
=
0
;
i
<
skipId
.
size
();
i
++
)
for
(
int
i
=
0
;
i
<
clusterIds
.
size
();
i
++
)
{
{
LayerParams
&
lp
=
map
[
skipId
[
i
]].
params
;
CV_DbgAssert
(
map
.
find
(
clusterIds
[
i
])
!=
map
.
end
());
const
LayerParams
&
lp
=
map
.
find
(
clusterIds
[
i
])
->
second
.
params
;
if
(
!
lp
.
name
.
empty
())
{
if
(
!
lp
.
name
.
empty
())
{
if
(
i
>
0
)
{
if
(
i
>
0
)
{
out
<<
" | "
;
out
<<
" | "
;
}
}
out
<<
lp
.
name
<<
"
\\
n"
<<
lp
.
type
<<
"
\\
n"
;
out
<<
lp
.
name
<<
"
\\
n"
<<
lp
.
type
<<
"
\\
n"
;
// align center
if
(
lp
.
has
(
"kernel_size"
))
{
if
(
lp
.
has
(
"kernel_size"
))
String
kernel
=
parseLayerParams
(
"kernel_size"
,
lp
);
{
string
kernel
=
dumpLayerParameterSize
(
"kernel_size"
,
lp
);
out
<<
kernel
;
out
<<
kernel
;
out
<<
"
\\
l"
;
// align left
}
else
if
(
lp
.
has
(
"kernel_h"
)
&&
lp
.
has
(
"kernel_w"
))
{
}
else
if
(
lp
.
has
(
"kernel_h"
)
&&
lp
.
has
(
"kernel_w"
))
{
DictValue
h
=
lp
.
get
(
"kernel_h"
);
DictValue
h
=
lp
.
get
(
"kernel_h"
);
DictValue
w
=
lp
.
get
(
"kernel_w"
);
DictValue
w
=
lp
.
get
(
"kernel_w"
);
out
<<
"kernel (HxW): "
<<
h
<<
" x "
<<
w
<<
"
\\
l"
;
out
<<
"kernel (HxW): "
<<
h
<<
" x "
<<
w
;
out
<<
"
\\
l"
;
// align left
}
}
if
(
lp
.
has
(
"stride"
))
{
if
(
lp
.
has
(
"stride"
))
{
String
stride
=
parseLayerParams
(
"stride"
,
lp
);
string
stride
=
dumpLayerParameterSize
(
"stride"
,
lp
);
out
<<
stride
;
out
<<
stride
;
out
<<
"
\\
l"
;
// align left
}
else
if
(
lp
.
has
(
"stride_h"
)
&&
lp
.
has
(
"stride_w"
))
{
}
else
if
(
lp
.
has
(
"stride_h"
)
&&
lp
.
has
(
"stride_w"
))
{
DictValue
h
=
lp
.
get
(
"stride_h"
);
DictValue
h
=
lp
.
get
(
"stride_h"
);
DictValue
w
=
lp
.
get
(
"stride_w"
);
DictValue
w
=
lp
.
get
(
"stride_w"
);
out
<<
"stride (HxW): "
<<
h
<<
" x "
<<
w
<<
"
\\
l"
;
out
<<
"stride (HxW): "
<<
h
<<
" x "
<<
w
;
out
<<
"
\\
l"
;
// align left
}
}
if
(
lp
.
has
(
"dilation"
))
{
if
(
lp
.
has
(
"dilation"
))
{
String
dilation
=
parseLayerParams
(
"dilation"
,
lp
);
string
dilation
=
dumpLayerParameterSize
(
"dilation"
,
lp
);
out
<<
dilation
;
out
<<
dilation
;
out
<<
"
\\
l"
;
// align left
}
else
if
(
lp
.
has
(
"dilation_h"
)
&&
lp
.
has
(
"dilation_w"
))
{
}
else
if
(
lp
.
has
(
"dilation_h"
)
&&
lp
.
has
(
"dilation_w"
))
{
DictValue
h
=
lp
.
get
(
"dilation_h"
);
DictValue
h
=
lp
.
get
(
"dilation_h"
);
DictValue
w
=
lp
.
get
(
"dilation_w"
);
DictValue
w
=
lp
.
get
(
"dilation_w"
);
out
<<
"dilation (HxW): "
<<
h
<<
" x "
<<
w
<<
"
\\
l"
;
out
<<
"dilation (HxW): "
<<
h
<<
" x "
<<
w
;
out
<<
"
\\
l"
;
// align left
}
}
if
(
lp
.
has
(
"pad"
))
{
if
(
lp
.
has
(
"pad"
))
{
DictValue
pad
=
lp
.
get
(
"pad"
);
DictValue
pad
=
lp
.
get
(
"pad"
);
out
<<
"pad "
;
out
<<
"pad "
;
switch
(
pad
.
size
())
{
switch
(
pad
.
size
())
case
1
:
out
<<
": "
<<
pad
<<
"
\\
l"
;
break
;
{
case
2
:
out
<<
"(HxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
" x "
<<
pad
.
get
<
int
>
(
1
)
<<
")"
<<
"
\\
l"
;
break
;
case
1
:
out
<<
": "
<<
pad
;
break
;
case
4
:
out
<<
"(HxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
", "
<<
pad
.
get
<
int
>
(
2
)
<<
") x ("
<<
pad
.
get
<
int
>
(
1
)
<<
", "
<<
pad
.
get
<
int
>
(
3
)
<<
")"
<<
"
\\
l"
;
break
;
case
2
:
case
6
:
out
<<
"(DxHxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
", "
<<
pad
.
get
<
int
>
(
3
)
<<
") x ("
<<
pad
.
get
<
int
>
(
1
)
<<
", "
<<
pad
.
get
<
int
>
(
4
)
out
<<
"(HxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
" x "
<<
pad
.
get
<
int
>
(
1
)
<<
")"
;
<<
") x ("
<<
pad
.
get
<
int
>
(
2
)
<<
", "
<<
pad
.
get
<
int
>
(
5
)
<<
")"
<<
"
\\
l"
;
break
;
break
;
case
4
:
out
<<
"(HxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
", "
<<
pad
.
get
<
int
>
(
2
)
<<
") x ("
<<
pad
.
get
<
int
>
(
1
)
<<
", "
<<
pad
.
get
<
int
>
(
3
)
<<
")"
;
break
;
case
6
:
out
<<
"(DxHxW): ("
<<
pad
.
get
<
int
>
(
0
)
<<
", "
<<
pad
.
get
<
int
>
(
3
)
<<
") x ("
<<
pad
.
get
<
int
>
(
1
)
<<
", "
<<
pad
.
get
<
int
>
(
4
)
<<
") x ("
<<
pad
.
get
<
int
>
(
2
)
<<
", "
<<
pad
.
get
<
int
>
(
5
)
<<
")"
;
break
;
default
:
CV_Error
(
Error
::
StsNotImplemented
,
format
(
"Unsupported pad size = %d"
,
pad
.
size
()));
default
:
CV_Error
(
Error
::
StsNotImplemented
,
format
(
"Unsupported pad size = %d"
,
pad
.
size
()));
}
}
}
else
if
(
lp
.
has
(
"pad_l"
)
&&
lp
.
has
(
"pad_t"
)
&&
lp
.
has
(
"pad_r"
)
&&
lp
.
has
(
"pad_b"
))
{
out
<<
"
\\
l"
;
// align left
DictValue
l
=
lp
.
get
(
"pad_l"
);
}
else
if
(
lp
.
has
(
"pad_l"
)
&&
lp
.
has
(
"pad_t"
)
&&
lp
.
has
(
"pad_r"
)
&&
lp
.
has
(
"pad_b"
))
{
DictValue
t
=
lp
.
get
(
"pad_t"
);
DictValue
l
=
lp
.
get
(
"pad_l"
);
DictValue
r
=
lp
.
get
(
"pad_r"
);
DictValue
t
=
lp
.
get
(
"pad_t"
);
DictValue
b
=
lp
.
get
(
"pad_b"
);
DictValue
r
=
lp
.
get
(
"pad_r"
);
out
<<
"pad (HxW): ("
<<
t
<<
", "
<<
b
<<
") x ("
<<
l
<<
", "
<<
r
<<
")"
<<
"
\\
l"
;
DictValue
b
=
lp
.
get
(
"pad_b"
);
}
out
<<
"pad (HxW): ("
<<
t
<<
", "
<<
b
<<
") x ("
<<
l
<<
", "
<<
r
<<
")"
;
else
if
(
lp
.
has
(
"pooled_w"
)
||
lp
.
has
(
"pooled_h"
))
{
out
<<
"
\\
l"
;
// align left
DictValue
h
=
lp
.
get
(
"pooled_h"
);
}
DictValue
w
=
lp
.
get
(
"pooled_w"
);
else
if
(
lp
.
has
(
"pooled_w"
)
||
lp
.
has
(
"pooled_h"
))
{
out
<<
"pad (HxW): "
<<
h
<<
" x "
<<
w
<<
"
\\
l"
;
DictValue
h
=
lp
.
get
(
"pooled_h"
);
}
DictValue
w
=
lp
.
get
(
"pooled_w"
);
if
(
lp
.
has
(
"pool"
))
{
out
<<
"pad pooled (HxW): "
<<
h
<<
" x "
<<
w
;
out
<<
"pool: "
<<
lp
.
get
(
"pool"
)
<<
"
\\
l"
;
out
<<
"
\\
l"
;
// align left
}
}
if
(
lp
.
has
(
"global_pooling"
))
{
if
(
lp
.
has
(
"pool"
))
{
out
<<
"global_pooling: "
<<
lp
.
get
(
"global_pooling"
)
<<
"
\\
l"
;
out
<<
"pool: "
<<
lp
.
get
(
"pool"
);
}
out
<<
"
\\
l"
;
// align left
if
(
lp
.
has
(
"group"
))
{
}
out
<<
"group: "
<<
lp
.
get
(
"group"
)
<<
"
\\
l"
;
if
(
lp
.
has
(
"global_pooling"
))
{
}
out
<<
"global_pooling: "
<<
lp
.
get
(
"global_pooling"
);
}
out
<<
"
\\
l"
;
// align left
}
}
if
(
!
it
->
second
.
outputBlobs
.
empty
())
if
(
lp
.
has
(
"group"
))
{
out
<<
"output: "
<<
it
->
second
.
outputBlobs
[
0
].
size
<<
"
\\
l"
;
out
<<
"group: "
<<
lp
.
get
(
"group"
);
out
<<
"
\\
l"
;
// align left
Ptr
<
BackendNode
>
layerBackend
=
it
->
second
.
backendNodes
[
prefBackend
];
}
out
<<
(
!
layerBackend
.
empty
()
?
backend
:
"OCV/"
);
}
int
colorId
=
0
;
}
switch
(
it
->
second
.
layerInstance
->
preferableTarget
)
{
if
(
!
ld
.
outputBlobs
.
empty
())
case
DNN_TARGET_CPU
:
out
<<
"CPU
\\
n"
;
colorId
=
layerBackend
.
empty
()
?
0
:
5
;
break
;
{
case
DNN_TARGET_OPENCL
:
out
<<
"OCL
\\
n"
;
colorId
=
1
;
break
;
out
<<
"output: "
<<
ld
.
outputBlobs
[
0
].
size
;
case
DNN_TARGET_OPENCL_FP16
:
out
<<
"OCL_FP16
\\
n"
;
colorId
=
2
;
break
;
out
<<
"
\\
l"
;
// align left
case
DNN_TARGET_MYRIAD
:
out
<<
"MYRIAD
\\
n"
;
colorId
=
3
;
break
;
}
case
DNN_TARGET_FPGA
:
out
<<
"FPGA
\\
n"
;
colorId
=
4
;
break
;
}
Ptr
<
BackendNode
>
layerBackend
;
out
<<
((
skipId
.
size
()
==
1
)
?
"
\"
"
:
" }
\"
"
);
std
::
map
<
int
,
Ptr
<
BackendNode
>
>::
const_iterator
ibn
=
ld
.
backendNodes
.
find
(
prefBackend
);
out
<<
"fillcolor=
\"
"
<<
colors
[
colorId
]
<<
"
\"
"
;
if
(
ibn
!=
ld
.
backendNodes
.
end
())
out
<<
"style=filled "
;
layerBackend
=
ibn
->
second
;
out
<<
"shape="
<<
((
skipId
.
size
()
==
1
)
?
"box"
:
"record"
)
<<
"]"
<<
'\n'
;
out
<<
(
!
layerBackend
.
empty
()
?
backend
:
"OCV/"
);
int
colorId
=
0
;
const
Target
target
=
ld
.
layerInstance
.
empty
()
?
DNN_TARGET_CPU
:
(
Target
)(
ld
.
layerInstance
->
preferableTarget
);
// TODO fix preferableTarget type
switch
(
target
)
{
case
DNN_TARGET_CPU
:
out
<<
"CPU"
;
colorId
=
layerBackend
.
empty
()
?
0
:
5
;
break
;
case
DNN_TARGET_OPENCL
:
out
<<
"OCL"
;
colorId
=
1
;
break
;
case
DNN_TARGET_OPENCL_FP16
:
out
<<
"OCL_FP16"
;
colorId
=
2
;
break
;
case
DNN_TARGET_MYRIAD
:
out
<<
"MYRIAD"
;
colorId
=
3
;
break
;
case
DNN_TARGET_FPGA
:
out
<<
"FPGA"
;
colorId
=
4
;
break
;
// don't use default:
}
out
<<
"
\\
n"
;
// align center
out
<<
((
clusterIds
.
size
()
==
1
)
?
"
\"
"
:
" }
\"
"
);
out
<<
"fillcolor=
\"
"
<<
colors
[
colorId
]
<<
"
\"
"
;
out
<<
"style=filled "
;
out
<<
"shape="
<<
((
clusterIds
.
size
()
==
1
)
?
"box"
:
"record"
)
<<
"]
\n
"
;
}
}
out
<<
'\n'
;
out
<<
'\n'
;
// Add edges
// Add edges
int
inputsSize
=
impl
->
netInputLayer
->
outNames
.
size
()
;
int
inputsSize
=
hasInput
?
netInputLayer
->
outNames
.
size
()
:
0
;
for
(
std
::
map
<
int
,
LayerData
>::
iterator
it
=
map
.
begin
();
it
!=
map
.
end
();
++
it
)
for
(
std
::
map
<
int
,
LayerData
>::
const_
iterator
it
=
map
.
begin
();
it
!=
map
.
end
();
++
it
)
{
{
const
LayerData
&
ld
=
it
->
second
;
if
(
allLayers
[
it
->
first
]
==
-
1
)
// node
if
(
allLayers
[
it
->
first
]
==
-
1
)
// node
{
{
for
(
int
i
=
0
;
i
<
it
->
secon
d
.
consumers
.
size
();
i
++
)
for
(
int
i
=
0
;
i
<
l
d
.
consumers
.
size
();
i
++
)
{
{
int
outId
=
it
->
secon
d
.
consumers
[
i
].
lid
;
int
outId
=
l
d
.
consumers
[
i
].
lid
;
if
(
it
==
map
.
begin
()
&&
inputsSize
>
1
)
if
(
it
==
map
.
begin
()
&&
inputsSize
>
1
)
out
<<
"
"
<<
"
\"
"
<<
it
->
secon
d
.
name
<<
"_"
<<
i
<<
"
\"
"
<<
" -> "
;
out
<<
"
\t\"
"
<<
l
d
.
name
<<
"_"
<<
i
<<
"
\"
"
<<
" -> "
;
else
else
out
<<
"
"
<<
"
\"
"
<<
it
->
secon
d
.
name
<<
"
\"
"
<<
" -> "
;
out
<<
"
\t\"
"
<<
l
d
.
name
<<
"
\"
"
<<
" -> "
;
if
(
allLayers
[
outId
]
==
-
1
)
// node
if
(
allLayers
[
outId
]
==
-
1
)
// node
out
<<
"
\"
"
<<
map
[
outId
].
name
<<
"
\"
"
<<
'\n'
;
{
CV_DbgAssert
(
map
.
find
(
outId
)
!=
map
.
end
());
out
<<
"
\"
"
<<
map
.
find
(
outId
)
->
second
.
name
<<
"
\"\n
"
;
}
else
// cluster
else
// cluster
out
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
outId
]
<<
"
\"
"
<<
'\n'
;
{
out
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
outId
]
<<
"
\"\n
"
;
}
}
}
}
}
else
if
(
it
->
first
==
skippedLayers
[
allLayers
[
it
->
first
]].
back
())
// edges from last layer in cluster
else
if
(
it
->
first
==
skippedLayers
[
allLayers
[
it
->
first
]].
back
())
// edges from last layer in cluster
{
{
for
(
int
i
=
0
;
i
<
it
->
secon
d
.
consumers
.
size
();
i
++
)
for
(
int
i
=
0
;
i
<
l
d
.
consumers
.
size
();
i
++
)
{
{
int
outId
=
it
->
second
.
consumers
[
i
].
lid
;
int
outId
=
ld
.
consumers
[
i
].
lid
;
if
(
allLayers
[
outId
]
==
-
1
)
{
// node
if
(
allLayers
[
outId
]
==
-
1
)
// node
out
<<
" "
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
it
->
first
]
<<
"
\"
"
<<
" -> "
;
{
out
<<
"
\"
"
<<
map
[
outId
].
name
<<
"
\"
"
<<
'\n'
;
CV_DbgAssert
(
map
.
find
(
outId
)
!=
map
.
end
());
out
<<
"
\t\"
"
<<
"cluster_"
<<
allLayers
[
it
->
first
]
<<
"
\"
"
<<
" -> "
;
out
<<
"
\"
"
<<
map
.
find
(
outId
)
->
second
.
name
<<
"
\"\n
"
;
}
}
else
if
(
allLayers
[
outId
]
!=
allLayers
[
it
->
first
])
{
// another cluster
else
if
(
allLayers
[
outId
]
!=
allLayers
[
it
->
first
])
{
// another cluster
out
<<
"
"
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
it
->
first
]
<<
"
\"
"
<<
" -> "
;
out
<<
"
\t
\"
"
<<
"cluster_"
<<
allLayers
[
it
->
first
]
<<
"
\"
"
<<
" -> "
;
out
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
outId
]
<<
"
\"
"
<<
'\n'
;
out
<<
"
\"
"
<<
"cluster_"
<<
allLayers
[
outId
]
<<
"
\"
\n
"
;
}
}
}
}
}
}
}
}
out
<<
"}"
;
out
<<
"}
\n
"
;
return
out
.
str
();
return
out
.
str
();
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment