Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in / Register
Toggle navigation
O
opencv
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Packages
Packages
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
submodule
opencv
Commits
8e1ccfae
Commit
8e1ccfae
authored
Sep 29, 2014
by
Ilya Lavrenov
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
cv::boxFilter
parent
1c491c42
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
255 additions
and
8 deletions
+255
-8
smooth.cpp
modules/imgproc/src/smooth.cpp
+255
-8
No files found.
modules/imgproc/src/smooth.cpp
View file @
8e1ccfae
...
...
@@ -132,8 +132,8 @@ struct ColumnSum :
SUM
=
&
sum
[
0
];
if
(
sumCount
==
0
)
{
for
(
i
=
0
;
i
<
width
;
i
++
)
SUM
[
i
]
=
0
;
memset
((
void
*
)
SUM
,
0
,
width
*
sizeof
(
ST
));
for
(
;
sumCount
<
ksize
-
1
;
sumCount
++
,
src
++
)
{
const
ST
*
Sp
=
(
const
ST
*
)
src
[
0
];
...
...
@@ -247,13 +247,16 @@ struct ColumnSum<int, uchar> :
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
4
;
i
+=
4
)
for
(
;
i
<
=
width
-
4
;
i
+=
4
)
{
__m128i
_sum
=
_mm_loadu_si128
((
const
__m128i
*
)(
SUM
+
i
));
__m128i
_sp
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sp
+
i
));
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_add_epi32
(
_sum
,
_sp
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
4
;
i
+=
4
)
vst1q_s32
(
SUM
+
i
,
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
)));
#endif
for
(
;
i
<
width
;
i
++
)
SUM
[
i
]
+=
Sp
[
i
];
...
...
@@ -277,7 +280,7 @@ struct ColumnSum<int, uchar> :
if
(
haveSSE2
)
{
const
__m128
scale4
=
_mm_set1_ps
((
float
)
_scale
);
for
(
;
i
<
width
-
8
;
i
+=
8
)
for
(
;
i
<
=
width
-
8
;
i
+=
8
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
__m128i
_sm1
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
+
4
));
...
...
@@ -298,6 +301,22 @@ struct ColumnSum<int, uchar> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
+
4
),
_mm_sub_epi32
(
_s01
,
_sm1
));
}
}
#elif CV_NEON
float32x4_t
v_scale
=
vdupq_n_f32
((
float
)
_scale
);
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
uint32x4_t
v_s0d
=
cv_vrndq_u32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s0
),
v_scale
));
uint32x4_t
v_s01d
=
cv_vrndq_u32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s01
),
v_scale
));
uint16x8_t
v_dst
=
vcombine_u16
(
vqmovn_u32
(
v_s0d
),
vqmovn_u32
(
v_s01d
));
vst1_u8
(
D
+
i
,
vqmovn_u16
(
v_dst
));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
{
...
...
@@ -312,7 +331,7 @@ struct ColumnSum<int, uchar> :
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
8
;
i
+=
8
)
for
(
;
i
<
=
width
-
8
;
i
+=
8
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
__m128i
_sm1
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
+
4
));
...
...
@@ -330,6 +349,18 @@ struct ColumnSum<int, uchar> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
+
4
),
_mm_sub_epi32
(
_s01
,
_sm1
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
uint16x8_t
v_dst
=
vcombine_u16
(
vqmovun_s32
(
v_s0
),
vqmovun_s32
(
v_s01
));
vst1_u8
(
D
+
i
,
vqmovn_u16
(
v_dst
));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
...
...
@@ -390,13 +421,16 @@ struct ColumnSum<int, short> :
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
4
;
i
+=
4
)
for
(
;
i
<
=
width
-
4
;
i
+=
4
)
{
__m128i
_sum
=
_mm_loadu_si128
((
const
__m128i
*
)(
SUM
+
i
));
__m128i
_sp
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sp
+
i
));
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_add_epi32
(
_sum
,
_sp
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
4
;
i
+=
4
)
vst1q_s32
(
SUM
+
i
,
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
)));
#endif
for
(
;
i
<
width
;
i
++
)
SUM
[
i
]
+=
Sp
[
i
];
...
...
@@ -420,7 +454,7 @@ struct ColumnSum<int, short> :
if
(
haveSSE2
)
{
const
__m128
scale4
=
_mm_set1_ps
((
float
)
_scale
);
for
(
;
i
<
width
-
8
;
i
+=
8
)
for
(
;
i
<
=
width
-
8
;
i
+=
8
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
__m128i
_sm1
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
+
4
));
...
...
@@ -439,6 +473,20 @@ struct ColumnSum<int, short> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
+
4
),
_mm_sub_epi32
(
_s01
,
_sm1
));
}
}
#elif CV_NEON
float32x4_t
v_scale
=
vdupq_n_f32
((
float
)
_scale
);
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
int32x4_t
v_s0d
=
cv_vrndq_s32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s0
),
v_scale
));
int32x4_t
v_s01d
=
cv_vrndq_s32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s01
),
v_scale
));
vst1q_s16
(
D
+
i
,
vcombine_s16
(
vqmovn_s32
(
v_s0d
),
vqmovn_s32
(
v_s01d
)));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
{
...
...
@@ -453,7 +501,7 @@ struct ColumnSum<int, short> :
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
8
;
i
+=
8
)
for
(
;
i
<
=
width
-
8
;
i
+=
8
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
...
...
@@ -470,6 +518,17 @@ struct ColumnSum<int, short> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
+
4
),
_mm_sub_epi32
(
_s01
,
_sm1
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
vst1q_s16
(
D
+
i
,
vcombine_s16
(
vqmovn_s32
(
v_s0
),
vqmovn_s32
(
v_s01
)));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
...
...
@@ -537,6 +596,9 @@ struct ColumnSum<int, ushort> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_add_epi32
(
_sum
,
_sp
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
4
;
i
+=
4
)
vst1q_s32
(
SUM
+
i
,
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
)));
#endif
for
(
;
i
<
width
;
i
++
)
SUM
[
i
]
+=
Sp
[
i
];
...
...
@@ -578,6 +640,20 @@ struct ColumnSum<int, ushort> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_sub_epi32
(
_s0
,
_sm
));
}
}
#elif CV_NEON
float32x4_t
v_scale
=
vdupq_n_f32
((
float
)
_scale
);
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
uint32x4_t
v_s0d
=
cv_vrndq_u32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s0
),
v_scale
));
uint32x4_t
v_s01d
=
cv_vrndq_u32_f32
(
vmulq_f32
(
vcvtq_f32_s32
(
v_s01
),
v_scale
));
vst1q_u16
(
D
+
i
,
vcombine_u16
(
vqmovn_u32
(
v_s0d
),
vqmovn_u32
(
v_s01d
)));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
{
...
...
@@ -608,6 +684,17 @@ struct ColumnSum<int, ushort> :
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_sub_epi32
(
_s0
,
_sm
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
vst1q_u16
(
D
+
i
,
vcombine_u16
(
vqmovun_s32
(
v_s0
),
vqmovun_s32
(
v_s01
)));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
...
...
@@ -626,6 +713,166 @@ struct ColumnSum<int, ushort> :
std
::
vector
<
int
>
sum
;
};
template
<>
struct
ColumnSum
<
int
,
float
>
:
public
BaseColumnFilter
{
ColumnSum
(
int
_ksize
,
int
_anchor
,
double
_scale
)
:
BaseColumnFilter
()
{
ksize
=
_ksize
;
anchor
=
_anchor
;
scale
=
_scale
;
sumCount
=
0
;
}
virtual
void
reset
()
{
sumCount
=
0
;
}
virtual
void
operator
()(
const
uchar
**
src
,
uchar
*
dst
,
int
dststep
,
int
count
,
int
width
)
{
int
i
;
int
*
SUM
;
bool
haveScale
=
scale
!=
1
;
double
_scale
=
scale
;
#if CV_SSE2
bool
haveSSE2
=
checkHardwareSupport
(
CV_CPU_SSE2
);
#endif
if
(
width
!=
(
int
)
sum
.
size
()
)
{
sum
.
resize
(
width
);
sumCount
=
0
;
}
SUM
=
&
sum
[
0
];
if
(
sumCount
==
0
)
{
memset
((
void
*
)
SUM
,
0
,
sizeof
(
int
)
*
width
);
for
(
;
sumCount
<
ksize
-
1
;
sumCount
++
,
src
++
)
{
const
int
*
Sp
=
(
const
int
*
)
src
[
0
];
i
=
0
;
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
4
;
i
+=
4
)
{
__m128i
_sum
=
_mm_loadu_si128
((
const
__m128i
*
)(
SUM
+
i
));
__m128i
_sp
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sp
+
i
));
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_add_epi32
(
_sum
,
_sp
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
4
;
i
+=
4
)
vst1q_s32
(
SUM
+
i
,
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
)));
#endif
for
(
;
i
<
width
;
i
++
)
SUM
[
i
]
+=
Sp
[
i
];
}
}
else
{
CV_Assert
(
sumCount
==
ksize
-
1
);
src
+=
ksize
-
1
;
}
for
(
;
count
--
;
src
++
)
{
const
int
*
Sp
=
(
const
int
*
)
src
[
0
];
const
int
*
Sm
=
(
const
int
*
)
src
[
1
-
ksize
];
float
*
D
=
(
float
*
)
dst
;
if
(
haveScale
)
{
i
=
0
;
#if CV_SSE2
if
(
haveSSE2
)
{
const
__m128
scale4
=
_mm_set1_ps
((
float
)
_scale
);
for
(
;
i
<
width
-
4
;
i
+=
4
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
__m128i
_s0
=
_mm_add_epi32
(
_mm_loadu_si128
((
const
__m128i
*
)(
SUM
+
i
)),
_mm_loadu_si128
((
const
__m128i
*
)(
Sp
+
i
)));
_mm_storeu_ps
(
D
+
i
,
_mm_mul_ps
(
scale4
,
_mm_cvtepi32_ps
(
_s0
)));
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_sub_epi32
(
_s0
,
_sm
));
}
}
#elif CV_NEON
float32x4_t
v_scale
=
vdupq_n_f32
((
float
)
_scale
);
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
vst1q_f32
(
D
+
i
,
vmulq_f32
(
vcvtq_f32_s32
(
v_s0
),
v_scale
));
vst1q_f32
(
D
+
i
+
4
,
vmulq_f32
(
vcvtq_f32_s32
(
v_s01
),
v_scale
));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
{
int
s0
=
SUM
[
i
]
+
Sp
[
i
];
D
[
i
]
=
(
float
)(
s0
*
_scale
);
SUM
[
i
]
=
s0
-
Sm
[
i
];
}
}
else
{
i
=
0
;
#if CV_SSE2
if
(
haveSSE2
)
{
for
(
;
i
<
width
-
4
;
i
+=
4
)
{
__m128i
_sm
=
_mm_loadu_si128
((
const
__m128i
*
)(
Sm
+
i
));
__m128i
_s0
=
_mm_add_epi32
(
_mm_loadu_si128
((
const
__m128i
*
)(
SUM
+
i
)),
_mm_loadu_si128
((
const
__m128i
*
)(
Sp
+
i
)));
_mm_storeu_ps
(
D
+
i
,
_mm_cvtepi32_ps
(
_s0
));
_mm_storeu_si128
((
__m128i
*
)(
SUM
+
i
),
_mm_sub_epi32
(
_s0
,
_sm
));
}
}
#elif CV_NEON
for
(
;
i
<=
width
-
8
;
i
+=
8
)
{
int32x4_t
v_s0
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
),
vld1q_s32
(
Sp
+
i
));
int32x4_t
v_s01
=
vaddq_s32
(
vld1q_s32
(
SUM
+
i
+
4
),
vld1q_s32
(
Sp
+
i
+
4
));
vst1q_f32
(
D
+
i
,
vcvtq_f32_s32
(
v_s0
));
vst1q_f32
(
D
+
i
+
4
,
vcvtq_f32_s32
(
v_s01
));
vst1q_s32
(
SUM
+
i
,
vsubq_s32
(
v_s0
,
vld1q_s32
(
Sm
+
i
)));
vst1q_s32
(
SUM
+
i
+
4
,
vsubq_s32
(
v_s01
,
vld1q_s32
(
Sm
+
i
+
4
)));
}
#endif
for
(
;
i
<
width
;
i
++
)
{
int
s0
=
SUM
[
i
]
+
Sp
[
i
];
D
[
i
]
=
(
float
)(
s0
);
SUM
[
i
]
=
s0
-
Sm
[
i
];
}
}
dst
+=
dststep
;
}
}
double
scale
;
int
sumCount
;
std
::
vector
<
int
>
sum
;
};
#ifdef HAVE_OPENCL
#define DIVUP(total, grain) ((total + grain - 1) / (grain))
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment