Commit 58295b07 authored by Alexander Alekhin's avatar Alexander Alekhin

Merge pull request #2356 from paroj:rapid

parents 65abc709 d3d5f2ad
set(the_description "rapid - silhouette based 3D object tracking")
ocv_define_module(rapid opencv_core opencv_imgproc opencv_calib3d WRAP python)
@inproceedings{harris1990rapid,
title={RAPID-a video rate object tracker.},
author={Harris, Chris and Stennett, Carl},
booktitle={BMVC},
pages={1--6},
year={1990}
}
@article{drummond2002real,
title={Real-time visual tracking of complex structures},
author={Drummond, Tom and Cipolla, Roberto},
journal={IEEE Transactions on pattern analysis and machine intelligence},
volume={24},
number={7},
pages={932--946},
year={2002},
publisher={IEEE}
}
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#ifndef OPENCV_RAPID_HPP_
#define OPENCV_RAPID_HPP_
#include <opencv2/core.hpp>
#include <opencv2/imgproc.hpp>
/**
@defgroup rapid silhouette based 3D object tracking
implements "RAPID-a video rate object tracker" @cite harris1990rapid with the dynamic control point extraction of @cite drummond2002real
*/
namespace cv
{
namespace rapid
{
//! @addtogroup rapid
//! @{
/**
* Debug draw markers of matched correspondences onto a lineBundle
* @param bundle the lineBundle
* @param srcLocations the according source locations
* @param newLocations matched source locations
* @param colors colors for the markers. Defaults to white.
*/
CV_EXPORTS_W void drawCorrespondencies(InputOutputArray bundle, InputArray srcLocations,
InputArray newLocations, InputArray colors = noArray());
/**
* Debug draw search lines onto an image
* @param img the output image
* @param locations the source locations of a line bundle
* @param color the line color
*/
CV_EXPORTS_W void drawSearchLines(InputOutputArray img, InputArray locations, const Scalar& color);
/**
* Draw a wireframe of a triangle mesh
* @param img the output image
* @param pts2d the 2d points obtained by @ref projectPoints
* @param tris triangle face connectivity
* @param color line color
* @param type line type. See @ref LineTypes.
* @param cullBackface enable back-face culling based on CCW order
*/
CV_EXPORTS_W void drawWireframe(InputOutputArray img, InputArray pts2d, InputArray tris,
const Scalar& color, int type = LINE_8, bool cullBackface = false);
/**
* Extract control points from the projected silhouette of a mesh
*
* see @cite drummond2002real Sec 2.1, Step b
* @param num number of control points
* @param len search radius (used to restrict the ROI)
* @param pts3d the 3D points of the mesh
* @param rvec rotation between mesh and camera
* @param tvec translation between mesh and camera
* @param K camera intrinsic
* @param imsize size of the video frame
* @param tris triangle face connectivity
* @param ctl2d the 2D locations of the control points
* @param ctl3d matching 3D points of the mesh
*/
CV_EXPORTS_W void extractControlPoints(int num, int len, InputArray pts3d, InputArray rvec, InputArray tvec,
InputArray K, const Size& imsize, InputArray tris, OutputArray ctl2d,
OutputArray ctl3d);
/**
* Extract the line bundle from an image
* @param len the search radius. The bundle will have `2*len + 1` columns.
* @param ctl2d the search lines will be centered at this points and orthogonal to the contour defined by
* them. The bundle will have as many rows.
* @param img the image to read the pixel intensities values from
* @param bundle line bundle image with size `ctl2d.rows() x (2 * len + 1)` and the same type as @p img
* @param srcLocations the source pixel locations of @p bundle in @p img as CV_16SC2
*/
CV_EXPORTS_W void extractLineBundle(int len, InputArray ctl2d, InputArray img, OutputArray bundle,
OutputArray srcLocations);
/**
* Find corresponding image locations by searching for a maximal sobel edge along the search line (a single
* row in the bundle)
* @param bundle the line bundle
* @param srcLocations the according source image location
* @param newLocations image locations with maximal edge along the search line
* @param response the sobel response for the selected point
*/
CV_EXPORTS_W void findCorrespondencies(InputArray bundle, InputArray srcLocations, OutputArray newLocations,
OutputArray response = noArray());
/**
* Filter corresponding 2d and 3d points based on mask
* @param pts2d 2d points
* @param pts3d 3d points
* @param mask mask containing non-zero values for the elements to be retained
*/
CV_EXPORTS_W void filterCorrespondencies(InputOutputArray pts2d, InputOutputArray pts3d, InputArray mask);
/**
* High level function to execute a single rapid @cite harris1990rapid iteration
*
* 1. @ref extractControlPoints
* 2. @ref extractLineBundle
* 3. @ref findCorrespondencies
* 4. @ref filterCorrespondencies
* 5. @ref solvePnPRefineLM
*
* @param img the video frame
* @param num number of search lines
* @param len search line radius
* @param pts3d the 3D points of the mesh
* @param tris triangle face connectivity
* @param K camera matrix
* @param rvec rotation between mesh and camera. Input values are used as an initial solution.
* @param tvec translation between mesh and camera. Input values are used as an initial solution.
* @return ratio of search lines that could be extracted and matched
*/
CV_EXPORTS_W float rapid(InputArray img, int num, int len, InputArray pts3d, InputArray tris, InputArray K,
InputOutputArray rvec, InputOutputArray tvec);
//! @}
} /* namespace rapid */
} /* namespace cv */
#endif /* OPENCV_RAPID_HPP_ */
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#ifndef __OPENCV_PRECOMP_H__
#define __OPENCV_PRECOMP_H__
#include "opencv2/rapid.hpp"
#include <vector>
#include <opencv2/calib3d.hpp>
#endif
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#include "precomp.hpp"
namespace cv
{
namespace rapid
{
static std::vector<int> getSilhoutteVertices(const Size& imsize, const std::vector<Point>& contour,
const Mat_<Point2f>& pts2d)
{
// store indices
Mat_<int> img1(imsize, 0);
Rect img_rect({0, 0}, imsize);
for (int i = 0; i < pts2d.rows; i++) {
if (img_rect.contains(pts2d(i))) {
img1(pts2d(i)) = i + 1;
}
}
std::vector<int> v_idx;
// look up indices on contour
for (size_t i = 0; i < contour.size(); i++) {
if (int idx = img1(contour[i])) {
v_idx.push_back(idx - 1);
}
}
return v_idx;
}
class Contour3DSampler {
std::vector<int> idx; // indices of points on contour
std::vector<float> cum_dist; // prefix sum
Mat_<Point2f> ipts2d;
Mat_<Point3f> ipts3d;
float lambda;
int pos;
public:
float perimeter;
Contour3DSampler(const Mat_<Point2f>& pts2d, const Mat_<Point3f>& pts3d,
const std::vector<Point>& contour, const Size& imsize)
: ipts2d(pts2d), ipts3d(pts3d)
{
idx = getSilhoutteVertices(imsize, contour, pts2d);
CV_Assert(!idx.empty());
// close the loop
idx.push_back(idx[0]);
cum_dist.resize(ipts2d.rows);
perimeter = 0.0f;
for (size_t i = 1; i < idx.size(); i++) {
perimeter += (float)norm(pts2d(idx[i]) - pts2d(idx[i - 1]));
cum_dist[i] = perimeter;
}
pos = 0;
lambda = 0;
}
void advanceTo(float dist)
{
while (pos < int(cum_dist.size() - 1) && dist >= cum_dist[pos]) {
pos++;
}
lambda = (dist - cum_dist[pos - 1]) / (cum_dist[pos] - cum_dist[pos - 1]);
}
Point3f current3D() const { return (1 - lambda) * ipts3d(idx[pos - 1]) + lambda * ipts3d(idx[pos]); }
Point2f current2D() const { return (1 - lambda) * ipts2d(idx[pos - 1]) + lambda * ipts2d(idx[pos]); }
};
void drawWireframe(InputOutputArray img, InputArray _pts2d, InputArray _tris,
const Scalar& color, int type, bool cullBackface)
{
CV_Assert(_tris.getMat().checkVector(3, CV_32S) > 0);
CV_Assert(_pts2d.getMat().checkVector(2, CV_32F) > 0);
Mat_<Vec3i> tris = _tris.getMat();
Mat_<Point2f> pts2d = _pts2d.getMat();
for (int i = 0; i < int(tris.total()); i++) {
const auto& idx = tris(i);
std::vector<Point> poly = {pts2d(idx[0]), pts2d(idx[1]), pts2d(idx[2])};
// skip back facing triangles
if (cullBackface && ((poly[2] - poly[0]).cross(poly[2] - poly[1]) >= 0))
continue;
polylines(img, poly, true, color, 1, type);
}
}
void drawSearchLines(InputOutputArray img, InputArray _locations, const Scalar& color)
{
Mat locations = _locations.getMat();
CV_CheckTypeEQ(_locations.type(), CV_16SC2, "Vec2s data type expected");
for (int i = 0; i < locations.rows; i++) {
Point pt1(locations.at<Vec2s>(i, 0));
Point pt2(locations.at<Vec2s>(i, locations.cols - 1));
line(img, pt1, pt2, color, 1);
}
}
static void sampleControlPoints(int num, Contour3DSampler& sampler, const Rect& roi, OutputArray _opts2d,
OutputArray _opts3d)
{
std::vector<Vec3f> opts3d;
opts3d.reserve(num);
std::vector<Vec2f> opts2d;
opts2d.reserve(num);
// sample at equal steps
float step = sampler.perimeter / num;
if (step == 0)
num = 0; // edge case -> skip loop
for (int i = 0; i < num; i++) {
sampler.advanceTo(step * i);
auto pt2d = sampler.current2D();
// skip points too close to border
if (!roi.contains(pt2d))
continue;
opts3d.push_back(sampler.current3D());
opts2d.push_back(pt2d);
}
Mat(opts3d).copyTo(_opts3d);
Mat(opts2d).copyTo(_opts2d);
}
void extractControlPoints(int num, int len, InputArray pts3d, InputArray rvec, InputArray tvec,
InputArray K, const Size& imsize, InputArray tris, OutputArray ctl2d,
OutputArray ctl3d)
{
CV_Assert(num);
Mat_<Point2f> pts2d(pts3d.rows(), 1);
projectPoints(pts3d, rvec, tvec, K, noArray(), pts2d);
Mat_<uchar> img(imsize, uchar(0));
drawWireframe(img, pts2d, tris.getMat(), 255, LINE_8, true);
// find contour
std::vector<std::vector<Point>> contours;
findContours(img, contours, RETR_EXTERNAL, CHAIN_APPROX_NONE);
CV_Assert(!contours.empty());
Contour3DSampler sampler(pts2d, pts3d.getMat(), contours[0], imsize);
Rect valid_roi(Point(len, len), imsize - Size(2 * len, 2 * len));
sampleControlPoints(num, sampler, valid_roi, ctl2d, ctl3d);
}
void extractLineBundle(int len, InputArray ctl2d, InputArray img, OutputArray bundle,
OutputArray srcLocations)
{
CV_Assert(len > 0);
Mat _img = img.getMat();
CV_Assert(ctl2d.getMat().checkVector(2, CV_32F) > 0);
Mat_<Point2f> contour = ctl2d.getMat();
const int N = contour.rows;
const int W = len * 2 + 1;
srcLocations.create(N, W, CV_16SC2);
Mat_<Vec2s> _srcLocations = srcLocations.getMat();
for (int i = 0; i < N; i++) {
// central difference
const Point2f diff = contour((i + 1) % N) - contour((i - 1 + N) % N);
Point2f n(normalize(Vec2f(-diff.y, diff.x))); // perpendicular to diff
// make it cover L pixels
n *= len / std::max(std::abs(n.x), std::abs(n.y));
LineIterator li(_img, contour(i) - n, contour(i) + n);
CV_DbgAssert(li.count == W);
for (int j = 0; j < li.count; j++, ++li) {
_srcLocations(i, j) = Vec2i(li.pos());
}
}
remap(img, bundle, srcLocations, noArray(),
INTER_NEAREST); // inter_nearest as we use integer locations
}
static void compute1DSobel(const Mat& src, Mat& dst)
{
CV_CheckDepthEQ(src.depth(), CV_8U, "only uchar images supported");
int channels = src.channels();
CV_Assert(channels == 1 || channels == 3);
dst.create(src.size(), CV_8U);
for (int i = 0; i < src.rows; i++) {
for (int j = 1; j < src.cols - 1; j++) {
// central difference kernel: [-1, 0, 1]
if (channels == 3) {
const Vec3s diff = Vec3s(src.at<Vec3b>(i, j + 1)) - Vec3s(src.at<Vec3b>(i, j - 1));
dst.at<uchar>(i, j) =
(uchar)std::max(std::max(std::abs(diff[0]), std::abs(diff[1])), std::abs(diff[2]));
} else {
dst.at<uchar>(i, j) = (uchar)std::abs(src.at<uchar>(i, j + 1) - src.at<uchar>(i, j - 1));
}
}
dst.at<uchar>(i, 0) = dst.at<uchar>(i, src.cols - 1) = 0; // border
}
}
void findCorrespondencies(InputArray bundle, InputArray _srcLocations, OutputArray _newLocations,
OutputArray _response)
{
CV_Assert(bundle.size() == _srcLocations.size());
CV_CheckTypeEQ(_srcLocations.type(), CV_16SC2, "Vec2s data type expected");
Mat_<uchar> sobel;
compute1DSobel(bundle.getMat(), sobel);
_newLocations.create(sobel.rows, 1, CV_16SC2);
Mat newLocations = _newLocations.getMat();
Mat srcLocations = _srcLocations.getMat();
Mat_<uchar> response;
if (_response.needed()) {
_response.create(sobel.rows, 1, CV_8U);
response = _response.getMat();
}
// sobel.cols = 2*len + 1
const int len = sobel.cols / 2;
const int ct = len + 1;
// find closest maximum to center
for (int i = 0; i < sobel.rows; i++) {
int pos = ct;
uchar mx = sobel.at<uchar>(i, ct);
for (int j = 0; j < len; j++) {
uchar right = sobel.at<uchar>(i, ct + j);
uchar left = sobel.at<uchar>(i, ct - j);
if (right > mx) {
mx = right;
pos = ct + j;
}
if (left > mx) {
mx = left;
pos = ct - j;
}
}
if (!response.empty())
response(i) = mx;
newLocations.at<Vec2s>(i, 0) = srcLocations.at<Vec2s>(i, pos);
}
}
void drawCorrespondencies(InputOutputArray _bundle, InputArray _srcLocations, InputArray _newLocations,
InputArray _colors)
{
CV_CheckTypeEQ(_srcLocations.type(), CV_16SC2, "Vec2s data type expected");
CV_CheckTypeEQ(_newLocations.type(), CV_16SC2, "Vec2s data type expected");
CV_Assert(_bundle.size() == _srcLocations.size());
CV_Assert(_colors.empty() || _colors.rows() == _srcLocations.rows());
Mat bundle = _bundle.getMat();
Mat_<Vec2s> srcLocations = _srcLocations.getMat();
Mat_<Vec2s> newLocations = _newLocations.getMat();
Mat_<Vec4d> colors = _colors.getMat();
for (int i = 0; i < bundle.rows; i++) {
const Vec2s& ref = newLocations(i);
for (int j = 1; j < bundle.cols - 1; j++) {
if (ref == srcLocations(i, j)) {
bundle(Rect(Point(j, i), Size(1, 1))) = colors.empty() ? Scalar::all(255) : colors(i);
}
}
}
}
void filterCorrespondencies(InputOutputArray _pts2d, InputOutputArray _pts3d, InputArray _mask)
{
CV_CheckTypeEQ(_mask.type(), CV_8UC1, "mask must be of uchar type");
CV_Assert(_pts2d.rows() == _pts3d.rows() && _pts2d.rows() == _mask.rows());
Mat pts2d = _pts2d.getMat();
Mat pts3d = _pts3d.getMat();
Mat_<uchar> mask = _mask.getMat();
Mat opts3d(0, 1, pts3d.type());
opts3d.reserve(mask.rows);
Mat opts2d(0, 1, pts2d.type());
opts2d.reserve(mask.rows);
for (int i = 0; i < mask.rows; i++) {
if (!mask(i))
continue;
opts2d.push_back(pts2d.row(i));
opts3d.push_back(pts3d.row(i));
}
Mat(opts3d).copyTo(_pts3d);
Mat(opts2d).copyTo(_pts2d);
}
float rapid(InputArray img, int num, int len, InputArray vtx, InputArray tris, InputArray K,
InputOutputArray rvec, InputOutputArray tvec)
{
CV_Assert(num >= 3);
Mat pts2d, pts3d, correspondencies;
extractControlPoints(num, len, vtx, rvec, tvec, K, img.size(), tris, pts2d, pts3d);
if (pts2d.empty())
return 0;
Mat lineBundle, imgLoc;
extractLineBundle(len, pts2d, img, lineBundle, imgLoc);
Mat response;
findCorrespondencies(lineBundle, imgLoc, correspondencies, response);
const uchar sobel_thresh = 20;
filterCorrespondencies(correspondencies, pts3d, response > sobel_thresh);
if (correspondencies.rows < 3)
return 0;
solvePnPRefineLM(pts3d, correspondencies, K, cv::noArray(), rvec, tvec);
return float(correspondencies.rows) / num;
}
} /* namespace rapid */
} /* namespace cv */
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#include "test_precomp.hpp"
CV_TEST_MAIN("cv")
namespace opencv_test { namespace {
TEST(CV_Rapid, rapid)
{
// a unit sized box
std::vector<Vec3f> vtx = {
{1, -1, -1}, {1, -1, 1}, {-1, -1, 1}, {-1, -1, -1}, {1, 1, -1}, {1, 1, 1}, {-1, 1, 1}, {-1, 1, -1},
};
std::vector<Vec3i> tris = {
{2, 4, 1}, {8, 6, 5}, {5, 2, 1}, {6, 3, 2}, {3, 8, 4}, {1, 8, 5},
{2, 3, 4}, {8, 7, 6}, {5, 6, 2}, {6, 7, 3}, {3, 7, 8}, {1, 4, 8},
};
Mat(tris) -= Scalar(1, 1, 1);
// camera setup
Size sz(1280, 720);
Mat K = getDefaultNewCameraMatrix(Matx33f::diag(Vec3f(800, 800, 1)), sz, true);
Vec3f trans = {0, 0, 5};
Vec3f rot = {0.7f, 0.6f, 0};
// draw something
Mat pts2d;
projectPoints(vtx, rot, trans, K, noArray(), pts2d);
Mat_<uchar> img(sz, uchar(0));
rapid::drawWireframe(img, pts2d, tris, Scalar(255), LINE_8);
// recover pose form different position
Vec3f t_init = Vec3f(0.1f, 0, 5);
for(int i = 0; i < 2; i++) // do two iteration
rapid::rapid(img, 100, 20, vtx, tris, K, rot, t_init);
// assert that it improved from init
ASSERT_LT(cv::norm(trans - t_init), 0.075);
}
}} // namespace
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#ifndef __OPENCV_TEST_PRECOMP_HPP__
#define __OPENCV_TEST_PRECOMP_HPP__
#include "opencv2/ts.hpp"
#include "opencv2/imgproc.hpp"
#include "opencv2/calib3d.hpp"
#include "opencv2/rapid.hpp"
#endif
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment