Commit 31e77a3b authored by Andrey Kamaev's avatar Andrey Kamaev

Android native camera: added BGR output format; added methods to configure output frame size.

parent 1965b297
...@@ -392,9 +392,8 @@ enum ...@@ -392,9 +392,8 @@ enum
//supported by Android camera output formats //supported by Android camera output formats
enum enum
{ {
CV_CAP_ANDROID_COLOR_FRAME = 1, //TODO: check RGB or BGR? CV_CAP_ANDROID_COLOR_FRAME = 0, //BGR
CV_CAP_ANDROID_GREY_FRAME = 0, CV_CAP_ANDROID_GREY_FRAME = 1 //Y
CV_CAP_ANDROID_YUV_FRAME = 2
}; };
/* retrieve or set capture properties */ /* retrieve or set capture properties */
......
/*M/////////////////////////////////////////////////////////////////////////////////////// /*M///////////////////////////////////////////////////////////////////////////////////////
// //
// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
// //
...@@ -58,7 +58,7 @@ class HighguiAndroidCameraActivity; ...@@ -58,7 +58,7 @@ class HighguiAndroidCameraActivity;
class CvCapture_Android : public CvCapture class CvCapture_Android : public CvCapture
{ {
public: public:
CvCapture_Android(); CvCapture_Android(int);
virtual ~CvCapture_Android(); virtual ~CvCapture_Android();
virtual double getProperty(int propIdx); virtual double getProperty(int propIdx);
...@@ -81,35 +81,44 @@ protected: ...@@ -81,35 +81,44 @@ protected:
CameraActivity* m_activity; CameraActivity* m_activity;
//raw from camera
int m_width;
int m_height;
unsigned char *m_frameYUV420i;
unsigned char *m_frameYUV420inext;
void setFrame(const void* buffer, int bufferSize);
private: private:
bool m_isOpened; bool m_isOpened;
bool m_CameraParamsChanged;
//frames counter for statistics
int m_framesGrabbed;
OutputMap *m_frameYUV; //cached converted frames
OutputMap *m_frameYUVnext;
OutputMap m_frameGray; OutputMap m_frameGray;
OutputMap m_frameColor; OutputMap m_frameColor;
bool m_hasGray; bool m_hasGray;
bool m_hasColor; bool m_hasColor;
//synchronization
pthread_mutex_t m_nextFrameMutex; pthread_mutex_t m_nextFrameMutex;
pthread_cond_t m_nextFrameCond; pthread_cond_t m_nextFrameCond;
volatile bool m_waitingNextFrame; volatile bool m_waitingNextFrame;
int m_framesGrabbed; void prepareCacheForYUV420i(int width, int height);
static bool convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
static bool convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
friend class HighguiAndroidCameraActivity;
void onFrame(const void* buffer, int bufferSize); friend class HighguiAndroidCameraActivity;
void convertBufferToYUV(const void* buffer, int size, int width, int height);
static bool convertYUVToGrey(const cv::Mat& yuv, cv::Mat& resmat);
static bool convertYUVToColor(const cv::Mat& yuv, cv::Mat& resmat);
}; };
class HighguiAndroidCameraActivity : public CameraActivity class HighguiAndroidCameraActivity : public CameraActivity
{ {
public: public:
HighguiAndroidCameraActivity(CvCapture_Android* capture) HighguiAndroidCameraActivity(CvCapture_Android* capture)
{ {
m_capture = capture; m_capture = capture;
...@@ -118,13 +127,12 @@ class HighguiAndroidCameraActivity : public CameraActivity ...@@ -118,13 +127,12 @@ class HighguiAndroidCameraActivity : public CameraActivity
virtual bool onFrameBuffer(void* buffer, int bufferSize) virtual bool onFrameBuffer(void* buffer, int bufferSize)
{ {
LOGD("buffer addr:%p size:%d",buffer, bufferSize);
if(isConnected() && buffer != 0 && bufferSize > 0) if(isConnected() && buffer != 0 && bufferSize > 0)
{ {
m_framesReceived++; m_framesReceived++;
if (m_capture->m_waitingNextFrame) if (m_capture->m_waitingNextFrame)
{ {
m_capture->onFrame(buffer, bufferSize); m_capture->setFrame(buffer, bufferSize);
pthread_mutex_lock(&m_capture->m_nextFrameMutex); pthread_mutex_lock(&m_capture->m_nextFrameMutex);
m_capture->m_waitingNextFrame = false;//set flag that no more frames required at this moment m_capture->m_waitingNextFrame = false;//set flag that no more frames required at this moment
pthread_cond_broadcast(&m_capture->m_nextFrameCond); pthread_cond_broadcast(&m_capture->m_nextFrameCond);
...@@ -140,7 +148,7 @@ class HighguiAndroidCameraActivity : public CameraActivity ...@@ -140,7 +148,7 @@ class HighguiAndroidCameraActivity : public CameraActivity
LOGI("FRAMES received: %d grabbed: %d", m_framesReceived, m_capture->m_framesGrabbed); LOGI("FRAMES received: %d grabbed: %d", m_framesReceived, m_capture->m_framesGrabbed);
} }
private: private:
CvCapture_Android* m_capture; CvCapture_Android* m_capture;
int m_framesReceived; int m_framesReceived;
}; };
...@@ -154,37 +162,33 @@ IplImage* CvCapture_Android::OutputMap::getIplImagePtr() ...@@ -154,37 +162,33 @@ IplImage* CvCapture_Android::OutputMap::getIplImagePtr()
return &iplHeader; return &iplHeader;
} }
bool CvCapture_Android::isOpened() const CvCapture_Android::CvCapture_Android(int cameraId)
{
return m_isOpened;
}
CvCapture_Android::CvCapture_Android()
{ {
//defaults //defaults
m_width = 0;
m_height = 0;
m_activity = 0; m_activity = 0;
m_isOpened = false; m_isOpened = false;
m_frameYUV = 0; m_frameYUV420i = 0;
m_frameYUVnext = 0; m_frameYUV420inext = 0;
m_hasGray = false; m_hasGray = false;
m_hasColor = false; m_hasColor = false;
m_waitingNextFrame = false; m_waitingNextFrame = false;
m_framesGrabbed = 0; m_framesGrabbed = 0;
m_CameraParamsChanged = false;
//try connect to camera //try connect to camera
m_activity = new HighguiAndroidCameraActivity(this); m_activity = new HighguiAndroidCameraActivity(this);
if (m_activity == 0) return; if (m_activity == 0) return;
pthread_mutex_init(&m_nextFrameMutex, NULL); pthread_mutex_init(&m_nextFrameMutex, NULL);
pthread_cond_init (&m_nextFrameCond, NULL); pthread_cond_init (&m_nextFrameCond, NULL);
CameraActivity::ErrorCode errcode = m_activity->connect(); CameraActivity::ErrorCode errcode = m_activity->connect(cameraId);
if(errcode == CameraActivity::NO_ERROR) if(errcode == CameraActivity::NO_ERROR)
{
m_isOpened = true; m_isOpened = true;
m_frameYUV = new OutputMap();
m_frameYUVnext = new OutputMap();
}
else else
{ {
LOGE("Native_camera returned opening error: %d", errcode); LOGE("Native_camera returned opening error: %d", errcode);
...@@ -193,6 +197,11 @@ CvCapture_Android::CvCapture_Android() ...@@ -193,6 +197,11 @@ CvCapture_Android::CvCapture_Android()
} }
} }
bool CvCapture_Android::isOpened() const
{
return m_isOpened;
}
CvCapture_Android::~CvCapture_Android() CvCapture_Android::~CvCapture_Android()
{ {
if (m_activity) if (m_activity)
...@@ -201,11 +210,11 @@ CvCapture_Android::~CvCapture_Android() ...@@ -201,11 +210,11 @@ CvCapture_Android::~CvCapture_Android()
//m_activity->disconnect() will be automatically called inside destructor; //m_activity->disconnect() will be automatically called inside destructor;
delete m_activity; delete m_activity;
delete m_frameYUV; delete m_frameYUV420i;
delete m_frameYUVnext; delete m_frameYUV420inext;
m_activity = 0; m_activity = 0;
m_frameYUV = 0; m_frameYUV420i = 0;
m_frameYUVnext = 0; m_frameYUV420inext = 0;
pthread_mutex_destroy(&m_nextFrameMutex); pthread_mutex_destroy(&m_nextFrameMutex);
pthread_cond_destroy(&m_nextFrameCond); pthread_cond_destroy(&m_nextFrameCond);
...@@ -217,11 +226,11 @@ double CvCapture_Android::getProperty( int propIdx ) ...@@ -217,11 +226,11 @@ double CvCapture_Android::getProperty( int propIdx )
switch ( propIdx ) switch ( propIdx )
{ {
case CV_CAP_PROP_FRAME_WIDTH: case CV_CAP_PROP_FRAME_WIDTH:
return (double)CameraActivity::getFrameWidth(); return (double)m_activity->getFrameWidth();
case CV_CAP_PROP_FRAME_HEIGHT: case CV_CAP_PROP_FRAME_HEIGHT:
return (double)CameraActivity::getFrameHeight(); return (double)m_activity->getFrameHeight();
default: default:
CV_Error( CV_StsError, "Failed attempt to GET unsupported camera property." ); CV_Error( CV_StsOutOfRange, "Failed attempt to GET unsupported camera property." );
break; break;
} }
return -1.0; return -1.0;
...@@ -234,10 +243,17 @@ bool CvCapture_Android::setProperty( int propIdx, double propValue ) ...@@ -234,10 +243,17 @@ bool CvCapture_Android::setProperty( int propIdx, double propValue )
{ {
switch ( propIdx ) switch ( propIdx )
{ {
case CV_CAP_PROP_FRAME_WIDTH:
m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEWIDTH, propValue);
break;
case CV_CAP_PROP_FRAME_HEIGHT:
m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEHEIGHT, propValue);
break;
default: default:
CV_Error( CV_StsError, "Failed attempt to SET unsupported camera property." ); CV_Error( CV_StsOutOfRange, "Failed attempt to SET unsupported camera property." );
break; break;
} }
m_CameraParamsChanged = true;
} }
return res; return res;
...@@ -249,53 +265,41 @@ bool CvCapture_Android::grabFrame() ...@@ -249,53 +265,41 @@ bool CvCapture_Android::grabFrame()
return false; return false;
pthread_mutex_lock(&m_nextFrameMutex); pthread_mutex_lock(&m_nextFrameMutex);
if (m_CameraParamsChanged)
{
m_activity->applyProperties();
m_CameraParamsChanged = false;
}
m_waitingNextFrame = true; m_waitingNextFrame = true;
pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex); pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex);
pthread_mutex_unlock(&m_nextFrameMutex); pthread_mutex_unlock(&m_nextFrameMutex);
m_framesGrabbed++; m_framesGrabbed++;
return true; return true;
} }
void CvCapture_Android::onFrame(const void* buffer, int bufferSize)
{
LOGD("Buffer available: %p + %d", buffer, bufferSize);
convertBufferToYUV(buffer, bufferSize, CameraActivity::getFrameWidth(), CameraActivity::getFrameHeight());
//swap current and new frames
OutputMap* tmp = m_frameYUV;
m_frameYUV = m_frameYUVnext;
m_frameYUVnext = tmp;
//discard cached frames
m_hasGray = false;
m_hasColor = false;
}
IplImage* CvCapture_Android::retrieveFrame( int outputType ) IplImage* CvCapture_Android::retrieveFrame( int outputType )
{ {
IplImage* image = 0; IplImage* image = 0;
if (0 != m_frameYUV && !m_frameYUV->mat.empty()) if (0 != m_frameYUV420i)
{ {
switch(outputType) switch(outputType)
{ {
case CV_CAP_ANDROID_YUV_FRAME: case CV_CAP_ANDROID_COLOR_FRAME:
image = m_frameYUV->getIplImagePtr(); if (!m_hasColor)
if (!(m_hasColor = convertYUV420i2BGR888(m_width, m_height, m_frameYUV420i, m_frameColor.mat)))
return 0;
image = m_frameColor.getIplImagePtr();
break; break;
case CV_CAP_ANDROID_GREY_FRAME: case CV_CAP_ANDROID_GREY_FRAME:
if (!m_hasGray) if (!m_hasGray)
if (!(m_hasGray = convertYUVToGrey(m_frameYUV->mat, m_frameGray.mat))) if (!(m_hasGray = convertYUV420i2Grey(m_width, m_height, m_frameYUV420i, m_frameGray.mat)))
image = 0; return 0;
image = m_frameGray.getIplImagePtr(); image = m_frameGray.getIplImagePtr();
break; break;
case CV_CAP_ANDROID_COLOR_FRAME:
if (!m_hasColor)
if (!(m_hasColor = convertYUVToColor(m_frameYUV->mat, m_frameColor.mat)))
image = 0;
image = m_frameColor.getIplImagePtr();
break;
default: default:
LOGE("Unsupported frame output format: %d", outputType); LOGE("Unsupported frame output format: %d", outputType);
CV_Error( CV_StsOutOfRange, "Output frame format is not supported." );
image = 0; image = 0;
break; break;
} }
...@@ -303,46 +307,147 @@ IplImage* CvCapture_Android::retrieveFrame( int outputType ) ...@@ -303,46 +307,147 @@ IplImage* CvCapture_Android::retrieveFrame( int outputType )
return image; return image;
} }
void CvCapture_Android::setFrame(const void* buffer, int bufferSize)
void CvCapture_Android::convertBufferToYUV(const void* buffer, int size, int width, int height)
{ {
cv::Size buffSize(width, height + (height / 2)); int width = m_activity->getFrameWidth();
if (buffSize.area() != size) int height = m_activity->getFrameHeight();
{ int expectedSize = (width * height * 3) >> 1;
LOGE("ERROR convertBufferToYuv_Mat: width=%d, height=%d, buffSize=%d x %d, buffSize.area()=%d, size=%d",
width, height, buffSize.width, buffSize.height, buffSize.area(), size);
if ( expectedSize != bufferSize)
{
LOGE("ERROR reading YUV420i buffer: width=%d, height=%d, size=%d, receivedSize=%d", width, height, expectedSize, bufferSize);
return; return;
} }
m_frameYUVnext->mat.create(buffSize, CV_8UC1); //allocate memery if needed
uchar* matBuff = m_frameYUVnext->mat.ptr<uchar> (0); prepareCacheForYUV420i(width, height);
memcpy(matBuff, buffer, size);
//copy data
memcpy(m_frameYUV420inext, buffer, bufferSize);
//swap current and new frames
unsigned char* tmp = m_frameYUV420i;
m_frameYUV420i = m_frameYUV420inext;
m_frameYUV420inext = tmp;
//discard cached frames
m_hasGray = false;
m_hasColor = false;
}
void CvCapture_Android::prepareCacheForYUV420i(int width, int height)
{
if (width != m_width || height != m_height)
{
m_width = width;
m_height = height;
unsigned char *tmp = m_frameYUV420inext;
m_frameYUV420inext = new unsigned char [width * height * 3 / 2];
delete[] tmp;
tmp = m_frameYUV420i;
m_frameYUV420i = new unsigned char [width * height * 3 / 2];
delete[] tmp;
}
} }
bool CvCapture_Android::convertYUVToGrey(const cv::Mat& yuv, cv::Mat& resmat) inline unsigned char clamp(int value)
{ {
if (yuv.empty()) if (value <= 0)
return false; return 0;
if (value >= 255)
return (unsigned char)255;
return (unsigned char)value;
}
resmat = yuv(cv::Range(0, yuv.rows * (2.0f / 3)), cv::Range::all());
bool CvCapture_Android::convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat)
{
if (yuv == 0) return false;
resmat.create(width, height, CV_8UC1);
unsigned char* matBuff = resmat.ptr<unsigned char> (0);
memcpy(matBuff, yuv, width * height);
return !resmat.empty(); return !resmat.empty();
} }
bool CvCapture_Android::convertYUVToColor(const cv::Mat& yuv, cv::Mat& resmat) bool CvCapture_Android::convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat)
{ {
if (yuv.empty()) if (yuv == 0) return false;
return false; CV_Assert(width % 2 == 0 && height % 2 == 0);
resmat.create(height, width, CV_8UC3);
unsigned char* y1 = (unsigned char*)yuv;
unsigned char* uv = y1 + width * height;
//B = 1.164(Y - 16) + 2.018(U - 128)
//G = 1.164(Y - 16) - 0.813(V - 128) - 0.391(U - 128)
//R = 1.164(Y - 16) + 1.596(V - 128)
for (int j = 0; j < height; j+=2, y1+=width*2, uv+=width)
{
unsigned char* row1 = resmat.ptr<unsigned char>(j);
unsigned char* row2 = resmat.ptr<unsigned char>(j+1);
unsigned char* y2 = y1 + width;
for(int i = 0; i < width; i+=2,row1+=6,row2+=6)
{
// unsigned char cr = uv[i];
// unsigned char cb = uv[i+1];
// row1[0] = y1[i];
// row1[1] = cr;
// row1[2] = cb;
// row1[3] = y1[i+1];
// row1[4] = cr;
// row1[5] = cb;
// row2[0] = y2[i];
// row2[1] = cr;
// row2[2] = cb;
// row2[3] = y2[i+1];
// row2[4] = cr;
// row2[5] = cb;
int cr = uv[i] - 128;
int cb = uv[i+1] - 128;
int ruv = 409 * cr + 128;
int guv = 128 - 100 * cb - 208 * cr;
int buv = 516 * cb + 128;
int y00 = (y1[i] - 16) * 298;
row1[0] = clamp((y00 + buv) >> 8);
row1[1] = clamp((y00 + guv) >> 8);
row1[2] = clamp((y00 + ruv) >> 8);
int y01 = (y1[i+1] - 16) * 298;
row1[3] = clamp((y01 + buv) >> 8);
row1[4] = clamp((y01 + guv) >> 8);
row1[5] = clamp((y01 + ruv) >> 8);
int y10 = (y2[i] - 16) * 298;
row2[0] = clamp((y10 + buv) >> 8);
row2[1] = clamp((y10 + guv) >> 8);
row2[2] = clamp((y10 + ruv) >> 8);
int y11 = (y2[i+1] - 16) * 298;
row2[3] = clamp((y11 + buv) >> 8);
row2[4] = clamp((y11 + guv) >> 8);
row2[5] = clamp((y11 + ruv) >> 8);
}
}
cv::cvtColor(yuv, resmat, CV_YUV2RGB);
return !resmat.empty(); return !resmat.empty();
} }
CvCapture* cvCreateCameraCapture_Android( int /*index*/ ) CvCapture* cvCreateCameraCapture_Android( int cameraId )
{ {
CvCapture_Android* capture = new CvCapture_Android(); CvCapture_Android* capture = new CvCapture_Android(cameraId);
if( capture->isOpened() ) if( capture->isOpened() )
return capture; return capture;
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment