/*M/////////////////////////////////////////////////////////////////////////////////////// // // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // // By downloading, copying, installing or using the software you agree to this license. // If you do not agree to this license, do not download, install, // copy or use the software. // // // Intel License Agreement // For Open Source Computer Vision Library // // Copyright (C) 2000, Intel Corporation, all rights reserved. // Third party copyrights are property of their respective owners. // // Redistribution and use in source and binary forms, with or without modification, // are permitted provided that the following conditions are met: // // * Redistribution's of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // // * Redistribution's in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // // * The name of Intel Corporation may not be used to endorse or promote products // derived from this software without specific prior written permission. // // This software is provided by the copyright holders and contributors "as is" and // any express or implied warranties, including, but not limited to, the implied // warranties of merchantability and fitness for a particular purpose are disclaimed. // In no event shall the Intel Corporation or contributors be liable for any direct, // indirect, incidental, special, exemplary, or consequential damages // (including, but not limited to, procurement of substitute goods or services; // loss of use, data, or profits; or business interruption) however caused // and on any theory of liability, whether in contract, strict liability, // or tort (including negligence or otherwise) arising in any way out of // the use of this software, even if advised of the possibility of such damage. // //M*/ #include "precomp.hpp" #ifdef HAVE_ANDROID_NATIVE_CAMERA #include #include #include #include #undef LOG_TAG #undef LOGD #undef LOGE #undef LOGI #define LOG_TAG "OpenCV::camera" #define LOGD(...) ((void)__android_log_print(ANDROID_LOG_DEBUG, LOG_TAG, __VA_ARGS__)) #define LOGI(...) ((void)__android_log_print(ANDROID_LOG_INFO, LOG_TAG, __VA_ARGS__)) #define LOGE(...) ((void)__android_log_print(ANDROID_LOG_ERROR, LOG_TAG, __VA_ARGS__)) class HighguiAndroidCameraActivity; class CvCapture_Android : public CvCapture { public: CvCapture_Android(int); virtual ~CvCapture_Android(); virtual double getProperty(int propIdx); virtual bool setProperty(int probIdx, double propVal); virtual bool grabFrame(); virtual IplImage* retrieveFrame(int outputType); virtual int getCaptureDomain() { return CV_CAP_ANDROID; } bool isOpened() const; protected: struct OutputMap { public: cv::Mat mat; IplImage* getIplImagePtr(); private: IplImage iplHeader; }; CameraActivity* m_activity; //raw from camera int m_width; int m_height; cv::Mat m_frameYUV420; cv::Mat m_frameYUV420next; enum YUVformat { noformat = 0, yuv420sp, yvu420sp, yuvUnknown }; YUVformat m_frameFormat; void setFrame(const void* buffer, int bufferSize); private: bool m_isOpened; bool m_CameraParamsChanged; //frames counter for statistics int m_framesGrabbed; //cached converted frames OutputMap m_frameGray; OutputMap m_frameColor; bool m_hasGray; bool m_hasColor; enum CvCapture_Android_DataState { CVCAPTURE_ANDROID_STATE_NO_FRAME=0, CVCAPTURE_ANDROID_STATE_HAS_NEW_FRAME_UNGRABBED, CVCAPTURE_ANDROID_STATE_HAS_FRAME_GRABBED }; volatile CvCapture_Android_DataState m_dataState; //synchronization pthread_mutex_t m_nextFrameMutex; pthread_cond_t m_nextFrameCond; volatile bool m_waitingNextFrame; volatile bool m_shouldAutoGrab; void prepareCacheForYUV(int width, int height); bool convertYUV2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat); bool convertYUV2BGR(int width, int height, const unsigned char* yuv, cv::Mat& resmat, bool inRGBorder, bool withAlpha); friend class HighguiAndroidCameraActivity; }; class HighguiAndroidCameraActivity : public CameraActivity { public: HighguiAndroidCameraActivity(CvCapture_Android* capture) { m_capture = capture; m_framesReceived = 0; } virtual bool onFrameBuffer(void* buffer, int bufferSize) { if(isConnected() && buffer != 0 && bufferSize > 0) { m_framesReceived++; if (m_capture->m_waitingNextFrame || m_capture->m_shouldAutoGrab) { pthread_mutex_lock(&m_capture->m_nextFrameMutex); m_capture->setFrame(buffer, bufferSize); pthread_cond_broadcast(&m_capture->m_nextFrameCond); pthread_mutex_unlock(&m_capture->m_nextFrameMutex); } return true; } return false; } void LogFramesRate() { LOGI("FRAMES received: %d grabbed: %d", m_framesReceived, m_capture->m_framesGrabbed); } private: CvCapture_Android* m_capture; int m_framesReceived; }; IplImage* CvCapture_Android::OutputMap::getIplImagePtr() { if( mat.empty() ) return 0; iplHeader = IplImage(mat); return &iplHeader; } CvCapture_Android::CvCapture_Android(int cameraId) { //defaults m_width = 0; m_height = 0; m_activity = 0; m_isOpened = false; // m_frameYUV420 = 0; // m_frameYUV420next = 0; m_hasGray = false; m_hasColor = false; m_dataState = CVCAPTURE_ANDROID_STATE_NO_FRAME; m_waitingNextFrame = false; m_shouldAutoGrab = false; m_framesGrabbed = 0; m_CameraParamsChanged = false; m_frameFormat = noformat; //try connect to camera LOGD("CvCapture_Android::CvCapture_Android(%i)", cameraId); m_activity = new HighguiAndroidCameraActivity(this); if (m_activity == 0) return; pthread_mutex_init(&m_nextFrameMutex, NULL); pthread_cond_init (&m_nextFrameCond, NULL); CameraActivity::ErrorCode errcode = m_activity->connect(cameraId); if(errcode == CameraActivity::NO_ERROR) m_isOpened = true; else { LOGE("Native_camera returned opening error: %d", errcode); delete m_activity; m_activity = 0; } } bool CvCapture_Android::isOpened() const { return m_isOpened; } CvCapture_Android::~CvCapture_Android() { if (m_activity) { ((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate(); pthread_mutex_lock(&m_nextFrameMutex); // unsigned char *tmp1=m_frameYUV420; // unsigned char *tmp2=m_frameYUV420next; // m_frameYUV420 = 0; // m_frameYUV420next = 0; // delete tmp1; // delete tmp2; m_dataState=CVCAPTURE_ANDROID_STATE_NO_FRAME; pthread_cond_broadcast(&m_nextFrameCond); pthread_mutex_unlock(&m_nextFrameMutex); //m_activity->disconnect() will be automatically called inside destructor; delete m_activity; m_activity = 0; pthread_mutex_destroy(&m_nextFrameMutex); pthread_cond_destroy(&m_nextFrameCond); } } double CvCapture_Android::getProperty( int propIdx ) { switch ( propIdx ) { case CV_CAP_PROP_FRAME_WIDTH: return (double)m_activity->getFrameWidth(); case CV_CAP_PROP_FRAME_HEIGHT: return (double)m_activity->getFrameHeight(); case CV_CAP_PROP_SUPPORTED_PREVIEW_SIZES_STRING: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_SUPPORTED_PREVIEW_SIZES_STRING); case CV_CAP_PROP_PREVIEW_FORMAT: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_PREVIEW_FORMAT_STRING); case CV_CAP_PROP_FPS: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FPS); case CV_CAP_PROP_EXPOSURE: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_EXPOSURE); case CV_CAP_PROP_ANDROID_FLASH_MODE: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FLASH_MODE); case CV_CAP_PROP_ANDROID_FOCUS_MODE: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FOCUS_MODE); case CV_CAP_PROP_ANDROID_WHITE_BALANCE: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_WHITE_BALANCE); case CV_CAP_PROP_ANDROID_ANTIBANDING: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_ANTIBANDING); case CV_CAP_PROP_ANDROID_FOCAL_LENGTH: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FOCAL_LENGTH); case CV_CAP_PROP_ANDROID_FOCUS_DISTANCE_NEAR: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FOCUS_DISTANCE_NEAR); case CV_CAP_PROP_ANDROID_FOCUS_DISTANCE_OPTIMAL: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FOCUS_DISTANCE_OPTIMAL); case CV_CAP_PROP_ANDROID_FOCUS_DISTANCE_FAR: return (double)m_activity->getProperty(ANDROID_CAMERA_PROPERTY_FOCUS_DISTANCE_FAR); default: CV_Error( CV_StsOutOfRange, "Failed attempt to GET unsupported camera property." ); break; } return -1.0; } bool CvCapture_Android::setProperty( int propIdx, double propValue ) { bool res = false; if( isOpened() ) { switch ( propIdx ) { case CV_CAP_PROP_FRAME_WIDTH: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEWIDTH, propValue); break; case CV_CAP_PROP_FRAME_HEIGHT: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEHEIGHT, propValue); break; case CV_CAP_PROP_AUTOGRAB: m_shouldAutoGrab=(propValue != 0); break; case CV_CAP_PROP_EXPOSURE: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_EXPOSURE, propValue); break; case CV_CAP_PROP_ANDROID_FLASH_MODE: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FLASH_MODE, propValue); break; case CV_CAP_PROP_ANDROID_FOCUS_MODE: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FOCUS_MODE, propValue); break; case CV_CAP_PROP_ANDROID_WHITE_BALANCE: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_WHITE_BALANCE, propValue); break; case CV_CAP_PROP_ANDROID_ANTIBANDING: m_activity->setProperty(ANDROID_CAMERA_PROPERTY_ANTIBANDING, propValue); break; default: CV_Error( CV_StsOutOfRange, "Failed attempt to SET unsupported camera property." ); return false; } if (propIdx != CV_CAP_PROP_AUTOGRAB) {// property for highgui class CvCapture_Android only m_CameraParamsChanged = true; } res = true; } return res; } bool CvCapture_Android::grabFrame() { if( !isOpened() ) { LOGE("CvCapture_Android::grabFrame(): camera is not opened"); return false; } bool res=false; pthread_mutex_lock(&m_nextFrameMutex); if (m_CameraParamsChanged) { m_activity->applyProperties(); m_CameraParamsChanged = false; m_dataState = CVCAPTURE_ANDROID_STATE_NO_FRAME;//we will wait new frame } if (m_dataState != CVCAPTURE_ANDROID_STATE_HAS_NEW_FRAME_UNGRABBED) { m_waitingNextFrame = true; pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex); } if (m_dataState == CVCAPTURE_ANDROID_STATE_HAS_NEW_FRAME_UNGRABBED) { //LOGD("CvCapture_Android::grabFrame: get new frame"); //swap current and new frames cv::swap(m_frameYUV420, m_frameYUV420next); //discard cached frames m_hasGray = false; m_hasColor = false; m_dataState=CVCAPTURE_ANDROID_STATE_HAS_FRAME_GRABBED; m_framesGrabbed++; res=true; } else { LOGE("CvCapture_Android::grabFrame: NO new frame"); } int res_unlock=pthread_mutex_unlock(&m_nextFrameMutex); if (res_unlock) { LOGE("Error in CvCapture_Android::grabFrame: pthread_mutex_unlock returned %d --- probably, this object has been destroyed", res_unlock); return false; } return res; } IplImage* CvCapture_Android::retrieveFrame( int outputType ) { IplImage* image = NULL; cv::Mat m_frameYUV420_ref = m_frameYUV420; unsigned char *current_frameYUV420=m_frameYUV420_ref.ptr(); //Attention! all the operations in this function below should occupy less time than the period between two frames from camera if (NULL != current_frameYUV420) { if (m_frameFormat == noformat) { union {double prop; const char* name;} u; u.prop = getProperty(CV_CAP_PROP_PREVIEW_FORMAT); if (0 == strcmp(u.name, "yuv420sp")) m_frameFormat = yuv420sp; else if (0 == strcmp(u.name, "yvu420sp")) m_frameFormat = yvu420sp; else m_frameFormat = yuvUnknown; } switch(outputType) { case CV_CAP_ANDROID_GREY_FRAME: if (!m_hasGray) if (!(m_hasGray = convertYUV2Grey(m_width, m_height, current_frameYUV420, m_frameGray.mat))) return NULL; image = m_frameGray.getIplImagePtr(); break; case CV_CAP_ANDROID_COLOR_FRAME_BGR: case CV_CAP_ANDROID_COLOR_FRAME_RGB: if (!m_hasColor) if (!(m_hasColor = convertYUV2BGR(m_width, m_height, current_frameYUV420, m_frameColor.mat, outputType == CV_CAP_ANDROID_COLOR_FRAME_RGB, false))) return NULL; image = m_frameColor.getIplImagePtr(); break; case CV_CAP_ANDROID_COLOR_FRAME_BGRA: case CV_CAP_ANDROID_COLOR_FRAME_RGBA: if (!m_hasColor) if (!(m_hasColor = convertYUV2BGR(m_width, m_height, current_frameYUV420, m_frameColor.mat, outputType == CV_CAP_ANDROID_COLOR_FRAME_RGBA, true))) return NULL; image = m_frameColor.getIplImagePtr(); break; default: LOGE("Unsupported frame output format: %d", outputType); CV_Error( CV_StsOutOfRange, "Output frame format is not supported." ); image = NULL; break; } } return image; } //Attention: this method should be called inside pthread_mutex_lock(m_nextFrameMutex) only void CvCapture_Android::setFrame(const void* buffer, int bufferSize) { int width = m_activity->getFrameWidth(); int height = m_activity->getFrameHeight(); int expectedSize = (width * height * 3) >> 1; if ( expectedSize != bufferSize) { LOGE("ERROR reading YUV buffer: width=%d, height=%d, size=%d, receivedSize=%d", width, height, expectedSize, bufferSize); return; } //allocate memory if needed prepareCacheForYUV(width, height); //copy data cv::Mat m_frameYUV420next_ref = m_frameYUV420next; memcpy(m_frameYUV420next_ref.ptr(), buffer, bufferSize); // LOGD("CvCapture_Android::setFrame -- memcpy is done"); // ((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate(); m_dataState = CVCAPTURE_ANDROID_STATE_HAS_NEW_FRAME_UNGRABBED; m_waitingNextFrame = false;//set flag that no more frames required at this moment } //Attention: this method should be called inside pthread_mutex_lock(m_nextFrameMutex) only void CvCapture_Android::prepareCacheForYUV(int width, int height) { if (width != m_width || height != m_height) { LOGD("CvCapture_Android::prepareCacheForYUV: Changing size of buffers: from width=%d height=%d to width=%d height=%d", m_width, m_height, width, height); m_width = width; m_height = height; /* unsigned char *tmp = m_frameYUV420next; m_frameYUV420next = new unsigned char [width * height * 3 / 2]; if (tmp != NULL) { delete[] tmp; } tmp = m_frameYUV420; m_frameYUV420 = new unsigned char [width * height * 3 / 2]; if (tmp != NULL) { delete[] tmp; }*/ m_frameYUV420.create(height * 3 / 2, width, CV_8UC1); m_frameYUV420next.create(height * 3 / 2, width, CV_8UC1); } } bool CvCapture_Android::convertYUV2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat) { if (yuv == 0) return false; if (m_frameFormat != yuv420sp && m_frameFormat != yvu420sp) return false; #define ALWAYS_COPY_GRAY 0 #if ALWAYS_COPY_GRAY resmat.create(height, width, CV_8UC1); unsigned char* matBuff = resmat.ptr (0); memcpy(matBuff, yuv, width * height); #else resmat = cv::Mat(height, width, CV_8UC1, (void*)yuv); #endif return !resmat.empty(); } bool CvCapture_Android::convertYUV2BGR(int width, int height, const unsigned char* yuv, cv::Mat& resmat, bool inRGBorder, bool withAlpha) { if (yuv == 0) return false; if (m_frameFormat != yuv420sp && m_frameFormat != yvu420sp) return false; CV_Assert(width % 2 == 0 && height % 2 == 0); cv::Mat src(height*3/2, width, CV_8UC1, (void*)yuv); if (m_frameFormat == yuv420sp) cv::cvtColor(src, resmat, inRGBorder ? CV_YUV420sp2RGB : CV_YUV420sp2BGR, withAlpha ? 4 : 3); else if (m_frameFormat == yvu420sp) cv::cvtColor(src, resmat, inRGBorder ? CV_YUV2RGB_NV21 : CV_YUV2BGR_NV12, withAlpha ? 4 : 3); return !resmat.empty(); } CvCapture* cvCreateCameraCapture_Android( int cameraId ) { CvCapture_Android* capture = new CvCapture_Android(cameraId); if( capture->isOpened() ) return capture; delete capture; return 0; } #endif