/* * Copyright (C) Texas Instruments - http://www.ti.com/ * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /** * @file OMXCapture.cpp * * This file contains functionality for handling image capture. * */ #include "CameraHal.h" #include "OMXCameraAdapter.h" #include "ErrorUtils.h" namespace Ti { namespace Camera { status_t OMXCameraAdapter::setParametersCapture(const android::CameraParameters ¶ms, BaseCameraAdapter::AdapterState state) { status_t ret = NO_ERROR; const char *str = NULL; int w, h; OMX_COLOR_FORMATTYPE pixFormat; CodingMode codingMode = mCodingMode; const char *valstr = NULL; int varint = 0; OMX_TI_STEREOFRAMELAYOUTTYPE capFrmLayout; bool inCaptureState = false; LOG_FUNCTION_NAME; OMXCameraPortParameters *cap; cap = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; capFrmLayout = cap->mFrameLayoutType; setParamS3D(mCameraAdapterParameters.mImagePortIndex, params.get(TICameraParameters::KEY_S3D_CAP_FRAME_LAYOUT)); if (capFrmLayout != cap->mFrameLayoutType) { mPendingCaptureSettings |= SetFormat; } params.getPictureSize(&w, &h); if ( ( w != ( int ) cap->mWidth ) || ( h != ( int ) cap->mHeight ) ) { mPendingCaptureSettings |= SetFormat; } cap->mWidth = w; cap->mHeight = h; //TODO: Support more pixelformats //cap->mStride = 2; CAMHAL_LOGVB("Image: cap.mWidth = %d", (int)cap->mWidth); CAMHAL_LOGVB("Image: cap.mHeight = %d", (int)cap->mHeight); if ((valstr = params.getPictureFormat()) != NULL) { if (strcmp(valstr, android::CameraParameters::PIXEL_FORMAT_YUV422I) == 0) { CAMHAL_LOGDA("CbYCrY format selected"); pixFormat = OMX_COLOR_FormatCbYCrY; mPictureFormatFromClient = android::CameraParameters::PIXEL_FORMAT_YUV422I; } else if(strcmp(valstr, android::CameraParameters::PIXEL_FORMAT_YUV420SP) == 0) { CAMHAL_LOGDA("YUV420SP format selected"); pixFormat = OMX_COLOR_FormatYUV420SemiPlanar; mPictureFormatFromClient = android::CameraParameters::PIXEL_FORMAT_YUV420SP; } else if(strcmp(valstr, android::CameraParameters::PIXEL_FORMAT_RGB565) == 0) { CAMHAL_LOGDA("RGB565 format selected"); pixFormat = OMX_COLOR_Format16bitRGB565; mPictureFormatFromClient = android::CameraParameters::PIXEL_FORMAT_RGB565; } else if (strcmp(valstr, android::CameraParameters::PIXEL_FORMAT_JPEG) == 0) { CAMHAL_LOGDA("JPEG format selected"); pixFormat = OMX_COLOR_FormatUnused; codingMode = CodingJPEG; mPictureFormatFromClient = android::CameraParameters::PIXEL_FORMAT_JPEG; } else if (strcmp(valstr, TICameraParameters::PIXEL_FORMAT_JPS) == 0) { CAMHAL_LOGDA("JPS format selected"); pixFormat = OMX_COLOR_FormatUnused; codingMode = CodingJPS; mPictureFormatFromClient = TICameraParameters::PIXEL_FORMAT_JPS; } else if (strcmp(valstr, TICameraParameters::PIXEL_FORMAT_MPO) == 0) { CAMHAL_LOGDA("MPO format selected"); pixFormat = OMX_COLOR_FormatUnused; codingMode = CodingMPO; mPictureFormatFromClient = TICameraParameters::PIXEL_FORMAT_MPO; } else if (strcmp(valstr, android::CameraParameters::PIXEL_FORMAT_BAYER_RGGB) == 0) { CAMHAL_LOGDA("RAW Picture format selected"); pixFormat = OMX_COLOR_FormatRawBayer10bit; mPictureFormatFromClient = android::CameraParameters::PIXEL_FORMAT_BAYER_RGGB; } else { CAMHAL_LOGEA("Invalid format, JPEG format selected as default"); pixFormat = OMX_COLOR_FormatUnused; codingMode = CodingJPEG; mPictureFormatFromClient = NULL; } } else { CAMHAL_LOGEA("Picture format is NULL, defaulting to JPEG"); pixFormat = OMX_COLOR_FormatUnused; codingMode = CodingJPEG; mPictureFormatFromClient = NULL; } #ifdef CAMERAHAL_USE_RAW_IMAGE_SAVING mRawCapture = false; mYuvCapture = false; valstr = params.get(TICameraParameters::KEY_CAP_MODE); if ( (!valstr || strcmp(valstr, TICameraParameters::HIGH_QUALITY_MODE) == 0) && access(kRawImagesOutputDirPath, F_OK) != -1 ) { mRawCapture = true; } if (mRawCapture && (access(kYuvImagesOutputDirPath, F_OK) != -1)) { pixFormat = OMX_COLOR_FormatCbYCrY; mYuvCapture = true; } #endif // JPEG capture is not supported in video mode by OMX Camera // Set capture format to yuv422i...jpeg encode will // be done on A9 valstr = params.get(TICameraParameters::KEY_CAP_MODE); if ( (valstr && ( strcmp(valstr, (const char *) TICameraParameters::VIDEO_MODE) == 0 || strcmp(valstr, (const char *) TICameraParameters::VIDEO_MODE_HQ) == 0 ) ) && (pixFormat == OMX_COLOR_FormatUnused) ) { CAMHAL_LOGDA("Capturing in video mode...selecting yuv422i"); pixFormat = OMX_COLOR_FormatCbYCrY; } if (pixFormat != cap->mColorFormat || codingMode != mCodingMode) { mPendingCaptureSettings |= SetFormat; cap->mColorFormat = pixFormat; mCodingMode = codingMode; } #ifdef OMAP_ENHANCEMENT str = params.get(TICameraParameters::KEY_TEMP_BRACKETING); if ( ( str != NULL ) && ( strcmp(str, android::CameraParameters::TRUE) == 0 ) ) { if ( !mBracketingSet ) { mPendingCaptureSettings |= SetBurstExpBracket; } mBracketingSet = true; } else { if ( mBracketingSet ) { mPendingCaptureSettings |= SetBurstExpBracket; } mBracketingSet = false; } if ( (str = params.get(TICameraParameters::KEY_EXP_BRACKETING_RANGE)) != NULL ) { parseExpRange(str, mExposureBracketingValues, NULL, mExposureGainBracketingModes, EXP_BRACKET_RANGE, mExposureBracketingValidEntries); if (mCapMode == OMXCameraAdapter::CP_CAM) { mExposureBracketMode = OMX_BracketVectorShot; } else { mExposureBracketMode = OMX_BracketExposureRelativeInEV; } mPendingCaptureSettings |= SetBurstExpBracket; } else if ( (str = params.get(TICameraParameters::KEY_EXP_GAIN_BRACKETING_RANGE)) != NULL) { parseExpRange(str, mExposureBracketingValues, mExposureGainBracketingValues, mExposureGainBracketingModes, EXP_BRACKET_RANGE, mExposureBracketingValidEntries); if (mCapMode == OMXCameraAdapter::CP_CAM) { mExposureBracketMode = OMX_BracketVectorShot; } else { mExposureBracketMode = OMX_BracketExposureGainAbsolute; } mPendingCaptureSettings |= SetBurstExpBracket; } else { // always set queued shot config in CPCAM mode if (mCapMode == OMXCameraAdapter::CP_CAM) { mExposureBracketMode = OMX_BracketVectorShot; mPendingCaptureSettings |= SetBurstExpBracket; } // if bracketing was previously set...we set again before capturing to clear if (mExposureBracketingValidEntries) { mPendingCaptureSettings |= SetBurstExpBracket; mExposureBracketingValidEntries = 0; } } str = params.get(TICameraParameters::KEY_ZOOM_BRACKETING_RANGE); if ( NULL != str ) { parseExpRange(str, mZoomBracketingValues, NULL, NULL, ZOOM_BRACKET_RANGE, mZoomBracketingValidEntries); mCurrentZoomBracketing = 0; mZoomBracketingEnabled = true; } else { if (mZoomBracketingValidEntries) { mZoomBracketingValidEntries = 0; } mZoomBracketingEnabled = false; } #endif // Flush config queue // If TRUE: Flush queue and abort processing before enqueing valstr = params.get(TICameraParameters::KEY_FLUSH_SHOT_CONFIG_QUEUE); if ( NULL != valstr ) { if ( 0 == strcmp(valstr, android::CameraParameters::TRUE) ) { mFlushShotConfigQueue = true; } else if ( 0 == strcmp(valstr, android::CameraParameters::FALSE) ) { mFlushShotConfigQueue = false; } else { CAMHAL_LOGE("Missing flush shot config parameter. Will use current (%s)", mFlushShotConfigQueue ? "true" : "false"); } } if ( params.getInt(android::CameraParameters::KEY_ROTATION) != -1 ) { if (params.getInt(android::CameraParameters::KEY_ROTATION) != (int) mPictureRotation) { mPendingCaptureSettings |= SetRotation; } if (cap->mFrameLayoutType == OMX_TI_StereoFrameLayout2D) { mPictureRotation = params.getInt(android::CameraParameters::KEY_ROTATION); } else { mPictureRotation = 0; } } else { if (mPictureRotation) mPendingCaptureSettings |= SetRotation; mPictureRotation = 0; } CAMHAL_LOGVB("Picture Rotation set %d", mPictureRotation); #ifdef OMAP_ENHANCEMENT // Read Sensor Orientation and set it based on perating mode varint = params.getInt(TICameraParameters::KEY_SENSOR_ORIENTATION); if ( varint != -1 ) { mSensorOrientation = varint; if (mSensorOrientation == 270 ||mSensorOrientation==90) { CAMHAL_LOGEA(" Orientation is 270/90. So setting counter rotation to Ducati"); mSensorOrientation +=180; mSensorOrientation%=360; } } else { mSensorOrientation = 0; } CAMHAL_LOGVB("Sensor Orientation set : %d", mSensorOrientation); #endif #ifdef OMAP_ENHANCEMENT_BURST_CAPTURE varint = params.getInt(TICameraParameters::KEY_BURST); if ( varint >= 1 ) { if (varint != (int) mBurstFrames) { mPendingCaptureSettings |= SetBurstExpBracket; } mBurstFrames = varint; } else { if (mBurstFrames != 1) mPendingCaptureSettings |= SetBurstExpBracket; mBurstFrames = 1; } CAMHAL_LOGVB("Burst Frames set %d", mBurstFrames); #endif varint = params.getInt(android::CameraParameters::KEY_JPEG_QUALITY); if ( varint >= MIN_JPEG_QUALITY && varint <= MAX_JPEG_QUALITY ) { if (varint != mPictureQuality) { mPendingCaptureSettings |= SetQuality; mPictureQuality = varint; } } else { if (mPictureQuality != MAX_JPEG_QUALITY) { mPendingCaptureSettings |= SetQuality; mPictureQuality = MAX_JPEG_QUALITY; } } CAMHAL_LOGVB("Picture Quality set %d", mPictureQuality); varint = params.getInt(android::CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH); if ( varint >= 0 ) { if (varint != mThumbWidth) { mPendingCaptureSettings |= SetThumb; mThumbWidth = varint; } } else { if (mThumbWidth != DEFAULT_THUMB_WIDTH) { mPendingCaptureSettings |= SetThumb; mThumbWidth = DEFAULT_THUMB_WIDTH; } } CAMHAL_LOGVB("Picture Thumb width set %d", mThumbWidth); varint = params.getInt(android::CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT); if ( varint >= 0 ) { if (varint != mThumbHeight) { mPendingCaptureSettings |= SetThumb; mThumbHeight = varint; } } else { if (mThumbHeight != DEFAULT_THUMB_HEIGHT) { mPendingCaptureSettings |= SetThumb; mThumbHeight = DEFAULT_THUMB_HEIGHT; } } CAMHAL_LOGVB("Picture Thumb height set %d", mThumbHeight); varint = params.getInt(android::CameraParameters::KEY_JPEG_THUMBNAIL_QUALITY); if ( varint >= MIN_JPEG_QUALITY && varint <= MAX_JPEG_QUALITY ) { if (varint != mThumbQuality) { mPendingCaptureSettings |= SetThumb; mThumbQuality = varint; } } else { if (mThumbQuality != MAX_JPEG_QUALITY) { mPendingCaptureSettings |= SetThumb; mThumbQuality = MAX_JPEG_QUALITY; } } CAMHAL_LOGDB("Thumbnail Quality set %d", mThumbQuality); if (mFirstTimeInit) { mPendingCaptureSettings = ECapturesettingsAll; } cap = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mVideoPortIndex]; cap->mWidth = params.getInt(TICameraParameters::RAW_WIDTH); cap->mHeight = params.getInt(TICameraParameters::RAW_HEIGHT); LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::getPictureBufferSize(CameraFrame &frame, size_t bufferCount) { status_t ret = NO_ERROR; OMXCameraPortParameters *imgCaptureData = NULL; OMX_ERRORTYPE eError = OMX_ErrorNone; LOG_FUNCTION_NAME; if ( NO_ERROR == ret ) { imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; // If any settings have changed that need to be set with SetParam, // we will need to disable the port to set them if ((mPendingCaptureSettings & ECaptureParamSettings)) { disableImagePort(); if ( NULL != mReleaseImageBuffersCallback ) { mReleaseImageBuffersCallback(mReleaseData); } } if (mPendingCaptureSettings & SetFormat) { ret = setFormat(OMX_CAMERA_PORT_IMAGE_OUT_IMAGE, *imgCaptureData); } if ( ret == NO_ERROR ) { frame.mLength = imgCaptureData->mBufSize; frame.mWidth = imgCaptureData->mWidth; frame.mHeight = imgCaptureData->mHeight; frame.mAlignment = imgCaptureData->mStride; CAMHAL_LOGDB("getPictureBufferSize: width:%u height:%u alignment:%u length:%u", frame.mWidth, frame.mHeight, frame.mAlignment, frame.mLength); } else { CAMHAL_LOGEB("setFormat() failed 0x%x", ret); } } LOG_FUNCTION_NAME_EXIT; return ret; } int OMXCameraAdapter::getBracketingValueMode(const char *a, const char *b) const { BracketingValueMode bvm = BracketingValueAbsolute; if ( (NULL != b) && (NULL != a) && (a < b) && ( (NULL != memchr(a, '+', b - a)) || (NULL != memchr(a, '-', b - a)) ) ) { bvm = BracketingValueRelative; } return bvm; } status_t OMXCameraAdapter::parseExpRange(const char *rangeStr, int *expRange, int *gainRange, int *expGainModes, size_t count, size_t &validEntries) { status_t ret = NO_ERROR; char *end = NULL; const char *startPtr = NULL; size_t i = 0; LOG_FUNCTION_NAME; if ( NULL == rangeStr ){ return -EINVAL; } if ( NULL == expRange ){ return -EINVAL; } if ( NO_ERROR == ret ) { startPtr = rangeStr; do { // Relative Exposure example: "-30,-10, 0, 10, 30" // Absolute Gain ex. (exposure,gain) pairs: "(100,300),(200,300),(400,300),(800,300),(1600,300)" // Relative Gain ex. (exposure,gain) pairs: "(-30,+0),(-10, +0),(+0,+0),(+10,+0),(+30,+0)" // Forced relative Exposure example: "-30F,-10F, 0F, 10F, 30F" // Forced absolute Gain ex. (exposure,gain) pairs: "(100,300)F,(200,300)F,(400,300)F,(800,300)F,(1600,300)F" // Forced relative Gain ex. (exposure,gain) pairs: "(-30,+0)F,(-10, +0)F,(+0,+0)F,(+10,+0)F,(+30,+0)F" // skip '(' and ',' while ((*startPtr == '(') || (*startPtr == ',')) startPtr++; expRange[i] = (int)strtol(startPtr, &end, 10); if (expGainModes) { // if gainRange is given rangeStr should be (exposure, gain) pair if (gainRange) { int bvm_exp = getBracketingValueMode(startPtr, end); startPtr = end + 1; // for the ',' gainRange[i] = (int)strtol(startPtr, &end, 10); if (BracketingValueAbsolute == bvm_exp) { expGainModes[i] = getBracketingValueMode(startPtr, end); } else { expGainModes[i] = bvm_exp; } } else { expGainModes[i] = BracketingValueCompensation; } } startPtr = end; // skip ')' while (*startPtr == ')') startPtr++; // Check for "forced" key if (expGainModes) { while ((*startPtr == 'F') || (*startPtr == 'f')) { if ( BracketingValueAbsolute == expGainModes[i] ) { expGainModes[i] = BracketingValueAbsoluteForced; } else if ( BracketingValueRelative == expGainModes[i] ) { expGainModes[i] = BracketingValueRelativeForced; } else if ( BracketingValueCompensation == expGainModes[i] ) { expGainModes[i] = BracketingValueCompensationForced; } else { CAMHAL_LOGE("Unexpected old mode 0x%x", expGainModes[i]); } startPtr++; } } i++; } while ((startPtr[0] != '\0') && (i < count)); validEntries = i; } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::doExposureBracketing(int *evValues, int *evValues2, int *evModes2, size_t evCount, size_t frameCount, bool flush, OMX_BRACKETMODETYPE bracketMode) { status_t ret = NO_ERROR; LOG_FUNCTION_NAME; if ( OMX_StateInvalid == mComponentState ) { CAMHAL_LOGEA("OMX component is in invalid state"); ret = -EINVAL; } if ( NULL == evValues ) { CAMHAL_LOGEA("Exposure compensation values pointer is invalid"); ret = -EINVAL; } if ( NO_ERROR == ret ) { #ifdef OMAP_ENHANCEMENT_CPCAM if (bracketMode == OMX_BracketVectorShot) { ret = setVectorShot(evValues, evValues2, evModes2, evCount, frameCount, flush, bracketMode); } else #endif ret = setExposureBracketing(evValues, evValues2, evCount, frameCount, bracketMode); } LOG_FUNCTION_NAME_EXIT; return ret; } #ifdef OMAP_ENHANCEMENT_CPCAM status_t OMXCameraAdapter::setVectorStop(bool toPreview) { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_TI_CONFIG_VECTSHOTSTOPMETHODTYPE vecShotStop; LOG_FUNCTION_NAME; OMX_INIT_STRUCT_PTR(&vecShotStop, OMX_TI_CONFIG_VECTSHOTSTOPMETHODTYPE); vecShotStop.nPortIndex = mCameraAdapterParameters.mImagePortIndex; if (toPreview) { vecShotStop.eStopMethod = OMX_TI_VECTSHOTSTOPMETHOD_GOTO_PREVIEW; } else { vecShotStop.eStopMethod = OMX_TI_VECTSHOTSTOPMETHOD_WAIT_IN_CAPTURE; } eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexConfigVectShotStopMethod, &vecShotStop); if (OMX_ErrorNone != eError) { CAMHAL_LOGEB("Error while configuring bracket shot 0x%x", eError); } else { CAMHAL_LOGDA("Bracket shot configured successfully"); } LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::initVectorShot() { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_CONFIG_CAPTUREMODETYPE expCapMode; OMX_CONFIG_EXTCAPTUREMODETYPE extExpCapMode; LOG_FUNCTION_NAME; if (NO_ERROR == ret) { OMX_INIT_STRUCT_PTR (&expCapMode, OMX_CONFIG_CAPTUREMODETYPE); expCapMode.nPortIndex = mCameraAdapterParameters.mImagePortIndex; expCapMode.bFrameLimited = OMX_FALSE; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, OMX_IndexConfigCaptureMode, &expCapMode); if (OMX_ErrorNone != eError) { CAMHAL_LOGEB("Error while configuring capture mode 0x%x", eError); goto exit; } else { CAMHAL_LOGDA("Camera capture mode configured successfully"); } } if (NO_ERROR == ret) { OMX_INIT_STRUCT_PTR (&extExpCapMode, OMX_CONFIG_EXTCAPTUREMODETYPE); extExpCapMode.nPortIndex = mCameraAdapterParameters.mImagePortIndex; extExpCapMode.bEnableBracketing = OMX_TRUE; extExpCapMode.tBracketConfigType.eBracketMode = OMX_BracketVectorShot; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_IndexConfigExtCaptureMode, &extExpCapMode); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring extended capture mode 0x%x", eError); goto exit; } else { CAMHAL_LOGDA("Extended camera capture mode configured successfully"); } } if (NO_ERROR == ret) { // set vector stop method to stop in capture ret = setVectorStop(false); } exit: LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::setVectorShot(int *evValues, int *evValues2, int *evModes2, size_t evCount, size_t frameCount, bool flush, OMX_BRACKETMODETYPE bracketMode) { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_TI_CONFIG_ENQUEUESHOTCONFIGS enqueueShotConfigs; OMX_TI_CONFIG_QUERYAVAILABLESHOTS queryAvailableShots; bool doFlush = flush; LOG_FUNCTION_NAME; OMX_INIT_STRUCT_PTR(&enqueueShotConfigs, OMX_TI_CONFIG_ENQUEUESHOTCONFIGS); OMX_INIT_STRUCT_PTR(&queryAvailableShots, OMX_TI_CONFIG_QUERYAVAILABLESHOTS); queryAvailableShots.nPortIndex = mCameraAdapterParameters.mImagePortIndex; eError = OMX_GetConfig(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexConfigQueryAvailableShots, &queryAvailableShots); if (OMX_ErrorNone != eError) { CAMHAL_LOGE("Error getting available shots 0x%x", eError); goto exit; } else { CAMHAL_LOGD("AVAILABLE SHOTS: %d", queryAvailableShots.nAvailableShots); if (queryAvailableShots.nAvailableShots < evCount) { // TODO(XXX): Need to implement some logic to handle this error CAMHAL_LOGE("Not enough available shots to fulfill this queue request"); ret = -ENOSPC; goto exit; } } for ( unsigned int confID = 0; confID < evCount; ) { unsigned int i; for ( i = 0 ; (i < ARRAY_SIZE(enqueueShotConfigs.nShotConfig)) && (confID < evCount); i++, confID++ ) { CAMHAL_LOGD("%2u: (%7d,%4d) mode: %d", confID, evValues[confID], evValues2[confID], evModes2[confID]); enqueueShotConfigs.nShotConfig[i].nConfigId = confID; enqueueShotConfigs.nShotConfig[i].nFrames = 1; if ( (BracketingValueCompensation == evModes2[confID]) || (BracketingValueCompensationForced == evModes2[confID]) ) { // EV compensation enqueueShotConfigs.nShotConfig[i].nEC = evValues[confID]; enqueueShotConfigs.nShotConfig[i].nExp = 0; enqueueShotConfigs.nShotConfig[i].nGain = 0; } else { // exposure,gain pair enqueueShotConfigs.nShotConfig[i].nEC = 0; enqueueShotConfigs.nShotConfig[i].nExp = evValues[confID]; enqueueShotConfigs.nShotConfig[i].nGain = evValues2[confID]; } enqueueShotConfigs.nShotConfig[i].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_ABSOLUTE; switch (evModes2[confID]) { case BracketingValueAbsolute: // (exp,gain) pairs directly program sensor values default : enqueueShotConfigs.nShotConfig[i].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_ABSOLUTE; break; case BracketingValueRelative: // (exp,gain) pairs relative to AE settings and constraints case BracketingValueCompensation: // EV compensation relative to AE settings and constraints enqueueShotConfigs.nShotConfig[i].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_RELATIVE; break; case BracketingValueAbsoluteForced: // (exp,gain) pairs directly program sensor values // are forced over constraints due to flicker, etc. enqueueShotConfigs.nShotConfig[i].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_FORCE_ABSOLUTE; break; case BracketingValueRelativeForced: // (exp, gain) pairs relative to AE settings AND settings case BracketingValueCompensationForced: // EV compensation relative to AE settings and constraints // are forced over constraints due to flicker, etc. enqueueShotConfigs.nShotConfig[i].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_FORCE_RELATIVE; break; } enqueueShotConfigs.nShotConfig[i].bNoSnapshot = OMX_FALSE; // TODO: Make this configurable } // Repeat last exposure and again if ((confID == evCount) && (evCount > 0) && (frameCount > evCount) && (0 != i)) { enqueueShotConfigs.nShotConfig[i-1].nFrames = frameCount - evCount; } enqueueShotConfigs.nPortIndex = mCameraAdapterParameters.mImagePortIndex; enqueueShotConfigs.bFlushQueue = doFlush ? OMX_TRUE : OMX_FALSE; enqueueShotConfigs.nNumConfigs = i; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_TI_IndexConfigEnqueueShotConfigs, &enqueueShotConfigs); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring enqueue shot 0x%x", eError); goto exit; } else { CAMHAL_LOGDA("Enqueue shot configured successfully"); } // Flush only first time doFlush = false; } // Handle burst capture (no any bracketing) case if (0 == evCount) { CAMHAL_LOGE("Handle burst capture (no any bracketing) case"); enqueueShotConfigs.nShotConfig[0].nConfigId = 0; enqueueShotConfigs.nShotConfig[0].nFrames = frameCount; enqueueShotConfigs.nShotConfig[0].nEC = 0; enqueueShotConfigs.nShotConfig[0].nExp = 0; enqueueShotConfigs.nShotConfig[0].nGain = 0; enqueueShotConfigs.nShotConfig[0].eExpGainApplyMethod = OMX_TI_EXPGAINAPPLYMETHOD_RELATIVE; enqueueShotConfigs.nShotConfig[0].bNoSnapshot = OMX_FALSE; // TODO: Make this configurable enqueueShotConfigs.nNumConfigs = 1; enqueueShotConfigs.nPortIndex = mCameraAdapterParameters.mImagePortIndex; enqueueShotConfigs.bFlushQueue = doFlush ? OMX_TRUE : OMX_FALSE; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_TI_IndexConfigEnqueueShotConfigs, &enqueueShotConfigs); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring enqueue shot 0x%x", eError); goto exit; } else { CAMHAL_LOGDA("Enqueue shot configured successfully"); } } exit: LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } #endif status_t OMXCameraAdapter::setExposureBracketing(int *evValues, int *evValues2, size_t evCount, size_t frameCount, OMX_BRACKETMODETYPE bracketMode) { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_CONFIG_CAPTUREMODETYPE expCapMode; OMX_CONFIG_EXTCAPTUREMODETYPE extExpCapMode; LOG_FUNCTION_NAME; if ( NO_ERROR == ret ) { OMX_INIT_STRUCT_PTR (&expCapMode, OMX_CONFIG_CAPTUREMODETYPE); expCapMode.nPortIndex = mCameraAdapterParameters.mImagePortIndex; /// If frameCount>0 but evCount<=0, then this is the case of HQ burst. //Otherwise, it is normal HQ capture ///If frameCount>0 and evCount>0 then this is the cause of HQ Exposure bracketing. if ( 0 == evCount && 0 == frameCount ) { expCapMode.bFrameLimited = OMX_FALSE; } else { expCapMode.bFrameLimited = OMX_TRUE; expCapMode.nFrameLimit = frameCount; } eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, OMX_IndexConfigCaptureMode, &expCapMode); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring capture mode 0x%x", eError); } else { CAMHAL_LOGDA("Camera capture mode configured successfully"); } } if ( NO_ERROR == ret ) { OMX_INIT_STRUCT_PTR (&extExpCapMode, OMX_CONFIG_EXTCAPTUREMODETYPE); extExpCapMode.nPortIndex = mCameraAdapterParameters.mImagePortIndex; if ( 0 == evCount ) { extExpCapMode.bEnableBracketing = OMX_FALSE; } else { extExpCapMode.bEnableBracketing = OMX_TRUE; extExpCapMode.tBracketConfigType.eBracketMode = bracketMode; extExpCapMode.tBracketConfigType.nNbrBracketingValues = evCount - 1; } for ( unsigned int i = 0 ; i < evCount ; i++ ) { if (bracketMode == OMX_BracketExposureGainAbsolute) { extExpCapMode.tBracketConfigType.nBracketValues[i] = evValues[i]; #ifndef CAMERAHAL_TUNA extExpCapMode.tBracketConfigType.nBracketValues2[i] = evValues2[i]; #endif } else { // assuming OMX_BracketExposureRelativeInEV extExpCapMode.tBracketConfigType.nBracketValues[i] = ( evValues[i] * ( 1 << Q16_OFFSET ) ) / 10; } } eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_IndexConfigExtCaptureMode, &extExpCapMode); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring extended capture mode 0x%x", eError); } else { CAMHAL_LOGDA("Extended camera capture mode configured successfully"); } } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::setShutterCallback(bool enabled) { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_CONFIG_CALLBACKREQUESTTYPE shutterRequstCallback; LOG_FUNCTION_NAME; if ( OMX_StateExecuting != mComponentState ) { CAMHAL_LOGEA("OMX component not in executing state"); ret = -1; } if ( NO_ERROR == ret ) { OMX_INIT_STRUCT_PTR (&shutterRequstCallback, OMX_CONFIG_CALLBACKREQUESTTYPE); shutterRequstCallback.nPortIndex = OMX_ALL; if ( enabled ) { shutterRequstCallback.bEnable = OMX_TRUE; shutterRequstCallback.nIndex = ( OMX_INDEXTYPE ) OMX_TI_IndexConfigShutterCallback; CAMHAL_LOGDA("Enabling shutter callback"); } else { shutterRequstCallback.bEnable = OMX_FALSE; shutterRequstCallback.nIndex = ( OMX_INDEXTYPE ) OMX_TI_IndexConfigShutterCallback; CAMHAL_LOGDA("Disabling shutter callback"); } eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_IndexConfigCallbackRequest, &shutterRequstCallback); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error registering shutter callback 0x%x", eError); ret = -1; } else { CAMHAL_LOGDB("Shutter callback for index 0x%x registered successfully", OMX_TI_IndexConfigShutterCallback); } } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::doBracketing(OMX_BUFFERHEADERTYPE *pBuffHeader, CameraFrame::FrameType typeOfFrame) { status_t ret = NO_ERROR; int currentBufferIdx, nextBufferIdx; OMXCameraPortParameters * imgCaptureData = NULL; LOG_FUNCTION_NAME; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; if ( OMX_StateExecuting != mComponentState ) { CAMHAL_LOGEA("OMX component is not in executing state"); ret = -EINVAL; } if ( NO_ERROR == ret ) { CameraBuffer *buffer = (CameraBuffer *)pBuffHeader->pAppPrivate; currentBufferIdx = buffer->index; if ( currentBufferIdx >= imgCaptureData->mNumBufs) { CAMHAL_LOGEB("Invalid bracketing buffer index 0x%x", currentBufferIdx); ret = -EINVAL; } } if ( NO_ERROR == ret ) { mBracketingBuffersQueued[currentBufferIdx] = false; mBracketingBuffersQueuedCount--; if ( 0 >= mBracketingBuffersQueuedCount ) { nextBufferIdx = ( currentBufferIdx + 1 ) % imgCaptureData->mNumBufs; mBracketingBuffersQueued[nextBufferIdx] = true; mBracketingBuffersQueuedCount++; mLastBracetingBufferIdx = nextBufferIdx; setFrameRefCountByType((CameraBuffer *)imgCaptureData->mBufferHeader[nextBufferIdx]->pAppPrivate, typeOfFrame, 1); returnFrame((CameraBuffer *)imgCaptureData->mBufferHeader[nextBufferIdx]->pAppPrivate, typeOfFrame); } } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::sendBracketFrames(size_t &framesSent) { status_t ret = NO_ERROR; int currentBufferIdx; OMXCameraPortParameters * imgCaptureData = NULL; LOG_FUNCTION_NAME; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; framesSent = 0; if ( OMX_StateExecuting != mComponentState ) { CAMHAL_LOGEA("OMX component is not in executing state"); ret = -EINVAL; } if ( NO_ERROR == ret ) { currentBufferIdx = mLastBracetingBufferIdx; do { currentBufferIdx++; currentBufferIdx %= imgCaptureData->mNumBufs; if (!mBracketingBuffersQueued[currentBufferIdx] ) { CameraFrame cameraFrame; sendCallBacks(cameraFrame, imgCaptureData->mBufferHeader[currentBufferIdx], imgCaptureData->mImageType, imgCaptureData); framesSent++; } } while ( currentBufferIdx != mLastBracetingBufferIdx ); } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::startBracketing(int range) { status_t ret = NO_ERROR; OMXCameraPortParameters * imgCaptureData = NULL; LOG_FUNCTION_NAME; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; if ( OMX_StateExecuting != mComponentState ) { CAMHAL_LOGEA("OMX component is not in executing state"); ret = -EINVAL; } { android::AutoMutex lock(mBracketingLock); if ( mBracketingEnabled ) { return ret; } } if ( 0 == imgCaptureData->mNumBufs ) { CAMHAL_LOGEB("Image capture buffers set to %d", imgCaptureData->mNumBufs); ret = -EINVAL; } if ( mPending3Asettings ) apply3Asettings(mParameters3A); if ( NO_ERROR == ret ) { android::AutoMutex lock(mBracketingLock); mBracketingRange = range; mBracketingBuffersQueued = new bool[imgCaptureData->mNumBufs]; if ( NULL == mBracketingBuffersQueued ) { CAMHAL_LOGEA("Unable to allocate bracketing management structures"); ret = -1; } if ( NO_ERROR == ret ) { mBracketingBuffersQueuedCount = imgCaptureData->mNumBufs; mBurstFramesAccum = imgCaptureData->mNumBufs; mLastBracetingBufferIdx = mBracketingBuffersQueuedCount - 1; for ( int i = 0 ; i < imgCaptureData->mNumBufs ; i++ ) { mBracketingBuffersQueued[i] = true; } } } if ( NO_ERROR == ret ) { CachedCaptureParameters* cap_params = cacheCaptureParameters(); ret = startImageCapture(true, cap_params); delete cap_params; { android::AutoMutex lock(mBracketingLock); if ( NO_ERROR == ret ) { mBracketingEnabled = true; } else { mBracketingEnabled = false; } } } LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::stopBracketing() { status_t ret = NO_ERROR; LOG_FUNCTION_NAME; ret = stopImageCapture(); android::AutoMutex lock(mBracketingLock); if ( NULL != mBracketingBuffersQueued ) { delete [] mBracketingBuffersQueued; } mBracketingBuffersQueued = NULL; mBracketingEnabled = false; mBracketingBuffersQueuedCount = 0; mLastBracetingBufferIdx = 0; LOG_FUNCTION_NAME_EXIT; return ret; } status_t OMXCameraAdapter::startImageCapture(bool bracketing, CachedCaptureParameters* capParams) { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMXCameraPortParameters * capData = NULL; OMX_CONFIG_BOOLEANTYPE bOMX; size_t bracketingSent = 0; LOG_FUNCTION_NAME; android::AutoMutex lock(mImageCaptureLock); if(!mCaptureConfigured) { ///Image capture was cancelled before we could start return NO_ERROR; } if ( 0 != mStartCaptureSem.Count() ) { CAMHAL_LOGEB("Error mStartCaptureSem semaphore count %d", mStartCaptureSem.Count()); return NO_INIT; } if ( !bracketing ) { if ((getNextState() & (CAPTURE_ACTIVE|BRACKETING_ACTIVE)) == 0) { CAMHAL_LOGDA("trying starting capture when already canceled"); return NO_ERROR; } } if (!capParams) { CAMHAL_LOGE("Invalid cached parameters sent!"); return BAD_VALUE; } // Camera framework doesn't expect face callbacks once capture is triggered pauseFaceDetection(true); //During bracketing image capture is already active { android::AutoMutex lock(mBracketingLock); if ( mBracketingEnabled ) { //Stop bracketing, activate normal burst for the remaining images mBracketingEnabled = false; ret = sendBracketFrames(bracketingSent); // Check if we accumulated enough buffers if ( bracketingSent < ( mBracketingRange - 1 ) ) { mCapturedFrames = mBracketingRange + ( ( mBracketingRange - 1 ) - bracketingSent ); } else { mCapturedFrames = mBracketingRange; } mBurstFramesQueued = 0; mBurstFramesAccum = mCapturedFrames; if(ret != NO_ERROR) goto EXIT; else return ret; } } if ( NO_ERROR == ret ) { if (capParams->mPendingCaptureSettings & SetRotation) { mPendingCaptureSettings &= ~SetRotation; ret = setPictureRotation(mPictureRotation); if ( NO_ERROR != ret ) { CAMHAL_LOGEB("Error configuring image rotation %x", ret); } } if (capParams->mPendingCaptureSettings & SetBurstExpBracket) { mPendingCaptureSettings &= ~SetBurstExpBracket; if ( mBracketingSet ) { ret = doExposureBracketing(capParams->mExposureBracketingValues, capParams->mExposureGainBracketingValues, capParams->mExposureGainBracketingModes, 0, 0, capParams->mFlushShotConfigQueue, capParams->mExposureBracketMode); } else { ret = doExposureBracketing(capParams->mExposureBracketingValues, capParams->mExposureGainBracketingValues, capParams->mExposureGainBracketingModes, capParams->mExposureBracketingValidEntries, capParams->mBurstFrames, capParams->mFlushShotConfigQueue, capParams->mExposureBracketMode); } if ( ret != NO_ERROR ) { CAMHAL_LOGEB("setExposureBracketing() failed %d", ret); goto EXIT; } } } #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("startImageCapture bracketing configs done: ", &mStartCapture); #endif capData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; //OMX shutter callback events are only available in hq mode if ( (HIGH_QUALITY == mCapMode) || (HIGH_QUALITY_ZSL== mCapMode)) { if ( NO_ERROR == ret ) { ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, (OMX_EVENTTYPE) OMX_EventIndexSettingChanged, OMX_ALL, OMX_TI_IndexConfigShutterCallback, mStartCaptureSem); } if ( NO_ERROR == ret ) { ret = setShutterCallback(true); } } if (mPending3Asettings) { apply3Asettings(mParameters3A); } if (ret == NO_ERROR) { int index = 0; int queued = 0; android::AutoMutex lock(mBurstLock); if (capParams->mFlushShotConfigQueue) { // reset shot queue mCapturedFrames = mBurstFrames; mBurstFramesAccum = mBurstFrames; mBurstFramesQueued = 0; for ( int index = 0 ; index < capData->mNumBufs ; index++ ) { if (OMXCameraPortParameters::FILL == capData->mStatus[index]) { mBurstFramesQueued++; } } } else { mCapturedFrames += mBurstFrames; mBurstFramesAccum += mBurstFrames; } CAMHAL_LOGD("mBurstFramesQueued = %d mBurstFramesAccum = %d index = %d " "capData->mNumBufs = %d queued = %d capData->mMaxQueueable = %d", mBurstFramesQueued,mBurstFramesAccum,index, capData->mNumBufs,queued,capData->mMaxQueueable); CAMHAL_LOGD("%d", (mBurstFramesQueued < mBurstFramesAccum) && (index < capData->mNumBufs) && (queued < capData->mMaxQueueable)); while ((mBurstFramesQueued < mBurstFramesAccum) && (index < capData->mNumBufs) && (queued < capData->mMaxQueueable)) { if (capData->mStatus[index] == OMXCameraPortParameters::IDLE) { CAMHAL_LOGDB("Queuing buffer on Capture port - %p", capData->mBufferHeader[index]->pBuffer); capData->mStatus[index] = OMXCameraPortParameters::FILL; eError = OMX_FillThisBuffer(mCameraAdapterParameters.mHandleComp, (OMX_BUFFERHEADERTYPE*)capData->mBufferHeader[index]); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); mBurstFramesQueued++; queued++; } else if (OMXCameraPortParameters::FILL == capData->mStatus[index]) { CAMHAL_LOGE("Not queueing index = %d", index); queued++; } index++; } #ifdef CAMERAHAL_USE_RAW_IMAGE_SAVING if (mRawCapture) { capData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mVideoPortIndex]; ///Queue all the buffers on capture port for ( int index = 0 ; index < capData->mNumBufs ; index++ ) { CAMHAL_LOGDB("Queuing buffer on Video port (for RAW capture) - 0x%x", ( unsigned int ) capData->mBufferHeader[index]->pBuffer); capData->mStatus[index] = OMXCameraPortParameters::FILL; eError = OMX_FillThisBuffer(mCameraAdapterParameters.mHandleComp, (OMX_BUFFERHEADERTYPE*)capData->mBufferHeader[index]); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); } } #endif mWaitingForSnapshot = true; mCaptureSignalled = false; // Capturing command is not needed when capturing in video mode // Only need to queue buffers on image ports if ( ( mCapMode != VIDEO_MODE ) && ( mCapMode != VIDEO_MODE_HQ ) ) { OMX_INIT_STRUCT_PTR (&bOMX, OMX_CONFIG_BOOLEANTYPE); bOMX.bEnabled = OMX_TRUE; /// sending Capturing Command to the component eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, OMX_IndexConfigCapturing, &bOMX); CAMHAL_LOGDB("Capture set - 0x%x", eError); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); } } #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("startImageCapture image buffers queued and capture enabled: ", &mStartCapture); #endif //OMX shutter callback events are only available in hq mode if ( (HIGH_QUALITY == mCapMode) || (HIGH_QUALITY_ZSL== mCapMode)) { if ( NO_ERROR == ret ) { ret = mStartCaptureSem.WaitTimeout(OMX_CAPTURE_TIMEOUT); } //If something bad happened while we wait if (mComponentState != OMX_StateExecuting) { CAMHAL_LOGEA("Invalid State after Image Capture Exitting!!!"); goto EXIT; } if ( NO_ERROR == ret ) { CAMHAL_LOGDA("Shutter callback received"); notifyShutterSubscribers(); } else { ret |= RemoveEvent(mCameraAdapterParameters.mHandleComp, (OMX_EVENTTYPE) OMX_EventIndexSettingChanged, OMX_ALL, OMX_TI_IndexConfigShutterCallback, NULL); CAMHAL_LOGEA("Timeout expired on shutter callback"); goto EXIT; } } #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("startImageCapture shutter event received: ", &mStartCapture); #endif return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); EXIT: CAMHAL_LOGEB("Exiting function %s because of ret %d eError=%x", __FUNCTION__, ret, eError); mWaitingForSnapshot = false; mCaptureSignalled = false; performCleanupAfterError(); LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::stopImageCapture() { status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMX_CONFIG_BOOLEANTYPE bOMX; OMXCameraPortParameters *imgCaptureData = NULL; LOG_FUNCTION_NAME; android::AutoMutex lock(mImageCaptureLock); if (!mCaptureConfigured) { //Capture is not ongoing, return from here return NO_ERROR; } if ( 0 != mStopCaptureSem.Count() ) { CAMHAL_LOGEB("Error mStopCaptureSem semaphore count %d", mStopCaptureSem.Count()); goto EXIT; } // TODO(XXX): Reprocessing is currently piggy-backing capture commands if (mAdapterState == REPROCESS_STATE) { ret = stopReprocess(); } //Disable the callback first mWaitingForSnapshot = false; // OMX shutter callback events are only available in hq mode if ((HIGH_QUALITY == mCapMode) || (HIGH_QUALITY_ZSL== mCapMode)) { //Disable the callback first ret = setShutterCallback(false); // if anybody is waiting on the shutter callback // signal them and then recreate the semaphore if ( 0 != mStartCaptureSem.Count() ) { for (int i = mStartCaptureSem.Count(); i < 0; i++) { ret |= SignalEvent(mCameraAdapterParameters.mHandleComp, (OMX_EVENTTYPE) OMX_EventIndexSettingChanged, OMX_ALL, OMX_TI_IndexConfigShutterCallback, NULL ); } mStartCaptureSem.Create(0); } } #ifdef OMAP_ENHANCEMENT_CPCAM else if (CP_CAM == mCapMode) { // Reset shot config queue OMX_TI_CONFIG_ENQUEUESHOTCONFIGS resetShotConfigs; OMX_INIT_STRUCT_PTR(&resetShotConfigs, OMX_TI_CONFIG_ENQUEUESHOTCONFIGS); resetShotConfigs.nPortIndex = mCameraAdapterParameters.mImagePortIndex; resetShotConfigs.bFlushQueue = OMX_TRUE; resetShotConfigs.nNumConfigs = 0; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, ( OMX_INDEXTYPE ) OMX_TI_IndexConfigEnqueueShotConfigs, &resetShotConfigs); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while reset shot config 0x%x", eError); goto EXIT; } else { CAMHAL_LOGDA("Shot config reset successfully"); } } #endif //Wait here for the capture to be done, in worst case timeout and proceed with cleanup mCaptureSem.WaitTimeout(OMX_CAPTURE_TIMEOUT); //If somethiing bad happened while we wait if (mComponentState == OMX_StateInvalid) { CAMHAL_LOGEA("Invalid State Image Capture Stop Exitting!!!"); goto EXIT; } // Disable image capture // Capturing command is not needed when capturing in video mode if ( ( mCapMode != VIDEO_MODE ) && ( mCapMode != VIDEO_MODE_HQ ) ) { OMX_INIT_STRUCT_PTR (&bOMX, OMX_CONFIG_BOOLEANTYPE); bOMX.bEnabled = OMX_FALSE; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, OMX_IndexConfigCapturing, &bOMX); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGDB("Error during SetConfig- 0x%x", eError); ret = -1; goto EXIT; } } CAMHAL_LOGDB("Capture set - 0x%x", eError); mCaptureSignalled = true; //set this to true if we exited because of timeout { android::AutoMutex lock(mFrameCountMutex); mFrameCount = 0; mFirstFrameCondition.broadcast(); } // Stop is always signalled externally in CPCAM mode // We need to make sure we really stop if ((mCapMode == CP_CAM)) { disableReprocess(); disableImagePort(); if ( NULL != mReleaseImageBuffersCallback ) { mReleaseImageBuffersCallback(mReleaseData); } } // Moving code for below commit here as an optimization for continuous capture, // so focus settings don't have to reapplied after each capture // c78fa2a CameraHAL: Always reset focus mode after capture // Workaround when doing many consecutive shots, CAF wasn't getting restarted. mPending3Asettings |= SetFocus; mCapturedFrames = 0; mBurstFramesAccum = 0; mBurstFramesQueued = 0; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); EXIT: CAMHAL_LOGEB("Exiting function %s because of ret %d eError=%x", __FUNCTION__, ret, eError); //Release image buffers if ( NULL != mReleaseImageBuffersCallback ) { mReleaseImageBuffersCallback(mReleaseData); } { android::AutoMutex lock(mFrameCountMutex); mFrameCount = 0; mFirstFrameCondition.broadcast(); } performCleanupAfterError(); LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::disableImagePort(){ status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMXCameraPortParameters *imgCaptureData = NULL; OMXCameraPortParameters *imgRawCaptureData = NULL; if (!mCaptureConfigured) { return NO_ERROR; } mCaptureConfigured = false; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; imgRawCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mVideoPortIndex]; // for RAW capture ///Register for Image port Disable event ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, OMX_EventCmdComplete, OMX_CommandPortDisable, mCameraAdapterParameters.mImagePortIndex, mStopCaptureSem); ///Disable Capture Port eError = OMX_SendCommand(mCameraAdapterParameters.mHandleComp, OMX_CommandPortDisable, mCameraAdapterParameters.mImagePortIndex, NULL); ///Free all the buffers on capture port if (imgCaptureData) { CAMHAL_LOGDB("Freeing buffer on Capture port - %d", imgCaptureData->mNumBufs); for ( int index = 0 ; index < imgCaptureData->mNumBufs ; index++) { CAMHAL_LOGDB("Freeing buffer on Capture port - 0x%x", ( unsigned int ) imgCaptureData->mBufferHeader[index]->pBuffer); eError = OMX_FreeBuffer(mCameraAdapterParameters.mHandleComp, mCameraAdapterParameters.mImagePortIndex, (OMX_BUFFERHEADERTYPE*)imgCaptureData->mBufferHeader[index]); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); } } CAMHAL_LOGDA("Waiting for port disable"); //Wait for the image port enable event ret = mStopCaptureSem.WaitTimeout(OMX_CMD_TIMEOUT); //If somethiing bad happened while we wait if (mComponentState == OMX_StateInvalid) { CAMHAL_LOGEA("Invalid State after Disable Image Port Exitting!!!"); goto EXIT; } if ( NO_ERROR == ret ) { CAMHAL_LOGDA("Port disabled"); } else { ret |= RemoveEvent(mCameraAdapterParameters.mHandleComp, OMX_EventCmdComplete, OMX_CommandPortDisable, mCameraAdapterParameters.mImagePortIndex, NULL); CAMHAL_LOGDA("Timeout expired on port disable"); goto EXIT; } deinitInternalBuffers(mCameraAdapterParameters.mImagePortIndex); // since port settings are not persistent after port is disabled... mPendingCaptureSettings |= SetFormat; #ifdef CAMERAHAL_USE_RAW_IMAGE_SAVING if (mRawCapture) { ///Register for Video port Disable event ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, OMX_EventCmdComplete, OMX_CommandPortDisable, mCameraAdapterParameters.mVideoPortIndex, mStopCaptureSem); ///Disable RawCapture Port eError = OMX_SendCommand(mCameraAdapterParameters.mHandleComp, OMX_CommandPortDisable, mCameraAdapterParameters.mVideoPortIndex, NULL); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); ///Free all the buffers on RawCapture port if (imgRawCaptureData) { CAMHAL_LOGDB("Freeing buffer on Capture port - %d", imgRawCaptureData->mNumBufs); for ( int index = 0 ; index < imgRawCaptureData->mNumBufs ; index++) { CAMHAL_LOGDB("Freeing buffer on Capture port - 0x%x", ( unsigned int ) imgRawCaptureData->mBufferHeader[index]->pBuffer); eError = OMX_FreeBuffer(mCameraAdapterParameters.mHandleComp, mCameraAdapterParameters.mVideoPortIndex, (OMX_BUFFERHEADERTYPE*)imgRawCaptureData->mBufferHeader[index]); GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); } } CAMHAL_LOGDA("Waiting for Video port disable"); //Wait for the image port enable event mStopCaptureSem.WaitTimeout(OMX_CMD_TIMEOUT); CAMHAL_LOGDA("Video Port disabled"); } #endif EXIT: return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::initInternalBuffers(OMX_U32 portIndex) { #ifndef CAMERAHAL_TUNA OMX_ERRORTYPE eError = OMX_ErrorNone; int index = 0; OMX_TI_PARAM_USEBUFFERDESCRIPTOR bufferdesc; /* Indicate to Ducati that we're planning to use dynamically-mapped buffers */ OMX_INIT_STRUCT_PTR (&bufferdesc, OMX_TI_PARAM_USEBUFFERDESCRIPTOR); bufferdesc.nPortIndex = portIndex; bufferdesc.bEnabled = OMX_FALSE; bufferdesc.eBufferType = OMX_TI_BufferTypePhysicalPageList; eError = OMX_SetParameter(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexUseBufferDescriptor, &bufferdesc); if (eError!=OMX_ErrorNone) { CAMHAL_LOGEB("OMX_SetParameter - %x", eError); return -EINVAL; } CAMHAL_LOGDA("Initializing internal buffers"); do { OMX_TI_PARAM_COMPONENTBUFALLOCTYPE bufferalloc; OMX_TI_PARAM_COMPONENTBUFALLOCTYPE bufferallocset; OMX_INIT_STRUCT_PTR (&bufferalloc, OMX_TI_PARAM_COMPONENTBUFALLOCTYPE); bufferalloc.nPortIndex = portIndex; bufferalloc.nIndex = index; eError = OMX_GetParameter (mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE)OMX_TI_IndexParamComponentBufferAllocation, &bufferalloc); if (eError == OMX_ErrorNoMore) { return NO_ERROR; } if (eError != OMX_ErrorNone) { CAMHAL_LOGE("GetParameter failed error = 0x%x", eError); break; } CAMHAL_LOGDB("Requesting buftype %d of size %dx%d", (int)bufferalloc.eBufType, (int)bufferalloc.nAllocWidth, (int)bufferalloc.nAllocLines); bufferalloc.eBufType = OMX_TI_BufferTypeHardwareReserved1D; OMX_INIT_STRUCT_PTR (&bufferallocset, OMX_TI_PARAM_COMPONENTBUFALLOCTYPE); bufferallocset.nPortIndex = portIndex; bufferallocset.nIndex = index; bufferallocset.eBufType = OMX_TI_BufferTypeHardwareReserved1D; bufferallocset.nAllocWidth = bufferalloc.nAllocWidth; bufferallocset.nAllocLines = bufferalloc.nAllocLines; eError = OMX_SetParameter (mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE)OMX_TI_IndexParamComponentBufferAllocation, &bufferallocset); if (eError != OMX_ErrorNone) { CAMHAL_LOGE("SetParameter failed, error=%08x", eError); if (eError == OMX_ErrorNoMore) return NO_ERROR; break; } index++; /* 1 is an arbitrary limit */ } while (index < 1); CAMHAL_LOGV("Ducati requested too many (>1) internal buffers"); return -EINVAL; #else return NO_ERROR; #endif } status_t OMXCameraAdapter::deinitInternalBuffers(OMX_U32 portIndex) { OMX_ERRORTYPE eError = OMX_ErrorNone; #ifndef CAMERAHAL_TUNA OMX_TI_PARAM_USEBUFFERDESCRIPTOR bufferdesc; OMX_INIT_STRUCT_PTR (&bufferdesc, OMX_TI_PARAM_USEBUFFERDESCRIPTOR); bufferdesc.nPortIndex = portIndex; bufferdesc.bEnabled = OMX_FALSE; bufferdesc.eBufferType = OMX_TI_BufferTypeDefault; eError = OMX_SetParameter(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexUseBufferDescriptor, &bufferdesc); if (eError!=OMX_ErrorNone) { CAMHAL_LOGEB("OMX_SetParameter - %x", eError); return -EINVAL; } OMX_TI_PARAM_COMPONENTBUFALLOCTYPE bufferalloc; OMX_INIT_STRUCT_PTR (&bufferalloc, OMX_TI_PARAM_COMPONENTBUFALLOCTYPE); bufferalloc.nPortIndex = portIndex; bufferalloc.eBufType = OMX_TI_BufferTypeDefault; bufferalloc.nAllocWidth = 1; bufferalloc.nAllocLines = 1; eError = OMX_SetParameter(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexParamComponentBufferAllocation, &bufferalloc); if (eError!=OMX_ErrorNone) { CAMHAL_LOGEB("OMX_SetParameter - %x", eError); return -EINVAL; } #endif return Utils::ErrorUtils::omxToAndroidError(eError); } status_t OMXCameraAdapter::UseBuffersCapture(CameraBuffer * bufArr, int num) { LOG_FUNCTION_NAME; status_t ret = NO_ERROR; OMX_ERRORTYPE eError = OMX_ErrorNone; OMXCameraPortParameters * imgCaptureData = NULL; OMXCameraPortParameters cap; imgCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mImagePortIndex]; if ( 0 != mUseCaptureSem.Count() ) { CAMHAL_LOGEB("Error mUseCaptureSem semaphore count %d", mUseCaptureSem.Count()); return BAD_VALUE; } CAMHAL_ASSERT(num > 0); // if some setting that requires a SetParameter (including // changing buffer types) then we need to disable the port // before being allowed to apply the settings if ((mPendingCaptureSettings & ECaptureParamSettings) || bufArr[0].type != imgCaptureData->mBufferType || imgCaptureData->mNumBufs != num) { if (mCaptureConfigured) { disableImagePort(); if ( NULL != mReleaseImageBuffersCallback ) { mReleaseImageBuffersCallback(mReleaseData); } } imgCaptureData->mBufferType = bufArr[0].type; imgCaptureData->mNumBufs = num; CAMHAL_LOGDB("Params Width = %d", (int)imgCaptureData->mWidth); CAMHAL_LOGDB("Params Height = %d", (int)imgCaptureData->mHeight); if (mPendingCaptureSettings & SetFormat) { mPendingCaptureSettings &= ~SetFormat; ret = setFormat(OMX_CAMERA_PORT_IMAGE_OUT_IMAGE, *imgCaptureData); if ( ret != NO_ERROR ) { CAMHAL_LOGEB("setFormat() failed %d", ret); LOG_FUNCTION_NAME_EXIT; return ret; } } if (mPendingCaptureSettings & SetThumb) { mPendingCaptureSettings &= ~SetThumb; ret = setThumbnailParams(mThumbWidth, mThumbHeight, mThumbQuality); if ( NO_ERROR != ret) { CAMHAL_LOGEB("Error configuring thumbnail size %x", ret); return ret; } } if (mPendingCaptureSettings & SetQuality) { mPendingCaptureSettings &= ~SetQuality; ret = setImageQuality(mPictureQuality); if ( NO_ERROR != ret) { CAMHAL_LOGEB("Error configuring image quality %x", ret); goto EXIT; } } // Configure DOMX to use either gralloc handles or vptrs { OMX_TI_PARAMUSENATIVEBUFFER domxUseGrallocHandles; OMX_INIT_STRUCT_PTR (&domxUseGrallocHandles, OMX_TI_PARAMUSENATIVEBUFFER); domxUseGrallocHandles.nPortIndex = mCameraAdapterParameters.mImagePortIndex; if (bufArr[0].type == CAMERA_BUFFER_ANW) { CAMHAL_LOGD ("Using ANW Buffers"); initInternalBuffers(mCameraAdapterParameters.mImagePortIndex); domxUseGrallocHandles.bEnable = OMX_TRUE; } else { CAMHAL_LOGD ("Using ION Buffers"); domxUseGrallocHandles.bEnable = OMX_FALSE; } eError = OMX_SetParameter(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE)OMX_TI_IndexUseNativeBuffers, &domxUseGrallocHandles); if (eError!=OMX_ErrorNone) { CAMHAL_LOGEB("OMX_SetParameter - %x", eError); } GOTO_EXIT_IF((eError!=OMX_ErrorNone), eError); } #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("Takepicture image port configuration: ", &bufArr->ppmStamp); #endif // Register for Image port ENABLE event ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, OMX_EventCmdComplete, OMX_CommandPortEnable, mCameraAdapterParameters.mImagePortIndex, mUseCaptureSem); // Enable Capture Port eError = OMX_SendCommand(mCameraAdapterParameters.mHandleComp, OMX_CommandPortEnable, mCameraAdapterParameters.mImagePortIndex, NULL); CAMHAL_LOGDB("OMX_UseBuffer = 0x%x", eError); GOTO_EXIT_IF(( eError != OMX_ErrorNone ), eError); for (int index = 0 ; index < imgCaptureData->mNumBufs ; index++) { OMX_BUFFERHEADERTYPE *pBufferHdr; CAMHAL_LOGDB("OMX_UseBuffer Capture address: 0x%x, size = %d", (unsigned int)bufArr[index].opaque, (int)imgCaptureData->mBufSize); eError = OMX_UseBuffer(mCameraAdapterParameters.mHandleComp, &pBufferHdr, mCameraAdapterParameters.mImagePortIndex, 0, imgCaptureData->mBufSize, (OMX_U8*)camera_buffer_get_omx_ptr(&bufArr[index])); CAMHAL_LOGDB("OMX_UseBuffer = 0x%x", eError); GOTO_EXIT_IF(( eError != OMX_ErrorNone ), eError); pBufferHdr->pAppPrivate = (OMX_PTR) &bufArr[index]; bufArr[index].index = index; pBufferHdr->nSize = sizeof(OMX_BUFFERHEADERTYPE); pBufferHdr->nVersion.s.nVersionMajor = 1 ; pBufferHdr->nVersion.s.nVersionMinor = 1 ; pBufferHdr->nVersion.s.nRevision = 0; pBufferHdr->nVersion.s.nStep = 0; imgCaptureData->mBufferHeader[index] = pBufferHdr; imgCaptureData->mStatus[index] = OMXCameraPortParameters::IDLE; } // Wait for the image port enable event CAMHAL_LOGDA("Waiting for port enable"); ret = mUseCaptureSem.WaitTimeout(OMX_CMD_TIMEOUT); // If somethiing bad happened while we wait if (mComponentState == OMX_StateInvalid) { CAMHAL_LOGEA("Invalid State after Enable Image Port Exitting!!!"); goto EXIT; } if (ret != NO_ERROR) { ret |= RemoveEvent(mCameraAdapterParameters.mHandleComp, OMX_EventCmdComplete, OMX_CommandPortEnable, mCameraAdapterParameters.mImagePortIndex, NULL); CAMHAL_LOGDA("Timeout expired on port enable"); goto EXIT; } CAMHAL_LOGDA("Port enabled"); #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("Takepicture image port enabled and buffers registered: ", &bufArr->ppmStamp); #endif if (mNextState != LOADED_REPROCESS_CAPTURE_STATE) { // Enable WB and vector shot extra data for metadata setExtraData(true, mCameraAdapterParameters.mImagePortIndex, OMX_WhiteBalance); #ifndef CAMERAHAL_TUNA setExtraData(true, mCameraAdapterParameters.mImagePortIndex, OMX_TI_LSCTable); #endif } #ifdef OMAP_ENHANCEMENT_CPCAM // CPCam mode only supports vector shot // Regular capture is not supported if ( (mCapMode == CP_CAM) && (mNextState != LOADED_REPROCESS_CAPTURE_STATE) ) { initVectorShot(); } #endif mCaptureBuffersAvailable.clear(); for (unsigned int i = 0; i < imgCaptureData->mMaxQueueable; i++ ) { mCaptureBuffersAvailable.add(&mCaptureBuffers[i], 0); } // initial ref count for undeqeueued buffers is 1 since buffer provider // is still holding on to it for (unsigned int i = imgCaptureData->mMaxQueueable; i < imgCaptureData->mNumBufs; i++ ) { mCaptureBuffersAvailable.add(&mCaptureBuffers[i], 1); } } if ( NO_ERROR == ret ) { ret = setupEXIF(); if ( NO_ERROR != ret ) { CAMHAL_LOGEB("Error configuring EXIF Buffer %x", ret); } } #ifdef OMAP_ENHANCEMENT_CPCAM // Choose proper single preview mode for cp capture (reproc or hs) if (( NO_ERROR == ret) && (OMXCameraAdapter::CP_CAM == mCapMode)) { OMX_TI_CONFIG_SINGLEPREVIEWMODETYPE singlePrevMode; OMX_INIT_STRUCT_PTR (&singlePrevMode, OMX_TI_CONFIG_SINGLEPREVIEWMODETYPE); if (mNextState == LOADED_CAPTURE_STATE) { singlePrevMode.eMode = OMX_TI_SinglePreviewMode_ImageCaptureHighSpeed; } else if (mNextState == LOADED_REPROCESS_CAPTURE_STATE) { singlePrevMode.eMode = OMX_TI_SinglePreviewMode_Reprocess; } else { CAMHAL_LOGE("Wrong state trying to start a capture in CPCAM mode?"); singlePrevMode.eMode = OMX_TI_SinglePreviewMode_ImageCaptureHighSpeed; } eError = OMX_SetConfig(mCameraAdapterParameters.mHandleComp, (OMX_INDEXTYPE) OMX_TI_IndexConfigSinglePreviewMode, &singlePrevMode); if ( OMX_ErrorNone != eError ) { CAMHAL_LOGEB("Error while configuring single preview mode 0x%x", eError); ret = Utils::ErrorUtils::omxToAndroidError(eError); } else { CAMHAL_LOGDA("single preview mode configured successfully"); } } #endif #if PPM_INSTRUMENTATION || PPM_INSTRUMENTATION_ABS CameraHal::PPM("Takepicture extra configs on image port done: ", &bufArr->ppmStamp); #endif mCaptureConfigured = true; #ifdef CAMERAHAL_USE_RAW_IMAGE_SAVING if (mRawCapture) { mCaptureConfigured = false; } #endif return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); EXIT: CAMHAL_LOGEB("Exiting function %s because of ret %d eError=%x", __FUNCTION__, ret, eError); setExtraData(false, mCameraAdapterParameters.mImagePortIndex, OMX_WhiteBalance); // TODO: WA: if domx client disables VectShotInfo metadata on the image port, this causes // VectShotInfo to be disabled internally on preview port also. Remove setting in OMXCapture // setExtraData(false, mCameraAdapterParameters.mImagePortIndex, OMX_TI_VectShotInfo); #ifndef CAMERAHAL_TUNA setExtraData(false, mCameraAdapterParameters.mImagePortIndex, OMX_TI_LSCTable); #endif //Release image buffers if ( NULL != mReleaseImageBuffersCallback ) { mReleaseImageBuffersCallback(mReleaseData); } performCleanupAfterError(); LOG_FUNCTION_NAME_EXIT; return (ret | Utils::ErrorUtils::omxToAndroidError(eError)); } status_t OMXCameraAdapter::UseBuffersRawCapture(CameraBuffer *bufArr, int num) { LOG_FUNCTION_NAME status_t ret; OMX_ERRORTYPE eError; OMXCameraPortParameters * imgRawCaptureData = NULL; Utils::Semaphore camSem; OMXCameraPortParameters cap; imgRawCaptureData = &mCameraAdapterParameters.mCameraPortParams[mCameraAdapterParameters.mVideoPortIndex]; if (mCaptureConfigured) { return NO_ERROR; } camSem.Create(); // mWaitingForSnapshot is true only when we're in the process of capturing if (mWaitingForSnapshot) { ///Register for Video port Disable event ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, (OMX_EVENTTYPE) OMX_EventCmdComplete, OMX_CommandPortDisable, mCameraAdapterParameters.mVideoPortIndex, camSem); ///Disable Capture Port eError = OMX_SendCommand(mCameraAdapterParameters.mHandleComp, OMX_CommandPortDisable, mCameraAdapterParameters.mVideoPortIndex, NULL); CAMHAL_LOGDA("Waiting for port disable"); //Wait for the image port enable event camSem.Wait(); CAMHAL_LOGDA("Port disabled"); } imgRawCaptureData->mNumBufs = num; CAMHAL_LOGDB("RAW Max sensor width = %d", (int)imgRawCaptureData->mWidth); CAMHAL_LOGDB("RAW Max sensor height = %d", (int)imgRawCaptureData->mHeight); ret = setFormat(OMX_CAMERA_PORT_VIDEO_OUT_VIDEO, *imgRawCaptureData); if (ret != NO_ERROR) { CAMHAL_LOGEB("setFormat() failed %d", ret); LOG_FUNCTION_NAME_EXIT return ret; } ///Register for Video port ENABLE event ret = RegisterForEvent(mCameraAdapterParameters.mHandleComp, (OMX_EVENTTYPE) OMX_EventCmdComplete, OMX_CommandPortEnable, mCameraAdapterParameters.mVideoPortIndex, camSem); ///Enable Video Capture Port eError = OMX_SendCommand(mCameraAdapterParameters.mHandleComp, OMX_CommandPortEnable, mCameraAdapterParameters.mVideoPortIndex, NULL); mCaptureBuffersLength = (int)imgRawCaptureData->mBufSize; for ( int index = 0 ; index < imgRawCaptureData->mNumBufs ; index++ ) { OMX_BUFFERHEADERTYPE *pBufferHdr; CAMHAL_LOGDB("OMX_UseBuffer rawCapture address: 0x%x, size = %d ", (unsigned int)bufArr[index].opaque, (int)imgRawCaptureData->mBufSize ); eError = OMX_UseBuffer( mCameraAdapterParameters.mHandleComp, &pBufferHdr, mCameraAdapterParameters.mVideoPortIndex, 0, mCaptureBuffersLength, (OMX_U8*)camera_buffer_get_omx_ptr(&bufArr[index])); if (eError != OMX_ErrorNone) { CAMHAL_LOGEB("OMX_UseBuffer = 0x%x", eError); } GOTO_EXIT_IF(( eError != OMX_ErrorNone ), eError); pBufferHdr->pAppPrivate = (OMX_PTR) &bufArr[index]; bufArr[index].index = index; pBufferHdr->nSize = sizeof(OMX_BUFFERHEADERTYPE); pBufferHdr->nVersion.s.nVersionMajor = 1 ; pBufferHdr->nVersion.s.nVersionMinor = 1 ; pBufferHdr->nVersion.s.nRevision = 0; pBufferHdr->nVersion.s.nStep = 0; imgRawCaptureData->mBufferHeader[index] = pBufferHdr; } //Wait for the image port enable event CAMHAL_LOGDA("Waiting for port enable"); camSem.Wait(); CAMHAL_LOGDA("Port enabled"); if (NO_ERROR == ret) { ret = setupEXIF(); if ( NO_ERROR != ret ) { CAMHAL_LOGEB("Error configuring EXIF Buffer %x", ret); } } mCapturedFrames = mBurstFrames; mBurstFramesQueued = 0; mCaptureConfigured = true; EXIT: if (eError != OMX_ErrorNone) { if ( NULL != mErrorNotifier ) { mErrorNotifier->errorNotify(eError); } } LOG_FUNCTION_NAME_EXIT return ret; } } // namespace Camera } // namespace Ti