/* ** ** Copyright 2008, The Android Open Source Project ** Copyright 2010, Samsung Electronics Co. LTD ** ** Licensed under the Apache License, Version 2.0 (the "License"); ** you may not use this file except in compliance with the License. ** You may obtain a copy of the License at ** ** http://www.apache.org/licenses/LICENSE-2.0 ** ** Unless required by applicable law or agreed to in writing, software ** distributed under the License is distributed on an "AS IS" BASIS, ** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ** See the License for the specific language governing permissions and ** limitations under the License. */ //#define LOG_NDEBUG 0 #define LOG_TAG "CameraHardwareSec" #include #include "SecCameraHWInterface.h" #include #include #include #include #include #define VIDEO_COMMENT_MARKER_H 0xFFBE #define VIDEO_COMMENT_MARKER_L 0xFFBF #define VIDEO_COMMENT_MARKER_LENGTH 4 #define JPEG_EOI_MARKER 0xFFD9 #define HIBYTE(x) (((x) >> 8) & 0xFF) #define LOBYTE(x) ((x) & 0xFF) #define BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR "0.10,1.20,Infinity" #define BACK_CAMERA_MACRO_FOCUS_DISTANCES_STR "0.10,0.20,Infinity" #define BACK_CAMERA_INFINITY_FOCUS_DISTANCES_STR "0.10,1.20,Infinity" #define FRONT_CAMERA_FOCUS_DISTANCES_STR "0.20,0.25,Infinity" // FIXME: // -- The actual preview color is set to YV12. The preview frames // returned via preview callback must be generated by color // conversion if the requested preview color format for the // preview frames is _not_ YV12. The reason that YV12 is used // for actual preview is because that is the only color format // supported by gralloc. Matching the preview cor format with // gralloc color format improves performance since no color // conversion is needed for preview. // // -- we only support two preview color formats that client // applications can set: NV21 and YUV420/YV12. namespace android { struct addrs { uint32_t type; // make sure that this is 4 byte. unsigned int addr_y; unsigned int addr_cbcr; unsigned int buf_index; unsigned int reserved; }; struct addrs_cap { unsigned int addr_y; unsigned int width; unsigned int height; }; static const int INITIAL_SKIP_FRAME = 3; static const int EFFECT_SKIP_FRAME = 1; gralloc_module_t const* CameraHardwareSec::mGrallocHal; CameraHardwareSec::CameraHardwareSec(int cameraId, camera_device_t *dev) : mCaptureInProgress(false), mParameters(), mFrameSizeDelta(0), mCameraSensorName(NULL), mSkipFrame(0), mNotifyCb(0), mDataCb(0), mDataCbTimestamp(0), mCallbackCookie(0), mMsgEnabled(0), mRecordRunning(false), mPostViewWidth(0), mPostViewHeight(0), mPostViewSize(0), mHalDevice(dev) { LOGV("%s :", __func__); int ret = 0; mPreviewWindow = NULL; mSecCamera = SecCamera::createInstance(); mRawHeap = NULL; mPreviewHeap = NULL; mRecordHeap = NULL; if (!mGrallocHal) { ret = hw_get_module(GRALLOC_HARDWARE_MODULE_ID, (const hw_module_t **)&mGrallocHal); if (ret) LOGE("ERR(%s):Fail on loading gralloc HAL", __func__); } ret = mSecCamera->initCamera(cameraId); if (ret < 0) { LOGE("ERR(%s):Fail on mSecCamera init", __func__); } mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize); LOGV("mPostViewWidth = %d mPostViewHeight = %d mPostViewSize = %d", mPostViewWidth,mPostViewHeight,mPostViewSize); initDefaultParameters(cameraId); mExitAutoFocusThread = false; mExitPreviewThread = false; /* whether the PreviewThread is active in preview or stopped. we * create the thread but it is initially in stopped state. */ mPreviewRunning = false; mPreviewStartDeferred = false; mPreviewThread = new PreviewThread(this); mAutoFocusThread = new AutoFocusThread(this); mPictureThread = new PictureThread(this); } int CameraHardwareSec::getCameraId() const { return mSecCamera->getCameraId(); } void CameraHardwareSec::initDefaultParameters(int cameraId) { if (mSecCamera == NULL) { LOGE("ERR(%s):mSecCamera object is NULL", __func__); return; } CameraParameters p; CameraParameters ip; mCameraSensorName = mSecCamera->getCameraSensorName(); LOGV("CameraSensorName: %s", mCameraSensorName); int preview_max_width = 0; int preview_max_height = 0; int snapshot_max_width = 0; int snapshot_max_height = 0; if (cameraId == SecCamera::CAMERA_ID_BACK) { p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_SIZES, "720x480,640x480,352x288,176x144"); p.set(CameraParameters::KEY_SUPPORTED_PICTURE_SIZES, "2560x1920,2048x1536,1600x1200,1280x960,640x480"); } else { p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_SIZES, "640x480,320x240,176x144"); p.set(CameraParameters::KEY_SUPPORTED_PICTURE_SIZES, "640x480"); } p.getSupportedPreviewSizes(mSupportedPreviewSizes); // If these fail, then we are using an invalid cameraId and we'll leave the // sizes at zero to catch the error. if (mSecCamera->getPreviewMaxSize(&preview_max_width, &preview_max_height) < 0) LOGE("getPreviewMaxSize fail (%d / %d) \n", preview_max_width, preview_max_height); if (mSecCamera->getSnapshotMaxSize(&snapshot_max_width, &snapshot_max_height) < 0) LOGE("getSnapshotMaxSize fail (%d / %d) \n", snapshot_max_width, snapshot_max_height); String8 previewColorString; previewColorString = CameraParameters::PIXEL_FORMAT_YUV420SP; previewColorString.append(","); previewColorString.append(CameraParameters::PIXEL_FORMAT_YUV420P); p.setPreviewFormat(CameraParameters::PIXEL_FORMAT_YUV420SP); mFrameSizeDelta = 16; p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FORMATS, previewColorString.string()); p.set(CameraParameters::KEY_VIDEO_FRAME_FORMAT, CameraParameters::PIXEL_FORMAT_YUV420P); p.setPreviewSize(preview_max_width, preview_max_height); p.setPictureFormat(CameraParameters::PIXEL_FORMAT_JPEG); p.setPictureSize(snapshot_max_width, snapshot_max_height); p.set(CameraParameters::KEY_JPEG_QUALITY, "100"); // maximum quality p.set(CameraParameters::KEY_SUPPORTED_PICTURE_FORMATS, CameraParameters::PIXEL_FORMAT_JPEG); String8 parameterString; if (cameraId == SecCamera::CAMERA_ID_BACK) { parameterString = CameraParameters::FOCUS_MODE_AUTO; parameterString.append(","); parameterString.append(CameraParameters::FOCUS_MODE_INFINITY); parameterString.append(","); parameterString.append(CameraParameters::FOCUS_MODE_MACRO); p.set(CameraParameters::KEY_SUPPORTED_FOCUS_MODES, parameterString.string()); p.set(CameraParameters::KEY_FOCUS_MODE, CameraParameters::FOCUS_MODE_AUTO); p.set(CameraParameters::KEY_FOCUS_DISTANCES, BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR); p.set(CameraParameters::KEY_SUPPORTED_JPEG_THUMBNAIL_SIZES, "320x240,0x0"); p.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, "320"); p.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, "240"); p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FRAME_RATES, "30"); p.setPreviewFrameRate(30); } else { parameterString = CameraParameters::FOCUS_MODE_FIXED; p.set(CameraParameters::KEY_SUPPORTED_FOCUS_MODES, parameterString.string()); p.set(CameraParameters::KEY_FOCUS_MODE, CameraParameters::FOCUS_MODE_FIXED); p.set(CameraParameters::KEY_FOCUS_DISTANCES, FRONT_CAMERA_FOCUS_DISTANCES_STR); p.set(CameraParameters::KEY_SUPPORTED_JPEG_THUMBNAIL_SIZES, "160x120,0x0"); p.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, "160"); p.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, "120"); p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FRAME_RATES, "15"); p.setPreviewFrameRate(15); } parameterString = CameraParameters::EFFECT_NONE; parameterString.append(","); parameterString.append(CameraParameters::EFFECT_MONO); parameterString.append(","); parameterString.append(CameraParameters::EFFECT_NEGATIVE); parameterString.append(","); parameterString.append(CameraParameters::EFFECT_SEPIA); p.set(CameraParameters::KEY_SUPPORTED_EFFECTS, parameterString.string()); if (cameraId == SecCamera::CAMERA_ID_BACK) { parameterString = CameraParameters::FLASH_MODE_ON; parameterString.append(","); parameterString.append(CameraParameters::FLASH_MODE_OFF); parameterString.append(","); parameterString.append(CameraParameters::FLASH_MODE_AUTO); parameterString.append(","); parameterString.append(CameraParameters::FLASH_MODE_TORCH); p.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, parameterString.string()); p.set(CameraParameters::KEY_FLASH_MODE, CameraParameters::FLASH_MODE_OFF); parameterString = CameraParameters::SCENE_MODE_AUTO; parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_PORTRAIT); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_LANDSCAPE); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_NIGHT); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_BEACH); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_SNOW); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_SUNSET); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_FIREWORKS); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_SPORTS); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_PARTY); parameterString.append(","); parameterString.append(CameraParameters::SCENE_MODE_CANDLELIGHT); p.set(CameraParameters::KEY_SUPPORTED_SCENE_MODES, parameterString.string()); p.set(CameraParameters::KEY_SCENE_MODE, CameraParameters::SCENE_MODE_AUTO); /* we have two ranges, 4-30fps for night mode and * 15-30fps for all others */ p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(15000,30000)"); p.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "15000,30000"); p.set(CameraParameters::KEY_FOCAL_LENGTH, "3.43"); } else { p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(7500,30000)"); p.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "7500,30000"); p.set(CameraParameters::KEY_FOCAL_LENGTH, "0.9"); } parameterString = CameraParameters::WHITE_BALANCE_AUTO; parameterString.append(","); parameterString.append(CameraParameters::WHITE_BALANCE_INCANDESCENT); parameterString.append(","); parameterString.append(CameraParameters::WHITE_BALANCE_FLUORESCENT); parameterString.append(","); parameterString.append(CameraParameters::WHITE_BALANCE_DAYLIGHT); parameterString.append(","); parameterString.append(CameraParameters::WHITE_BALANCE_CLOUDY_DAYLIGHT); p.set(CameraParameters::KEY_SUPPORTED_WHITE_BALANCE, parameterString.string()); ip.set("sharpness-min", 0); ip.set("sharpness-max", 4); ip.set("saturation-min", 0); ip.set("saturation-max", 4); ip.set("contrast-min", 0); ip.set("contrast-max", 4); p.set(CameraParameters::KEY_JPEG_THUMBNAIL_QUALITY, "100"); p.set(CameraParameters::KEY_ROTATION, 0); p.set(CameraParameters::KEY_WHITE_BALANCE, CameraParameters::WHITE_BALANCE_AUTO); p.set(CameraParameters::KEY_EFFECT, CameraParameters::EFFECT_NONE); ip.set("sharpness", SHARPNESS_DEFAULT); ip.set("contrast", CONTRAST_DEFAULT); ip.set("saturation", SATURATION_DEFAULT); ip.set("iso", "auto"); ip.set("metering", "center"); ip.set("wdr", 0); ip.set("chk_dataline", 0); if (cameraId == SecCamera::CAMERA_ID_FRONT) { ip.set("vtmode", 0); ip.set("blur", 0); } p.set(CameraParameters::KEY_HORIZONTAL_VIEW_ANGLE, "51.2"); p.set(CameraParameters::KEY_VERTICAL_VIEW_ANGLE, "39.4"); p.set(CameraParameters::KEY_EXPOSURE_COMPENSATION, "0"); p.set(CameraParameters::KEY_MAX_EXPOSURE_COMPENSATION, "4"); p.set(CameraParameters::KEY_MIN_EXPOSURE_COMPENSATION, "-4"); p.set(CameraParameters::KEY_EXPOSURE_COMPENSATION_STEP, "0.5"); mParameters = p; mInternalParameters = ip; /* make sure mSecCamera has all the settings we do. applications * aren't required to call setParameters themselves (only if they * want to change something. */ setParameters(p); mSecCamera->setISO(ISO_AUTO); mSecCamera->setMetering(METERING_CENTER); mSecCamera->setContrast(CONTRAST_DEFAULT); mSecCamera->setSharpness(SHARPNESS_DEFAULT); mSecCamera->setSaturation(SATURATION_DEFAULT); if (cameraId == SecCamera::CAMERA_ID_BACK) mSecCamera->setFrameRate(30); else mSecCamera->setFrameRate(15); } CameraHardwareSec::~CameraHardwareSec() { LOGV("%s", __func__); mSecCamera->DeinitCamera(); } status_t CameraHardwareSec::setPreviewWindow(preview_stream_ops *w) { int min_bufs; mPreviewWindow = w; LOGV("%s: mPreviewWindow %p", __func__, mPreviewWindow); if (!w) { LOGE("preview window is NULL!"); return OK; } mPreviewLock.lock(); if (mPreviewRunning && !mPreviewStartDeferred) { LOGI("stop preview (window change)"); stopPreviewInternal(); } if (w->get_min_undequeued_buffer_count(w, &min_bufs)) { LOGE("%s: could not retrieve min undequeued buffer count", __func__); return INVALID_OPERATION; } if (min_bufs >= kBufferCount) { LOGE("%s: min undequeued buffer count %d is too high (expecting at most %d)", __func__, min_bufs, kBufferCount - 1); } LOGV("%s: setting buffer count to %d", __func__, kBufferCount); if (w->set_buffer_count(w, kBufferCount)) { LOGE("%s: could not set buffer count", __func__); return INVALID_OPERATION; } int preview_width; int preview_height; mParameters.getPreviewSize(&preview_width, &preview_height); int hal_pixel_format = HAL_PIXEL_FORMAT_YV12; const char *str_preview_format = mParameters.getPreviewFormat(); LOGV("%s: preview format %s", __func__, str_preview_format); mFrameSizeDelta = 16; if (w->set_usage(w, GRALLOC_USAGE_SW_WRITE_OFTEN)) { LOGE("%s: could not set usage on gralloc buffer", __func__); return INVALID_OPERATION; } if (w->set_buffers_geometry(w, preview_width, preview_height, hal_pixel_format)) { LOGE("%s: could not set buffers geometry to %s", __func__, str_preview_format); return INVALID_OPERATION; } if (mPreviewRunning && mPreviewStartDeferred) { LOGV("start/resume preview"); status_t ret = startPreviewInternal(); if (ret == OK) { mPreviewStartDeferred = false; mPreviewCondition.signal(); } } mPreviewLock.unlock(); return OK; } void CameraHardwareSec::setCallbacks(camera_notify_callback notify_cb, camera_data_callback data_cb, camera_data_timestamp_callback data_cb_timestamp, camera_request_memory get_memory, void *user) { mNotifyCb = notify_cb; mDataCb = data_cb; mDataCbTimestamp = data_cb_timestamp; mGetMemoryCb = get_memory; mCallbackCookie = user; } void CameraHardwareSec::enableMsgType(int32_t msgType) { LOGV("%s : msgType = 0x%x, mMsgEnabled before = 0x%x", __func__, msgType, mMsgEnabled); mMsgEnabled |= msgType; LOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled); } void CameraHardwareSec::disableMsgType(int32_t msgType) { LOGV("%s : msgType = 0x%x, mMsgEnabled before = 0x%x", __func__, msgType, mMsgEnabled); mMsgEnabled &= ~msgType; LOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled); } bool CameraHardwareSec::msgTypeEnabled(int32_t msgType) { return (mMsgEnabled & msgType); } // --------------------------------------------------------------------------- void CameraHardwareSec::setSkipFrame(int frame) { Mutex::Autolock lock(mSkipFrameLock); if (frame < mSkipFrame) return; mSkipFrame = frame; } int CameraHardwareSec::previewThreadWrapper() { LOGI("%s: starting", __func__); while (1) { mPreviewLock.lock(); while (!mPreviewRunning) { LOGI("%s: calling mSecCamera->stopPreview() and waiting", __func__); mSecCamera->stopPreview(); /* signal that we're stopping */ mPreviewStoppedCondition.signal(); mPreviewCondition.wait(mPreviewLock); LOGI("%s: return from wait", __func__); } mPreviewLock.unlock(); if (mExitPreviewThread) { LOGI("%s: exiting", __func__); mSecCamera->stopPreview(); return 0; } previewThread(); } } int CameraHardwareSec::previewThread() { int index; nsecs_t timestamp; unsigned int phyYAddr; unsigned int phyCAddr; struct addrs *addrs; index = mSecCamera->getPreview(); if (index < 0) { LOGE("ERR(%s):Fail on SecCamera->getPreview()", __func__); return UNKNOWN_ERROR; } // LOGV("%s: index %d", __func__, index); mSkipFrameLock.lock(); if (mSkipFrame > 0) { mSkipFrame--; mSkipFrameLock.unlock(); LOGV("%s: index %d skipping frame", __func__, index); return NO_ERROR; } mSkipFrameLock.unlock(); timestamp = systemTime(SYSTEM_TIME_MONOTONIC); phyYAddr = mSecCamera->getPhyAddrY(index); phyCAddr = mSecCamera->getPhyAddrC(index); if (phyYAddr == 0xffffffff || phyCAddr == 0xffffffff) { LOGE("ERR(%s):Fail on SecCamera getPhyAddr Y addr = %0x C addr = %0x", __func__, phyYAddr, phyCAddr); return UNKNOWN_ERROR; } int width, height, frame_size, offset; mSecCamera->getPreviewSize(&width, &height, &frame_size); offset = (frame_size + mFrameSizeDelta) * index; #if 0 // FIXME: this does not seem to be necessary. Is it? memcpy((char *)mPreviewHeap->data + offset + frame_size, &phyYAddr, 4); memcpy((char *)mPreviewHeap->data + offset + frame_size + 4, &phyCAddr, 4); #endif if (mPreviewWindow && mGrallocHal) { buffer_handle_t *buf_handle; int stride; if (0 != mPreviewWindow->dequeue_buffer(mPreviewWindow, &buf_handle, &stride)) { LOGE("Could not dequeue gralloc buffer!\n"); goto callbacks; } void *vaddr; if (!mGrallocHal->lock(mGrallocHal, *buf_handle, GRALLOC_USAGE_SW_WRITE_OFTEN, 0, 0, width, height, &vaddr)) { char *frame = ((char *)mPreviewHeap->data) + offset; int total = frame_size + mFrameSizeDelta; // the code below assumes YUV, not RGB { int h; char *src = frame; char *ptr = (char *)vaddr; // Copy the Y plane, while observing the stride for (h = 0; h < height; h++) { memcpy(ptr, src, width); ptr += stride; src += width; } { // U char *v = ptr; ptr += stride * height / 4; for (h = 0; h < height / 2; h++) { memcpy(ptr, src, width / 2); ptr += stride / 2; src += width / 2; } // V ptr = v; for (h = 0; h < height / 2; h++) { memcpy(ptr, src, width / 2); ptr += stride / 2; src += width / 2; } } } mGrallocHal->unlock(mGrallocHal, *buf_handle); } else LOGE("%s: could not obtain gralloc buffer", __func__); if (0 != mPreviewWindow->enqueue_buffer(mPreviewWindow, buf_handle)) { LOGE("Could not enqueue gralloc buffer!\n"); goto callbacks; } } callbacks: // Notify the client of a new frame. if (mMsgEnabled & CAMERA_MSG_PREVIEW_FRAME) { const char * preview_format = mParameters.getPreviewFormat(); if (!strcmp(preview_format, CameraParameters::PIXEL_FORMAT_YUV420SP)) { // Color conversion from YUV420 to NV21 char *vu = ((char *)mPreviewHeap->data) + offset + width * height; const int uv_size = (width * height) >> 1; char saved_uv[uv_size]; memcpy(saved_uv, vu, uv_size); char *u = saved_uv; char *v = u + (uv_size >> 1); int h = 0; while (h < width * height / 4) { *vu++ = *v++; *vu++ = *u++; ++h; } } mDataCb(CAMERA_MSG_PREVIEW_FRAME, mPreviewHeap, index, NULL, mCallbackCookie); } Mutex::Autolock lock(mRecordLock); if (mRecordRunning == true) { index = mSecCamera->getRecordFrame(); if (index < 0) { LOGE("ERR(%s):Fail on SecCamera->getRecord()", __func__); return UNKNOWN_ERROR; } phyYAddr = mSecCamera->getRecPhyAddrY(index); phyCAddr = mSecCamera->getRecPhyAddrC(index); if (phyYAddr == 0xffffffff || phyCAddr == 0xffffffff) { LOGE("ERR(%s):Fail on SecCamera getRectPhyAddr Y addr = %0x C addr = %0x", __func__, phyYAddr, phyCAddr); return UNKNOWN_ERROR; } addrs = (struct addrs *)mRecordHeap->data; addrs[index].type = kMetadataBufferTypeCameraSource; addrs[index].addr_y = phyYAddr; addrs[index].addr_cbcr = phyCAddr; addrs[index].buf_index = index; // Notify the client of a new frame. if (mMsgEnabled & CAMERA_MSG_VIDEO_FRAME) { mDataCbTimestamp(timestamp, CAMERA_MSG_VIDEO_FRAME, mRecordHeap, index, mCallbackCookie); } else { mSecCamera->releaseRecordFrame(index); } } return NO_ERROR; } status_t CameraHardwareSec::startPreview() { int ret = 0; //s1 [Apply factory standard] LOGV("%s :", __func__); if (waitCaptureCompletion() != NO_ERROR) { return TIMED_OUT; } mPreviewLock.lock(); if (mPreviewRunning) { // already running LOGE("%s : preview thread already running", __func__); mPreviewLock.unlock(); return INVALID_OPERATION; } mPreviewRunning = true; mPreviewStartDeferred = false; if (!mPreviewWindow) { LOGI("%s : deferring", __func__); mPreviewStartDeferred = true; mPreviewLock.unlock(); return NO_ERROR; } ret = startPreviewInternal(); if (ret == OK) mPreviewCondition.signal(); mPreviewLock.unlock(); return ret; } status_t CameraHardwareSec::startPreviewInternal() { LOGV("%s", __func__); int ret = mSecCamera->startPreview(); LOGV("%s : mSecCamera->startPreview() returned %d", __func__, ret); if (ret < 0) { LOGE("ERR(%s):Fail on mSecCamera->startPreview()", __func__); return UNKNOWN_ERROR; } setSkipFrame(INITIAL_SKIP_FRAME); int width, height, frame_size; mSecCamera->getPreviewSize(&width, &height, &frame_size); LOGD("mPreviewHeap(fd(%d), size(%d), width(%d), height(%d))", mSecCamera->getCameraFd(), frame_size + mFrameSizeDelta, width, height); if (mPreviewHeap) { mPreviewHeap->release(mPreviewHeap); mPreviewHeap = 0; } mPreviewHeap = mGetMemoryCb((int)mSecCamera->getCameraFd(), frame_size, kBufferCount, 0); // no cookie mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize); LOGV("CameraHardwareSec: mPostViewWidth = %d mPostViewHeight = %d mPostViewSize = %d", mPostViewWidth,mPostViewHeight,mPostViewSize); return NO_ERROR; } void CameraHardwareSec::stopPreviewInternal() { LOGV("%s :", __func__); /* request that the preview thread stop. */ if (mPreviewRunning) { mPreviewRunning = false; if (!mPreviewStartDeferred) { mPreviewCondition.signal(); /* wait until preview thread is stopped */ mPreviewStoppedCondition.wait(mPreviewLock); } else LOGV("%s : preview running but deferred, doing nothing", __func__); } else LOGI("%s : preview not running, doing nothing", __func__); } void CameraHardwareSec::stopPreview() { LOGV("%s :", __func__); /* request that the preview thread stop. */ mPreviewLock.lock(); stopPreviewInternal(); mPreviewLock.unlock(); } bool CameraHardwareSec::previewEnabled() { Mutex::Autolock lock(mPreviewLock); LOGV("%s : %d", __func__, mPreviewRunning); return mPreviewRunning; } // --------------------------------------------------------------------------- status_t CameraHardwareSec::startRecording() { LOGV("%s :", __func__); Mutex::Autolock lock(mRecordLock); if (mRecordHeap) { mRecordHeap->release(mRecordHeap); mRecordHeap = 0; } mRecordHeap = mGetMemoryCb(-1, sizeof(struct addrs), kBufferCount, NULL); if (!mRecordHeap) { LOGE("ERR(%s): Record heap creation fail", __func__); return UNKNOWN_ERROR; } if (mRecordRunning == false) { if (mSecCamera->startRecord() < 0) { LOGE("ERR(%s):Fail on mSecCamera->startRecord()", __func__); return UNKNOWN_ERROR; } mRecordRunning = true; } return NO_ERROR; } void CameraHardwareSec::stopRecording() { LOGV("%s :", __func__); Mutex::Autolock lock(mRecordLock); if (mRecordRunning == true) { if (mSecCamera->stopRecord() < 0) { LOGE("ERR(%s):Fail on mSecCamera->stopRecord()", __func__); return; } mRecordRunning = false; } } bool CameraHardwareSec::recordingEnabled() { LOGV("%s :", __func__); return mRecordRunning; } void CameraHardwareSec::releaseRecordingFrame(const void *opaque) { struct addrs *addrs = (struct addrs *)opaque; mSecCamera->releaseRecordFrame(addrs->buf_index); } // --------------------------------------------------------------------------- int CameraHardwareSec::autoFocusThread() { int count =0; int af_status =0 ; LOGV("%s : starting", __func__); /* block until we're told to start. we don't want to use * a restartable thread and requestExitAndWait() in cancelAutoFocus() * because it would cause deadlock between our callbacks and the * caller of cancelAutoFocus() which both want to grab the same lock * in CameraServices layer. */ mFocusLock.lock(); /* check early exit request */ if (mExitAutoFocusThread) { mFocusLock.unlock(); LOGV("%s : exiting on request0", __func__); return NO_ERROR; } mFocusCondition.wait(mFocusLock); /* check early exit request */ if (mExitAutoFocusThread) { mFocusLock.unlock(); LOGV("%s : exiting on request1", __func__); return NO_ERROR; } mFocusLock.unlock(); LOGV("%s : calling setAutoFocus", __func__); if (mSecCamera->setAutofocus() < 0) { LOGE("ERR(%s):Fail on mSecCamera->setAutofocus()", __func__); return UNKNOWN_ERROR; } af_status = mSecCamera->getAutoFocusResult(); if (af_status == 0x01) { LOGV("%s : AF Success!!", __func__); if (mMsgEnabled & CAMERA_MSG_FOCUS) mNotifyCb(CAMERA_MSG_FOCUS, true, 0, mCallbackCookie); } else if (af_status == 0x02) { LOGV("%s : AF Cancelled !!", __func__); if (mMsgEnabled & CAMERA_MSG_FOCUS) { /* CAMERA_MSG_FOCUS only takes a bool. true for * finished and false for failure. cancel is still * considered a true result. */ mNotifyCb(CAMERA_MSG_FOCUS, true, 0, mCallbackCookie); } } else { LOGV("%s : AF Fail !!", __func__); LOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled); if (mMsgEnabled & CAMERA_MSG_FOCUS) mNotifyCb(CAMERA_MSG_FOCUS, false, 0, mCallbackCookie); } LOGV("%s : exiting with no error", __func__); return NO_ERROR; } status_t CameraHardwareSec::autoFocus() { LOGV("%s :", __func__); /* signal autoFocusThread to run once */ mFocusCondition.signal(); return NO_ERROR; } /* 2009.10.14 by icarus for added interface */ status_t CameraHardwareSec::cancelAutoFocus() { LOGV("%s :", __func__); // cancelAutoFocus should be allowed after preview is started. But if // the preview is deferred, cancelAutoFocus will fail. Ignore it if that is // the case. if (mPreviewRunning && mPreviewStartDeferred) return NO_ERROR; if (mSecCamera->cancelAutofocus() < 0) { LOGE("ERR(%s):Fail on mSecCamera->cancelAutofocus()", __func__); return UNKNOWN_ERROR; } return NO_ERROR; } int CameraHardwareSec::save_jpeg( unsigned char *real_jpeg, int jpeg_size) { FILE *yuv_fp = NULL; char filename[100], *buffer = NULL; /* file create/open, note to "wb" */ yuv_fp = fopen("/data/camera_dump.jpeg", "wb"); if (yuv_fp == NULL) { LOGE("Save jpeg file open error"); return -1; } LOGV("[BestIQ] real_jpeg size ========> %d\n", jpeg_size); buffer = (char *) malloc(jpeg_size); if (buffer == NULL) { LOGE("Save YUV] buffer alloc failed"); if (yuv_fp) fclose(yuv_fp); return -1; } memcpy(buffer, real_jpeg, jpeg_size); fflush(stdout); fwrite(buffer, 1, jpeg_size, yuv_fp); fflush(yuv_fp); if (yuv_fp) fclose(yuv_fp); if (buffer) free(buffer); return 0; } void CameraHardwareSec::save_postview(const char *fname, uint8_t *buf, uint32_t size) { int nw; int cnt = 0; uint32_t written = 0; LOGD("opening file [%s]\n", fname); int fd = open(fname, O_RDWR | O_CREAT); if (fd < 0) { LOGE("failed to create file [%s]: %s", fname, strerror(errno)); return; } LOGD("writing %d bytes to file [%s]\n", size, fname); while (written < size) { nw = ::write(fd, buf + written, size - written); if (nw < 0) { LOGE("failed to write to file %d [%s]: %s",written,fname, strerror(errno)); break; } written += nw; cnt++; } LOGD("done writing %d bytes to file [%s] in %d passes\n",size, fname, cnt); ::close(fd); } bool CameraHardwareSec::scaleDownYuv422(char *srcBuf, uint32_t srcWidth, uint32_t srcHeight, char *dstBuf, uint32_t dstWidth, uint32_t dstHeight) { int32_t step_x, step_y; int32_t iXsrc, iXdst; int32_t x, y, src_y_start_pos, dst_pos, src_pos; if (dstWidth % 2 != 0 || dstHeight % 2 != 0){ LOGE("scale_down_yuv422: invalid width, height for scaling"); return false; } step_x = srcWidth / dstWidth; step_y = srcHeight / dstHeight; dst_pos = 0; for (uint32_t y = 0; y < dstHeight; y++) { src_y_start_pos = (y * step_y * (srcWidth * 2)); for (uint32_t x = 0; x < dstWidth; x += 2) { src_pos = src_y_start_pos + (x * (step_x * 2)); dstBuf[dst_pos++] = srcBuf[src_pos ]; dstBuf[dst_pos++] = srcBuf[src_pos + 1]; dstBuf[dst_pos++] = srcBuf[src_pos + 2]; dstBuf[dst_pos++] = srcBuf[src_pos + 3]; } } return true; } bool CameraHardwareSec::YUY2toNV21(void *srcBuf, void *dstBuf, uint32_t srcWidth, uint32_t srcHeight) { int32_t x, y, src_y_start_pos, dst_cbcr_pos, dst_pos, src_pos; unsigned char *srcBufPointer = (unsigned char *)srcBuf; unsigned char *dstBufPointer = (unsigned char *)dstBuf; dst_pos = 0; dst_cbcr_pos = srcWidth*srcHeight; for (uint32_t y = 0; y < srcHeight; y++) { src_y_start_pos = (y * (srcWidth * 2)); for (uint32_t x = 0; x < (srcWidth * 2); x += 2) { src_pos = src_y_start_pos + x; dstBufPointer[dst_pos++] = srcBufPointer[src_pos]; } } for (uint32_t y = 0; y < srcHeight; y += 2) { src_y_start_pos = (y * (srcWidth * 2)); for (uint32_t x = 0; x < (srcWidth * 2); x += 4) { src_pos = src_y_start_pos + x; dstBufPointer[dst_cbcr_pos++] = srcBufPointer[src_pos + 3]; dstBufPointer[dst_cbcr_pos++] = srcBufPointer[src_pos + 1]; } } return true; } int CameraHardwareSec::pictureThread() { LOGV("%s :", __func__); int jpeg_size = 0; int ret = NO_ERROR; unsigned char *jpeg_data = NULL; int postview_offset = 0; unsigned char *postview_data = NULL; unsigned char *addr = NULL; int mPostViewWidth, mPostViewHeight, mPostViewSize; int mThumbWidth, mThumbHeight, mThumbSize; int cap_width, cap_height, cap_frame_size; int JpegImageSize, JpegExifSize; bool isLSISensor = false; unsigned int output_size = 0; mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize); mSecCamera->getThumbnailConfig(&mThumbWidth, &mThumbHeight, &mThumbSize); int postviewHeapSize = mPostViewSize; mSecCamera->getSnapshotSize(&cap_width, &cap_height, &cap_frame_size); int mJpegHeapSize; if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) mJpegHeapSize = cap_frame_size * SecCamera::getJpegRatio(); else mJpegHeapSize = cap_frame_size; LOG_TIME_DEFINE(0) LOG_TIME_START(0) // sp buffer = new MemoryBase(mRawHeap, 0, mPostViewSize + 8); struct addrs_cap *addrs = (struct addrs_cap *)mRawHeap->data; addrs[0].width = mPostViewWidth; addrs[0].height = mPostViewHeight; LOGV("[5B] mPostViewWidth = %d mPostViewHeight = %d\n",mPostViewWidth,mPostViewHeight); camera_memory_t *JpegHeap = mGetMemoryCb(-1, mJpegHeapSize, 1, 0); sp PostviewHeap = new MemoryHeapBase(mPostViewSize); sp ThumbnailHeap = new MemoryHeapBase(mThumbSize); LOG_TIME_DEFINE(1) LOG_TIME_START(1) int picture_size, picture_width, picture_height; mSecCamera->getSnapshotSize(&picture_width, &picture_height, &picture_size); int picture_format = mSecCamera->getSnapshotPixelFormat(); unsigned int phyAddr; // Modified the shutter sound timing for Jpeg capture if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) mSecCamera->setSnapshotCmd(); if (mMsgEnabled & CAMERA_MSG_SHUTTER) { mNotifyCb(CAMERA_MSG_SHUTTER, 0, 0, mCallbackCookie); } if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK){ jpeg_data = mSecCamera->getJpeg(&jpeg_size, &phyAddr); if (jpeg_data == NULL) { LOGE("ERR(%s):Fail on SecCamera->getSnapshot()", __func__); ret = UNKNOWN_ERROR; goto out; } } else { if (mSecCamera->getSnapshotAndJpeg((unsigned char*)PostviewHeap->base(), (unsigned char*)JpegHeap->data, &output_size) < 0) { ret = UNKNOWN_ERROR; goto out; } LOGI("snapshotandjpeg done\n"); } LOG_TIME_END(1) LOG_CAMERA("getSnapshotAndJpeg interval: %lu us", LOG_TIME(1)); if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) { isLSISensor = !strncmp((const char*)mCameraSensorName, "S5K4ECGX", 8); if(isLSISensor) { LOGI("== Camera Sensor Detect %s - Samsung LSI SOC 5M ==\n", mCameraSensorName); // LSI 5M SOC if (!SplitFrame(jpeg_data, SecCamera::getInterleaveDataSize(), SecCamera::getJpegLineLength(), mPostViewWidth * 2, mPostViewWidth, JpegHeap->data, &JpegImageSize, PostviewHeap->base(), &mPostViewSize)) { ret = UNKNOWN_ERROR; goto out; } } else { LOGI("== Camera Sensor Detect %s Sony SOC 5M ==\n", mCameraSensorName); decodeInterleaveData(jpeg_data, SecCamera::getInterleaveDataSize(), mPostViewWidth, mPostViewHeight, &JpegImageSize, JpegHeap->data, PostviewHeap->base()); } } else { JpegImageSize = static_cast(output_size); } scaleDownYuv422((char *)PostviewHeap->base(), mPostViewWidth, mPostViewHeight, (char *)ThumbnailHeap->base(), mThumbWidth, mThumbHeight); memcpy(mRawHeap->data, PostviewHeap->base(), postviewHeapSize); if (mMsgEnabled & CAMERA_MSG_RAW_IMAGE) { mDataCb(CAMERA_MSG_RAW_IMAGE, mRawHeap, 0, NULL, mCallbackCookie); } else if (mMsgEnabled & CAMERA_MSG_RAW_IMAGE_NOTIFY) { mNotifyCb(CAMERA_MSG_RAW_IMAGE_NOTIFY, 0, 0, mCallbackCookie); } if (mMsgEnabled & CAMERA_MSG_COMPRESSED_IMAGE) { camera_memory_t *ExifHeap = mGetMemoryCb(-1, EXIF_FILE_SIZE + JPG_STREAM_BUF_SIZE, 1, 0); JpegExifSize = mSecCamera->getExif((unsigned char *)ExifHeap->data, (unsigned char *)ThumbnailHeap->base()); LOGV("JpegExifSize=%d", JpegExifSize); if (JpegExifSize < 0) { ret = UNKNOWN_ERROR; ExifHeap->release(ExifHeap); goto out; } camera_memory_t *mem = mGetMemoryCb(-1, JpegImageSize + JpegExifSize, 1, 0); uint8_t *ptr = (uint8_t *) mem->data; memcpy(ptr, JpegHeap->data, 2); ptr += 2; memcpy(ptr, ExifHeap->data, JpegExifSize); ptr += JpegExifSize; memcpy(ptr, (uint8_t *) JpegHeap->data + 2, JpegImageSize - 2); mDataCb(CAMERA_MSG_COMPRESSED_IMAGE, mem, 0, NULL, mCallbackCookie); mem->release(mem); ExifHeap->release(ExifHeap); } LOG_TIME_END(0) LOG_CAMERA("pictureThread interval: %lu us", LOG_TIME(0)); LOGV("%s : pictureThread end", __func__); out: JpegHeap->release(JpegHeap); mSecCamera->endSnapshot(); mCaptureLock.lock(); mCaptureInProgress = false; mCaptureCondition.broadcast(); mCaptureLock.unlock(); return ret; } status_t CameraHardwareSec::waitCaptureCompletion() { // 5 seconds timeout nsecs_t endTime = 5000000000LL + systemTime(SYSTEM_TIME_MONOTONIC); Mutex::Autolock lock(mCaptureLock); while (mCaptureInProgress) { nsecs_t remainingTime = endTime - systemTime(SYSTEM_TIME_MONOTONIC); if (remainingTime <= 0) { LOGE("Timed out waiting picture thread."); return TIMED_OUT; } LOGD("Waiting for picture thread to complete."); mCaptureCondition.waitRelative(mCaptureLock, remainingTime); } return NO_ERROR; } status_t CameraHardwareSec::takePicture() { LOGV("%s :", __func__); stopPreview(); if (!mRawHeap) { int rawHeapSize = mPostViewSize; LOGV("mRawHeap : MemoryHeapBase(previewHeapSize(%d))", rawHeapSize); mRawHeap = mGetMemoryCb(-1, rawHeapSize, 1, 0); if (!mRawHeap) { LOGE("ERR(%s): Raw heap creation fail", __func__); } } if (waitCaptureCompletion() != NO_ERROR) { return TIMED_OUT; } if (mPictureThread->run("CameraPictureThread", PRIORITY_DEFAULT) != NO_ERROR) { LOGE("%s : couldn't run picture thread", __func__); return INVALID_OPERATION; } mCaptureLock.lock(); mCaptureInProgress = true; mCaptureLock.unlock(); return NO_ERROR; } status_t CameraHardwareSec::cancelPicture() { LOGV("%s", __func__); if (mPictureThread.get()) { LOGV("%s: waiting for picture thread to exit", __func__); mPictureThread->requestExitAndWait(); LOGV("%s: picture thread has exited", __func__); } return NO_ERROR; } bool CameraHardwareSec::CheckVideoStartMarker(unsigned char *pBuf) { if (!pBuf) { LOGE("CheckVideoStartMarker() => pBuf is NULL\n"); return false; } if (HIBYTE(VIDEO_COMMENT_MARKER_H) == * pBuf && LOBYTE(VIDEO_COMMENT_MARKER_H) == *(pBuf + 1) && HIBYTE(VIDEO_COMMENT_MARKER_L) == *(pBuf + 2) && LOBYTE(VIDEO_COMMENT_MARKER_L) == *(pBuf + 3)) return true; return false; } bool CameraHardwareSec::CheckEOIMarker(unsigned char *pBuf) { if (!pBuf) { LOGE("CheckEOIMarker() => pBuf is NULL\n"); return false; } // EOI marker [FF D9] if (HIBYTE(JPEG_EOI_MARKER) == *pBuf && LOBYTE(JPEG_EOI_MARKER) == *(pBuf + 1)) return true; return false; } bool CameraHardwareSec::FindEOIMarkerInJPEG(unsigned char *pBuf, int dwBufSize, int *pnJPEGsize) { if (NULL == pBuf || 0 >= dwBufSize) { LOGE("FindEOIMarkerInJPEG() => There is no contents."); return false; } unsigned char *pBufEnd = pBuf + dwBufSize; while (pBuf < pBufEnd) { if (CheckEOIMarker(pBuf++)) return true; (*pnJPEGsize)++; } return false; } bool CameraHardwareSec::SplitFrame(unsigned char *pFrame, int dwSize, int dwJPEGLineLength, int dwVideoLineLength, int dwVideoHeight, void *pJPEG, int *pdwJPEGSize, void *pVideo, int *pdwVideoSize) { LOGV("===========SplitFrame Start=============="); if (NULL == pFrame || 0 >= dwSize) { LOGE("There is no contents (pFrame=%p, dwSize=%d", pFrame, dwSize); return false; } if (0 == dwJPEGLineLength || 0 == dwVideoLineLength) { LOGE("There in no input information for decoding interleaved jpeg"); return false; } unsigned char *pSrc = pFrame; unsigned char *pSrcEnd = pFrame + dwSize; unsigned char *pJ = (unsigned char *)pJPEG; int dwJSize = 0; unsigned char *pV = (unsigned char *)pVideo; int dwVSize = 0; bool bRet = false; bool isFinishJpeg = false; while (pSrc < pSrcEnd) { // Check video start marker if (CheckVideoStartMarker(pSrc)) { int copyLength; if (pSrc + dwVideoLineLength <= pSrcEnd) copyLength = dwVideoLineLength; else copyLength = pSrcEnd - pSrc - VIDEO_COMMENT_MARKER_LENGTH; // Copy video data if (pV) { memcpy(pV, pSrc + VIDEO_COMMENT_MARKER_LENGTH, copyLength); pV += copyLength; dwVSize += copyLength; } pSrc += copyLength + VIDEO_COMMENT_MARKER_LENGTH; } else { // Copy pure JPEG data int size = 0; int dwCopyBufLen = dwJPEGLineLength <= pSrcEnd-pSrc ? dwJPEGLineLength : pSrcEnd - pSrc; if (FindEOIMarkerInJPEG((unsigned char *)pSrc, dwCopyBufLen, &size)) { isFinishJpeg = true; size += 2; // to count EOF marker size } else { if ((dwCopyBufLen == 1) && (pJPEG < pJ)) { unsigned char checkBuf[2] = { *(pJ - 1), *pSrc }; if (CheckEOIMarker(checkBuf)) isFinishJpeg = true; } size = dwCopyBufLen; } memcpy(pJ, pSrc, size); dwJSize += size; pJ += dwCopyBufLen; pSrc += dwCopyBufLen; } if (isFinishJpeg) break; } if (isFinishJpeg) { bRet = true; if(pdwJPEGSize) *pdwJPEGSize = dwJSize; if(pdwVideoSize) *pdwVideoSize = dwVSize; } else { LOGE("DecodeInterleaveJPEG_WithOutDT() => Can not find EOI"); bRet = false; if(pdwJPEGSize) *pdwJPEGSize = 0; if(pdwVideoSize) *pdwVideoSize = 0; } LOGV("===========SplitFrame end=============="); return bRet; } int CameraHardwareSec::decodeInterleaveData(unsigned char *pInterleaveData, int interleaveDataSize, int yuvWidth, int yuvHeight, int *pJpegSize, void *pJpegData, void *pYuvData) { if (pInterleaveData == NULL) return false; bool ret = true; unsigned int *interleave_ptr = (unsigned int *)pInterleaveData; unsigned char *jpeg_ptr = (unsigned char *)pJpegData; unsigned char *yuv_ptr = (unsigned char *)pYuvData; unsigned char *p; int jpeg_size = 0; int yuv_size = 0; int i = 0; LOGV("decodeInterleaveData Start~~~"); while (i < interleaveDataSize) { if ((*interleave_ptr == 0xFFFFFFFF) || (*interleave_ptr == 0x02FFFFFF) || (*interleave_ptr == 0xFF02FFFF)) { // Padding Data // LOGE("%d(%x) padding data\n", i, *interleave_ptr); interleave_ptr++; i += 4; } else if ((*interleave_ptr & 0xFFFF) == 0x05FF) { // Start-code of YUV Data // LOGE("%d(%x) yuv data\n", i, *interleave_ptr); p = (unsigned char *)interleave_ptr; p += 2; i += 2; // Extract YUV Data if (pYuvData != NULL) { memcpy(yuv_ptr, p, yuvWidth * 2); yuv_ptr += yuvWidth * 2; yuv_size += yuvWidth * 2; } p += yuvWidth * 2; i += yuvWidth * 2; // Check End-code of YUV Data if ((*p == 0xFF) && (*(p + 1) == 0x06)) { interleave_ptr = (unsigned int *)(p + 2); i += 2; } else { ret = false; break; } } else { // Extract JPEG Data // LOGE("%d(%x) jpg data, jpeg_size = %d bytes\n", i, *interleave_ptr, jpeg_size); if (pJpegData != NULL) { memcpy(jpeg_ptr, interleave_ptr, 4); jpeg_ptr += 4; jpeg_size += 4; } interleave_ptr++; i += 4; } } if (ret) { if (pJpegData != NULL) { // Remove Padding after EOI for (i = 0; i < 3; i++) { if (*(--jpeg_ptr) != 0xFF) { break; } jpeg_size--; } *pJpegSize = jpeg_size; } // Check YUV Data Size if (pYuvData != NULL) { if (yuv_size != (yuvWidth * yuvHeight * 2)) { ret = false; } } } LOGV("decodeInterleaveData End~~~"); return ret; } status_t CameraHardwareSec::dump(int fd) const { const size_t SIZE = 256; char buffer[SIZE]; String8 result; const Vector args; if (mSecCamera != 0) { mSecCamera->dump(fd); mParameters.dump(fd, args); mInternalParameters.dump(fd, args); snprintf(buffer, 255, " preview running(%s)\n", mPreviewRunning?"true": "false"); result.append(buffer); } else { result.append("No camera client yet.\n"); } write(fd, result.string(), result.size()); return NO_ERROR; } bool CameraHardwareSec::isSupportedPreviewSize(const int width, const int height) const { unsigned int i; for (i = 0; i < mSupportedPreviewSizes.size(); i++) { if (mSupportedPreviewSizes[i].width == width && mSupportedPreviewSizes[i].height == height) return true; } return false; } bool CameraHardwareSec::isSupportedParameter(const char * const parm, const char * const supported_parm) const { const char *pStart; const char *pEnd; if (!parm || !supported_parm) return false; pStart = supported_parm; while (true) { pEnd = strchr(pStart, ','); if (!pEnd) { if (!strcmp(parm, pStart)) return true; else return false; } if (!strncmp(parm, pStart, pEnd - pStart)) { return true; } pStart = pEnd + 1; } /* NOTREACHED */ } status_t CameraHardwareSec::setParameters(const CameraParameters& params) { LOGV("%s :", __func__); status_t ret = NO_ERROR; /* if someone calls us while picture thread is running, it could screw * up the sensor quite a bit so return error. */ if (waitCaptureCompletion() != NO_ERROR) { return TIMED_OUT; } // preview size int new_preview_width = 0; int new_preview_height = 0; params.getPreviewSize(&new_preview_width, &new_preview_height); const char *new_str_preview_format = params.getPreviewFormat(); LOGV("%s : new_preview_width x new_preview_height = %dx%d, format = %s", __func__, new_preview_width, new_preview_height, new_str_preview_format); if (strcmp(new_str_preview_format, CameraParameters::PIXEL_FORMAT_YUV420SP) && strcmp(new_str_preview_format, CameraParameters::PIXEL_FORMAT_YUV420P)) { LOGE("Unsupported preview color format: %s", new_str_preview_format); return BAD_VALUE; } if (0 < new_preview_width && 0 < new_preview_height && new_str_preview_format != NULL && isSupportedPreviewSize(new_preview_width, new_preview_height)) { int new_preview_format = V4L2_PIX_FMT_YUV420; mFrameSizeDelta = 16; int current_preview_width, current_preview_height, current_frame_size; mSecCamera->getPreviewSize(¤t_preview_width, ¤t_preview_height, ¤t_frame_size); int current_pixel_format = mSecCamera->getPreviewPixelFormat(); if (current_preview_width != new_preview_width || current_preview_height != new_preview_height || current_pixel_format != new_preview_format) { if (mSecCamera->setPreviewSize(new_preview_width, new_preview_height, new_preview_format) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setPreviewSize(width(%d), height(%d), format(%d))", __func__, new_preview_width, new_preview_height, new_preview_format); ret = UNKNOWN_ERROR; } else { if (mPreviewWindow) { if (mPreviewRunning && !mPreviewStartDeferred) { LOGE("ERR(%s): preview is running, cannot change size and format!", __func__); ret = INVALID_OPERATION; } LOGV("%s: mPreviewWindow (%p) set_buffers_geometry", __func__, mPreviewWindow); LOGV("%s: mPreviewWindow->set_buffers_geometry (%p)", __func__, mPreviewWindow->set_buffers_geometry); mPreviewWindow->set_buffers_geometry(mPreviewWindow, new_preview_width, new_preview_height, new_preview_format); LOGV("%s: DONE mPreviewWindow (%p) set_buffers_geometry", __func__, mPreviewWindow); } mParameters.setPreviewSize(new_preview_width, new_preview_height); mParameters.setPreviewFormat(new_str_preview_format); } } else LOGV("%s: preview size and format has not changed", __func__); } else { LOGE("%s: Invalid preview size(%dx%d)", __func__, new_preview_width, new_preview_height); ret = INVALID_OPERATION; } int new_picture_width = 0; int new_picture_height = 0; params.getPictureSize(&new_picture_width, &new_picture_height); LOGV("%s : new_picture_width x new_picture_height = %dx%d", __func__, new_picture_width, new_picture_height); if (0 < new_picture_width && 0 < new_picture_height) { LOGV("%s: setSnapshotSize", __func__); if (mSecCamera->setSnapshotSize(new_picture_width, new_picture_height) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setSnapshotSize(width(%d), height(%d))", __func__, new_picture_width, new_picture_height); ret = UNKNOWN_ERROR; } else { mParameters.setPictureSize(new_picture_width, new_picture_height); } } // picture format const char *new_str_picture_format = params.getPictureFormat(); LOGV("%s : new_str_picture_format %s", __func__, new_str_picture_format); if (new_str_picture_format != NULL) { int new_picture_format = 0; if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_RGB565)) new_picture_format = V4L2_PIX_FMT_RGB565; else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_RGBA8888)) new_picture_format = V4L2_PIX_FMT_RGB32; else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_YUV420SP)) new_picture_format = V4L2_PIX_FMT_NV21; else if (!strcmp(new_str_picture_format, "yuv420sp_custom")) new_picture_format = V4L2_PIX_FMT_NV12T; else if (!strcmp(new_str_picture_format, "yuv420p")) new_picture_format = V4L2_PIX_FMT_YUV420; else if (!strcmp(new_str_picture_format, "yuv422i")) new_picture_format = V4L2_PIX_FMT_YUYV; else if (!strcmp(new_str_picture_format, "uyv422i_custom")) //Zero copy UYVY format new_picture_format = V4L2_PIX_FMT_UYVY; else if (!strcmp(new_str_picture_format, "uyv422i")) //Non-zero copy UYVY format new_picture_format = V4L2_PIX_FMT_UYVY; else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_JPEG)) new_picture_format = V4L2_PIX_FMT_YUYV; else if (!strcmp(new_str_picture_format, "yuv422p")) new_picture_format = V4L2_PIX_FMT_YUV422P; else new_picture_format = V4L2_PIX_FMT_NV21; //for 3rd party if (mSecCamera->setSnapshotPixelFormat(new_picture_format) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setSnapshotPixelFormat(format(%d))", __func__, new_picture_format); ret = UNKNOWN_ERROR; } else { mParameters.setPictureFormat(new_str_picture_format); } } //JPEG image quality int new_jpeg_quality = params.getInt(CameraParameters::KEY_JPEG_QUALITY); LOGV("%s : new_jpeg_quality %d", __func__, new_jpeg_quality); /* we ignore bad values */ if (new_jpeg_quality >=1 && new_jpeg_quality <= 100) { if (mSecCamera->setJpegQuality(new_jpeg_quality) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setJpegQuality(quality(%d))", __func__, new_jpeg_quality); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_JPEG_QUALITY, new_jpeg_quality); } } // JPEG thumbnail size int new_jpeg_thumbnail_width = params.getInt(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH); int new_jpeg_thumbnail_height= params.getInt(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT); if (0 <= new_jpeg_thumbnail_width && 0 <= new_jpeg_thumbnail_height) { if (mSecCamera->setJpegThumbnailSize(new_jpeg_thumbnail_width, new_jpeg_thumbnail_height) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setJpegThumbnailSize(width(%d), height(%d))", __func__, new_jpeg_thumbnail_width, new_jpeg_thumbnail_height); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, new_jpeg_thumbnail_width); mParameters.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, new_jpeg_thumbnail_height); } } // frame rate int new_frame_rate = params.getPreviewFrameRate(); /* ignore any fps request, we're determine fps automatically based * on scene mode. don't return an error because it causes CTS failure. */ if (new_frame_rate != mParameters.getPreviewFrameRate()) { LOGW("WARN(%s): request for preview frame %d not allowed, != %d\n", __func__, new_frame_rate, mParameters.getPreviewFrameRate()); } // rotation int new_rotation = params.getInt(CameraParameters::KEY_ROTATION); LOGV("%s : new_rotation %d", __func__, new_rotation); if (0 <= new_rotation) { LOGV("%s : set orientation:%d\n", __func__, new_rotation); if (mSecCamera->setExifOrientationInfo(new_rotation) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setExifOrientationInfo(%d)", __func__, new_rotation); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_ROTATION, new_rotation); } } // brightness int new_exposure_compensation = params.getInt(CameraParameters::KEY_EXPOSURE_COMPENSATION); int max_exposure_compensation = params.getInt(CameraParameters::KEY_MAX_EXPOSURE_COMPENSATION); int min_exposure_compensation = params.getInt(CameraParameters::KEY_MIN_EXPOSURE_COMPENSATION); LOGV("%s : new_exposure_compensation %d", __func__, new_exposure_compensation); if ((min_exposure_compensation <= new_exposure_compensation) && (max_exposure_compensation >= new_exposure_compensation)) { if (mSecCamera->setBrightness(new_exposure_compensation) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setBrightness(brightness(%d))", __func__, new_exposure_compensation); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_EXPOSURE_COMPENSATION, new_exposure_compensation); } } // whitebalance const char *new_white_str = params.get(CameraParameters::KEY_WHITE_BALANCE); LOGV("%s : new_white_str %s", __func__, new_white_str); if (new_white_str != NULL) { int new_white = -1; if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_AUTO)) new_white = WHITE_BALANCE_AUTO; else if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_DAYLIGHT)) new_white = WHITE_BALANCE_SUNNY; else if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_CLOUDY_DAYLIGHT)) new_white = WHITE_BALANCE_CLOUDY; else if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_FLUORESCENT)) new_white = WHITE_BALANCE_FLUORESCENT; else if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_INCANDESCENT)) new_white = WHITE_BALANCE_TUNGSTEN; else { LOGE("ERR(%s):Invalid white balance(%s)", __func__, new_white_str); //twilight, shade, warm_flourescent ret = UNKNOWN_ERROR; } if (0 <= new_white) { if (mSecCamera->setWhiteBalance(new_white) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setWhiteBalance(white(%d))", __func__, new_white); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_WHITE_BALANCE, new_white_str); } } } // scene mode const char *new_scene_mode_str = params.get(CameraParameters::KEY_SCENE_MODE); const char *current_scene_mode_str = mParameters.get(CameraParameters::KEY_SCENE_MODE); // fps range int new_min_fps = 0; int new_max_fps = 0; int current_min_fps, current_max_fps; params.getPreviewFpsRange(&new_min_fps, &new_max_fps); mParameters.getPreviewFpsRange(¤t_min_fps, ¤t_max_fps); /* our fps range is determined by the sensor, reject any request * that isn't exactly what we're already at. * but the check is performed when requesting only changing fps range */ if (new_scene_mode_str && current_scene_mode_str) { if (!strcmp(new_scene_mode_str, current_scene_mode_str)) { if ((new_min_fps != current_min_fps) || (new_max_fps != current_max_fps)) { LOGW("%s : requested new_min_fps = %d, new_max_fps = %d not allowed", __func__, new_min_fps, new_max_fps); LOGE("%s : current_min_fps = %d, current_max_fps = %d", __func__, current_min_fps, current_max_fps); ret = UNKNOWN_ERROR; } } } else { /* Check basic validation if scene mode is different */ if ((new_min_fps > new_max_fps) || (new_min_fps < 0) || (new_max_fps < 0)) ret = UNKNOWN_ERROR; } const char *new_focus_mode_str = params.get(CameraParameters::KEY_FOCUS_MODE); if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) { int new_scene_mode = -1; const char *new_flash_mode_str = params.get(CameraParameters::KEY_FLASH_MODE); // fps range is (15000,30000) by default. mParameters.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(15000,30000)"); mParameters.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "15000,30000"); if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_AUTO)) { new_scene_mode = SCENE_MODE_NONE; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "on,off,auto,torch"); } else { // defaults for non-auto scene modes if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) { new_focus_mode_str = CameraParameters::FOCUS_MODE_AUTO; } new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_PORTRAIT)) { new_scene_mode = SCENE_MODE_PORTRAIT; new_flash_mode_str = CameraParameters::FLASH_MODE_AUTO; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "auto"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_LANDSCAPE)) { new_scene_mode = SCENE_MODE_LANDSCAPE; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_SPORTS)) { new_scene_mode = SCENE_MODE_SPORTS; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_PARTY)) { new_scene_mode = SCENE_MODE_PARTY_INDOOR; new_flash_mode_str = CameraParameters::FLASH_MODE_AUTO; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "auto"); } else if ((!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_BEACH)) || (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_SNOW))) { new_scene_mode = SCENE_MODE_BEACH_SNOW; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_SUNSET)) { new_scene_mode = SCENE_MODE_SUNSET; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_NIGHT)) { new_scene_mode = SCENE_MODE_NIGHTSHOT; mParameters.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(4000,30000)"); mParameters.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "4000,30000"); new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_FIREWORKS)) { new_scene_mode = SCENE_MODE_FIREWORKS; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_CANDLELIGHT)) { new_scene_mode = SCENE_MODE_CANDLE_LIGHT; new_flash_mode_str = CameraParameters::FLASH_MODE_OFF; mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off"); } else { LOGE("%s::unmatched scene_mode(%s)", __func__, new_scene_mode_str); //action, night-portrait, theatre, steadyphoto ret = UNKNOWN_ERROR; } } // focus mode if (new_focus_mode_str != NULL) { int new_focus_mode = -1; if (!strcmp(new_focus_mode_str, CameraParameters::FOCUS_MODE_AUTO)) { new_focus_mode = FOCUS_MODE_AUTO; mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES, BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR); } else if (!strcmp(new_focus_mode_str, CameraParameters::FOCUS_MODE_MACRO)) { new_focus_mode = FOCUS_MODE_MACRO; mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES, BACK_CAMERA_MACRO_FOCUS_DISTANCES_STR); } else if (!strcmp(new_focus_mode_str, CameraParameters::FOCUS_MODE_INFINITY)) { new_focus_mode = FOCUS_MODE_INFINITY; mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES, BACK_CAMERA_INFINITY_FOCUS_DISTANCES_STR); } else { LOGE("%s::unmatched focus_mode(%s)", __func__, new_focus_mode_str); ret = UNKNOWN_ERROR; } if (0 <= new_focus_mode) { if (mSecCamera->setFocusMode(new_focus_mode) < 0) { LOGE("%s::mSecCamera->setFocusMode(%d) fail", __func__, new_focus_mode); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_FOCUS_MODE, new_focus_mode_str); } } } // flash.. if (new_flash_mode_str != NULL) { int new_flash_mode = -1; if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_OFF)) new_flash_mode = FLASH_MODE_OFF; else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_AUTO)) new_flash_mode = FLASH_MODE_AUTO; else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_ON)) new_flash_mode = FLASH_MODE_ON; else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_TORCH)) new_flash_mode = FLASH_MODE_TORCH; else { LOGE("%s::unmatched flash_mode(%s)", __func__, new_flash_mode_str); //red-eye ret = UNKNOWN_ERROR; } if (0 <= new_flash_mode) { if (mSecCamera->setFlashMode(new_flash_mode) < 0) { LOGE("%s::mSecCamera->setFlashMode(%d) fail", __func__, new_flash_mode); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_FLASH_MODE, new_flash_mode_str); } } } // scene.. if (0 <= new_scene_mode) { if (mSecCamera->setSceneMode(new_scene_mode) < 0) { LOGE("%s::mSecCamera->setSceneMode(%d) fail", __func__, new_scene_mode); ret = UNKNOWN_ERROR; } else { mParameters.set(CameraParameters::KEY_SCENE_MODE, new_scene_mode_str); } } } else { if (!isSupportedParameter(new_focus_mode_str, mParameters.get(CameraParameters::KEY_SUPPORTED_FOCUS_MODES))) { LOGE("%s: Unsupported focus mode: %s", __func__, new_focus_mode_str); ret = UNKNOWN_ERROR; } } // --------------------------------------------------------------------------- // image effect const char *new_image_effect_str = params.get(CameraParameters::KEY_EFFECT); if (new_image_effect_str != NULL) { int new_image_effect = -1; if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_NONE)) new_image_effect = IMAGE_EFFECT_NONE; else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_MONO)) new_image_effect = IMAGE_EFFECT_BNW; else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_SEPIA)) new_image_effect = IMAGE_EFFECT_SEPIA; else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_AQUA)) new_image_effect = IMAGE_EFFECT_AQUA; else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_NEGATIVE)) new_image_effect = IMAGE_EFFECT_NEGATIVE; else { //posterize, whiteboard, blackboard, solarize LOGE("ERR(%s):Invalid effect(%s)", __func__, new_image_effect_str); ret = UNKNOWN_ERROR; } if (new_image_effect >= 0) { if (mSecCamera->setImageEffect(new_image_effect) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setImageEffect(effect(%d))", __func__, new_image_effect); ret = UNKNOWN_ERROR; } else { const char *old_image_effect_str = mParameters.get(CameraParameters::KEY_EFFECT); if (old_image_effect_str) { if (strcmp(old_image_effect_str, new_image_effect_str)) { setSkipFrame(EFFECT_SKIP_FRAME); } } mParameters.set(CameraParameters::KEY_EFFECT, new_image_effect_str); } } } //vt mode int new_vtmode = mInternalParameters.getInt("vtmode"); if (0 <= new_vtmode) { if (mSecCamera->setVTmode(new_vtmode) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setVTMode(%d)", __func__, new_vtmode); ret = UNKNOWN_ERROR; } } //contrast int new_contrast = mInternalParameters.getInt("contrast"); if (0 <= new_contrast) { if (mSecCamera->setContrast(new_contrast) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setContrast(%d)", __func__, new_contrast); ret = UNKNOWN_ERROR; } } //WDR int new_wdr = mInternalParameters.getInt("wdr"); if (0 <= new_wdr) { if (mSecCamera->setWDR(new_wdr) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setWDR(%d)", __func__, new_wdr); ret = UNKNOWN_ERROR; } } //anti shake int new_anti_shake = mInternalParameters.getInt("anti-shake"); if (0 <= new_anti_shake) { if (mSecCamera->setAntiShake(new_anti_shake) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setWDR(%d)", __func__, new_anti_shake); ret = UNKNOWN_ERROR; } } // gps latitude const char *new_gps_latitude_str = params.get(CameraParameters::KEY_GPS_LATITUDE); if (mSecCamera->setGPSLatitude(new_gps_latitude_str) < 0) { LOGE("%s::mSecCamera->setGPSLatitude(%s) fail", __func__, new_gps_latitude_str); ret = UNKNOWN_ERROR; } else { if (new_gps_latitude_str) { mParameters.set(CameraParameters::KEY_GPS_LATITUDE, new_gps_latitude_str); } else { mParameters.remove(CameraParameters::KEY_GPS_LATITUDE); } } // gps longitude const char *new_gps_longitude_str = params.get(CameraParameters::KEY_GPS_LONGITUDE); if (mSecCamera->setGPSLongitude(new_gps_longitude_str) < 0) { LOGE("%s::mSecCamera->setGPSLongitude(%s) fail", __func__, new_gps_longitude_str); ret = UNKNOWN_ERROR; } else { if (new_gps_longitude_str) { mParameters.set(CameraParameters::KEY_GPS_LONGITUDE, new_gps_longitude_str); } else { mParameters.remove(CameraParameters::KEY_GPS_LONGITUDE); } } // gps altitude const char *new_gps_altitude_str = params.get(CameraParameters::KEY_GPS_ALTITUDE); if (mSecCamera->setGPSAltitude(new_gps_altitude_str) < 0) { LOGE("%s::mSecCamera->setGPSAltitude(%s) fail", __func__, new_gps_altitude_str); ret = UNKNOWN_ERROR; } else { if (new_gps_altitude_str) { mParameters.set(CameraParameters::KEY_GPS_ALTITUDE, new_gps_altitude_str); } else { mParameters.remove(CameraParameters::KEY_GPS_ALTITUDE); } } // gps timestamp const char *new_gps_timestamp_str = params.get(CameraParameters::KEY_GPS_TIMESTAMP); if (mSecCamera->setGPSTimeStamp(new_gps_timestamp_str) < 0) { LOGE("%s::mSecCamera->setGPSTimeStamp(%s) fail", __func__, new_gps_timestamp_str); ret = UNKNOWN_ERROR; } else { if (new_gps_timestamp_str) { mParameters.set(CameraParameters::KEY_GPS_TIMESTAMP, new_gps_timestamp_str); } else { mParameters.remove(CameraParameters::KEY_GPS_TIMESTAMP); } } // gps processing method const char *new_gps_processing_method_str = params.get(CameraParameters::KEY_GPS_PROCESSING_METHOD); if (mSecCamera->setGPSProcessingMethod(new_gps_processing_method_str) < 0) { LOGE("%s::mSecCamera->setGPSProcessingMethod(%s) fail", __func__, new_gps_processing_method_str); ret = UNKNOWN_ERROR; } else { if (new_gps_processing_method_str) { mParameters.set(CameraParameters::KEY_GPS_PROCESSING_METHOD, new_gps_processing_method_str); } else { mParameters.remove(CameraParameters::KEY_GPS_PROCESSING_METHOD); } } // Recording size int new_recording_width = mInternalParameters.getInt("recording-size-width"); int new_recording_height= mInternalParameters.getInt("recording-size-height"); if (0 < new_recording_width && 0 < new_recording_height) { if (mSecCamera->setRecordingSize(new_recording_width, new_recording_height) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setRecordingSize(width(%d), height(%d))", __func__, new_recording_width, new_recording_height); ret = UNKNOWN_ERROR; } } else { if (mSecCamera->setRecordingSize(new_preview_width, new_preview_height) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setRecordingSize(width(%d), height(%d))", __func__, new_preview_width, new_preview_height); ret = UNKNOWN_ERROR; } } //gamma const char *new_gamma_str = mInternalParameters.get("video_recording_gamma"); if (new_gamma_str != NULL) { int new_gamma = -1; if (!strcmp(new_gamma_str, "off")) new_gamma = GAMMA_OFF; else if (!strcmp(new_gamma_str, "on")) new_gamma = GAMMA_ON; else { LOGE("%s::unmatched gamma(%s)", __func__, new_gamma_str); ret = UNKNOWN_ERROR; } if (0 <= new_gamma) { if (mSecCamera->setGamma(new_gamma) < 0) { LOGE("%s::mSecCamera->setGamma(%d) fail", __func__, new_gamma); ret = UNKNOWN_ERROR; } } } //slow ae const char *new_slow_ae_str = mInternalParameters.get("slow_ae"); if (new_slow_ae_str != NULL) { int new_slow_ae = -1; if (!strcmp(new_slow_ae_str, "off")) new_slow_ae = SLOW_AE_OFF; else if (!strcmp(new_slow_ae_str, "on")) new_slow_ae = SLOW_AE_ON; else { LOGE("%s::unmatched slow_ae(%s)", __func__, new_slow_ae_str); ret = UNKNOWN_ERROR; } if (0 <= new_slow_ae) { if (mSecCamera->setSlowAE(new_slow_ae) < 0) { LOGE("%s::mSecCamera->setSlowAE(%d) fail", __func__, new_slow_ae); ret = UNKNOWN_ERROR; } } } /*Camcorder fix fps*/ int new_sensor_mode = mInternalParameters.getInt("cam_mode"); if (0 <= new_sensor_mode) { if (mSecCamera->setSensorMode(new_sensor_mode) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setSensorMode(%d)", __func__, new_sensor_mode); ret = UNKNOWN_ERROR; } } else { new_sensor_mode=0; } /*Shot mode*/ int new_shot_mode = mInternalParameters.getInt("shot_mode"); if (0 <= new_shot_mode) { if (mSecCamera->setShotMode(new_shot_mode) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setShotMode(%d)", __func__, new_shot_mode); ret = UNKNOWN_ERROR; } } else { new_shot_mode=0; } //blur for Video call int new_blur_level = mInternalParameters.getInt("blur"); if (0 <= new_blur_level) { if (mSecCamera->setBlur(new_blur_level) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setBlur(%d)", __func__, new_blur_level); ret = UNKNOWN_ERROR; } } // chk_dataline int new_dataline = mInternalParameters.getInt("chk_dataline"); if (0 <= new_dataline) { if (mSecCamera->setDataLineCheck(new_dataline) < 0) { LOGE("ERR(%s):Fail on mSecCamera->setDataLineCheck(%d)", __func__, new_dataline); ret = UNKNOWN_ERROR; } } LOGV("%s return ret = %d", __func__, ret); return ret; } CameraParameters CameraHardwareSec::getParameters() const { LOGV("%s :", __func__); return mParameters; } status_t CameraHardwareSec::sendCommand(int32_t command, int32_t arg1, int32_t arg2) { return BAD_VALUE; } void CameraHardwareSec::release() { LOGV("%s", __func__); /* shut down any threads we have that might be running. do it here * instead of the destructor. we're guaranteed to be on another thread * than the ones below. if we used the destructor, since the threads * have a reference to this object, we could wind up trying to wait * for ourself to exit, which is a deadlock. */ if (mPreviewThread != NULL) { /* this thread is normally already in it's threadLoop but blocked * on the condition variable or running. signal it so it wakes * up and can exit. */ mPreviewThread->requestExit(); mExitPreviewThread = true; mPreviewRunning = true; /* let it run so it can exit */ mPreviewCondition.signal(); mPreviewThread->requestExitAndWait(); mPreviewThread.clear(); } if (mAutoFocusThread != NULL) { /* this thread is normally already in it's threadLoop but blocked * on the condition variable. signal it so it wakes up and can exit. */ mFocusLock.lock(); mAutoFocusThread->requestExit(); mExitAutoFocusThread = true; mFocusCondition.signal(); mFocusLock.unlock(); mAutoFocusThread->requestExitAndWait(); mAutoFocusThread.clear(); } if (mPictureThread != NULL) { mPictureThread->requestExitAndWait(); mPictureThread.clear(); } if (mRawHeap) { mRawHeap->release(mRawHeap); mRawHeap = 0; } if (mPreviewHeap) { mPreviewHeap->release(mPreviewHeap); mPreviewHeap = 0; } if (mRecordHeap) { mRecordHeap->release(mRecordHeap); mRecordHeap = 0; } /* close after all the heaps are cleared since those * could have dup'd our file descriptor. */ mSecCamera->DeinitCamera(); } status_t CameraHardwareSec::storeMetaDataInBuffers(bool enable) { // FIXME: // metadata buffer mode can be turned on or off. // Samsung needs to fix this. if (!enable) { LOGE("Non-metadata buffer mode is not supported!"); return INVALID_OPERATION; } return OK; } static CameraInfo sCameraInfo[] = { { CAMERA_FACING_BACK, 90, /* orientation */ }, { CAMERA_FACING_FRONT, 270, /* orientation */ } }; /** Close this device */ static camera_device_t *g_cam_device; static int HAL_camera_device_close(struct hw_device_t* device) { LOGI("%s", __func__); if (device) { camera_device_t *cam_device = (camera_device_t *)device; delete static_cast(cam_device->priv); free(cam_device); g_cam_device = 0; } return 0; } static inline CameraHardwareSec *obj(struct camera_device *dev) { return reinterpret_cast(dev->priv); } /** Set the preview_stream_ops to which preview frames are sent */ static int HAL_camera_device_set_preview_window(struct camera_device *dev, struct preview_stream_ops *buf) { LOGV("%s", __func__); return obj(dev)->setPreviewWindow(buf); } /** Set the notification and data callbacks */ static void HAL_camera_device_set_callbacks(struct camera_device *dev, camera_notify_callback notify_cb, camera_data_callback data_cb, camera_data_timestamp_callback data_cb_timestamp, camera_request_memory get_memory, void* user) { LOGV("%s", __func__); obj(dev)->setCallbacks(notify_cb, data_cb, data_cb_timestamp, get_memory, user); } /** * The following three functions all take a msg_type, which is a bitmask of * the messages defined in include/ui/Camera.h */ /** * Enable a message, or set of messages. */ static void HAL_camera_device_enable_msg_type(struct camera_device *dev, int32_t msg_type) { LOGV("%s", __func__); obj(dev)->enableMsgType(msg_type); } /** * Disable a message, or a set of messages. * * Once received a call to disableMsgType(CAMERA_MSG_VIDEO_FRAME), camera * HAL should not rely on its client to call releaseRecordingFrame() to * release video recording frames sent out by the cameral HAL before and * after the disableMsgType(CAMERA_MSG_VIDEO_FRAME) call. Camera HAL * clients must not modify/access any video recording frame after calling * disableMsgType(CAMERA_MSG_VIDEO_FRAME). */ static void HAL_camera_device_disable_msg_type(struct camera_device *dev, int32_t msg_type) { LOGV("%s", __func__); obj(dev)->disableMsgType(msg_type); } /** * Query whether a message, or a set of messages, is enabled. Note that * this is operates as an AND, if any of the messages queried are off, this * will return false. */ static int HAL_camera_device_msg_type_enabled(struct camera_device *dev, int32_t msg_type) { LOGV("%s", __func__); return obj(dev)->msgTypeEnabled(msg_type); } /** * Start preview mode. */ static int HAL_camera_device_start_preview(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->startPreview(); } /** * Stop a previously started preview. */ static void HAL_camera_device_stop_preview(struct camera_device *dev) { LOGV("%s", __func__); obj(dev)->stopPreview(); } /** * Returns true if preview is enabled. */ static int HAL_camera_device_preview_enabled(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->previewEnabled(); } /** * Request the camera HAL to store meta data or real YUV data in the video * buffers sent out via CAMERA_MSG_VIDEO_FRAME for a recording session. If * it is not called, the default camera HAL behavior is to store real YUV * data in the video buffers. * * This method should be called before startRecording() in order to be * effective. * * If meta data is stored in the video buffers, it is up to the receiver of * the video buffers to interpret the contents and to find the actual frame * data with the help of the meta data in the buffer. How this is done is * outside of the scope of this method. * * Some camera HALs may not support storing meta data in the video buffers, * but all camera HALs should support storing real YUV data in the video * buffers. If the camera HAL does not support storing the meta data in the * video buffers when it is requested to do do, INVALID_OPERATION must be * returned. It is very useful for the camera HAL to pass meta data rather * than the actual frame data directly to the video encoder, since the * amount of the uncompressed frame data can be very large if video size is * large. * * @param enable if true to instruct the camera HAL to store * meta data in the video buffers; false to instruct * the camera HAL to store real YUV data in the video * buffers. * * @return OK on success. */ static int HAL_camera_device_store_meta_data_in_buffers(struct camera_device *dev, int enable) { LOGV("%s", __func__); return obj(dev)->storeMetaDataInBuffers(enable); } /** * Start record mode. When a record image is available, a * CAMERA_MSG_VIDEO_FRAME message is sent with the corresponding * frame. Every record frame must be released by a camera HAL client via * releaseRecordingFrame() before the client calls * disableMsgType(CAMERA_MSG_VIDEO_FRAME). After the client calls * disableMsgType(CAMERA_MSG_VIDEO_FRAME), it is the camera HAL's * responsibility to manage the life-cycle of the video recording frames, * and the client must not modify/access any video recording frames. */ static int HAL_camera_device_start_recording(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->startRecording(); } /** * Stop a previously started recording. */ static void HAL_camera_device_stop_recording(struct camera_device *dev) { LOGV("%s", __func__); obj(dev)->stopRecording(); } /** * Returns true if recording is enabled. */ static int HAL_camera_device_recording_enabled(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->recordingEnabled(); } /** * Release a record frame previously returned by CAMERA_MSG_VIDEO_FRAME. * * It is camera HAL client's responsibility to release video recording * frames sent out by the camera HAL before the camera HAL receives a call * to disableMsgType(CAMERA_MSG_VIDEO_FRAME). After it receives the call to * disableMsgType(CAMERA_MSG_VIDEO_FRAME), it is the camera HAL's * responsibility to manage the life-cycle of the video recording frames. */ static void HAL_camera_device_release_recording_frame(struct camera_device *dev, const void *opaque) { LOGV("%s", __func__); obj(dev)->releaseRecordingFrame(opaque); } /** * Start auto focus, the notification callback routine is called with * CAMERA_MSG_FOCUS once when focusing is complete. autoFocus() will be * called again if another auto focus is needed. */ static int HAL_camera_device_auto_focus(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->autoFocus(); } /** * Cancels auto-focus function. If the auto-focus is still in progress, * this function will cancel it. Whether the auto-focus is in progress or * not, this function will return the focus position to the default. If * the camera does not support auto-focus, this is a no-op. */ static int HAL_camera_device_cancel_auto_focus(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->cancelAutoFocus(); } /** * Take a picture. */ static int HAL_camera_device_take_picture(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->takePicture(); } /** * Cancel a picture that was started with takePicture. Calling this method * when no picture is being taken is a no-op. */ static int HAL_camera_device_cancel_picture(struct camera_device *dev) { LOGV("%s", __func__); return obj(dev)->cancelPicture(); } /** * Set the camera parameters. This returns BAD_VALUE if any parameter is * invalid or not supported. */ static int HAL_camera_device_set_parameters(struct camera_device *dev, const char *parms) { LOGV("%s", __func__); String8 str(parms); CameraParameters p(str); return obj(dev)->setParameters(p); } /** Return the camera parameters. */ char *HAL_camera_device_get_parameters(struct camera_device *dev) { LOGV("%s", __func__); String8 str; CameraParameters parms = obj(dev)->getParameters(); str = parms.flatten(); return strdup(str.string()); } void HAL_camera_device_put_parameters(struct camera_device *dev, char *parms) { LOGV("%s", __func__); free(parms); } /** * Send command to camera driver. */ static int HAL_camera_device_send_command(struct camera_device *dev, int32_t cmd, int32_t arg1, int32_t arg2) { LOGV("%s", __func__); return obj(dev)->sendCommand(cmd, arg1, arg2); } /** * Release the hardware resources owned by this object. Note that this is * *not* done in the destructor. */ static void HAL_camera_device_release(struct camera_device *dev) { LOGV("%s", __func__); obj(dev)->release(); } /** * Dump state of the camera hardware */ static int HAL_camera_device_dump(struct camera_device *dev, int fd) { LOGV("%s", __func__); return obj(dev)->dump(fd); } static int HAL_getNumberOfCameras() { LOGV("%s", __func__); return sizeof(sCameraInfo) / sizeof(sCameraInfo[0]); } static int HAL_getCameraInfo(int cameraId, struct camera_info *cameraInfo) { LOGV("%s", __func__); memcpy(cameraInfo, &sCameraInfo[cameraId], sizeof(CameraInfo)); return 0; } #define SET_METHOD(m) m : HAL_camera_device_##m static camera_device_ops_t camera_device_ops = { SET_METHOD(set_preview_window), SET_METHOD(set_callbacks), SET_METHOD(enable_msg_type), SET_METHOD(disable_msg_type), SET_METHOD(msg_type_enabled), SET_METHOD(start_preview), SET_METHOD(stop_preview), SET_METHOD(preview_enabled), SET_METHOD(store_meta_data_in_buffers), SET_METHOD(start_recording), SET_METHOD(stop_recording), SET_METHOD(recording_enabled), SET_METHOD(release_recording_frame), SET_METHOD(auto_focus), SET_METHOD(cancel_auto_focus), SET_METHOD(take_picture), SET_METHOD(cancel_picture), SET_METHOD(set_parameters), SET_METHOD(get_parameters), SET_METHOD(put_parameters), SET_METHOD(send_command), SET_METHOD(release), SET_METHOD(dump), }; #undef SET_METHOD static int HAL_camera_device_open(const struct hw_module_t* module, const char *id, struct hw_device_t** device) { LOGV("%s", __func__); int cameraId = atoi(id); if (cameraId < 0 || cameraId >= HAL_getNumberOfCameras()) { LOGE("Invalid camera ID %s", id); return -EINVAL; } if (g_cam_device) { if (obj(g_cam_device)->getCameraId() == cameraId) { LOGV("returning existing camera ID %s", id); goto done; } else { LOGE("Cannot open camera %d. camera %d is already running!", cameraId, obj(g_cam_device)->getCameraId()); return -ENOSYS; } } g_cam_device = (camera_device_t *)malloc(sizeof(camera_device_t)); if (!g_cam_device) return -ENOMEM; g_cam_device->common.tag = HARDWARE_DEVICE_TAG; g_cam_device->common.version = 1; g_cam_device->common.module = const_cast(module); g_cam_device->common.close = HAL_camera_device_close; g_cam_device->ops = &camera_device_ops; LOGI("%s: open camera %s", __func__, id); g_cam_device->priv = new CameraHardwareSec(cameraId, g_cam_device); done: *device = (hw_device_t *)g_cam_device; LOGI("%s: opened camera %s (%p)", __func__, id, *device); return 0; } static hw_module_methods_t camera_module_methods = { open : HAL_camera_device_open }; extern "C" { struct camera_module HAL_MODULE_INFO_SYM = { common : { tag : HARDWARE_MODULE_TAG, version_major : 1, version_minor : 0, id : CAMERA_HARDWARE_MODULE_ID, name : "Crespo camera HAL", author : "Samsung Corporation", methods : &camera_module_methods, }, get_number_of_cameras : HAL_getNumberOfCameras, get_camera_info : HAL_getCameraInfo }; } }; // namespace android