blob: 40d53b3d4053cac14ec55d397d6f1fd7810aea6f [file] [log] [blame]
/*
* Copyright (C) 2012 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#define LOG_TAG "Camera2-FrameProcessor"
#define ATRACE_TAG ATRACE_TAG_CAMERA
//#define LOG_NDEBUG 0
#include <utils/Log.h>
#include <utils/Trace.h>
#include "common/CameraDeviceBase.h"
#include "api1/Camera2Client.h"
#include "api1/client2/FrameProcessor.h"
namespace android {
namespace camera2 {
FrameProcessor::FrameProcessor(wp<CameraDeviceBase> device,
sp<Camera2Client> client) :
FrameProcessorBase(device),
mClient(client),
mLastFrameNumberOfFaces(0),
mLast3AFrameNumber(-1) {
sp<CameraDeviceBase> d = device.promote();
mSynthesize3ANotify = !(d->willNotify3A());
{
SharedParameters::Lock l(client->getParameters());
if (client->getCameraDeviceVersion() >= CAMERA_DEVICE_API_VERSION_3_2) {
mUsePartialResult = (mNumPartialResults > 1);
} else {
mUsePartialResult = l.mParameters.quirks.partialResults;
}
// Initialize starting 3A state
m3aState.afTriggerId = l.mParameters.afTriggerCounter;
m3aState.aeTriggerId = l.mParameters.precaptureTriggerCounter;
// Check if lens is fixed-focus
if (l.mParameters.focusMode == Parameters::FOCUS_MODE_FIXED) {
m3aState.afMode = ANDROID_CONTROL_AF_MODE_OFF;
}
}
}
FrameProcessor::~FrameProcessor() {
}
bool FrameProcessor::processSingleFrame(CaptureResult &frame,
const sp<CameraDeviceBase> &device) {
sp<Camera2Client> client = mClient.promote();
if (!client.get()) {
return false;
}
bool isPartialResult = false;
if (mUsePartialResult) {
if (client->getCameraDeviceVersion() >= CAMERA_DEVICE_API_VERSION_3_2) {
isPartialResult = frame.mResultExtras.partialResultCount < mNumPartialResults;
} else {
camera_metadata_entry_t entry;
entry = frame.mMetadata.find(ANDROID_QUIRKS_PARTIAL_RESULT);
if (entry.count > 0 &&
entry.data.u8[0] == ANDROID_QUIRKS_PARTIAL_RESULT_PARTIAL) {
isPartialResult = true;
}
}
}
if (!isPartialResult && processFaceDetect(frame.mMetadata, client) != OK) {
return false;
}
if (mSynthesize3ANotify) {
process3aState(frame, client);
}
return FrameProcessorBase::processSingleFrame(frame, device);
}
status_t FrameProcessor::processFaceDetect(const CameraMetadata &frame,
const sp<Camera2Client> &client) {
status_t res = BAD_VALUE;
ATRACE_CALL();
camera_metadata_ro_entry_t entry;
bool enableFaceDetect;
{
SharedParameters::Lock l(client->getParameters());
enableFaceDetect = l.mParameters.enableFaceDetect;
}
entry = frame.find(ANDROID_STATISTICS_FACE_DETECT_MODE);
// TODO: This should be an error once implementations are compliant
if (entry.count == 0) {
return OK;
}
uint8_t faceDetectMode = entry.data.u8[0];
camera_frame_metadata metadata;
Vector<camera_face_t> faces;
metadata.number_of_faces = 0;
if (enableFaceDetect &&
faceDetectMode != ANDROID_STATISTICS_FACE_DETECT_MODE_OFF) {
SharedParameters::Lock l(client->getParameters());
entry = frame.find(ANDROID_STATISTICS_FACE_RECTANGLES);
if (entry.count == 0) {
// No faces this frame
/* warning: locks SharedCameraCallbacks */
callbackFaceDetection(client, metadata);
return OK;
}
metadata.number_of_faces = entry.count / 4;
if (metadata.number_of_faces >
l.mParameters.fastInfo.maxFaces) {
ALOGE("%s: Camera %d: More faces than expected! (Got %d, max %d)",
__FUNCTION__, client->getCameraId(),
metadata.number_of_faces, l.mParameters.fastInfo.maxFaces);
return res;
}
const int32_t *faceRects = entry.data.i32;
entry = frame.find(ANDROID_STATISTICS_FACE_SCORES);
if (entry.count == 0) {
ALOGE("%s: Camera %d: Unable to read face scores",
__FUNCTION__, client->getCameraId());
return res;
}
const uint8_t *faceScores = entry.data.u8;
const int32_t *faceLandmarks = NULL;
const int32_t *faceIds = NULL;
if (faceDetectMode == ANDROID_STATISTICS_FACE_DETECT_MODE_FULL) {
entry = frame.find(ANDROID_STATISTICS_FACE_LANDMARKS);
if (entry.count == 0) {
ALOGE("%s: Camera %d: Unable to read face landmarks",
__FUNCTION__, client->getCameraId());
return res;
}
faceLandmarks = entry.data.i32;
entry = frame.find(ANDROID_STATISTICS_FACE_IDS);
if (entry.count == 0) {
ALOGE("%s: Camera %d: Unable to read face IDs",
__FUNCTION__, client->getCameraId());
return res;
}
faceIds = entry.data.i32;
}
entry = frame.find(ANDROID_SCALER_CROP_REGION);
if (entry.count < 4) {
ALOGE("%s: Camera %d: Unable to read crop region (count = %d)",
__FUNCTION__, client->getCameraId(), entry.count);
return res;
}
Parameters::CropRegion scalerCrop = {
static_cast<float>(entry.data.i32[0]),
static_cast<float>(entry.data.i32[1]),
static_cast<float>(entry.data.i32[2]),
static_cast<float>(entry.data.i32[3])};
faces.setCapacity(metadata.number_of_faces);
size_t maxFaces = metadata.number_of_faces;
for (size_t i = 0; i < maxFaces; i++) {
if (faceScores[i] == 0) {
metadata.number_of_faces--;
continue;
}
if (faceScores[i] > 100) {
ALOGW("%s: Face index %zu with out of range score %d",
__FUNCTION__, i, faceScores[i]);
}
camera_face_t face;
face.rect[0] = l.mParameters.arrayXToNormalizedWithCrop(
faceRects[i*4 + 0], scalerCrop);
face.rect[1] = l.mParameters.arrayYToNormalizedWithCrop(
faceRects[i*4 + 1], scalerCrop);
face.rect[2] = l.mParameters.arrayXToNormalizedWithCrop(
faceRects[i*4 + 2], scalerCrop);
face.rect[3] = l.mParameters.arrayYToNormalizedWithCrop(
faceRects[i*4 + 3], scalerCrop);
face.score = faceScores[i];
if (faceDetectMode == ANDROID_STATISTICS_FACE_DETECT_MODE_FULL) {
face.id = faceIds[i];
face.left_eye[0] = l.mParameters.arrayXToNormalizedWithCrop(
faceLandmarks[i*6 + 0], scalerCrop);
face.left_eye[1] = l.mParameters.arrayYToNormalizedWithCrop(
faceLandmarks[i*6 + 1], scalerCrop);
face.right_eye[0] = l.mParameters.arrayXToNormalizedWithCrop(
faceLandmarks[i*6 + 2], scalerCrop);
face.right_eye[1] = l.mParameters.arrayYToNormalizedWithCrop(
faceLandmarks[i*6 + 3], scalerCrop);
face.mouth[0] = l.mParameters.arrayXToNormalizedWithCrop(
faceLandmarks[i*6 + 4], scalerCrop);
face.mouth[1] = l.mParameters.arrayYToNormalizedWithCrop(
faceLandmarks[i*6 + 5], scalerCrop);
} else {
face.id = 0;
face.left_eye[0] = face.left_eye[1] = -2000;
face.right_eye[0] = face.right_eye[1] = -2000;
face.mouth[0] = face.mouth[1] = -2000;
}
faces.push_back(face);
}
metadata.faces = faces.editArray();
}
/* warning: locks SharedCameraCallbacks */
callbackFaceDetection(client, metadata);
return OK;
}
status_t FrameProcessor::process3aState(const CaptureResult &frame,
const sp<Camera2Client> &client) {
ATRACE_CALL();
const CameraMetadata &metadata = frame.mMetadata;
camera_metadata_ro_entry_t entry;
int cameraId = client->getCameraId();
entry = metadata.find(ANDROID_REQUEST_FRAME_COUNT);
int32_t frameNumber = entry.data.i32[0];
// Don't send 3A notifications for the same frame number twice
if (frameNumber <= mLast3AFrameNumber) {
ALOGV("%s: Already sent 3A for frame number %d, skipping",
__FUNCTION__, frameNumber);
return OK;
}
mLast3AFrameNumber = frameNumber;
// Get 3A states from result metadata
bool gotAllStates = true;
AlgState new3aState;
// TODO: Also use AE mode, AE trigger ID
gotAllStates &= get3aResult<uint8_t>(metadata, ANDROID_CONTROL_AF_MODE,
&new3aState.afMode, frameNumber, cameraId);
gotAllStates &= get3aResult<uint8_t>(metadata, ANDROID_CONTROL_AWB_MODE,
&new3aState.awbMode, frameNumber, cameraId);
gotAllStates &= get3aResult<uint8_t>(metadata, ANDROID_CONTROL_AE_STATE,
&new3aState.aeState, frameNumber, cameraId);
gotAllStates &= get3aResult<uint8_t>(metadata, ANDROID_CONTROL_AF_STATE,
&new3aState.afState, frameNumber, cameraId);
gotAllStates &= get3aResult<uint8_t>(metadata, ANDROID_CONTROL_AWB_STATE,
&new3aState.awbState, frameNumber, cameraId);
if (client->getCameraDeviceVersion() >= CAMERA_DEVICE_API_VERSION_3_2) {
new3aState.afTriggerId = frame.mResultExtras.afTriggerId;
new3aState.aeTriggerId = frame.mResultExtras.precaptureTriggerId;
} else {
gotAllStates &= get3aResult<int32_t>(metadata, ANDROID_CONTROL_AF_TRIGGER_ID,
&new3aState.afTriggerId, frameNumber, cameraId);
gotAllStates &= get3aResult<int32_t>(metadata, ANDROID_CONTROL_AE_PRECAPTURE_ID,
&new3aState.aeTriggerId, frameNumber, cameraId);
}
if (!gotAllStates) return BAD_VALUE;
if (new3aState.aeState != m3aState.aeState) {
ALOGV("%s: Camera %d: AE state %d->%d",
__FUNCTION__, cameraId,
m3aState.aeState, new3aState.aeState);
client->notifyAutoExposure(new3aState.aeState, new3aState.aeTriggerId);
}
if (new3aState.afState != m3aState.afState ||
new3aState.afMode != m3aState.afMode ||
new3aState.afTriggerId != m3aState.afTriggerId) {
ALOGV("%s: Camera %d: AF state %d->%d. AF mode %d->%d. Trigger %d->%d",
__FUNCTION__, cameraId,
m3aState.afState, new3aState.afState,
m3aState.afMode, new3aState.afMode,
m3aState.afTriggerId, new3aState.afTriggerId);
client->notifyAutoFocus(new3aState.afState, new3aState.afTriggerId);
}
if (new3aState.awbState != m3aState.awbState ||
new3aState.awbMode != m3aState.awbMode) {
ALOGV("%s: Camera %d: AWB state %d->%d. AWB mode %d->%d",
__FUNCTION__, cameraId,
m3aState.awbState, new3aState.awbState,
m3aState.awbMode, new3aState.awbMode);
client->notifyAutoWhitebalance(new3aState.awbState,
new3aState.aeTriggerId);
}
m3aState = new3aState;
return OK;
}
template<typename Src, typename T>
bool FrameProcessor::get3aResult(const CameraMetadata& result, int32_t tag,
T* value, int32_t frameNumber, int cameraId) {
camera_metadata_ro_entry_t entry;
if (value == NULL) {
ALOGE("%s: Camera %d: Value to write to is NULL",
__FUNCTION__, cameraId);
return false;
}
entry = result.find(tag);
if (entry.count == 0) {
ALOGE("%s: Camera %d: No %s provided by HAL for frame %d!",
__FUNCTION__, cameraId,
get_camera_metadata_tag_name(tag), frameNumber);
return false;
} else {
switch(sizeof(Src)){
case sizeof(uint8_t):
*value = static_cast<T>(entry.data.u8[0]);
break;
case sizeof(int32_t):
*value = static_cast<T>(entry.data.i32[0]);
break;
default:
ALOGE("%s: Camera %d: Unsupported source",
__FUNCTION__, cameraId);
return false;
}
}
return true;
}
void FrameProcessor::callbackFaceDetection(sp<Camera2Client> client,
const camera_frame_metadata &metadata) {
camera_frame_metadata *metadata_ptr =
const_cast<camera_frame_metadata*>(&metadata);
/**
* Filter out repeated 0-face callbacks,
* but not when the last frame was >0
*/
if (metadata.number_of_faces != 0 ||
mLastFrameNumberOfFaces != metadata.number_of_faces) {
Camera2Client::SharedCameraCallbacks::Lock
l(client->mSharedCameraCallbacks);
if (l.mRemoteCallback != NULL) {
l.mRemoteCallback->dataCallback(CAMERA_MSG_PREVIEW_METADATA,
NULL,
metadata_ptr);
}
}
mLastFrameNumberOfFaces = metadata.number_of_faces;
}
}; // namespace camera2
}; // namespace android