/* * Copyright (C) 2010 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #define LOG_TAG "NuPlayer" #include #include "NuPlayer.h" #include "HTTPLiveSource.h" #include "NuPlayerDecoder.h" #include "NuPlayerDriver.h" #include "NuPlayerRenderer.h" #include "NuPlayerSource.h" #include "StreamingSource.h" #include "ATSParser.h" #include #include #include #include #include #include #include #include #include #include #include "avc_utils.h" namespace android { //////////////////////////////////////////////////////////////////////////////// NuPlayer::NuPlayer() : mUIDValid(false), mVideoIsAVC(false), mAudioEOS(false), mVideoEOS(false), mScanSourcesPending(false), mScanSourcesGeneration(0), mFlushingAudio(NONE), mFlushingVideo(NONE), mResetInProgress(false), mResetPostponed(false), mSkipRenderingAudioUntilMediaTimeUs(-1ll), mSkipRenderingVideoUntilMediaTimeUs(-1ll), mVideoLateByUs(0ll), mNumFramesTotal(0ll), mNumFramesDropped(0ll) { } NuPlayer::~NuPlayer() { } void NuPlayer::setUID(uid_t uid) { mUIDValid = true; mUID = uid; } void NuPlayer::setDriver(const wp &driver) { mDriver = driver; } void NuPlayer::setDataSource(const sp &source) { sp msg = new AMessage(kWhatSetDataSource, id()); msg->setObject("source", new StreamingSource(source)); msg->post(); } void NuPlayer::setDataSource( const char *url, const KeyedVector *headers) { sp msg = new AMessage(kWhatSetDataSource, id()); msg->setObject("source", new HTTPLiveSource(url, headers, mUIDValid, mUID)); msg->post(); } void NuPlayer::setVideoSurface(const sp &surface) { sp msg = new AMessage(kWhatSetVideoNativeWindow, id()); msg->setObject("native-window", new NativeWindowWrapper(surface)); msg->post(); } void NuPlayer::setVideoSurfaceTexture(const sp &surfaceTexture) { sp msg = new AMessage(kWhatSetVideoNativeWindow, id()); sp surfaceTextureClient(surfaceTexture != NULL ? new SurfaceTextureClient(surfaceTexture) : NULL); msg->setObject("native-window", new NativeWindowWrapper(surfaceTextureClient)); msg->post(); } void NuPlayer::setAudioSink(const sp &sink) { sp msg = new AMessage(kWhatSetAudioSink, id()); msg->setObject("sink", sink); msg->post(); } void NuPlayer::start() { (new AMessage(kWhatStart, id()))->post(); } void NuPlayer::pause() { (new AMessage(kWhatPause, id()))->post(); } void NuPlayer::resume() { (new AMessage(kWhatResume, id()))->post(); } void NuPlayer::resetAsync() { (new AMessage(kWhatReset, id()))->post(); } void NuPlayer::seekToAsync(int64_t seekTimeUs) { sp msg = new AMessage(kWhatSeek, id()); msg->setInt64("seekTimeUs", seekTimeUs); msg->post(); } // static bool NuPlayer::IsFlushingState(FlushStatus state, bool *needShutdown) { switch (state) { case FLUSHING_DECODER: if (needShutdown != NULL) { *needShutdown = false; } return true; case FLUSHING_DECODER_SHUTDOWN: if (needShutdown != NULL) { *needShutdown = true; } return true; default: return false; } } void NuPlayer::onMessageReceived(const sp &msg) { switch (msg->what()) { case kWhatSetDataSource: { LOGV("kWhatSetDataSource"); CHECK(mSource == NULL); sp obj; CHECK(msg->findObject("source", &obj)); mSource = static_cast(obj.get()); break; } case kWhatSetVideoNativeWindow: { LOGV("kWhatSetVideoNativeWindow"); sp obj; CHECK(msg->findObject("native-window", &obj)); mNativeWindow = static_cast(obj.get()); break; } case kWhatSetAudioSink: { LOGV("kWhatSetAudioSink"); sp obj; CHECK(msg->findObject("sink", &obj)); mAudioSink = static_cast(obj.get()); break; } case kWhatStart: { LOGV("kWhatStart"); mVideoIsAVC = false; mAudioEOS = false; mVideoEOS = false; mSkipRenderingAudioUntilMediaTimeUs = -1; mSkipRenderingVideoUntilMediaTimeUs = -1; mVideoLateByUs = 0; mNumFramesTotal = 0; mNumFramesDropped = 0; mSource->start(); mRenderer = new Renderer( mAudioSink, new AMessage(kWhatRendererNotify, id())); looper()->registerHandler(mRenderer); postScanSources(); break; } case kWhatScanSources: { int32_t generation; CHECK(msg->findInt32("generation", &generation)); if (generation != mScanSourcesGeneration) { // Drop obsolete msg. break; } mScanSourcesPending = false; LOGV("scanning sources haveAudio=%d, haveVideo=%d", mAudioDecoder != NULL, mVideoDecoder != NULL); instantiateDecoder(false, &mVideoDecoder); if (mAudioSink != NULL) { instantiateDecoder(true, &mAudioDecoder); } if (!mSource->feedMoreTSData()) { if (mAudioDecoder == NULL && mVideoDecoder == NULL) { // We're not currently decoding anything (no audio or // video tracks found) and we just ran out of input data. notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); } break; } if (mAudioDecoder == NULL || mVideoDecoder == NULL) { msg->post(100000ll); mScanSourcesPending = true; } break; } case kWhatVideoNotify: case kWhatAudioNotify: { bool audio = msg->what() == kWhatAudioNotify; sp codecRequest; CHECK(msg->findMessage("codec-request", &codecRequest)); int32_t what; CHECK(codecRequest->findInt32("what", &what)); if (what == ACodec::kWhatFillThisBuffer) { status_t err = feedDecoderInputData( audio, codecRequest); if (err == -EWOULDBLOCK) { if (mSource->feedMoreTSData()) { msg->post(); } } } else if (what == ACodec::kWhatEOS) { mRenderer->queueEOS(audio, ERROR_END_OF_STREAM); } else if (what == ACodec::kWhatFlushCompleted) { bool needShutdown; if (audio) { CHECK(IsFlushingState(mFlushingAudio, &needShutdown)); mFlushingAudio = FLUSHED; } else { CHECK(IsFlushingState(mFlushingVideo, &needShutdown)); mFlushingVideo = FLUSHED; mVideoLateByUs = 0; } LOGV("decoder %s flush completed", audio ? "audio" : "video"); if (needShutdown) { LOGV("initiating %s decoder shutdown", audio ? "audio" : "video"); (audio ? mAudioDecoder : mVideoDecoder)->initiateShutdown(); if (audio) { mFlushingAudio = SHUTTING_DOWN_DECODER; } else { mFlushingVideo = SHUTTING_DOWN_DECODER; } } finishFlushIfPossible(); } else if (what == ACodec::kWhatOutputFormatChanged) { if (audio) { int32_t numChannels; CHECK(codecRequest->findInt32("channel-count", &numChannels)); int32_t sampleRate; CHECK(codecRequest->findInt32("sample-rate", &sampleRate)); LOGV("Audio output format changed to %d Hz, %d channels", sampleRate, numChannels); mAudioSink->close(); CHECK_EQ(mAudioSink->open( sampleRate, numChannels, AUDIO_FORMAT_PCM_16_BIT, 8 /* bufferCount */), (status_t)OK); mAudioSink->start(); mRenderer->signalAudioSinkChanged(); } else { // video int32_t width, height; CHECK(codecRequest->findInt32("width", &width)); CHECK(codecRequest->findInt32("height", &height)); int32_t cropLeft, cropTop, cropRight, cropBottom; CHECK(codecRequest->findRect( "crop", &cropLeft, &cropTop, &cropRight, &cropBottom)); LOGV("Video output format changed to %d x %d " "(crop: %d x %d @ (%d, %d))", width, height, (cropRight - cropLeft + 1), (cropBottom - cropTop + 1), cropLeft, cropTop); notifyListener( MEDIA_SET_VIDEO_SIZE, cropRight - cropLeft + 1, cropBottom - cropTop + 1); } } else if (what == ACodec::kWhatShutdownCompleted) { LOGV("%s shutdown completed", audio ? "audio" : "video"); if (audio) { mAudioDecoder.clear(); CHECK_EQ((int)mFlushingAudio, (int)SHUTTING_DOWN_DECODER); mFlushingAudio = SHUT_DOWN; } else { mVideoDecoder.clear(); CHECK_EQ((int)mFlushingVideo, (int)SHUTTING_DOWN_DECODER); mFlushingVideo = SHUT_DOWN; } finishFlushIfPossible(); } else if (what == ACodec::kWhatError) { LOGE("Received error from %s decoder, aborting playback.", audio ? "audio" : "video"); mRenderer->queueEOS(audio, UNKNOWN_ERROR); } else { CHECK_EQ((int)what, (int)ACodec::kWhatDrainThisBuffer); renderBuffer(audio, codecRequest); } break; } case kWhatRendererNotify: { int32_t what; CHECK(msg->findInt32("what", &what)); if (what == Renderer::kWhatEOS) { int32_t audio; CHECK(msg->findInt32("audio", &audio)); int32_t finalResult; CHECK(msg->findInt32("finalResult", &finalResult)); if (audio) { mAudioEOS = true; } else { mVideoEOS = true; } if (finalResult == ERROR_END_OF_STREAM) { LOGV("reached %s EOS", audio ? "audio" : "video"); } else { LOGE("%s track encountered an error (0x%08x)", audio ? "audio" : "video", finalResult); notifyListener( MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, finalResult); } if ((mAudioEOS || mAudioDecoder == NULL) && (mVideoEOS || mVideoDecoder == NULL)) { notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); } } else if (what == Renderer::kWhatPosition) { int64_t positionUs; CHECK(msg->findInt64("positionUs", &positionUs)); CHECK(msg->findInt64("videoLateByUs", &mVideoLateByUs)); if (mDriver != NULL) { sp driver = mDriver.promote(); if (driver != NULL) { driver->notifyPosition(positionUs); driver->notifyFrameStats( mNumFramesTotal, mNumFramesDropped); } } } else if (what == Renderer::kWhatFlushComplete) { CHECK_EQ(what, (int32_t)Renderer::kWhatFlushComplete); int32_t audio; CHECK(msg->findInt32("audio", &audio)); LOGV("renderer %s flush completed.", audio ? "audio" : "video"); } break; } case kWhatMoreDataQueued: { break; } case kWhatReset: { LOGV("kWhatReset"); if (mFlushingAudio != NONE || mFlushingVideo != NONE) { // We're currently flushing, postpone the reset until that's // completed. LOGV("postponing reset"); mResetPostponed = true; break; } if (mAudioDecoder == NULL && mVideoDecoder == NULL) { finishReset(); break; } if (mAudioDecoder != NULL) { flushDecoder(true /* audio */, true /* needShutdown */); } if (mVideoDecoder != NULL) { flushDecoder(false /* audio */, true /* needShutdown */); } mResetInProgress = true; break; } case kWhatSeek: { int64_t seekTimeUs; CHECK(msg->findInt64("seekTimeUs", &seekTimeUs)); LOGV("kWhatSeek seekTimeUs=%lld us (%.2f secs)", seekTimeUs, seekTimeUs / 1E6); mSource->seekTo(seekTimeUs); if (mDriver != NULL) { sp driver = mDriver.promote(); if (driver != NULL) { driver->notifySeekComplete(); } } break; } case kWhatPause: { CHECK(mRenderer != NULL); mRenderer->pause(); break; } case kWhatResume: { CHECK(mRenderer != NULL); mRenderer->resume(); break; } default: TRESPASS(); break; } } void NuPlayer::finishFlushIfPossible() { if (mFlushingAudio != FLUSHED && mFlushingAudio != SHUT_DOWN) { return; } if (mFlushingVideo != FLUSHED && mFlushingVideo != SHUT_DOWN) { return; } LOGV("both audio and video are flushed now."); mRenderer->signalTimeDiscontinuity(); if (mAudioDecoder != NULL) { mAudioDecoder->signalResume(); } if (mVideoDecoder != NULL) { mVideoDecoder->signalResume(); } mFlushingAudio = NONE; mFlushingVideo = NONE; if (mResetInProgress) { LOGV("reset completed"); mResetInProgress = false; finishReset(); } else if (mResetPostponed) { (new AMessage(kWhatReset, id()))->post(); mResetPostponed = false; } else if (mAudioDecoder == NULL || mVideoDecoder == NULL) { postScanSources(); } } void NuPlayer::finishReset() { CHECK(mAudioDecoder == NULL); CHECK(mVideoDecoder == NULL); mRenderer.clear(); mSource.clear(); if (mDriver != NULL) { sp driver = mDriver.promote(); if (driver != NULL) { driver->notifyResetComplete(); } } } void NuPlayer::postScanSources() { if (mScanSourcesPending) { return; } sp msg = new AMessage(kWhatScanSources, id()); msg->setInt32("generation", mScanSourcesGeneration); msg->post(); mScanSourcesPending = true; } status_t NuPlayer::instantiateDecoder(bool audio, sp *decoder) { if (*decoder != NULL) { return OK; } sp meta = mSource->getFormat(audio); if (meta == NULL) { return -EWOULDBLOCK; } if (!audio) { const char *mime; CHECK(meta->findCString(kKeyMIMEType, &mime)); mVideoIsAVC = !strcasecmp(MEDIA_MIMETYPE_VIDEO_AVC, mime); } sp notify = new AMessage(audio ? kWhatAudioNotify : kWhatVideoNotify, id()); *decoder = audio ? new Decoder(notify) : new Decoder(notify, mNativeWindow); looper()->registerHandler(*decoder); (*decoder)->configure(meta); int64_t durationUs; if (mDriver != NULL && mSource->getDuration(&durationUs) == OK) { sp driver = mDriver.promote(); if (driver != NULL) { driver->notifyDuration(durationUs); } } return OK; } status_t NuPlayer::feedDecoderInputData(bool audio, const sp &msg) { sp reply; CHECK(msg->findMessage("reply", &reply)); if ((audio && IsFlushingState(mFlushingAudio)) || (!audio && IsFlushingState(mFlushingVideo))) { reply->setInt32("err", INFO_DISCONTINUITY); reply->post(); return OK; } sp accessUnit; bool dropAccessUnit; do { status_t err = mSource->dequeueAccessUnit(audio, &accessUnit); if (err == -EWOULDBLOCK) { return err; } else if (err != OK) { if (err == INFO_DISCONTINUITY) { int32_t type; CHECK(accessUnit->meta()->findInt32("discontinuity", &type)); bool formatChange = type == ATSParser::DISCONTINUITY_FORMATCHANGE; LOGV("%s discontinuity (formatChange=%d)", audio ? "audio" : "video", formatChange); if (audio) { mSkipRenderingAudioUntilMediaTimeUs = -1; } else { mSkipRenderingVideoUntilMediaTimeUs = -1; } sp extra; if (accessUnit->meta()->findMessage("extra", &extra) && extra != NULL) { int64_t resumeAtMediaTimeUs; if (extra->findInt64( "resume-at-mediatimeUs", &resumeAtMediaTimeUs)) { LOGI("suppressing rendering of %s until %lld us", audio ? "audio" : "video", resumeAtMediaTimeUs); if (audio) { mSkipRenderingAudioUntilMediaTimeUs = resumeAtMediaTimeUs; } else { mSkipRenderingVideoUntilMediaTimeUs = resumeAtMediaTimeUs; } } } flushDecoder(audio, formatChange); } reply->setInt32("err", err); reply->post(); return OK; } if (!audio) { ++mNumFramesTotal; } dropAccessUnit = false; if (!audio && mVideoLateByUs > 100000ll && mVideoIsAVC && !IsAVCReferenceFrame(accessUnit)) { dropAccessUnit = true; ++mNumFramesDropped; } } while (dropAccessUnit); // LOGV("returned a valid buffer of %s data", audio ? "audio" : "video"); #if 0 int64_t mediaTimeUs; CHECK(accessUnit->meta()->findInt64("timeUs", &mediaTimeUs)); LOGV("feeding %s input buffer at media time %.2f secs", audio ? "audio" : "video", mediaTimeUs / 1E6); #endif reply->setObject("buffer", accessUnit); reply->post(); return OK; } void NuPlayer::renderBuffer(bool audio, const sp &msg) { // LOGV("renderBuffer %s", audio ? "audio" : "video"); sp reply; CHECK(msg->findMessage("reply", &reply)); if (IsFlushingState(audio ? mFlushingAudio : mFlushingVideo)) { // We're currently attempting to flush the decoder, in order // to complete this, the decoder wants all its buffers back, // so we don't want any output buffers it sent us (from before // we initiated the flush) to be stuck in the renderer's queue. LOGV("we're still flushing the %s decoder, sending its output buffer" " right back.", audio ? "audio" : "video"); reply->post(); return; } sp obj; CHECK(msg->findObject("buffer", &obj)); sp buffer = static_cast(obj.get()); int64_t &skipUntilMediaTimeUs = audio ? mSkipRenderingAudioUntilMediaTimeUs : mSkipRenderingVideoUntilMediaTimeUs; if (skipUntilMediaTimeUs >= 0) { int64_t mediaTimeUs; CHECK(buffer->meta()->findInt64("timeUs", &mediaTimeUs)); if (mediaTimeUs < skipUntilMediaTimeUs) { LOGV("dropping %s buffer at time %lld as requested.", audio ? "audio" : "video", mediaTimeUs); reply->post(); return; } skipUntilMediaTimeUs = -1; } mRenderer->queueBuffer(audio, buffer, reply); } void NuPlayer::notifyListener(int msg, int ext1, int ext2) { if (mDriver == NULL) { return; } sp driver = mDriver.promote(); if (driver == NULL) { return; } driver->sendEvent(msg, ext1, ext2); } void NuPlayer::flushDecoder(bool audio, bool needShutdown) { // Make sure we don't continue to scan sources until we finish flushing. ++mScanSourcesGeneration; mScanSourcesPending = false; (audio ? mAudioDecoder : mVideoDecoder)->signalFlush(); mRenderer->flush(audio); FlushStatus newStatus = needShutdown ? FLUSHING_DECODER_SHUTDOWN : FLUSHING_DECODER; if (audio) { CHECK(mFlushingAudio == NONE || mFlushingAudio == AWAITING_DISCONTINUITY); mFlushingAudio = newStatus; if (mFlushingVideo == NONE) { mFlushingVideo = (mVideoDecoder != NULL) ? AWAITING_DISCONTINUITY : FLUSHED; } } else { CHECK(mFlushingVideo == NONE || mFlushingVideo == AWAITING_DISCONTINUITY); mFlushingVideo = newStatus; if (mFlushingAudio == NONE) { mFlushingAudio = (mAudioDecoder != NULL) ? AWAITING_DISCONTINUITY : FLUSHED; } } } } // namespace android