/* * Copyright (C) 2018 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #ifdef MPEG4 #define LOG_TAG "C2SoftMpeg4Dec" #else #define LOG_TAG "C2SoftH263Dec" #endif #include <log/log.h> #include <media/stagefright/foundation/AUtils.h> #include <media/stagefright/foundation/MediaDefs.h> #include <C2Debug.h> #include <C2PlatformSupport.h> #include <SimpleC2Interface.h> #include "C2SoftMpeg4Dec.h" #include "mp4dec_api.h" namespace android { #ifdef MPEG4 constexpr char COMPONENT_NAME[] = "c2.android.mpeg4.decoder"; #else constexpr char COMPONENT_NAME[] = "c2.android.h263.decoder"; #endif class C2SoftMpeg4Dec::IntfImpl : public SimpleInterface<void>::BaseParams { public: explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper> &helper) : SimpleInterface<void>::BaseParams( helper, COMPONENT_NAME, C2Component::KIND_DECODER, C2Component::DOMAIN_VIDEO, #ifdef MPEG4 MEDIA_MIMETYPE_VIDEO_MPEG4 #else MEDIA_MIMETYPE_VIDEO_H263 #endif ) { noPrivateBuffers(); // TODO: account for our buffers here noInputReferences(); noOutputReferences(); noInputLatency(); noTimeStretch(); // TODO: Proper support for reorder depth. addParameter( DefineParam(mActualOutputDelay, C2_PARAMKEY_OUTPUT_DELAY) .withConstValue(new C2PortActualDelayTuning::output(1u)) .build()); addParameter( DefineParam(mAttrib, C2_PARAMKEY_COMPONENT_ATTRIBUTES) .withConstValue(new C2ComponentAttributesSetting(C2Component::ATTRIB_IS_TEMPORAL)) .build()); addParameter( DefineParam(mSize, C2_PARAMKEY_PICTURE_SIZE) .withDefault(new C2StreamPictureSizeInfo::output(0u, 176, 144)) .withFields({ #ifdef MPEG4 C2F(mSize, width).inRange(2, 1920, 2), C2F(mSize, height).inRange(2, 1088, 2), #else C2F(mSize, width).inRange(2, 352, 2), C2F(mSize, height).inRange(2, 288, 2), #endif }) .withSetter(SizeSetter) .build()); #ifdef MPEG4 addParameter( DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL) .withDefault(new C2StreamProfileLevelInfo::input(0u, C2Config::PROFILE_MP4V_SIMPLE, C2Config::LEVEL_MP4V_3)) .withFields({ C2F(mProfileLevel, profile).equalTo( C2Config::PROFILE_MP4V_SIMPLE), C2F(mProfileLevel, level).oneOf({ C2Config::LEVEL_MP4V_0, C2Config::LEVEL_MP4V_0B, C2Config::LEVEL_MP4V_1, C2Config::LEVEL_MP4V_2, C2Config::LEVEL_MP4V_3, C2Config::LEVEL_MP4V_3B, C2Config::LEVEL_MP4V_4, C2Config::LEVEL_MP4V_4A, C2Config::LEVEL_MP4V_5, C2Config::LEVEL_MP4V_6}) }) .withSetter(ProfileLevelSetter, mSize) .build()); #else addParameter( DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL) .withDefault(new C2StreamProfileLevelInfo::input(0u, C2Config::PROFILE_H263_BASELINE, C2Config::LEVEL_H263_30)) .withFields({ C2F(mProfileLevel, profile).oneOf({ C2Config::PROFILE_H263_BASELINE, C2Config::PROFILE_H263_ISWV2}), C2F(mProfileLevel, level).oneOf({ C2Config::LEVEL_H263_10, C2Config::LEVEL_H263_20, C2Config::LEVEL_H263_30, C2Config::LEVEL_H263_40, C2Config::LEVEL_H263_45}) }) .withSetter(ProfileLevelSetter, mSize) .build()); #endif addParameter( DefineParam(mMaxSize, C2_PARAMKEY_MAX_PICTURE_SIZE) #ifdef MPEG4 .withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 1920, 1088)) #else .withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 352, 288)) #endif .withFields({ #ifdef MPEG4 C2F(mSize, width).inRange(2, 1920, 2), C2F(mSize, height).inRange(2, 1088, 2), #else C2F(mSize, width).inRange(2, 352, 2), C2F(mSize, height).inRange(2, 288, 2), #endif }) .withSetter(MaxPictureSizeSetter, mSize) .build()); addParameter( DefineParam(mMaxInputSize, C2_PARAMKEY_INPUT_MAX_BUFFER_SIZE) #ifdef MPEG4 .withDefault(new C2StreamMaxBufferSizeInfo::input(0u, 1920 * 1088 * 3 / 2)) #else .withDefault(new C2StreamMaxBufferSizeInfo::input(0u, 352 * 288 * 3 / 2)) #endif .withFields({ C2F(mMaxInputSize, value).any(), }) .calculatedAs(MaxInputSizeSetter, mMaxSize) .build()); C2ChromaOffsetStruct locations[1] = { C2ChromaOffsetStruct::ITU_YUV_420_0() }; std::shared_ptr<C2StreamColorInfo::output> defaultColorInfo = C2StreamColorInfo::output::AllocShared( 1u, 0u, 8u /* bitDepth */, C2Color::YUV_420); memcpy(defaultColorInfo->m.locations, locations, sizeof(locations)); defaultColorInfo = C2StreamColorInfo::output::AllocShared( { C2ChromaOffsetStruct::ITU_YUV_420_0() }, 0u, 8u /* bitDepth */, C2Color::YUV_420); helper->addStructDescriptors<C2ChromaOffsetStruct>(); addParameter( DefineParam(mColorInfo, C2_PARAMKEY_CODED_COLOR_INFO) .withConstValue(defaultColorInfo) .build()); // TODO: support more formats? addParameter( DefineParam(mPixelFormat, C2_PARAMKEY_PIXEL_FORMAT) .withConstValue(new C2StreamPixelFormatInfo::output( 0u, HAL_PIXEL_FORMAT_YCBCR_420_888)) .build()); } static C2R SizeSetter(bool mayBlock, const C2P<C2StreamPictureSizeInfo::output> &oldMe, C2P<C2StreamPictureSizeInfo::output> &me) { (void)mayBlock; C2R res = C2R::Ok(); if (!me.F(me.v.width).supportsAtAll(me.v.width)) { res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width))); me.set().width = oldMe.v.width; } if (!me.F(me.v.height).supportsAtAll(me.v.height)) { res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height))); me.set().height = oldMe.v.height; } return res; } static C2R MaxPictureSizeSetter(bool mayBlock, C2P<C2StreamMaxPictureSizeTuning::output> &me, const C2P<C2StreamPictureSizeInfo::output> &size) { (void)mayBlock; // TODO: get max width/height from the size's field helpers vs. hardcoding #ifdef MPEG4 me.set().width = c2_min(c2_max(me.v.width, size.v.width), 1920u); me.set().height = c2_min(c2_max(me.v.height, size.v.height), 1088u); #else me.set().width = c2_min(c2_max(me.v.width, size.v.width), 352u); me.set().height = c2_min(c2_max(me.v.height, size.v.height), 288u); #endif return C2R::Ok(); } static C2R MaxInputSizeSetter(bool mayBlock, C2P<C2StreamMaxBufferSizeInfo::input> &me, const C2P<C2StreamMaxPictureSizeTuning::output> &maxSize) { (void)mayBlock; // assume compression ratio of 1 me.set().value = (((maxSize.v.width + 15) / 16) * ((maxSize.v.height + 15) / 16) * 384); return C2R::Ok(); } static C2R ProfileLevelSetter(bool mayBlock, C2P<C2StreamProfileLevelInfo::input> &me, const C2P<C2StreamPictureSizeInfo::output> &size) { (void)mayBlock; (void)size; (void)me; // TODO: validate return C2R::Ok(); } uint32_t getMaxWidth() const { return mMaxSize->width; } uint32_t getMaxHeight() const { return mMaxSize->height; } private: std::shared_ptr<C2StreamProfileLevelInfo::input> mProfileLevel; std::shared_ptr<C2StreamPictureSizeInfo::output> mSize; std::shared_ptr<C2StreamMaxPictureSizeTuning::output> mMaxSize; std::shared_ptr<C2StreamMaxBufferSizeInfo::input> mMaxInputSize; std::shared_ptr<C2StreamColorInfo::output> mColorInfo; std::shared_ptr<C2StreamPixelFormatInfo::output> mPixelFormat; }; C2SoftMpeg4Dec::C2SoftMpeg4Dec( const char *name, c2_node_id_t id, const std::shared_ptr<IntfImpl> &intfImpl) : SimpleC2Component(std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)), mIntf(intfImpl), mDecHandle(nullptr), mOutputBuffer{}, mInitialized(false) { } C2SoftMpeg4Dec::~C2SoftMpeg4Dec() { onRelease(); } c2_status_t C2SoftMpeg4Dec::onInit() { status_t err = initDecoder(); return err == OK ? C2_OK : C2_CORRUPTED; } c2_status_t C2SoftMpeg4Dec::onStop() { if (mInitialized) { if (mDecHandle) { PVCleanUpVideoDecoder(mDecHandle); } mInitialized = false; } for (int32_t i = 0; i < kNumOutputBuffers; ++i) { if (mOutputBuffer[i]) { free(mOutputBuffer[i]); mOutputBuffer[i] = nullptr; } } mNumSamplesOutput = 0; mFramesConfigured = false; mSignalledOutputEos = false; mSignalledError = false; return C2_OK; } void C2SoftMpeg4Dec::onReset() { (void)onStop(); (void)onInit(); } void C2SoftMpeg4Dec::onRelease() { if (mInitialized) { if (mDecHandle) { PVCleanUpVideoDecoder(mDecHandle); delete mDecHandle; mDecHandle = nullptr; } mInitialized = false; } if (mOutBlock) { mOutBlock.reset(); } for (int32_t i = 0; i < kNumOutputBuffers; ++i) { if (mOutputBuffer[i]) { free(mOutputBuffer[i]); mOutputBuffer[i] = nullptr; } } } c2_status_t C2SoftMpeg4Dec::onFlush_sm() { if (mInitialized) { if (PV_TRUE != PVResetVideoDecoder(mDecHandle)) { return C2_CORRUPTED; } } mSignalledOutputEos = false; mSignalledError = false; return C2_OK; } status_t C2SoftMpeg4Dec::initDecoder() { #ifdef MPEG4 mIsMpeg4 = true; #else mIsMpeg4 = false; #endif if (!mDecHandle) { mDecHandle = new tagvideoDecControls; } if (!mDecHandle) { ALOGE("mDecHandle is null"); return NO_MEMORY; } memset(mDecHandle, 0, sizeof(tagvideoDecControls)); /* TODO: bring these values to 352 and 288. It cannot be done as of now * because, h263 doesn't seem to allow port reconfiguration. In OMX, the * problem of larger width and height than default width and height is * overcome by adaptivePlayBack() api call. This call gets width and height * information from extractor. Such a thing is not possible here. * So we are configuring to larger values.*/ mWidth = 1408; mHeight = 1152; mNumSamplesOutput = 0; mInitialized = false; mFramesConfigured = false; mSignalledOutputEos = false; mSignalledError = false; return OK; } void fillEmptyWork(const std::unique_ptr<C2Work> &work) { uint32_t flags = 0; if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) { flags |= C2FrameData::FLAG_END_OF_STREAM; ALOGV("signalling eos"); } work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; work->worklets.front()->output.buffers.clear(); work->worklets.front()->output.ordinal = work->input.ordinal; work->workletsProcessed = 1u; } void C2SoftMpeg4Dec::finishWork(uint64_t index, const std::unique_ptr<C2Work> &work) { std::shared_ptr<C2Buffer> buffer = createGraphicBuffer(std::move(mOutBlock), C2Rect(mWidth, mHeight)); mOutBlock = nullptr; auto fillWork = [buffer, index](const std::unique_ptr<C2Work> &work) { uint32_t flags = 0; if ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) && (c2_cntr64_t(index) == work->input.ordinal.frameIndex)) { flags |= C2FrameData::FLAG_END_OF_STREAM; ALOGV("signalling eos"); } work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; work->worklets.front()->output.buffers.clear(); work->worklets.front()->output.buffers.push_back(buffer); work->worklets.front()->output.ordinal = work->input.ordinal; work->workletsProcessed = 1u; }; if (work && c2_cntr64_t(index) == work->input.ordinal.frameIndex) { fillWork(work); } else { finish(index, fillWork); } } c2_status_t C2SoftMpeg4Dec::ensureDecoderState(const std::shared_ptr<C2BlockPool> &pool) { if (!mDecHandle) { ALOGE("not supposed to be here, invalid decoder context"); return C2_CORRUPTED; } mOutputBufferSize = align(mIntf->getMaxWidth(), 16) * align(mIntf->getMaxHeight(), 16) * 3 / 2; for (int32_t i = 0; i < kNumOutputBuffers; ++i) { if (!mOutputBuffer[i]) { mOutputBuffer[i] = (uint8_t *)malloc(mOutputBufferSize); if (!mOutputBuffer[i]) { return C2_NO_MEMORY; } } } if (mOutBlock && (mOutBlock->width() != align(mWidth, 16) || mOutBlock->height() != mHeight)) { mOutBlock.reset(); } if (!mOutBlock) { uint32_t format = HAL_PIXEL_FORMAT_YV12; C2MemoryUsage usage = { C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE }; c2_status_t err = pool->fetchGraphicBlock(align(mWidth, 16), mHeight, format, usage, &mOutBlock); if (err != C2_OK) { ALOGE("fetchGraphicBlock for Output failed with status %d", err); return err; } ALOGV("provided (%dx%d) required (%dx%d)", mOutBlock->width(), mOutBlock->height(), mWidth, mHeight); } return C2_OK; } bool C2SoftMpeg4Dec::handleResChange(const std::unique_ptr<C2Work> &work) { uint32_t disp_width, disp_height; PVGetVideoDimensions(mDecHandle, (int32 *)&disp_width, (int32 *)&disp_height); uint32_t buf_width, buf_height; PVGetBufferDimensions(mDecHandle, (int32 *)&buf_width, (int32 *)&buf_height); CHECK_LE(disp_width, buf_width); CHECK_LE(disp_height, buf_height); ALOGV("display size (%dx%d), buffer size (%dx%d)", disp_width, disp_height, buf_width, buf_height); bool resChanged = false; if (disp_width != mWidth || disp_height != mHeight) { mWidth = disp_width; mHeight = disp_height; resChanged = true; for (int32_t i = 0; i < kNumOutputBuffers; ++i) { if (mOutputBuffer[i]) { free(mOutputBuffer[i]); mOutputBuffer[i] = nullptr; } } if (!mIsMpeg4) { PVCleanUpVideoDecoder(mDecHandle); uint8_t *vol_data[1]{}; int32_t vol_size = 0; if (!PVInitVideoDecoder( mDecHandle, vol_data, &vol_size, 1, mIntf->getMaxWidth(), mIntf->getMaxHeight(), H263_MODE)) { ALOGE("Error in PVInitVideoDecoder H263_MODE while resChanged was set to true"); mSignalledError = true; work->result = C2_CORRUPTED; return true; } } mFramesConfigured = false; } return resChanged; } /* TODO: can remove temporary copy after library supports writing to display * buffer Y, U and V plane pointers using stride info. */ static void copyOutputBufferToYuvPlanarFrame( uint8_t *dst, uint8_t *src, size_t dstYStride, size_t dstUVStride, size_t srcYStride, uint32_t width, uint32_t height) { size_t srcUVStride = srcYStride / 2; uint8_t *srcStart = src; uint8_t *dstStart = dst; size_t vStride = align(height, 16); for (size_t i = 0; i < height; ++i) { memcpy(dst, src, width); src += srcYStride; dst += dstYStride; } /* U buffer */ src = srcStart + vStride * srcYStride; dst = dstStart + (dstYStride * height) + (dstUVStride * height / 2); for (size_t i = 0; i < height / 2; ++i) { memcpy(dst, src, width / 2); src += srcUVStride; dst += dstUVStride; } /* V buffer */ src = srcStart + vStride * srcYStride * 5 / 4; dst = dstStart + (dstYStride * height); for (size_t i = 0; i < height / 2; ++i) { memcpy(dst, src, width / 2); src += srcUVStride; dst += dstUVStride; } } void C2SoftMpeg4Dec::process( const std::unique_ptr<C2Work> &work, const std::shared_ptr<C2BlockPool> &pool) { // Initialize output work work->result = C2_OK; work->workletsProcessed = 1u; work->worklets.front()->output.configUpdate.clear(); work->worklets.front()->output.flags = work->input.flags; if (mSignalledError || mSignalledOutputEos) { work->result = C2_BAD_VALUE; return; } size_t inOffset = 0u; size_t inSize = 0u; uint32_t workIndex = work->input.ordinal.frameIndex.peeku() & 0xFFFFFFFF; C2ReadView rView = mDummyReadView; if (!work->input.buffers.empty()) { rView = work->input.buffers[0]->data().linearBlocks().front().map().get(); inSize = rView.capacity(); if (inSize && rView.error()) { ALOGE("read view map failed %d", rView.error()); work->result = C2_CORRUPTED; return; } } ALOGV("in buffer attr. size %zu timestamp %d frameindex %d, flags %x", inSize, (int)work->input.ordinal.timestamp.peeku(), (int)work->input.ordinal.frameIndex.peeku(), work->input.flags); bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0); if (inSize == 0) { fillEmptyWork(work); if (eos) { mSignalledOutputEos = true; } return; } uint8_t *bitstream = const_cast<uint8_t *>(rView.data() + inOffset); uint32_t *start_code = (uint32_t *)bitstream; bool volHeader = *start_code == 0xB0010000; if (volHeader) { PVCleanUpVideoDecoder(mDecHandle); mInitialized = false; } if (!mInitialized) { uint8_t *vol_data[1]{}; int32_t vol_size = 0; bool codecConfig = (work->input.flags & C2FrameData::FLAG_CODEC_CONFIG) != 0; if (codecConfig || volHeader) { vol_data[0] = bitstream; vol_size = inSize; } MP4DecodingMode mode = (mIsMpeg4) ? MPEG4_MODE : H263_MODE; if (!PVInitVideoDecoder( mDecHandle, vol_data, &vol_size, 1, mIntf->getMaxWidth(), mIntf->getMaxHeight(), mode)) { ALOGE("PVInitVideoDecoder failed. Unsupported content?"); mSignalledError = true; work->result = C2_CORRUPTED; return; } mInitialized = true; MP4DecodingMode actualMode = PVGetDecBitstreamMode(mDecHandle); if (mode != actualMode) { ALOGE("Decoded mode not same as actual mode of the decoder"); mSignalledError = true; work->result = C2_CORRUPTED; return; } PVSetPostProcType(mDecHandle, 0); if (handleResChange(work)) { ALOGI("Setting width and height"); C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight); std::vector<std::unique_ptr<C2SettingResult>> failures; c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures); if (err == OK) { work->worklets.front()->output.configUpdate.push_back( C2Param::Copy(size)); } else { ALOGE("Config update size failed"); mSignalledError = true; work->result = C2_CORRUPTED; return; } } if (codecConfig) { fillEmptyWork(work); return; } } size_t inPos = 0; while (inPos < inSize) { c2_status_t err = ensureDecoderState(pool); if (C2_OK != err) { mSignalledError = true; work->result = err; return; } C2GraphicView wView = mOutBlock->map().get(); if (wView.error()) { ALOGE("graphic view map failed %d", wView.error()); work->result = C2_CORRUPTED; return; } uint32_t yFrameSize = sizeof(uint8) * mDecHandle->size; if (mOutputBufferSize < yFrameSize * 3 / 2){ ALOGE("Too small output buffer: %zu bytes", mOutputBufferSize); mSignalledError = true; work->result = C2_NO_MEMORY; return; } if (!mFramesConfigured) { PVSetReferenceYUV(mDecHandle,mOutputBuffer[1]); mFramesConfigured = true; } // Need to check if header contains new info, e.g., width/height, etc. VopHeaderInfo header_info; uint32_t useExtTimestamp = (inPos == 0); int32_t tmpInSize = (int32_t)inSize; uint8_t *bitstreamTmp = bitstream; uint32_t timestamp = workIndex; if (PVDecodeVopHeader( mDecHandle, &bitstreamTmp, ×tamp, &tmpInSize, &header_info, &useExtTimestamp, mOutputBuffer[mNumSamplesOutput & 1]) != PV_TRUE) { ALOGE("failed to decode vop header."); mSignalledError = true; work->result = C2_CORRUPTED; return; } // H263 doesn't have VOL header, the frame size information is in short header, i.e. the // decoder may detect size change after PVDecodeVopHeader. bool resChange = handleResChange(work); if (mIsMpeg4 && resChange) { mSignalledError = true; work->result = C2_CORRUPTED; return; } else if (resChange) { ALOGI("Setting width and height"); C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight); std::vector<std::unique_ptr<C2SettingResult>> failures; c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures); if (err == OK) { work->worklets.front()->output.configUpdate.push_back(C2Param::Copy(size)); } else { ALOGE("Config update size failed"); mSignalledError = true; work->result = C2_CORRUPTED; return; } continue; } if (PVDecodeVopBody(mDecHandle, &tmpInSize) != PV_TRUE) { ALOGE("failed to decode video frame."); mSignalledError = true; work->result = C2_CORRUPTED; return; } if (handleResChange(work)) { mSignalledError = true; work->result = C2_CORRUPTED; return; } uint8_t *outputBufferY = wView.data()[C2PlanarLayout::PLANE_Y]; C2PlanarLayout layout = wView.layout(); size_t dstYStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc; size_t dstUVStride = layout.planes[C2PlanarLayout::PLANE_U].rowInc; (void)copyOutputBufferToYuvPlanarFrame( outputBufferY, mOutputBuffer[mNumSamplesOutput & 1], dstYStride, dstUVStride, align(mWidth, 16), mWidth, mHeight); inPos += inSize - (size_t)tmpInSize; finishWork(workIndex, work); ++mNumSamplesOutput; if (inSize - inPos != 0) { ALOGD("decoded frame, ignoring further trailing bytes %d", (int)inSize - (int)inPos); break; } } } c2_status_t C2SoftMpeg4Dec::drain( uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool) { (void)pool; if (drainMode == NO_DRAIN) { ALOGW("drain with NO_DRAIN: no-op"); return C2_OK; } if (drainMode == DRAIN_CHAIN) { ALOGW("DRAIN_CHAIN not supported"); return C2_OMITTED; } return C2_OK; } class C2SoftMpeg4DecFactory : public C2ComponentFactory { public: C2SoftMpeg4DecFactory() : mHelper(std::static_pointer_cast<C2ReflectorHelper>( GetCodec2PlatformComponentStore()->getParamReflector())) { } virtual c2_status_t createComponent( c2_node_id_t id, std::shared_ptr<C2Component>* const component, std::function<void(C2Component*)> deleter) override { *component = std::shared_ptr<C2Component>( new C2SoftMpeg4Dec(COMPONENT_NAME, id, std::make_shared<C2SoftMpeg4Dec::IntfImpl>(mHelper)), deleter); return C2_OK; } virtual c2_status_t createInterface( c2_node_id_t id, std::shared_ptr<C2ComponentInterface>* const interface, std::function<void(C2ComponentInterface*)> deleter) override { *interface = std::shared_ptr<C2ComponentInterface>( new SimpleInterface<C2SoftMpeg4Dec::IntfImpl>( COMPONENT_NAME, id, std::make_shared<C2SoftMpeg4Dec::IntfImpl>(mHelper)), deleter); return C2_OK; } virtual ~C2SoftMpeg4DecFactory() override = default; private: std::shared_ptr<C2ReflectorHelper> mHelper; }; } // namespace android extern "C" ::C2ComponentFactory* CreateCodec2Factory() { ALOGV("in %s", __func__); return new ::android::C2SoftMpeg4DecFactory(); } extern "C" void DestroyCodec2Factory(::C2ComponentFactory* factory) { ALOGV("in %s", __func__); delete factory; }