/*
* Copyright (c) 2009-2011 Intel Corporation. All rights reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <va/va.h>
#include "VideoDecoderBase.h"
#include "VideoDecoderAVC.h"
#include "VideoDecoderTrace.h"
#include "vbp_loader.h"
#include "VideoDecoderAVCSecure.h"
#include "VideoFrameInfo.h"
#define MAX_SLICEHEADER_BUFFER_SIZE 4096
#define STARTCODE_PREFIX_LEN 3
#define NALU_TYPE_MASK 0x1F
#define MAX_NALU_HEADER_BUFFER 8192
static const uint8_t startcodePrefix[STARTCODE_PREFIX_LEN] = {0x00, 0x00, 0x01};
/* H264 start code values */
typedef enum _h264_nal_unit_type
{
h264_NAL_UNIT_TYPE_unspecified = 0,
h264_NAL_UNIT_TYPE_SLICE,
h264_NAL_UNIT_TYPE_DPA,
h264_NAL_UNIT_TYPE_DPB,
h264_NAL_UNIT_TYPE_DPC,
h264_NAL_UNIT_TYPE_IDR,
h264_NAL_UNIT_TYPE_SEI,
h264_NAL_UNIT_TYPE_SPS,
h264_NAL_UNIT_TYPE_PPS,
h264_NAL_UNIT_TYPE_Acc_unit_delimiter,
h264_NAL_UNIT_TYPE_EOSeq,
h264_NAL_UNIT_TYPE_EOstream,
h264_NAL_UNIT_TYPE_filler_data,
h264_NAL_UNIT_TYPE_SPS_extension,
h264_NAL_UNIT_TYPE_ACP = 19,
h264_NAL_UNIT_TYPE_Slice_extension = 20
} h264_nal_unit_type_t;
VideoDecoderAVCSecure::VideoDecoderAVCSecure(const char *mimeType)
: VideoDecoderAVC(mimeType){
mFrameSize = 0;
mFrameData = NULL;
mIsEncryptData = 0;
mClearData = NULL;
mCachedHeader = NULL;
setParserType(VBP_H264SECURE);
mFrameIdx = 0;
mModularMode = 0;
mSliceNum = 0;
}
Decode_Status VideoDecoderAVCSecure::start(VideoConfigBuffer *buffer) {
VTRACE("VideoDecoderAVCSecure::start");
Decode_Status status = VideoDecoderAVC::start(buffer);
if (status != DECODE_SUCCESS) {
return status;
}
mClearData = new uint8_t [MAX_NALU_HEADER_BUFFER];
if (mClearData == NULL) {
ETRACE("Failed to allocate memory for mClearData");
return DECODE_MEMORY_FAIL;
}
mCachedHeader= new uint8_t [MAX_SLICEHEADER_BUFFER_SIZE];
if (mCachedHeader == NULL) {
ETRACE("Failed to allocate memory for mCachedHeader");
return DECODE_MEMORY_FAIL;
}
return status;
}
void VideoDecoderAVCSecure::stop(void) {
VTRACE("VideoDecoderAVCSecure::stop");
VideoDecoderAVC::stop();
if (mClearData) {
delete [] mClearData;
mClearData = NULL;
}
if (mCachedHeader) {
delete [] mCachedHeader;
mCachedHeader = NULL;
}
}
Decode_Status VideoDecoderAVCSecure::processModularInputBuffer(VideoDecodeBuffer *buffer, vbp_data_h264 **data)
{
VTRACE("processModularInputBuffer +++");
Decode_Status status;
int32_t clear_data_size = 0;
uint8_t *clear_data = NULL;
int32_t nalu_num = 0;
uint8_t nalu_type = 0;
int32_t nalu_offset = 0;
uint32_t nalu_size = 0;
uint8_t naluType = 0;
uint8_t *nalu_data = NULL;
uint32_t sliceidx = 0;
frame_info_t *pFrameInfo = NULL;
mSliceNum = 0;
memset(&mSliceInfo, 0, sizeof(mSliceInfo));
mIsEncryptData = 0;
if (buffer->flag & IS_SECURE_DATA) {
VTRACE("Decoding protected video ...");
pFrameInfo = (frame_info_t *) buffer->data;
if (pFrameInfo == NULL) {
ETRACE("Invalid parameter: pFrameInfo is NULL!");
return DECODE_MEMORY_FAIL;
}
mFrameData = pFrameInfo->data;
mFrameSize = pFrameInfo->size;
VTRACE("mFrameData = %p, mFrameSize = %d", mFrameData, mFrameSize);
nalu_num = pFrameInfo->num_nalus;
VTRACE("nalu_num = %d", nalu_num);
if (nalu_num <= 0 || nalu_num >= MAX_NUM_NALUS) {
ETRACE("Invalid parameter: nalu_num = %d", nalu_num);
return DECODE_MEMORY_FAIL;
}
for (int32_t i = 0; i < nalu_num; i++) {
nalu_size = pFrameInfo->nalus[i].length;
nalu_type = pFrameInfo->nalus[i].type;
nalu_offset = pFrameInfo->nalus[i].offset;
nalu_data = pFrameInfo->nalus[i].data;
naluType = nalu_type & NALU_TYPE_MASK;
VTRACE("nalu_type = 0x%x, nalu_size = %d, nalu_offset = 0x%x", nalu_type, nalu_size, nalu_offset);
if (naluType >= h264_NAL_UNIT_TYPE_SLICE && naluType <= h264_NAL_UNIT_TYPE_IDR) {
mIsEncryptData = 1;
VTRACE("slice idx = %d", sliceidx);
mSliceInfo[sliceidx].sliceHeaderByte = nalu_type;
mSliceInfo[sliceidx].sliceStartOffset = (nalu_offset >> 4) << 4;
mSliceInfo[sliceidx].sliceByteOffset = nalu_offset - mSliceInfo[sliceidx].sliceStartOffset;
mSliceInfo[sliceidx].sliceLength = mSliceInfo[sliceidx].sliceByteOffset + nalu_size;
mSliceInfo[sliceidx].sliceSize = (mSliceInfo[sliceidx].sliceByteOffset + nalu_size + 0xF) & ~0xF;
VTRACE("sliceHeaderByte = 0x%x", mSliceInfo[sliceidx].sliceHeaderByte);
VTRACE("sliceStartOffset = %d", mSliceInfo[sliceidx].sliceStartOffset);
VTRACE("sliceByteOffset = %d", mSliceInfo[sliceidx].sliceByteOffset);
VTRACE("sliceSize = %d", mSliceInfo[sliceidx].sliceSize);
VTRACE("sliceLength = %d", mSliceInfo[sliceidx].sliceLength);
#if 0
uint32_t testsize;
uint8_t *testdata;
testsize = mSliceInfo[sliceidx].sliceSize > 64 ? 64 : mSliceInfo[sliceidx].sliceSize ;
testdata = (uint8_t *)(mFrameData);
for (int i = 0; i < testsize; i++) {
VTRACE("testdata[%d] = 0x%x", i, testdata[i]);
}
#endif
sliceidx++;
} else if (naluType == h264_NAL_UNIT_TYPE_SPS || naluType == h264_NAL_UNIT_TYPE_PPS) {
if (nalu_data == NULL) {
ETRACE("Invalid parameter: nalu_data = NULL for naluType 0x%x", naluType);
return DECODE_MEMORY_FAIL;
}
memcpy(mClearData + clear_data_size,
nalu_data,
nalu_size);
clear_data_size += nalu_size;
} else {
ITRACE("Nalu type = 0x%x is skipped", naluType);
continue;
}
}
clear_data = mClearData;
mSliceNum = sliceidx;
} else {
VTRACE("Decoding clear video ...");
mIsEncryptData = 0;
mFrameSize = buffer->size;
mFrameData = buffer->data;
clear_data = buffer->data;
clear_data_size = buffer->size;
}
if (clear_data_size > 0) {
status = VideoDecoderBase::parseBuffer(
clear_data,
clear_data_size,
false,
(void**)data);
CHECK_STATUS("VideoDecoderBase::parseBuffer");
} else {
status = VideoDecoderBase::queryBuffer((void**)data);
CHECK_STATUS("VideoDecoderBase::queryBuffer");
}
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::processClassicInputBuffer(VideoDecodeBuffer *buffer, vbp_data_h264 **data)
{
Decode_Status status;
int32_t clear_data_size = 0;
uint8_t *clear_data = NULL;
uint8_t naluType = 0;
int32_t num_nalus;
int32_t nalu_offset;
int32_t offset;
uint8_t *data_src;
uint8_t *nalu_data;
uint32_t nalu_size;
if (buffer->flag & IS_SECURE_DATA) {
VTRACE("Decoding protected video ...");
mIsEncryptData = 1;
mFrameData = buffer->data;
mFrameSize = buffer->size;
VTRACE("mFrameData = %p, mFrameSize = %d", mFrameData, mFrameSize);
num_nalus = *(uint32_t *)(buffer->data + buffer->size + sizeof(uint32_t));
VTRACE("num_nalus = %d", num_nalus);
offset = 4;
for (int32_t i = 0; i < num_nalus; i++) {
VTRACE("%d nalu, offset = %d", i, offset);
data_src = buffer->data + buffer->size + sizeof(uint32_t) + offset;
nalu_size = *(uint32_t *)(data_src + 2 * sizeof(uint32_t));
nalu_size = (nalu_size + 0x03) & (~0x03);
nalu_data = data_src + 3 *sizeof(uint32_t);
naluType = nalu_data[0] & NALU_TYPE_MASK;
offset += nalu_size + 3 *sizeof(uint32_t);
VTRACE("naluType = 0x%x", naluType);
VTRACE("nalu_size = %d, nalu_data = %p", nalu_size, nalu_data);
if (naluType >= h264_NAL_UNIT_TYPE_SLICE && naluType <= h264_NAL_UNIT_TYPE_IDR) {
ETRACE("Slice NALU received!");
return DECODE_INVALID_DATA;
}
else if (naluType >= h264_NAL_UNIT_TYPE_SEI && naluType <= h264_NAL_UNIT_TYPE_PPS) {
memcpy(mClearData + clear_data_size,
startcodePrefix,
STARTCODE_PREFIX_LEN);
clear_data_size += STARTCODE_PREFIX_LEN;
memcpy(mClearData + clear_data_size,
nalu_data,
nalu_size);
clear_data_size += nalu_size;
} else {
ETRACE("Failure: DECODE_FRAME_DROPPED");
return DECODE_FRAME_DROPPED;
}
}
clear_data = mClearData;
} else {
VTRACE("Decoding clear video ...");
mIsEncryptData = 0;
mFrameSize = buffer->size;
mFrameData = buffer->data;
clear_data = buffer->data;
clear_data_size = buffer->size;
}
if (clear_data_size > 0) {
status = VideoDecoderBase::parseBuffer(
clear_data,
clear_data_size,
false,
(void**)data);
CHECK_STATUS("VideoDecoderBase::parseBuffer");
} else {
status = VideoDecoderBase::queryBuffer((void**)data);
CHECK_STATUS("VideoDecoderBase::queryBuffer");
}
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::decode(VideoDecodeBuffer *buffer) {
VTRACE("VideoDecoderAVCSecure::decode");
Decode_Status status;
vbp_data_h264 *data = NULL;
if (buffer == NULL) {
return DECODE_INVALID_DATA;
}
#if 0
uint32_t testsize;
uint8_t *testdata;
testsize = buffer->size > 16 ? 16:buffer->size ;
testdata = (uint8_t *)(buffer->data);
for (int i = 0; i < 16; i++) {
VTRACE("testdata[%d] = 0x%x", i, testdata[i]);
}
#endif
if (buffer->flag & IS_SUBSAMPLE_ENCRYPTION) {
mModularMode = 1;
}
if (mModularMode) {
status = processModularInputBuffer(buffer,&data);
CHECK_STATUS("processModularInputBuffer");
}
else {
status = processClassicInputBuffer(buffer,&data);
CHECK_STATUS("processClassicInputBuffer");
}
if (!mVAStarted) {
if (data->has_sps && data->has_pps) {
status = startVA(data);
CHECK_STATUS("startVA");
} else {
WTRACE("Can't start VA as either SPS or PPS is still not available.");
return DECODE_SUCCESS;
}
}
status = decodeFrame(buffer, data);
return status;
}
Decode_Status VideoDecoderAVCSecure::decodeFrame(VideoDecodeBuffer *buffer, vbp_data_h264 *data) {
VTRACE("VideoDecoderAVCSecure::decodeFrame");
Decode_Status status;
VTRACE("data->has_sps = %d, data->has_pps = %d", data->has_sps, data->has_pps);
#if 0
// Don't remove the following codes, it can be enabled for debugging DPB.
for (unsigned int i = 0; i < data->num_pictures; i++) {
VAPictureH264 &pic = data->pic_data[i].pic_parms->CurrPic;
VTRACE("%d: decoding frame %.2f, poc top = %d, poc bottom = %d, flags = %d, reference = %d",
i,
buffer->timeStamp/1E6,
pic.TopFieldOrderCnt,
pic.BottomFieldOrderCnt,
pic.flags,
(pic.flags & VA_PICTURE_H264_SHORT_TERM_REFERENCE) ||
(pic.flags & VA_PICTURE_H264_LONG_TERM_REFERENCE));
}
#endif
if (data->new_sps || data->new_pps) {
status = handleNewSequence(data);
CHECK_STATUS("handleNewSequence");
}
if (mModularMode && (!mIsEncryptData)) {
if (data->pic_data[0].num_slices == 0) {
ITRACE("No slice available for decoding.");
status = mSizeChanged ? DECODE_FORMAT_CHANGE : DECODE_SUCCESS;
mSizeChanged = false;
return status;
}
}
uint64_t lastPTS = mCurrentPTS;
mCurrentPTS = buffer->timeStamp;
// start decoding a new frame
status = acquireSurfaceBuffer();
CHECK_STATUS("acquireSurfaceBuffer");
if (mModularMode) {
parseModularSliceHeader(buffer,data);
}
else {
parseClassicSliceHeader(buffer,data);
}
if (status != DECODE_SUCCESS) {
endDecodingFrame(true);
return status;
}
status = beginDecodingFrame(data);
CHECK_STATUS("beginDecodingFrame");
// finish decoding the last frame
status = endDecodingFrame(false);
CHECK_STATUS("endDecodingFrame");
if (isNewFrame(data, lastPTS == mCurrentPTS) == 0) {
ETRACE("Can't handle interlaced frames yet");
return DECODE_FAIL;
}
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::beginDecodingFrame(vbp_data_h264 *data) {
VTRACE("VideoDecoderAVCSecure::beginDecodingFrame");
Decode_Status status;
VAPictureH264 *picture = &(data->pic_data[0].pic_parms->CurrPic);
if ((picture->flags & VA_PICTURE_H264_SHORT_TERM_REFERENCE) ||
(picture->flags & VA_PICTURE_H264_LONG_TERM_REFERENCE)) {
mAcquiredBuffer->referenceFrame = true;
} else {
mAcquiredBuffer->referenceFrame = false;
}
if (picture->flags & VA_PICTURE_H264_TOP_FIELD) {
mAcquiredBuffer->renderBuffer.scanFormat = VA_BOTTOM_FIELD | VA_TOP_FIELD;
} else {
mAcquiredBuffer->renderBuffer.scanFormat = VA_FRAME_PICTURE;
}
mAcquiredBuffer->renderBuffer.flag = 0;
mAcquiredBuffer->renderBuffer.timeStamp = mCurrentPTS;
mAcquiredBuffer->pictureOrder = getPOC(picture);
if (mSizeChanged) {
mAcquiredBuffer->renderBuffer.flag |= IS_RESOLUTION_CHANGE;
mSizeChanged = false;
}
status = continueDecodingFrame(data);
return status;
}
Decode_Status VideoDecoderAVCSecure::continueDecodingFrame(vbp_data_h264 *data) {
VTRACE("VideoDecoderAVCSecure::continueDecodingFrame");
Decode_Status status;
vbp_picture_data_h264 *picData = data->pic_data;
if (mAcquiredBuffer == NULL || mAcquiredBuffer->renderBuffer.surface == VA_INVALID_SURFACE) {
ETRACE("mAcquiredBuffer is NULL. Implementation bug.");
return DECODE_FAIL;
}
VTRACE("data->num_pictures = %d", data->num_pictures);
for (uint32_t picIndex = 0; picIndex < data->num_pictures; picIndex++, picData++) {
if (picData == NULL || picData->pic_parms == NULL || picData->slc_data == NULL || picData->num_slices == 0) {
return DECODE_PARSER_FAIL;
}
if (picIndex > 0 &&
(picData->pic_parms->CurrPic.flags & (VA_PICTURE_H264_TOP_FIELD | VA_PICTURE_H264_BOTTOM_FIELD)) == 0) {
ETRACE("Packed frame is not supported yet!");
return DECODE_FAIL;
}
VTRACE("picData->num_slices = %d", picData->num_slices);
for (uint32_t sliceIndex = 0; sliceIndex < picData->num_slices; sliceIndex++) {
status = decodeSlice(data, picIndex, sliceIndex);
if (status != DECODE_SUCCESS) {
endDecodingFrame(true);
// remove current frame from DPB as it can't be decoded.
removeReferenceFromDPB(picData->pic_parms);
return status;
}
}
}
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::parseClassicSliceHeader(VideoDecodeBuffer *buffer, vbp_data_h264 *data) {
Decode_Status status;
VAStatus vaStatus;
VABufferID sliceheaderbufferID;
VABufferID pictureparameterparsingbufferID;
VABufferID mSlicebufferID;
if (mFrameSize <= 0) {
return DECODE_SUCCESS;
}
vaStatus = vaBeginPicture(mVADisplay, mVAContext, mAcquiredBuffer->renderBuffer.surface);
CHECK_VA_STATUS("vaBeginPicture");
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAParseSliceHeaderGroupBufferType,
MAX_SLICEHEADER_BUFFER_SIZE,
1,
NULL,
&sliceheaderbufferID);
CHECK_VA_STATUS("vaCreateSliceHeaderGroupBuffer");
void *sliceheaderbuf;
vaStatus = vaMapBuffer(
mVADisplay,
sliceheaderbufferID,
&sliceheaderbuf);
CHECK_VA_STATUS("vaMapBuffer");
memset(sliceheaderbuf, 0, MAX_SLICEHEADER_BUFFER_SIZE);
vaStatus = vaUnmapBuffer(
mVADisplay,
sliceheaderbufferID);
CHECK_VA_STATUS("vaUnmapBuffer");
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VASliceDataBufferType,
mFrameSize, //size
1, //num_elements
mFrameData,
&mSlicebufferID);
CHECK_VA_STATUS("vaCreateSliceDataBuffer");
data->pic_parse_buffer->frame_buf_id = mSlicebufferID;
data->pic_parse_buffer->slice_headers_buf_id = sliceheaderbufferID;
data->pic_parse_buffer->frame_size = mFrameSize;
data->pic_parse_buffer->slice_headers_size = MAX_SLICEHEADER_BUFFER_SIZE;
#if 0
VTRACE("flags.bits.frame_mbs_only_flag = %d", data->pic_parse_buffer->flags.bits.frame_mbs_only_flag);
VTRACE("flags.bits.pic_order_present_flag = %d", data->pic_parse_buffer->flags.bits.pic_order_present_flag);
VTRACE("flags.bits.delta_pic_order_always_zero_flag = %d", data->pic_parse_buffer->flags.bits.delta_pic_order_always_zero_flag);
VTRACE("flags.bits.redundant_pic_cnt_present_flag = %d", data->pic_parse_buffer->flags.bits.redundant_pic_cnt_present_flag);
VTRACE("flags.bits.weighted_pred_flag = %d", data->pic_parse_buffer->flags.bits.weighted_pred_flag);
VTRACE("flags.bits.entropy_coding_mode_flag = %d", data->pic_parse_buffer->flags.bits.entropy_coding_mode_flag);
VTRACE("flags.bits.deblocking_filter_control_present_flag = %d", data->pic_parse_buffer->flags.bits.deblocking_filter_control_present_flag);
VTRACE("flags.bits.weighted_bipred_idc = %d", data->pic_parse_buffer->flags.bits.weighted_bipred_idc);
VTRACE("pic_parse_buffer->expected_pic_parameter_set_id = %d", data->pic_parse_buffer->expected_pic_parameter_set_id);
VTRACE("pic_parse_buffer->num_slice_groups_minus1 = %d", data->pic_parse_buffer->num_slice_groups_minus1);
VTRACE("pic_parse_buffer->chroma_format_idc = %d", data->pic_parse_buffer->chroma_format_idc);
VTRACE("pic_parse_buffer->log2_max_pic_order_cnt_lsb_minus4 = %d", data->pic_parse_buffer->log2_max_pic_order_cnt_lsb_minus4);
VTRACE("pic_parse_buffer->pic_order_cnt_type = %d", data->pic_parse_buffer->pic_order_cnt_type);
VTRACE("pic_parse_buffer->residual_colour_transform_flag = %d", data->pic_parse_buffer->residual_colour_transform_flag);
VTRACE("pic_parse_buffer->num_ref_idc_l0_active_minus1 = %d", data->pic_parse_buffer->num_ref_idc_l0_active_minus1);
VTRACE("pic_parse_buffer->num_ref_idc_l1_active_minus1 = %d", data->pic_parse_buffer->num_ref_idc_l1_active_minus1);
#endif
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAParsePictureParameterBufferType,
sizeof(VAParsePictureParameterBuffer),
1,
data->pic_parse_buffer,
&pictureparameterparsingbufferID);
CHECK_VA_STATUS("vaCreatePictureParameterParsingBuffer");
vaStatus = vaRenderPicture(
mVADisplay,
mVAContext,
&pictureparameterparsingbufferID,
1);
CHECK_VA_STATUS("vaRenderPicture");
vaStatus = vaMapBuffer(
mVADisplay,
sliceheaderbufferID,
&sliceheaderbuf);
CHECK_VA_STATUS("vaMapBuffer");
status = updateSliceParameter(data,sliceheaderbuf);
CHECK_STATUS("processSliceHeader");
vaStatus = vaUnmapBuffer(
mVADisplay,
sliceheaderbufferID);
CHECK_VA_STATUS("vaUnmapBuffer");
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::parseModularSliceHeader(VideoDecodeBuffer *buffer, vbp_data_h264 *data) {
Decode_Status status;
VAStatus vaStatus;
VABufferID sliceheaderbufferID;
VABufferID pictureparameterparsingbufferID;
VABufferID mSlicebufferID;
int32_t sliceIdx;
vaStatus = vaBeginPicture(mVADisplay, mVAContext, mAcquiredBuffer->renderBuffer.surface);
CHECK_VA_STATUS("vaBeginPicture");
if (mFrameSize <= 0 || mSliceNum <=0) {
return DECODE_SUCCESS;
}
void *sliceheaderbuf;
memset(mCachedHeader, 0, MAX_SLICEHEADER_BUFFER_SIZE);
int32_t offset = 0;
int32_t size = 0;
for (sliceIdx = 0; sliceIdx < mSliceNum; sliceIdx++) {
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAParseSliceHeaderGroupBufferType,
MAX_SLICEHEADER_BUFFER_SIZE,
1,
NULL,
&sliceheaderbufferID);
CHECK_VA_STATUS("vaCreateSliceHeaderGroupBuffer");
vaStatus = vaMapBuffer(
mVADisplay,
sliceheaderbufferID,
&sliceheaderbuf);
CHECK_VA_STATUS("vaMapBuffer");
memset(sliceheaderbuf, 0, MAX_SLICEHEADER_BUFFER_SIZE);
vaStatus = vaUnmapBuffer(
mVADisplay,
sliceheaderbufferID);
CHECK_VA_STATUS("vaUnmapBuffer");
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VASliceDataBufferType,
mSliceInfo[sliceIdx].sliceSize, //size
1, //num_elements
mFrameData + mSliceInfo[sliceIdx].sliceStartOffset,
&mSlicebufferID);
CHECK_VA_STATUS("vaCreateSliceDataBuffer");
data->pic_parse_buffer->frame_buf_id = mSlicebufferID;
data->pic_parse_buffer->slice_headers_buf_id = sliceheaderbufferID;
data->pic_parse_buffer->frame_size = mSliceInfo[sliceIdx].sliceLength;
data->pic_parse_buffer->slice_headers_size = MAX_SLICEHEADER_BUFFER_SIZE;
data->pic_parse_buffer->nalu_header.value = mSliceInfo[sliceIdx].sliceHeaderByte;
data->pic_parse_buffer->slice_offset = mSliceInfo[sliceIdx].sliceByteOffset;
#if 0
VTRACE("data->pic_parse_buffer->slice_offset = 0x%x", data->pic_parse_buffer->slice_offset);
VTRACE("pic_parse_buffer->nalu_header.value = %x", data->pic_parse_buffer->nalu_header.value = mSliceInfo[sliceIdx].sliceHeaderByte);
VTRACE("flags.bits.frame_mbs_only_flag = %d", data->pic_parse_buffer->flags.bits.frame_mbs_only_flag);
VTRACE("flags.bits.pic_order_present_flag = %d", data->pic_parse_buffer->flags.bits.pic_order_present_flag);
VTRACE("flags.bits.delta_pic_order_always_zero_flag = %d", data->pic_parse_buffer->flags.bits.delta_pic_order_always_zero_flag);
VTRACE("flags.bits.redundant_pic_cnt_present_flag = %d", data->pic_parse_buffer->flags.bits.redundant_pic_cnt_present_flag);
VTRACE("flags.bits.weighted_pred_flag = %d", data->pic_parse_buffer->flags.bits.weighted_pred_flag);
VTRACE("flags.bits.entropy_coding_mode_flag = %d", data->pic_parse_buffer->flags.bits.entropy_coding_mode_flag);
VTRACE("flags.bits.deblocking_filter_control_present_flag = %d", data->pic_parse_buffer->flags.bits.deblocking_filter_control_present_flag);
VTRACE("flags.bits.weighted_bipred_idc = %d", data->pic_parse_buffer->flags.bits.weighted_bipred_idc);
VTRACE("pic_parse_buffer->expected_pic_parameter_set_id = %d", data->pic_parse_buffer->expected_pic_parameter_set_id);
VTRACE("pic_parse_buffer->num_slice_groups_minus1 = %d", data->pic_parse_buffer->num_slice_groups_minus1);
VTRACE("pic_parse_buffer->chroma_format_idc = %d", data->pic_parse_buffer->chroma_format_idc);
VTRACE("pic_parse_buffer->log2_max_pic_order_cnt_lsb_minus4 = %d", data->pic_parse_buffer->log2_max_pic_order_cnt_lsb_minus4);
VTRACE("pic_parse_buffer->pic_order_cnt_type = %d", data->pic_parse_buffer->pic_order_cnt_type);
VTRACE("pic_parse_buffer->residual_colour_transform_flag = %d", data->pic_parse_buffer->residual_colour_transform_flag);
VTRACE("pic_parse_buffer->num_ref_idc_l0_active_minus1 = %d", data->pic_parse_buffer->num_ref_idc_l0_active_minus1);
VTRACE("pic_parse_buffer->num_ref_idc_l1_active_minus1 = %d", data->pic_parse_buffer->num_ref_idc_l1_active_minus1);
#endif
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAParsePictureParameterBufferType,
sizeof(VAParsePictureParameterBuffer),
1,
data->pic_parse_buffer,
&pictureparameterparsingbufferID);
CHECK_VA_STATUS("vaCreatePictureParameterParsingBuffer");
vaStatus = vaRenderPicture(
mVADisplay,
mVAContext,
&pictureparameterparsingbufferID,
1);
CHECK_VA_STATUS("vaRenderPicture");
vaStatus = vaMapBuffer(
mVADisplay,
sliceheaderbufferID,
&sliceheaderbuf);
CHECK_VA_STATUS("vaMapBuffer");
size = *(uint32 *)((uint8 *)sliceheaderbuf + 4) + 4;
VTRACE("slice header size = 0x%x, offset = 0x%x", size, offset);
if (offset + size <= MAX_SLICEHEADER_BUFFER_SIZE - 4) {
memcpy(mCachedHeader+offset, sliceheaderbuf, size);
offset += size;
} else {
WTRACE("Cached slice header is not big enough!");
}
vaStatus = vaUnmapBuffer(
mVADisplay,
sliceheaderbufferID);
CHECK_VA_STATUS("vaUnmapBuffer");
}
memset(mCachedHeader + offset, 0xFF, 4);
status = updateSliceParameter(data,mCachedHeader);
CHECK_STATUS("processSliceHeader");
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::updateSliceParameter(vbp_data_h264 *data, void *sliceheaderbuf) {
VTRACE("VideoDecoderAVCSecure::updateSliceParameter");
Decode_Status status;
status = VideoDecoderBase::updateBuffer(
(uint8_t *)sliceheaderbuf,
MAX_SLICEHEADER_BUFFER_SIZE,
(void**)&data);
CHECK_STATUS("updateBuffer");
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::decodeSlice(vbp_data_h264 *data, uint32_t picIndex, uint32_t sliceIndex) {
Decode_Status status;
VAStatus vaStatus;
uint32_t bufferIDCount = 0;
// maximum 3 buffers to render a slice: picture parameter, IQMatrix, slice parameter
VABufferID bufferIDs[3];
vbp_picture_data_h264 *picData = &(data->pic_data[picIndex]);
vbp_slice_data_h264 *sliceData = &(picData->slc_data[sliceIndex]);
VAPictureParameterBufferH264 *picParam = picData->pic_parms;
VASliceParameterBufferH264 *sliceParam = &(sliceData->slc_parms);
uint32_t slice_data_size = 0;
uint8_t* slice_data_addr = NULL;
if (sliceParam->first_mb_in_slice == 0 || mDecodingFrame == false) {
// either condition indicates start of a new frame
if (sliceParam->first_mb_in_slice != 0) {
WTRACE("The first slice is lost.");
}
VTRACE("Current frameidx = %d", mFrameIdx++);
// Update the reference frames and surface IDs for DPB and current frame
status = updateDPB(picParam);
CHECK_STATUS("updateDPB");
//We have to provide a hacked DPB rather than complete DPB for libva as workaround
status = updateReferenceFrames(picData);
CHECK_STATUS("updateReferenceFrames");
mDecodingFrame = true;
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAPictureParameterBufferType,
sizeof(VAPictureParameterBufferH264),
1,
picParam,
&bufferIDs[bufferIDCount]);
CHECK_VA_STATUS("vaCreatePictureParameterBuffer");
bufferIDCount++;
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VAIQMatrixBufferType,
sizeof(VAIQMatrixBufferH264),
1,
data->IQ_matrix_buf,
&bufferIDs[bufferIDCount]);
CHECK_VA_STATUS("vaCreateIQMatrixBuffer");
bufferIDCount++;
}
status = setReference(sliceParam);
CHECK_STATUS("setReference");
if (mModularMode) {
if (mIsEncryptData) {
sliceParam->slice_data_size = mSliceInfo[sliceIndex].sliceSize;
slice_data_size = mSliceInfo[sliceIndex].sliceSize;
slice_data_addr = mFrameData + mSliceInfo[sliceIndex].sliceStartOffset;
} else {
slice_data_size = sliceData->slice_size;
slice_data_addr = sliceData->buffer_addr + sliceData->slice_offset;
}
} else {
sliceParam->slice_data_size = mFrameSize;
slice_data_size = mFrameSize;
slice_data_addr = mFrameData;
}
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VASliceParameterBufferType,
sizeof(VASliceParameterBufferH264),
1,
sliceParam,
&bufferIDs[bufferIDCount]);
CHECK_VA_STATUS("vaCreateSliceParameterBuffer");
bufferIDCount++;
vaStatus = vaRenderPicture(
mVADisplay,
mVAContext,
bufferIDs,
bufferIDCount);
CHECK_VA_STATUS("vaRenderPicture");
VABufferID slicebufferID;
vaStatus = vaCreateBuffer(
mVADisplay,
mVAContext,
VASliceDataBufferType,
slice_data_size, //size
1, //num_elements
slice_data_addr,
&slicebufferID);
CHECK_VA_STATUS("vaCreateSliceDataBuffer");
vaStatus = vaRenderPicture(
mVADisplay,
mVAContext,
&slicebufferID,
1);
CHECK_VA_STATUS("vaRenderPicture");
return DECODE_SUCCESS;
}
Decode_Status VideoDecoderAVCSecure::getCodecSpecificConfigs(
VAProfile profile, VAConfigID *config)
{
VAStatus vaStatus;
VAConfigAttrib attrib[2];
if (config == NULL) {
ETRACE("Invalid parameter!");
return DECODE_FAIL;
}
attrib[0].type = VAConfigAttribRTFormat;
attrib[0].value = VA_RT_FORMAT_YUV420;
attrib[1].type = VAConfigAttribDecSliceMode;
attrib[1].value = VA_DEC_SLICE_MODE_NORMAL;
if (mModularMode) {
attrib[1].value = VA_DEC_SLICE_MODE_SUBSAMPLE;
}
vaStatus = vaCreateConfig(
mVADisplay,
profile,
VAEntrypointVLD,
&attrib[0],
2,
config);
CHECK_VA_STATUS("vaCreateConfig");
return DECODE_SUCCESS;
}