Archived
1
0
This repository has been archived on 2024-10-17. You can view files and clone it, but cannot push or open issues or pull requests.
winamp/Src/h264dec/ldecod/src/image.c
2024-09-24 14:54:57 +02:00

1700 lines
57 KiB
C

/*!
***********************************************************************
* \file image.c
*
* \brief
* Decode a Slice
*
* \author
* Main contributors (see contributors.h for copyright, address and affiliation details)
* - Inge Lille-Langoy <inge.lille-langoy@telenor.com>
* - Rickard Sjoberg <rickard.sjoberg@era.ericsson.se>
* - Jani Lainema <jani.lainema@nokia.com>
* - Sebastian Purreiter <sebastian.purreiter@mch.siemens.de>
* - Byeong-Moon Jeon <jeonbm@lge.com>
* - Thomas Wedi <wedi@tnt.uni-hannover.de>
* - Gabi Blaettermann
* - Ye-Kui Wang <wyk@ieee.org>
* - Antti Hallapuro <antti.hallapuro@nokia.com>
* - Alexis Tourapis <alexismt@ieee.org>
* - Jill Boyce <jill.boyce@thomson.net>
* - Saurav K Bandyopadhyay <saurav@ieee.org>
* - Zhenyu Wu <Zhenyu.Wu@thomson.net
* - Purvin Pandit <Purvin.Pandit@thomson.net>
*
***********************************************************************
*/
#include "contributors.h"
#include <math.h>
#include <limits.h>
#include "global.h"
#include "image.h"
#include "fmo.h"
#include "nalu.h"
#include "parset.h"
#include "header.h"
#include "sei.h"
#include "output.h"
#include "mb_access.h"
#include "memalloc.h"
#include "macroblock.h"
#include "loopfilter.h"
#include "biaridecod.h"
#include "context_ini.h"
#include "cabac.h"
#include "vlc.h"
#include "quant.h"
#include "errorconcealment.h"
#include "erc_api.h"
/*!
************************************************************************
* \brief
* checks if the System is big- or little-endian
* \return
* 0, little-endian (e.g. Intel architectures)
* 1, big-endian (e.g. SPARC, MIPS, PowerPC)
************************************************************************
*/
int testEndian(void)
{
short s;
byte *p;
p=(byte*)&s;
s=1;
return (*p==0);
}
static int read_new_slice(Slice *currSlice, uint64_t time_code);
/*!
************************************************************************
* \brief
* Initializes the parameters for a new picture
************************************************************************
*/
// benski> time_code is user-passed value
static void init_picture(VideoParameters *p_Vid, Slice *currSlice, InputParameters *p_Inp, uint64_t time_code)
{
int i;
int nplane;
if (p_Vid->dec_picture)
{
// this may only happen on slice loss
exit_picture(p_Vid, &p_Vid->dec_picture);
}
if (p_Vid->recovery_point)
p_Vid->recovery_frame_num = (p_Vid->frame_num + p_Vid->recovery_frame_cnt) % p_Vid->MaxFrameNum;
if (p_Vid->idr_flag)
p_Vid->recovery_frame_num = p_Vid->frame_num;
if (p_Vid->recovery_point == 0 &&
p_Vid->pre_frame_num != INT_MIN &&
p_Vid->frame_num != p_Vid->pre_frame_num &&
p_Vid->frame_num != (p_Vid->pre_frame_num + 1) % p_Vid->MaxFrameNum)
{
if (p_Vid->active_sps->gaps_in_frame_num_value_allowed_flag == 0)
{
/* Advanced Error Concealment would be called here to combat unintentional loss of pictures. */
error("An unintentional loss of pictures occurs! Exit\n", 100);
}
fill_frame_num_gap(p_Vid);
}
if(p_Vid->nal_reference_idc)
{
p_Vid->pre_frame_num = p_Vid->frame_num;
}
p_Vid->num_dec_mb = 0;
//calculate POC
decode_poc(p_Vid);
if (p_Vid->recovery_frame_num == p_Vid->frame_num &&
p_Vid->recovery_poc == 0x7fffffff)
p_Vid->recovery_poc = p_Vid->framepoc;
if(p_Vid->nal_reference_idc)
p_Vid->last_ref_pic_poc = p_Vid->framepoc;
// dumppoc (p_Vid);
p_Vid->dec_picture = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr);
p_Vid->dec_picture->time_code = time_code;
p_Vid->dec_picture->top_poc=p_Vid->toppoc;
p_Vid->dec_picture->bottom_poc=p_Vid->bottompoc;
p_Vid->dec_picture->frame_poc=p_Vid->framepoc;
p_Vid->dec_picture->qp = p_Vid->qp;
p_Vid->dec_picture->slice_qp_delta = currSlice->slice_qp_delta;
p_Vid->dec_picture->chroma_qp_offset[0] = p_Vid->active_pps->chroma_qp_index_offset;
p_Vid->dec_picture->chroma_qp_offset[1] = p_Vid->active_pps->second_chroma_qp_index_offset;
// reset all variables of the error concealment instance before decoding of every frame.
// here the third parameter should, if perfectly, be equal to the number of slices per frame.
// using little value is ok, the code will allocate more memory if the slice number is larger
ercReset(p_Vid->erc_errorVar, p_Vid->PicSizeInMbs, p_Vid->PicSizeInMbs, p_Vid->dec_picture->size_x);
p_Vid->erc_mvperMB = 0;
switch (p_Vid->structure )
{
case TOP_FIELD:
{
p_Vid->dec_picture->poc=p_Vid->toppoc;
p_Vid->number *= 2;
break;
}
case BOTTOM_FIELD:
{
p_Vid->dec_picture->poc=p_Vid->bottompoc;
p_Vid->number = p_Vid->number * 2 + 1;
break;
}
case FRAME:
{
p_Vid->dec_picture->poc=p_Vid->framepoc;
break;
}
default:
error("p_Vid->structure not initialized", 235);
}
p_Vid->current_slice_nr=0;
if (p_Vid->type > SI_SLICE)
{
set_ec_flag(p_Vid, SE_PTYPE);
p_Vid->type = P_SLICE; // concealed element
}
// CAVLC init
if (p_Vid->active_pps->entropy_coding_mode_flag == CAVLC)
{
memset(p_Vid->nz_coeff[0], -1, p_Vid->PicSizeInMbs * 48 *sizeof(byte)); // 3 * 4 * 4
}
if(p_Vid->active_pps->constrained_intra_pred_flag)
{
for (i=0; i<(int)p_Vid->PicSizeInMbs; ++i)
{
p_Vid->intra_block[i] = 1;
}
}
// Set the slice_nr member of each MB to -1, to ensure correct when packet loss occurs
// TO set Macroblock Map (mark all MBs as 'have to be concealed')
if( IS_INDEPENDENT(p_Vid) )
{
for( nplane=0; nplane<MAX_PLANE; ++nplane )
{
for(i=0; i<(int)p_Vid->PicSizeInMbs; ++i)
{
p_Vid->mb_data_JV[nplane][i].slice_nr = -1;
p_Vid->mb_data_JV[nplane][i].ei_flag = 1;
p_Vid->mb_data_JV[nplane][i].dpl_flag = 0;
}
}
}
else
{
for(i=0; i<(int)p_Vid->PicSizeInMbs; ++i)
{
p_Vid->mb_data[i].slice_nr = -1;
p_Vid->mb_data[i].ei_flag = 1;
p_Vid->mb_data[i].dpl_flag = 0;
}
}
p_Vid->dec_picture->slice_type = p_Vid->type;
p_Vid->dec_picture->used_for_reference = (p_Vid->nal_reference_idc != 0);
p_Vid->dec_picture->idr_flag = p_Vid->idr_flag;
p_Vid->dec_picture->no_output_of_prior_pics_flag = p_Vid->no_output_of_prior_pics_flag;
p_Vid->dec_picture->long_term_reference_flag = p_Vid->long_term_reference_flag;
p_Vid->dec_picture->adaptive_ref_pic_buffering_flag = p_Vid->adaptive_ref_pic_buffering_flag;
p_Vid->dec_picture->dec_ref_pic_marking_buffer = p_Vid->dec_ref_pic_marking_buffer;
p_Vid->dec_ref_pic_marking_buffer = NULL;
p_Vid->dec_picture->mb_aff_frame_flag = p_Vid->mb_aff_frame_flag;
p_Vid->dec_picture->PicWidthInMbs = p_Vid->PicWidthInMbs;
if (p_Vid->dec_picture->mb_aff_frame_flag)
{
p_Vid->get_mb_block_pos = get_mb_block_pos_mbaff;
p_Vid->getNeighbour = getAffNeighbour;
p_Vid->getNeighbourXP_NoPos = getAffNeighbour;
p_Vid->getNeighbourPX_NoPos = getAffNeighbour;
p_Vid->getNeighbourLuma = getAffNeighbourLuma;
p_Vid->getNeighbourPXLuma = getAffNeighbourLuma;
p_Vid->getNeighbourXPLuma = getAffNeighbourXPLuma;
p_Vid->getNeighbourLeftLuma = getAffNeighbourN0Luma;
p_Vid->getNeighbourNXLuma = getAffNeighbourNXLuma;
p_Vid->getNeighbourLeft = getAffNeighbourN0;
p_Vid->getNeighbourUp = getAffNeighbour0N;
p_Vid->getNeighbourUpLuma = getAffNeighbour0NLuma;
p_Vid->getNeighbourNX = getAffNeighbourNX;
p_Vid->getNeighbourNP = getAffNeighbourNX;
p_Vid->getNeighbourNPChromaNB = getAffNeighbourNX;
p_Vid->getNeighbour0X = getAffNeighbour0X;
p_Vid->getNeighbour0XLuma = getAffNeighbour0XLuma;
p_Vid->getNeighbourX0 = getAffNeighbourX0;
p_Vid->getNeighbourNPLumaNB = getAffNeighbourNPLuma;
p_Vid->getNeighbourPXLumaNB = getAffNeighbourPXLumaNB;
p_Vid->getNeighbourPXLumaNB_NoPos = getAffNeighbourPXLumaNB_NoPos;
p_Vid->getNeighbourXPLumaNB = getAffNeighbourXPLuma;
p_Vid->getNeighbourPPLumaNB = getAffNeighbourPPLumaNB;
p_Vid->getNeighbourXPLumaNB_NoPos = getAffNeighbourXPLuma;
}
else
{
p_Vid->get_mb_block_pos = get_mb_block_pos_normal;
p_Vid->getNeighbour = getNonAffNeighbour;
p_Vid->getNeighbourXP_NoPos = getNonAffNeighbourXP_NoPos;
p_Vid->getNeighbourPX_NoPos = getNonAffNeighbourPX_NoPos;
p_Vid->getNeighbourLuma = getNonAffNeighbourLuma;
p_Vid->getNeighbourPXLuma = getNonAffNeighbourPXLuma;
p_Vid->getNeighbourXPLuma = getNonAffNeighbourXPLuma;
p_Vid->getNeighbourLeftLuma = getNonAffNeighbourN0Luma;
p_Vid->getNeighbourNXLuma = getNonAffNeighbourNXLuma;
p_Vid->getNeighbourLeft = getNonAffNeighbourN0;
p_Vid->getNeighbourUp = getNonAffNeighbour0N;
p_Vid->getNeighbourUpLuma = getNonAffNeighbour0NLuma;
p_Vid->getNeighbourNX = getNonAffNeighbourNX;
p_Vid->getNeighbourNP = getNonAffNeighbourNP;
p_Vid->getNeighbourNPChromaNB = getNonAffNeighbourNPChromaNB;
p_Vid->getNeighbour0X = getNonAffNeighbour0X;
p_Vid->getNeighbour0XLuma = getNonAffNeighbour0XLuma;
p_Vid->getNeighbourX0 = getNonAffNeighbourX0;
p_Vid->getNeighbourNPLumaNB = getNonAffNeighbourNPLumaNB;
p_Vid->getNeighbourPXLumaNB = getNonAffNeighbourPXLumaNB;
p_Vid->getNeighbourPXLumaNB_NoPos = getNonAffNeighbourPXLumaNB_NoPos;
p_Vid->getNeighbourXPLumaNB = getNonAffNeighbourXPLumaNB;
p_Vid->getNeighbourPPLumaNB = getNonAffNeighbourPPLumaNB;
p_Vid->getNeighbourXPLumaNB_NoPos = getNonAffNeighbourXPLumaNB_NoPos;
}
p_Vid->dec_picture->pic_num = p_Vid->frame_num;
p_Vid->dec_picture->frame_num = p_Vid->frame_num;
p_Vid->dec_picture->recovery_frame = (unsigned int) (p_Vid->frame_num == p_Vid->recovery_frame_num);
p_Vid->dec_picture->coded_frame = (p_Vid->structure==FRAME);
p_Vid->dec_picture->chroma_format_idc = p_Vid->active_sps->chroma_format_idc;
p_Vid->dec_picture->frame_mbs_only_flag = p_Vid->active_sps->frame_mbs_only_flag;
p_Vid->dec_picture->frame_cropping_flag = p_Vid->active_sps->frame_cropping_flag;
if (p_Vid->dec_picture->frame_cropping_flag)
{
p_Vid->dec_picture->frame_cropping_rect_left_offset = p_Vid->active_sps->frame_cropping_rect_left_offset;
p_Vid->dec_picture->frame_cropping_rect_right_offset = p_Vid->active_sps->frame_cropping_rect_right_offset;
p_Vid->dec_picture->frame_cropping_rect_top_offset = p_Vid->active_sps->frame_cropping_rect_top_offset;
p_Vid->dec_picture->frame_cropping_rect_bottom_offset = p_Vid->active_sps->frame_cropping_rect_bottom_offset;
}
#if (ENABLE_OUTPUT_TONEMAPPING)
// store the necessary tone mapping sei into StorablePicture structure
p_Vid->dec_picture->seiHasTone_mapping = 0;
if (p_Vid->seiToneMapping->seiHasTone_mapping)
{
p_Vid->dec_picture->seiHasTone_mapping = 1;
p_Vid->dec_picture->tone_mapping_model_id = p_Vid->seiToneMapping->model_id;
p_Vid->dec_picture->tonemapped_bit_depth = p_Vid->seiToneMapping->sei_bit_depth;
p_Vid->dec_picture->tone_mapping_lut = malloc(sizeof(int)*(1<<p_Vid->seiToneMapping->coded_data_bit_depth));
if (NULL == p_Vid->dec_picture->tone_mapping_lut)
{
no_mem_exit("init_picture: tone_mapping_lut");
}
memcpy(p_Vid->dec_picture->tone_mapping_lut, p_Vid->seiToneMapping->lut, sizeof(imgpel)*(1<<p_Vid->seiToneMapping->coded_data_bit_depth));
update_tone_mapping_sei(p_Vid->seiToneMapping);
}
#endif
if( IS_INDEPENDENT(p_Vid) )
{
p_Vid->dec_picture_JV[0] = p_Vid->dec_picture;
p_Vid->dec_picture_JV[1] = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr);
copy_dec_picture_JV( p_Vid, p_Vid->dec_picture_JV[1], p_Vid->dec_picture_JV[0] );
p_Vid->dec_picture_JV[2] = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr);
copy_dec_picture_JV( p_Vid, p_Vid->dec_picture_JV[2], p_Vid->dec_picture_JV[0] );
}
}
void MbAffPostProc(VideoParameters *p_Vid)
{
imgpel temp[32][16];
StorablePicture *dec_picture = p_Vid->dec_picture;
imgpel ** imgY = dec_picture->imgY->img;
imgpel **imgUV[2] = {dec_picture->imgUV[0]->img, dec_picture->imgUV[1]->img};
short i, y, x0, y0, uv;
for (i=0; i<(int)dec_picture->PicSizeInMbs; i+=2)
{
if (dec_picture->motion.mb_field[i])
{
get_mb_pos(p_Vid, i, p_Vid->mb_size[IS_LUMA], &x0, &y0);
for (y=0; y<(2*MB_BLOCK_SIZE);++y)
memcpy(temp[y], &imgY[y0+y][x0], MB_BLOCK_SIZE * sizeof(imgpel));
for (y=0; y<MB_BLOCK_SIZE; ++y)
{
memcpy(&imgY[y0+(2*y )][x0], temp[y ], MB_BLOCK_SIZE * sizeof(imgpel));
memcpy(&imgY[y0+(2*y+1)][x0], temp[y+MB_BLOCK_SIZE], MB_BLOCK_SIZE * sizeof(imgpel));
}
if (dec_picture->chroma_format_idc != YUV400)
{
x0 = (short) (x0 / (16/p_Vid->mb_cr_size_x));
y0 = (short) (y0 / (16/p_Vid->mb_cr_size_y));
for (uv=0; uv<2; ++uv)
{
for (y=0; y<(2*p_Vid->mb_cr_size_y);++y)
memcpy(temp[y], &imgUV[uv][y0+y][x0], p_Vid->mb_cr_size_x * sizeof(imgpel));
for (y=0; y<p_Vid->mb_cr_size_y;++y)
{
memcpy(&imgUV[uv][y0+(2*y )][x0], temp[y ], p_Vid->mb_cr_size_x * sizeof(imgpel));
memcpy(&imgUV[uv][y0+(2*y+1)][x0], temp[y+p_Vid->mb_cr_size_y], p_Vid->mb_cr_size_x * sizeof(imgpel));
}
}
}
}
}
}
static void fill_wp_params(Slice *currSlice)
{
VideoParameters *p_Vid = currSlice->p_Vid;
int i, j, k;
int comp;
int log_weight_denom;
int tb, td;
int tx,DistScaleFactor;
int max_l0_ref = currSlice->num_ref_idx_l0_active;
int max_l1_ref = currSlice->num_ref_idx_l1_active;
if (p_Vid->active_pps->weighted_bipred_idc == 2)
{
currSlice->luma_log2_weight_denom = 5;
currSlice->chroma_log2_weight_denom = 5;
currSlice->wp_round_luma = 16;
currSlice->wp_round_chroma = 16;
for (i=0; i<MAX_REFERENCE_PICTURES; ++i)
{
for (comp=0; comp<3; ++comp)
{
log_weight_denom = (comp == 0) ? currSlice->luma_log2_weight_denom : currSlice->chroma_log2_weight_denom;
currSlice->wp_weight[0][i][comp] = 1<<log_weight_denom;
currSlice->wp_weight[1][i][comp] = 1<<log_weight_denom;
currSlice->wp_offset[0][i][comp] = 0;
currSlice->wp_offset[1][i][comp] = 0;
}
}
}
for (i=0; i<max_l0_ref; ++i)
{
for (j=0; j<max_l1_ref; ++j)
{
for (comp = 0; comp<3; ++comp)
{
log_weight_denom = (comp == 0) ? currSlice->luma_log2_weight_denom : currSlice->chroma_log2_weight_denom;
if (p_Vid->active_pps->weighted_bipred_idc == 1)
{
currSlice->wbp_weight[0][i][j][comp] = currSlice->wp_weight[0][i][comp];
currSlice->wbp_weight[1][i][j][comp] = currSlice->wp_weight[1][j][comp];
}
else if (p_Vid->active_pps->weighted_bipred_idc == 2)
{
td = iClip3(-128,127,p_Vid->listX[LIST_1][j]->poc - p_Vid->listX[LIST_0][i]->poc);
if (td == 0 || p_Vid->listX[LIST_1][j]->is_long_term || p_Vid->listX[LIST_0][i]->is_long_term)
{
currSlice->wbp_weight[0][i][j][comp] = 32;
currSlice->wbp_weight[1][i][j][comp] = 32;
}
else
{
tb = iClip3(-128,127,p_Vid->ThisPOC - p_Vid->listX[LIST_0][i]->poc);
tx = (16384 + iabs(td/2))/td;
DistScaleFactor = iClip3(-1024, 1023, (tx*tb + 32 )>>6);
currSlice->wbp_weight[1][i][j][comp] = DistScaleFactor >> 2;
currSlice->wbp_weight[0][i][j][comp] = 64 - currSlice->wbp_weight[1][i][j][comp];
if (currSlice->wbp_weight[1][i][j][comp] < -64 || currSlice->wbp_weight[1][i][j][comp] > 128)
{
currSlice->wbp_weight[0][i][j][comp] = 32;
currSlice->wbp_weight[1][i][j][comp] = 32;
currSlice->wp_offset[0][i][comp] = 0;
currSlice->wp_offset[1][j][comp] = 0;
}
}
}
}
}
}
if (currSlice->mb_aff_frame_flag)
{
for (i=0; i<2*max_l0_ref; ++i)
{
for (j=0; j<2*max_l1_ref; ++j)
{
for (comp = 0; comp<3; ++comp)
{
for (k=2; k<6; k+=2)
{
currSlice->wp_offset[k+0][i][comp] = currSlice->wp_offset[0][i>>1][comp];
currSlice->wp_offset[k+1][j][comp] = currSlice->wp_offset[1][j>>1][comp];
log_weight_denom = (comp == 0) ? currSlice->luma_log2_weight_denom : currSlice->chroma_log2_weight_denom;
if (p_Vid->active_pps->weighted_bipred_idc == 1)
{
currSlice->wbp_weight[k+0][i][j][comp] = currSlice->wp_weight[0][i>>1][comp];
currSlice->wbp_weight[k+1][i][j][comp] = currSlice->wp_weight[1][j>>1][comp];
}
else if (p_Vid->active_pps->weighted_bipred_idc == 2)
{
td = iClip3(-128,127,p_Vid->listX[k+LIST_1][j]->poc - p_Vid->listX[k+LIST_0][i]->poc);
if (td == 0 || p_Vid->listX[k+LIST_1][j]->is_long_term || p_Vid->listX[k+LIST_0][i]->is_long_term)
{
currSlice->wbp_weight[k+0][i][j][comp] = 32;
currSlice->wbp_weight[k+1][i][j][comp] = 32;
}
else
{
tb = iClip3(-128,127,((k==2)?p_Vid->toppoc:p_Vid->bottompoc) - p_Vid->listX[k+LIST_0][i]->poc);
tx = (16384 + iabs(td/2))/td;
DistScaleFactor = iClip3(-1024, 1023, (tx*tb + 32 )>>6);
currSlice->wbp_weight[k+1][i][j][comp] = DistScaleFactor >> 2;
currSlice->wbp_weight[k+0][i][j][comp] = 64 - currSlice->wbp_weight[k+1][i][j][comp];
if (currSlice->wbp_weight[k+1][i][j][comp] < -64 || currSlice->wbp_weight[k+1][i][j][comp] > 128)
{
currSlice->wbp_weight[k+1][i][j][comp] = 32;
currSlice->wbp_weight[k+0][i][j][comp] = 32;
currSlice->wp_offset[k+0][i][comp] = 0;
currSlice->wp_offset[k+1][j][comp] = 0;
}
}
}
}
}
}
}
}
}
static void decode_slice(Slice *currSlice, int current_header)
{
VideoParameters *p_Vid = currSlice->p_Vid;
if (p_Vid->active_pps->entropy_coding_mode_flag)
{
init_contexts (currSlice);
cabac_new_slice(currSlice);
}
if ( (p_Vid->active_pps->weighted_bipred_idc > 0 && (currSlice->slice_type == B_SLICE)) || (p_Vid->active_pps->weighted_pred_flag && currSlice->slice_type !=I_SLICE))
fill_wp_params(currSlice);
//printf("frame picture %d %d %d\n",p_Vid->structure,p_Vid->ThisPOC,currSlice->direct_spatial_mv_pred_flag);
// decode main slice information
if (current_header == SOP || current_header == SOS)
decode_one_slice(currSlice);
// setMB-Nr in case this slice was lost
// if(currSlice->ei_flag)
// p_Vid->current_mb_nr = currSlice->last_mb_nr + 1;
}
/*!
************************************************************************
* \brief
* Error tracking: if current frame is lost or any reference frame of
* current frame is lost, current frame is incorrect.
************************************************************************
*/
static void Error_tracking(VideoParameters *p_Vid)
{
Slice *currSlice = p_Vid->currentSlice;
int i;
if(p_Vid->redundant_pic_cnt == 0)
{
p_Vid->Is_primary_correct = p_Vid->Is_redundant_correct = 1;
}
if(p_Vid->redundant_pic_cnt == 0 && p_Vid->type != I_SLICE)
{
for(i=0;i<currSlice->num_ref_idx_l0_active;++i)
{
if(p_Vid->ref_flag[i] == 0) // any reference of primary slice is incorrect
{
p_Vid->Is_primary_correct = 0; // primary slice is incorrect
}
}
}
else if(p_Vid->redundant_pic_cnt != 0 && p_Vid->type != I_SLICE)
{
if(p_Vid->ref_flag[p_Vid->redundant_slice_ref_idx] == 0) // reference of redundant slice is incorrect
{
p_Vid->Is_redundant_correct = 0; // redundant slice is incorrect
}
}
}
/*!
***********************************************************************
* \brief
* decodes one I- or P-frame
*
***********************************************************************
*/
int decode_one_frame(VideoParameters *p_Vid, uint64_t time_code)
{
InputParameters *p_Inp = p_Vid->p_Inp;
int current_header;
int i;
while (1)
{
Slice *currSlice = p_Vid->currentSlice;
currSlice->p_Vid = p_Vid;
currSlice->p_Inp = p_Inp;
current_header = read_new_slice(p_Vid->currentSlice, time_code);
if (current_header == EOS)
return EOS;
// error tracking of primary and redundant slices.
Error_tracking(p_Vid);
// If primary and redundant are received and primary is correct, discard the redundant
// else, primary slice will be replaced with redundant slice.
if(p_Vid->frame_num == p_Vid->previous_frame_num && p_Vid->redundant_pic_cnt !=0
&& p_Vid->Is_primary_correct !=0 && current_header != EOS)
{
continue;
}
// update reference flags and set current p_Vid->ref_flag
if(!(p_Vid->redundant_pic_cnt != 0 && p_Vid->previous_frame_num == p_Vid->frame_num))
{
for(i=16;i>0;i--)
{
p_Vid->ref_flag[i] = p_Vid->ref_flag[i-1];
}
}
p_Vid->ref_flag[0] = p_Vid->redundant_pic_cnt==0 ? p_Vid->Is_primary_correct : p_Vid->Is_redundant_correct;
p_Vid->previous_frame_num = p_Vid->frame_num;
if (current_header == EOS)
{
// exit_picture(p_Vid, &p_Vid->dec_picture);
return EOS;
}
if((p_Vid->active_sps->chroma_format_idc==0)||(p_Vid->active_sps->chroma_format_idc==3))
{
currSlice->linfo_cbp_intra = linfo_cbp_intra_other;
currSlice->linfo_cbp_inter = linfo_cbp_inter_other;
}
else
{
currSlice->linfo_cbp_intra = linfo_cbp_intra_normal;
currSlice->linfo_cbp_inter = linfo_cbp_inter_normal;
}
decode_slice(currSlice, current_header);
++(p_Vid->current_slice_nr);
}
exit_picture(p_Vid, &p_Vid->dec_picture);
return (SOP);
}
/*!
************************************************************************
* \brief
* Convert file read buffer to source picture structure
* \param imgX
* Pointer to image plane
* \param buf
* Buffer for file output
* \param size_x
* horizontal image size in pixel
* \param size_y
* vertical image size in pixel
* \param symbol_size_in_bytes
* number of bytes used per pel
************************************************************************
*/
void buffer2img (imgpel** imgX, unsigned char* buf, int size_x, int size_y, int symbol_size_in_bytes)
{
int i,j;
uint16 tmp16, ui16;
unsigned long tmp32, ui32;
if (symbol_size_in_bytes> sizeof(imgpel))
{
error ("Source picture has higher bit depth than imgpel data type. \nPlease recompile with larger data type for imgpel.", 500);
}
if (( sizeof(char) == sizeof (imgpel)) && ( sizeof(char) == symbol_size_in_bytes))
{
// imgpel == pixel_in_file == 1 byte -> simple copy
memcpy(&imgX[0][0], buf, size_x * size_y);
}
else
{
// sizeof (imgpel) > sizeof(char)
if (testEndian())
{
// big endian
switch (symbol_size_in_bytes)
{
case 1:
{
for(j = 0; j < size_y; ++j)
for(i = 0; i < size_x; ++i)
{
imgX[j][i]= buf[i+j*size_x];
}
break;
}
case 2:
{
for(j=0;j<size_y;++j)
for(i=0;i<size_x;++i)
{
memcpy(&tmp16, buf+((i+j*size_x)*2), 2);
ui16 = (uint16) ((tmp16 >> 8) | ((tmp16&0xFF)<<8));
imgX[j][i] = (imgpel) ui16;
}
break;
}
case 4:
{
for(j=0;j<size_y;++j)
for(i=0;i<size_x;++i)
{
memcpy(&tmp32, buf+((i+j*size_x)*4), 4);
ui32 = ((tmp32&0xFF00)<<8) | ((tmp32&0xFF)<<24) | ((tmp32&0xFF0000)>>8) | ((tmp32&0xFF000000)>>24);
imgX[j][i] = (imgpel) ui32;
}
}
default:
{
error ("reading only from formats of 8, 16 or 32 bit allowed on big endian architecture", 500);
break;
}
}
}
else
{
// little endian
if (symbol_size_in_bytes == 1)
{
for (j=0; j < size_y; ++j)
{
for (i=0; i < size_x; ++i)
{
imgX[j][i]=*(buf++);
}
}
}
else
{
for (j=0; j < size_y; ++j)
{
int jpos = j*size_x;
for (i=0; i < size_x; ++i)
{
imgX[j][i]=0;
memcpy(&(imgX[j][i]), buf +((i+jpos)*symbol_size_in_bytes), symbol_size_in_bytes);
}
}
}
}
}
}
void reorder_lists(Slice *currSlice)
{
VideoParameters *p_Vid = currSlice->p_Vid;
if ((currSlice->slice_type != I_SLICE)&&(currSlice->slice_type != SI_SLICE))
{
if (currSlice->ref_pic_list_reordering_flag_l0)
{
reorder_ref_pic_list(p_Vid, p_Vid->listX[0], &p_Vid->listXsize[0],
currSlice->num_ref_idx_l0_active - 1,
currSlice->reordering_of_pic_nums_idc_l0,
currSlice->abs_diff_pic_num_minus1_l0,
currSlice->long_term_pic_idx_l0);
}
if (p_Vid->no_reference_picture == p_Vid->listX[0][currSlice->num_ref_idx_l0_active-1])
{
if (p_Vid->non_conforming_stream)
printf("RefPicList0[ num_ref_idx_l0_active_minus1 ] is equal to 'no reference picture'\n");
else
error("RefPicList0[ num_ref_idx_l0_active_minus1 ] is equal to 'no reference picture', invalid bitstream",500);
}
// that's a definition
p_Vid->listXsize[0] = (char) currSlice->num_ref_idx_l0_active;
}
if (currSlice->slice_type == B_SLICE)
{
if (currSlice->ref_pic_list_reordering_flag_l1)
{
reorder_ref_pic_list(p_Vid, p_Vid->listX[1], &p_Vid->listXsize[1],
currSlice->num_ref_idx_l1_active - 1,
currSlice->reordering_of_pic_nums_idc_l1,
currSlice->abs_diff_pic_num_minus1_l1,
currSlice->long_term_pic_idx_l1);
}
if (p_Vid->no_reference_picture == p_Vid->listX[1][currSlice->num_ref_idx_l1_active-1])
{
if (p_Vid->non_conforming_stream)
printf("RefPicList1[ num_ref_idx_l1_active_minus1 ] is equal to 'no reference picture'\n");
else
error("RefPicList1[ num_ref_idx_l1_active_minus1 ] is equal to 'no reference picture', invalid bitstream",500);
}
// that's a definition
p_Vid->listXsize[1] = (char) currSlice->num_ref_idx_l1_active;
}
free_ref_pic_list_reordering_buffer(currSlice);
}
/*!
************************************************************************
* \brief
* initialize ref_pic_num array
************************************************************************
*/
void set_ref_pic_num(VideoParameters *p_Vid)
{
int i,j;
StorablePicture *dec_picture = p_Vid->dec_picture;
int slice_id=p_Vid->current_slice_nr;
for (i=0;i<p_Vid->listXsize[LIST_0];++i)
{
dec_picture->ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->poc * 2 + ((p_Vid->listX[LIST_0][i]->structure==BOTTOM_FIELD)?1:0) ;
dec_picture->frm_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->frame_poc * 2;
dec_picture->top_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->top_poc * 2;
dec_picture->bottom_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->bottom_poc * 2 + 1;
//printf("POCS %d %d %d %d ",p_Vid->listX[LIST_0][i]->frame_poc,p_Vid->listX[LIST_0][i]->bottom_poc,p_Vid->listX[LIST_0][i]->top_poc,p_Vid->listX[LIST_0][i]->poc);
//printf("refid %d %d %d %d\n",(int) dec_picture->frm_ref_pic_num[LIST_0][i],(int) dec_picture->top_ref_pic_num[LIST_0][i],(int) dec_picture->bottom_ref_pic_num[LIST_0][i],(int) dec_picture->ref_pic_num[LIST_0][i]);
}
for (i=0;i<p_Vid->listXsize[LIST_1];++i)
{
dec_picture->ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->poc *2 + ((p_Vid->listX[LIST_1][i]->structure==BOTTOM_FIELD)?1:0);
dec_picture->frm_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->frame_poc * 2;
dec_picture->top_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->top_poc * 2;
dec_picture->bottom_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->bottom_poc * 2 + 1;
}
if (!p_Vid->active_sps->frame_mbs_only_flag)
{
if (p_Vid->structure==FRAME)
{
for (j=2;j<6;++j)
{
for (i=0;i<p_Vid->listXsize[j];++i)
{
dec_picture->ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->poc * 2 + ((p_Vid->listX[j][i]->structure==BOTTOM_FIELD)?1:0);
dec_picture->frm_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->frame_poc * 2 ;
dec_picture->top_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->top_poc * 2 ;
dec_picture->bottom_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->bottom_poc * 2 + 1;
}
}
}
}
}
/*!
************************************************************************
* \brief
* Reads new slice from bit_stream
************************************************************************
*/
// benski> time_code is user-passed data
static int read_new_slice(Slice *currSlice, uint64_t time_code)
{
VideoParameters *p_Vid = currSlice->p_Vid;
InputParameters *p_Inp = currSlice->p_Inp;
NALU_t *nalu = p_Vid->nalu;
int current_header = 0;
Bitstream *currStream;
int slice_id_a, slice_id_b, slice_id_c;
int redundant_pic_cnt_b, redundant_pic_cnt_c;
if (!nalu)
return EOS;
for (;;)
{
if (0 == read_next_nalu(p_Vid, nalu))
return EOS;
process_nalu:
switch (nalu->nal_unit_type)
{
case NALU_TYPE_SLICE:
case NALU_TYPE_IDR:
if (p_Vid->recovery_point || nalu->nal_unit_type == NALU_TYPE_IDR)
{
if (p_Vid->recovery_point_found == 0)
{
if (nalu->nal_unit_type != NALU_TYPE_IDR)
{
printf("Warning: Decoding does not start with an IDR picture.\n");
p_Vid->non_conforming_stream = 1;
}
else
p_Vid->non_conforming_stream = 0;
}
p_Vid->recovery_point_found = 1;
}
if (p_Vid->recovery_point_found == 0)
break;
p_Vid->idr_flag = (nalu->nal_unit_type == NALU_TYPE_IDR);
p_Vid->nal_reference_idc = nalu->nal_reference_idc;
currSlice->dp_mode = PAR_DP_1;
currSlice->max_part_nr = 1;
currStream = currSlice->partArr[0].bitstream;
currStream->frame_bitoffset = currStream->read_len = 0;
//memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1);
currStream->streamBuffer = &nalu->buf[1];
currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1);
// Some syntax of the Slice Header depends on the parameter set, which depends on
// the parameter set ID of the SLice header. Hence, read the pic_parameter_set_id
// of the slice header first, then setup the active parameter sets, and then read
// the rest of the slice header
FirstPartOfSliceHeader(currSlice);
UseParameterSet (currSlice, currSlice->pic_parameter_set_id);
RestOfSliceHeader (currSlice);
fmo_init (p_Vid);
currSlice->active_sps = p_Vid->active_sps;
currSlice->active_pps = p_Vid->active_pps;
assign_quant_params (currSlice);
// if primary slice is replaced with redundant slice, set the correct image type
if(p_Vid->redundant_pic_cnt && p_Vid->Is_primary_correct==0 && p_Vid->Is_redundant_correct)
{
p_Vid->dec_picture->slice_type = p_Vid->type;
}
if(is_new_picture(p_Vid->dec_picture, currSlice, p_Vid->old_slice))
{
init_picture(p_Vid, currSlice, p_Inp, time_code);
current_header = SOP;
//check zero_byte if it is also the first NAL unit in the access unit
CheckZeroByteVCL(p_Vid, nalu);
}
else
current_header = SOS;
if (currSlice->slice_type == B_SLICE)
{
if( IS_INDEPENDENT(p_Vid) )
{
int nplane = 0;
for( nplane=0; nplane<MAX_PLANE; nplane++ )
{
if( NULL != currSlice->Co_located_JV[nplane] )
{
free_colocated(p_Vid, currSlice->Co_located_JV[nplane]);
currSlice->Co_located_JV[nplane] = NULL;
}
currSlice->Co_located_JV[nplane] = alloc_colocated (p_Vid, p_Vid->width, p_Vid->height, p_Vid->active_sps->mb_adaptive_frame_field_flag);
}
}
else
{
if (NULL != currSlice->p_colocated)
{
free_colocated(p_Vid, currSlice->p_colocated);
currSlice->p_colocated = NULL;
}
currSlice->p_colocated = alloc_colocated (p_Vid, p_Vid->width, p_Vid->height,p_Vid->active_sps->mb_adaptive_frame_field_flag);
}
}
init_lists(currSlice);
reorder_lists (currSlice);
if (p_Vid->structure==FRAME)
{
init_mbaff_lists(p_Vid);
}
// From here on, p_Vid->active_sps, p_Vid->active_pps and the slice header are valid
if (currSlice->mb_aff_frame_flag)
p_Vid->current_mb_nr = currSlice->start_mb_nr << 1;
else
p_Vid->current_mb_nr = currSlice->start_mb_nr;
if (p_Vid->active_pps->entropy_coding_mode_flag)
{
int ByteStartPosition = currStream->frame_bitoffset/8;
if (currStream->frame_bitoffset%8 != 0)
{
++ByteStartPosition;
}
arideco_start_decoding (&currSlice->partArr[0].de_cabac, currStream->streamBuffer, ByteStartPosition, &currStream->read_len);
}
// printf ("read_new_slice: returning %s\n", current_header == SOP?"SOP":"SOS");
p_Vid->recovery_point = 0;
return current_header;
break;
case NALU_TYPE_DPA:
// read DP_A
currSlice->dpB_NotPresent =1;
currSlice->dpC_NotPresent =1;
p_Vid->idr_flag = FALSE;
p_Vid->nal_reference_idc = nalu->nal_reference_idc;
currSlice->dp_mode = PAR_DP_3;
currSlice->max_part_nr = 3;
currStream = currSlice->partArr[0].bitstream;
currStream->frame_bitoffset = currStream->read_len = 0;
//memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1);
currStream->streamBuffer = &nalu->buf[1];
currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1);
FirstPartOfSliceHeader(currSlice);
UseParameterSet (currSlice, currSlice->pic_parameter_set_id);
RestOfSliceHeader (currSlice);
fmo_init (p_Vid);
if(is_new_picture(p_Vid->dec_picture, currSlice, p_Vid->old_slice))
{
init_picture(p_Vid, currSlice, p_Inp, time_code);
current_header = SOP;
CheckZeroByteVCL(p_Vid, nalu);
}
else
current_header = SOS;
init_lists(currSlice);
reorder_lists (currSlice);
if (p_Vid->structure==FRAME)
{
init_mbaff_lists(p_Vid);
}
// From here on, p_Vid->active_sps, p_Vid->active_pps and the slice header are valid
if (currSlice->mb_aff_frame_flag)
p_Vid->current_mb_nr = currSlice->start_mb_nr << 1;
else
p_Vid->current_mb_nr = currSlice->start_mb_nr;
// Now I need to read the slice ID, which depends on the value of
// redundant_pic_cnt_present_flag
slice_id_a = ue_v("NALU: DP_A slice_id", currStream);
if (p_Vid->active_pps->entropy_coding_mode_flag)
error ("received data partition with CABAC, this is not allowed", 500);
// continue with reading next DP
if (0 == read_next_nalu(p_Vid, nalu))
return current_header;
if ( NALU_TYPE_DPB == nalu->nal_unit_type)
{
// we got a DPB
currStream = currSlice->partArr[1].bitstream;
currStream->frame_bitoffset = currStream->read_len = 0;
//memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1);
currStream->streamBuffer = &nalu->buf[1];
currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1);
slice_id_b = ue_v("NALU: DP_B slice_id", currStream);
currSlice->dpB_NotPresent = 0;
if ((slice_id_b != slice_id_a) || (nalu->lost_packets))
{
printf ("Waning: got a data partition B which does not match DP_A (DP loss!)\n");
currSlice->dpB_NotPresent =1;
currSlice->dpC_NotPresent =1;
}
else
{
if (p_Vid->active_pps->redundant_pic_cnt_present_flag)
redundant_pic_cnt_b = ue_v("NALU: DP_B redudant_pic_cnt", currStream);
else
redundant_pic_cnt_b = 0;
// we're finished with DP_B, so let's continue with next DP
if (0 == read_next_nalu(p_Vid, nalu))
return current_header;
}
}
else
{
currSlice->dpB_NotPresent =1;
}
// check if we got DP_C
if ( NALU_TYPE_DPC == nalu->nal_unit_type)
{
currStream = currSlice->partArr[2].bitstream;
currStream->frame_bitoffset = currStream->read_len = 0;
//memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1);
currStream->streamBuffer = &nalu->buf[1];
currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1);
currSlice->dpC_NotPresent = 0;
slice_id_c = ue_v("NALU: DP_C slice_id", currStream);
if ((slice_id_c != slice_id_a)|| (nalu->lost_packets))
{
printf ("Warning: got a data partition C which does not match DP_A(DP loss!)\n");
//currSlice->dpB_NotPresent =1;
currSlice->dpC_NotPresent =1;
}
if (p_Vid->active_pps->redundant_pic_cnt_present_flag)
redundant_pic_cnt_c = ue_v("NALU:SLICE_C redudand_pic_cnt", currStream);
else
redundant_pic_cnt_c = 0;
}
else
{
currSlice->dpC_NotPresent =1;
}
// check if we read anything else than the expected partitions
if ((nalu->nal_unit_type != NALU_TYPE_DPB) && (nalu->nal_unit_type != NALU_TYPE_DPC))
{
// we have a NALI that we can't process here, so restart processing
goto process_nalu;
// yes, "goto" should not be used, but it's really the best way here before we restructure the decoding loop
// (which should be taken care of anyway)
}
return current_header;
break;
case NALU_TYPE_DPB:
printf ("found data partition B without matching DP A, discarding\n");
break;
case NALU_TYPE_DPC:
printf ("found data partition C without matching DP A, discarding\n");
break;
case NALU_TYPE_SEI:
//printf ("read_new_slice: Found NALU_TYPE_SEI, len %d\n", nalu->len);
InterpretSEIMessage(nalu->buf,nalu->len,p_Vid);
break;
case NALU_TYPE_PPS:
ProcessPPS(p_Vid, nalu);
break;
case NALU_TYPE_SPS:
ProcessSPS(p_Vid, nalu);
break;
case NALU_TYPE_AUD:
// printf ("read_new_slice: Found 'Access Unit Delimiter' NAL unit, len %d, ignored\n", nalu->len);
break;
case NALU_TYPE_EOSEQ:
// printf ("read_new_slice: Found 'End of Sequence' NAL unit, len %d, ignored\n", nalu->len);
break;
case NALU_TYPE_EOSTREAM:
// printf ("read_new_slice: Found 'End of Stream' NAL unit, len %d, ignored\n", nalu->len);
break;
case NALU_TYPE_FILL:
printf ("read_new_slice: Found NALU_TYPE_FILL, len %d\n", (int) nalu->len);
printf ("Skipping these filling bits, proceeding w/ next NALU\n");
break;
default:
printf ("Found NALU type %d, len %d undefined, ignore NALU, moving on\n", (int) nalu->nal_unit_type, (int) nalu->len);
break;
}
}
return current_header;
}
/*!
************************************************************************
* \brief
* finish decoding of a picture, conceal errors and store it
* into the DPB
************************************************************************
*/
void exit_picture(VideoParameters *p_Vid, StorablePicture **dec_picture)
{
InputParameters *p_Inp = p_Vid->p_Inp;
int ercStartMB;
int ercSegment;
frame recfr;
unsigned int i;
int structure, frame_poc, slice_type, refpic, qp, pic_num, chroma_format_idc, is_idr;
int nplane;
// return if the last picture has already been finished
if (*dec_picture==NULL)
{
return;
}
recfr.p_Vid = p_Vid;
recfr.yptr = (*dec_picture)->imgY->base_address;
if ((*dec_picture)->chroma_format_idc != YUV400)
{
recfr.uptr = (*dec_picture)->imgUV[0]->base_address;
recfr.vptr = (*dec_picture)->imgUV[1]->base_address;
}
//! this is always true at the beginning of a picture
ercStartMB = 0;
ercSegment = 0;
//! mark the start of the first segment
if (!(*dec_picture)->mb_aff_frame_flag)
{
ercStartSegment(0, ercSegment, 0 , p_Vid->erc_errorVar);
//! generate the segments according to the macroblock map
for(i = 1; i<(*dec_picture)->PicSizeInMbs; ++i)
{
if(p_Vid->mb_data[i].ei_flag != p_Vid->mb_data[i-1].ei_flag)
{
ercStopSegment(i-1, ercSegment, 0, p_Vid->erc_errorVar); //! stop current segment
//! mark current segment as lost or OK
if(p_Vid->mb_data[i-1].ei_flag)
ercMarkCurrSegmentLost((*dec_picture)->size_x, p_Vid->erc_errorVar);
else
ercMarkCurrSegmentOK((*dec_picture)->size_x, p_Vid->erc_errorVar);
++ercSegment; //! next segment
ercStartSegment(i, ercSegment, 0 , p_Vid->erc_errorVar); //! start new segment
ercStartMB = i;//! save start MB for this segment
}
}
//! mark end of the last segment
ercStopSegment((*dec_picture)->PicSizeInMbs-1, ercSegment, 0, p_Vid->erc_errorVar);
if(p_Vid->mb_data[i-1].ei_flag)
ercMarkCurrSegmentLost((*dec_picture)->size_x, p_Vid->erc_errorVar);
else
ercMarkCurrSegmentOK((*dec_picture)->size_x, p_Vid->erc_errorVar);
//! call the right error concealment function depending on the frame type.
p_Vid->erc_mvperMB /= (*dec_picture)->PicSizeInMbs;
p_Vid->erc_img = p_Vid;
if((*dec_picture)->slice_type == I_SLICE || (*dec_picture)->slice_type == SI_SLICE) // I-frame
ercConcealIntraFrame(p_Vid, &recfr, (*dec_picture)->size_x, (*dec_picture)->size_y, p_Vid->erc_errorVar);
else
ercConcealInterFrame(&recfr, p_Vid->erc_object_list, (*dec_picture)->size_x, (*dec_picture)->size_y, p_Vid->erc_errorVar, (*dec_picture)->chroma_format_idc);
}
//deblocking for frame or field
if( IS_INDEPENDENT(p_Vid) )
{
int colour_plane_id = p_Vid->colour_plane_id;
for( nplane=0; nplane<MAX_PLANE; ++nplane )
{
change_plane_JV( p_Vid, nplane );
DeblockPicture( p_Vid, *dec_picture );
}
p_Vid->colour_plane_id = colour_plane_id;
make_frame_picture_JV(p_Vid);
}
else
{
DeblockPicture( p_Vid, *dec_picture );
}
if ((*dec_picture)->mb_aff_frame_flag)
MbAffPostProc(p_Vid);
if (p_Vid->structure == FRAME) // buffer mgt. for frame mode
frame_postprocessing(p_Vid);
else
field_postprocessing(p_Vid); // reset all interlaced variables
structure = (*dec_picture)->structure;
slice_type = (*dec_picture)->slice_type;
frame_poc = (*dec_picture)->frame_poc;
refpic = (*dec_picture)->used_for_reference;
qp = (*dec_picture)->qp;
pic_num = (*dec_picture)->pic_num;
is_idr = (*dec_picture)->idr_flag;
chroma_format_idc = (*dec_picture)->chroma_format_idc;
store_picture_in_dpb(p_Vid, *dec_picture);
*dec_picture=NULL;
if (p_Vid->last_has_mmco_5)
{
p_Vid->pre_frame_num = 0;
}
if ((structure==FRAME)||structure==BOTTOM_FIELD)
{
if(slice_type == I_SLICE || slice_type == SI_SLICE || slice_type == P_SLICE || refpic) // I or P pictures
++(p_Vid->number);
}
p_Vid->current_mb_nr = -4712; // impossible value for debugging, StW
p_Vid->current_slice_nr = 0;
}
/*!
************************************************************************
* \brief
* write the encoding mode and motion vectors of current
* MB to the buffer of the error concealment module.
************************************************************************
*/
void ercWriteMBMODEandMV(Macroblock *currMB)
{
VideoParameters *p_Vid = currMB->p_Vid;
int i, ii, jj, currMBNum = p_Vid->current_mb_nr;
StorablePicture *dec_picture = p_Vid->dec_picture;
int mbx = 4*xPosMB(currMBNum, dec_picture->size_x), mby = 4*yPosMB(currMBNum, dec_picture->size_x);
objectBuffer_t *currRegion, *pRegion;
currRegion = p_Vid->erc_object_list + (currMBNum<<2);
if(p_Vid->type != B_SLICE) //non-B frame
{
for (i=0; i<4; ++i)
{
pRegion = currRegion + i;
pRegion->regionMode = (currMB->mb_type ==I16MB ? REGMODE_INTRA :
currMB->b8mode[i]==IBLOCK ? REGMODE_INTRA_8x8 :
currMB->b8mode[i]==0 ? REGMODE_INTER_COPY :
currMB->b8mode[i]==1 ? REGMODE_INTER_PRED : REGMODE_INTER_PRED_8x8);
if (currMB->b8mode[i]==0 || currMB->b8mode[i]==IBLOCK) // INTRA OR COPY
{
pRegion->mv[0] = 0;
pRegion->mv[1] = 0;
pRegion->mv[2] = 0;
}
else
{
PicMotion *motion00;
PicMotion *motion01;
ii = mbx + (i & 0x01)*2;// + BLOCK_SIZE;
jj = mby + (i & ~1 );
motion00 = &dec_picture->motion.motion[LIST_0][jj][ii];
if (currMB->b8mode[i]>=5 && currMB->b8mode[i]<=7) // SMALL BLOCKS
{
motion01 = &dec_picture->motion.motion[LIST_0][jj+1][ii];
pRegion->mv[0] = (motion00[0].mv[0] + motion00[1].mv[0] + motion01[0].mv[0] + motion01[1].mv[0] + 2)/4;
pRegion->mv[1] = (motion00[0].mv[1] + motion00[1].mv[1] + motion01[0].mv[1] + motion01[1].mv[1] + 2)/4;
}
else // 16x16, 16x8, 8x16, 8x8
{
pRegion->mv[0] = motion00->mv[0];
pRegion->mv[1] = motion00->mv[1];
}
p_Vid->erc_mvperMB += abs(pRegion->mv[0]) + abs(pRegion->mv[1]);
pRegion->mv[2] = motion00->ref_idx;
}
}
}
else //B-frame
{
for (i=0; i<4; ++i)
{
ii = mbx + (i%2)*2;// + BLOCK_SIZE;
jj = mby + (i/2)*2;
pRegion = currRegion + i;
pRegion->regionMode = (currMB->mb_type ==I16MB ? REGMODE_INTRA :
currMB->b8mode[i]==IBLOCK ? REGMODE_INTRA_8x8 : REGMODE_INTER_PRED_8x8);
if (currMB->mb_type==I16MB || currMB->b8mode[i]==IBLOCK) // INTRA
{
pRegion->mv[0] = 0;
pRegion->mv[1] = 0;
pRegion->mv[2] = 0;
}
else
{
PicMotion *motion0= & dec_picture->motion.motion[LIST_0][jj][ii];
PicMotion *motion1;
if (motion0->ref_idx<0)
{
motion0 = & dec_picture->motion.motion[LIST_1][jj][ii];
motion1 = & dec_picture->motion.motion[LIST_1][jj+1][ii];
}
else
{
motion1 = & dec_picture->motion.motion[LIST_0][jj+1][ii];
}
// int idx = (currMB->b8mode[i]==0 && currMB->b8pdir[i]==2 ? LIST_0 : currMB->b8pdir[i]==1 ? LIST_1 : LIST_0);
// int idx = currMB->b8pdir[i]==0 ? LIST_0 : LIST_1;
pRegion->mv[0] = (motion0[0].mv[0] + motion0[1].mv[0] + motion1[0].mv[0] + motion1[1].mv[0] + 2)/4;
pRegion->mv[1] = (motion0[0].mv[1] + motion0[1].mv[1] + motion1[0].mv[1] + motion1[1].mv[1] + 2)/4;
p_Vid->erc_mvperMB += iabs(pRegion->mv[0]) + iabs(pRegion->mv[1]);
pRegion->mv[2] = (motion0[0].ref_idx);
/*
if (currMB->b8pdir[i]==0 || (currMB->b8pdir[i]==2 && currMB->b8mode[i]!=0)) // forward or bidirect
{
pRegion->mv[2] = (dec_picture->motion.ref_idx[LIST_0][jj][ii]);
///???? is it right, not only "p_Vid->fw_refFrArr[jj][ii-4]"
}
else
{
pRegion->mv[2] = (dec_picture->motion.ref_idx[LIST_1][jj][ii]);
// pRegion->mv[2] = 0;
}
*/
}
}
}
}
/*!
************************************************************************
* \brief
* set defaults for old_slice
* NAL unit of a picture"
************************************************************************
*/
void init_old_slice(OldSliceParams *p_old_slice)
{
p_old_slice->field_pic_flag = 0;
p_old_slice->pps_id = INT_MAX;
p_old_slice->frame_num = INT_MAX;
p_old_slice->nal_ref_idc = INT_MAX;
p_old_slice->idr_flag = FALSE;
p_old_slice->pic_oder_cnt_lsb = UINT_MAX;
p_old_slice->delta_pic_oder_cnt_bottom = INT_MAX;
p_old_slice->delta_pic_order_cnt[0] = INT_MAX;
p_old_slice->delta_pic_order_cnt[1] = INT_MAX;
}
/*!
************************************************************************
* \brief
* save slice parameters that are needed for checking of "first VCL
* NAL unit of a picture"
************************************************************************
*/
static void exit_slice(Slice *currSlice, OldSliceParams *p_old_slice)
{
VideoParameters *p_Vid = currSlice->p_Vid;
if (currSlice->slice_type == B_SLICE)
{
if( IS_INDEPENDENT(p_Vid) )
{
int nplane;
for( nplane = 0; nplane < MAX_PLANE; ++nplane )
{
free_colocated(p_Vid, currSlice->Co_located_JV[nplane]);
currSlice->Co_located_JV[nplane] = NULL;
}
}
else
{
free_colocated(p_Vid, currSlice->p_colocated);
currSlice->p_colocated = NULL;
}
}
p_old_slice->pps_id = currSlice->pic_parameter_set_id;
p_old_slice->frame_num = p_Vid->frame_num;
p_old_slice->field_pic_flag = p_Vid->field_pic_flag;
if(p_Vid->field_pic_flag)
{
p_old_slice->bottom_field_flag = p_Vid->bottom_field_flag;
}
p_old_slice->nal_ref_idc = p_Vid->nal_reference_idc;
p_old_slice->idr_flag = (byte) p_Vid->idr_flag;
if (p_Vid->idr_flag)
{
p_old_slice->idr_pic_id = p_Vid->idr_pic_id;
}
if (p_Vid->active_sps->pic_order_cnt_type == 0)
{
p_old_slice->pic_oder_cnt_lsb = p_Vid->pic_order_cnt_lsb;
p_old_slice->delta_pic_oder_cnt_bottom = p_Vid->delta_pic_order_cnt_bottom;
}
if (p_Vid->active_sps->pic_order_cnt_type == 1)
{
p_old_slice->delta_pic_order_cnt[0] = p_Vid->delta_pic_order_cnt[0];
p_old_slice->delta_pic_order_cnt[1] = p_Vid->delta_pic_order_cnt[1];
}
}
/*!
************************************************************************
* \brief
* detect if current slice is "first VCL NAL unit of a picture"
************************************************************************
*/
int is_new_picture(StorablePicture *dec_picture, Slice *currSlice, OldSliceParams *p_old_slice)
{
VideoParameters *p_Vid = currSlice->p_Vid;
int result=0;
result |= (NULL==dec_picture);
result |= (p_old_slice->pps_id != currSlice->pic_parameter_set_id);
result |= (p_old_slice->frame_num != p_Vid->frame_num);
result |= (p_old_slice->field_pic_flag != p_Vid->field_pic_flag);
if(p_Vid->field_pic_flag && p_old_slice->field_pic_flag)
{
result |= (p_old_slice->bottom_field_flag != p_Vid->bottom_field_flag);
}
result |= (p_old_slice->nal_ref_idc != p_Vid->nal_reference_idc) && ((p_old_slice->nal_ref_idc == 0) || (p_Vid->nal_reference_idc == 0));
result |= ( p_old_slice->idr_flag != p_Vid->idr_flag);
if (p_Vid->idr_flag && p_old_slice->idr_flag)
{
result |= (p_old_slice->idr_pic_id != p_Vid->idr_pic_id);
}
if (p_Vid->active_sps->pic_order_cnt_type == 0)
{
result |= (p_old_slice->pic_oder_cnt_lsb != p_Vid->pic_order_cnt_lsb);
result |= (p_old_slice->delta_pic_oder_cnt_bottom != p_Vid->delta_pic_order_cnt_bottom);
}
if (p_Vid->active_sps->pic_order_cnt_type == 1)
{
result |= (p_old_slice->delta_pic_order_cnt[0] != p_Vid->delta_pic_order_cnt[0]);
result |= (p_old_slice->delta_pic_order_cnt[1] != p_Vid->delta_pic_order_cnt[1]);
}
return result;
}
/*!
************************************************************************
* \brief
* Prepare field and frame buffer after frame decoding
************************************************************************
*/
void frame_postprocessing(VideoParameters *p_Vid)
{
}
/*!
************************************************************************
* \brief
* Prepare field and frame buffer after field decoding
************************************************************************
*/
void field_postprocessing(VideoParameters *p_Vid)
{
p_Vid->number /= 2;
}
/*!
************************************************************************
* \brief
* copy StorablePicture *src -> StorablePicture *dst
* for 4:4:4 Independent mode
************************************************************************
*/
void copy_dec_picture_JV( VideoParameters *p_Vid, StorablePicture *dst, StorablePicture *src )
{
dst->top_poc = src->top_poc;
dst->bottom_poc = src->bottom_poc;
dst->frame_poc = src->frame_poc;
dst->qp = src->qp;
dst->slice_qp_delta = src->slice_qp_delta;
dst->chroma_qp_offset[0] = src->chroma_qp_offset[0];
dst->chroma_qp_offset[1] = src->chroma_qp_offset[1];
dst->poc = src->poc;
dst->slice_type = src->slice_type;
dst->used_for_reference = src->used_for_reference;
dst->idr_flag = src->idr_flag;
dst->no_output_of_prior_pics_flag = src->no_output_of_prior_pics_flag;
dst->long_term_reference_flag = src->long_term_reference_flag;
dst->adaptive_ref_pic_buffering_flag = src->adaptive_ref_pic_buffering_flag;
dst->dec_ref_pic_marking_buffer = src->dec_ref_pic_marking_buffer;
dst->mb_aff_frame_flag = src->mb_aff_frame_flag;
dst->PicWidthInMbs = src->PicWidthInMbs;
dst->pic_num = src->pic_num;
dst->frame_num = src->frame_num;
dst->recovery_frame = src->recovery_frame;
dst->coded_frame = src->coded_frame;
dst->chroma_format_idc = src->chroma_format_idc;
dst->frame_mbs_only_flag = src->frame_mbs_only_flag;
dst->frame_cropping_flag = src->frame_cropping_flag;
dst->frame_cropping_rect_left_offset = src->frame_cropping_rect_left_offset;
dst->frame_cropping_rect_right_offset = src->frame_cropping_rect_right_offset;
dst->frame_cropping_rect_top_offset = src->frame_cropping_rect_top_offset;
dst->frame_cropping_rect_bottom_offset = src->frame_cropping_rect_bottom_offset;
#if (ENABLE_OUTPUT_TONEMAPPING)
// store the necessary tone mapping sei into StorablePicture structure
dst->seiHasTone_mapping = src->seiHasTone_mapping;
dst->seiHasTone_mapping = src->seiHasTone_mapping;
dst->tone_mapping_model_id = src->tone_mapping_model_id;
dst->tonemapped_bit_depth = src->tonemapped_bit_depth;
if( src->tone_mapping_lut )
{
dst->tone_mapping_lut = malloc(sizeof(int)*(1 << p_Vid->seiToneMapping->coded_data_bit_depth));
if (NULL == dst->tone_mapping_lut)
{
no_mem_exit("copy_dec_picture_JV: tone_mapping_lut");
}
memcpy(dst->tone_mapping_lut, src->tone_mapping_lut, sizeof(imgpel)*(1<<p_Vid->seiToneMapping->coded_data_bit_depth));
}
#endif
}
/*!
************************************************************************
* \brief
* decodes one slice
************************************************************************
*/
void decode_one_slice(Slice *currSlice)
{
VideoParameters *p_Vid = currSlice->p_Vid;
Boolean end_of_slice = FALSE;
Macroblock *currMB = NULL;
p_Vid->cod_counter=-1;
setup_slice_methods(currSlice);
if( IS_INDEPENDENT(p_Vid) )
{
change_plane_JV( p_Vid, p_Vid->colour_plane_id );
}
set_ref_pic_num(p_Vid);
if (currSlice->slice_type == B_SLICE)
{
currSlice->compute_colocated(currSlice, currSlice->p_colocated, p_Vid->listX);
}
//reset_ec_flags(p_Vid);
while (end_of_slice == FALSE) // loop over macroblocks
{
#if TRACE
fprintf(p_trace,"\n*********** POC: %i (I/P) MB: %i Slice: %i Type %d **********\n", p_Vid->ThisPOC, p_Vid->current_mb_nr, p_Vid->current_slice_nr, currSlice->slice_type);
#endif
// Initializes the current macroblock
start_macroblock(currSlice, &currMB);
// Get the syntax elements from the NAL
currSlice->read_one_macroblock(currMB);
decode_one_macroblock(currMB, p_Vid->dec_picture);
if(currSlice->mb_aff_frame_flag && p_Vid->dec_picture->motion.mb_field[p_Vid->current_mb_nr])
{
currSlice->num_ref_idx_l0_active >>= 1;
currSlice->num_ref_idx_l1_active >>= 1;
}
ercWriteMBMODEandMV(currMB);
end_of_slice = exit_macroblock(currSlice, (!currSlice->mb_aff_frame_flag||p_Vid->current_mb_nr%2));
}
exit_slice(currSlice, p_Vid->old_slice);
//reset_ec_flags(p_Vid);
}