/*! *********************************************************************** * \file image.c * * \brief * Decode a Slice * * \author * Main contributors (see contributors.h for copyright, address and affiliation details) * - Inge Lille-Langoy * - Rickard Sjoberg * - Jani Lainema * - Sebastian Purreiter * - Byeong-Moon Jeon * - Thomas Wedi * - Gabi Blaettermann * - Ye-Kui Wang * - Antti Hallapuro * - Alexis Tourapis * - Jill Boyce * - Saurav K Bandyopadhyay * - Zhenyu Wu * *********************************************************************** */ #include "contributors.h" #include #include #include "global.h" #include "image.h" #include "fmo.h" #include "nalu.h" #include "parset.h" #include "header.h" #include "sei.h" #include "output.h" #include "mb_access.h" #include "memalloc.h" #include "macroblock.h" #include "loopfilter.h" #include "biaridecod.h" #include "context_ini.h" #include "cabac.h" #include "vlc.h" #include "quant.h" #include "errorconcealment.h" #include "erc_api.h" /*! ************************************************************************ * \brief * checks if the System is big- or little-endian * \return * 0, little-endian (e.g. Intel architectures) * 1, big-endian (e.g. SPARC, MIPS, PowerPC) ************************************************************************ */ int testEndian(void) { short s; byte *p; p=(byte*)&s; s=1; return (*p==0); } static int read_new_slice(Slice *currSlice, uint64_t time_code); /*! ************************************************************************ * \brief * Initializes the parameters for a new picture ************************************************************************ */ // benski> time_code is user-passed value static void init_picture(VideoParameters *p_Vid, Slice *currSlice, InputParameters *p_Inp, uint64_t time_code) { int i; int nplane; if (p_Vid->dec_picture) { // this may only happen on slice loss exit_picture(p_Vid, &p_Vid->dec_picture); } if (p_Vid->recovery_point) p_Vid->recovery_frame_num = (p_Vid->frame_num + p_Vid->recovery_frame_cnt) % p_Vid->MaxFrameNum; if (p_Vid->idr_flag) p_Vid->recovery_frame_num = p_Vid->frame_num; if (p_Vid->recovery_point == 0 && p_Vid->pre_frame_num != INT_MIN && p_Vid->frame_num != p_Vid->pre_frame_num && p_Vid->frame_num != (p_Vid->pre_frame_num + 1) % p_Vid->MaxFrameNum) { if (p_Vid->active_sps->gaps_in_frame_num_value_allowed_flag == 0) { /* Advanced Error Concealment would be called here to combat unintentional loss of pictures. */ error("An unintentional loss of pictures occurs! Exit\n", 100); } fill_frame_num_gap(p_Vid); } if(p_Vid->nal_reference_idc) { p_Vid->pre_frame_num = p_Vid->frame_num; } p_Vid->num_dec_mb = 0; //calculate POC decode_poc(p_Vid); if (p_Vid->recovery_frame_num == p_Vid->frame_num && p_Vid->recovery_poc == 0x7fffffff) p_Vid->recovery_poc = p_Vid->framepoc; if(p_Vid->nal_reference_idc) p_Vid->last_ref_pic_poc = p_Vid->framepoc; // dumppoc (p_Vid); p_Vid->dec_picture = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr); p_Vid->dec_picture->time_code = time_code; p_Vid->dec_picture->top_poc=p_Vid->toppoc; p_Vid->dec_picture->bottom_poc=p_Vid->bottompoc; p_Vid->dec_picture->frame_poc=p_Vid->framepoc; p_Vid->dec_picture->qp = p_Vid->qp; p_Vid->dec_picture->slice_qp_delta = currSlice->slice_qp_delta; p_Vid->dec_picture->chroma_qp_offset[0] = p_Vid->active_pps->chroma_qp_index_offset; p_Vid->dec_picture->chroma_qp_offset[1] = p_Vid->active_pps->second_chroma_qp_index_offset; // reset all variables of the error concealment instance before decoding of every frame. // here the third parameter should, if perfectly, be equal to the number of slices per frame. // using little value is ok, the code will allocate more memory if the slice number is larger ercReset(p_Vid->erc_errorVar, p_Vid->PicSizeInMbs, p_Vid->PicSizeInMbs, p_Vid->dec_picture->size_x); p_Vid->erc_mvperMB = 0; switch (p_Vid->structure ) { case TOP_FIELD: { p_Vid->dec_picture->poc=p_Vid->toppoc; p_Vid->number *= 2; break; } case BOTTOM_FIELD: { p_Vid->dec_picture->poc=p_Vid->bottompoc; p_Vid->number = p_Vid->number * 2 + 1; break; } case FRAME: { p_Vid->dec_picture->poc=p_Vid->framepoc; break; } default: error("p_Vid->structure not initialized", 235); } p_Vid->current_slice_nr=0; if (p_Vid->type > SI_SLICE) { set_ec_flag(p_Vid, SE_PTYPE); p_Vid->type = P_SLICE; // concealed element } // CAVLC init if (p_Vid->active_pps->entropy_coding_mode_flag == CAVLC) { memset(p_Vid->nz_coeff[0], -1, p_Vid->PicSizeInMbs * 48 *sizeof(byte)); // 3 * 4 * 4 } if(p_Vid->active_pps->constrained_intra_pred_flag) { for (i=0; i<(int)p_Vid->PicSizeInMbs; ++i) { p_Vid->intra_block[i] = 1; } } // Set the slice_nr member of each MB to -1, to ensure correct when packet loss occurs // TO set Macroblock Map (mark all MBs as 'have to be concealed') if( IS_INDEPENDENT(p_Vid) ) { for( nplane=0; nplanePicSizeInMbs; ++i) { p_Vid->mb_data_JV[nplane][i].slice_nr = -1; p_Vid->mb_data_JV[nplane][i].ei_flag = 1; p_Vid->mb_data_JV[nplane][i].dpl_flag = 0; } } } else { for(i=0; i<(int)p_Vid->PicSizeInMbs; ++i) { p_Vid->mb_data[i].slice_nr = -1; p_Vid->mb_data[i].ei_flag = 1; p_Vid->mb_data[i].dpl_flag = 0; } } p_Vid->dec_picture->slice_type = p_Vid->type; p_Vid->dec_picture->used_for_reference = (p_Vid->nal_reference_idc != 0); p_Vid->dec_picture->idr_flag = p_Vid->idr_flag; p_Vid->dec_picture->no_output_of_prior_pics_flag = p_Vid->no_output_of_prior_pics_flag; p_Vid->dec_picture->long_term_reference_flag = p_Vid->long_term_reference_flag; p_Vid->dec_picture->adaptive_ref_pic_buffering_flag = p_Vid->adaptive_ref_pic_buffering_flag; p_Vid->dec_picture->dec_ref_pic_marking_buffer = p_Vid->dec_ref_pic_marking_buffer; p_Vid->dec_ref_pic_marking_buffer = NULL; p_Vid->dec_picture->mb_aff_frame_flag = p_Vid->mb_aff_frame_flag; p_Vid->dec_picture->PicWidthInMbs = p_Vid->PicWidthInMbs; if (p_Vid->dec_picture->mb_aff_frame_flag) { p_Vid->get_mb_block_pos = get_mb_block_pos_mbaff; p_Vid->getNeighbour = getAffNeighbour; p_Vid->getNeighbourXP_NoPos = getAffNeighbour; p_Vid->getNeighbourPX_NoPos = getAffNeighbour; p_Vid->getNeighbourLuma = getAffNeighbourLuma; p_Vid->getNeighbourPXLuma = getAffNeighbourLuma; p_Vid->getNeighbourXPLuma = getAffNeighbourXPLuma; p_Vid->getNeighbourLeftLuma = getAffNeighbourN0Luma; p_Vid->getNeighbourNXLuma = getAffNeighbourNXLuma; p_Vid->getNeighbourLeft = getAffNeighbourN0; p_Vid->getNeighbourUp = getAffNeighbour0N; p_Vid->getNeighbourUpLuma = getAffNeighbour0NLuma; p_Vid->getNeighbourNX = getAffNeighbourNX; p_Vid->getNeighbourNP = getAffNeighbourNX; p_Vid->getNeighbourNPChromaNB = getAffNeighbourNX; p_Vid->getNeighbour0X = getAffNeighbour0X; p_Vid->getNeighbour0XLuma = getAffNeighbour0XLuma; p_Vid->getNeighbourX0 = getAffNeighbourX0; p_Vid->getNeighbourNPLumaNB = getAffNeighbourNPLuma; p_Vid->getNeighbourPXLumaNB = getAffNeighbourPXLumaNB; p_Vid->getNeighbourPXLumaNB_NoPos = getAffNeighbourPXLumaNB_NoPos; p_Vid->getNeighbourXPLumaNB = getAffNeighbourXPLuma; p_Vid->getNeighbourPPLumaNB = getAffNeighbourPPLumaNB; p_Vid->getNeighbourXPLumaNB_NoPos = getAffNeighbourXPLuma; } else { p_Vid->get_mb_block_pos = get_mb_block_pos_normal; p_Vid->getNeighbour = getNonAffNeighbour; p_Vid->getNeighbourXP_NoPos = getNonAffNeighbourXP_NoPos; p_Vid->getNeighbourPX_NoPos = getNonAffNeighbourPX_NoPos; p_Vid->getNeighbourLuma = getNonAffNeighbourLuma; p_Vid->getNeighbourPXLuma = getNonAffNeighbourPXLuma; p_Vid->getNeighbourXPLuma = getNonAffNeighbourXPLuma; p_Vid->getNeighbourLeftLuma = getNonAffNeighbourN0Luma; p_Vid->getNeighbourNXLuma = getNonAffNeighbourNXLuma; p_Vid->getNeighbourLeft = getNonAffNeighbourN0; p_Vid->getNeighbourUp = getNonAffNeighbour0N; p_Vid->getNeighbourUpLuma = getNonAffNeighbour0NLuma; p_Vid->getNeighbourNX = getNonAffNeighbourNX; p_Vid->getNeighbourNP = getNonAffNeighbourNP; p_Vid->getNeighbourNPChromaNB = getNonAffNeighbourNPChromaNB; p_Vid->getNeighbour0X = getNonAffNeighbour0X; p_Vid->getNeighbour0XLuma = getNonAffNeighbour0XLuma; p_Vid->getNeighbourX0 = getNonAffNeighbourX0; p_Vid->getNeighbourNPLumaNB = getNonAffNeighbourNPLumaNB; p_Vid->getNeighbourPXLumaNB = getNonAffNeighbourPXLumaNB; p_Vid->getNeighbourPXLumaNB_NoPos = getNonAffNeighbourPXLumaNB_NoPos; p_Vid->getNeighbourXPLumaNB = getNonAffNeighbourXPLumaNB; p_Vid->getNeighbourPPLumaNB = getNonAffNeighbourPPLumaNB; p_Vid->getNeighbourXPLumaNB_NoPos = getNonAffNeighbourXPLumaNB_NoPos; } p_Vid->dec_picture->pic_num = p_Vid->frame_num; p_Vid->dec_picture->frame_num = p_Vid->frame_num; p_Vid->dec_picture->recovery_frame = (unsigned int) (p_Vid->frame_num == p_Vid->recovery_frame_num); p_Vid->dec_picture->coded_frame = (p_Vid->structure==FRAME); p_Vid->dec_picture->chroma_format_idc = p_Vid->active_sps->chroma_format_idc; p_Vid->dec_picture->frame_mbs_only_flag = p_Vid->active_sps->frame_mbs_only_flag; p_Vid->dec_picture->frame_cropping_flag = p_Vid->active_sps->frame_cropping_flag; if (p_Vid->dec_picture->frame_cropping_flag) { p_Vid->dec_picture->frame_cropping_rect_left_offset = p_Vid->active_sps->frame_cropping_rect_left_offset; p_Vid->dec_picture->frame_cropping_rect_right_offset = p_Vid->active_sps->frame_cropping_rect_right_offset; p_Vid->dec_picture->frame_cropping_rect_top_offset = p_Vid->active_sps->frame_cropping_rect_top_offset; p_Vid->dec_picture->frame_cropping_rect_bottom_offset = p_Vid->active_sps->frame_cropping_rect_bottom_offset; } #if (ENABLE_OUTPUT_TONEMAPPING) // store the necessary tone mapping sei into StorablePicture structure p_Vid->dec_picture->seiHasTone_mapping = 0; if (p_Vid->seiToneMapping->seiHasTone_mapping) { p_Vid->dec_picture->seiHasTone_mapping = 1; p_Vid->dec_picture->tone_mapping_model_id = p_Vid->seiToneMapping->model_id; p_Vid->dec_picture->tonemapped_bit_depth = p_Vid->seiToneMapping->sei_bit_depth; p_Vid->dec_picture->tone_mapping_lut = malloc(sizeof(int)*(1<seiToneMapping->coded_data_bit_depth)); if (NULL == p_Vid->dec_picture->tone_mapping_lut) { no_mem_exit("init_picture: tone_mapping_lut"); } memcpy(p_Vid->dec_picture->tone_mapping_lut, p_Vid->seiToneMapping->lut, sizeof(imgpel)*(1<seiToneMapping->coded_data_bit_depth)); update_tone_mapping_sei(p_Vid->seiToneMapping); } #endif if( IS_INDEPENDENT(p_Vid) ) { p_Vid->dec_picture_JV[0] = p_Vid->dec_picture; p_Vid->dec_picture_JV[1] = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr); copy_dec_picture_JV( p_Vid, p_Vid->dec_picture_JV[1], p_Vid->dec_picture_JV[0] ); p_Vid->dec_picture_JV[2] = alloc_storable_picture (p_Vid, (PictureStructure) p_Vid->structure, p_Vid->width, p_Vid->height, p_Vid->width_cr, p_Vid->height_cr); copy_dec_picture_JV( p_Vid, p_Vid->dec_picture_JV[2], p_Vid->dec_picture_JV[0] ); } } void MbAffPostProc(VideoParameters *p_Vid) { imgpel temp[32][16]; StorablePicture *dec_picture = p_Vid->dec_picture; imgpel ** imgY = dec_picture->imgY->img; imgpel **imgUV[2] = {dec_picture->imgUV[0]->img, dec_picture->imgUV[1]->img}; short i, y, x0, y0, uv; for (i=0; i<(int)dec_picture->PicSizeInMbs; i+=2) { if (dec_picture->motion.mb_field[i]) { get_mb_pos(p_Vid, i, p_Vid->mb_size[IS_LUMA], &x0, &y0); for (y=0; y<(2*MB_BLOCK_SIZE);++y) memcpy(temp[y], &imgY[y0+y][x0], MB_BLOCK_SIZE * sizeof(imgpel)); for (y=0; ychroma_format_idc != YUV400) { x0 = (short) (x0 / (16/p_Vid->mb_cr_size_x)); y0 = (short) (y0 / (16/p_Vid->mb_cr_size_y)); for (uv=0; uv<2; ++uv) { for (y=0; y<(2*p_Vid->mb_cr_size_y);++y) memcpy(temp[y], &imgUV[uv][y0+y][x0], p_Vid->mb_cr_size_x * sizeof(imgpel)); for (y=0; ymb_cr_size_y;++y) { memcpy(&imgUV[uv][y0+(2*y )][x0], temp[y ], p_Vid->mb_cr_size_x * sizeof(imgpel)); memcpy(&imgUV[uv][y0+(2*y+1)][x0], temp[y+p_Vid->mb_cr_size_y], p_Vid->mb_cr_size_x * sizeof(imgpel)); } } } } } } static void fill_wp_params(Slice *currSlice) { VideoParameters *p_Vid = currSlice->p_Vid; int i, j, k; int comp; int log_weight_denom; int tb, td; int tx,DistScaleFactor; int max_l0_ref = currSlice->num_ref_idx_l0_active; int max_l1_ref = currSlice->num_ref_idx_l1_active; if (p_Vid->active_pps->weighted_bipred_idc == 2) { currSlice->luma_log2_weight_denom = 5; currSlice->chroma_log2_weight_denom = 5; currSlice->wp_round_luma = 16; currSlice->wp_round_chroma = 16; for (i=0; iluma_log2_weight_denom : currSlice->chroma_log2_weight_denom; currSlice->wp_weight[0][i][comp] = 1<wp_weight[1][i][comp] = 1<wp_offset[0][i][comp] = 0; currSlice->wp_offset[1][i][comp] = 0; } } } for (i=0; iluma_log2_weight_denom : currSlice->chroma_log2_weight_denom; if (p_Vid->active_pps->weighted_bipred_idc == 1) { currSlice->wbp_weight[0][i][j][comp] = currSlice->wp_weight[0][i][comp]; currSlice->wbp_weight[1][i][j][comp] = currSlice->wp_weight[1][j][comp]; } else if (p_Vid->active_pps->weighted_bipred_idc == 2) { td = iClip3(-128,127,p_Vid->listX[LIST_1][j]->poc - p_Vid->listX[LIST_0][i]->poc); if (td == 0 || p_Vid->listX[LIST_1][j]->is_long_term || p_Vid->listX[LIST_0][i]->is_long_term) { currSlice->wbp_weight[0][i][j][comp] = 32; currSlice->wbp_weight[1][i][j][comp] = 32; } else { tb = iClip3(-128,127,p_Vid->ThisPOC - p_Vid->listX[LIST_0][i]->poc); tx = (16384 + iabs(td/2))/td; DistScaleFactor = iClip3(-1024, 1023, (tx*tb + 32 )>>6); currSlice->wbp_weight[1][i][j][comp] = DistScaleFactor >> 2; currSlice->wbp_weight[0][i][j][comp] = 64 - currSlice->wbp_weight[1][i][j][comp]; if (currSlice->wbp_weight[1][i][j][comp] < -64 || currSlice->wbp_weight[1][i][j][comp] > 128) { currSlice->wbp_weight[0][i][j][comp] = 32; currSlice->wbp_weight[1][i][j][comp] = 32; currSlice->wp_offset[0][i][comp] = 0; currSlice->wp_offset[1][j][comp] = 0; } } } } } } if (currSlice->mb_aff_frame_flag) { for (i=0; i<2*max_l0_ref; ++i) { for (j=0; j<2*max_l1_ref; ++j) { for (comp = 0; comp<3; ++comp) { for (k=2; k<6; k+=2) { currSlice->wp_offset[k+0][i][comp] = currSlice->wp_offset[0][i>>1][comp]; currSlice->wp_offset[k+1][j][comp] = currSlice->wp_offset[1][j>>1][comp]; log_weight_denom = (comp == 0) ? currSlice->luma_log2_weight_denom : currSlice->chroma_log2_weight_denom; if (p_Vid->active_pps->weighted_bipred_idc == 1) { currSlice->wbp_weight[k+0][i][j][comp] = currSlice->wp_weight[0][i>>1][comp]; currSlice->wbp_weight[k+1][i][j][comp] = currSlice->wp_weight[1][j>>1][comp]; } else if (p_Vid->active_pps->weighted_bipred_idc == 2) { td = iClip3(-128,127,p_Vid->listX[k+LIST_1][j]->poc - p_Vid->listX[k+LIST_0][i]->poc); if (td == 0 || p_Vid->listX[k+LIST_1][j]->is_long_term || p_Vid->listX[k+LIST_0][i]->is_long_term) { currSlice->wbp_weight[k+0][i][j][comp] = 32; currSlice->wbp_weight[k+1][i][j][comp] = 32; } else { tb = iClip3(-128,127,((k==2)?p_Vid->toppoc:p_Vid->bottompoc) - p_Vid->listX[k+LIST_0][i]->poc); tx = (16384 + iabs(td/2))/td; DistScaleFactor = iClip3(-1024, 1023, (tx*tb + 32 )>>6); currSlice->wbp_weight[k+1][i][j][comp] = DistScaleFactor >> 2; currSlice->wbp_weight[k+0][i][j][comp] = 64 - currSlice->wbp_weight[k+1][i][j][comp]; if (currSlice->wbp_weight[k+1][i][j][comp] < -64 || currSlice->wbp_weight[k+1][i][j][comp] > 128) { currSlice->wbp_weight[k+1][i][j][comp] = 32; currSlice->wbp_weight[k+0][i][j][comp] = 32; currSlice->wp_offset[k+0][i][comp] = 0; currSlice->wp_offset[k+1][j][comp] = 0; } } } } } } } } } static void decode_slice(Slice *currSlice, int current_header) { VideoParameters *p_Vid = currSlice->p_Vid; if (p_Vid->active_pps->entropy_coding_mode_flag) { init_contexts (currSlice); cabac_new_slice(currSlice); } if ( (p_Vid->active_pps->weighted_bipred_idc > 0 && (currSlice->slice_type == B_SLICE)) || (p_Vid->active_pps->weighted_pred_flag && currSlice->slice_type !=I_SLICE)) fill_wp_params(currSlice); //printf("frame picture %d %d %d\n",p_Vid->structure,p_Vid->ThisPOC,currSlice->direct_spatial_mv_pred_flag); // decode main slice information if (current_header == SOP || current_header == SOS) decode_one_slice(currSlice); // setMB-Nr in case this slice was lost // if(currSlice->ei_flag) // p_Vid->current_mb_nr = currSlice->last_mb_nr + 1; } /*! ************************************************************************ * \brief * Error tracking: if current frame is lost or any reference frame of * current frame is lost, current frame is incorrect. ************************************************************************ */ static void Error_tracking(VideoParameters *p_Vid) { Slice *currSlice = p_Vid->currentSlice; int i; if(p_Vid->redundant_pic_cnt == 0) { p_Vid->Is_primary_correct = p_Vid->Is_redundant_correct = 1; } if(p_Vid->redundant_pic_cnt == 0 && p_Vid->type != I_SLICE) { for(i=0;inum_ref_idx_l0_active;++i) { if(p_Vid->ref_flag[i] == 0) // any reference of primary slice is incorrect { p_Vid->Is_primary_correct = 0; // primary slice is incorrect } } } else if(p_Vid->redundant_pic_cnt != 0 && p_Vid->type != I_SLICE) { if(p_Vid->ref_flag[p_Vid->redundant_slice_ref_idx] == 0) // reference of redundant slice is incorrect { p_Vid->Is_redundant_correct = 0; // redundant slice is incorrect } } } /*! *********************************************************************** * \brief * decodes one I- or P-frame * *********************************************************************** */ int decode_one_frame(VideoParameters *p_Vid, uint64_t time_code) { InputParameters *p_Inp = p_Vid->p_Inp; int current_header; int i; while (1) { Slice *currSlice = p_Vid->currentSlice; currSlice->p_Vid = p_Vid; currSlice->p_Inp = p_Inp; current_header = read_new_slice(p_Vid->currentSlice, time_code); if (current_header == EOS) return EOS; // error tracking of primary and redundant slices. Error_tracking(p_Vid); // If primary and redundant are received and primary is correct, discard the redundant // else, primary slice will be replaced with redundant slice. if(p_Vid->frame_num == p_Vid->previous_frame_num && p_Vid->redundant_pic_cnt !=0 && p_Vid->Is_primary_correct !=0 && current_header != EOS) { continue; } // update reference flags and set current p_Vid->ref_flag if(!(p_Vid->redundant_pic_cnt != 0 && p_Vid->previous_frame_num == p_Vid->frame_num)) { for(i=16;i>0;i--) { p_Vid->ref_flag[i] = p_Vid->ref_flag[i-1]; } } p_Vid->ref_flag[0] = p_Vid->redundant_pic_cnt==0 ? p_Vid->Is_primary_correct : p_Vid->Is_redundant_correct; p_Vid->previous_frame_num = p_Vid->frame_num; if (current_header == EOS) { // exit_picture(p_Vid, &p_Vid->dec_picture); return EOS; } if((p_Vid->active_sps->chroma_format_idc==0)||(p_Vid->active_sps->chroma_format_idc==3)) { currSlice->linfo_cbp_intra = linfo_cbp_intra_other; currSlice->linfo_cbp_inter = linfo_cbp_inter_other; } else { currSlice->linfo_cbp_intra = linfo_cbp_intra_normal; currSlice->linfo_cbp_inter = linfo_cbp_inter_normal; } decode_slice(currSlice, current_header); ++(p_Vid->current_slice_nr); } exit_picture(p_Vid, &p_Vid->dec_picture); return (SOP); } /*! ************************************************************************ * \brief * Convert file read buffer to source picture structure * \param imgX * Pointer to image plane * \param buf * Buffer for file output * \param size_x * horizontal image size in pixel * \param size_y * vertical image size in pixel * \param symbol_size_in_bytes * number of bytes used per pel ************************************************************************ */ void buffer2img (imgpel** imgX, unsigned char* buf, int size_x, int size_y, int symbol_size_in_bytes) { int i,j; uint16 tmp16, ui16; unsigned long tmp32, ui32; if (symbol_size_in_bytes> sizeof(imgpel)) { error ("Source picture has higher bit depth than imgpel data type. \nPlease recompile with larger data type for imgpel.", 500); } if (( sizeof(char) == sizeof (imgpel)) && ( sizeof(char) == symbol_size_in_bytes)) { // imgpel == pixel_in_file == 1 byte -> simple copy memcpy(&imgX[0][0], buf, size_x * size_y); } else { // sizeof (imgpel) > sizeof(char) if (testEndian()) { // big endian switch (symbol_size_in_bytes) { case 1: { for(j = 0; j < size_y; ++j) for(i = 0; i < size_x; ++i) { imgX[j][i]= buf[i+j*size_x]; } break; } case 2: { for(j=0;j> 8) | ((tmp16&0xFF)<<8)); imgX[j][i] = (imgpel) ui16; } break; } case 4: { for(j=0;j>8) | ((tmp32&0xFF000000)>>24); imgX[j][i] = (imgpel) ui32; } } default: { error ("reading only from formats of 8, 16 or 32 bit allowed on big endian architecture", 500); break; } } } else { // little endian if (symbol_size_in_bytes == 1) { for (j=0; j < size_y; ++j) { for (i=0; i < size_x; ++i) { imgX[j][i]=*(buf++); } } } else { for (j=0; j < size_y; ++j) { int jpos = j*size_x; for (i=0; i < size_x; ++i) { imgX[j][i]=0; memcpy(&(imgX[j][i]), buf +((i+jpos)*symbol_size_in_bytes), symbol_size_in_bytes); } } } } } } void reorder_lists(Slice *currSlice) { VideoParameters *p_Vid = currSlice->p_Vid; if ((currSlice->slice_type != I_SLICE)&&(currSlice->slice_type != SI_SLICE)) { if (currSlice->ref_pic_list_reordering_flag_l0) { reorder_ref_pic_list(p_Vid, p_Vid->listX[0], &p_Vid->listXsize[0], currSlice->num_ref_idx_l0_active - 1, currSlice->reordering_of_pic_nums_idc_l0, currSlice->abs_diff_pic_num_minus1_l0, currSlice->long_term_pic_idx_l0); } if (p_Vid->no_reference_picture == p_Vid->listX[0][currSlice->num_ref_idx_l0_active-1]) { if (p_Vid->non_conforming_stream) printf("RefPicList0[ num_ref_idx_l0_active_minus1 ] is equal to 'no reference picture'\n"); else error("RefPicList0[ num_ref_idx_l0_active_minus1 ] is equal to 'no reference picture', invalid bitstream",500); } // that's a definition p_Vid->listXsize[0] = (char) currSlice->num_ref_idx_l0_active; } if (currSlice->slice_type == B_SLICE) { if (currSlice->ref_pic_list_reordering_flag_l1) { reorder_ref_pic_list(p_Vid, p_Vid->listX[1], &p_Vid->listXsize[1], currSlice->num_ref_idx_l1_active - 1, currSlice->reordering_of_pic_nums_idc_l1, currSlice->abs_diff_pic_num_minus1_l1, currSlice->long_term_pic_idx_l1); } if (p_Vid->no_reference_picture == p_Vid->listX[1][currSlice->num_ref_idx_l1_active-1]) { if (p_Vid->non_conforming_stream) printf("RefPicList1[ num_ref_idx_l1_active_minus1 ] is equal to 'no reference picture'\n"); else error("RefPicList1[ num_ref_idx_l1_active_minus1 ] is equal to 'no reference picture', invalid bitstream",500); } // that's a definition p_Vid->listXsize[1] = (char) currSlice->num_ref_idx_l1_active; } free_ref_pic_list_reordering_buffer(currSlice); } /*! ************************************************************************ * \brief * initialize ref_pic_num array ************************************************************************ */ void set_ref_pic_num(VideoParameters *p_Vid) { int i,j; StorablePicture *dec_picture = p_Vid->dec_picture; int slice_id=p_Vid->current_slice_nr; for (i=0;ilistXsize[LIST_0];++i) { dec_picture->ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->poc * 2 + ((p_Vid->listX[LIST_0][i]->structure==BOTTOM_FIELD)?1:0) ; dec_picture->frm_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->frame_poc * 2; dec_picture->top_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->top_poc * 2; dec_picture->bottom_ref_pic_num [slice_id][LIST_0][i] = p_Vid->listX[LIST_0][i]->bottom_poc * 2 + 1; //printf("POCS %d %d %d %d ",p_Vid->listX[LIST_0][i]->frame_poc,p_Vid->listX[LIST_0][i]->bottom_poc,p_Vid->listX[LIST_0][i]->top_poc,p_Vid->listX[LIST_0][i]->poc); //printf("refid %d %d %d %d\n",(int) dec_picture->frm_ref_pic_num[LIST_0][i],(int) dec_picture->top_ref_pic_num[LIST_0][i],(int) dec_picture->bottom_ref_pic_num[LIST_0][i],(int) dec_picture->ref_pic_num[LIST_0][i]); } for (i=0;ilistXsize[LIST_1];++i) { dec_picture->ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->poc *2 + ((p_Vid->listX[LIST_1][i]->structure==BOTTOM_FIELD)?1:0); dec_picture->frm_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->frame_poc * 2; dec_picture->top_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->top_poc * 2; dec_picture->bottom_ref_pic_num [slice_id][LIST_1][i] = p_Vid->listX[LIST_1][i]->bottom_poc * 2 + 1; } if (!p_Vid->active_sps->frame_mbs_only_flag) { if (p_Vid->structure==FRAME) { for (j=2;j<6;++j) { for (i=0;ilistXsize[j];++i) { dec_picture->ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->poc * 2 + ((p_Vid->listX[j][i]->structure==BOTTOM_FIELD)?1:0); dec_picture->frm_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->frame_poc * 2 ; dec_picture->top_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->top_poc * 2 ; dec_picture->bottom_ref_pic_num [slice_id][j][i] = p_Vid->listX[j][i]->bottom_poc * 2 + 1; } } } } } /*! ************************************************************************ * \brief * Reads new slice from bit_stream ************************************************************************ */ // benski> time_code is user-passed data static int read_new_slice(Slice *currSlice, uint64_t time_code) { VideoParameters *p_Vid = currSlice->p_Vid; InputParameters *p_Inp = currSlice->p_Inp; NALU_t *nalu = p_Vid->nalu; int current_header = 0; Bitstream *currStream; int slice_id_a, slice_id_b, slice_id_c; int redundant_pic_cnt_b, redundant_pic_cnt_c; if (!nalu) return EOS; for (;;) { if (0 == read_next_nalu(p_Vid, nalu)) return EOS; process_nalu: switch (nalu->nal_unit_type) { case NALU_TYPE_SLICE: case NALU_TYPE_IDR: if (p_Vid->recovery_point || nalu->nal_unit_type == NALU_TYPE_IDR) { if (p_Vid->recovery_point_found == 0) { if (nalu->nal_unit_type != NALU_TYPE_IDR) { printf("Warning: Decoding does not start with an IDR picture.\n"); p_Vid->non_conforming_stream = 1; } else p_Vid->non_conforming_stream = 0; } p_Vid->recovery_point_found = 1; } if (p_Vid->recovery_point_found == 0) break; p_Vid->idr_flag = (nalu->nal_unit_type == NALU_TYPE_IDR); p_Vid->nal_reference_idc = nalu->nal_reference_idc; currSlice->dp_mode = PAR_DP_1; currSlice->max_part_nr = 1; currStream = currSlice->partArr[0].bitstream; currStream->frame_bitoffset = currStream->read_len = 0; //memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1); currStream->streamBuffer = &nalu->buf[1]; currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1); // Some syntax of the Slice Header depends on the parameter set, which depends on // the parameter set ID of the SLice header. Hence, read the pic_parameter_set_id // of the slice header first, then setup the active parameter sets, and then read // the rest of the slice header FirstPartOfSliceHeader(currSlice); UseParameterSet (currSlice, currSlice->pic_parameter_set_id); RestOfSliceHeader (currSlice); fmo_init (p_Vid); currSlice->active_sps = p_Vid->active_sps; currSlice->active_pps = p_Vid->active_pps; assign_quant_params (currSlice); // if primary slice is replaced with redundant slice, set the correct image type if(p_Vid->redundant_pic_cnt && p_Vid->Is_primary_correct==0 && p_Vid->Is_redundant_correct) { p_Vid->dec_picture->slice_type = p_Vid->type; } if(is_new_picture(p_Vid->dec_picture, currSlice, p_Vid->old_slice)) { init_picture(p_Vid, currSlice, p_Inp, time_code); current_header = SOP; //check zero_byte if it is also the first NAL unit in the access unit CheckZeroByteVCL(p_Vid, nalu); } else current_header = SOS; if (currSlice->slice_type == B_SLICE) { if( IS_INDEPENDENT(p_Vid) ) { int nplane = 0; for( nplane=0; nplaneCo_located_JV[nplane] ) { free_colocated(p_Vid, currSlice->Co_located_JV[nplane]); currSlice->Co_located_JV[nplane] = NULL; } currSlice->Co_located_JV[nplane] = alloc_colocated (p_Vid, p_Vid->width, p_Vid->height, p_Vid->active_sps->mb_adaptive_frame_field_flag); } } else { if (NULL != currSlice->p_colocated) { free_colocated(p_Vid, currSlice->p_colocated); currSlice->p_colocated = NULL; } currSlice->p_colocated = alloc_colocated (p_Vid, p_Vid->width, p_Vid->height,p_Vid->active_sps->mb_adaptive_frame_field_flag); } } init_lists(currSlice); reorder_lists (currSlice); if (p_Vid->structure==FRAME) { init_mbaff_lists(p_Vid); } // From here on, p_Vid->active_sps, p_Vid->active_pps and the slice header are valid if (currSlice->mb_aff_frame_flag) p_Vid->current_mb_nr = currSlice->start_mb_nr << 1; else p_Vid->current_mb_nr = currSlice->start_mb_nr; if (p_Vid->active_pps->entropy_coding_mode_flag) { int ByteStartPosition = currStream->frame_bitoffset/8; if (currStream->frame_bitoffset%8 != 0) { ++ByteStartPosition; } arideco_start_decoding (&currSlice->partArr[0].de_cabac, currStream->streamBuffer, ByteStartPosition, &currStream->read_len); } // printf ("read_new_slice: returning %s\n", current_header == SOP?"SOP":"SOS"); p_Vid->recovery_point = 0; return current_header; break; case NALU_TYPE_DPA: // read DP_A currSlice->dpB_NotPresent =1; currSlice->dpC_NotPresent =1; p_Vid->idr_flag = FALSE; p_Vid->nal_reference_idc = nalu->nal_reference_idc; currSlice->dp_mode = PAR_DP_3; currSlice->max_part_nr = 3; currStream = currSlice->partArr[0].bitstream; currStream->frame_bitoffset = currStream->read_len = 0; //memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1); currStream->streamBuffer = &nalu->buf[1]; currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1); FirstPartOfSliceHeader(currSlice); UseParameterSet (currSlice, currSlice->pic_parameter_set_id); RestOfSliceHeader (currSlice); fmo_init (p_Vid); if(is_new_picture(p_Vid->dec_picture, currSlice, p_Vid->old_slice)) { init_picture(p_Vid, currSlice, p_Inp, time_code); current_header = SOP; CheckZeroByteVCL(p_Vid, nalu); } else current_header = SOS; init_lists(currSlice); reorder_lists (currSlice); if (p_Vid->structure==FRAME) { init_mbaff_lists(p_Vid); } // From here on, p_Vid->active_sps, p_Vid->active_pps and the slice header are valid if (currSlice->mb_aff_frame_flag) p_Vid->current_mb_nr = currSlice->start_mb_nr << 1; else p_Vid->current_mb_nr = currSlice->start_mb_nr; // Now I need to read the slice ID, which depends on the value of // redundant_pic_cnt_present_flag slice_id_a = ue_v("NALU: DP_A slice_id", currStream); if (p_Vid->active_pps->entropy_coding_mode_flag) error ("received data partition with CABAC, this is not allowed", 500); // continue with reading next DP if (0 == read_next_nalu(p_Vid, nalu)) return current_header; if ( NALU_TYPE_DPB == nalu->nal_unit_type) { // we got a DPB currStream = currSlice->partArr[1].bitstream; currStream->frame_bitoffset = currStream->read_len = 0; //memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1); currStream->streamBuffer = &nalu->buf[1]; currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1); slice_id_b = ue_v("NALU: DP_B slice_id", currStream); currSlice->dpB_NotPresent = 0; if ((slice_id_b != slice_id_a) || (nalu->lost_packets)) { printf ("Waning: got a data partition B which does not match DP_A (DP loss!)\n"); currSlice->dpB_NotPresent =1; currSlice->dpC_NotPresent =1; } else { if (p_Vid->active_pps->redundant_pic_cnt_present_flag) redundant_pic_cnt_b = ue_v("NALU: DP_B redudant_pic_cnt", currStream); else redundant_pic_cnt_b = 0; // we're finished with DP_B, so let's continue with next DP if (0 == read_next_nalu(p_Vid, nalu)) return current_header; } } else { currSlice->dpB_NotPresent =1; } // check if we got DP_C if ( NALU_TYPE_DPC == nalu->nal_unit_type) { currStream = currSlice->partArr[2].bitstream; currStream->frame_bitoffset = currStream->read_len = 0; //memcpy (currStream->streamBuffer, &nalu->buf[1], nalu->len-1); currStream->streamBuffer = &nalu->buf[1]; currStream->code_len = currStream->bitstream_length = RBSPtoSODB(currStream->streamBuffer, nalu->len-1); currSlice->dpC_NotPresent = 0; slice_id_c = ue_v("NALU: DP_C slice_id", currStream); if ((slice_id_c != slice_id_a)|| (nalu->lost_packets)) { printf ("Warning: got a data partition C which does not match DP_A(DP loss!)\n"); //currSlice->dpB_NotPresent =1; currSlice->dpC_NotPresent =1; } if (p_Vid->active_pps->redundant_pic_cnt_present_flag) redundant_pic_cnt_c = ue_v("NALU:SLICE_C redudand_pic_cnt", currStream); else redundant_pic_cnt_c = 0; } else { currSlice->dpC_NotPresent =1; } // check if we read anything else than the expected partitions if ((nalu->nal_unit_type != NALU_TYPE_DPB) && (nalu->nal_unit_type != NALU_TYPE_DPC)) { // we have a NALI that we can't process here, so restart processing goto process_nalu; // yes, "goto" should not be used, but it's really the best way here before we restructure the decoding loop // (which should be taken care of anyway) } return current_header; break; case NALU_TYPE_DPB: printf ("found data partition B without matching DP A, discarding\n"); break; case NALU_TYPE_DPC: printf ("found data partition C without matching DP A, discarding\n"); break; case NALU_TYPE_SEI: //printf ("read_new_slice: Found NALU_TYPE_SEI, len %d\n", nalu->len); InterpretSEIMessage(nalu->buf,nalu->len,p_Vid); break; case NALU_TYPE_PPS: ProcessPPS(p_Vid, nalu); break; case NALU_TYPE_SPS: ProcessSPS(p_Vid, nalu); break; case NALU_TYPE_AUD: // printf ("read_new_slice: Found 'Access Unit Delimiter' NAL unit, len %d, ignored\n", nalu->len); break; case NALU_TYPE_EOSEQ: // printf ("read_new_slice: Found 'End of Sequence' NAL unit, len %d, ignored\n", nalu->len); break; case NALU_TYPE_EOSTREAM: // printf ("read_new_slice: Found 'End of Stream' NAL unit, len %d, ignored\n", nalu->len); break; case NALU_TYPE_FILL: printf ("read_new_slice: Found NALU_TYPE_FILL, len %d\n", (int) nalu->len); printf ("Skipping these filling bits, proceeding w/ next NALU\n"); break; default: printf ("Found NALU type %d, len %d undefined, ignore NALU, moving on\n", (int) nalu->nal_unit_type, (int) nalu->len); break; } } return current_header; } /*! ************************************************************************ * \brief * finish decoding of a picture, conceal errors and store it * into the DPB ************************************************************************ */ void exit_picture(VideoParameters *p_Vid, StorablePicture **dec_picture) { InputParameters *p_Inp = p_Vid->p_Inp; int ercStartMB; int ercSegment; frame recfr; unsigned int i; int structure, frame_poc, slice_type, refpic, qp, pic_num, chroma_format_idc, is_idr; int nplane; // return if the last picture has already been finished if (*dec_picture==NULL) { return; } recfr.p_Vid = p_Vid; recfr.yptr = (*dec_picture)->imgY->base_address; if ((*dec_picture)->chroma_format_idc != YUV400) { recfr.uptr = (*dec_picture)->imgUV[0]->base_address; recfr.vptr = (*dec_picture)->imgUV[1]->base_address; } //! this is always true at the beginning of a picture ercStartMB = 0; ercSegment = 0; //! mark the start of the first segment if (!(*dec_picture)->mb_aff_frame_flag) { ercStartSegment(0, ercSegment, 0 , p_Vid->erc_errorVar); //! generate the segments according to the macroblock map for(i = 1; i<(*dec_picture)->PicSizeInMbs; ++i) { if(p_Vid->mb_data[i].ei_flag != p_Vid->mb_data[i-1].ei_flag) { ercStopSegment(i-1, ercSegment, 0, p_Vid->erc_errorVar); //! stop current segment //! mark current segment as lost or OK if(p_Vid->mb_data[i-1].ei_flag) ercMarkCurrSegmentLost((*dec_picture)->size_x, p_Vid->erc_errorVar); else ercMarkCurrSegmentOK((*dec_picture)->size_x, p_Vid->erc_errorVar); ++ercSegment; //! next segment ercStartSegment(i, ercSegment, 0 , p_Vid->erc_errorVar); //! start new segment ercStartMB = i;//! save start MB for this segment } } //! mark end of the last segment ercStopSegment((*dec_picture)->PicSizeInMbs-1, ercSegment, 0, p_Vid->erc_errorVar); if(p_Vid->mb_data[i-1].ei_flag) ercMarkCurrSegmentLost((*dec_picture)->size_x, p_Vid->erc_errorVar); else ercMarkCurrSegmentOK((*dec_picture)->size_x, p_Vid->erc_errorVar); //! call the right error concealment function depending on the frame type. p_Vid->erc_mvperMB /= (*dec_picture)->PicSizeInMbs; p_Vid->erc_img = p_Vid; if((*dec_picture)->slice_type == I_SLICE || (*dec_picture)->slice_type == SI_SLICE) // I-frame ercConcealIntraFrame(p_Vid, &recfr, (*dec_picture)->size_x, (*dec_picture)->size_y, p_Vid->erc_errorVar); else ercConcealInterFrame(&recfr, p_Vid->erc_object_list, (*dec_picture)->size_x, (*dec_picture)->size_y, p_Vid->erc_errorVar, (*dec_picture)->chroma_format_idc); } //deblocking for frame or field if( IS_INDEPENDENT(p_Vid) ) { int colour_plane_id = p_Vid->colour_plane_id; for( nplane=0; nplanecolour_plane_id = colour_plane_id; make_frame_picture_JV(p_Vid); } else { DeblockPicture( p_Vid, *dec_picture ); } if ((*dec_picture)->mb_aff_frame_flag) MbAffPostProc(p_Vid); if (p_Vid->structure == FRAME) // buffer mgt. for frame mode frame_postprocessing(p_Vid); else field_postprocessing(p_Vid); // reset all interlaced variables structure = (*dec_picture)->structure; slice_type = (*dec_picture)->slice_type; frame_poc = (*dec_picture)->frame_poc; refpic = (*dec_picture)->used_for_reference; qp = (*dec_picture)->qp; pic_num = (*dec_picture)->pic_num; is_idr = (*dec_picture)->idr_flag; chroma_format_idc = (*dec_picture)->chroma_format_idc; store_picture_in_dpb(p_Vid, *dec_picture); *dec_picture=NULL; if (p_Vid->last_has_mmco_5) { p_Vid->pre_frame_num = 0; } if ((structure==FRAME)||structure==BOTTOM_FIELD) { if(slice_type == I_SLICE || slice_type == SI_SLICE || slice_type == P_SLICE || refpic) // I or P pictures ++(p_Vid->number); } p_Vid->current_mb_nr = -4712; // impossible value for debugging, StW p_Vid->current_slice_nr = 0; } /*! ************************************************************************ * \brief * write the encoding mode and motion vectors of current * MB to the buffer of the error concealment module. ************************************************************************ */ void ercWriteMBMODEandMV(Macroblock *currMB) { VideoParameters *p_Vid = currMB->p_Vid; int i, ii, jj, currMBNum = p_Vid->current_mb_nr; StorablePicture *dec_picture = p_Vid->dec_picture; int mbx = 4*xPosMB(currMBNum, dec_picture->size_x), mby = 4*yPosMB(currMBNum, dec_picture->size_x); objectBuffer_t *currRegion, *pRegion; currRegion = p_Vid->erc_object_list + (currMBNum<<2); if(p_Vid->type != B_SLICE) //non-B frame { for (i=0; i<4; ++i) { pRegion = currRegion + i; pRegion->regionMode = (currMB->mb_type ==I16MB ? REGMODE_INTRA : currMB->b8mode[i]==IBLOCK ? REGMODE_INTRA_8x8 : currMB->b8mode[i]==0 ? REGMODE_INTER_COPY : currMB->b8mode[i]==1 ? REGMODE_INTER_PRED : REGMODE_INTER_PRED_8x8); if (currMB->b8mode[i]==0 || currMB->b8mode[i]==IBLOCK) // INTRA OR COPY { pRegion->mv[0] = 0; pRegion->mv[1] = 0; pRegion->mv[2] = 0; } else { PicMotion *motion00; PicMotion *motion01; ii = mbx + (i & 0x01)*2;// + BLOCK_SIZE; jj = mby + (i & ~1 ); motion00 = &dec_picture->motion.motion[LIST_0][jj][ii]; if (currMB->b8mode[i]>=5 && currMB->b8mode[i]<=7) // SMALL BLOCKS { motion01 = &dec_picture->motion.motion[LIST_0][jj+1][ii]; pRegion->mv[0] = (motion00[0].mv[0] + motion00[1].mv[0] + motion01[0].mv[0] + motion01[1].mv[0] + 2)/4; pRegion->mv[1] = (motion00[0].mv[1] + motion00[1].mv[1] + motion01[0].mv[1] + motion01[1].mv[1] + 2)/4; } else // 16x16, 16x8, 8x16, 8x8 { pRegion->mv[0] = motion00->mv[0]; pRegion->mv[1] = motion00->mv[1]; } p_Vid->erc_mvperMB += abs(pRegion->mv[0]) + abs(pRegion->mv[1]); pRegion->mv[2] = motion00->ref_idx; } } } else //B-frame { for (i=0; i<4; ++i) { ii = mbx + (i%2)*2;// + BLOCK_SIZE; jj = mby + (i/2)*2; pRegion = currRegion + i; pRegion->regionMode = (currMB->mb_type ==I16MB ? REGMODE_INTRA : currMB->b8mode[i]==IBLOCK ? REGMODE_INTRA_8x8 : REGMODE_INTER_PRED_8x8); if (currMB->mb_type==I16MB || currMB->b8mode[i]==IBLOCK) // INTRA { pRegion->mv[0] = 0; pRegion->mv[1] = 0; pRegion->mv[2] = 0; } else { PicMotion *motion0= & dec_picture->motion.motion[LIST_0][jj][ii]; PicMotion *motion1; if (motion0->ref_idx<0) { motion0 = & dec_picture->motion.motion[LIST_1][jj][ii]; motion1 = & dec_picture->motion.motion[LIST_1][jj+1][ii]; } else { motion1 = & dec_picture->motion.motion[LIST_0][jj+1][ii]; } // int idx = (currMB->b8mode[i]==0 && currMB->b8pdir[i]==2 ? LIST_0 : currMB->b8pdir[i]==1 ? LIST_1 : LIST_0); // int idx = currMB->b8pdir[i]==0 ? LIST_0 : LIST_1; pRegion->mv[0] = (motion0[0].mv[0] + motion0[1].mv[0] + motion1[0].mv[0] + motion1[1].mv[0] + 2)/4; pRegion->mv[1] = (motion0[0].mv[1] + motion0[1].mv[1] + motion1[0].mv[1] + motion1[1].mv[1] + 2)/4; p_Vid->erc_mvperMB += iabs(pRegion->mv[0]) + iabs(pRegion->mv[1]); pRegion->mv[2] = (motion0[0].ref_idx); /* if (currMB->b8pdir[i]==0 || (currMB->b8pdir[i]==2 && currMB->b8mode[i]!=0)) // forward or bidirect { pRegion->mv[2] = (dec_picture->motion.ref_idx[LIST_0][jj][ii]); ///???? is it right, not only "p_Vid->fw_refFrArr[jj][ii-4]" } else { pRegion->mv[2] = (dec_picture->motion.ref_idx[LIST_1][jj][ii]); // pRegion->mv[2] = 0; } */ } } } } /*! ************************************************************************ * \brief * set defaults for old_slice * NAL unit of a picture" ************************************************************************ */ void init_old_slice(OldSliceParams *p_old_slice) { p_old_slice->field_pic_flag = 0; p_old_slice->pps_id = INT_MAX; p_old_slice->frame_num = INT_MAX; p_old_slice->nal_ref_idc = INT_MAX; p_old_slice->idr_flag = FALSE; p_old_slice->pic_oder_cnt_lsb = UINT_MAX; p_old_slice->delta_pic_oder_cnt_bottom = INT_MAX; p_old_slice->delta_pic_order_cnt[0] = INT_MAX; p_old_slice->delta_pic_order_cnt[1] = INT_MAX; } /*! ************************************************************************ * \brief * save slice parameters that are needed for checking of "first VCL * NAL unit of a picture" ************************************************************************ */ static void exit_slice(Slice *currSlice, OldSliceParams *p_old_slice) { VideoParameters *p_Vid = currSlice->p_Vid; if (currSlice->slice_type == B_SLICE) { if( IS_INDEPENDENT(p_Vid) ) { int nplane; for( nplane = 0; nplane < MAX_PLANE; ++nplane ) { free_colocated(p_Vid, currSlice->Co_located_JV[nplane]); currSlice->Co_located_JV[nplane] = NULL; } } else { free_colocated(p_Vid, currSlice->p_colocated); currSlice->p_colocated = NULL; } } p_old_slice->pps_id = currSlice->pic_parameter_set_id; p_old_slice->frame_num = p_Vid->frame_num; p_old_slice->field_pic_flag = p_Vid->field_pic_flag; if(p_Vid->field_pic_flag) { p_old_slice->bottom_field_flag = p_Vid->bottom_field_flag; } p_old_slice->nal_ref_idc = p_Vid->nal_reference_idc; p_old_slice->idr_flag = (byte) p_Vid->idr_flag; if (p_Vid->idr_flag) { p_old_slice->idr_pic_id = p_Vid->idr_pic_id; } if (p_Vid->active_sps->pic_order_cnt_type == 0) { p_old_slice->pic_oder_cnt_lsb = p_Vid->pic_order_cnt_lsb; p_old_slice->delta_pic_oder_cnt_bottom = p_Vid->delta_pic_order_cnt_bottom; } if (p_Vid->active_sps->pic_order_cnt_type == 1) { p_old_slice->delta_pic_order_cnt[0] = p_Vid->delta_pic_order_cnt[0]; p_old_slice->delta_pic_order_cnt[1] = p_Vid->delta_pic_order_cnt[1]; } } /*! ************************************************************************ * \brief * detect if current slice is "first VCL NAL unit of a picture" ************************************************************************ */ int is_new_picture(StorablePicture *dec_picture, Slice *currSlice, OldSliceParams *p_old_slice) { VideoParameters *p_Vid = currSlice->p_Vid; int result=0; result |= (NULL==dec_picture); result |= (p_old_slice->pps_id != currSlice->pic_parameter_set_id); result |= (p_old_slice->frame_num != p_Vid->frame_num); result |= (p_old_slice->field_pic_flag != p_Vid->field_pic_flag); if(p_Vid->field_pic_flag && p_old_slice->field_pic_flag) { result |= (p_old_slice->bottom_field_flag != p_Vid->bottom_field_flag); } result |= (p_old_slice->nal_ref_idc != p_Vid->nal_reference_idc) && ((p_old_slice->nal_ref_idc == 0) || (p_Vid->nal_reference_idc == 0)); result |= ( p_old_slice->idr_flag != p_Vid->idr_flag); if (p_Vid->idr_flag && p_old_slice->idr_flag) { result |= (p_old_slice->idr_pic_id != p_Vid->idr_pic_id); } if (p_Vid->active_sps->pic_order_cnt_type == 0) { result |= (p_old_slice->pic_oder_cnt_lsb != p_Vid->pic_order_cnt_lsb); result |= (p_old_slice->delta_pic_oder_cnt_bottom != p_Vid->delta_pic_order_cnt_bottom); } if (p_Vid->active_sps->pic_order_cnt_type == 1) { result |= (p_old_slice->delta_pic_order_cnt[0] != p_Vid->delta_pic_order_cnt[0]); result |= (p_old_slice->delta_pic_order_cnt[1] != p_Vid->delta_pic_order_cnt[1]); } return result; } /*! ************************************************************************ * \brief * Prepare field and frame buffer after frame decoding ************************************************************************ */ void frame_postprocessing(VideoParameters *p_Vid) { } /*! ************************************************************************ * \brief * Prepare field and frame buffer after field decoding ************************************************************************ */ void field_postprocessing(VideoParameters *p_Vid) { p_Vid->number /= 2; } /*! ************************************************************************ * \brief * copy StorablePicture *src -> StorablePicture *dst * for 4:4:4 Independent mode ************************************************************************ */ void copy_dec_picture_JV( VideoParameters *p_Vid, StorablePicture *dst, StorablePicture *src ) { dst->top_poc = src->top_poc; dst->bottom_poc = src->bottom_poc; dst->frame_poc = src->frame_poc; dst->qp = src->qp; dst->slice_qp_delta = src->slice_qp_delta; dst->chroma_qp_offset[0] = src->chroma_qp_offset[0]; dst->chroma_qp_offset[1] = src->chroma_qp_offset[1]; dst->poc = src->poc; dst->slice_type = src->slice_type; dst->used_for_reference = src->used_for_reference; dst->idr_flag = src->idr_flag; dst->no_output_of_prior_pics_flag = src->no_output_of_prior_pics_flag; dst->long_term_reference_flag = src->long_term_reference_flag; dst->adaptive_ref_pic_buffering_flag = src->adaptive_ref_pic_buffering_flag; dst->dec_ref_pic_marking_buffer = src->dec_ref_pic_marking_buffer; dst->mb_aff_frame_flag = src->mb_aff_frame_flag; dst->PicWidthInMbs = src->PicWidthInMbs; dst->pic_num = src->pic_num; dst->frame_num = src->frame_num; dst->recovery_frame = src->recovery_frame; dst->coded_frame = src->coded_frame; dst->chroma_format_idc = src->chroma_format_idc; dst->frame_mbs_only_flag = src->frame_mbs_only_flag; dst->frame_cropping_flag = src->frame_cropping_flag; dst->frame_cropping_rect_left_offset = src->frame_cropping_rect_left_offset; dst->frame_cropping_rect_right_offset = src->frame_cropping_rect_right_offset; dst->frame_cropping_rect_top_offset = src->frame_cropping_rect_top_offset; dst->frame_cropping_rect_bottom_offset = src->frame_cropping_rect_bottom_offset; #if (ENABLE_OUTPUT_TONEMAPPING) // store the necessary tone mapping sei into StorablePicture structure dst->seiHasTone_mapping = src->seiHasTone_mapping; dst->seiHasTone_mapping = src->seiHasTone_mapping; dst->tone_mapping_model_id = src->tone_mapping_model_id; dst->tonemapped_bit_depth = src->tonemapped_bit_depth; if( src->tone_mapping_lut ) { dst->tone_mapping_lut = malloc(sizeof(int)*(1 << p_Vid->seiToneMapping->coded_data_bit_depth)); if (NULL == dst->tone_mapping_lut) { no_mem_exit("copy_dec_picture_JV: tone_mapping_lut"); } memcpy(dst->tone_mapping_lut, src->tone_mapping_lut, sizeof(imgpel)*(1<seiToneMapping->coded_data_bit_depth)); } #endif } /*! ************************************************************************ * \brief * decodes one slice ************************************************************************ */ void decode_one_slice(Slice *currSlice) { VideoParameters *p_Vid = currSlice->p_Vid; Boolean end_of_slice = FALSE; Macroblock *currMB = NULL; p_Vid->cod_counter=-1; setup_slice_methods(currSlice); if( IS_INDEPENDENT(p_Vid) ) { change_plane_JV( p_Vid, p_Vid->colour_plane_id ); } set_ref_pic_num(p_Vid); if (currSlice->slice_type == B_SLICE) { currSlice->compute_colocated(currSlice, currSlice->p_colocated, p_Vid->listX); } //reset_ec_flags(p_Vid); while (end_of_slice == FALSE) // loop over macroblocks { #if TRACE fprintf(p_trace,"\n*********** POC: %i (I/P) MB: %i Slice: %i Type %d **********\n", p_Vid->ThisPOC, p_Vid->current_mb_nr, p_Vid->current_slice_nr, currSlice->slice_type); #endif // Initializes the current macroblock start_macroblock(currSlice, &currMB); // Get the syntax elements from the NAL currSlice->read_one_macroblock(currMB); decode_one_macroblock(currMB, p_Vid->dec_picture); if(currSlice->mb_aff_frame_flag && p_Vid->dec_picture->motion.mb_field[p_Vid->current_mb_nr]) { currSlice->num_ref_idx_l0_active >>= 1; currSlice->num_ref_idx_l1_active >>= 1; } ercWriteMBMODEandMV(currMB); end_of_slice = exit_macroblock(currSlice, (!currSlice->mb_aff_frame_flag||p_Vid->current_mb_nr%2)); } exit_slice(currSlice, p_Vid->old_slice); //reset_ec_flags(p_Vid); }