From e4bf0302a32bb253a53215fff533beec8e065dc4 Mon Sep 17 00:00:00 2001 From: Kostya Shishkov Date: Mon, 3 Jul 2006 02:38:08 +0000 Subject: 4-MV decoding support Originally committed as revision 5591 to svn://svn.ffmpeg.org/ffmpeg/trunk --- libavcodec/vc1.c | 449 +++++++++++++++++++++++++++++++++++++++++++++---------- 1 file changed, 368 insertions(+), 81 deletions(-) diff --git a/libavcodec/vc1.c b/libavcodec/vc1.c index e43f926866..95879cffe7 100644 --- a/libavcodec/vc1.c +++ b/libavcodec/vc1.c @@ -288,6 +288,7 @@ typedef struct VC1Context{ int codingset2; ///< index of current table set from 11.8 to use for chroma block decoding int pqindex; ///< raw pqindex used in coding set selection int a_avail, c_avail; + uint8_t *mb_type_base, *mb_type[3]; /** Luma compensation parameters */ @@ -845,10 +846,10 @@ static void vc1_v_overlap(uint8_t* src, int stride) c = src[0]; d = src[stride]; - src[-2*stride] = (7*a + d) >> 3; - src[-stride] = (-a + 7*b + c + d) >> 3; - src[0] = (a + b + 7*c - d) >> 3; - src[stride] = (a + 7*d) >> 3; + src[-2*stride] = clip_uint8((7*a + d + 3) >> 3); + src[-stride] = clip_uint8((-a + 7*b + c + d + 3) >> 3); + src[0] = clip_uint8((a + b + 7*c - d + 3) >> 3); + src[stride] = clip_uint8((a + 7*d + 3) >> 3); src++; } } @@ -863,10 +864,10 @@ static void vc1_h_overlap(uint8_t* src, int stride) c = src[0]; d = src[1]; - src[-2] = (7*a + d) >> 3; - src[-1] = (-a + 7*b + c + d) >> 3; - src[0] = (a + b + 7*c - d) >> 3; - src[1] = (a + 7*d) >> 3; + src[-2] = clip_uint8((7*a + d + 3) >> 3); + src[-1] = clip_uint8((-a + 7*b + c + d + 3) >> 3); + src[0] = clip_uint8((a + b + 7*c - d + 3) >> 3); + src[1] = clip_uint8((a + 7*d + 3) >> 3); src += stride; } } @@ -973,6 +974,149 @@ static void vc1_mc_1mv(VC1Context *v) // dsp->put_mspel_pixels_tab[uvdxy](s->dest[2], srcV, s->uvlinesize); } +/** Do motion compensation for 4-MV macroblock - luminance block + */ +static void vc1_mc_4mv_luma(VC1Context *v, int n) +{ + MpegEncContext *s = &v->s; + DSPContext *dsp = &v->s.dsp; + uint8_t *srcY; + int dxy, mx, my, src_x, src_y; + int off; + + if(!v->s.last_picture.data[0])return; + mx = s->mv[0][n][0]; + my = s->mv[0][n][1]; + srcY = s->last_picture.data[0]; + + off = s->linesize * 4 * (n&2) + (n&1) * 8; + + src_x = s->mb_x * 16 + (n&1) * 8 + (mx >> 2); + src_y = s->mb_y * 16 + (n&2) * 4 + (my >> 2); + + CLIP_RANGE(src_x, s->mb_x, s->mb_width * 16, 16); + CLIP_RANGE(src_y, s->mb_y, s->mb_height * 16, 16); + + srcY += src_y * s->linesize + src_x; + + if((unsigned)src_x > s->h_edge_pos - (mx&3) - 16 + || (unsigned)src_y > s->v_edge_pos - (my&3) - 16){ + ff_emulated_edge_mc(s->edge_emu_buffer, srcY, s->linesize, 16+1, 16+1, + src_x, src_y, s->h_edge_pos, s->v_edge_pos); + srcY = s->edge_emu_buffer; + } + + if(!s->quarter_sample) { // hpel mc + mx >>= 1; + my >>= 1; + dxy = ((my & 1) << 1) | (mx & 1); + + dsp->put_no_rnd_pixels_tab[1][dxy](s->dest[0] + off, srcY, s->linesize, 8); + } else { + dxy = ((my & 3) << 2) | (mx & 3); + + dsp->put_no_rnd_qpel_pixels_tab[1][dxy](s->dest[0] + off, srcY, s->linesize); + } +} + +#define SETMAXMIN(var) \ + if(var > ma) ma = var; \ + if(var < mi) mi = var; + +static inline int median4(int a, int b, int c, int d) +{ + int ma, mi; + + ma = mi = a; + SETMAXMIN(b); + SETMAXMIN(c); + SETMAXMIN(d); + + return (a + b + c + d - ma - mi) >> 1; +} + + +/** Do motion compensation for 4-MV macroblock - both chroma blocks + */ +static void vc1_mc_4mv_chroma(VC1Context *v) +{ + MpegEncContext *s = &v->s; + DSPContext *dsp = &v->s.dsp; + uint8_t *srcU, *srcV; + int uvdxy, uvmx, uvmy, uvsrc_x, uvsrc_y; + int i, idx, tx = 0, ty = 0; + int mvx[4], mvy[4], intra[4]; + static const int count[16] = { 0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3, 4}; + + if(!v->s.last_picture.data[0])return; + + for(i = 0; i < 4; i++) { + mvx[i] = s->mv[0][i][0]; + mvy[i] = s->mv[0][i][1]; + intra[i] = v->mb_type[0][s->block_index[i]]; + } + + /* calculate chroma MV vector from four luma MVs */ + idx = (intra[0] << 3) | (intra[1] << 2) | (intra[2] << 1) | intra[3]; + if(!idx) { // all blocks are inter + tx = median4(mvx[0], mvx[1], mvx[2], mvx[3]); + ty = median4(mvy[0], mvy[1], mvy[2], mvy[3]); + } else if(count[idx] == 1) { // 3 inter blocks + switch(idx) { + case 0x1: + tx = mid_pred(mvx[1], mvx[2], mvx[3]); + tx = mid_pred(mvy[1], mvy[2], mvy[3]); + break; + case 0x2: + tx = mid_pred(mvx[0], mvx[2], mvx[3]); + tx = mid_pred(mvy[0], mvy[2], mvy[3]); + break; + case 0x4: + tx = mid_pred(mvx[0], mvx[1], mvx[3]); + tx = mid_pred(mvy[0], mvy[1], mvy[3]); + break; + case 0x8: + tx = mid_pred(mvx[0], mvx[1], mvx[2]); + tx = mid_pred(mvy[0], mvy[1], mvy[2]); + break; + } + } else if(count[idx] == 2) { + int t1 = 0, t2 = 0; + for(i=0; i<3;i++) if(!intra[i]) {t1 = i; break;} + for(i= t1+1; i<4; i++)if(!intra[i]) {t2 = i; break;} + tx = (mvx[t1] + mvx[t2]) >> 1; + ty = (mvy[t1] + mvy[t2]) >> 1; + } else + return; //no need to do MC for inter blocks + + uvmx = (tx + ((tx&3) == 3)) >> 1; + uvmy = (ty + ((ty&3) == 3)) >> 1; + + uvsrc_x = s->mb_x * 8 + (uvmx >> 2); + uvsrc_y = s->mb_y * 8 + (uvmy >> 2); + + CLIP_RANGE(uvsrc_x, s->mb_x, s->mb_width * 8, 8); + CLIP_RANGE(uvsrc_y, s->mb_y, s->mb_height * 8, 8); + srcU = s->last_picture.data[1] + uvsrc_y * s->uvlinesize + uvsrc_x; + srcV = s->last_picture.data[2] + uvsrc_y * s->uvlinesize + uvsrc_x; + if((unsigned)uvsrc_x > (s->h_edge_pos >> 1) - ((uvmx >> 1)&1) - 8 + || (unsigned)uvsrc_y > (s->v_edge_pos >> 1) - ((uvmy >> 1)&1) - 8){ + ff_emulated_edge_mc(s->edge_emu_buffer , srcU, s->uvlinesize, 8+1, 8+1, + uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1); + ff_emulated_edge_mc(s->edge_emu_buffer + 16, srcV, s->uvlinesize, 8+1, 8+1, + uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1); + srcU = s->edge_emu_buffer; + srcV = s->edge_emu_buffer + 16; + } + + if(!s->quarter_sample) // hpel mc + uvdxy = 0; + else + uvdxy = ((uvmy & 1) << 1) | (uvmx & 1); + dsp->put_no_rnd_pixels_tab[1][uvdxy](s->dest[1], srcU, s->uvlinesize, 8); + dsp->put_no_rnd_pixels_tab[1][uvdxy](s->dest[2], srcV, s->uvlinesize, 8); +} + /** * Decode Simple/Main Profiles sequence header * @see Figure 7-8, p16-17 @@ -1189,10 +1333,6 @@ static int vc1_parse_frame_header(VC1Context *v, GetBitContext* gb) else v->s.quarter_sample = 1; -if(v->mv_mode != MV_PMODE_1MV && v->mv_mode != MV_PMODE_1MV_HPEL && v->mv_mode != MV_PMODE_1MV_HPEL_BILIN) { - av_log(v->s.avctx, AV_LOG_ERROR, "Only 1MV P-frames are supported by now\n"); - return -1; -} if ((v->mv_mode == MV_PMODE_INTENSITY_COMP && v->mv_mode2 == MV_PMODE_MIXED_MV) || v->mv_mode == MV_PMODE_MIXED_MV) @@ -1276,6 +1416,48 @@ if(v->mv_mode != MV_PMODE_1MV && v->mv_mode != MV_PMODE_1MV_HPEL && v->mv_mode ! else mquant = get_bits(gb, 5); \ } \ } \ + else if(v->dqprofile == DQPROFILE_SINGLE_EDGE) \ + { \ + switch(v->dqsbedge){ \ + case 0: /* left */ \ + mquant = (s->mb_x) ? v->pq : v->altpq; \ + break; \ + case 1: /* top */ \ + mquant = (s->mb_y) ? v->pq : v->altpq; \ + break; \ + case 2: /* right */ \ + mquant = (s->mb_x != (s->mb_width - 1)) ? v->pq : v->altpq; \ + break; \ + case 3: /* bottom */ \ + mquant = (s->mb_y != (s->mb_height-1)) ? v->pq : v->altpq; \ + break; \ + default: \ + mquant = v->pq; \ + } \ + } \ + else if(v->dqprofile == DQPROFILE_DOUBLE_EDGES) \ + { \ + switch(v->dqsbedge){ \ + case 0: /* left and top */ \ + mquant = (s->mb_x && s->mb_y) ? v->pq : v->altpq; \ + break; \ + case 1: /* top and right */ \ + mquant = (s->mb_y && s->mb_x != (s->mb_width - 1)) ? v->pq : v->altpq; \ + break; \ + case 2: /* right and bottom */ \ + mquant = (s->mb_x != (s->mb_width - 1) && s->mb_y != (s->mb_height-1)) ? v->pq : v->altpq; \ + break; \ + case 3: /* bottom and left */ \ + mquant = (s->mb_x && s->mb_y != (s->mb_height-1)) ? v->pq : v->altpq; \ + break; \ + default: \ + mquant = v->pq; \ + } \ + } \ + else if(v->dqprofile == DQPROFILE_FOUR_EDGES) \ + { \ + mquant = (s->mb_x && s->mb_y && s->mb_x != (s->mb_width - 1) && s->mb_y != (s->mb_height-1)) ? v->pq : v->altpq; \ + } \ else mquant = v->pq; \ } @@ -1332,27 +1514,57 @@ if(v->mv_mode != MV_PMODE_1MV && v->mv_mode != MV_PMODE_1MV_HPEL && v->mv_mode ! /** Predict and set motion vector */ -static inline void vc1_pred_mv(MpegEncContext *s, int dmv_x, int dmv_y, int mv1, int r_x, int r_y) +static inline void vc1_pred_mv(MpegEncContext *s, int n, int dmv_x, int dmv_y, int mv1, int r_x, int r_y, uint8_t* is_intra) { - int xy, wrap, off; + int xy, wrap, off = 0; int16_t *A, *B, *C; int px, py; int sum; - int mb_pos = s->mb_x + s->mb_y * s->mb_stride; /* scale MV difference to be quad-pel */ dmv_x <<= 1 - s->quarter_sample; dmv_y <<= 1 - s->quarter_sample; wrap = s->b8_stride; - xy = s->block_index[0]; + xy = s->block_index[n]; - C = s->current_picture.motion_val[0][xy - (1 << mv1)]; - A = s->current_picture.motion_val[0][xy - (wrap << mv1)]; - off = (s->mb_x == (s->mb_width - 1)) ? -1 : 1; - B = s->current_picture.motion_val[0][xy + ((off - wrap) << mv1)]; + if(s->mb_intra){ + s->mv[0][n][0] = s->current_picture.motion_val[0][xy][0] = 0; + s->mv[0][n][1] = s->current_picture.motion_val[0][xy][1] = 0; + if(mv1) { /* duplicate motion data for 1-MV block */ + s->current_picture.motion_val[0][xy + 1][0] = 0; + s->current_picture.motion_val[0][xy + 1][1] = 0; + s->current_picture.motion_val[0][xy + wrap][0] = 0; + s->current_picture.motion_val[0][xy + wrap][1] = 0; + s->current_picture.motion_val[0][xy + wrap + 1][0] = 0; + s->current_picture.motion_val[0][xy + wrap + 1][1] = 0; + } + return; + } + + C = s->current_picture.motion_val[0][xy - 1]; + A = s->current_picture.motion_val[0][xy - wrap]; + if(mv1) + off = (s->mb_x == (s->mb_width - 1)) ? -1 : 2; + else { + //in 4-MV mode different blocks have different B predictor position + switch(n){ + case 0: + off = (s->mb_x > 0) ? -1 : 1; + break; + case 1: + off = (s->mb_x == (s->mb_width - 1)) ? -1 : 1; + break; + case 2: + off = 1; + break; + case 3: + off = -1; + } + } + B = s->current_picture.motion_val[0][xy - wrap + off]; - if(!s->first_slice_line) { // predictor A is not out of bounds + if(!s->first_slice_line || (n==2 || n==3)) { // predictor A is not out of bounds if(s->mb_width == 1) { px = A[0]; py = A[1]; @@ -1360,14 +1572,12 @@ static inline void vc1_pred_mv(MpegEncContext *s, int dmv_x, int dmv_y, int mv1, px = mid_pred(A[0], B[0], C[0]); py = mid_pred(A[1], B[1], C[1]); } - } else if(s->mb_x) { // predictor C is not out of bounds + } else if(s->mb_x || (n==1 || n==3)) { // predictor C is not out of bounds px = C[0]; py = C[1]; } else { px = py = 0; } - if(s->mb_intra) px = py = 0; - /* Pullback MV as specified in 8.3.5.3.4 */ { int qx, qy, X, Y; @@ -1386,8 +1596,8 @@ static inline void vc1_pred_mv(MpegEncContext *s, int dmv_x, int dmv_y, int mv1, if(qy + py > Y) py = Y - qy; } /* Calculate hybrid prediction as specified in 8.3.5.3.5 */ - if(!s->mb_intra && !s->first_slice_line && s->mb_x) { - if(IS_INTRA(s->current_picture.mb_type[mb_pos - s->mb_stride])) + if((!s->first_slice_line || (n==2 || n==3)) && (s->mb_x || (n==1 || n==3))) { + if(is_intra[xy - wrap]) sum = ABS(px) + ABS(py); else sum = ABS(px - A[0]) + ABS(py - A[1]); @@ -1400,7 +1610,7 @@ static inline void vc1_pred_mv(MpegEncContext *s, int dmv_x, int dmv_y, int mv1, py = C[1]; } } else { - if(IS_INTRA(s->current_picture.mb_type[mb_pos - 1])) + if(is_intra[xy - 1]) sum = ABS(px) + ABS(py); else sum = ABS(px - C[0]) + ABS(py - C[1]); @@ -1416,8 +1626,16 @@ static inline void vc1_pred_mv(MpegEncContext *s, int dmv_x, int dmv_y, int mv1, } } /* store MV using signed modulus of MV range defined in 4.11 */ - s->mv[0][0][0] = s->current_picture.motion_val[0][xy][0] = ((px + dmv_x + r_x) & ((r_x << 1) - 1)) - r_x; - s->mv[0][0][1] = s->current_picture.motion_val[0][xy][1] = ((py + dmv_y + r_y) & ((r_y << 1) - 1)) - r_y; + s->mv[0][n][0] = s->current_picture.motion_val[0][xy][0] = ((px + dmv_x + r_x) & ((r_x << 1) - 1)) - r_x; + s->mv[0][n][1] = s->current_picture.motion_val[0][xy][1] = ((py + dmv_y + r_y) & ((r_y << 1) - 1)) - r_y; + if(mv1) { /* duplicate motion data for 1-MV block */ + s->current_picture.motion_val[0][xy + 1][0] = s->current_picture.motion_val[0][xy][0]; + s->current_picture.motion_val[0][xy + 1][1] = s->current_picture.motion_val[0][xy][1]; + s->current_picture.motion_val[0][xy + wrap][0] = s->current_picture.motion_val[0][xy][0]; + s->current_picture.motion_val[0][xy + wrap][1] = s->current_picture.motion_val[0][xy][1]; + s->current_picture.motion_val[0][xy + wrap + 1][0] = s->current_picture.motion_val[0][xy][0]; + s->current_picture.motion_val[0][xy + wrap + 1][1] = s->current_picture.motion_val[0][xy][1]; + } } /** Get predicted DC value for I-frames only @@ -1529,15 +1747,13 @@ static inline int vc1_pred_dc(MpegEncContext *s, int overlap, int pq, int n, *dir_ptr = 1;//left } - /* scale coeffs if needed */ + /* scale coeffs if needed mb_pos2 = mb_pos - *dir_ptr - (1 - *dir_ptr) * s->mb_stride; - q1 = s->current_picture.qscale_table[mb_pos]; - q2 = s->current_picture.qscale_table[mb_pos2]; - if(0 && q1 && q2 && q1 != q2) { - q1 = s->y_dc_scale_table[q1]; - q2 = s->y_dc_scale_table[q2]; + q1 = s->y_dc_scale_table[s->current_picture.qscale_table[mb_pos]]; + q2 = s->y_dc_scale_table[s->current_picture.qscale_table[mb_pos2]]; + if(q2 && q1!=q2 && ((*dir_ptr && c_avail) || (!*dir_ptr && a_avail))) { pred = (pred * q2 * vc1_dqscale[q1 - 1] + 0x20000) >> 18; - } + } */ /* update predictor */ *dc_val_ptr = &dc_val[0]; @@ -1695,7 +1911,6 @@ static int vc1_decode_i_block(VC1Context *v, DCTELEM block[64], int n, int coded *dc_val = dcdiff; /* Store the quantized DC coeff, used for prediction */ - if (n < 4) { block[0] = dcdiff * s->y_dc_scale; } else { @@ -1926,7 +2141,7 @@ static int vc1_decode_intra_block(VC1Context *v, DCTELEM block[64], int n, int c q1 = s->current_picture.qscale_table[mb_pos]; q2 = s->current_picture.qscale_table[mb_pos2]; - if(q2 && q1 != q2) { + if(0 && q2 && q1!=q2 && ((dc_pred_dir && c_avail) || (!dc_pred_dir && a_avail))) { q1 = q1 * 2 - 1; q2 = q2 * 2 - 1; @@ -1968,6 +2183,11 @@ not_coded: if(!coded) { int k, scale; int use_pred = s->ac_pred; + int mb_pos2, q1, q2; + + mb_pos2 = mb_pos - dc_pred_dir - (1 - dc_pred_dir) * s->mb_stride; + q1 = s->current_picture.qscale_table[mb_pos]; + q2 = s->current_picture.qscale_table[mb_pos2]; ac_val = s->ac_val[0][0] + s->block_index[n] * 16; ac_val2 = ac_val; @@ -1987,25 +2207,39 @@ not_coded: memset(ac_val2, 0, 16 * 2); if(dc_pred_dir) {//left ac_val -= 16; - if(use_pred) + if(use_pred) { memcpy(ac_val2, ac_val, 8 * 2); + if(0 && q2 && q1!=q2 && c_avail) { + q1 = q1 * 2 - 1; + q2 = q2 * 2 - 1; + for(k = 1; k < 8; k++) + ac_val2[k] = (ac_val2[k] * q2 * vc1_dqscale[q1 - 1] + 0x20000) >> 18; + } + } } else {//top ac_val -= 16 * s->block_wrap[n]; - if(use_pred) + if(use_pred) { memcpy(ac_val2 + 8, ac_val + 8, 8 * 2); + if(0 && q2 && q1!=q2 && a_avail) { + q1 = q1 * 2 - 1; + q2 = q2 * 2 - 1; + for(k = 1; k < 8; k++) + ac_val2[k + 8] = (ac_val2[k + 8] * q2 * vc1_dqscale[q1 - 1] + 0x20000) >> 18; + } + } } /* apply AC prediction if needed */ if(use_pred) { if(dc_pred_dir) { //left for(k = 1; k < 8; k++) { - block[k << 3] = ac_val[k] * scale; + block[k << 3] = ac_val2[k] * scale; if(!v->pquantizer) block[k << 3] += (block[k << 3] < 0) ? -mquant : mquant; } } else { //top for(k = 1; k < 8; k++) { - block[k] = ac_val[k + 8] * scale; + block[k] = ac_val2[k + 8] * scale; if(!v->pquantizer) block[k] += (block[k] < 0) ? -mquant : mquant; } @@ -2133,7 +2367,6 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) int i, j; int mb_pos = s->mb_x + s->mb_y * s->mb_stride; int cbp; /* cbp decoding stuff */ - int hybrid_pred; /* Prediction types */ int mqdiff, mquant; /* MB quantization */ int ttmb = v->ttmb; /* MB Transform type */ int status; @@ -2159,6 +2392,8 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) else skipped = v->s.mbskip_table[mb_pos]; + s->dsp.clear_blocks(s->block[0]); + if (!fourmv) /* 1MV mode */ { if (!skipped) @@ -2166,7 +2401,7 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) GET_MVDATA(dmv_x, dmv_y); s->current_picture.mb_type[mb_pos] = s->mb_intra ? MB_TYPE_INTRA : MB_TYPE_16x16; - vc1_pred_mv(s, dmv_x, dmv_y, 1, v->range_x, v->range_y); + vc1_pred_mv(s, 0, dmv_x, dmv_y, 1, v->range_x, v->range_y, v->mb_type[0]); /* FIXME Set DC val for inter block ? */ if (s->mb_intra && !mb_has_coeffs) @@ -2191,8 +2426,7 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) if (!v->ttmbf && !s->mb_intra && mb_has_coeffs) ttmb = get_vlc2(gb, vc1_ttmb_vlc[v->tt_index].table, VC1_TTMB_VLC_BITS, 2); - s->dsp.clear_blocks(block[0]); - vc1_mc_1mv(v); + if(!s->mb_intra) vc1_mc_1mv(v); dst_idx = 0; for (i=0; i<6; i++) { @@ -2200,13 +2434,14 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) dst_idx += i >> 2; val = ((cbp >> (5 - i)) & 1); off = (i & 4) ? 0 : ((i & 1) * 8 + (i & 2) * 4 * s->linesize); + v->mb_type[0][s->block_index[i]] = s->mb_intra; if(s->mb_intra) { /* check if prediction blocks A and C are available */ v->a_avail = v->c_avail = 0; - if((i == 2 || i == 3) || (s->mb_y && IS_INTRA(s->current_picture.mb_type[mb_pos - s->mb_stride]))) - v->a_avail = 1; - if((i == 1 || i == 3) || (s->mb_x && IS_INTRA(s->current_picture.mb_type[mb_pos - 1]))) - v->c_avail = 1; + if(i == 2 || i == 3 || s->mb_y) + v->a_avail = v->mb_type[0][s->block_index[i] - s->block_wrap[i]]; + if(i == 1 || i == 3 || s->mb_x) + v->c_avail = v->mb_type[0][s->block_index[i] - 1]; vc1_decode_intra_block(v, block[i], i, val, mquant, (i&4)?v->codingset2:v->codingset); vc1_inv_trans(block[i], 8, 8); @@ -2230,62 +2465,107 @@ static int vc1_decode_p_mb(VC1Context *v, DCTELEM block[6][64]) else //Skipped { s->mb_intra = 0; + for(i = 0; i < 6; i++) v->mb_type[0][s->block_index[i]] = 0; s->current_picture.mb_type[mb_pos] = MB_TYPE_SKIP; - vc1_pred_mv(s, 0, 0, 1, v->range_x, v->range_y); + s->current_picture.qscale_table[mb_pos] = 0; + vc1_pred_mv(s, 0, 0, 0, 1, v->range_x, v->range_y, v->mb_type[0]); vc1_mc_1mv(v); return 0; } } //1MV mode else //4MV mode - {//FIXME: looks not conforming to standard and is not even theoretically complete + { if (!skipped /* unskipped MB */) { - int blk_intra[4], blk_coded[4]; + int intra_count = 0, coded_inter = 0; + int is_intra[6], is_coded[6]; /* Get CBPCY */ cbp = get_vlc2(&v->s.gb, v->cbpcy_vlc->table, VC1_CBPCY_P_VLC_BITS, 2); - for (i=0; i<4; i++) + for (i=0; i<6; i++) { val = ((cbp >> (5 - i)) & 1); - blk_intra[i] = 0; - blk_coded[i] = val; - if(val) { - GET_MVDATA(dmv_x, dmv_y); - blk_intra[i] = s->mb_intra; + s->dc_val[0][s->block_index[i]] = 0; + s->mb_intra = 0; + if(i < 4) { + dmv_x = dmv_y = 0; + s->mb_intra = 0; + mb_has_coeffs = 0; + if(val) { + GET_MVDATA(dmv_x, dmv_y); + } + vc1_pred_mv(s, i, dmv_x, dmv_y, 0, v->range_x, v->range_y, v->mb_type[0]); + if(!s->mb_intra) vc1_mc_4mv_luma(v, i); + intra_count += s->mb_intra; + is_intra[i] = s->mb_intra; + is_coded[i] = mb_has_coeffs; + } + if(i&4){ + is_intra[i] = (intra_count >= 3); + is_coded[i] = val; } - if (v->mv_mode == MV_PMODE_MIXED_MV /* Hybrid pred */) - hybrid_pred = get_bits(gb, 1); + if(i == 4) vc1_mc_4mv_chroma(v); + v->mb_type[0][s->block_index[i]] = is_intra[i]; + if(!coded_inter) coded_inter = !is_intra[i] & is_coded[i]; } - if((blk_intra[0] | blk_intra[1] | blk_intra[2] | blk_intra[3]) || - (blk_coded[0] | blk_coded[1] | blk_coded[2] | blk_coded[3])) { - GET_MQUANT(); + dst_idx = 0; + GET_MQUANT(); + s->current_picture.qscale_table[mb_pos] = mquant; + /* test if block is intra and has pred */ + { + int intrapred = 0; + for(i=0; i<6; i++) + if(is_intra[i]) { + if(v->mb_type[0][s->block_index[i] - s->block_wrap[i]] || v->mb_type[0][s->block_index[i] - 1]) { + intrapred = 1; + break; + } + } + if(intrapred)s->ac_pred = get_bits(gb, 1); + else s->ac_pred = 0; + } + if (!v->ttmbf && coded_inter) + ttmb = get_vlc2(gb, vc1_ttmb_vlc[v->tt_index].table, VC1_TTMB_VLC_BITS, 12); + for (i=0; i<6; i++) + { + dst_idx += i >> 2; + off = (i & 4) ? 0 : ((i & 1) * 8 + (i & 2) * 4 * s->linesize); + s->mb_intra = is_intra[i]; + if (is_intra[i]) { + /* check if prediction blocks A and C are available */ + v->a_avail = v->c_avail = 0; + if(i == 2 || i == 3 || s->mb_y) + v->a_avail = v->mb_type[0][s->block_index[i] - s->block_wrap[i]]; + if(i == 1 || i == 3 || s->mb_x) + v->c_avail = v->mb_type[0][s->block_index[i] - 1]; - if (s->mb_intra /* One of the 4 blocks is intra */ - /* non-zero pred for that block */) - s->ac_pred = get_bits(gb, 1); - if (!v->ttmbf) - ttmb = get_vlc2(gb, vc1_ttmb_vlc[v->tt_index].table, - VC1_TTMB_VLC_BITS, 12); - for(i = 0; i < 6; i++) { - val = ((cbp >> (5 - i)) & 1); - if(i & 4 || blk_intra[i] || val) { - if(i < 4 && blk_intra[i]) - status = vc1_decode_intra_block(v, block[i], i, val, mquant, (i&4)?v->codingset2:v->codingset); - else - status = vc1_decode_p_block(v, block[i], i, mquant, ttmb, 0); + vc1_decode_intra_block(v, s->block[i], i, is_coded[i], mquant, (i&4)?v->codingset2:v->codingset); + //s->dsp.put_pixels_clamped(s->block[i], s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize); + /* TODO: proper loop filtering */ + if(v->pq >= 9 && v->overlap) { + if(v->a_avail) + s->dsp.h263_v_loop_filter(s->dest[dst_idx] + off, s->linesize >> ((i & 4) >> 2), s->y_dc_scale); + if(v->c_avail) + s->dsp.h263_h_loop_filter(s->dest[dst_idx] + off, s->linesize >> ((i & 4) >> 2), s->y_dc_scale); } + } else if(is_coded[i]) { + status = vc1_decode_p_block(v, s->block[i], i, mquant, ttmb, first_block); + if(!v->ttmbf && ttmb < 8) ttmb = -1; + first_block = 0; + //s->dsp.add_pixels_clamped(s->block[i], s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize); } } return status; } else //Skipped MB { - /* XXX: Skipped => cbp=0 and mquant doesn't matter ? */ for (i=0; i<4; i++) { - if (v->mv_mode == MV_PMODE_MIXED_MV /* Hybrid pred */) - hybrid_pred = get_bits(gb, 1); + vc1_pred_mv(s, i, 0, 0, 0, v->range_x, v->range_y, v->mb_type[0]); + vc1_mc_4mv_luma(v, i); } - /* TODO: blah */ + vc1_mc_4mv_chroma(v); + for(i = 0; i < 6; i++) v->mb_type[0][s->block_index[i]] = 0; + s->current_picture.qscale_table[mb_pos] = 0; return 0; } } @@ -2519,6 +2799,12 @@ static int vc1_decode_init(AVCodecContext *avctx) /* Allocate mb bitplanes */ v->mv_type_mb_plane = av_malloc(s->mb_stride * s->mb_height); + /* allocate block type info in that way so it could be used with s->block_index[] */ + v->mb_type_base = av_malloc(s->b8_stride * (s->mb_height * 2 + 1) + s->mb_stride * (s->mb_height + 1) * 2); + v->mb_type[0] = v->mb_type_base + s->b8_stride + 1; + v->mb_type[1] = v->mb_type_base + s->b8_stride * (s->mb_height * 2 + 1) + s->mb_stride + 1; + v->mb_type[2] = v->mb_type[1] + s->mb_stride * (s->mb_height + 1); + /* Init coded blocks info */ if (v->profile == PROFILE_ADVANCED) { @@ -2640,6 +2926,7 @@ static int vc1_decode_end(AVCodecContext *avctx) av_freep(&v->hrd_buffer); MPV_common_end(&v->s); av_freep(&v->mv_type_mb_plane); + av_freep(&v->mb_type_base); return 0; } -- cgit v1.2.3