3 * VP6 compatible video decoder
5 * Copyright (C) 2006 Aurelien Jacobs <aurel@gnuage.org>
7 * The VP6F decoder accepts an optional 1 byte extradata. It is composed of:
8 * - upper 4bits: difference between encoded width and visible width
9 * - lower 4bits: difference between encoded height and visible height
11 * This file is part of FFmpeg.
13 * FFmpeg is free software; you can redistribute it and/or
14 * modify it under the terms of the GNU Lesser General Public
15 * License as published by the Free Software Foundation; either
16 * version 2.1 of the License, or (at your option) any later version.
18 * FFmpeg is distributed in the hope that it will be useful,
19 * but WITHOUT ANY WARRANTY; without even the implied warranty of
20 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
21 * Lesser General Public License for more details.
23 * You should have received a copy of the GNU Lesser General Public
24 * License along with FFmpeg; if not, write to the Free Software
25 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
32 #include "bitstream.h"
33 #include "mpegvideo.h"
40 static int vp6_parse_header(vp56_context_t *s, uint8_t *buf, int buf_size,
43 vp56_range_coder_t *c = &s->c;
44 int parse_filter_info = 0;
50 int separated_coeff = buf[0] & 1;
52 s->framep[VP56_FRAME_CURRENT]->key_frame = !(buf[0] & 0x80);
53 vp56_init_dequant(s, (buf[0] >> 1) & 0x3F);
55 if (s->framep[VP56_FRAME_CURRENT]->key_frame) {
56 sub_version = buf[1] >> 3;
59 s->filter_header = buf[1] & 0x06;
61 av_log(s->avctx, AV_LOG_ERROR, "interlacing not supported\n");
64 if (separated_coeff || !s->filter_header) {
65 coeff_offset = AV_RB16(buf+2) - 2;
70 rows = buf[2]; /* number of stored macroblock rows */
71 cols = buf[3]; /* number of stored macroblock cols */
72 /* buf[4] is number of displayed macroblock rows */
73 /* buf[5] is number of displayed macroblock cols */
75 if (16*cols != s->avctx->coded_width ||
76 16*rows != s->avctx->coded_height) {
77 avcodec_set_dimensions(s->avctx, 16*cols, 16*rows);
78 if (s->avctx->extradata_size == 1) {
79 s->avctx->width -= s->avctx->extradata[0] >> 4;
80 s->avctx->height -= s->avctx->extradata[0] & 0x0F;
85 vp56_init_range_decoder(c, buf+6, buf_size-6);
88 parse_filter_info = s->filter_header;
91 s->sub_version = sub_version;
96 if (separated_coeff || !s->filter_header) {
97 coeff_offset = AV_RB16(buf+1) - 2;
101 vp56_init_range_decoder(c, buf+1, buf_size-1);
103 *golden_frame = vp56_rac_get(c);
104 if (s->filter_header) {
105 s->deblock_filtering = vp56_rac_get(c);
106 if (s->deblock_filtering)
108 if (s->sub_version > 7)
109 parse_filter_info = vp56_rac_get(c);
113 if (parse_filter_info) {
114 if (vp56_rac_get(c)) {
116 s->sample_variance_threshold = vp56_rac_gets(c, 5) << vrt_shift;
117 s->max_vector_length = 2 << vp56_rac_gets(c, 3);
118 } else if (vp56_rac_get(c)) {
123 if (s->sub_version > 7)
124 s->filter_selection = vp56_rac_gets(c, 4);
126 s->filter_selection = 16;
130 av_log(s->avctx, AV_LOG_WARNING,
131 "alternative entropy decoding not supported\n");
134 vp56_init_range_decoder(&s->cc, buf+coeff_offset,
135 buf_size-coeff_offset);
144 static void vp6_coeff_order_table_init(vp56_context_t *s)
148 s->coeff_index_to_pos[0] = 0;
150 for (pos=1; pos<64; pos++)
151 if (s->coeff_reorder[pos] == i)
152 s->coeff_index_to_pos[idx++] = pos;
155 static void vp6_default_models_init(vp56_context_t *s)
157 s->vector_model_dct[0] = 0xA2;
158 s->vector_model_dct[1] = 0xA4;
159 s->vector_model_sig[0] = 0x80;
160 s->vector_model_sig[1] = 0x80;
162 memcpy(s->mb_types_stats, vp56_def_mb_types_stats, sizeof(s->mb_types_stats));
163 memcpy(s->vector_model_fdv, vp6_def_fdv_vector_model, sizeof(s->vector_model_fdv));
164 memcpy(s->vector_model_pdv, vp6_def_pdv_vector_model, sizeof(s->vector_model_pdv));
165 memcpy(s->coeff_model_runv, vp6_def_runv_coeff_model, sizeof(s->coeff_model_runv));
166 memcpy(s->coeff_reorder, vp6_def_coeff_reorder, sizeof(s->coeff_reorder));
168 vp6_coeff_order_table_init(s);
171 static void vp6_parse_vector_models(vp56_context_t *s)
173 vp56_range_coder_t *c = &s->c;
176 for (comp=0; comp<2; comp++) {
177 if (vp56_rac_get_prob(c, vp6_sig_dct_pct[comp][0]))
178 s->vector_model_dct[comp] = vp56_rac_gets_nn(c, 7);
179 if (vp56_rac_get_prob(c, vp6_sig_dct_pct[comp][1]))
180 s->vector_model_sig[comp] = vp56_rac_gets_nn(c, 7);
183 for (comp=0; comp<2; comp++)
184 for (node=0; node<7; node++)
185 if (vp56_rac_get_prob(c, vp6_pdv_pct[comp][node]))
186 s->vector_model_pdv[comp][node] = vp56_rac_gets_nn(c, 7);
188 for (comp=0; comp<2; comp++)
189 for (node=0; node<8; node++)
190 if (vp56_rac_get_prob(c, vp6_fdv_pct[comp][node]))
191 s->vector_model_fdv[comp][node] = vp56_rac_gets_nn(c, 7);
194 static void vp6_parse_coeff_models(vp56_context_t *s)
196 vp56_range_coder_t *c = &s->c;
198 int node, cg, ctx, pos;
199 int ct; /* code type */
200 int pt; /* plane type (0 for Y, 1 for U or V) */
202 memset(def_prob, 0x80, sizeof(def_prob));
204 for (pt=0; pt<2; pt++)
205 for (node=0; node<11; node++)
206 if (vp56_rac_get_prob(c, vp6_dccv_pct[pt][node])) {
207 def_prob[node] = vp56_rac_gets_nn(c, 7);
208 s->coeff_model_dccv[pt][node] = def_prob[node];
209 } else if (s->framep[VP56_FRAME_CURRENT]->key_frame) {
210 s->coeff_model_dccv[pt][node] = def_prob[node];
213 if (vp56_rac_get(c)) {
214 for (pos=1; pos<64; pos++)
215 if (vp56_rac_get_prob(c, vp6_coeff_reorder_pct[pos]))
216 s->coeff_reorder[pos] = vp56_rac_gets(c, 4);
217 vp6_coeff_order_table_init(s);
220 for (cg=0; cg<2; cg++)
221 for (node=0; node<14; node++)
222 if (vp56_rac_get_prob(c, vp6_runv_pct[cg][node]))
223 s->coeff_model_runv[cg][node] = vp56_rac_gets_nn(c, 7);
225 for (ct=0; ct<3; ct++)
226 for (pt=0; pt<2; pt++)
227 for (cg=0; cg<6; cg++)
228 for (node=0; node<11; node++)
229 if (vp56_rac_get_prob(c, vp6_ract_pct[ct][pt][cg][node])) {
230 def_prob[node] = vp56_rac_gets_nn(c, 7);
231 s->coeff_model_ract[pt][ct][cg][node] = def_prob[node];
232 } else if (s->framep[VP56_FRAME_CURRENT]->key_frame) {
233 s->coeff_model_ract[pt][ct][cg][node] = def_prob[node];
236 /* coeff_model_dcct is a linear combination of coeff_model_dccv */
237 for (pt=0; pt<2; pt++)
238 for (ctx=0; ctx<3; ctx++)
239 for (node=0; node<5; node++)
240 s->coeff_model_dcct[pt][ctx][node] = av_clip(((s->coeff_model_dccv[pt][node] * vp6_dccv_lc[ctx][node][0] + 128) >> 8) + vp6_dccv_lc[ctx][node][1], 1, 255);
243 static void vp6_parse_vector_adjustment(vp56_context_t *s, vp56_mv_t *vect)
245 vp56_range_coder_t *c = &s->c;
248 *vect = (vp56_mv_t) {0,0};
249 if (s->vector_candidate_pos < 2)
250 *vect = s->vector_candidate[0];
252 for (comp=0; comp<2; comp++) {
255 if (vp56_rac_get_prob(c, s->vector_model_dct[comp])) {
256 static const uint8_t prob_order[] = {0, 1, 2, 7, 6, 5, 4};
257 for (i=0; i<sizeof(prob_order); i++) {
258 int j = prob_order[i];
259 delta |= vp56_rac_get_prob(c, s->vector_model_fdv[comp][j])<<j;
262 delta |= vp56_rac_get_prob(c, s->vector_model_fdv[comp][3])<<3;
266 delta = vp56_rac_get_tree(c, vp56_pva_tree,
267 s->vector_model_pdv[comp]);
270 if (delta && vp56_rac_get_prob(c, s->vector_model_sig[comp]))
280 static void vp6_parse_coeff(vp56_context_t *s)
282 vp56_range_coder_t *c = s->ccp;
283 uint8_t *permute = s->scantable.permutated;
284 uint8_t *model, *model2, *model3;
285 int coeff, sign, coeff_idx;
286 int b, i, cg, idx, ctx;
287 int pt = 0; /* plane type (0 for Y, 1 for U or V) */
289 for (b=0; b<6; b++) {
290 int ct = 1; /* code type */
295 ctx = s->left_block[vp56_b6to4[b]].not_null_dc
296 + s->above_blocks[s->above_block_idx[b]].not_null_dc;
297 model = s->coeff_model_dccv[pt];
298 model2 = s->coeff_model_dcct[pt][ctx];
300 for (coeff_idx=0; coeff_idx<64; ) {
301 if ((coeff_idx>1 && ct==0) || vp56_rac_get_prob(c, model2[0])) {
303 if (vp56_rac_get_prob(c, model2[2])) {
304 if (vp56_rac_get_prob(c, model2[3])) {
305 idx = vp56_rac_get_tree(c, vp56_pc_tree, model);
306 coeff = vp56_coeff_bias[idx];
307 for (i=vp56_coeff_bit_length[idx]; i>=0; i--)
308 coeff += vp56_rac_get_prob(c, vp56_coeff_parse_table[idx][i]) << i;
310 if (vp56_rac_get_prob(c, model2[4]))
311 coeff = 3 + vp56_rac_get_prob(c, model[5]);
320 sign = vp56_rac_get(c);
321 coeff = (coeff ^ -sign) + sign;
323 coeff *= s->dequant_ac;
324 idx = s->coeff_index_to_pos[coeff_idx];
325 s->block_coeff[b][permute[idx]] = coeff;
331 if (!vp56_rac_get_prob(c, model2[1]))
334 model3 = s->coeff_model_runv[coeff_idx >= 6];
335 run = vp56_rac_get_tree(c, vp6_pcr_tree, model3);
337 for (run=9, i=0; i<6; i++)
338 run += vp56_rac_get_prob(c, model3[i+8]) << i;
342 cg = vp6_coeff_groups[coeff_idx+=run];
343 model = model2 = s->coeff_model_ract[pt][ct][cg];
346 s->left_block[vp56_b6to4[b]].not_null_dc =
347 s->above_blocks[s->above_block_idx[b]].not_null_dc = !!s->block_coeff[b][0];
351 static int vp6_adjust(int v, int t)
353 int V = v, s = v >> 31;
356 if (V-t-1 >= (unsigned)(t-1))
364 static int vp6_block_variance(uint8_t *src, int stride)
366 int sum = 0, square_sum = 0;
369 for (y=0; y<8; y+=2) {
370 for (x=0; x<8; x+=2) {
372 square_sum += src[x]*src[x];
376 return (16*square_sum - sum*sum) >> 8;
379 static void vp6_filter_hv4(uint8_t *dst, uint8_t *src, int stride,
380 int delta, const int16_t *weights)
384 for (y=0; y<8; y++) {
385 for (x=0; x<8; x++) {
386 dst[x] = av_clip_uint8(( src[x-delta ] * weights[0]
387 + src[x ] * weights[1]
388 + src[x+delta ] * weights[2]
389 + src[x+2*delta] * weights[3] + 64) >> 7);
396 static void vp6_filter_diag2(vp56_context_t *s, uint8_t *dst, uint8_t *src,
397 int stride, int h_weight, int v_weight)
399 uint8_t *tmp = s->edge_emu_buffer+16;
400 s->dsp.put_h264_chroma_pixels_tab[0](tmp, src, stride, 9, h_weight, 0);
401 s->dsp.put_h264_chroma_pixels_tab[0](dst, tmp, stride, 8, 0, v_weight);
404 static void vp6_filter_diag4(uint8_t *dst, uint8_t *src, int stride,
405 const int16_t *h_weights,const int16_t *v_weights)
413 for (y=0; y<11; y++) {
414 for (x=0; x<8; x++) {
415 t[x] = av_clip_uint8(( src[x-1] * h_weights[0]
416 + src[x ] * h_weights[1]
417 + src[x+1] * h_weights[2]
418 + src[x+2] * h_weights[3] + 64) >> 7);
425 for (y=0; y<8; y++) {
426 for (x=0; x<8; x++) {
427 dst[x] = av_clip_uint8(( t[x-8 ] * v_weights[0]
428 + t[x ] * v_weights[1]
429 + t[x+8 ] * v_weights[2]
430 + t[x+16] * v_weights[3] + 64) >> 7);
437 static void vp6_filter(vp56_context_t *s, uint8_t *dst, uint8_t *src,
438 int offset1, int offset2, int stride,
439 vp56_mv_t mv, int mask, int select, int luma)
442 int x8 = mv.x & mask;
443 int y8 = mv.y & mask;
448 filter4 = s->filter_mode;
450 if (s->max_vector_length &&
451 (FFABS(mv.x) > s->max_vector_length ||
452 FFABS(mv.y) > s->max_vector_length)) {
454 } else if (s->sample_variance_threshold
455 && (vp6_block_variance(src+offset1, stride)
456 < s->sample_variance_threshold)) {
462 if ((y8 && (offset2-offset1)*s->flip<0) || (!y8 && offset1 > offset2)) {
467 if (!y8) { /* left or right combine */
468 vp6_filter_hv4(dst, src+offset1, stride, 1,
469 vp6_block_copy_filter[select][x8]);
470 } else if (!x8) { /* above or below combine */
471 vp6_filter_hv4(dst, src+offset1, stride, stride,
472 vp6_block_copy_filter[select][y8]);
474 vp6_filter_diag4(dst, src+offset1 + ((mv.x^mv.y)>>31), stride,
475 vp6_block_copy_filter[select][x8],
476 vp6_block_copy_filter[select][y8]);
480 s->dsp.put_h264_chroma_pixels_tab[0](dst, src+offset1, stride, 8, x8, y8);
482 vp6_filter_diag2(s, dst, src+offset1 + ((mv.x^mv.y)>>31), stride, x8, y8);
487 static int vp6_decode_init(AVCodecContext *avctx)
489 vp56_context_t *s = avctx->priv_data;
491 vp56_init(s, avctx, avctx->codec->id == CODEC_ID_VP6);
492 s->vp56_coord_div = vp6_coord_div;
493 s->parse_vector_adjustment = vp6_parse_vector_adjustment;
494 s->adjust = vp6_adjust;
495 s->filter = vp6_filter;
496 s->parse_coeff = vp6_parse_coeff;
497 s->default_models_init = vp6_default_models_init;
498 s->parse_vector_models = vp6_parse_vector_models;
499 s->parse_coeff_models = vp6_parse_coeff_models;
500 s->parse_header = vp6_parse_header;
505 AVCodec vp6_decoder = {
509 sizeof(vp56_context_t),
517 /* flash version, not flipped upside-down */
518 AVCodec vp6f_decoder = {
522 sizeof(vp56_context_t),