for (i = 0; i < 12; i++) {
s->pblocks[i] = &s->block[i];
}
+
+ FF_ALLOCZ_OR_GOTO(s->avctx, s->block32, sizeof(*s->block32), fail)
+
if (s->avctx->codec_tag == AV_RL32("VCR2")) {
// exchange uv
FFSWAP(void *, s->pblocks[4], s->pblocks[5]);
av_freep(&s->me.map);
av_freep(&s->me.score_map);
av_freep(&s->blocks);
+ av_freep(&s->block32);
av_freep(&s->ac_val_base);
s->block = NULL;
}
COPY(me.score_map);
COPY(blocks);
COPY(block);
+ COPY(block32);
COPY(start_mb_y);
COPY(end_mb_y);
COPY(me.map_generation);
s->dct_error_sum = NULL;
s->block = NULL;
s->blocks = NULL;
+ s->block32 = NULL;
memset(s->pblocks, 0, sizeof(s->pblocks));
s->ac_val_base = NULL;
s->ac_val[0] =
ff_thread_report_progress(&s->current_picture_ptr->tf, INT_MAX, 0);
}
-static int add_mb(AVMotionVector *mb, uint32_t mb_type,
- int dst_x, int dst_y,
- int motion_x, int motion_y, int motion_scale,
- int direction)
-{
- mb->w = IS_8X8(mb_type) || IS_8X16(mb_type) ? 8 : 16;
- mb->h = IS_8X8(mb_type) || IS_16X8(mb_type) ? 8 : 16;
- mb->motion_x = motion_x;
- mb->motion_y = motion_y;
- mb->motion_scale = motion_scale;
- mb->dst_x = dst_x;
- mb->dst_y = dst_y;
- mb->src_x = dst_x + motion_x / motion_scale;
- mb->src_y = dst_y + motion_y / motion_scale;
- mb->source = direction ? 1 : -1;
- mb->flags = 0; // XXX: does mb_type contain extra information that could be exported here?
- return 1;
-}
-
-/**
- * Print debugging info for the given picture.
- */
-void ff_print_debug_info2(AVCodecContext *avctx, AVFrame *pict, uint8_t *mbskip_table,
- uint32_t *mbtype_table, int8_t *qscale_table, int16_t (*motion_val[2])[2],
- int *low_delay,
- int mb_width, int mb_height, int mb_stride, int quarter_sample)
-{
- if ((avctx->flags2 & AV_CODEC_FLAG2_EXPORT_MVS) && mbtype_table && motion_val[0]) {
- const int shift = 1 + quarter_sample;
- const int scale = 1 << shift;
- const int mv_sample_log2 = avctx->codec_id == AV_CODEC_ID_H264 || avctx->codec_id == AV_CODEC_ID_SVQ3 ? 2 : 1;
- const int mv_stride = (mb_width << mv_sample_log2) +
- (avctx->codec->id == AV_CODEC_ID_H264 ? 0 : 1);
- int mb_x, mb_y, mbcount = 0;
-
- /* size is width * height * 2 * 4 where 2 is for directions and 4 is
- * for the maximum number of MB (4 MB in case of IS_8x8) */
- AVMotionVector *mvs = av_malloc_array(mb_width * mb_height, 2 * 4 * sizeof(AVMotionVector));
- if (!mvs)
- return;
-
- for (mb_y = 0; mb_y < mb_height; mb_y++) {
- for (mb_x = 0; mb_x < mb_width; mb_x++) {
- int i, direction, mb_type = mbtype_table[mb_x + mb_y * mb_stride];
- for (direction = 0; direction < 2; direction++) {
- if (!USES_LIST(mb_type, direction))
- continue;
- if (IS_8X8(mb_type)) {
- for (i = 0; i < 4; i++) {
- int sx = mb_x * 16 + 4 + 8 * (i & 1);
- int sy = mb_y * 16 + 4 + 8 * (i >> 1);
- int xy = (mb_x * 2 + (i & 1) +
- (mb_y * 2 + (i >> 1)) * mv_stride) << (mv_sample_log2 - 1);
- int mx = motion_val[direction][xy][0];
- int my = motion_val[direction][xy][1];
- mbcount += add_mb(mvs + mbcount, mb_type, sx, sy, mx, my, scale, direction);
- }
- } else if (IS_16X8(mb_type)) {
- for (i = 0; i < 2; i++) {
- int sx = mb_x * 16 + 8;
- int sy = mb_y * 16 + 4 + 8 * i;
- int xy = (mb_x * 2 + (mb_y * 2 + i) * mv_stride) << (mv_sample_log2 - 1);
- int mx = motion_val[direction][xy][0];
- int my = motion_val[direction][xy][1];
-
- if (IS_INTERLACED(mb_type))
- my *= 2;
-
- mbcount += add_mb(mvs + mbcount, mb_type, sx, sy, mx, my, scale, direction);
- }
- } else if (IS_8X16(mb_type)) {
- for (i = 0; i < 2; i++) {
- int sx = mb_x * 16 + 4 + 8 * i;
- int sy = mb_y * 16 + 8;
- int xy = (mb_x * 2 + i + mb_y * 2 * mv_stride) << (mv_sample_log2 - 1);
- int mx = motion_val[direction][xy][0];
- int my = motion_val[direction][xy][1];
-
- if (IS_INTERLACED(mb_type))
- my *= 2;
-
- mbcount += add_mb(mvs + mbcount, mb_type, sx, sy, mx, my, scale, direction);
- }
- } else {
- int sx = mb_x * 16 + 8;
- int sy = mb_y * 16 + 8;
- int xy = (mb_x + mb_y * mv_stride) << mv_sample_log2;
- int mx = motion_val[direction][xy][0];
- int my = motion_val[direction][xy][1];
- mbcount += add_mb(mvs + mbcount, mb_type, sx, sy, mx, my, scale, direction);
- }
- }
- }
- }
-
- if (mbcount) {
- AVFrameSideData *sd;
-
- av_log(avctx, AV_LOG_DEBUG, "Adding %d MVs info to frame %d\n", mbcount, avctx->frame_number);
- sd = av_frame_new_side_data(pict, AV_FRAME_DATA_MOTION_VECTORS, mbcount * sizeof(AVMotionVector));
- if (!sd) {
- av_freep(&mvs);
- return;
- }
- memcpy(sd->data, mvs, mbcount * sizeof(AVMotionVector));
- }
-
- av_freep(&mvs);
- }
-
- /* TODO: export all the following to make them accessible for users (and filters) */
- if (avctx->hwaccel || !mbtype_table)
- return;
-
-
- if (avctx->debug & (FF_DEBUG_SKIP | FF_DEBUG_QP | FF_DEBUG_MB_TYPE)) {
- int x,y;
-
- av_log(avctx, AV_LOG_DEBUG, "New frame, type: %c\n",
- av_get_picture_type_char(pict->pict_type));
- for (y = 0; y < mb_height; y++) {
- for (x = 0; x < mb_width; x++) {
- if (avctx->debug & FF_DEBUG_SKIP) {
- int count = mbskip_table ? mbskip_table[x + y * mb_stride] : 0;
- if (count > 9)
- count = 9;
- av_log(avctx, AV_LOG_DEBUG, "%1d", count);
- }
- if (avctx->debug & FF_DEBUG_QP) {
- av_log(avctx, AV_LOG_DEBUG, "%2d",
- qscale_table[x + y * mb_stride]);
- }
- if (avctx->debug & FF_DEBUG_MB_TYPE) {
- int mb_type = mbtype_table[x + y * mb_stride];
- // Type & MV direction
- if (IS_PCM(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "P");
- else if (IS_INTRA(mb_type) && IS_ACPRED(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "A");
- else if (IS_INTRA4x4(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "i");
- else if (IS_INTRA16x16(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "I");
- else if (IS_DIRECT(mb_type) && IS_SKIP(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "d");
- else if (IS_DIRECT(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "D");
- else if (IS_GMC(mb_type) && IS_SKIP(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "g");
- else if (IS_GMC(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "G");
- else if (IS_SKIP(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "S");
- else if (!USES_LIST(mb_type, 1))
- av_log(avctx, AV_LOG_DEBUG, ">");
- else if (!USES_LIST(mb_type, 0))
- av_log(avctx, AV_LOG_DEBUG, "<");
- else {
- av_assert2(USES_LIST(mb_type, 0) && USES_LIST(mb_type, 1));
- av_log(avctx, AV_LOG_DEBUG, "X");
- }
-
- // segmentation
- if (IS_8X8(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "+");
- else if (IS_16X8(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "-");
- else if (IS_8X16(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "|");
- else if (IS_INTRA(mb_type) || IS_16X16(mb_type))
- av_log(avctx, AV_LOG_DEBUG, " ");
- else
- av_log(avctx, AV_LOG_DEBUG, "?");
-
-
- if (IS_INTERLACED(mb_type))
- av_log(avctx, AV_LOG_DEBUG, "=");
- else
- av_log(avctx, AV_LOG_DEBUG, " ");
- }
- }
- av_log(avctx, AV_LOG_DEBUG, "\n");
- }
- }
-
-#if FF_API_DEBUG_MV
- if ((avctx->debug & (FF_DEBUG_VIS_QP | FF_DEBUG_VIS_MB_TYPE)) ||
- (avctx->debug_mv)) {
- int mb_y;
- int i, ret;
- int h_chroma_shift, v_chroma_shift, block_height;
- const int mv_sample_log2 = avctx->codec_id == AV_CODEC_ID_H264 || avctx->codec_id == AV_CODEC_ID_SVQ3 ? 2 : 1;
- const int mv_stride = (mb_width << mv_sample_log2) +
- (avctx->codec->id == AV_CODEC_ID_H264 ? 0 : 1);
-
- if (low_delay)
- *low_delay = 0; // needed to see the vectors without trashing the buffers
-
- ret = av_pix_fmt_get_chroma_sub_sample (avctx->pix_fmt, &h_chroma_shift, &v_chroma_shift);
- if (ret)
- return ret;
-
- av_frame_make_writable(pict);
-
- pict->opaque = NULL;
- block_height = 16 >> v_chroma_shift;
-
- for (mb_y = 0; mb_y < mb_height; mb_y++) {
- int mb_x;
- for (mb_x = 0; mb_x < mb_width; mb_x++) {
- const int mb_index = mb_x + mb_y * mb_stride;
- if ((avctx->debug & FF_DEBUG_VIS_QP)) {
- uint64_t c = (qscale_table[mb_index] * 128 / 31) *
- 0x0101010101010101ULL;
- int y;
- for (y = 0; y < block_height; y++) {
- *(uint64_t *)(pict->data[1] + 8 * mb_x +
- (block_height * mb_y + y) *
- pict->linesize[1]) = c;
- *(uint64_t *)(pict->data[2] + 8 * mb_x +
- (block_height * mb_y + y) *
- pict->linesize[2]) = c;
- }
- }
- if ((avctx->debug & FF_DEBUG_VIS_MB_TYPE) &&
- motion_val[0]) {
- int mb_type = mbtype_table[mb_index];
- uint64_t u,v;
- int y;
-#define COLOR(theta, r) \
- u = (int)(128 + r * cos(theta * M_PI / 180)); \
- v = (int)(128 + r * sin(theta * M_PI / 180));
-
-
- u = v = 128;
- if (IS_PCM(mb_type)) {
- COLOR(120, 48)
- } else if ((IS_INTRA(mb_type) && IS_ACPRED(mb_type)) ||
- IS_INTRA16x16(mb_type)) {
- COLOR(30, 48)
- } else if (IS_INTRA4x4(mb_type)) {
- COLOR(90, 48)
- } else if (IS_DIRECT(mb_type) && IS_SKIP(mb_type)) {
- // COLOR(120, 48)
- } else if (IS_DIRECT(mb_type)) {
- COLOR(150, 48)
- } else if (IS_GMC(mb_type) && IS_SKIP(mb_type)) {
- COLOR(170, 48)
- } else if (IS_GMC(mb_type)) {
- COLOR(190, 48)
- } else if (IS_SKIP(mb_type)) {
- // COLOR(180, 48)
- } else if (!USES_LIST(mb_type, 1)) {
- COLOR(240, 48)
- } else if (!USES_LIST(mb_type, 0)) {
- COLOR(0, 48)
- } else {
- av_assert2(USES_LIST(mb_type, 0) && USES_LIST(mb_type, 1));
- COLOR(300,48)
- }
-
- u *= 0x0101010101010101ULL;
- v *= 0x0101010101010101ULL;
- for (y = 0; y < block_height; y++) {
- *(uint64_t *)(pict->data[1] + 8 * mb_x +
- (block_height * mb_y + y) * pict->linesize[1]) = u;
- *(uint64_t *)(pict->data[2] + 8 * mb_x +
- (block_height * mb_y + y) * pict->linesize[2]) = v;
- }
-
- // segmentation
- if (IS_8X8(mb_type) || IS_16X8(mb_type)) {
- *(uint64_t *)(pict->data[0] + 16 * mb_x + 0 +
- (16 * mb_y + 8) * pict->linesize[0]) ^= 0x8080808080808080ULL;
- *(uint64_t *)(pict->data[0] + 16 * mb_x + 8 +
- (16 * mb_y + 8) * pict->linesize[0]) ^= 0x8080808080808080ULL;
- }
- if (IS_8X8(mb_type) || IS_8X16(mb_type)) {
- for (y = 0; y < 16; y++)
- pict->data[0][16 * mb_x + 8 + (16 * mb_y + y) *
- pict->linesize[0]] ^= 0x80;
- }
- if (IS_8X8(mb_type) && mv_sample_log2 >= 2) {
- int dm = 1 << (mv_sample_log2 - 2);
- for (i = 0; i < 4; i++) {
- int sx = mb_x * 16 + 8 * (i & 1);
- int sy = mb_y * 16 + 8 * (i >> 1);
- int xy = (mb_x * 2 + (i & 1) +
- (mb_y * 2 + (i >> 1)) * mv_stride) << (mv_sample_log2 - 1);
- // FIXME bidir
- int32_t *mv = (int32_t *) &motion_val[0][xy];
- if (mv[0] != mv[dm] ||
- mv[dm * mv_stride] != mv[dm * (mv_stride + 1)])
- for (y = 0; y < 8; y++)
- pict->data[0][sx + 4 + (sy + y) * pict->linesize[0]] ^= 0x80;
- if (mv[0] != mv[dm * mv_stride] || mv[dm] != mv[dm * (mv_stride + 1)])
- *(uint64_t *)(pict->data[0] + sx + (sy + 4) *
- pict->linesize[0]) ^= 0x8080808080808080ULL;
- }
- }
-
- if (IS_INTERLACED(mb_type) &&
- avctx->codec->id == AV_CODEC_ID_H264) {
- // hmm
- }
- }
- if (mbskip_table)
- mbskip_table[mb_index] = 0;
- }
- }
- }
-#endif
-}
-
void ff_print_debug_info(MpegEncContext *s, Picture *p, AVFrame *pict)
{
ff_print_debug_info2(s->avctx, pict, s->mbskip_table, p->mb_type,
ff_wmv2_add_mb(s, block, dest_y, dest_cb, dest_cr);
}
} else {
+ /* Only MPEG-4 Simple Studio Profile is supported in > 8-bit mode.
+ TODO: Integrate 10-bit properly into mpegvideo.c so that ER works properly */
+ if (s->avctx->bits_per_raw_sample > 8){
+ const int act_block_size = block_size * 2;
+ s->idsp.idct_put(dest_y, dct_linesize, (int16_t*)(*s->block32)[0]);
+ s->idsp.idct_put(dest_y + act_block_size, dct_linesize, (int16_t*)(*s->block32)[1]);
+ s->idsp.idct_put(dest_y + dct_offset, dct_linesize, (int16_t*)(*s->block32)[2]);
+ s->idsp.idct_put(dest_y + dct_offset + act_block_size, dct_linesize, (int16_t*)(*s->block32)[3]);
+
+ dct_linesize = uvlinesize << s->interlaced_dct;
+ dct_offset = s->interlaced_dct ? uvlinesize : uvlinesize*block_size;
+
+ s->idsp.idct_put(dest_cb, dct_linesize, (int16_t*)(*s->block32)[4]);
+ s->idsp.idct_put(dest_cr, dct_linesize, (int16_t*)(*s->block32)[5]);
+ s->idsp.idct_put(dest_cb + dct_offset, dct_linesize, (int16_t*)(*s->block32)[6]);
+ s->idsp.idct_put(dest_cr + dct_offset, dct_linesize, (int16_t*)(*s->block32)[7]);
+ if(!s->chroma_x_shift){//Chroma444
+ s->idsp.idct_put(dest_cb + act_block_size, dct_linesize, (int16_t*)(*s->block32)[8]);
+ s->idsp.idct_put(dest_cr + act_block_size, dct_linesize, (int16_t*)(*s->block32)[9]);
+ s->idsp.idct_put(dest_cb + act_block_size + dct_offset, dct_linesize, (int16_t*)(*s->block32)[10]);
+ s->idsp.idct_put(dest_cr + act_block_size + dct_offset, dct_linesize, (int16_t*)(*s->block32)[11]);
+ }
+ }
/* dct only in intra block */
- if(s->encoding || !(s->codec_id==AV_CODEC_ID_MPEG1VIDEO || s->codec_id==AV_CODEC_ID_MPEG2VIDEO)){
+ else if(s->encoding || !(s->codec_id==AV_CODEC_ID_MPEG1VIDEO || s->codec_id==AV_CODEC_ID_MPEG2VIDEO)){
put_dct(s, block[0], 0, dest_y , dct_linesize, s->qscale);
put_dct(s, block[1], 1, dest_y + block_size, dct_linesize, s->qscale);
put_dct(s, block[2], 2, dest_y + dct_offset , dct_linesize, s->qscale);
void ff_init_block_index(MpegEncContext *s){ //FIXME maybe rename
const int linesize = s->current_picture.f->linesize[0]; //not s->linesize as this would be wrong for field pics
const int uvlinesize = s->current_picture.f->linesize[1];
- const int mb_size= 4 - s->avctx->lowres;
+ const int width_of_mb = (4 + (s->avctx->bits_per_raw_sample > 8)) - s->avctx->lowres;
+ const int height_of_mb = 4 - s->avctx->lowres;
s->block_index[0]= s->b8_stride*(s->mb_y*2 ) - 2 + s->mb_x*2;
s->block_index[1]= s->b8_stride*(s->mb_y*2 ) - 1 + s->mb_x*2;
s->block_index[5]= s->mb_stride*(s->mb_y + s->mb_height + 2) + s->b8_stride*s->mb_height*2 + s->mb_x - 1;
//block_index is not used by mpeg2, so it is not affected by chroma_format
- s->dest[0] = s->current_picture.f->data[0] + (int)((s->mb_x - 1U) << mb_size);
- s->dest[1] = s->current_picture.f->data[1] + (int)((s->mb_x - 1U) << (mb_size - s->chroma_x_shift));
- s->dest[2] = s->current_picture.f->data[2] + (int)((s->mb_x - 1U) << (mb_size - s->chroma_x_shift));
+ s->dest[0] = s->current_picture.f->data[0] + (int)((s->mb_x - 1U) << width_of_mb);
+ s->dest[1] = s->current_picture.f->data[1] + (int)((s->mb_x - 1U) << (width_of_mb - s->chroma_x_shift));
+ s->dest[2] = s->current_picture.f->data[2] + (int)((s->mb_x - 1U) << (width_of_mb - s->chroma_x_shift));
if(!(s->pict_type==AV_PICTURE_TYPE_B && s->avctx->draw_horiz_band && s->picture_structure==PICT_FRAME))
{
if(s->picture_structure==PICT_FRAME){
- s->dest[0] += s->mb_y * linesize << mb_size;
- s->dest[1] += s->mb_y * uvlinesize << (mb_size - s->chroma_y_shift);
- s->dest[2] += s->mb_y * uvlinesize << (mb_size - s->chroma_y_shift);
+ s->dest[0] += s->mb_y * linesize << height_of_mb;
+ s->dest[1] += s->mb_y * uvlinesize << (height_of_mb - s->chroma_y_shift);
+ s->dest[2] += s->mb_y * uvlinesize << (height_of_mb - s->chroma_y_shift);
}else{
- s->dest[0] += (s->mb_y>>1) * linesize << mb_size;
- s->dest[1] += (s->mb_y>>1) * uvlinesize << (mb_size - s->chroma_y_shift);
- s->dest[2] += (s->mb_y>>1) * uvlinesize << (mb_size - s->chroma_y_shift);
+ s->dest[0] += (s->mb_y>>1) * linesize << height_of_mb;
+ s->dest[1] += (s->mb_y>>1) * uvlinesize << (height_of_mb - s->chroma_y_shift);
+ s->dest[2] += (s->mb_y>>1) * uvlinesize << (height_of_mb - s->chroma_y_shift);
av_assert1((s->mb_y&1) == (s->picture_structure == PICT_BOTTOM_FIELD));
}
}