X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=libavcodec%2Fproresenc_anatoliy.c;h=8bc13fd576780a432e2c9b6eb6d774fe81bff7de;hb=60e7021064ea9e99854c585d452e89f38635480e;hp=b9524b3d997b2f3f5ba2a52faad796df2b000554;hpb=bbbbf237597e116f6af9094b57e979a7acc4afcb;p=ffmpeg diff --git a/libavcodec/proresenc_anatoliy.c b/libavcodec/proresenc_anatoliy.c index b9524b3d997..8bc13fd5767 100644 --- a/libavcodec/proresenc_anatoliy.c +++ b/libavcodec/proresenc_anatoliy.c @@ -45,20 +45,22 @@ static const AVProfile profiles[] = { { FF_PROFILE_PRORES_STANDARD, "apcn"}, { FF_PROFILE_PRORES_HQ, "apch"}, { FF_PROFILE_PRORES_4444, "ap4h"}, + { FF_PROFILE_PRORES_XQ, "ap4x"}, { FF_PROFILE_UNKNOWN } }; -static const int qp_start_table[5] = { 8, 3, 2, 1, 1}; -static const int qp_end_table[5] = { 13, 9, 6, 6, 5}; -static const int bitrate_table[5] = { 1000, 2100, 3500, 5400, 7000}; +static const int qp_start_table[] = { 8, 3, 2, 1, 1, 1}; +static const int qp_end_table[] = { 13, 9, 6, 6, 5, 4}; +static const int bitrate_table[] = { 1000, 2100, 3500, 5400, 7000, 10000}; -static const int valid_primaries[9] = { AVCOL_PRI_RESERVED0, AVCOL_PRI_BT709, AVCOL_PRI_UNSPECIFIED, AVCOL_PRI_BT470BG, - AVCOL_PRI_SMPTE170M, AVCOL_PRI_BT2020, AVCOL_PRI_SMPTE431, AVCOL_PRI_SMPTE432,INT_MAX }; -static const int valid_trc[4] = { AVCOL_TRC_RESERVED0, AVCOL_TRC_BT709, AVCOL_TRC_UNSPECIFIED, INT_MAX }; -static const int valid_colorspace[5] = { AVCOL_SPC_BT709, AVCOL_SPC_UNSPECIFIED, AVCOL_SPC_SMPTE170M, - AVCOL_SPC_BT2020_NCL, INT_MAX }; +static const int valid_primaries[] = { AVCOL_PRI_RESERVED0, AVCOL_PRI_BT709, AVCOL_PRI_UNSPECIFIED, AVCOL_PRI_BT470BG, + AVCOL_PRI_SMPTE170M, AVCOL_PRI_BT2020, AVCOL_PRI_SMPTE431, AVCOL_PRI_SMPTE432, INT_MAX }; +static const int valid_trc[] = { AVCOL_TRC_RESERVED0, AVCOL_TRC_BT709, AVCOL_TRC_UNSPECIFIED, AVCOL_TRC_SMPTE2084, + AVCOL_TRC_ARIB_STD_B67, INT_MAX }; +static const int valid_colorspace[] = { AVCOL_SPC_BT709, AVCOL_SPC_UNSPECIFIED, AVCOL_SPC_SMPTE170M, + AVCOL_SPC_BT2020_NCL, INT_MAX }; -static const uint8_t QMAT_LUMA[5][64] = { +static const uint8_t QMAT_LUMA[6][64] = { { 4, 7, 9, 11, 13, 14, 15, 63, 7, 7, 11, 12, 14, 15, 63, 63, @@ -104,10 +106,19 @@ static const uint8_t QMAT_LUMA[5][64] = { 4, 4, 4, 4, 4, 5, 5, 6, 4, 4, 4, 4, 5, 5, 6, 7, 4, 4, 4, 4, 5, 6, 7, 7 + }, { /* 444 XQ */ + 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 3, + 2, 2, 2, 2, 2, 2, 3, 3, + 2, 2, 2, 2, 2, 3, 3, 3, + 2, 2, 2, 2, 3, 3, 3, 4, + 2, 2, 2, 2, 3, 3, 4, 4, } }; -static const uint8_t QMAT_CHROMA[5][64] = { +static const uint8_t QMAT_CHROMA[6][64] = { { 4, 7, 9, 11, 13, 14, 63, 63, 7, 7, 11, 12, 14, 63, 63, 63, @@ -153,6 +164,15 @@ static const uint8_t QMAT_CHROMA[5][64] = { 4, 4, 4, 4, 4, 5, 5, 6, 4, 4, 4, 4, 5, 5, 6, 7, 4, 4, 4, 4, 5, 6, 7, 7 + }, { /* 444 xq */ + 4, 4, 4, 4, 4, 4, 4, 4, + 4, 4, 4, 4, 4, 4, 4, 4, + 4, 4, 4, 4, 4, 4, 4, 4, + 4, 4, 4, 4, 4, 4, 4, 5, + 4, 4, 4, 4, 4, 4, 5, 5, + 4, 4, 4, 4, 4, 5, 5, 6, + 4, 4, 4, 4, 5, 5, 6, 7, + 4, 4, 4, 4, 5, 6, 7, 7 } }; @@ -167,9 +187,11 @@ typedef struct { int qmat_luma[16][64]; int qmat_chroma[16][64]; + const uint8_t *scantable; int is_422; int need_alpha; + int is_interlaced; char *vendor; } ProresContext; @@ -203,7 +225,7 @@ static void encode_codeword(PutBitContext *pb, int val, int codebook) } #define QSCALE(qmat,ind,val) ((val) / ((qmat)[ind])) -#define TO_GOLOMB(val) (((val) << 1) ^ ((val) >> 31)) +#define TO_GOLOMB(val) (((val) * 2) ^ ((val) >> 31)) #define DIFF_SIGN(val, sign) (((val) >> 31) ^ (sign)) #define IS_NEGATIVE(val) ((((val) >> 31) ^ -1) + 1) #define TO_GOLOMB2(val,sign) ((val)==0 ? 0 : ((val) << 1) + (sign)) @@ -250,14 +272,14 @@ static const uint8_t lev_to_cb[10] = { 0x04, 0x0A, 0x05, 0x06, 0x04, 0x28, 0x28, 0x28, 0x28, 0x4C }; static void encode_ac_coeffs(PutBitContext *pb, - int16_t *in, int blocks_per_slice, int *qmat) + int16_t *in, int blocks_per_slice, int *qmat, const uint8_t ff_prores_scan[64]) { int prev_run = 4; int prev_level = 2; int run = 0, level, code, i, j; for (i = 1; i < 64; i++) { - int indp = ff_prores_progressive_scan[i]; + int indp = ff_prores_scan[i]; for (j = 0; j < blocks_per_slice; j++) { int val = QSCALE(qmat, indp, in[(j << 6) + indp]); if (val) { @@ -335,7 +357,8 @@ static void calc_plane_dct(FDCTDSPContext *fdsp, uint8_t *src, int16_t * blocks, } } -static int encode_slice_plane(int16_t *blocks, int mb_count, uint8_t *buf, unsigned buf_size, int *qmat, int sub_sample_chroma) +static int encode_slice_plane(int16_t *blocks, int mb_count, uint8_t *buf, unsigned buf_size, int *qmat, int sub_sample_chroma, + const uint8_t ff_prores_scan[64]) { int blocks_per_slice; PutBitContext pb; @@ -344,7 +367,7 @@ static int encode_slice_plane(int16_t *blocks, int mb_count, uint8_t *buf, unsig init_put_bits(&pb, buf, buf_size); encode_dc_coeffs(&pb, blocks, blocks_per_slice, qmat); - encode_ac_coeffs(&pb, blocks, blocks_per_slice, qmat); + encode_ac_coeffs(&pb, blocks, blocks_per_slice, qmat, ff_prores_scan); flush_put_bits(&pb); return put_bits_ptr(&pb) - pb.buf; @@ -359,15 +382,15 @@ static av_always_inline unsigned encode_slice_data(AVCodecContext *avctx, ProresContext* ctx = avctx->priv_data; *y_data_size = encode_slice_plane(blocks_y, mb_count, - buf, data_size, ctx->qmat_luma[qp - 1], 0); + buf, data_size, ctx->qmat_luma[qp - 1], 0, ctx->scantable); if (!(avctx->flags & AV_CODEC_FLAG_GRAY)) { *u_data_size = encode_slice_plane(blocks_u, mb_count, buf + *y_data_size, data_size - *y_data_size, - ctx->qmat_chroma[qp - 1], ctx->is_422); + ctx->qmat_chroma[qp - 1], ctx->is_422, ctx->scantable); *v_data_size = encode_slice_plane(blocks_v, mb_count, buf + *y_data_size + *u_data_size, data_size - *y_data_size - *u_data_size, - ctx->qmat_chroma[qp - 1], ctx->is_422); + ctx->qmat_chroma[qp - 1], ctx->is_422, ctx->scantable); } return *y_data_size + *u_data_size + *v_data_size; @@ -447,22 +470,40 @@ static av_always_inline int encode_alpha_slice_data(AVCodecContext *avctx, int8_ } } -static void subimage_with_fill(uint16_t *src, unsigned x, unsigned y, - unsigned stride, unsigned width, unsigned height, uint16_t *dst, - unsigned dst_width, unsigned dst_height) +static inline void subimage_with_fill_template(uint16_t *src, unsigned x, unsigned y, + unsigned stride, unsigned width, unsigned height, uint16_t *dst, + unsigned dst_width, unsigned dst_height, int is_alpha_plane, + int is_interlaced, int is_top_field) { - int box_width = FFMIN(width - x, dst_width); - int box_height = FFMIN(height - y, dst_height); - int i, j, src_stride = stride >> 1; + int i, j, src_stride, box_height; uint16_t last_pix, *last_line; - src += y * src_stride + x; + if (!is_interlaced) { + src_stride = stride >> 1; + src += y * src_stride + x; + box_height = FFMIN(height - y, dst_height); + } else { + src_stride = stride; /* 2 lines stride */ + src += y * src_stride + x; + box_height = FFMIN(height/2 - y, dst_height); + if (!is_top_field) + src += stride >> 1; + } + for (i = 0; i < box_height; ++i) { for (j = 0; j < box_width; ++j) { - dst[j] = src[j]; + if (!is_alpha_plane) { + dst[j] = src[j]; + } else { + dst[j] = src[j] << 6; /* alpha 10b to 16b */ + } + } + if (!is_alpha_plane) { + last_pix = dst[j - 1]; + } else { + last_pix = dst[j - 1] << 6; /* alpha 10b to 16b */ } - last_pix = dst[j - 1]; for (; j < dst_width; j++) dst[j] = last_pix; src += src_stride; @@ -477,39 +518,24 @@ static void subimage_with_fill(uint16_t *src, unsigned x, unsigned y, } } +static void subimage_with_fill(uint16_t *src, unsigned x, unsigned y, + unsigned stride, unsigned width, unsigned height, uint16_t *dst, + unsigned dst_width, unsigned dst_height, int is_interlaced, int is_top_field) +{ + subimage_with_fill_template(src, x, y, stride, width, height, dst, dst_width, dst_height, 0, is_interlaced, is_top_field); +} + /* reorganize alpha data and convert 10b -> 16b */ static void subimage_alpha_with_fill(uint16_t *src, unsigned x, unsigned y, unsigned stride, unsigned width, unsigned height, uint16_t *dst, - unsigned dst_width, unsigned dst_height) + unsigned dst_width, unsigned dst_height, int is_interlaced, int is_top_field) { - int box_width = FFMIN(width - x, dst_width); - int box_height = FFMIN(height - y, dst_height); - int i, j, src_stride = stride >> 1; - uint16_t last_pix, *last_line; - - src += y * src_stride + x; - for (i = 0; i < box_height; ++i) { - for (j = 0; j < box_width; ++j) { - dst[j] = src[j] << 6; /* 10b to 16b */ - } - last_pix = dst[j - 1] << 6; /* 10b to 16b */ - for (; j < dst_width; j++) - dst[j] = last_pix; - src += src_stride; - dst += dst_width; - } - last_line = dst - dst_width; - for (; i < dst_height; i++) { - for (j = 0; j < dst_width; ++j) { - dst[j] = last_line[j]; - } - dst += dst_width; - } + subimage_with_fill_template(src, x, y, stride, width, height, dst, dst_width, dst_height, 1, is_interlaced, is_top_field); } static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, int mb_x, int mb_y, unsigned mb_count, uint8_t *buf, unsigned data_size, - int unsafe, int *qp) + int unsafe, int *qp, int is_interlaced, int is_top_field) { int luma_stride, chroma_stride, alpha_stride = 0; ProresContext* ctx = avctx->priv_data; @@ -532,21 +558,33 @@ static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, int mb_x, if (ctx->need_alpha) alpha_stride = pic->linesize[3]; - dest_y = pic->data[0] + (mb_y << 4) * luma_stride + (mb_x << 5); - dest_u = pic->data[1] + (mb_y << 4) * chroma_stride + (mb_x << (5 - ctx->is_422)); - dest_v = pic->data[2] + (mb_y << 4) * chroma_stride + (mb_x << (5 - ctx->is_422)); + if (!is_interlaced) { + dest_y = pic->data[0] + (mb_y << 4) * luma_stride + (mb_x << 5); + dest_u = pic->data[1] + (mb_y << 4) * chroma_stride + (mb_x << (5 - ctx->is_422)); + dest_v = pic->data[2] + (mb_y << 4) * chroma_stride + (mb_x << (5 - ctx->is_422)); + } else { + dest_y = pic->data[0] + (mb_y << 4) * luma_stride * 2 + (mb_x << 5); + dest_u = pic->data[1] + (mb_y << 4) * chroma_stride * 2 + (mb_x << (5 - ctx->is_422)); + dest_v = pic->data[2] + (mb_y << 4) * chroma_stride * 2 + (mb_x << (5 - ctx->is_422)); + if (!is_top_field){ /* bottom field, offset dest */ + dest_y += luma_stride; + dest_u += chroma_stride; + dest_v += chroma_stride; + } + } if (unsafe) { subimage_with_fill((uint16_t *) pic->data[0], mb_x << 4, mb_y << 4, luma_stride, avctx->width, avctx->height, - (uint16_t *) ctx->fill_y, mb_count << 4, 16); + (uint16_t *) ctx->fill_y, mb_count << 4, 16, is_interlaced, is_top_field); subimage_with_fill((uint16_t *) pic->data[1], mb_x << (4 - ctx->is_422), mb_y << 4, chroma_stride, avctx->width >> ctx->is_422, avctx->height, - (uint16_t *) ctx->fill_u, mb_count << (4 - ctx->is_422), 16); + (uint16_t *) ctx->fill_u, mb_count << (4 - ctx->is_422), 16, is_interlaced, is_top_field); subimage_with_fill((uint16_t *) pic->data[2], mb_x << (4 - ctx->is_422), mb_y << 4, chroma_stride, avctx->width >> ctx->is_422, avctx->height, - (uint16_t *) ctx->fill_v, mb_count << (4 - ctx->is_422), 16); + (uint16_t *) ctx->fill_v, mb_count << (4 - ctx->is_422), 16, is_interlaced, is_top_field); + /* no need for interlaced special case, data already reorganized in subimage_with_fill */ calc_plane_dct(fdsp, ctx->fill_y, blocks_y, mb_count << 5, mb_count, 0, 0); calc_plane_dct(fdsp, ctx->fill_u, blocks_u, mb_count << (5 - ctx->is_422), mb_count, 1, ctx->is_422); calc_plane_dct(fdsp, ctx->fill_v, blocks_v, mb_count << (5 - ctx->is_422), mb_count, 1, ctx->is_422); @@ -556,9 +594,15 @@ static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, int mb_x, &y_data_size, &u_data_size, &v_data_size, *qp); } else { - calc_plane_dct(fdsp, dest_y, blocks_y, luma_stride, mb_count, 0, 0); - calc_plane_dct(fdsp, dest_u, blocks_u, chroma_stride, mb_count, 1, ctx->is_422); - calc_plane_dct(fdsp, dest_v, blocks_v, chroma_stride, mb_count, 1, ctx->is_422); + if (!is_interlaced) { + calc_plane_dct(fdsp, dest_y, blocks_y, luma_stride, mb_count, 0, 0); + calc_plane_dct(fdsp, dest_u, blocks_u, chroma_stride, mb_count, 1, ctx->is_422); + calc_plane_dct(fdsp, dest_v, blocks_v, chroma_stride, mb_count, 1, ctx->is_422); + } else { + calc_plane_dct(fdsp, dest_y, blocks_y, luma_stride * 2, mb_count, 0, 0); + calc_plane_dct(fdsp, dest_u, blocks_u, chroma_stride * 2, mb_count, 1, ctx->is_422); + calc_plane_dct(fdsp, dest_v, blocks_v, chroma_stride * 2, mb_count, 1, ctx->is_422); + } slice_size = encode_slice_data(avctx, blocks_y, blocks_u, blocks_v, mb_count, buf + hdr_size, data_size - hdr_size, @@ -595,7 +639,7 @@ static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, int mb_x, subimage_alpha_with_fill((uint16_t *) pic->data[3], mb_x << 4, mb_y << 4, alpha_stride, avctx->width, avctx->height, - (uint16_t *) ctx->fill_a, mb_count << 4, 16); + (uint16_t *) ctx->fill_a, mb_count << 4, 16, is_interlaced, is_top_field); ret = encode_alpha_slice_data(avctx, ctx->fill_a, mb_count, buf + hdr_size + slice_size, data_size - hdr_size - slice_size, &a_data_size); @@ -608,16 +652,29 @@ static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, int mb_x, } static int prores_encode_picture(AVCodecContext *avctx, const AVFrame *pic, - uint8_t *buf, const int buf_size) + uint8_t *buf, const int buf_size, const int picture_index, const int is_top_field) { + ProresContext *ctx = avctx->priv_data; int mb_width = (avctx->width + 15) >> 4; - int mb_height = (avctx->height + 15) >> 4; int hdr_size, sl_size, i; - int mb_y, sl_data_size, qp; + int mb_y, sl_data_size, qp, mb_height, picture_height, unsafe_mb_height_limit; int unsafe_bot, unsafe_right; uint8_t *sl_data, *sl_data_sizes; int slice_per_line = 0, rem = mb_width; + if (!ctx->is_interlaced) { /* progressive encoding */ + mb_height = (avctx->height + 15) >> 4; + unsafe_mb_height_limit = mb_height; + } else { + if (is_top_field) { + picture_height = (avctx->height + 1) / 2; + } else { + picture_height = avctx->height / 2; + } + mb_height = (picture_height + 15) >> 4; + unsafe_mb_height_limit = mb_height; + } + for (i = av_log2(DEFAULT_SLICE_MB_WIDTH); i >= 0; --i) { slice_per_line += rem >> i; rem &= (1 << i) - 1; @@ -634,11 +691,11 @@ static int prores_encode_picture(AVCodecContext *avctx, const AVFrame *pic, while (mb_width - mb_x < slice_mb_count) slice_mb_count >>= 1; - unsafe_bot = (avctx->height & 0xf) && (mb_y == mb_height - 1); + unsafe_bot = (avctx->height & 0xf) && (mb_y == unsafe_mb_height_limit - 1); unsafe_right = (avctx->width & 0xf) && (mb_x + slice_mb_count == mb_width); sl_size = encode_slice(avctx, pic, mb_x, mb_y, slice_mb_count, - sl_data, sl_data_size, unsafe_bot || unsafe_right, &qp); + sl_data, sl_data_size, unsafe_bot || unsafe_right, &qp, ctx->is_interlaced, is_top_field); if (sl_size < 0){ return sl_size; } @@ -652,8 +709,8 @@ static int prores_encode_picture(AVCodecContext *avctx, const AVFrame *pic, buf[0] = hdr_size << 3; AV_WB32(buf + 1, sl_data - buf); - AV_WB16(buf + 5, slice_per_line * mb_height); - buf[7] = av_log2(DEFAULT_SLICE_MB_WIDTH) << 4; + AV_WB16(buf + 5, slice_per_line * mb_height); /* picture size */ + buf[7] = av_log2(DEFAULT_SLICE_MB_WIDTH) << 4; /* number of slices */ return sl_data - buf; } @@ -664,7 +721,8 @@ static int prores_encode_frame(AVCodecContext *avctx, AVPacket *pkt, ProresContext *ctx = avctx->priv_data; int header_size = 148; uint8_t *buf; - int pic_size, ret; + int compress_frame_size, pic_size, ret, is_top_field_first = 0; + uint8_t frame_flags; int frame_size = FFALIGN(avctx->width, 16) * FFALIGN(avctx->height, 16)*16 + 500 + AV_INPUT_BUFFER_MIN_SIZE; //FIXME choose tighter limit @@ -672,13 +730,9 @@ static int prores_encode_frame(AVCodecContext *avctx, AVPacket *pkt, return ret; buf = pkt->data; - pic_size = prores_encode_picture(avctx, pict, buf + header_size + 8, - pkt->size - header_size - 8); - if (pic_size < 0) { - return pic_size; - } + compress_frame_size = 8 + header_size; - bytestream_put_be32(&buf, pic_size + 8 + header_size); + bytestream_put_be32(&buf, compress_frame_size);/* frame size will be update after picture(s) encoding */ bytestream_put_buffer(&buf, "icpf", 4); bytestream_put_be16(&buf, header_size); @@ -686,11 +740,22 @@ static int prores_encode_frame(AVCodecContext *avctx, AVPacket *pkt, bytestream_put_buffer(&buf, ctx->vendor, 4); bytestream_put_be16(&buf, avctx->width); bytestream_put_be16(&buf, avctx->height); - if (avctx->profile == FF_PROFILE_PRORES_4444) { - *buf++ = 0xC2; // 444, not interlaced + frame_flags = 0x82; /* 422 not interlaced */ + if (avctx->profile >= FF_PROFILE_PRORES_4444) /* 4444 or 4444 Xq */ + frame_flags |= 0x40; /* 444 chroma */ + if (ctx->is_interlaced) { + if (pict->top_field_first || !pict->interlaced_frame) { /* tff frame or progressive frame interpret as tff */ + av_log(avctx, AV_LOG_DEBUG, "use interlaced encoding, top field first\n"); + frame_flags |= 0x04; /* interlaced tff */ + is_top_field_first = 1; + } else { + av_log(avctx, AV_LOG_DEBUG, "use interlaced encoding, bottom field first\n"); + frame_flags |= 0x08; /* interlaced bff */ + } } else { - *buf++ = 0x82; // 422, not interlaced + av_log(avctx, AV_LOG_DEBUG, "use progressive encoding\n"); } + *buf++ = frame_flags; *buf++ = 0; /* reserved */ /* only write color properties, if valid value. set to unspecified otherwise */ *buf++ = ff_int_from_list_or_default(avctx, "frame color primaries", pict->color_primaries, valid_primaries, 0); @@ -711,8 +776,25 @@ static int prores_encode_frame(AVCodecContext *avctx, AVPacket *pkt, bytestream_put_buffer(&buf, QMAT_LUMA[avctx->profile], 64); bytestream_put_buffer(&buf, QMAT_CHROMA[avctx->profile], 64); + pic_size = prores_encode_picture(avctx, pict, buf, + pkt->size - compress_frame_size, 0, is_top_field_first);/* encode progressive or first field */ + if (pic_size < 0) { + return pic_size; + } + compress_frame_size += pic_size; + + if (ctx->is_interlaced) { /* encode second field */ + pic_size = prores_encode_picture(avctx, pict, pkt->data + compress_frame_size, + pkt->size - compress_frame_size, 1, !is_top_field_first); + if (pic_size < 0) { + return pic_size; + } + compress_frame_size += pic_size; + } + + AV_WB32(pkt->data, compress_frame_size);/* update frame size */ pkt->flags |= AV_PKT_FLAG_KEY; - pkt->size = pic_size + 8 + header_size; + pkt->size = compress_frame_size; *got_packet = 1; return 0; @@ -732,6 +814,12 @@ static av_cold int prores_encode_init(AVCodecContext *avctx) avctx->bits_per_raw_sample = 10; ctx->need_alpha = 0; + ctx->is_interlaced = !!(avctx->flags & AV_CODEC_FLAG_INTERLACED_DCT); + if (ctx->is_interlaced) { + ctx->scantable = ff_prores_interlaced_scan; + } else { + ctx->scantable = ff_prores_progressive_scan; + } if (avctx->width & 0x1) { av_log(avctx, AV_LOG_ERROR, @@ -768,16 +856,16 @@ static av_cold int prores_encode_init(AVCodecContext *avctx) return AVERROR(EINVAL); } } else if (avctx->profile < FF_PROFILE_PRORES_PROXY - || avctx->profile > FF_PROFILE_PRORES_4444) { + || avctx->profile > FF_PROFILE_PRORES_XQ) { av_log( avctx, AV_LOG_ERROR, - "unknown profile %d, use [0 - apco, 1 - apcs, 2 - apcn (default), 3 - apch, 4 - ap4h]\n", + "unknown profile %d, use [0 - apco, 1 - apcs, 2 - apcn (default), 3 - apch, 4 - ap4h, 5 - ap4x]\n", avctx->profile); return AVERROR(EINVAL); } else if ((avctx->pix_fmt == AV_PIX_FMT_YUV422P10) && (avctx->profile > FF_PROFILE_PRORES_HQ)){ av_log(avctx, AV_LOG_ERROR, - "encoding with ProRes 444 (ap4h) profile, need YUV444P10 input\n"); + "encoding with ProRes 444/Xq (ap4h/ap4x) profile, need YUV444P10 input\n"); return AVERROR(EINVAL); } else if ((avctx->pix_fmt == AV_PIX_FMT_YUV444P10 || avctx->pix_fmt == AV_PIX_FMT_YUVA444P10) && (avctx->profile < FF_PROFILE_PRORES_4444)){ @@ -837,7 +925,7 @@ static av_cold int prores_encode_close(AVCodecContext *avctx) #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM static const AVOption options[] = { - { "vendor", "vendor ID", OFFSET(vendor), AV_OPT_TYPE_STRING, { .str = "fmpg" }, CHAR_MIN, CHAR_MAX, VE }, + { "vendor", "vendor ID", OFFSET(vendor), AV_OPT_TYPE_STRING, { .str = "fmpg" }, 0, 0, VE }, { NULL } }; @@ -865,7 +953,7 @@ AVCodec ff_prores_aw_encoder = { .close = prores_encode_close, .encode2 = prores_encode_frame, .pix_fmts = (const enum AVPixelFormat[]){AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, AV_PIX_FMT_YUVA444P10, AV_PIX_FMT_NONE}, - .capabilities = AV_CODEC_CAP_FRAME_THREADS | AV_CODEC_CAP_INTRA_ONLY, + .capabilities = AV_CODEC_CAP_FRAME_THREADS, .priv_class = &proresaw_enc_class, .profiles = NULL_IF_CONFIG_SMALL(ff_prores_profiles), }; @@ -880,7 +968,7 @@ AVCodec ff_prores_encoder = { .close = prores_encode_close, .encode2 = prores_encode_frame, .pix_fmts = (const enum AVPixelFormat[]){AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, AV_PIX_FMT_YUVA444P10, AV_PIX_FMT_NONE}, - .capabilities = AV_CODEC_CAP_FRAME_THREADS | AV_CODEC_CAP_INTRA_ONLY, + .capabilities = AV_CODEC_CAP_FRAME_THREADS, .priv_class = &prores_enc_class, .profiles = NULL_IF_CONFIG_SMALL(ff_prores_profiles), };