}
avcodec_get_chroma_sub_sample(avc_context->pix_fmt, &h->uv_hshift, &h->uv_vshift);
- if (avc_context->flags & CODEC_FLAG_QSCALE) {
+ if (avc_context->flags & AV_CODEC_FLAG_QSCALE) {
/* Clip global_quality in QP units to the [0 - 10] range
to be consistent with the libvorbis implementation.
Theora accepts a quality parameter which is an int value in
}
// need to enable 2 pass (via TH_ENCCTL_2PASS_) before encoding headers
- if (avc_context->flags & CODEC_FLAG_PASS1) {
+ if (avc_context->flags & AV_CODEC_FLAG_PASS1) {
if ((ret = get_stats(avc_context, 0)) < 0)
return ret;
- } else if (avc_context->flags & CODEC_FLAG_PASS2) {
+ } else if (avc_context->flags & AV_CODEC_FLAG_PASS2) {
if ((ret = submit_stats(avc_context)) < 0)
return ret;
}
th_comment_clear(&t_comment);
- /* Set up the output AVFrame */
- avc_context->coded_frame = av_frame_alloc();
- if (!avc_context->coded_frame)
- return AVERROR(ENOMEM);
-
return 0;
}
// EOS, finish and get 1st pass stats if applicable
if (!frame) {
th_encode_packetout(h->t_state, 1, &o_packet);
- if (avc_context->flags & CODEC_FLAG_PASS1)
+ if (avc_context->flags & AV_CODEC_FLAG_PASS1)
if ((ret = get_stats(avc_context, 1)) < 0)
return ret;
return 0;
t_yuv_buffer[i].data = frame->data[i];
}
- if (avc_context->flags & CODEC_FLAG_PASS2)
+ if (avc_context->flags & AV_CODEC_FLAG_PASS2)
if ((ret = submit_stats(avc_context)) < 0)
return ret;
return AVERROR_EXTERNAL;
}
- if (avc_context->flags & CODEC_FLAG_PASS1)
+ if (avc_context->flags & AV_CODEC_FLAG_PASS1)
if ((ret = get_stats(avc_context, 0)) < 0)
return ret;
}
/* Copy ogg_packet content out to buffer */
- if ((ret = ff_alloc_packet2(avc_context, pkt, o_packet.bytes)) < 0)
+ if ((ret = ff_alloc_packet2(avc_context, pkt, o_packet.bytes, 0)) < 0)
return ret;
memcpy(pkt->data, o_packet.packet, o_packet.bytes);
// HACK: assumes no encoder delay, this is true until libtheora becomes
// multithreaded (which will be disabled unless explicitly requested)
pkt->pts = pkt->dts = frame->pts;
+#if FF_API_CODED_FRAME
+FF_DISABLE_DEPRECATION_WARNINGS
avc_context->coded_frame->key_frame = !(o_packet.granulepos & h->keyframe_mask);
+FF_ENABLE_DEPRECATION_WARNINGS
+#endif
if (!(o_packet.granulepos & h->keyframe_mask))
pkt->flags |= AV_PKT_FLAG_KEY;
*got_packet = 1;
th_encode_free(h->t_state);
av_freep(&h->stats);
- av_frame_free(&avc_context->coded_frame);
av_freep(&avc_context->stats_out);
av_freep(&avc_context->extradata);
avc_context->extradata_size = 0;
.init = encode_init,
.close = encode_close,
.encode2 = encode_frame,
- .capabilities = CODEC_CAP_DELAY, // needed to get the statsfile summary
+ .capabilities = AV_CODEC_CAP_DELAY, // needed to get the statsfile summary
.pix_fmts = (const enum AVPixelFormat[]){
AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_NONE
},