X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=ffmpeg_capture.cpp;h=1ffa9d5418dc8d16cf215fb84892346c89084cb8;hb=fa54f2630c56a1df0046923d6a77b1bd58abf240;hp=3de7d5efa056e77708c5b0bd7925c4dbfea1ead4;hpb=b0ce4383b7d64760bbfccf4e0e769b293f0db0cd;p=nageru diff --git a/ffmpeg_capture.cpp b/ffmpeg_capture.cpp index 3de7d5e..1ffa9d5 100644 --- a/ffmpeg_capture.cpp +++ b/ffmpeg_capture.cpp @@ -32,6 +32,7 @@ extern "C" { #include "ffmpeg_util.h" #include "flags.h" #include "image_input.h" +#include "ref_counted_frame.h" #include "timebase.h" #define FRAME_SIZE (8 << 20) // 8 MB. @@ -408,13 +409,13 @@ bool FFmpegCapture::play_video(const string &pathname) if (process_queued_commands(format_ctx.get(), pathname, last_modified, /*rewound=*/nullptr)) { return true; } - FrameAllocator::Frame audio_frame = audio_frame_allocator->alloc_frame(); + UniqueFrame audio_frame = audio_frame_allocator->alloc_frame(); AudioFormat audio_format; int64_t audio_pts; bool error; AVFrameWithDeleter frame = decode_frame(format_ctx.get(), video_codec_ctx.get(), audio_codec_ctx.get(), - pathname, video_stream_index, audio_stream_index, &audio_frame, &audio_format, &audio_pts, &error); + pathname, video_stream_index, audio_stream_index, audio_frame.get(), &audio_format, &audio_pts, &error); if (error) { return false; } @@ -424,6 +425,12 @@ bool FFmpegCapture::play_video(const string &pathname) fprintf(stderr, "%s: Rewind failed, not looping.\n", pathname.c_str()); return true; } + if (video_codec_ctx != nullptr) { + avcodec_flush_buffers(video_codec_ctx.get()); + } + if (audio_codec_ctx != nullptr) { + avcodec_flush_buffers(audio_codec_ctx.get()); + } // If the file has changed since last time, return to get it reloaded. // Note that depending on how you move the file into place, you might // end up corrupting the one you're already playing, so this path @@ -436,7 +443,7 @@ bool FFmpegCapture::play_video(const string &pathname) } VideoFormat video_format = construct_video_format(frame.get(), video_timebase); - FrameAllocator::Frame video_frame = make_video_frame(frame.get(), pathname, &error); + UniqueFrame video_frame = make_video_frame(frame.get(), pathname, &error); if (error) { return false; } @@ -446,15 +453,15 @@ bool FFmpegCapture::play_video(const string &pathname) pts_origin = frame->pts; } next_frame_start = compute_frame_start(frame->pts, pts_origin, video_timebase, start, rate); - video_frame.received_timestamp = next_frame_start; + video_frame->received_timestamp = next_frame_start; bool finished_wakeup = producer_thread_should_quit.sleep_until(next_frame_start); if (finished_wakeup) { - if (audio_frame.len > 0) { + if (audio_frame->len > 0) { assert(audio_pts != -1); } frame_callback(frame->pts, video_timebase, audio_pts, audio_timebase, timecode++, - video_frame, 0, video_format, - audio_frame, 0, audio_format); + video_frame.get_and_release(), 0, video_format, + audio_frame.get_and_release(), 0, audio_format); break; } else { if (producer_thread_should_quit.should_quit()) break; @@ -465,7 +472,6 @@ bool FFmpegCapture::play_video(const string &pathname) } // If we just rewound, drop this frame on the floor and be done. if (rewound) { - video_frame_allocator->release_frame(video_frame); break; } // OK, we didn't, so probably a rate change. Recalculate next_frame_start, @@ -639,10 +645,15 @@ void FFmpegCapture::convert_audio(const AVFrame *audio_avframe, FrameAllocator:: } audio_format->num_channels = 2; + int64_t channel_layout = audio_avframe->channel_layout; + if (channel_layout == 0) { + channel_layout = av_get_default_channel_layout(audio_avframe->channels); + } + if (resampler == nullptr || audio_avframe->format != last_src_format || dst_format != last_dst_format || - av_frame_get_channel_layout(audio_avframe) != last_channel_layout || + channel_layout != last_channel_layout || av_frame_get_sample_rate(audio_avframe) != last_sample_rate) { avresample_free(&resampler); resampler = avresample_alloc_context(); @@ -651,7 +662,7 @@ void FFmpegCapture::convert_audio(const AVFrame *audio_avframe, FrameAllocator:: exit(1); } - av_opt_set_int(resampler, "in_channel_layout", av_frame_get_channel_layout(audio_avframe), 0); + av_opt_set_int(resampler, "in_channel_layout", channel_layout, 0); av_opt_set_int(resampler, "out_channel_layout", AV_CH_LAYOUT_STEREO, 0); av_opt_set_int(resampler, "in_sample_rate", av_frame_get_sample_rate(audio_avframe), 0); av_opt_set_int(resampler, "out_sample_rate", OUTPUT_FREQUENCY, 0); @@ -665,7 +676,7 @@ void FFmpegCapture::convert_audio(const AVFrame *audio_avframe, FrameAllocator:: last_src_format = AVSampleFormat(audio_avframe->format); last_dst_format = dst_format; - last_channel_layout = av_frame_get_channel_layout(audio_avframe); + last_channel_layout = channel_layout; last_sample_rate = av_frame_get_sample_rate(audio_avframe); } @@ -674,7 +685,7 @@ void FFmpegCapture::convert_audio(const AVFrame *audio_avframe, FrameAllocator:: uint8_t *data = audio_frame->data + audio_frame->len; int out_samples = avresample_convert(resampler, &data, 0, num_samples_room, - audio_avframe->data, audio_avframe->linesize[0], audio_avframe->nb_samples); + const_cast(audio_avframe->data), audio_avframe->linesize[0], audio_avframe->nb_samples); if (out_samples < 0) { fprintf(stderr, "Audio conversion failed.\n"); exit(1); @@ -708,12 +719,12 @@ VideoFormat FFmpegCapture::construct_video_format(const AVFrame *frame, AVRation return video_format; } -FrameAllocator::Frame FFmpegCapture::make_video_frame(const AVFrame *frame, const string &pathname, bool *error) +UniqueFrame FFmpegCapture::make_video_frame(const AVFrame *frame, const string &pathname, bool *error) { *error = false; - FrameAllocator::Frame video_frame = video_frame_allocator->alloc_frame(); - if (video_frame.data == nullptr) { + UniqueFrame video_frame(video_frame_allocator->alloc_frame()); + if (video_frame->data == nullptr) { return video_frame; } @@ -739,17 +750,17 @@ FrameAllocator::Frame FFmpegCapture::make_video_frame(const AVFrame *frame, cons uint8_t *pic_data[4] = { nullptr, nullptr, nullptr, nullptr }; int linesizes[4] = { 0, 0, 0, 0 }; if (pixel_format == bmusb::PixelFormat_8BitBGRA) { - pic_data[0] = video_frame.data; + pic_data[0] = video_frame->data; linesizes[0] = width * 4; - video_frame.len = (width * 4) * height; + video_frame->len = (width * 4) * height; } else if (pixel_format == PixelFormat_NV12) { - pic_data[0] = video_frame.data; + pic_data[0] = video_frame->data; linesizes[0] = width; pic_data[1] = pic_data[0] + width * height; linesizes[1] = width; - video_frame.len = (width * 2) * height; + video_frame->len = (width * 2) * height; const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(sws_dst_format); current_frame_ycbcr_format = decode_ycbcr_format(desc, frame); @@ -760,7 +771,7 @@ FrameAllocator::Frame FFmpegCapture::make_video_frame(const AVFrame *frame, cons int chroma_width = AV_CEIL_RSHIFT(int(width), desc->log2_chroma_w); int chroma_height = AV_CEIL_RSHIFT(int(height), desc->log2_chroma_h); - pic_data[0] = video_frame.data; + pic_data[0] = video_frame->data; linesizes[0] = width; pic_data[1] = pic_data[0] + width * height; @@ -769,7 +780,7 @@ FrameAllocator::Frame FFmpegCapture::make_video_frame(const AVFrame *frame, cons pic_data[2] = pic_data[1] + chroma_width * chroma_height; linesizes[2] = chroma_width; - video_frame.len = width * height + 2 * chroma_width * chroma_height; + video_frame->len = width * height + 2 * chroma_width * chroma_height; current_frame_ycbcr_format = decode_ycbcr_format(desc, frame); }