]> git.sesse.net Git - nageru/blobdiff - video_encoder.cpp
Fix an issue where the mixer lagging too much behind CEF would cause us to display...
[nageru] / video_encoder.cpp
index dbb2aa774502f67fb4469a8b7b462594aef5908a..efc9732c654093f4a00240039b5de55527382862 100644 (file)
@@ -1,16 +1,28 @@
 #include "video_encoder.h"
 
 #include <assert.h>
-
+#include <stdio.h>
+#include <time.h>
+#include <unistd.h>
 #include <string>
+#include <thread>
+
+extern "C" {
+#include <libavutil/mem.h>
+}
 
+#include "audio_encoder.h"
 #include "defs.h"
+#include "ffmpeg_raii.h"
 #include "flags.h"
 #include "httpd.h"
-#include "timebase.h"
+#include "mux.h"
 #include "quicksync_encoder.h"
+#include "timebase.h"
 #include "x264_encoder.h"
 
+class RefCountedFrame;
+
 using namespace std;
 using namespace movit;
 
@@ -28,15 +40,16 @@ string generate_local_dump_filename(int frame)
        // Use the frame number to disambiguate between two cuts starting
        // on the same second.
        char filename[256];
-       snprintf(filename, sizeof(filename), "%s%s-f%02d%s",
+       snprintf(filename, sizeof(filename), "%s/%s%s-f%02d%s",
+               global_flags.recording_dir.c_str(),
                LOCAL_DUMP_PREFIX, timestamp, frame % 100, LOCAL_DUMP_SUFFIX);
        return filename;
 }
 
 }  // namespace
 
-VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd)
-       : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd)
+VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd, DiskSpaceEstimator *disk_space_estimator)
+       : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd), disk_space_estimator(disk_space_estimator)
 {
        oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr);
        assert(oformat != nullptr);
@@ -45,23 +58,26 @@ VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const
        } else {
                stream_audio_encoder.reset(new AudioEncoder(global_flags.stream_audio_codec_name, global_flags.stream_audio_codec_bitrate, oformat));
        }
-       if (global_flags.x264_video_to_http) {
+       if (global_flags.x264_video_to_http || global_flags.x264_video_to_disk) {
                x264_encoder.reset(new X264Encoder(oformat));
        }
 
        string filename = generate_local_dump_filename(/*frame=*/0);
-       quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get()));
+       quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator));
 
        open_output_stream();
        stream_audio_encoder->add_mux(stream_mux.get());
        quicksync_encoder->set_stream_mux(stream_mux.get());
        if (global_flags.x264_video_to_http) {
-               x264_encoder->set_mux(stream_mux.get());
+               x264_encoder->add_mux(stream_mux.get());
        }
 }
 
 VideoEncoder::~VideoEncoder()
 {
+       quicksync_encoder->shutdown();
+       x264_encoder.reset(nullptr);
+       quicksync_encoder->close_file();
        quicksync_encoder.reset(nullptr);
        while (quicksync_encoders_in_shutdown.load() > 0) {
                usleep(10000);
@@ -80,10 +96,17 @@ void VideoEncoder::do_cut(int frame)
        // the same time, it means pts could come out of order to the stream mux,
        // and we need to plug it until the shutdown is complete.
        stream_mux->plug();
-       lock_guard<mutex> lock(qs_mu);
+       lock(qs_mu, qs_audio_mu);
+       lock_guard<mutex> lock1(qs_mu, adopt_lock), lock2(qs_audio_mu, adopt_lock);
        QuickSyncEncoder *old_encoder = quicksync_encoder.release();  // When we go C++14, we can use move capture instead.
-       thread([old_encoder, this]{
+       X264Encoder *old_x264_encoder = nullptr;
+       if (global_flags.x264_video_to_disk) {
+               old_x264_encoder = x264_encoder.release();
+       }
+       thread([old_encoder, old_x264_encoder, this]{
                old_encoder->shutdown();
+               delete old_x264_encoder;
+               old_encoder->close_file();
                stream_mux->unplug();
 
                // We cannot delete the encoder here, as this thread has no OpenGL context.
@@ -92,33 +115,58 @@ void VideoEncoder::do_cut(int frame)
                qs_needing_cleanup.emplace_back(old_encoder);
        }).detach();
 
-       quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get()));
+       if (global_flags.x264_video_to_disk) {
+               x264_encoder.reset(new X264Encoder(oformat));
+               if (global_flags.x264_video_to_http) {
+                       x264_encoder->add_mux(stream_mux.get());
+               }
+               if (overriding_bitrate != 0) {
+                       x264_encoder->change_bitrate(overriding_bitrate);
+               }
+       }
+
+       quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator));
        quicksync_encoder->set_stream_mux(stream_mux.get());
 }
 
 void VideoEncoder::change_x264_bitrate(unsigned rate_kbit)
 {
+       overriding_bitrate = rate_kbit;
        x264_encoder->change_bitrate(rate_kbit);
 }
 
 void VideoEncoder::add_audio(int64_t pts, std::vector<float> audio)
 {
-       lock_guard<mutex> lock(qs_mu);
-       quicksync_encoder->add_audio(pts, audio);
+       // Take only qs_audio_mu, since add_audio() is thread safe
+       // (we can only conflict with do_cut(), which takes qs_audio_mu)
+       // and we don't want to contend with begin_frame().
+       {
+               lock_guard<mutex> lock(qs_audio_mu);
+               quicksync_encoder->add_audio(pts, audio);
+       }
        stream_audio_encoder->encode_audio(audio, pts + quicksync_encoder->global_delay());
 }
 
-bool VideoEncoder::begin_frame(GLuint *y_tex, GLuint *cbcr_tex)
+bool VideoEncoder::is_zerocopy() const
+{
+       // Explicitly do _not_ take qs_mu; this is called from the mixer,
+       // and qs_mu might be contended. is_zerocopy() is thread safe
+       // and never called in parallel with do_cut() (both happen only
+       // from the mixer thread).
+       return quicksync_encoder->is_zerocopy();
+}
+
+bool VideoEncoder::begin_frame(int64_t pts, int64_t duration, movit::YCbCrLumaCoefficients ycbcr_coefficients, const std::vector<RefCountedFrame> &input_frames, GLuint *y_tex, GLuint *cbcr_tex)
 {
        lock_guard<mutex> lock(qs_mu);
        qs_needing_cleanup.clear();  // Since we have an OpenGL context here, and are called regularly.
-       return quicksync_encoder->begin_frame(y_tex, cbcr_tex);
+       return quicksync_encoder->begin_frame(pts, duration, ycbcr_coefficients, input_frames, y_tex, cbcr_tex);
 }
 
-RefCountedGLsync VideoEncoder::end_frame(int64_t pts, int64_t duration, const std::vector<RefCountedFrame> &input_frames)
+RefCountedGLsync VideoEncoder::end_frame()
 {
        lock_guard<mutex> lock(qs_mu);
-       return quicksync_encoder->end_frame(pts, duration, input_frames);
+       return quicksync_encoder->end_frame();
 }
 
 void VideoEncoder::open_output_stream()
@@ -141,12 +189,14 @@ void VideoEncoder::open_output_stream()
        avctx->flags = AVFMT_FLAG_CUSTOM_IO;
 
        string video_extradata;
-       if (global_flags.x264_video_to_http) {
+       if (global_flags.x264_video_to_http || global_flags.x264_video_to_disk) {
                video_extradata = x264_encoder->get_global_headers();
        }
 
        int time_base = global_flags.stream_coarse_timebase ? COARSE_TIMEBASE : TIMEBASE;
-       stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_ctx(), time_base));
+       stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_codec_parameters().get(), time_base,
+               /*write_callback=*/nullptr, Mux::WRITE_FOREGROUND, { &stream_mux_metrics }));
+       stream_mux_metrics.init({{ "destination", "http" }});
 }
 
 int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time)
@@ -157,6 +207,14 @@ int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size,
 
 int VideoEncoder::write_packet2(uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time)
 {
+       if (type == AVIO_DATA_MARKER_SYNC_POINT || type == AVIO_DATA_MARKER_BOUNDARY_POINT) {
+               seen_sync_markers = true;
+       } else if (type == AVIO_DATA_MARKER_UNKNOWN && !seen_sync_markers) {
+               // We don't know if this is a keyframe or not (the muxer could
+               // avoid marking it), so we just have to make the best of it.
+               type = AVIO_DATA_MARKER_SYNC_POINT;
+       }
+
        if (type == AVIO_DATA_MARKER_HEADER) {
                stream_mux_header.append((char *)buf, buf_size);
                httpd->set_header(stream_mux_header);