X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=video_encoder.cpp;h=a622ba2af6640425d72dc5a72de44b4cc8f218e3;hb=96cb6414f85e0ef4d660b7bd56267303e80fcd05;hp=07bb1c95a1fb8aa72c35d3615f9c75676697bf2c;hpb=355343173dcb39cff811baff55e18977d6b49472;p=nageru diff --git a/video_encoder.cpp b/video_encoder.cpp index 07bb1c9..a622ba2 100644 --- a/video_encoder.cpp +++ b/video_encoder.cpp @@ -1,16 +1,28 @@ #include "video_encoder.h" #include - +#include +#include +#include #include +#include + +extern "C" { +#include +} +#include "audio_encoder.h" #include "defs.h" +#include "ffmpeg_raii.h" #include "flags.h" #include "httpd.h" -#include "timebase.h" +#include "mux.h" #include "quicksync_encoder.h" +#include "timebase.h" #include "x264_encoder.h" +class RefCountedFrame; + using namespace std; using namespace movit; @@ -28,15 +40,16 @@ string generate_local_dump_filename(int frame) // Use the frame number to disambiguate between two cuts starting // on the same second. char filename[256]; - snprintf(filename, sizeof(filename), "%s%s-f%02d%s", + snprintf(filename, sizeof(filename), "%s/%s%s-f%02d%s", + global_flags.recording_dir.c_str(), LOCAL_DUMP_PREFIX, timestamp, frame % 100, LOCAL_DUMP_SUFFIX); return filename; } } // namespace -VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd) - : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd) +VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd, DiskSpaceEstimator *disk_space_estimator) + : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd), disk_space_estimator(disk_space_estimator) { oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr); assert(oformat != nullptr); @@ -45,23 +58,26 @@ VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const } else { stream_audio_encoder.reset(new AudioEncoder(global_flags.stream_audio_codec_name, global_flags.stream_audio_codec_bitrate, oformat)); } - if (global_flags.x264_video_to_http) { + if (global_flags.x264_video_to_http || global_flags.x264_video_to_disk) { x264_encoder.reset(new X264Encoder(oformat)); } string filename = generate_local_dump_filename(/*frame=*/0); - quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get())); + quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator)); open_output_stream(); stream_audio_encoder->add_mux(stream_mux.get()); quicksync_encoder->set_stream_mux(stream_mux.get()); if (global_flags.x264_video_to_http) { - x264_encoder->set_mux(stream_mux.get()); + x264_encoder->add_mux(stream_mux.get()); } } VideoEncoder::~VideoEncoder() { + quicksync_encoder->shutdown(); + x264_encoder.reset(nullptr); + quicksync_encoder->close_file(); quicksync_encoder.reset(nullptr); while (quicksync_encoders_in_shutdown.load() > 0) { usleep(10000); @@ -82,8 +98,14 @@ void VideoEncoder::do_cut(int frame) stream_mux->plug(); lock_guard lock(qs_mu); QuickSyncEncoder *old_encoder = quicksync_encoder.release(); // When we go C++14, we can use move capture instead. - thread([old_encoder, this]{ + X264Encoder *old_x264_encoder = nullptr; + if (global_flags.x264_video_to_disk) { + old_x264_encoder = x264_encoder.release(); + } + thread([old_encoder, old_x264_encoder, this]{ old_encoder->shutdown(); + delete old_x264_encoder; + old_encoder->close_file(); stream_mux->unplug(); // We cannot delete the encoder here, as this thread has no OpenGL context. @@ -92,10 +114,26 @@ void VideoEncoder::do_cut(int frame) qs_needing_cleanup.emplace_back(old_encoder); }).detach(); - quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get())); + if (global_flags.x264_video_to_disk) { + x264_encoder.reset(new X264Encoder(oformat)); + if (global_flags.x264_video_to_http) { + x264_encoder->add_mux(stream_mux.get()); + } + if (overriding_bitrate != 0) { + x264_encoder->change_bitrate(overriding_bitrate); + } + } + + quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator)); quicksync_encoder->set_stream_mux(stream_mux.get()); } +void VideoEncoder::change_x264_bitrate(unsigned rate_kbit) +{ + overriding_bitrate = rate_kbit; + x264_encoder->change_bitrate(rate_kbit); +} + void VideoEncoder::add_audio(int64_t pts, std::vector audio) { lock_guard lock(qs_mu); @@ -103,17 +141,23 @@ void VideoEncoder::add_audio(int64_t pts, std::vector audio) stream_audio_encoder->encode_audio(audio, pts + quicksync_encoder->global_delay()); } -bool VideoEncoder::begin_frame(GLuint *y_tex, GLuint *cbcr_tex) +bool VideoEncoder::is_zerocopy() const +{ + lock_guard lock(qs_mu); + return quicksync_encoder->is_zerocopy(); +} + +bool VideoEncoder::begin_frame(int64_t pts, int64_t duration, movit::YCbCrLumaCoefficients ycbcr_coefficients, const std::vector &input_frames, GLuint *y_tex, GLuint *cbcr_tex) { lock_guard lock(qs_mu); qs_needing_cleanup.clear(); // Since we have an OpenGL context here, and are called regularly. - return quicksync_encoder->begin_frame(y_tex, cbcr_tex); + return quicksync_encoder->begin_frame(pts, duration, ycbcr_coefficients, input_frames, y_tex, cbcr_tex); } -RefCountedGLsync VideoEncoder::end_frame(int64_t pts, int64_t duration, const std::vector &input_frames) +RefCountedGLsync VideoEncoder::end_frame() { lock_guard lock(qs_mu); - return quicksync_encoder->end_frame(pts, duration, input_frames); + return quicksync_encoder->end_frame(); } void VideoEncoder::open_output_stream() @@ -136,12 +180,14 @@ void VideoEncoder::open_output_stream() avctx->flags = AVFMT_FLAG_CUSTOM_IO; string video_extradata; - if (global_flags.x264_video_to_http) { + if (global_flags.x264_video_to_http || global_flags.x264_video_to_disk) { video_extradata = x264_encoder->get_global_headers(); } int time_base = global_flags.stream_coarse_timebase ? COARSE_TIMEBASE : TIMEBASE; - stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_ctx(), time_base)); + stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_codec_parameters().get(), time_base, + /*write_callback=*/nullptr, { &stream_mux_metrics })); + stream_mux_metrics.init({{ "destination", "http" }}); } int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time) @@ -152,6 +198,14 @@ int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, int VideoEncoder::write_packet2(uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time) { + if (type == AVIO_DATA_MARKER_SYNC_POINT || type == AVIO_DATA_MARKER_BOUNDARY_POINT) { + seen_sync_markers = true; + } else if (type == AVIO_DATA_MARKER_UNKNOWN && !seen_sync_markers) { + // We don't know if this is a keyframe or not (the muxer could + // avoid marking it), so we just have to make the best of it. + type = AVIO_DATA_MARKER_SYNC_POINT; + } + if (type == AVIO_DATA_MARKER_HEADER) { stream_mux_header.append((char *)buf, buf_size); httpd->set_header(stream_mux_header);