X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=video_encoder.cpp;h=dfbf565437d5dcfc311f84c7245eaa06eaee68be;hb=bea1be71fd06888ac03fcddb47f6d497f8e07aab;hp=cae4328f2a5e4fe974ffde3dd52494fabc43223a;hpb=e1a58f0f4e9cd05441f1e1b43fc4c83d1f862dd7;p=nageru diff --git a/video_encoder.cpp b/video_encoder.cpp index cae4328..dfbf565 100644 --- a/video_encoder.cpp +++ b/video_encoder.cpp @@ -35,64 +35,101 @@ string generate_local_dump_filename(int frame) } // namespace -VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd) - : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd) +VideoEncoder::VideoEncoder(ResourcePool *resource_pool, QSurface *surface, const std::string &va_display, int width, int height, HTTPD *httpd, DiskSpaceEstimator *disk_space_estimator) + : resource_pool(resource_pool), surface(surface), va_display(va_display), width(width), height(height), httpd(httpd), disk_space_estimator(disk_space_estimator) { - open_output_stream(); - + oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr); + assert(oformat != nullptr); if (global_flags.stream_audio_codec_name.empty()) { - stream_audio_encoder.reset(new AudioEncoder(AUDIO_OUTPUT_CODEC_NAME, DEFAULT_AUDIO_OUTPUT_BIT_RATE)); + stream_audio_encoder.reset(new AudioEncoder(AUDIO_OUTPUT_CODEC_NAME, DEFAULT_AUDIO_OUTPUT_BIT_RATE, oformat)); } else { - stream_audio_encoder.reset(new AudioEncoder(global_flags.stream_audio_codec_name, global_flags.stream_audio_codec_bitrate)); + stream_audio_encoder.reset(new AudioEncoder(global_flags.stream_audio_codec_name, global_flags.stream_audio_codec_bitrate, oformat)); } - stream_audio_encoder->add_mux(stream_mux.get()); - if (global_flags.x264_video_to_http) { - x264_encoder.reset(new X264Encoder(stream_mux.get())); + x264_encoder.reset(new X264Encoder(oformat)); } string filename = generate_local_dump_filename(/*frame=*/0); - quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, stream_mux.get(), stream_audio_encoder.get(), x264_encoder.get())); + quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator)); + + open_output_stream(); + stream_audio_encoder->add_mux(stream_mux.get()); + quicksync_encoder->set_stream_mux(stream_mux.get()); + if (global_flags.x264_video_to_http) { + x264_encoder->set_mux(stream_mux.get()); + } } VideoEncoder::~VideoEncoder() { quicksync_encoder.reset(nullptr); - close_output_stream(); + while (quicksync_encoders_in_shutdown.load() > 0) { + usleep(10000); + } } void VideoEncoder::do_cut(int frame) { string filename = generate_local_dump_filename(frame); printf("Starting new recording: %s\n", filename.c_str()); - quicksync_encoder->shutdown(); - quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, stream_mux.get(), stream_audio_encoder.get(), x264_encoder.get())); + + // Do the shutdown of the old encoder in a separate thread, since it can + // take some time (it needs to wait for all the frames in the queue to be + // done encoding, for one) and we are running on the main mixer thread. + // However, since this means both encoders could be sending packets at + // the same time, it means pts could come out of order to the stream mux, + // and we need to plug it until the shutdown is complete. + stream_mux->plug(); + lock_guard lock(qs_mu); + QuickSyncEncoder *old_encoder = quicksync_encoder.release(); // When we go C++14, we can use move capture instead. + thread([old_encoder, this]{ + old_encoder->shutdown(); + stream_mux->unplug(); + + // We cannot delete the encoder here, as this thread has no OpenGL context. + // We'll deal with it in begin_frame(). + lock_guard lock(qs_mu); + qs_needing_cleanup.emplace_back(old_encoder); + }).detach(); + + quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, x264_encoder.get(), disk_space_estimator)); + quicksync_encoder->set_stream_mux(stream_mux.get()); +} + +void VideoEncoder::change_x264_bitrate(unsigned rate_kbit) +{ + x264_encoder->change_bitrate(rate_kbit); } void VideoEncoder::add_audio(int64_t pts, std::vector audio) { + lock_guard lock(qs_mu); quicksync_encoder->add_audio(pts, audio); + stream_audio_encoder->encode_audio(audio, pts + quicksync_encoder->global_delay()); } bool VideoEncoder::begin_frame(GLuint *y_tex, GLuint *cbcr_tex) { + lock_guard lock(qs_mu); + qs_needing_cleanup.clear(); // Since we have an OpenGL context here, and are called regularly. return quicksync_encoder->begin_frame(y_tex, cbcr_tex); } RefCountedGLsync VideoEncoder::end_frame(int64_t pts, int64_t duration, const std::vector &input_frames) { + lock_guard lock(qs_mu); return quicksync_encoder->end_frame(pts, duration, input_frames); } void VideoEncoder::open_output_stream() { AVFormatContext *avctx = avformat_alloc_context(); - AVOutputFormat *oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr); - assert(oformat != nullptr); avctx->oformat = oformat; uint8_t *buf = (uint8_t *)av_malloc(MUX_BUFFER_SIZE); - avctx->pb = avio_alloc_context(buf, MUX_BUFFER_SIZE, 1, this, nullptr, &VideoEncoder::write_packet_thunk, nullptr); + avctx->pb = avio_alloc_context(buf, MUX_BUFFER_SIZE, 1, this, nullptr, nullptr, nullptr); + avctx->pb->write_data_type = &VideoEncoder::write_packet2_thunk; + avctx->pb->ignore_boundary_point = 1; Mux::Codec video_codec; if (global_flags.uncompressed_video_to_http) { @@ -103,32 +140,29 @@ void VideoEncoder::open_output_stream() avctx->flags = AVFMT_FLAG_CUSTOM_IO; - int time_base = global_flags.stream_coarse_timebase ? COARSE_TIMEBASE : TIMEBASE; - stream_mux_writing_header = true; - stream_mux.reset(new Mux(avctx, width, height, video_codec, stream_audio_encoder->get_ctx(), time_base, this)); - stream_mux_writing_header = false; - httpd->set_header(stream_mux_header); - stream_mux_header.clear(); -} + string video_extradata; + if (global_flags.x264_video_to_http) { + video_extradata = x264_encoder->get_global_headers(); + } -void VideoEncoder::close_output_stream() -{ - stream_mux.reset(); + int time_base = global_flags.stream_coarse_timebase ? COARSE_TIMEBASE : TIMEBASE; + stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_codec_parameters().get(), time_base, + /*write_callback=*/nullptr)); } -int VideoEncoder::write_packet_thunk(void *opaque, uint8_t *buf, int buf_size) +int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time) { VideoEncoder *video_encoder = (VideoEncoder *)opaque; - return video_encoder->write_packet(buf, buf_size); + return video_encoder->write_packet2(buf, buf_size, type, time); } -int VideoEncoder::write_packet(uint8_t *buf, int buf_size) +int VideoEncoder::write_packet2(uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time) { - if (stream_mux_writing_header) { + if (type == AVIO_DATA_MARKER_HEADER) { stream_mux_header.append((char *)buf, buf_size); + httpd->set_header(stream_mux_header); } else { - httpd->add_data((char *)buf, buf_size, stream_mux_writing_keyframes); - stream_mux_writing_keyframes = false; + httpd->add_data((char *)buf, buf_size, type == AVIO_DATA_MARKER_SYNC_POINT); } return buf_size; }