quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, http_encoder, disk_encoder, disk_space_estimator));
open_output_stream();
- stream_audio_encoder->add_mux(stream_mux.get());
- quicksync_encoder->set_stream_mux(stream_mux.get());
+ stream_audio_encoder->add_mux(http_mux.get());
+ quicksync_encoder->set_http_mux(http_mux.get());
if (global_flags.x264_video_to_http) {
- x264_encoder->add_mux(stream_mux.get());
+ x264_encoder->add_mux(http_mux.get());
}
#ifdef HAVE_AV1
if (global_flags.av1_video_to_http) {
- av1_encoder->add_mux(stream_mux.get());
+ av1_encoder->add_mux(http_mux.get());
}
#endif
}
// However, since this means both encoders could be sending packets at
// the same time, it means pts could come out of order to the stream mux,
// and we need to plug it until the shutdown is complete.
- stream_mux->plug();
+ http_mux->plug();
lock(qs_mu, qs_audio_mu);
lock_guard<mutex> lock1(qs_mu, adopt_lock), lock2(qs_audio_mu, adopt_lock);
QuickSyncEncoder *old_encoder = quicksync_encoder.release(); // When we go C++14, we can use move capture instead.
delete old_x264_encoder;
delete old_x264_disk_encoder;
old_encoder->close_file();
- stream_mux->unplug();
+ http_mux->unplug();
// We cannot delete the encoder here, as this thread has no OpenGL context.
// We'll deal with it in begin_frame().
x264_encoder.reset(new X264Encoder(oformat, /*use_separate_disk_params=*/false));
assert(global_flags.x264_video_to_http);
if (global_flags.x264_video_to_http) {
- x264_encoder->add_mux(stream_mux.get());
+ x264_encoder->add_mux(http_mux.get());
}
if (overriding_bitrate != 0) {
x264_encoder->change_bitrate(overriding_bitrate);
}
quicksync_encoder.reset(new QuickSyncEncoder(filename, resource_pool, surface, va_display, width, height, oformat, http_encoder, disk_encoder, disk_space_estimator));
- quicksync_encoder->set_stream_mux(stream_mux.get());
+ quicksync_encoder->set_http_mux(http_mux.get());
}
void VideoEncoder::change_x264_bitrate(unsigned rate_kbit)
void VideoEncoder::open_output_stream()
{
AVFormatContext *avctx = avformat_alloc_context();
- avctx->oformat = const_cast<decltype(avctx->oformat)>(oformat); // const_cast is a hack to work in FFmpeg both before and after 5.0.
+ avctx->oformat = oformat;
uint8_t *buf = (uint8_t *)av_malloc(MUX_BUFFER_SIZE);
avctx->pb = avio_alloc_context(buf, MUX_BUFFER_SIZE, 1, this, nullptr, nullptr, nullptr);
avctx->pb->ignore_boundary_point = 1;
Mux::Codec video_codec;
- if (global_flags.uncompressed_video_to_http) {
- video_codec = Mux::CODEC_NV12;
- } else if (global_flags.av1_video_to_http) {
+ if (global_flags.av1_video_to_http) {
video_codec = Mux::CODEC_AV1;
} else {
video_codec = Mux::CODEC_H264;
#endif
}
- stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_codec_parameters().get(),
+ http_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, stream_audio_encoder->get_codec_parameters().get(),
get_color_space(global_flags.ycbcr_rec709_coefficients), COARSE_TIMEBASE,
- /*write_callback=*/nullptr, Mux::WRITE_FOREGROUND, { &stream_mux_metrics }));
- stream_mux_metrics.init({{ "destination", "http" }});
+ /*write_callback=*/nullptr, Mux::WRITE_FOREGROUND, { &http_mux_metrics }));
+ http_mux_metrics.init({{ "destination", "http" }});
}
int VideoEncoder::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time)
}
if (type == AVIO_DATA_MARKER_HEADER) {
- stream_mux_header.append((char *)buf, buf_size);
- httpd->set_header(HTTPD::StreamID{ HTTPD::MAIN_STREAM, 0 }, stream_mux_header);
+ http_mux_header.append((char *)buf, buf_size);
+ httpd->set_header(HTTPD::StreamID{ HTTPD::MAIN_STREAM, 0 }, http_mux_header);
} else {
httpd->add_data(HTTPD::StreamID{ HTTPD::MAIN_STREAM, 0 }, (char *)buf, buf_size, type == AVIO_DATA_MARKER_SYNC_POINT, time, AVRational{ AV_TIME_BASE, 1 });
}