]> git.sesse.net Git - nageru/blobdiff - nageru/kaeru.cpp
Add CEF support to Kaeru.
[nageru] / nageru / kaeru.cpp
index 9d9738391082a484cf3338361c2746aff1aa5ec7..16ad079789a02b68c46bf109bed69d3d71ae11bf 100644 (file)
@@ -2,21 +2,60 @@
 
 #include "audio_encoder.h"
 #include "basic_stats.h"
+#ifdef HAVE_CEF
+#include "cef_capture.h"
+#endif
 #include "defs.h"
 #include "flags.h"
 #include "ffmpeg_capture.h"
 #include "mixer.h"
+#include "print_latency.h"
+#include "shared/ffmpeg_raii.h"
+#include "shared/httpd.h"
 #include "shared/mux.h"
 #include "quittable_sleeper.h"
+#include "shared/shared_defs.h"
 #include "shared/timebase.h"
 #include "x264_encoder.h"
 
 #include <assert.h>
-#include <fcntl.h>
-#include <signal.h>
-#include <unistd.h>
+#include <bmusb/bmusb.h>
 #include <chrono>
+#include <endian.h>
+#include <errno.h>
+#include <functional>
+#include <memory>
+#include <movit/image_format.h>
+#include <signal.h>
+#include <stddef.h>
+#include <stdint.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
 #include <string>
+#include <vector>
+
+extern "C" {
+#include <libavcodec/bsf.h>
+#include <libavcodec/codec_par.h>
+#include <libavcodec/packet.h>
+#include <libavformat/avformat.h>
+#include <libavformat/avio.h>
+#include <libavformat/version.h>
+#include <libavutil/avutil.h>
+#include <libavutil/common.h>
+#include <libavutil/error.h>
+#include <libavutil/mathematics.h>
+#include <libavutil/mem.h>
+#include <libavutil/rational.h>
+#include <libavutil/version.h>
+}
+
+#ifdef HAVE_CEF
+#include "cef_encoder_adapter.h"
+#include "nageru_cef_app.h"
+CefRefPtr<NageruCefApp> cef_app;
+#endif
 
 using namespace bmusb;
 using namespace movit;
@@ -47,21 +86,22 @@ int write_packet(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType ty
                type = AVIO_DATA_MARKER_SYNC_POINT;
        }
 
+       HTTPD::StreamID stream_id{ HTTPD::MAIN_STREAM, 0 };
        if (type == AVIO_DATA_MARKER_HEADER) {
                stream_mux_header.append((char *)buf, buf_size);
-               httpd->set_header(HTTPD::MAIN_STREAM, stream_mux_header);
+               httpd->set_header(stream_id, stream_mux_header);
        } else {
-               httpd->add_data(HTTPD::MAIN_STREAM, (char *)buf, buf_size, type == AVIO_DATA_MARKER_SYNC_POINT, time, AVRational{ AV_TIME_BASE, 1 });
+               httpd->add_data(stream_id, (char *)buf, buf_size, type == AVIO_DATA_MARKER_SYNC_POINT, time, AVRational{ AV_TIME_BASE, 1 });
        }
        return buf_size;
 }
 
 }  // namespace
 
-unique_ptr<Mux> create_mux(HTTPD *httpd, AVOutputFormat *oformat, X264Encoder *x264_encoder, AudioEncoder *audio_encoder)
+unique_ptr<Mux> create_mux(HTTPD *httpd, const AVOutputFormat *oformat, X264Encoder *x264_encoder, AudioEncoder *audio_encoder)
 {
        AVFormatContext *avctx = avformat_alloc_context();
-       avctx->oformat = oformat;
+       avctx->oformat = oformat;  // const_cast is a hack to work in FFmpeg both before and after 5.0.
 
        uint8_t *buf = (uint8_t *)av_malloc(MUX_BUFFER_SIZE);
        avctx->pb = avio_alloc_context(buf, MUX_BUFFER_SIZE, 1, httpd, nullptr, nullptr, nullptr);
@@ -71,20 +111,26 @@ unique_ptr<Mux> create_mux(HTTPD *httpd, AVOutputFormat *oformat, X264Encoder *x
 
        string video_extradata = x264_encoder->get_global_headers();
 
+       // If audio is disabled (ie., we won't ever see any audio packets),
+       // set nullptr here to also not include the stream in the mux.
+       AVCodecParameters *audio_codecpar =
+               global_flags.enable_audio ? audio_encoder->get_codec_parameters().release() : nullptr;
+
        unique_ptr<Mux> mux;
-       mux.reset(new Mux(avctx, global_flags.width, global_flags.height, Mux::CODEC_H264, video_extradata, audio_encoder->get_codec_parameters().get(),
+       mux.reset(new Mux(avctx, global_flags.width, global_flags.height, Mux::CODEC_H264, video_extradata, audio_codecpar,
                get_color_space(global_flags.ycbcr_rec709_coefficients), COARSE_TIMEBASE,
                /*write_callback=*/nullptr, Mux::WRITE_FOREGROUND, { &stream_mux_metrics }));
        stream_mux_metrics.init({{ "destination", "http" }});
        return mux;
 }
 
+// NOTE: If we start using the timecode for anything, CEFEncoderAdapter will need adjustment.
 void video_frame_callback(FFmpegCapture *video, X264Encoder *x264_encoder, AudioEncoder *audio_encoder,
                           int64_t video_pts, AVRational video_timebase,
                           int64_t audio_pts, AVRational audio_timebase,
                           uint16_t timecode,
-                         FrameAllocator::Frame video_frame, size_t video_offset, VideoFormat video_format,
-                         FrameAllocator::Frame audio_frame, size_t audio_offset, AudioFormat audio_format)
+                          FrameAllocator::Frame video_frame, size_t video_offset, VideoFormat video_format,
+                          FrameAllocator::Frame audio_frame, size_t audio_offset, AudioFormat audio_format)
 {
        if (video_pts >= 0 && video_frame.len > 0) {
                ReceivedTimestamps ts;
@@ -92,7 +138,8 @@ void video_frame_callback(FFmpegCapture *video, X264Encoder *x264_encoder, Audio
 
                video_pts = av_rescale_q(video_pts, video_timebase, AVRational{ 1, TIMEBASE });
                int64_t frame_duration = int64_t(TIMEBASE) * video_format.frame_rate_den / video_format.frame_rate_nom;
-               x264_encoder->add_frame(video_pts, frame_duration, video->get_current_frame_ycbcr_format().luma_coefficients, video_frame.data + video_offset, ts);
+               YCbCrLumaCoefficients luma_coefficients = video ? video->get_current_frame_ycbcr_format().luma_coefficients : YCBCR_REC_709;
+               x264_encoder->add_frame(video_pts, frame_duration, luma_coefficients, video_frame.data + video_offset, ts);
                global_basic_stats->update(frame_num++, /*dropped_frames=*/0);
        }
        if (audio_frame.len > 0) {
@@ -104,24 +151,25 @@ void video_frame_callback(FFmpegCapture *video, X264Encoder *x264_encoder, Audio
                size_t num_samples = audio_frame.len / (audio_format.bits_per_sample / 8);
                vector<float> float_samples;
                float_samples.resize(num_samples);
+
                if (audio_format.bits_per_sample == 16) {
                        const int16_t *src = (const int16_t *)audio_frame.data;
                        float *dst = &float_samples[0];
                        for (size_t i = 0; i < num_samples; ++i) {
-                               *dst++ = le16toh(*src++) * (1.0f / 32768.0f);
+                               *dst++ = int16_t(le16toh(*src++)) * (1.0f / 32768.0f);
                        }
                } else if (audio_format.bits_per_sample == 32) {
                        const int32_t *src = (const int32_t *)audio_frame.data;
                        float *dst = &float_samples[0];
                        for (size_t i = 0; i < num_samples; ++i) {
-                               *dst++ = le32toh(*src++) * (1.0f / 2147483648.0f);
+                               *dst++ = int32_t(le32toh(*src++)) * (1.0f / 2147483648.0f);
                        }
                } else {
                        assert(false);
                }
                audio_pts = av_rescale_q(audio_pts, audio_timebase, AVRational{ 1, TIMEBASE });
                audio_encoder->encode_audio(float_samples, audio_pts);
-        }
+       }
 
        if (video_frame.owner) {
                video_frame.owner->release_frame(video_frame);
@@ -131,9 +179,37 @@ void video_frame_callback(FFmpegCapture *video, X264Encoder *x264_encoder, Audio
        }
 }
 
-void audio_frame_callback(Mux *mux, const AVPacket *pkt, AVRational timebase)
+void raw_packet_callback(Mux *mux, int stream_index, const AVPacket *pkt, AVRational timebase)
+{
+       mux->add_packet(*pkt, pkt->pts, pkt->dts == AV_NOPTS_VALUE ? pkt->pts : pkt->dts, timebase, stream_index);
+}
+
+void filter_packet_callback(Mux *mux, int stream_index, AVBSFContext *bsfctx, const AVPacket *pkt, AVRational timebase)
 {
-       mux->add_packet(*pkt, pkt->pts, pkt->dts == AV_NOPTS_VALUE ? pkt->pts : pkt->dts, timebase, /*stream_index=*/1);
+       if (pkt->size <= 2 || pkt->data[0] != 0xff || (pkt->data[1] & 0xf0) != 0xf0) {
+               // Not ADTS data, so just pass it through.
+               mux->add_packet(*pkt, pkt->pts, pkt->dts == AV_NOPTS_VALUE ? pkt->pts : pkt->dts, timebase, stream_index);
+               return;
+       }
+
+       AVPacket *in_pkt = av_packet_clone(pkt);
+       unique_ptr<AVPacket, decltype(av_packet_unref) *> in_pkt_cleanup(in_pkt, av_packet_unref);
+       int err = av_bsf_send_packet(bsfctx, in_pkt);
+       if (err < 0) {
+               fprintf(stderr, "av_bsf_send_packet() failed with %d, ignoring\n", err);
+       }
+       for ( ;; ) {
+               AVPacketWithDeleter out_pkt = av_packet_alloc_unique();
+               err = av_bsf_receive_packet(bsfctx, out_pkt.get());
+               if (err == AVERROR(EAGAIN)) {
+                       break;
+               }
+               if (err < 0) {
+                       fprintf(stderr, "av_bsf_receive_packet() failed with %d, ignoring\n", err);
+                       return;
+               }
+               mux->add_packet(*out_pkt, out_pkt->pts, out_pkt->dts == AV_NOPTS_VALUE ? out_pkt->pts : out_pkt->dts, timebase, stream_index);
+       }
 }
 
 void adjust_bitrate(int signal)
@@ -169,12 +245,23 @@ void request_quit(int signal)
 
 int main(int argc, char *argv[])
 {
+#ifdef HAVE_CEF
+       // Let CEF have first priority on parsing the command line, because we might be
+       // launched as a CEF sub-process.
+       CefMainArgs main_args(argc, argv);
+       cef_app = CefRefPtr<NageruCefApp>(new NageruCefApp());
+       int err = CefExecuteProcess(main_args, cef_app.get(), nullptr);
+       if (err >= 0) {
+               return err;
+       }
+#endif
+
        parse_flags(PROGRAM_KAERU, argc, argv);
        if (optind + 1 != argc) {
                usage(PROGRAM_KAERU);
-               exit(1);
+               abort();
        }
-       global_flags.num_cards = 1;  // For latency metrics.
+       global_flags.max_num_cards = 1;  // For latency metrics.
 
 #if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(58, 9, 100)
        av_register_all();
@@ -183,7 +270,7 @@ int main(int argc, char *argv[])
 
        HTTPD httpd;
 
-       AVOutputFormat *oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr);
+       const AVOutputFormat *oformat = av_guess_format(global_flags.stream_mux_name.c_str(), nullptr, nullptr);
        assert(oformat != nullptr);
 
        unique_ptr<AudioEncoder> audio_encoder;
@@ -193,23 +280,60 @@ int main(int argc, char *argv[])
                audio_encoder.reset(new AudioEncoder(global_flags.stream_audio_codec_name, global_flags.stream_audio_codec_bitrate, oformat));
        }
 
-       unique_ptr<X264Encoder> x264_encoder(new X264Encoder(oformat));
+       unique_ptr<X264Encoder> x264_encoder(new X264Encoder(oformat, /*use_separate_disk_params=*/false));
        unique_ptr<Mux> http_mux = create_mux(&httpd, oformat, x264_encoder.get(), audio_encoder.get());
        if (global_flags.transcode_audio) {
                audio_encoder->add_mux(http_mux.get());
        }
-       x264_encoder->add_mux(http_mux.get());
+       if (global_flags.transcode_video) {
+               x264_encoder->add_mux(http_mux.get());
+       }
        global_x264_encoder = x264_encoder.get();
 
-       FFmpegCapture video(argv[optind], global_flags.width, global_flags.height);
-       video.set_pixel_format(FFmpegCapture::PixelFormat_NV12);
-       video.set_frame_callback(bind(video_frame_callback, &video, x264_encoder.get(), audio_encoder.get(), _1, _2, _3, _4, _5, _6, _7, _8, _9, _10, _11));
-       if (!global_flags.transcode_audio) {
-               video.set_audio_callback(bind(audio_frame_callback, http_mux.get(), _1, _2));
+       CaptureInterface *video;
+       unique_ptr<FFmpegCapture> ffmpeg_video;
+#ifdef HAVE_CEF
+       unique_ptr<CEFCapture> cef_video;
+       unique_ptr<CEFEncoderAdapter> cef_encoder_adapter;
+       if (global_flags.use_cef) {
+               cef_encoder_adapter.reset(new CEFEncoderAdapter(global_flags.width, global_flags.height, x264_encoder.get(), audio_encoder.get()));
+               cef_video.reset(new CEFCapture(argv[optind], global_flags.width, global_flags.height));
+               cef_video->set_pixel_format(bmusb::PixelFormat_8BitBGRA);
+               cef_video->set_frame_callback(bind(&CEFEncoderAdapter::video_frame_callback, cef_encoder_adapter.get(), _1, _2, _3, _4, _5, _6, _7));
+               // NOTE: No CEF audio support yet.
+               video = cef_video.get();
+       } else
+#endif
+       {
+              ffmpeg_video.reset(new FFmpegCapture(argv[optind], global_flags.width, global_flags.height));
+              ffmpeg_video->set_pixel_format(FFmpegCapture::PixelFormat_NV12);
+              if (global_flags.transcode_video) {
+                      ffmpeg_video->set_frame_callback(bind(video_frame_callback, ffmpeg_video.get(), x264_encoder.get(), audio_encoder.get(), _1, _2, _3, _4, _5, _6, _7, _8, _9, _10, _11));
+              } else {
+                      ffmpeg_video->set_video_callback(bind(raw_packet_callback, http_mux.get(), /*stream_index=*/0, _1, _2));
+              }
+              if (!global_flags.transcode_audio && global_flags.enable_audio) {
+                      AVBSFContext *bsfctx = nullptr;
+                      if (strcmp(oformat->name, "mp4") == 0 && strcmp(audio_encoder->get_codec()->name, "aac") == 0) {
+                              // We need to insert the aac_adtstoasc filter, seemingly (or we will get warnings to do so).
+                              const AVBitStreamFilter *filter = av_bsf_get_by_name("aac_adtstoasc");
+                              int err = av_bsf_alloc(filter, &bsfctx);
+                              if (err < 0) {
+                                      fprintf(stderr, "av_bsf_alloc() failed with %d\n", err);
+                                      exit(1);
+                              }
+                      }
+                      if (bsfctx == nullptr) {
+                              ffmpeg_video->set_audio_callback(bind(raw_packet_callback, http_mux.get(), /*stream_index=*/1, _1, _2));
+                      } else {
+                              ffmpeg_video->set_audio_callback(bind(filter_packet_callback, http_mux.get(), /*stream_index=*/1, bsfctx, _1, _2));
+                      }
+              }
+              ffmpeg_video->change_rate(10.0);  // Play as fast as possible.
+              video = ffmpeg_video.get();
        }
-       video.configure_card();
-       video.start_bm_capture();
-       video.change_rate(2.0);  // Be sure never to really fall behind, but also don't dump huge amounts of stuff onto x264.
+       video->configure_card();
+       video->start_bm_capture();
 
        BasicStats basic_stats(/*verbose=*/false, /*use_opengl=*/false);
        global_basic_stats = &basic_stats;
@@ -220,10 +344,16 @@ int main(int argc, char *argv[])
        signal(SIGINT, request_quit);
 
        while (!should_quit.should_quit()) {
+#ifdef HAVE_CEF
+               if (global_flags.use_cef) {
+                       cef_encoder_adapter->duplicate_frame_if_needed(&should_quit);
+                       continue;
+               }
+#endif
                should_quit.sleep_for(hours(1000));
        }
 
-       video.stop_dequeue_thread();
+       video->stop_dequeue_thread();
        // Stop the x264 encoder before killing the mux it's writing to.
        global_x264_encoder = nullptr;
        x264_encoder.reset();