X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=mixer.cpp;h=cbc66d423ccd94682b3eb67304f0d04744961ba4;hb=4a0187ffb4075b4d217b8d9e9c96cac548b199d8;hp=f1d116fe497bdeb0e6f020aef2a7b468624ede2a;hpb=868450678563ed2c9fd5240dca704d87371d1478;p=nageru diff --git a/mixer.cpp b/mixer.cpp index f1d116f..cbc66d4 100644 --- a/mixer.cpp +++ b/mixer.cpp @@ -33,6 +33,9 @@ #include "basic_stats.h" #include "bmusb/bmusb.h" #include "bmusb/fake_capture.h" +#ifdef HAVE_CEF +#include "cef_capture.h" +#endif #include "chroma_subsampler.h" #include "context.h" #include "decklink_capture.h" @@ -51,6 +54,10 @@ #include "v210_converter.h" #include "video_encoder.h" +#undef Status +#include +#include "json.pb.h" + class IDeckLink; class QOpenGLContext; @@ -150,6 +157,8 @@ void ensure_texture_resolution(PBOFrameAllocator::Userdata *userdata, unsigned f } check_error(); break; + default: + assert(false); } userdata->last_width[field] = width; userdata->last_height[field] = height; @@ -239,7 +248,7 @@ double JitterHistory::estimate_max_jitter() const if (percentile <= 0.5) { return *next(orders.begin(), elem_idx) * multiplier; } else { - return *prev(orders.end(), elem_idx + 1) * multiplier; + return *prev(orders.end(), orders.size() - elem_idx) * multiplier; } } @@ -352,8 +361,15 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) // Must be instantiated after VideoEncoder has initialized global_flags.use_zerocopy. theme.reset(new Theme(global_flags.theme_filename, global_flags.theme_dirs, resource_pool.get(), num_cards)); + httpd.add_endpoint("/channels", bind(&Mixer::get_channels_json, this), HTTPD::ALLOW_ALL_ORIGINS); + for (int channel_idx = 2; channel_idx < theme->get_num_channels(); ++channel_idx) { + char url[256]; + snprintf(url, sizeof(url), "/channels/%d/color", channel_idx); + httpd.add_endpoint(url, bind(&Mixer::get_channel_color_http, this, unsigned(channel_idx)), HTTPD::ALLOW_ALL_ORIGINS); + } + // Start listening for clients only once VideoEncoder has written its header, if any. - httpd.start(9095); + httpd.start(global_flags.http_port); // First try initializing the then PCI devices, then USB, then // fill up with fake cards until we have the desired number of cards. @@ -371,7 +387,10 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) DeckLinkCapture *capture = new DeckLinkCapture(decklink, card_index); DeckLinkOutput *output = new DeckLinkOutput(resource_pool.get(), decklink_output_surface, global_flags.width, global_flags.height, card_index); - output->set_device(decklink); + if (!output->set_device(decklink)) { + delete output; + output = nullptr; + } configure_card(card_index, capture, CardType::LIVE_CARD, output); ++num_pci_devices; } @@ -413,10 +432,24 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) } num_video_inputs = video_inputs.size(); +#ifdef HAVE_CEF + // Same, for HTML inputs. + std::vector html_inputs = theme->get_html_inputs(); + for (unsigned html_card_index = 0; html_card_index < html_inputs.size(); ++card_index, ++html_card_index) { + if (card_index >= MAX_VIDEO_CARDS) { + fprintf(stderr, "ERROR: Not enough card slots available for the HTML inputs the theme requested.\n"); + exit(1); + } + configure_card(card_index, html_inputs[html_card_index], CardType::CEF_INPUT, /*output=*/nullptr); + html_inputs[html_card_index]->set_card_index(card_index); + } + num_html_inputs = html_inputs.size(); +#endif + BMUSBCapture::set_card_connected_callback(bind(&Mixer::bm_hotplug_add, this, _1)); BMUSBCapture::start_bm_thread(); - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { cards[card_index].queue_length_policy.reset(card_index); } @@ -466,7 +499,7 @@ Mixer::~Mixer() { BMUSBCapture::stop_bm_thread(); - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { { unique_lock lock(card_mutex); cards[card_index].should_quit = true; // Unblock thread. @@ -492,6 +525,8 @@ void Mixer::configure_card(unsigned card_index, CaptureInterface *capture, CardT } card->capture.reset(capture); card->is_fake_capture = (card_type == CardType::FAKE_CAPTURE); + card->is_cef_capture = (card_type == CardType::CEF_INPUT); + card->may_have_dropped_last_frame = false; card->type = card_type; if (card->output.get() != output) { card->output.reset(output); @@ -500,6 +535,8 @@ void Mixer::configure_card(unsigned card_index, CaptureInterface *capture, CardT PixelFormat pixel_format; if (card_type == CardType::FFMPEG_INPUT) { pixel_format = capture->get_current_pixel_format(); + } else if (card_type == CardType::CEF_INPUT) { + pixel_format = PixelFormat_8BitBGRA; } else if (global_flags.ten_bit_input) { pixel_format = PixelFormat_10BitYCbCr; } else { @@ -564,6 +601,9 @@ void Mixer::configure_card(unsigned card_index, CaptureInterface *capture, CardT case CardType::FFMPEG_INPUT: labels.emplace_back("cardtype", "ffmpeg"); break; + case CardType::CEF_INPUT: + labels.emplace_back("cardtype", "cef"); + break; default: assert(false); } @@ -684,8 +724,8 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, size_t num_samples = (audio_frame.len > audio_offset) ? (audio_frame.len - audio_offset) / audio_format.num_channels / (audio_format.bits_per_sample / 8) : 0; if (num_samples > OUTPUT_FREQUENCY / 10) { - printf("Card %d: Dropping frame with implausible audio length (len=%d, offset=%d) [timecode=0x%04x video_len=%d video_offset=%d video_format=%x)\n", - card_index, int(audio_frame.len), int(audio_offset), + printf("%s: Dropping frame with implausible audio length (len=%d, offset=%d) [timecode=0x%04x video_len=%d video_offset=%d video_format=%x)\n", + spec_to_string(device).c_str(), int(audio_frame.len), int(audio_offset), timecode, int(video_frame.len), int(video_offset), video_format.id); if (video_frame.owner) { video_frame.owner->release_frame(video_frame); @@ -706,15 +746,15 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, const int silence_samples = OUTPUT_FREQUENCY * video_format.frame_rate_den / video_format.frame_rate_nom; if (dropped_frames > MAX_FPS * 2) { - fprintf(stderr, "Card %d lost more than two seconds (or time code jumping around; from 0x%04x to 0x%04x), resetting resampler\n", - card_index, card->last_timecode, timecode); + fprintf(stderr, "%s lost more than two seconds (or time code jumping around; from 0x%04x to 0x%04x), resetting resampler\n", + spec_to_string(device).c_str(), card->last_timecode, timecode); audio_mixer.reset_resampler(device); dropped_frames = 0; ++card->metric_input_resets; } else if (dropped_frames > 0) { // Insert silence as needed. - fprintf(stderr, "Card %d dropped %d frame(s) (before timecode 0x%04x), inserting silence.\n", - card_index, dropped_frames, timecode); + fprintf(stderr, "%s dropped %d frame(s) (before timecode 0x%04x), inserting silence.\n", + spec_to_string(device).c_str(), dropped_frames, timecode); card->metric_input_dropped_frames_error += dropped_frames; bool success; @@ -759,8 +799,8 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, if (video_frame.len - video_offset == 0 || video_frame.len - video_offset != expected_length) { if (video_frame.len != 0) { - printf("Card %d: Dropping video frame with wrong length (%ld; expected %ld)\n", - card_index, video_frame.len - video_offset, expected_length); + printf("%s: Dropping video frame with wrong length (%ld; expected %ld)\n", + spec_to_string(device).c_str(), video_frame.len - video_offset, expected_length); } if (video_frame.owner) { video_frame.owner->release_frame(video_frame); @@ -904,6 +944,7 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, new_frame.received_timestamp = video_frame.received_timestamp; // Ignore the audio timestamp. card->new_frames.push_back(move(new_frame)); card->jitter_history.frame_arrived(video_frame.received_timestamp, frame_length, dropped_frames); + card->may_have_dropped_last_frame = false; } card->new_frames_changed.notify_all(); } @@ -933,7 +974,7 @@ void Mixer::thread_func() // Start the actual capture. (We don't want to do it before we're actually ready // to process output frames.) - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { if (int(card_index) != output_card_index) { cards[card_index].capture->start_bm_capture(); } @@ -974,7 +1015,7 @@ void Mixer::thread_func() handle_hotplugged_cards(); - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { if (card_index == master_card_index || !has_new_frame[card_index]) { continue; } @@ -995,7 +1036,7 @@ void Mixer::thread_func() continue; } - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { if (!has_new_frame[card_index] || new_frames[card_index].frame->len == 0) continue; @@ -1080,6 +1121,10 @@ void Mixer::trim_queue(CaptureCard *card, size_t safe_queue_length) card->new_frames_changed.notify_all(); --queue_length; ++dropped_frames; + + if (queue_length == 0 && card->is_cef_capture) { + card->may_have_dropped_last_frame = true; + } } card->metric_input_dropped_frames_jitter += dropped_frames; @@ -1093,6 +1138,24 @@ void Mixer::trim_queue(CaptureCard *card, size_t safe_queue_length) #endif } +pair Mixer::get_channels_json() +{ + Channels ret; + for (int channel_idx = 2; channel_idx < theme->get_num_channels(); ++channel_idx) { + Channel *channel = ret.add_channel(); + channel->set_index(channel_idx); + channel->set_name(theme->get_channel_name(channel_idx)); + channel->set_color(theme->get_channel_color(channel_idx)); + } + string contents; + google::protobuf::util::MessageToJsonString(ret, &contents); // Ignore any errors. + return make_pair(contents, "text/json"); +} + +pair Mixer::get_channel_color_http(unsigned channel_idx) +{ + return make_pair(theme->get_channel_color(channel_idx), "text/plain"); +} Mixer::OutputFrameInfo Mixer::get_one_frame_from_each_card(unsigned master_card_index, bool master_card_is_output, CaptureCard::NewFrame new_frames[MAX_VIDEO_CARDS], bool has_new_frame[MAX_VIDEO_CARDS]) { @@ -1122,10 +1185,20 @@ start: goto start; } - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { CaptureCard *card = &cards[card_index]; if (card->new_frames.empty()) { // Starvation. ++card->metric_input_duped_frames; +#ifdef HAVE_CEF + if (card->is_cef_capture && card->may_have_dropped_last_frame) { + // Unlike other sources, CEF is not guaranteed to send us a steady + // stream of frames, so we'll have to ask it to repaint the frame + // we dropped. (may_have_dropped_last_frame is set whenever we + // trim the queue completely away, and cleared when we actually + // get a new frame.) + ((CEFCapture *)card->capture.get())->request_new_frame(); + } +#endif } else { new_frames[card_index] = move(card->new_frames.front()); has_new_frame[card_index] = true; @@ -1144,7 +1217,7 @@ start: output_jitter_history.frame_arrived(output_frame_info.frame_timestamp, output_frame_info.frame_duration, output_frame_info.dropped_frames); } - for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) { + for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) { CaptureCard *card = &cards[card_index]; if (has_new_frame[card_index] && !input_card_is_master_clock(card_index, master_card_index) && @@ -1277,12 +1350,6 @@ void Mixer::render_one_frame(int64_t duration) theme_main_chain.setup_chain(); //theme_main_chain.chain->enable_phase_timing(true); - // The theme can't (or at least shouldn't!) call connect_signal() on - // each FFmpeg input, so we'll do it here. - for (const pair &conn : theme->get_signal_connections()) { - conn.first->connect_signal_raw(conn.second->get_card_index(), input_state); - } - // If HDMI/SDI output is active and the user has requested auto mode, // its mode overrides the existing Y'CbCr setting for the chain. YCbCrLumaCoefficients ycbcr_output_coefficients; @@ -1387,18 +1454,18 @@ void Mixer::render_one_frame(int64_t duration) live_frame.ready_fence = fence; live_frame.input_frames = {}; live_frame.temp_textures = { y_display_tex, cbcr_display_tex }; - output_channel[OUTPUT_LIVE].output_frame(live_frame); + output_channel[OUTPUT_LIVE].output_frame(move(live_frame)); // Set up preview and any additional channels. for (int i = 1; i < theme->get_num_channels() + 2; ++i) { DisplayFrame display_frame; Theme::Chain chain = theme->get_chain(i, pts(), global_flags.width, global_flags.height, input_state); // FIXME: dimensions - display_frame.chain = chain.chain; - display_frame.setup_chain = chain.setup_chain; + display_frame.chain = move(chain.chain); + display_frame.setup_chain = move(chain.setup_chain); display_frame.ready_fence = fence; - display_frame.input_frames = chain.input_frames; + display_frame.input_frames = move(chain.input_frames); display_frame.temp_textures = {}; - output_channel[i].output_frame(display_frame); + output_channel[i].output_frame(move(display_frame)); } } @@ -1518,7 +1585,7 @@ Mixer::OutputChannel::~OutputChannel() } } -void Mixer::OutputChannel::output_frame(DisplayFrame frame) +void Mixer::OutputChannel::output_frame(DisplayFrame &&frame) { // Store this frame for display. Remove the ready frame if any // (it was seemingly never used). @@ -1527,7 +1594,7 @@ void Mixer::OutputChannel::output_frame(DisplayFrame frame) if (has_ready_frame) { parent->release_display_frame(&ready_frame); } - ready_frame = frame; + ready_frame = move(frame); has_ready_frame = true; // Call the callbacks under the mutex (they should be short), @@ -1590,7 +1657,7 @@ bool Mixer::OutputChannel::get_display_frame(DisplayFrame *frame) } if (has_ready_frame) { assert(!has_current_frame); - current_frame = ready_frame; + current_frame = move(ready_frame); ready_frame.ready_fence.reset(); // Drop the refcount. ready_frame.input_frames.clear(); // Drop the refcounts. has_current_frame = true;