#include "basic_stats.h"
#include "bmusb/bmusb.h"
#include "bmusb/fake_capture.h"
+#ifdef HAVE_CEF
+#include "cef_capture.h"
+#endif
#include "chroma_subsampler.h"
#include "context.h"
#include "decklink_capture.h"
#include "v210_converter.h"
#include "video_encoder.h"
+#undef Status
+#include <google/protobuf/util/json_util.h>
+#include "json.pb.h"
+
class IDeckLink;
class QOpenGLContext;
}
check_error();
break;
+ default:
+ assert(false);
}
userdata->last_width[field] = width;
userdata->last_height[field] = height;
if (percentile <= 0.5) {
return *next(orders.begin(), elem_idx) * multiplier;
} else {
- return *prev(orders.end(), elem_idx + 1) * multiplier;
+ return *prev(orders.end(), orders.size() - elem_idx) * multiplier;
}
}
// Must be instantiated after VideoEncoder has initialized global_flags.use_zerocopy.
theme.reset(new Theme(global_flags.theme_filename, global_flags.theme_dirs, resource_pool.get(), num_cards));
+ httpd.add_endpoint("/channels", bind(&Mixer::get_channels_json, this), HTTPD::ALLOW_ALL_ORIGINS);
+ for (int channel_idx = 2; channel_idx < theme->get_num_channels(); ++channel_idx) {
+ char url[256];
+ snprintf(url, sizeof(url), "/channels/%d/color", channel_idx);
+ httpd.add_endpoint(url, bind(&Mixer::get_channel_color_http, this, unsigned(channel_idx)), HTTPD::ALLOW_ALL_ORIGINS);
+ }
+
// Start listening for clients only once VideoEncoder has written its header, if any.
httpd.start(global_flags.http_port);
}
num_video_inputs = video_inputs.size();
+#ifdef HAVE_CEF
+ // Same, for HTML inputs.
+ std::vector<CEFCapture *> html_inputs = theme->get_html_inputs();
+ for (unsigned html_card_index = 0; html_card_index < html_inputs.size(); ++card_index, ++html_card_index) {
+ if (card_index >= MAX_VIDEO_CARDS) {
+ fprintf(stderr, "ERROR: Not enough card slots available for the HTML inputs the theme requested.\n");
+ exit(1);
+ }
+ configure_card(card_index, html_inputs[html_card_index], CardType::CEF_INPUT, /*output=*/nullptr);
+ html_inputs[html_card_index]->set_card_index(card_index);
+ }
+ num_html_inputs = html_inputs.size();
+#endif
+
BMUSBCapture::set_card_connected_callback(bind(&Mixer::bm_hotplug_add, this, _1));
BMUSBCapture::start_bm_thread();
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
cards[card_index].queue_length_policy.reset(card_index);
}
{
BMUSBCapture::stop_bm_thread();
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
{
unique_lock<mutex> lock(card_mutex);
cards[card_index].should_quit = true; // Unblock thread.
}
card->capture.reset(capture);
card->is_fake_capture = (card_type == CardType::FAKE_CAPTURE);
+ card->is_cef_capture = (card_type == CardType::CEF_INPUT);
+ card->may_have_dropped_last_frame = false;
card->type = card_type;
if (card->output.get() != output) {
card->output.reset(output);
PixelFormat pixel_format;
if (card_type == CardType::FFMPEG_INPUT) {
pixel_format = capture->get_current_pixel_format();
+ } else if (card_type == CardType::CEF_INPUT) {
+ pixel_format = PixelFormat_8BitBGRA;
} else if (global_flags.ten_bit_input) {
pixel_format = PixelFormat_10BitYCbCr;
} else {
case CardType::FFMPEG_INPUT:
labels.emplace_back("cardtype", "ffmpeg");
break;
+ case CardType::CEF_INPUT:
+ labels.emplace_back("cardtype", "cef");
+ break;
default:
assert(false);
}
new_frame.received_timestamp = video_frame.received_timestamp; // Ignore the audio timestamp.
card->new_frames.push_back(move(new_frame));
card->jitter_history.frame_arrived(video_frame.received_timestamp, frame_length, dropped_frames);
+ card->may_have_dropped_last_frame = false;
}
card->new_frames_changed.notify_all();
}
// Start the actual capture. (We don't want to do it before we're actually ready
// to process output frames.)
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
if (int(card_index) != output_card_index) {
cards[card_index].capture->start_bm_capture();
}
handle_hotplugged_cards();
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
if (card_index == master_card_index || !has_new_frame[card_index]) {
continue;
}
continue;
}
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
if (!has_new_frame[card_index] || new_frames[card_index].frame->len == 0)
continue;
card->new_frames_changed.notify_all();
--queue_length;
++dropped_frames;
+
+ if (queue_length == 0 && card->is_cef_capture) {
+ card->may_have_dropped_last_frame = true;
+ }
}
card->metric_input_dropped_frames_jitter += dropped_frames;
#endif
}
+pair<string, string> Mixer::get_channels_json()
+{
+ Channels ret;
+ for (int channel_idx = 2; channel_idx < theme->get_num_channels(); ++channel_idx) {
+ Channel *channel = ret.add_channel();
+ channel->set_index(channel_idx);
+ channel->set_name(theme->get_channel_name(channel_idx));
+ channel->set_color(theme->get_channel_color(channel_idx));
+ }
+ string contents;
+ google::protobuf::util::MessageToJsonString(ret, &contents); // Ignore any errors.
+ return make_pair(contents, "text/json");
+}
+
+pair<string, string> Mixer::get_channel_color_http(unsigned channel_idx)
+{
+ return make_pair(theme->get_channel_color(channel_idx), "text/plain");
+}
Mixer::OutputFrameInfo Mixer::get_one_frame_from_each_card(unsigned master_card_index, bool master_card_is_output, CaptureCard::NewFrame new_frames[MAX_VIDEO_CARDS], bool has_new_frame[MAX_VIDEO_CARDS])
{
goto start;
}
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
CaptureCard *card = &cards[card_index];
if (card->new_frames.empty()) { // Starvation.
++card->metric_input_duped_frames;
+#ifdef HAVE_CEF
+ if (card->is_cef_capture && card->may_have_dropped_last_frame) {
+ // Unlike other sources, CEF is not guaranteed to send us a steady
+ // stream of frames, so we'll have to ask it to repaint the frame
+ // we dropped. (may_have_dropped_last_frame is set whenever we
+ // trim the queue completely away, and cleared when we actually
+ // get a new frame.)
+ ((CEFCapture *)card->capture.get())->request_new_frame();
+ }
+#endif
} else {
new_frames[card_index] = move(card->new_frames.front());
has_new_frame[card_index] = true;
output_jitter_history.frame_arrived(output_frame_info.frame_timestamp, output_frame_info.frame_duration, output_frame_info.dropped_frames);
}
- for (unsigned card_index = 0; card_index < num_cards + num_video_inputs; ++card_index) {
+ for (unsigned card_index = 0; card_index < num_cards + num_video_inputs + num_html_inputs; ++card_index) {
CaptureCard *card = &cards[card_index];
if (has_new_frame[card_index] &&
!input_card_is_master_clock(card_index, master_card_index) &&
//theme_main_chain.chain->enable_phase_timing(true);
// The theme can't (or at least shouldn't!) call connect_signal() on
- // each FFmpeg input, so we'll do it here.
- for (const pair<LiveInputWrapper *, FFmpegCapture *> &conn : theme->get_signal_connections()) {
+ // each FFmpeg or CEF input, so we'll do it here.
+ for (const pair<LiveInputWrapper *, FFmpegCapture *> &conn : theme->get_video_signal_connections()) {
conn.first->connect_signal_raw(conn.second->get_card_index(), input_state);
}
+#ifdef HAVE_CEF
+ for (const pair<LiveInputWrapper *, CEFCapture *> &conn : theme->get_html_signal_connections()) {
+ conn.first->connect_signal_raw(conn.second->get_card_index(), input_state);
+ }
+#endif
// If HDMI/SDI output is active and the user has requested auto mode,
// its mode overrides the existing Y'CbCr setting for the chain.
live_frame.ready_fence = fence;
live_frame.input_frames = {};
live_frame.temp_textures = { y_display_tex, cbcr_display_tex };
- output_channel[OUTPUT_LIVE].output_frame(live_frame);
+ output_channel[OUTPUT_LIVE].output_frame(move(live_frame));
// Set up preview and any additional channels.
for (int i = 1; i < theme->get_num_channels() + 2; ++i) {
display_frame.ready_fence = fence;
display_frame.input_frames = move(chain.input_frames);
display_frame.temp_textures = {};
- output_channel[i].output_frame(display_frame);
+ output_channel[i].output_frame(move(display_frame));
}
}
}
}
-void Mixer::OutputChannel::output_frame(DisplayFrame frame)
+void Mixer::OutputChannel::output_frame(DisplayFrame &&frame)
{
// Store this frame for display. Remove the ready frame if any
// (it was seemingly never used).