]> git.sesse.net Git - nageru/blobdiff - nageru/mixer.cpp
Unbreak showing the first two channels in the tally JSON.
[nageru] / nageru / mixer.cpp
index c9f7ed78861b7bf3136c913038eff9c13a0254c5..255f28a91b315b550a130e4d67ebd48f44f231a1 100644 (file)
@@ -4,6 +4,7 @@
 
 #include <assert.h>
 #include <epoxy/egl.h>
+#include <movit/effect.h>
 #include <movit/effect_chain.h>
 #include <movit/effect_util.h>
 #include <movit/flat_input.h>
@@ -347,6 +348,11 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards)
        ycbcr_format.cb_y_position = 0.5f;
        ycbcr_format.cr_y_position = 0.5f;
 
+       // Initialize the neutral colors to sane values.
+       for (unsigned i = 0; i < MAX_VIDEO_CARDS; ++i) {
+               last_received_neutral_color[i] = RGBTriplet(1.0f, 1.0f, 1.0f);
+       }
+
        // Display chain; shows the live output produced by the main chain (or rather, a copy of it).
        display_chain.reset(new EffectChain(global_flags.width, global_flags.height, resource_pool.get()));
        check_error();
@@ -795,7 +801,8 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode,
                //
                // Only bother doing MJPEG encoding if there are any connected clients
                // that want the stream.
-               if (httpd.get_num_connected_multicam_clients() > 0) {
+               if (httpd.get_num_connected_multicam_clients() > 0 ||
+                   httpd.get_num_connected_siphon_clients(card_index) > 0) {
                        vector<int32_t> converted_samples = convert_audio_to_fixed32(audio_frame.data + audio_offset, num_samples, audio_format, 2);
                        lock_guard<mutex> lock(card_mutex);
                        if (card->new_raw_audio.empty()) {
@@ -1003,6 +1010,10 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode,
                        new_frame.video_format = video_format;
                        new_frame.y_offset = y_offset;
                        new_frame.cbcr_offset = cbcr_offset;
+                       if (card->type == CardType::FFMPEG_INPUT) {
+                               FFmpegCapture *ffmpeg_capture = static_cast<FFmpegCapture *>(card->capture.get());
+                               new_frame.neutral_color = ffmpeg_capture->get_last_neutral_color();
+                       }
                        card->new_frames.push_back(move(new_frame));
                        card->jitter_history.frame_arrived(video_frame.received_timestamp, frame_length, dropped_frames);
                        card->may_have_dropped_last_frame = false;
@@ -1114,12 +1125,24 @@ void Mixer::thread_func()
                                new_frame->upload_func = nullptr;
                        }
 
-                       if (new_frame->frame->data_copy != nullptr) {
-                               int mjpeg_card_index = mjpeg_encoder->get_mjpeg_stream_for_card(card_index);
-                               if (mjpeg_card_index != -1) {
-                                       mjpeg_encoder->upload_frame(pts_int, mjpeg_card_index, new_frame->frame, new_frame->video_format, new_frame->y_offset, new_frame->cbcr_offset, move(raw_audio[card_index]));
-                               }
+                       // Only set the white balance if it actually changed. This means that the user
+                       // is free to override the white balance in a video with no white balance information
+                       // actually set (ie. r=g=b=1 all the time), or one where the white point is wrong,
+                       // but frame-to-frame decisions will be heeded. We do this pretty much as late
+                       // as possible (ie., after picking out the frame from the buffer), so that we are sure
+                       // that the change takes effect on exactly the right frame.
+                       if (fabs(new_frame->neutral_color.r - last_received_neutral_color[card_index].r) > 1e-3 ||
+                           fabs(new_frame->neutral_color.g - last_received_neutral_color[card_index].g) > 1e-3 ||
+                           fabs(new_frame->neutral_color.b - last_received_neutral_color[card_index].b) > 1e-3) {
+                               theme->set_wb_for_signal(card_index, new_frame->neutral_color.r, new_frame->neutral_color.g, new_frame->neutral_color.b);
+                               last_received_neutral_color[card_index] = new_frame->neutral_color;
                        }
+
+                       if (new_frame->frame->data_copy != nullptr && mjpeg_encoder->should_encode_mjpeg_for_card(card_index)) {
+                               RGBTriplet neutral_color = theme->get_white_balance_for_signal(card_index);
+                               mjpeg_encoder->upload_frame(pts_int, card_index, new_frame->frame, new_frame->video_format, new_frame->y_offset, new_frame->cbcr_offset, move(raw_audio[card_index]), neutral_color);
+                       }
+
                }
 
                int64_t frame_duration = output_frame_info.frame_duration;
@@ -1215,11 +1238,11 @@ void Mixer::trim_queue(CaptureCard *card, size_t safe_queue_length)
 pair<string, string> Mixer::get_channels_json()
 {
        Channels ret;
-       for (int channel_idx = 2; channel_idx < theme->get_num_channels(); ++channel_idx) {
+       for (int channel_idx = 0; channel_idx < theme->get_num_channels(); ++channel_idx) {
                Channel *channel = ret.add_channel();
-               channel->set_index(channel_idx);
-               channel->set_name(theme->get_channel_name(channel_idx));
-               channel->set_color(theme->get_channel_color(channel_idx));
+               channel->set_index(channel_idx + 2);
+               channel->set_name(theme->get_channel_name(channel_idx + 2));
+               channel->set_color(theme->get_channel_color(channel_idx + 2));
        }
        string contents;
        google::protobuf::util::MessageToJsonString(ret, &contents);  // Ignore any errors.