X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=nageru%2Fmixer.cpp;h=c91857e62fbab09b3677fcebba41402235feabc6;hb=9ffd4f03f314cc6e0254449593def95c9bc203d6;hp=c25b22dc711fa528c2720493504f9bebcfe12f3b;hpb=647810a468e01fab5d7f0fcd122565ab000a2e73;p=nageru diff --git a/nageru/mixer.cpp b/nageru/mixer.cpp index c25b22d..c91857e 100644 --- a/nageru/mixer.cpp +++ b/nageru/mixer.cpp @@ -432,7 +432,7 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) for (unsigned video_card_index = 0; video_card_index < video_inputs.size(); ++card_index, ++video_card_index) { if (card_index >= MAX_VIDEO_CARDS) { fprintf(stderr, "ERROR: Not enough card slots available for the videos the theme requested.\n"); - exit(1); + abort(); } configure_card(card_index, video_inputs[video_card_index], CardType::FFMPEG_INPUT, /*output=*/nullptr); video_inputs[video_card_index]->set_card_index(card_index); @@ -445,7 +445,7 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) for (unsigned html_card_index = 0; html_card_index < html_inputs.size(); ++card_index, ++html_card_index) { if (card_index >= MAX_VIDEO_CARDS) { fprintf(stderr, "ERROR: Not enough card slots available for the HTML inputs the theme requested.\n"); - exit(1); + abort(); } configure_card(card_index, html_inputs[html_card_index], CardType::CEF_INPUT, /*output=*/nullptr); html_inputs[html_card_index]->set_card_index(card_index); @@ -466,7 +466,7 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) if (!v210Converter::has_hardware_support()) { fprintf(stderr, "ERROR: --ten-bit-input requires support for OpenGL compute shaders\n"); fprintf(stderr, " (OpenGL 4.3, or GL_ARB_compute_shader + GL_ARB_shader_image_load_store).\n"); - exit(1); + abort(); } v210_converter.reset(new v210Converter()); @@ -483,7 +483,7 @@ Mixer::Mixer(const QSurfaceFormat &format, unsigned num_cards) if (!v210Converter::has_hardware_support()) { fprintf(stderr, "ERROR: --ten-bit-output requires support for OpenGL compute shaders\n"); fprintf(stderr, " (OpenGL 4.3, or GL_ARB_compute_shader + GL_ARB_shader_image_load_store).\n"); - exit(1); + abort(); } } @@ -551,7 +551,7 @@ void Mixer::configure_card(unsigned card_index, CaptureInterface *capture, CardT card->capture->set_frame_callback(bind(&Mixer::bm_frame, this, card_index, _1, _2, _3, _4, _5, _6, _7)); if (card->frame_allocator == nullptr) { - card->frame_allocator.reset(new PBOFrameAllocator(pixel_format, 8 << 20, global_flags.width, global_flags.height)); // 8 MB. + card->frame_allocator.reset(new PBOFrameAllocator(pixel_format, 8 << 20, global_flags.width, global_flags.height, card_index, mjpeg_encoder.get())); // 8 MB. } card->capture->set_video_frame_allocator(card->frame_allocator.get()); if (card->surface == nullptr) { @@ -779,12 +779,12 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, bool success; do { - success = audio_mixer->add_silence(device, silence_samples, dropped_frames, frame_length); + success = audio_mixer->add_silence(device, silence_samples, dropped_frames); } while (!success); } if (num_samples > 0) { - audio_mixer->add_audio(device, audio_frame.data + audio_offset, num_samples, audio_format, frame_length, audio_frame.received_timestamp); + audio_mixer->add_audio(device, audio_frame.data + audio_offset, num_samples, audio_format, audio_frame.received_timestamp); } // Done with the audio, so release it. @@ -795,10 +795,15 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, card->last_timecode = timecode; PBOFrameAllocator::Userdata *userdata = (PBOFrameAllocator::Userdata *)video_frame.userdata; + if (card->type == CardType::FFMPEG_INPUT && userdata != nullptr) { + FFmpegCapture *ffmpeg_capture = static_cast(card->capture.get()); + userdata->has_last_subtitle = ffmpeg_capture->get_has_last_subtitle(); + userdata->last_subtitle = ffmpeg_capture->get_last_subtitle(); + } size_t cbcr_width, cbcr_height, cbcr_offset, y_offset; size_t expected_length = video_format.stride * (video_format.height + video_format.extra_lines_top + video_format.extra_lines_bottom); - if (userdata->pixel_format == PixelFormat_8BitYCbCrPlanar) { + if (userdata != nullptr && userdata->pixel_format == PixelFormat_8BitYCbCrPlanar) { // The calculation above is wrong for planar Y'CbCr, so just override it. assert(card->type == CardType::FFMPEG_INPUT); assert(video_offset == 0); @@ -819,7 +824,7 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, if (video_frame.len - video_offset == 0 || video_frame.len - video_offset != expected_length) { if (video_frame.len != 0) { - printf("%s: Dropping video frame with wrong length (%ld; expected %ld)\n", + printf("%s: Dropping video frame with wrong length (%zu; expected %zu)\n", spec_to_string(device).c_str(), video_frame.len - video_offset, expected_length); } if (video_frame.owner) { @@ -829,7 +834,7 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, // Still send on the information that we _had_ a frame, even though it's corrupted, // so that pts can go up accordingly. { - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); CaptureCard::NewFrame new_frame; new_frame.frame = RefCountedFrame(FrameAllocator::Frame()); new_frame.length = frame_length; @@ -860,6 +865,7 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, } frame_upload_start = steady_clock::now(); } + assert(userdata != nullptr); userdata->last_interlaced = video_format.interlaced; userdata->last_has_signal = video_format.has_signal; userdata->last_is_connected = video_format.is_connected; @@ -953,7 +959,7 @@ void Mixer::bm_frame(unsigned card_index, uint16_t timecode, } { - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); CaptureCard::NewFrame new_frame; new_frame.frame = frame; new_frame.length = frame_length; @@ -992,7 +998,7 @@ void Mixer::thread_func() QOpenGLContext *context = create_context(mixer_surface); if (!make_current(context, mixer_surface)) { printf("oops\n"); - exit(1); + abort(); } // Start the actual capture. (We don't want to do it before we're actually ready @@ -1075,12 +1081,10 @@ void Mixer::thread_func() new_frame->upload_func = nullptr; } - // Only bother doing MJPEG encoding if there are any connected clients - // that want the stream. - if (httpd.get_num_connected_multicam_clients() > 0) { - auto stream_it = global_flags.card_to_mjpeg_stream_export.find(card_index); - if (stream_it != global_flags.card_to_mjpeg_stream_export.end()) { - mjpeg_encoder->upload_frame(pts_int, stream_it->second, new_frame->frame, new_frame->video_format, new_frame->y_offset, new_frame->cbcr_offset); + if (new_frame->frame->data_copy != nullptr) { + int mjpeg_card_index = mjpeg_encoder->get_mjpeg_stream_for_card(card_index); + if (mjpeg_card_index != -1) { + mjpeg_encoder->upload_frame(pts_int, mjpeg_card_index, new_frame->frame, new_frame->video_format, new_frame->y_offset, new_frame->cbcr_offset); } } } @@ -1348,7 +1352,7 @@ void Mixer::schedule_audio_resampling_tasks(unsigned dropped_frames, int num_sam // non-dropped frame; perhaps we should just discard that as well, // since dropped frames are expected to be rare, and it might be // better to just wait until we have a slightly more normal situation). - unique_lock lock(audio_mutex); + lock_guard lock(audio_mutex); bool adjust_rate = !dropped_frame && !is_preroll; audio_task_queue.push(AudioTask{pts_int, num_samples_per_frame, adjust_rate, frame_timestamp}); audio_task_queue_changed.notify_one(); @@ -1372,7 +1376,7 @@ void Mixer::render_one_frame(int64_t duration) // Update Y'CbCr settings for all cards. { - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); for (unsigned card_index = 0; card_index < num_cards; ++card_index) { YCbCrInterpretation *interpretation = &ycbcr_interpretation[card_index]; input_state.ycbcr_coefficients_auto[card_index] = interpretation->ycbcr_coefficients_auto; @@ -1578,13 +1582,13 @@ void Mixer::channel_clicked(int preview_num) YCbCrInterpretation Mixer::get_input_ycbcr_interpretation(unsigned card_index) const { - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); return ycbcr_interpretation[card_index]; } void Mixer::set_input_ycbcr_interpretation(unsigned card_index, const YCbCrInterpretation &interpretation) { - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); ycbcr_interpretation[card_index] = interpretation; } @@ -1608,7 +1612,7 @@ void Mixer::start_mode_scanning(unsigned card_index) map Mixer::get_available_output_video_modes() const { assert(desired_output_card_index != -1); - unique_lock lock(card_mutex); + lock_guard lock(card_mutex); return cards[desired_output_card_index].output->get_available_video_modes(); } @@ -1646,7 +1650,7 @@ void Mixer::OutputChannel::output_frame(DisplayFrame &&frame) // Store this frame for display. Remove the ready frame if any // (it was seemingly never used). { - unique_lock lock(frame_mutex); + lock_guard lock(frame_mutex); if (has_ready_frame) { parent->release_display_frame(&ready_frame); } @@ -1701,7 +1705,7 @@ void Mixer::OutputChannel::output_frame(DisplayFrame &&frame) bool Mixer::OutputChannel::get_display_frame(DisplayFrame *frame) { - unique_lock lock(frame_mutex); + lock_guard lock(frame_mutex); if (!has_current_frame && !has_ready_frame) { return false; } @@ -1726,13 +1730,13 @@ bool Mixer::OutputChannel::get_display_frame(DisplayFrame *frame) void Mixer::OutputChannel::add_frame_ready_callback(void *key, Mixer::new_frame_ready_callback_t callback) { - unique_lock lock(frame_mutex); + lock_guard lock(frame_mutex); new_frame_ready_callbacks[key] = callback; } void Mixer::OutputChannel::remove_frame_ready_callback(void *key) { - unique_lock lock(frame_mutex); + lock_guard lock(frame_mutex); new_frame_ready_callbacks.erase(key); }