cbcr_width != userdata->last_cbcr_width[field] ||
cbcr_height != userdata->last_cbcr_height[field];
const bool recreate_v210_texture =
- global_flags.ten_bit_input &&
+ global_flags.bit_depth > 8 &&
(first || v210_width != userdata->last_v210_width[field] || height != userdata->last_height[field]);
if (recreate_main_texture) {
void JitterHistory::frame_arrived(steady_clock::time_point now, int64_t frame_duration, size_t dropped_frames)
{
+ if (frame_duration != last_duration) {
+ // If the frame rate changed, the input clock is also going to change,
+ // so our historical data doesn't make much sense anymore.
+ // Also, format changes typically introduce blips that are not representative
+ // of the typical frame stream. (We make the assumption that format changes
+ // don't happen all the time in regular use; if they did, we should probably
+ // rather keep the history so that we take jitter they may introduce into account.)
+ clear();
+ last_duration = frame_duration;
+ }
if (expected_timestamp > steady_clock::time_point::min()) {
expected_timestamp += dropped_frames * nanoseconds(frame_duration * 1000000000 / TIMEBASE);
double jitter_seconds = fabs(duration<double>(expected_timestamp - now).count());
}
void QueueLengthPolicy::update_policy(steady_clock::time_point now,
- steady_clock::time_point expected_next_frame,
+ steady_clock::time_point expected_next_input_frame,
int64_t input_frame_duration,
int64_t master_frame_duration,
double max_input_card_jitter_seconds,
// Figure out when we can expect the next frame for this card, assuming
// worst-case jitter (ie., the frame is maximally late).
- double seconds_until_next_frame = max(duration<double>(expected_next_frame - now).count() + max_input_card_jitter_seconds, 0.0);
+ double seconds_until_next_frame = max(duration<double>(expected_next_input_frame - now).count() + max_input_card_jitter_seconds, 0.0);
// How many times are the master card expected to tick in that time?
// We assume the master clock has worst-case jitter but not any rate
ycbcr_format.luma_coefficients = YCBCR_REC_601;
}
ycbcr_format.full_range = false;
- ycbcr_format.num_levels = 1 << global_flags.x264_bit_depth;
+ ycbcr_format.num_levels = 1 << global_flags.bit_depth;
ycbcr_format.cb_x_position = 0.0f;
ycbcr_format.cr_x_position = 0.0f;
ycbcr_format.cb_y_position = 0.5f;
// Display chain; shows the live output produced by the main chain (or rather, a copy of it).
display_chain.reset(new EffectChain(global_flags.width, global_flags.height, resource_pool.get()));
check_error();
- GLenum type = global_flags.x264_bit_depth > 8 ? GL_UNSIGNED_SHORT : GL_UNSIGNED_BYTE;
+ GLenum type = global_flags.bit_depth > 8 ? GL_UNSIGNED_SHORT : GL_UNSIGNED_BYTE;
display_input = new YCbCrInput(inout_format, ycbcr_format, global_flags.width, global_flags.height, YCBCR_INPUT_SPLIT_Y_AND_CBCR, type);
display_chain->add_input(display_input);
display_chain->add_output(inout_format, OUTPUT_ALPHA_FORMAT_POSTMULTIPLIED);
DeckLinkCapture *capture = new DeckLinkCapture(decklink, card_index);
DeckLinkOutput *output = new DeckLinkOutput(resource_pool.get(), decklink_output_surface, global_flags.width, global_flags.height, card_index);
- if (!output->set_device(decklink)) {
+ if (!output->set_device(decklink, capture->get_input())) {
delete output;
output = nullptr;
}
chroma_subsampler.reset(new ChromaSubsampler(resource_pool.get()));
- if (global_flags.ten_bit_input) {
+ if (global_flags.bit_depth > 8) {
if (!v210Converter::has_hardware_support()) {
fprintf(stderr, "ERROR: --ten-bit-input requires support for OpenGL compute shaders\n");
fprintf(stderr, " (OpenGL 4.3, or GL_ARB_compute_shader + GL_ARB_shader_image_load_store).\n");
v210_converter->precompile_shader(3840);
v210_converter->precompile_shader(4096);
}
- if (global_flags.ten_bit_output) {
+ if (global_flags.bit_depth > 8) {
if (!v210Converter::has_hardware_support()) {
fprintf(stderr, "ERROR: --ten-bit-output requires support for OpenGL compute shaders\n");
fprintf(stderr, " (OpenGL 4.3, or GL_ARB_compute_shader + GL_ARB_shader_image_load_store).\n");
if (global_flags.enable_alsa_output) {
alsa.reset(new ALSAOutput(OUTPUT_FREQUENCY, /*num_channels=*/2));
}
+ output_card_is_master = global_flags.output_card_is_master;
if (global_flags.output_card != -1) {
desired_output_card_index = global_flags.output_card;
set_output_card_internal(global_flags.output_card);
pixel_format = capture->get_current_pixel_format();
} else if (card_type == CardType::CEF_INPUT) {
pixel_format = PixelFormat_8BitBGRA;
- } else if (global_flags.ten_bit_input) {
+ } else if (global_flags.bit_depth > 8) {
pixel_format = PixelFormat_10BitYCbCr;
} else {
pixel_format = PixelFormat_8BitYCbCr;
card->jitter_history.clear();
card->capture->start_bm_capture();
desired_output_video_mode = output_video_mode = card->output->pick_video_mode(desired_output_video_mode);
- card->output->start_output(desired_output_video_mode, pts_int);
+ card->output->start_output(desired_output_video_mode, pts_int, /*is_master_card=*/output_card_is_master);
}
output_card_index = card_index;
output_jitter_history.clear();
// (Could be nonintegral, but resampling will save us then.)
const int silence_samples = OUTPUT_FREQUENCY * video_format.frame_rate_den / video_format.frame_rate_nom;
- if (dropped_frames > MAX_FPS * 2) {
+ if (dropped_frames > TYPICAL_FPS * 2) {
fprintf(stderr, "%s lost more than two seconds (or time code jumping around; from 0x%04x to 0x%04x), resetting resampler\n",
description_for_card(card_index).c_str(), card->last_timecode, timecode);
audio_mixer->reset_resampler(device);
DeckLinkOutput *output = cards[output_card_index].output.get();
output->end_output();
desired_output_video_mode = output_video_mode = output->pick_video_mode(desired_output_video_mode);
- output->start_output(desired_output_video_mode, pts_int);
+ output->start_output(desired_output_video_mode, pts_int, /*is_master_card=*/output_card_is_master);
}
{
bool master_card_is_output;
unsigned master_card_index;
- if (output_card_index != -1) {
+ if (output_card_index != -1 && output_card_is_master) {
master_card_is_output = true;
master_card_index = output_card_index;
} else {
bool Mixer::input_card_is_master_clock(unsigned card_index, unsigned master_card_index) const
{
- if (output_card_index != -1) {
+ if (output_card_index != -1 && output_card_is_master) {
// The output card (ie., cards[output_card_index].output) is the master clock,
// so no input card (ie., cards[card_index].capture) is.
return false;
output_ycbcr_format.chroma_subsampling_y = 1;
output_ycbcr_format.luma_coefficients = ycbcr_output_coefficients;
output_ycbcr_format.full_range = false;
- output_ycbcr_format.num_levels = 1 << global_flags.x264_bit_depth;
+ output_ycbcr_format.num_levels = 1 << global_flags.bit_depth;
chain->change_ycbcr_output_format(output_ycbcr_format);
// Render main chain. If we're using zerocopy Quick Sync encoding
GLuint y_tex, cbcr_full_tex, cbcr_tex;
GLuint y_copy_tex, cbcr_copy_tex = 0;
GLuint y_display_tex, cbcr_display_tex;
- GLenum y_type = (global_flags.x264_bit_depth > 8) ? GL_R16 : GL_R8;
- GLenum cbcr_type = (global_flags.x264_bit_depth > 8) ? GL_RG16 : GL_RG8;
+ GLenum y_type = (global_flags.bit_depth > 8) ? GL_R16 : GL_R8;
+ GLenum cbcr_type = (global_flags.bit_depth > 8) ? GL_RG16 : GL_RG8;
const bool is_zerocopy = video_encoder->is_zerocopy();
if (is_zerocopy) {
cbcr_full_tex = resource_pool->create_2d_texture(cbcr_type, global_flags.width, global_flags.height);
void Mixer::start_mode_scanning(unsigned card_index)
{
assert(card_index < MAX_VIDEO_CARDS);
- if (cards[card_index].capture != nullptr) {
+ if (cards[card_index].capture == nullptr) {
// Inactive card. Should never happen.
return;
}