check_error();
resource_pool.reset(new ResourcePool);
- chain.reset(new EffectChain(WIDTH, HEIGHT, resource_pool.get()));
- check_error();
+ theme.reset(new Theme("theme.lua", resource_pool.get()));
+ output_channel[OUTPUT_LIVE].parent = this;
+ output_channel[OUTPUT_PREVIEW].parent = this;
+ output_channel[OUTPUT_INPUT0].parent = this;
+ output_channel[OUTPUT_INPUT1].parent = this;
ImageFormat inout_format;
inout_format.color_space = COLORSPACE_sRGB;
inout_format.gamma_curve = GAMMA_sRGB;
- YCbCrFormat ycbcr_format;
- ycbcr_format.chroma_subsampling_x = 2;
- ycbcr_format.chroma_subsampling_y = 1;
- ycbcr_format.cb_x_position = 0.0;
- ycbcr_format.cr_x_position = 0.0;
- ycbcr_format.cb_y_position = 0.5;
- ycbcr_format.cr_y_position = 0.5;
- ycbcr_format.luma_coefficients = YCBCR_REC_601;
- ycbcr_format.full_range = false;
-
- input[0] = new YCbCrInput(inout_format, ycbcr_format, WIDTH, HEIGHT, YCBCR_INPUT_SPLIT_Y_AND_CBCR);
- chain->add_input(input[0]);
- input[1] = new YCbCrInput(inout_format, ycbcr_format, WIDTH, HEIGHT, YCBCR_INPUT_SPLIT_Y_AND_CBCR);
- chain->add_input(input[1]);
- resample_effect = chain->add_effect(new ResampleEffect(), input[0]);
- padding_effect = chain->add_effect(new IntegralPaddingEffect());
- float border_color[] = { 0.0f, 0.0f, 0.0f, 1.0f };
- CHECK(padding_effect->set_vec4("border_color", border_color));
-
- resample2_effect = chain->add_effect(new ResampleEffect(), input[1]);
- Effect *saturation_effect = chain->add_effect(new SaturationEffect());
- CHECK(saturation_effect->set_float("saturation", 0.3f));
- Effect *wb_effect = chain->add_effect(new WhiteBalanceEffect());
- CHECK(wb_effect->set_float("output_color_temperature", 3500.0));
- padding2_effect = chain->add_effect(new IntegralPaddingEffect());
-
- chain->add_effect(new OverlayEffect(), padding_effect, padding2_effect);
-
- ycbcr_format.chroma_subsampling_x = 1;
-
- chain->add_output(inout_format, OUTPUT_ALPHA_FORMAT_POSTMULTIPLIED);
- chain->add_ycbcr_output(inout_format, OUTPUT_ALPHA_FORMAT_POSTMULTIPLIED, ycbcr_format, YCBCR_OUTPUT_SPLIT_Y_AND_CBCR);
- chain->set_dither_bits(8);
- chain->set_output_origin(OUTPUT_ORIGIN_TOP_LEFT);
- chain->finalize();
+ // Display chain; shows the live output produced by the main chain (its RGBA version).
+ display_chain.reset(new EffectChain(WIDTH, HEIGHT, resource_pool.get()));
+ check_error();
+ display_input = new FlatInput(inout_format, FORMAT_RGB, GL_UNSIGNED_BYTE, WIDTH, HEIGHT); // FIXME: GL_UNSIGNED_BYTE is really wrong.
+ display_chain->add_input(display_input);
+ display_chain->add_output(inout_format, OUTPUT_ALPHA_FORMAT_POSTMULTIPLIED);
+ display_chain->set_dither_bits(0); // Don't bother.
+ display_chain->finalize();
h264_encoder.reset(new H264Encoder(h264_encoder_surface, WIDTH, HEIGHT, "test.mp4"));
printf("Configuring first card...\n");
cards[0].usb = new BMUSBCapture(0x1edb, 0xbd3b); // 0xbd4f
cards[0].usb->set_frame_callback(std::bind(&Mixer::bm_frame, this, 0, _1, _2, _3, _4, _5, _6, _7));
- cards[0].frame_allocator.reset(new PBOFrameAllocator(1280 * 750 * 2 + 44));
+ cards[0].frame_allocator.reset(new PBOFrameAllocator(1280 * 750 * 2 + 44, 1280, 720));
cards[0].usb->set_video_frame_allocator(cards[0].frame_allocator.get());
cards[0].usb->configure_card();
cards[0].surface = create_surface(format);
printf("Configuring second card...\n");
cards[1].usb = new BMUSBCapture(0x1edb, 0xbd4f);
cards[1].usb->set_frame_callback(std::bind(&Mixer::bm_frame, this, 1, _1, _2, _3, _4, _5, _6, _7));
- cards[1].frame_allocator.reset(new PBOFrameAllocator(1280 * 750 * 2 + 44));
+ cards[1].frame_allocator.reset(new PBOFrameAllocator(1280 * 750 * 2 + 44, 1280, 720));
cards[1].usb->set_video_frame_allocator(cards[1].frame_allocator.get());
cards[1].usb->configure_card();
}
for (int card_index = 0; card_index < NUM_CARDS; ++card_index) {
cards[card_index].usb->start_bm_capture();
- input[card_index]->set_pixel_data(0, nullptr, 0);
- input[card_index]->set_pixel_data(1, nullptr, 0);
}
//chain->enable_phase_timing(true);
std::unique_lock<std::mutex> lock(bmusb_mutex);
card->new_data_ready_changed.wait(lock, [card]{ return !card->new_data_ready; });
}
- GLuint pbo = (GLint)(intptr_t)video_frame.userdata;
+ const PBOFrameAllocator::Userdata *userdata = (const PBOFrameAllocator::Userdata *)video_frame.userdata;
+ GLuint pbo = userdata->pbo;
check_error();
glBindBuffer(GL_PIXEL_UNPACK_BUFFER_ARB, pbo);
check_error();
check_error();
//glMemoryBarrier(GL_CLIENT_MAPPED_BUFFER_BARRIER_BIT);
//check_error();
+
+ // Upload the textures.
+ glBindTexture(GL_TEXTURE_2D, userdata->tex_y);
+ check_error();
+ glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 1280, 720, GL_RED, GL_UNSIGNED_BYTE, BUFFER_OFFSET((1280 * 750 * 2 + 44) / 2 + 1280 * 25 + 22));
+ check_error();
+ glBindTexture(GL_TEXTURE_2D, userdata->tex_cbcr);
+ check_error();
+ glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 1280/2, 720, GL_RG, GL_UNSIGNED_BYTE, BUFFER_OFFSET(1280 * 25 + 22));
+ check_error();
+ glBindTexture(GL_TEXTURE_2D, 0);
+ check_error();
GLsync fence = glFenceSync(GL_SYNC_GPU_COMMANDS_COMPLETE, /*flags=*/0);
check_error();
assert(fence != nullptr);
{
std::unique_lock<std::mutex> lock(bmusb_mutex);
card->new_data_ready = true;
- card->new_frame = video_frame;
+ card->new_frame = RefCountedFrame(video_frame);
card->new_data_ready_fence = fence;
card->new_data_ready_changed.notify_all();
}
- // Video frame will be released later.
+ // Video frame will be released when last user of card->new_frame goes out of scope.
card->usb->get_audio_frame_allocator()->release_frame(audio_frame);
}
while (!should_quit) {
++frame;
+#if 0
//int width0 = lrintf(848 * (1.0 + 0.2 * sin(frame * 0.02)));
int width0 = 848;
int height0 = lrintf(width0 * 9.0 / 16.0);
place_rectangle(resample_effect, padding_effect, left0, top0, right0, bottom0);
place_rectangle(resample2_effect, padding2_effect, left1, top1, right1, bottom1);
+#endif
CaptureCard card_copy[NUM_CARDS];
}
}
- vector<FrameAllocator::Frame> input_frames_to_release;
-
for (int card_index = 0; card_index < NUM_CARDS; ++card_index) {
CaptureCard *card = &card_copy[card_index];
if (!card->new_data_ready)
continue;
- // Now we're done with the previous frame, so we can definitely
- // release it when this is done rendering. (Actually, we could do
- // it one frame earlier, but before we have a new one, there's no
- // knowing when the current one is released.)
- if (bmusb_current_rendering_frame[card_index].owner != nullptr) {
- input_frames_to_release.push_back(bmusb_current_rendering_frame[card_index]);
- }
+ assert(card->new_frame != nullptr);
bmusb_current_rendering_frame[card_index] = card->new_frame;
check_error();
check_error();
glDeleteSync(card->new_data_ready_fence);
check_error();
- GLint input_tex_pbo = (GLint)(intptr_t)card->new_frame.userdata;
- input[card_index]->set_pixel_data(0, (unsigned char *)BUFFER_OFFSET((1280 * 750 * 2 + 44) / 2 + 1280 * 25 + 22), input_tex_pbo);
- input[card_index]->set_pixel_data(1, (unsigned char *)BUFFER_OFFSET(1280 * 25 + 22), input_tex_pbo);
-
- if (NUM_CARDS == 1) {
- // Set to the other one, too.
- input[1]->set_pixel_data(0, (unsigned char *)BUFFER_OFFSET((1280 * 750 * 2 + 44) / 2 + 1280 * 25 + 22), input_tex_pbo);
- input[1]->set_pixel_data(1, (unsigned char *)BUFFER_OFFSET(1280 * 25 + 22), input_tex_pbo);
- }
+ const PBOFrameAllocator::Userdata *userdata = (const PBOFrameAllocator::Userdata *)card->new_frame->userdata;
+ theme->set_input_textures(card_index, userdata->tex_y, userdata->tex_cbcr);
}
+ // Get the main chain from the theme, and set its state immediately.
+ pair<EffectChain *, function<void()>> theme_main_chain = theme->get_chain(0, frame / 60.0f, WIDTH, HEIGHT);
+ EffectChain *chain = theme_main_chain.first;
+ theme_main_chain.second();
+
GLuint y_tex, cbcr_tex;
bool got_frame = h264_encoder->begin_frame(&y_tex, &cbcr_tex);
assert(got_frame);
- // Render chain.
+ // Render main chain.
GLuint cbcr_full_tex = resource_pool->create_2d_texture(GL_RG8, WIDTH, HEIGHT);
- GLuint rgba_tex = resource_pool->create_2d_texture(GL_RGBA8, WIDTH, HEIGHT);
+ GLuint rgba_tex = resource_pool->create_2d_texture(GL_RGB565, WIDTH, HEIGHT); // Saves texture bandwidth, although dithering gets messed up.
GLuint fbo = resource_pool->create_fbo(y_tex, cbcr_full_tex, rgba_tex);
chain->render_to_fbo(fbo, WIDTH, HEIGHT);
resource_pool->release_fbo(fbo);
subsample_chroma(cbcr_full_tex, cbcr_tex);
resource_pool->release_2d_texture(cbcr_full_tex);
+ // Set the right state for rgba_tex.
+ glBindFramebuffer(GL_FRAMEBUFFER, 0);
+ glBindTexture(GL_TEXTURE_2D, rgba_tex);
+ glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR);
+ glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
+ glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
+
RefCountedGLsync fence(GL_SYNC_GPU_COMMANDS_COMPLETE, /*flags=*/0);
check_error();
- h264_encoder->end_frame(fence, input_frames_to_release);
- // Store this frame for display. Remove the ready frame if any
- // (it was seemingly never used).
- {
- std::unique_lock<std::mutex> lock(display_frame_mutex);
- if (has_ready_display_frame) {
- resource_pool->release_2d_texture(ready_display_frame.texnum);
- ready_display_frame.ready_fence.reset();
- }
- ready_display_frame.texnum = rgba_tex;
- ready_display_frame.ready_fence = fence;
- has_ready_display_frame = true;
+ // Make sure the H.264 gets a reference to all the
+ // input frames needed, so that they are not released back
+ // until the rendering is done.
+ vector<RefCountedFrame> input_frames;
+ for (int card_index = 0; card_index < NUM_CARDS; ++card_index) {
+ input_frames.push_back(bmusb_current_rendering_frame[card_index]);
}
-
- if (has_new_frame_ready_callback) {
- new_frame_ready_callback();
+ h264_encoder->end_frame(fence, input_frames);
+
+ // The live frame just shows the RGBA texture we just rendered.
+ // It owns rgba_tex now.
+ DisplayFrame live_frame;
+ live_frame.chain = display_chain.get();
+ live_frame.setup_chain = [this, rgba_tex]{
+ display_input->set_texture_num(rgba_tex);
+ };
+ live_frame.ready_fence = fence;
+ live_frame.input_frames = {};
+ live_frame.temp_textures = { rgba_tex };
+ output_channel[OUTPUT_LIVE].output_frame(live_frame);
+
+ // Set up preview and any additional channels.
+ for (unsigned i = 1; i < theme->get_num_channels() + 2; ++i) {
+ DisplayFrame display_frame;
+ pair<EffectChain *, function<void()>> chain = theme->get_chain(i, frame / 60.0f, WIDTH, HEIGHT); // FIXME: dimensions
+ display_frame.chain = chain.first;
+ display_frame.setup_chain = chain.second;
+ display_frame.ready_fence = fence;
+ display_frame.input_frames = { bmusb_current_rendering_frame[0], bmusb_current_rendering_frame[1] }; // FIXME: possible to do better?
+ display_frame.temp_textures = {};
+ output_channel[i].output_frame(display_frame);
}
clock_gettime(CLOCK_MONOTONIC, &now);
glDeleteVertexArrays(1, &vao);
}
-bool Mixer::get_display_frame(DisplayFrame *frame)
+void Mixer::release_display_frame(DisplayFrame *frame)
{
- std::unique_lock<std::mutex> lock(display_frame_mutex);
- if (!has_current_display_frame && !has_ready_display_frame) {
- return false;
+ for (GLuint texnum : frame->temp_textures) {
+ resource_pool->release_2d_texture(texnum);
}
-
- if (has_current_display_frame && has_ready_display_frame) {
- // We have a new ready frame. Toss the current one.
- resource_pool->release_2d_texture(current_display_frame.texnum);
- current_display_frame.ready_fence.reset();
- has_current_display_frame = false;
- }
- if (has_ready_display_frame) {
- assert(!has_current_display_frame);
- current_display_frame = ready_display_frame;
- ready_display_frame.ready_fence.reset(); // Drop the refcount.
- has_current_display_frame = true;
- has_ready_display_frame = false;
- }
-
- *frame = current_display_frame;
- return true;
-}
-
-void Mixer::set_frame_ready_fallback(new_frame_ready_callback_t callback)
-{
- new_frame_ready_callback = callback;
- has_new_frame_ready_callback = true;
+ frame->temp_textures.clear();
+ frame->ready_fence.reset();
+ frame->input_frames.clear();
}
void Mixer::start()
mixer_thread.join();
}
-void Mixer::cut(Source source)
+void Mixer::transition_clicked(int transition_num, float t)
+{
+ theme->transition_clicked(transition_num, t);
+}
+
+void Mixer::OutputChannel::output_frame(DisplayFrame frame)
+{
+ // Store this frame for display. Remove the ready frame if any
+ // (it was seemingly never used).
+ {
+ std::unique_lock<std::mutex> lock(frame_mutex);
+ if (has_ready_frame) {
+ parent->release_display_frame(&ready_frame);
+ }
+ ready_frame = frame;
+ has_ready_frame = true;
+ }
+
+ if (has_new_frame_ready_callback) {
+ new_frame_ready_callback();
+ }
+}
+
+bool Mixer::OutputChannel::get_display_frame(DisplayFrame *frame)
{
- current_source = source;
+ std::unique_lock<std::mutex> lock(frame_mutex);
+ if (!has_current_frame && !has_ready_frame) {
+ return false;
+ }
+
+ if (has_current_frame && has_ready_frame) {
+ // We have a new ready frame. Toss the current one.
+ parent->release_display_frame(¤t_frame);
+ has_current_frame = false;
+ }
+ if (has_ready_frame) {
+ assert(!has_current_frame);
+ current_frame = ready_frame;
+ ready_frame.ready_fence.reset(); // Drop the refcount.
+ ready_frame.input_frames.clear(); // Drop the refcounts.
+ has_current_frame = true;
+ has_ready_frame = false;
+ }
+
+ *frame = current_frame;
+ return true;
+}
+
+void Mixer::OutputChannel::set_frame_ready_callback(Mixer::new_frame_ready_callback_t callback)
+{
+ new_frame_ready_callback = callback;
+ has_new_frame_ready_callback = true;
}