X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=mixer.cpp;h=0f8cf903d326d87fc41b00ec65b07afa6ce1c8b7;hb=af35bb637146622dc43416a5f2e801b1efaf6953;hp=4a758c0defc7c567caad281c5dc0097e200ea297;hpb=02083dca89292d6ba95642cf8957f1cee8826313;p=nageru diff --git a/mixer.cpp b/mixer.cpp index 4a758c0..0f8cf90 100644 --- a/mixer.cpp +++ b/mixer.cpp @@ -1,5 +1,6 @@ #define WIDTH 1280 #define HEIGHT 720 +#define EXTRAHEIGHT 30 #undef Success @@ -70,12 +71,19 @@ void convert_fixed24_to_fp32(float *dst, size_t out_channels, const uint8_t *src } // namespace Mixer::Mixer(const QSurfaceFormat &format) - : mixer_surface(create_surface(format)), + : httpd("test.ts", WIDTH, HEIGHT), + mixer_surface(create_surface(format)), h264_encoder_surface(create_surface(format)) { + httpd.start(9095); + CHECK(init_movit(MOVIT_SHADER_DIR, MOVIT_DEBUG_OFF)); check_error(); + // Since we allow non-bouncing 4:2:2 YCbCrInputs, effective subpixel precision + // will be halved when sampling them, and we need to compensate here. + movit_texel_subpixel_precision /= 2.0; + resource_pool.reset(new ResourcePool); theme.reset(new Theme("theme.lua", resource_pool.get())); output_channel[OUTPUT_LIVE].parent = this; @@ -96,14 +104,14 @@ Mixer::Mixer(const QSurfaceFormat &format) display_chain->set_dither_bits(0); // Don't bother. display_chain->finalize(); - h264_encoder.reset(new H264Encoder(h264_encoder_surface, WIDTH, HEIGHT, "test.mp4")); + h264_encoder.reset(new H264Encoder(h264_encoder_surface, WIDTH, HEIGHT, &httpd)); for (int card_index = 0; card_index < NUM_CARDS; ++card_index) { printf("Configuring card %d...\n", card_index); CaptureCard *card = &cards[card_index]; - card->usb = new BMUSBCapture(0x1edb, card_index == 0 ? 0xbd3b : 0xbd4f); + card->usb = new BMUSBCapture(card_index); card->usb->set_frame_callback(bind(&Mixer::bm_frame, this, card_index, _1, _2, _3, _4, _5, _6, _7)); - card->frame_allocator.reset(new PBOFrameAllocator(1280 * 750 * 2 + 44, 1280, 720)); + card->frame_allocator.reset(new PBOFrameAllocator(WIDTH * (HEIGHT+EXTRAHEIGHT) * 2 + 44, WIDTH, HEIGHT)); card->usb->set_video_frame_allocator(card->frame_allocator.get()); card->surface = create_surface(format); card->usb->set_dequeue_thread_callbacks( @@ -114,7 +122,6 @@ Mixer::Mixer(const QSurfaceFormat &format) printf("failed to create bmusb context\n"); exit(1); } - printf("inited!\n"); }, [this]{ resource_pool->clean_context(); @@ -143,6 +150,9 @@ Mixer::Mixer(const QSurfaceFormat &format) " gl_FragColor = texture2D(cbcr_tex, tc0); \n" "} \n"; cbcr_program_num = resource_pool->compile_glsl_program(cbcr_vert_shader, cbcr_frag_shader); + + r128.init(2, 48000); + r128.integr_start(); } Mixer::~Mixer() @@ -172,6 +182,30 @@ int unwrap_timecode(uint16_t current_wrapped, int last) } } +float find_peak(const vector &samples) +{ + float m = fabs(samples[0]); + for (size_t i = 1; i < samples.size(); ++i) { + m = std::max(m, fabs(samples[i])); + } + return m; +} + +void deinterleave_samples(const vector &in, vector *out_l, vector *out_r) +{ + size_t num_samples = in.size() / 2; + out_l->resize(num_samples); + out_r->resize(num_samples); + + const float *inptr = in.data(); + float *lptr = &(*out_l)[0]; + float *rptr = &(*out_r)[0]; + for (size_t i = 0; i < num_samples; ++i) { + *lptr++ = *inptr++; + *rptr++ = *inptr++; + } +} + } // namespace void Mixer::bm_frame(int card_index, uint16_t timecode, @@ -241,7 +275,7 @@ void Mixer::bm_frame(int card_index, uint16_t timecode, if (card->should_quit) return; } - if (video_frame.len - video_offset != 1280 * 750 * 2) { + if (video_frame.len - video_offset != WIDTH * (HEIGHT+EXTRAHEIGHT) * 2) { if (video_frame.len != 0) { printf("Card %d: Dropping video frame with wrong length (%ld)\n", card_index, video_frame.len - video_offset); @@ -276,11 +310,11 @@ void Mixer::bm_frame(int card_index, uint16_t timecode, // Upload the textures. glBindTexture(GL_TEXTURE_2D, userdata->tex_y); check_error(); - glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 1280, 720, GL_RED, GL_UNSIGNED_BYTE, BUFFER_OFFSET((1280 * 750 * 2 + 44) / 2 + 1280 * 25 + 22)); + glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, WIDTH, HEIGHT, GL_RED, GL_UNSIGNED_BYTE, BUFFER_OFFSET((WIDTH * (HEIGHT+EXTRAHEIGHT) * 2 + 44) / 2 + WIDTH * 25 + 22)); check_error(); glBindTexture(GL_TEXTURE_2D, userdata->tex_cbcr); check_error(); - glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 1280/2, 720, GL_RG, GL_UNSIGNED_BYTE, BUFFER_OFFSET(1280 * 25 + 22)); + glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, WIDTH/2, HEIGHT, GL_RG, GL_UNSIGNED_BYTE, BUFFER_OFFSET(WIDTH * 25 + 22)); check_error(); glBindTexture(GL_TEXTURE_2D, 0); check_error(); @@ -349,6 +383,11 @@ void Mixer::thread_func() } } if (card_index == 0) { + vector left, right; + peak = std::max(peak, find_peak(samples_out)); + deinterleave_samples(samples_out, &left, &right); + float *ptrs[] = { left.data(), right.data() }; + r128.process(left.size(), ptrs); h264_encoder->add_audio(pts_int, move(samples_out)); } } @@ -359,6 +398,16 @@ void Mixer::thread_func() } } + if (audio_level_callback != nullptr) { + double loudness_s = r128.loudness_S(); + double loudness_i = r128.integrated(); + double loudness_range_low = r128.range_min(); + double loudness_range_high = r128.range_max(); + + audio_level_callback(loudness_s, 20.0 * log10(peak), + loudness_i, loudness_range_low, loudness_range_high); + } + // If the first card is reporting a corrupted or otherwise dropped frame, // just increase the pts (skipping over this frame) and don't try to compute anything new. if (card_copy[0].new_frame->len == 0) { @@ -401,6 +450,7 @@ void Mixer::thread_func() GLuint cbcr_full_tex = resource_pool->create_2d_texture(GL_RG8, WIDTH, HEIGHT); GLuint rgba_tex = resource_pool->create_2d_texture(GL_RGB565, WIDTH, HEIGHT); // Saves texture bandwidth, although dithering gets messed up. GLuint fbo = resource_pool->create_fbo(y_tex, cbcr_full_tex, rgba_tex); + check_error(); chain->render_to_fbo(fbo, WIDTH, HEIGHT); resource_pool->release_fbo(fbo); @@ -424,7 +474,8 @@ void Mixer::thread_func() for (int card_index = 0; card_index < NUM_CARDS; ++card_index) { input_frames.push_back(bmusb_current_rendering_frame[card_index]); } - h264_encoder->end_frame(fence, pts_int, input_frames); + const int64_t av_delay = TIMEBASE / 10; // Corresponds to the fixed delay in resampler.h. TODO: Make less hard-coded. + h264_encoder->end_frame(fence, pts_int + av_delay, input_frames); ++frame; pts_int += TIMEBASE / 60;