#include "bmusb/bmusb.h"
#include "alsa_output.h"
#include "ebu_r128_proc.h"
-#include "h264encode.h"
+#include "video_encoder.h"
#include "httpd.h"
#include "pbo_frame_allocator.h"
#include "ref_counted_frame.h"
#include "input_state.h"
#include "correlation_measurer.h"
-class H264Encoder;
+class QuickSyncEncoder;
class QSurface;
namespace movit {
class Effect;
//
// N is reduced as follows: If the queue has had at least one spare frame for
// at least 50 (master) frames (ie., it's been too conservative for a second),
-// we reduce N by 1 and reset the timers.
+// we reduce N by 1 and reset the timers. TODO: Only do this if N ever actually
+// touched the limit.
//
-// Whenever the queue is starved (we needed a frame but there was none), N was
-// obviously too low, so we increment N. We will never set N above 5, though.
+// Whenever the queue is starved (we needed a frame but there was none),
+// and we've been at N since the last starvation, N was obviously too low,
+// so we increment it. We will never set N above 5, though.
class QueueLengthPolicy {
public:
QueueLengthPolicy() {}
this->card_index = card_index;
safe_queue_length = 0;
frames_with_at_least_one = 0;
+ been_at_safe_point_since_last_starvation = false;
}
void update_policy(int queue_length); // Give in -1 for starvation.
unsigned card_index; // For debugging only.
unsigned safe_queue_length = 0; // Called N in the comments.
unsigned frames_with_at_least_one = 0;
+ bool been_at_safe_point_since_last_starvation = false;
};
class Mixer {
return theme->get_channel_name(channel);
}
+ std::string get_channel_color(unsigned channel) const
+ {
+ return theme->get_channel_color(channel);
+ }
+
int get_channel_signal(unsigned channel) const
{
return theme->get_channel_signal(channel);
audio_source_channel = channel;
}
+ unsigned get_master_clock() const
+ {
+ return master_clock_channel;
+ }
+
+ void set_master_clock(unsigned channel)
+ {
+ master_clock_channel = channel;
+ }
+
void set_signal_mapping(int signal, int card)
{
return theme->set_signal_mapping(signal, card);
FrameAllocator::Frame audio_frame, size_t audio_offset, AudioFormat audio_format);
void place_rectangle(movit::Effect *resample_effect, movit::Effect *padding_effect, float x0, float y0, float x1, float y1);
void thread_func();
- void render_one_frame();
+ void schedule_audio_resampling_tasks(unsigned dropped_frames, int num_samples_per_frame, int length_per_frame);
+ void render_one_frame(int64_t duration);
+ void send_audio_level_callback();
void audio_thread_func();
void process_audio_one_frame(int64_t frame_pts_int, int num_samples);
void subsample_chroma(GLuint src_tex, GLuint dst_dst);
std::unique_ptr<movit::ResourcePool> resource_pool;
std::unique_ptr<Theme> theme;
std::atomic<unsigned> audio_source_channel{0};
+ std::atomic<unsigned> master_clock_channel{0};
std::unique_ptr<movit::EffectChain> display_chain;
GLuint cbcr_program_num; // Owned by <resource_pool>.
GLuint cbcr_vbo; // Holds position and texcoord data.
GLuint cbcr_position_attribute_index, cbcr_texcoord_attribute_index;
- std::unique_ptr<H264Encoder> h264_encoder;
+ std::unique_ptr<VideoEncoder> video_encoder;
// Effects part of <display_chain>. Owned by <display_chain>.
movit::FlatInput *display_input;
int64_t length; // In TIMEBASE units.
bool interlaced;
unsigned field; // Which field (0 or 1) of the frame to use. Always 0 for progressive.
- GLsync ready_fence; // Whether frame is ready for rendering.
+ std::function<void()> upload_func; // Needs to be called to actually upload the texture to OpenGL.
unsigned dropped_frames = 0; // Number of dropped frames before this one.
};
std::queue<NewFrame> new_frames;
int64_t next_local_pts = 0; // Beginning of next frame, in TIMEBASE units.
};
CaptureCard cards[MAX_CARDS]; // protected by <bmusb_mutex>
+ void get_one_frame_from_each_card(unsigned master_card_index, CaptureCard::NewFrame new_frames[MAX_CARDS], bool has_new_frame[MAX_CARDS], int num_samples[MAX_CARDS]);
InputState input_state;