X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=video_stream.h;h=2dae6e06f01c7a626556090fa9050daf7e982727;hb=6745f125b516ffca30dc2e7fbc5d396f82149f53;hp=3a8a2f26968ecd9841226d6a406eeb9f350b7bae;hpb=9a5e0639486a26685ea623c7c85e9ae87f8ae656;p=nageru diff --git a/video_stream.h b/video_stream.h index 3a8a2f2..2dae6e0 100644 --- a/video_stream.h +++ b/video_stream.h @@ -1,28 +1,44 @@ #ifndef _VIDEO_STREAM_H #define _VIDEO_STREAM_H 1 -#include #include +#include extern "C" { #include } +#include "jpeg_frame_view.h" +#include "ref_counted_gl_sync.h" + #include #include +#include +#include +#include #include #include #include +class ChromaSubsampler; +class DISComputeFlow; +class Interpolate; class Mux; +class QSurface; +class QSurfaceFormat; +class YCbCrConverter; class VideoStream { public: + VideoStream(); + ~VideoStream(); void start(); void stop(); void schedule_original_frame(int64_t output_pts, unsigned stream_idx, int64_t input_pts); - void schedule_interpolated_frame(int64_t output_pts, unsigned stream_idx, int64_t input_first_pts, int64_t input_second_pts, float alpha); + void schedule_faded_frame(int64_t output_pts, unsigned stream_idx, int64_t input_pts, int secondary_stream_idx, int64_t secondary_input_pts, float fade_alpha); + void schedule_interpolated_frame(int64_t output_pts, unsigned stream_idx, int64_t input_first_pts, int64_t input_second_pts, float alpha, int secondary_stream_idx = -1, int64_t secondary_inputs_pts = -1, float fade_alpha = 0.0f); // -1 = no secondary frame. + void schedule_refresh_frame(int64_t output_pts); private: void encode_thread_func(); @@ -31,17 +47,42 @@ private: static int write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time); int write_packet2(uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time); + // Allocated at the very start; if we're empty, we start dropping frames + // (so that we don't build up an infinite interpolation backlog). + struct InterpolatedFrameResources { + GLuint input_tex; // Layered (contains both input frames), Y'CbCr. + GLuint gray_tex; // Same, but Y only. + GLuint input_fbos[2]; // For rendering to the two layers of input_tex. + + // Destination textures and FBO if there is a fade. + GLuint fade_y_output_tex, fade_cbcr_output_tex; + GLuint fade_fbo; + + GLuint cb_tex, cr_tex; // Subsampled, final output. + + GLuint pbo; // For reading the data back. + void *pbo_contents; // Persistently mapped. + }; + std::deque interpolate_resources; // Under . + static constexpr size_t num_interpolate_slots = 10; + struct QueuedFrame { int64_t output_pts; - enum Type { ORIGINAL, INTERPOLATED } type; + enum Type { ORIGINAL, FADED, INTERPOLATED, FADED_INTERPOLATED, REFRESH } type; unsigned stream_idx; - int64_t input_first_pts; // The only pts for original frames. + int64_t input_first_pts; // The only pts for original frames. + + // For fades only (including fades against interpolated frames). + int secondary_stream_idx = -1; + int64_t secondary_input_pts; // For interpolated frames only. int64_t input_second_pts; float alpha; - GLuint flow_tex; - GLuint fence; // Set when the flow is done computing. + InterpolatedFrameResources resources; + RefCountedGLsync fence; // Set when the interpolated image is read back to the CPU. + GLuint flow_tex, output_tex, cbcr_tex; // Released in the receiving thread; not really used for anything else. + JPEGID id; }; std::deque frame_queue; // Under . std::mutex queue_lock; @@ -50,6 +91,16 @@ private: std::unique_ptr stream_mux; // To HTTP. std::string stream_mux_header; bool seen_sync_markers = false; + + std::unique_ptr ycbcr_converter; + std::unique_ptr ycbcr_semiplanar_converter; + + // Frame interpolation. + std::unique_ptr compute_flow; + std::unique_ptr interpolate, interpolate_no_split; + std::unique_ptr chroma_subsampler; + + std::vector last_frame; }; #endif // !defined(_VIDEO_STREAM_H)