#ifndef _VIDEO_STREAM_H
#define _VIDEO_STREAM_H 1
-#include <stdint.h>
#include <epoxy/gl.h>
+#include <stdint.h>
extern "C" {
#include <libavformat/avio.h>
}
+#include "jpeg_frame_view.h"
+#include "ref_counted_gl_sync.h"
+#include "queue_spot_holder.h"
+
+#include <chrono>
#include <condition_variable>
#include <deque>
-#include <mutex>
-#include <string>
-#include <thread>
-
+#include <functional>
#include <movit/effect_chain.h>
#include <movit/mix_effect.h>
#include <movit/ycbcr_input.h>
-
-#include "ref_counted_gl_sync.h"
+#include <mutex>
+#include <string>
+#include <thread>
class ChromaSubsampler;
class DISComputeFlow;
~VideoStream();
void start();
void stop();
-
- void schedule_original_frame(int64_t output_pts, unsigned stream_idx, int64_t input_pts);
- void schedule_faded_frame(int64_t output_pts, unsigned stream_idx, int64_t input_pts, int secondary_stream_idx, int64_t secondary_input_pts, float fade_alpha);
- void schedule_interpolated_frame(int64_t output_pts, unsigned stream_idx, int64_t input_first_pts, int64_t input_second_pts, float alpha, int secondary_stream_idx = -1, int64_t secondary_inputs_pts = -1, float fade_alpha = 0.0f); // -1 = no secondary frame.
+ void clear_queue();
+
+ // “display_func” is called after the frame has been calculated (if needed)
+ // and has gone out to the stream.
+ void schedule_original_frame(std::chrono::steady_clock::time_point,
+ int64_t output_pts, std::function<void()> &&display_func,
+ QueueSpotHolder &&queue_spot_holder,
+ unsigned stream_idx, int64_t input_pts);
+ void schedule_faded_frame(std::chrono::steady_clock::time_point, int64_t output_pts,
+ std::function<void()> &&display_func,
+ QueueSpotHolder &&queue_spot_holder,
+ unsigned stream_idx, int64_t input_pts, int secondary_stream_idx,
+ int64_t secondary_input_pts, float fade_alpha);
+ void schedule_interpolated_frame(std::chrono::steady_clock::time_point, int64_t output_pts,
+ std::function<void(std::shared_ptr<Frame>)> &&display_func,
+ QueueSpotHolder &&queue_spot_holder,
+ unsigned stream_idx, int64_t input_first_pts, int64_t input_second_pts,
+ float alpha, int secondary_stream_idx = -1, int64_t secondary_inputs_pts = -1,
+ float fade_alpha = 0.0f); // -1 = no secondary frame.
+ void schedule_refresh_frame(std::chrono::steady_clock::time_point, int64_t output_pts,
+ std::function<void()> &&display_func,
+ QueueSpotHolder &&queue_spot_holder);
private:
-
void encode_thread_func();
std::thread encode_thread;
// Allocated at the very start; if we're empty, we start dropping frames
// (so that we don't build up an infinite interpolation backlog).
struct InterpolatedFrameResources {
+ VideoStream *owner; // Used only for IFRReleaser, below.
+
GLuint input_tex; // Layered (contains both input frames), Y'CbCr.
GLuint gray_tex; // Same, but Y only.
GLuint input_fbos[2]; // For rendering to the two layers of input_tex.
GLuint pbo; // For reading the data back.
void *pbo_contents; // Persistently mapped.
};
- std::deque<InterpolatedFrameResources> interpolate_resources; // Under <queue_lock>.
- static constexpr size_t num_interpolate_slots = 10;
+ std::mutex queue_lock;
+ std::deque<std::unique_ptr<InterpolatedFrameResources>> interpolate_resources; // Under <queue_lock>.
+ static constexpr size_t num_interpolate_slots = 15; // Should be larger than Player::max_queued_frames, or we risk mass-dropping frames.
+
+ struct IFRReleaser {
+ void operator() (InterpolatedFrameResources *ifr) const
+ {
+ if (ifr != nullptr) {
+ std::unique_lock<std::mutex> lock(ifr->owner->queue_lock);
+ ifr->owner->interpolate_resources.emplace_back(ifr);
+ }
+ }
+ };
+ using BorrowedInterpolatedFrameResources = std::unique_ptr<InterpolatedFrameResources, IFRReleaser>;
struct QueuedFrame {
+ std::chrono::steady_clock::time_point local_pts;
+
int64_t output_pts;
- enum Type { ORIGINAL, FADED, INTERPOLATED, FADED_INTERPOLATED } type;
+ enum Type { ORIGINAL, FADED, INTERPOLATED, FADED_INTERPOLATED, REFRESH } type;
unsigned stream_idx;
- int64_t input_first_pts; // The only pts for original frames.
+ int64_t input_first_pts; // The only pts for original frames.
// For fades only (including fades against interpolated frames).
int secondary_stream_idx = -1;
// For interpolated frames only.
int64_t input_second_pts;
float alpha;
- InterpolatedFrameResources resources;
+ BorrowedInterpolatedFrameResources resources;
RefCountedGLsync fence; // Set when the interpolated image is read back to the CPU.
GLuint flow_tex, output_tex, cbcr_tex; // Released in the receiving thread; not really used for anything else.
+ JPEGID id;
+
+ std::function<void()> display_func; // Called when the image is done decoding.
+ std::function<void(std::shared_ptr<Frame>)> display_decoded_func; // Same, except for INTERPOLATED and FADED_INTERPOLATED.
+
+ QueueSpotHolder queue_spot_holder;
};
std::deque<QueuedFrame> frame_queue; // Under <queue_lock>.
- std::mutex queue_lock;
- std::condition_variable queue_nonempty;
+ std::condition_variable queue_changed;
std::unique_ptr<Mux> stream_mux; // To HTTP.
std::string stream_mux_header;
std::unique_ptr<DISComputeFlow> compute_flow;
std::unique_ptr<Interpolate> interpolate, interpolate_no_split;
std::unique_ptr<ChromaSubsampler> chroma_subsampler;
+
+ std::vector<uint8_t> last_frame;
};
#endif // !defined(_VIDEO_STREAM_H)