-#include <algorithm>
-#include <chrono>
-#include <condition_variable>
-#include <mutex>
-#include <thread>
-#include <vector>
-
-#include <stdio.h>
-
-extern "C" {
-#include <libavformat/avformat.h>
-#include <libavformat/avio.h>
-}
+#include "player.h"
#include "clip_list.h"
+#include "context.h"
#include "defs.h"
#include "ffmpeg_raii.h"
#include "httpd.h"
#include "jpeg_frame_view.h"
#include "mux.h"
-#include "player.h"
+#include "timebase.h"
+#include "video_stream.h"
+
+#include <algorithm>
+#include <chrono>
+#include <condition_variable>
+#include <movit/util.h>
+#include <mutex>
+#include <stdio.h>
+#include <thread>
+#include <vector>
using namespace std;
using namespace std::chrono;
extern vector<int64_t> frames[MAX_STREAMS];
extern HTTPD *global_httpd;
-namespace {
-
-string read_file(const string &filename)
+void Player::thread_func(bool also_output_to_stream)
{
- FILE *fp = fopen(filename.c_str(), "rb");
- if (fp == nullptr) {
- perror(filename.c_str());
- return "";
+ pthread_setname_np(pthread_self(), "Player");
+
+ QSurface *surface = create_surface();
+ QOpenGLContext *context = create_context(surface);
+ if (!make_current(context, surface)) {
+ printf("oops\n");
+ exit(1);
}
- fseek(fp, 0, SEEK_END);
- long len = ftell(fp);
- rewind(fp);
+ check_error();
- string ret;
- ret.resize(len);
- fread(&ret[0], len, 1, fp);
- fclose(fp);
- return ret;
-}
+ // Create the VideoStream object, now that we have an OpenGL context.
+ if (also_output_to_stream) {
+ video_stream.reset(new VideoStream);
+ video_stream->start();
+ }
-} // namespace
+ check_error();
+
+ constexpr double output_framerate = 60000.0 / 1001.0; // FIXME: make configurable
+ int64_t pts = 0;
+ Clip next_clip;
+ bool got_next_clip = false;
+ double next_clip_fade_time = -1.0;
-void Player::thread_func()
-{
for ( ;; ) {
+wait_for_clip:
+ bool clip_ready;
+ steady_clock::time_point before_sleep = steady_clock::now();
+
// Wait until we're supposed to play something.
{
unique_lock<mutex> lock(queue_state_mu);
- new_clip_changed.wait(lock, [this]{
+ clip_ready = new_clip_changed.wait_for(lock, milliseconds(100), [this] {
return new_clip_ready && current_clip.pts_in != -1;
});
new_clip_ready = false;
playing = true;
}
+ steady_clock::duration time_slept = steady_clock::now() - before_sleep;
+ pts += duration_cast<duration<size_t, TimebaseRatio>>(time_slept).count();
+
+ if (!clip_ready) {
+ if (video_stream != nullptr) {
+ video_stream->schedule_refresh_frame(pts, /*display_func=*/nullptr);
+ }
+ continue;
+ }
+
Clip clip;
unsigned stream_idx;
{
stream_idx = current_stream_idx;
}
steady_clock::time_point origin = steady_clock::now();
- int64_t pts_origin = clip.pts_in;
+ int64_t in_pts_origin = clip.pts_in;
+got_clip:
+ int64_t out_pts_origin = pts;
- int64_t next_pts = pts_origin - 1; // Make sure we play the frame at clip.pts_in if it exists.
+ // Start playing exactly at a frame.
+ // TODO: Snap secondary (fade-to) clips in the same fashion
+ // so that we don't get jank here).
+ {
+ lock_guard<mutex> lock(frame_mu);
+
+ // Find the first frame such that frame.pts <= in_pts.
+ auto it = lower_bound(frames[stream_idx].begin(),
+ frames[stream_idx].end(),
+ in_pts_origin);
+ if (it != frames[stream_idx].end()) {
+ in_pts_origin = *it;
+ }
+ }
- bool aborted = false;
- for ( ;; ) {
- // Find the next frame.
- {
- lock_guard<mutex> lock(frame_mu);
- auto it = upper_bound(frames[stream_idx].begin(),
- frames[stream_idx].end(),
- next_pts);
- if (it == frames[stream_idx].end() || *it >= clip.pts_out) {
- break;
+ // TODO: Lock to a rational multiple of the frame rate if possible.
+ double speed = 0.5;
+
+ int64_t in_pts_start_next_clip = -1;
+ for (int frameno = 0; ; ++frameno) { // Ends when the clip ends.
+ double out_pts = out_pts_origin + TIMEBASE * frameno / output_framerate;
+ steady_clock::time_point next_frame_start =
+ origin + microseconds(lrint((out_pts - out_pts_origin) * 1e6 / TIMEBASE));
+ int64_t in_pts = lrint(in_pts_origin + TIMEBASE * frameno * speed / output_framerate);
+ pts = lrint(out_pts);
+
+ if (in_pts >= clip.pts_out) {
+ break;
+ }
+
+ steady_clock::duration time_behind = steady_clock::now() - next_frame_start;
+ if (time_behind >= milliseconds(200)) {
+ fprintf(stderr, "WARNING: %ld ms behind, dropping a frame (no matter the type).\n",
+ lrint(1e3 * duration<double>(time_behind).count()));
+ continue;
+ }
+
+ double time_left_this_clip = double(clip.pts_out - in_pts) / TIMEBASE / speed;
+ if (!got_next_clip && next_clip_callback != nullptr && time_left_this_clip <= clip.fade_time_seconds) {
+ // Find the next clip so that we can begin a fade.
+ next_clip = next_clip_callback();
+ if (next_clip.pts_in != -1) {
+ got_next_clip = true;
+
+ double duration_next_clip = (next_clip.pts_out - next_clip.pts_in) / TIMEBASE / speed;
+ next_clip_fade_time = std::min(time_left_this_clip, duration_next_clip);
+ in_pts_start_next_clip = next_clip.pts_in + lrint(next_clip_fade_time * TIMEBASE * speed);
}
- next_pts = *it;
}
- // FIXME: assumes a given timebase.
- double speed = 0.5;
- steady_clock::time_point next_frame_start =
- origin + microseconds((next_pts - pts_origin) * int(1000000 / speed) / 12800);
+ int primary_stream_idx = stream_idx;
+ int secondary_stream_idx = -1;
+ int64_t secondary_pts = -1;
+ int64_t in_pts_secondary = -1;
+ float fade_alpha = 0.0f;
+ if (got_next_clip) {
+ secondary_stream_idx = next_clip.stream_idx;
+ in_pts_secondary = lrint(next_clip.pts_in + (next_clip_fade_time - time_left_this_clip) * TIMEBASE * speed);
+ fade_alpha = 1.0f - time_left_this_clip / next_clip_fade_time;
+
+ // If more than half-way through the fade, interpolate the next clip
+ // instead of the current one, since it's more visible.
+ if (fade_alpha >= 0.5f) {
+ swap(primary_stream_idx, secondary_stream_idx);
+ swap(in_pts, in_pts_secondary);
+ fade_alpha = 1.0f - fade_alpha;
+ }
+
+ int64_t in_pts_lower, in_pts_upper;
+ bool ok = find_surrounding_frames(in_pts_secondary, secondary_stream_idx, &in_pts_lower, &in_pts_upper);
+ if (ok) {
+ secondary_pts = in_pts_lower;
+ } else {
+ secondary_stream_idx = -1;
+ }
+ }
+
+ if (progress_callback != nullptr) {
+ // NOTE: None of this will take into account any snapping done below.
+ double played_this_clip = double(in_pts - clip.pts_in) / TIMEBASE / speed;
+ double total_length = double(clip.pts_out - clip.pts_in) / TIMEBASE / speed;
+ progress_callback(played_this_clip, total_length);
+ }
+
+ int64_t in_pts_lower, in_pts_upper;
+ bool ok = find_surrounding_frames(in_pts, primary_stream_idx, &in_pts_lower, &in_pts_upper);
+ if (!ok) {
+ break;
+ }
// Sleep until the next frame start, or until there's a new clip we're supposed to play.
{
new_clip_changed.wait_until(lock, next_frame_start, [this]{
return new_clip_ready || override_stream_idx != -1;
});
- if (new_clip_ready) break;
+ if (new_clip_ready) {
+ goto wait_for_clip;
+ }
if (override_stream_idx != -1) {
stream_idx = override_stream_idx;
override_stream_idx = -1;
}
}
- destination->setFrame(stream_idx, next_pts);
+ if (in_pts_lower == in_pts_upper) {
+ auto display_func = [this, primary_stream_idx, in_pts_lower, secondary_stream_idx, secondary_pts, fade_alpha]{
+ destination->setFrame(primary_stream_idx, in_pts_lower, /*interpolated=*/false, secondary_stream_idx, secondary_pts, fade_alpha);
+ };
+ if (video_stream == nullptr) {
+ display_func();
+ } else {
+ if (secondary_stream_idx == -1) {
+ video_stream->schedule_original_frame(pts, display_func, primary_stream_idx, in_pts_lower);
+ } else {
+ video_stream->schedule_faded_frame(pts, display_func, primary_stream_idx, in_pts_lower, secondary_stream_idx, secondary_pts, fade_alpha);
+ }
+ }
+ continue;
+ }
+
+ // Snap to input frame: If we can do so with less than 1% jitter
+ // (ie., move less than 1% of an _output_ frame), do so.
+ // TODO: Snap secondary (fade-to) clips in the same fashion.
+ bool snapped = false;
+ for (int64_t snap_pts : { in_pts_lower, in_pts_upper }) {
+ double snap_pts_as_frameno = (snap_pts - in_pts_origin) * output_framerate / TIMEBASE / speed;
+ if (fabs(snap_pts_as_frameno - frameno) < 0.01) {
+ auto display_func = [this, primary_stream_idx, snap_pts, secondary_stream_idx, secondary_pts, fade_alpha]{
+ destination->setFrame(primary_stream_idx, snap_pts, /*interpolated=*/false, secondary_stream_idx, secondary_pts, fade_alpha);
+ };
+ if (video_stream == nullptr) {
+ display_func();
+ } else {
+ if (secondary_stream_idx == -1) {
+ video_stream->schedule_original_frame(pts, display_func, primary_stream_idx, snap_pts);
+ } else {
+ video_stream->schedule_faded_frame(pts, display_func, primary_stream_idx, snap_pts, secondary_stream_idx, secondary_pts, fade_alpha);
+ }
+ }
+ in_pts_origin += snap_pts - in_pts;
+ snapped = true;
+ break;
+ }
+ }
+ if (snapped) {
+ continue;
+ }
+
+ if (time_behind >= milliseconds(100)) {
+ fprintf(stderr, "WARNING: %ld ms behind, dropping an interpolated frame.\n",
+ lrint(1e3 * duration<double>(time_behind).count()));
+ continue;
+ }
+
+ double alpha = double(in_pts - in_pts_lower) / (in_pts_upper - in_pts_lower);
+
+ if (video_stream == nullptr) {
+ // Previews don't do any interpolation.
+ assert(secondary_stream_idx == -1);
+ destination->setFrame(primary_stream_idx, in_pts_lower, /*interpolated=*/false);
+ } else {
+ // Calculate the interpolated frame. When it's done, the destination
+ // will be unblocked.
+ auto display_func = [this, primary_stream_idx, pts, secondary_stream_idx, secondary_pts, fade_alpha]{
+ destination->setFrame(primary_stream_idx, pts, /*interpolated=*/true, secondary_stream_idx, secondary_pts, fade_alpha);
+ };
+ video_stream->schedule_interpolated_frame(pts, display_func, primary_stream_idx, in_pts_lower, in_pts_upper, alpha, secondary_stream_idx, secondary_pts, fade_alpha);
+ }
+ }
+
+ // The clip ended.
+
+ // Last-ditch effort to get the next clip (if e.g. the fade time was zero seconds).
+ if (!got_next_clip && next_clip_callback != nullptr) {
+ next_clip = next_clip_callback();
+ if (next_clip.pts_in != -1) {
+ got_next_clip = true;
+ in_pts_start_next_clip = next_clip.pts_in;
+ }
+ }
+
+ // Switch to next clip if we got it.
+ if (got_next_clip) {
+ clip = next_clip;
+ stream_idx = next_clip.stream_idx; // Override is used for previews only, and next_clip is used for live ony.
+ if (done_callback != nullptr) {
+ done_callback();
+ }
+ got_next_clip = false;
- // Send the frame to the stream.
- // FIXME: Vaguely less crazy pts, perhaps.
- double pts_float = fmod(duration<double>(next_frame_start.time_since_epoch()).count(), 86400.0f);
- int64_t pts = lrint(pts_float * TIMEBASE);
- string jpeg = read_file(filename_for_frame(stream_idx, next_pts));
- AVPacket pkt;
- av_init_packet(&pkt);
- pkt.stream_index = 0;
- pkt.data = (uint8_t *)jpeg.data();
- pkt.size = jpeg.size();
- stream_mux->add_packet(pkt, pts, pts);
+ // Start the next clip from the point where the fade went out.
+ origin = steady_clock::now();
+ in_pts_origin = in_pts_start_next_clip;
+ goto got_clip;
}
{
unique_lock<mutex> lock(queue_state_mu);
playing = false;
}
- if (done_callback != nullptr && !aborted) {
+ if (done_callback != nullptr) {
done_callback();
}
}
}
-Player::Player(JPEGFrameView *destination)
+// Find the frame immediately before and after this point.
+bool Player::find_surrounding_frames(int64_t pts, int stream_idx, int64_t *pts_lower, int64_t *pts_upper)
+{
+ lock_guard<mutex> lock(frame_mu);
+
+ // Find the first frame such that frame.pts >= pts.
+ auto it = lower_bound(frames[stream_idx].begin(),
+ frames[stream_idx].end(),
+ pts);
+ if (it == frames[stream_idx].end()) {
+ return false;
+ }
+ *pts_upper = *it;
+
+ // Find the last frame such that in_pts <= frame.pts (if any).
+ if (it == frames[stream_idx].begin()) {
+ *pts_lower = *it;
+ } else {
+ *pts_lower = *(it - 1);
+ }
+ assert(pts >= *pts_lower);
+ assert(pts <= *pts_upper);
+ return true;
+}
+
+Player::Player(JPEGFrameView *destination, bool also_output_to_stream)
: destination(destination)
{
- open_output_stream();
- thread(&Player::thread_func, this).detach();
+ thread(&Player::thread_func, this, also_output_to_stream).detach();
}
void Player::play_clip(const Clip &clip, unsigned stream_idx)
void Player::override_angle(unsigned stream_idx)
{
- // Corner case: If a new clip is waiting to be played, change its stream and then we're done.
+ // Corner case: If a new clip is waiting to be played, change its stream and then we're done.
{
unique_lock<mutex> lock(queue_state_mu);
if (new_clip_ready) {
}
pts_out = current_clip.pts_out;
}
-
+
lock_guard<mutex> lock(frame_mu);
auto it = upper_bound(frames[stream_idx].begin(), frames[stream_idx].end(), pts_out);
if (it == frames[stream_idx].end()) {
return;
}
- destination->setFrame(stream_idx, *it);
-}
-
-void Player::open_output_stream()
-{
- AVFormatContext *avctx = avformat_alloc_context();
- avctx->oformat = av_guess_format("nut", nullptr, nullptr);
-
- uint8_t *buf = (uint8_t *)av_malloc(MUX_BUFFER_SIZE);
- avctx->pb = avio_alloc_context(buf, MUX_BUFFER_SIZE, 1, this, nullptr, nullptr, nullptr);
- avctx->pb->write_data_type = &Player::write_packet2_thunk;
- avctx->pb->ignore_boundary_point = 1;
-
- Mux::Codec video_codec = Mux::CODEC_MJPEG;
-
- avctx->flags = AVFMT_FLAG_CUSTOM_IO;
-
- string video_extradata;
-
- constexpr int width = 1280, height = 720; // Doesn't matter for MJPEG.
- stream_mux.reset(new Mux(avctx, width, height, video_codec, video_extradata, /*audio_codec_parameters=*/nullptr, COARSE_TIMEBASE,
- /*write_callback=*/nullptr, Mux::WRITE_FOREGROUND, {}));
-}
-
-int Player::write_packet2_thunk(void *opaque, uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time)
-{
- Player *player = (Player *)opaque;
- return player->write_packet2(buf, buf_size, type, time);
-}
-
-int Player::write_packet2(uint8_t *buf, int buf_size, AVIODataMarkerType type, int64_t time)
-{
- if (type == AVIO_DATA_MARKER_SYNC_POINT || type == AVIO_DATA_MARKER_BOUNDARY_POINT) {
- seen_sync_markers = true;
- } else if (type == AVIO_DATA_MARKER_UNKNOWN && !seen_sync_markers) {
- // We don't know if this is a keyframe or not (the muxer could
- // avoid marking it), so we just have to make the best of it.
- type = AVIO_DATA_MARKER_SYNC_POINT;
- }
-
- if (type == AVIO_DATA_MARKER_HEADER) {
- stream_mux_header.append((char *)buf, buf_size);
- global_httpd->set_header(stream_mux_header);
- } else {
- global_httpd->add_data((char *)buf, buf_size, type == AVIO_DATA_MARKER_SYNC_POINT, time, AVRational{ AV_TIME_BASE, 1 });
- }
- return buf_size;
+ destination->setFrame(stream_idx, *it, /*interpolated=*/false);
}