#include "x264_encoder.h"
+#include <assert.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include "defs.h"
#include "flags.h"
#include "mux.h"
+#include "print_latency.h"
#include "timebase.h"
#include "x264_speed_control.h"
}
using namespace std;
+using namespace std::chrono;
namespace {
X264Encoder::X264Encoder(AVOutputFormat *oformat)
: wants_global_headers(oformat->flags & AVFMT_GLOBALHEADER)
{
- frame_pool.reset(new uint8_t[WIDTH * HEIGHT * 2 * X264_QUEUE_LENGTH]);
+ frame_pool.reset(new uint8_t[global_flags.width * global_flags.height * 2 * X264_QUEUE_LENGTH]);
for (unsigned i = 0; i < X264_QUEUE_LENGTH; ++i) {
- free_frames.push(frame_pool.get() + i * (WIDTH * HEIGHT * 2));
+ free_frames.push(frame_pool.get() + i * (global_flags.width * global_flags.height * 2));
}
encoder_thread = thread(&X264Encoder::encoder_thread_func, this);
}
encoder_thread.join();
}
-void X264Encoder::add_frame(int64_t pts, int64_t duration, const uint8_t *data)
+void X264Encoder::add_frame(int64_t pts, int64_t duration, const uint8_t *data, const ReceivedTimestamps &received_ts)
{
QueuedFrame qf;
qf.pts = pts;
qf.duration = duration;
+ qf.received_ts = received_ts;
{
lock_guard<mutex> lock(mu);
free_frames.pop();
}
- memcpy(qf.data, data, WIDTH * HEIGHT * 2);
+ memcpy(qf.data, data, global_flags.width * global_flags.height * 2);
{
lock_guard<mutex> lock(mu);
x264_param_t param;
x264_param_default_preset(¶m, global_flags.x264_preset.c_str(), global_flags.x264_tune.c_str());
- param.i_width = WIDTH;
- param.i_height = HEIGHT;
+ param.i_width = global_flags.width;
+ param.i_height = global_flags.height;
param.i_csp = X264_CSP_NV12;
param.b_vfr_input = 1;
param.i_timebase_num = 1;
// No exit; it's not fatal.
}
init_x264();
+ x264_init_done = true;
bool frames_left;
pic.img.i_csp = X264_CSP_NV12;
pic.img.i_plane = 2;
pic.img.plane[0] = qf.data;
- pic.img.i_stride[0] = WIDTH;
- pic.img.plane[1] = qf.data + WIDTH * HEIGHT;
- pic.img.i_stride[1] = WIDTH / 2 * sizeof(uint16_t);
+ pic.img.i_stride[0] = global_flags.width;
+ pic.img.plane[1] = qf.data + global_flags.width * global_flags.height;
+ pic.img.i_stride[1] = global_flags.width / 2 * sizeof(uint16_t);
pic.opaque = reinterpret_cast<void *>(intptr_t(qf.duration));
input_pic = &pic;
+
+ frames_being_encoded[qf.pts] = qf.received_ts;
}
// See if we have a new bitrate to change to.
speed_control->after_frame();
}
+ if (num_nal == 0) return;
+
+ if (frames_being_encoded.count(pic.i_pts)) {
+ ReceivedTimestamps received_ts = frames_being_encoded[pic.i_pts];
+ frames_being_encoded.erase(pic.i_pts);
+
+ static int frameno = 0;
+ print_latency("Current x264 latency (video inputs → network mux):",
+ received_ts, (pic.i_type == X264_TYPE_B || pic.i_type == X264_TYPE_BREF),
+ &frameno);
+ } else {
+ assert(false);
+ }
+
// We really need one AVPacket for the entire frame, it seems,
// so combine it all.
size_t num_bytes = buffered_sei.size();