-#include "../StdAfx.h"
-
-#include "ffmpeg_consumer.h"
-
-#include "../ffmpeg_error.h"
-#include "../producer/util/util.h"
-#include "../producer/filter/filter.h"
-#include "../producer/filter/audio_filter.h"
-
-#include <common/except.h>
-#include <common/executor.h>
-#include <common/assert.h>
-#include <common/utf.h>
-#include <common/future.h>
-#include <common/diagnostics/graph.h>
-#include <common/env.h>
-#include <common/scope_exit.h>
-#include <common/ptree.h>
-#include <common/param.h>
-#include <common/semaphore.h>
-
-#include <core/consumer/frame_consumer.h>
-#include <core/frame/frame.h>
-#include <core/frame/audio_channel_layout.h>
-#include <core/video_format.h>
-#include <core/monitor/monitor.h>
-#include <core/help/help_repository.h>
-#include <core/help/help_sink.h>
-
-#include <boost/noncopyable.hpp>
-#include <boost/rational.hpp>
-#include <boost/format.hpp>
-#include <boost/algorithm/string/predicate.hpp>
-#include <boost/property_tree/ptree.hpp>
-
-#pragma warning(push)
-#pragma warning(disable: 4244)
-#pragma warning(disable: 4245)
-#include <boost/crc.hpp>
-#pragma warning(pop)
-
-#include <tbb/atomic.h>
-#include <tbb/concurrent_queue.h>
-#include <tbb/parallel_invoke.h>
-#include <tbb/parallel_for.h>
-
-#include <numeric>
-
-#pragma warning(push)
-#pragma warning(disable: 4244)
-
-extern "C"
-{
- #define __STDC_CONSTANT_MACROS
- #define __STDC_LIMIT_MACROS
- #include <libavformat/avformat.h>
- #include <libavcodec/avcodec.h>
- #include <libavutil/avutil.h>
- #include <libavutil/frame.h>
- #include <libavutil/opt.h>
- #include <libavutil/imgutils.h>
- #include <libavutil/parseutils.h>
- #include <libavfilter/avfilter.h>
- #include <libavfilter/buffersink.h>
- #include <libavfilter/buffersrc.h>
-}
-
-#pragma warning(pop)
-
-namespace caspar { namespace ffmpeg { namespace {
-
-void set_pixel_format(AVFilterContext* sink, AVPixelFormat pix_fmt)
-{
-#pragma warning (push)
-#pragma warning (disable : 4245)
-
- FF(av_opt_set_int_list(
- sink,
- "pix_fmts",
- std::vector<AVPixelFormat>({ pix_fmt, AVPixelFormat::AV_PIX_FMT_NONE }).data(),
- -1,
- AV_OPT_SEARCH_CHILDREN));
-
-#pragma warning (pop)
-}
-
-void adjust_video_filter(const AVCodec& codec, const core::video_format_desc& in_format, AVFilterContext* sink, std::string& filter)
-{
- switch (codec.id)
- {
- case AV_CODEC_ID_DVVIDEO:
- // Crop
- if (in_format.format == core::video_format::ntsc)
- filter = u8(append_filter(u16(filter), L"crop=720:480:0:2"));
-
- // Pixel format selection
- if (in_format.format == core::video_format::ntsc)
- set_pixel_format(sink, AVPixelFormat::AV_PIX_FMT_YUV411P);
- else if (in_format.format == core::video_format::pal)
- set_pixel_format(sink, AVPixelFormat::AV_PIX_FMT_YUV420P);
- else
- set_pixel_format(sink, AVPixelFormat::AV_PIX_FMT_YUV422P);
-
- // Scale
- if (in_format.height == 1080)
- filter = u8(append_filter(u16(filter), in_format.duration == 1001
- ? L"scale=1280:1080"
- : L"scale=1440:1080"));
- else if (in_format.height == 720)
- filter = u8(append_filter(u16(filter), L"scale=960:720"));
-
- break;
- }
-}
-
-void setup_codec_defaults(AVCodecContext& encoder)
-{
- static const int MEGABIT = 1000000;
-
- switch (encoder.codec_id)
- {
- case AV_CODEC_ID_DNXHD:
- encoder.bit_rate = 220 * MEGABIT;
-
- break;
- case AV_CODEC_ID_PRORES:
- encoder.bit_rate = encoder.width < 1280
- ? 63 * MEGABIT
- : 220 * MEGABIT;
-
- break;
- case AV_CODEC_ID_H264:
- av_opt_set(encoder.priv_data, "preset", "ultrafast", 0);
- av_opt_set(encoder.priv_data, "tune", "fastdecode", 0);
- av_opt_set(encoder.priv_data, "crf", "5", 0);
-
- break;
- }
-}
-
-bool is_pcm_s24le_not_supported(const AVFormatContext& container)
-{
- auto name = std::string(container.oformat->name);
-
- if (name == "mp4" || name == "dv")
- return true;
-
- return false;
-}
-
-template<typename Out, typename In>
-std::vector<Out> from_terminated_array(const In* array, In terminator)
-{
- std::vector<Out> result;
-
- while (array != nullptr && *array != terminator)
- {
- In val = *array;
- Out casted = static_cast<Out>(val);
-
- result.push_back(casted);
-
- ++array;
- }
-
- return result;
-}
-
-class ffmpeg_consumer
-{
-private:
- const spl::shared_ptr<diagnostics::graph> graph_;
- core::monitor::subject subject_;
- std::string path_;
- boost::filesystem::path full_path_;
-
- std::map<std::string, std::string> options_;
- bool mono_streams_;
-
- core::video_format_desc in_video_format_;
- core::audio_channel_layout in_channel_layout_ = core::audio_channel_layout::invalid();
-
- std::shared_ptr<AVFormatContext> oc_;
- tbb::atomic<bool> abort_request_;
-
- std::shared_ptr<AVStream> video_st_;
- std::vector<std::shared_ptr<AVStream>> audio_sts_;
-
- std::int64_t video_pts_ = 0;
- std::int64_t audio_pts_ = 0;
-
- std::unique_ptr<audio_filter> audio_filter_;
-
- // TODO: make use of already existent avfilter abstraction for video also
- AVFilterContext* video_graph_in_;
- AVFilterContext* video_graph_out_;
- std::shared_ptr<AVFilterGraph> video_graph_;
-
- executor video_encoder_executor_;
- executor audio_encoder_executor_;
-
- semaphore tokens_ { 0 };
-
- tbb::atomic<int64_t> current_encoding_delay_;
-
- executor write_executor_;
-
-public:
-
- ffmpeg_consumer(
- std::string path,
- std::string options,
- bool mono_streams)
- : path_(path)
- , full_path_(path)
- , mono_streams_(mono_streams)
- , audio_encoder_executor_(print() + L" audio_encoder")
- , video_encoder_executor_(print() + L" video_encoder")
- , write_executor_(print() + L" io")
- {
- abort_request_ = false;
- current_encoding_delay_ = 0;
-
- for(auto it =
- boost::sregex_iterator(
- options.begin(),
- options.end(),
- boost::regex("-(?<NAME>[^-\\s]+)(\\s+(?<VALUE>[^\\s]+))?"));
- it != boost::sregex_iterator();
- ++it)
- {
- options_[(*it)["NAME"].str()] = (*it)["VALUE"].matched ? (*it)["VALUE"].str() : "";
- }
-
- if (options_.find("threads") == options_.end())
- options_["threads"] = "auto";
-
- tokens_.release(
- std::max(
- 1,
- try_remove_arg<int>(
- options_,
- boost::regex("tokens")).get_value_or(2)));
- }
-
- ~ffmpeg_consumer()
- {
- if(oc_)
- {
- video_encoder_executor_.begin_invoke([&] { encode_video(core::const_frame::empty(), nullptr); });
- audio_encoder_executor_.begin_invoke([&] { encode_audio(core::const_frame::empty(), nullptr); });
-
- video_encoder_executor_.stop();
- audio_encoder_executor_.stop();
- video_encoder_executor_.join();
- audio_encoder_executor_.join();
-
- video_graph_.reset();
- audio_filter_.reset();
- video_st_.reset();
- audio_sts_.clear();
-
- write_packet(nullptr, nullptr);
-
- write_executor_.stop();
- write_executor_.join();
-
- FF(av_write_trailer(oc_.get()));
-
- if (!(oc_->oformat->flags & AVFMT_NOFILE) && oc_->pb)
- avio_close(oc_->pb);
-
- oc_.reset();
- }
- }
-
- void initialize(
- const core::video_format_desc& format_desc,
- const core::audio_channel_layout& channel_layout)
- {
- try
- {
- static boost::regex prot_exp("^.+:.*" );
-
- if(!boost::regex_match(
- path_,
- prot_exp))
- {
- if(!full_path_.is_complete())
- {
- full_path_ =
- u8(
- env::media_folder()) +
- path_;
- }
-
- if(boost::filesystem::exists(full_path_))
- boost::filesystem::remove(full_path_);
-
- boost::filesystem::create_directories(full_path_.parent_path());
- }
-
- graph_->set_color("frame-time", diagnostics::color(0.1f, 1.0f, 0.1f));
- graph_->set_color("dropped-frame", diagnostics::color(0.3f, 0.6f, 0.3f));
- graph_->set_text(print());
- diagnostics::register_graph(graph_);
-
- const auto oformat_name =
- try_remove_arg<std::string>(
- options_,
- boost::regex("^f|format$"));
-
- AVFormatContext* oc;
-
- FF(avformat_alloc_output_context2(
- &oc,
- nullptr,
- oformat_name && !oformat_name->empty() ? oformat_name->c_str() : nullptr,
- full_path_.string().c_str()));
-
- oc_.reset(
- oc,
- avformat_free_context);
-
- CASPAR_VERIFY(oc_->oformat);
-
- oc_->interrupt_callback.callback = ffmpeg_consumer::interrupt_cb;
- oc_->interrupt_callback.opaque = this;
-
- CASPAR_VERIFY(format_desc.format != core::video_format::invalid);
-
- in_video_format_ = format_desc;
- in_channel_layout_ = channel_layout;
-
- CASPAR_VERIFY(oc_->oformat);
-
- const auto video_codec_name =
- try_remove_arg<std::string>(
- options_,
- boost::regex("^c:v|codec:v|vcodec$"));
-
- const auto video_codec =
- video_codec_name
- ? avcodec_find_encoder_by_name(video_codec_name->c_str())
- : avcodec_find_encoder(oc_->oformat->video_codec);
-
- const auto audio_codec_name =
- try_remove_arg<std::string>(
- options_,
- boost::regex("^c:a|codec:a|acodec$"));
-
- const auto audio_codec =
- audio_codec_name
- ? avcodec_find_encoder_by_name(audio_codec_name->c_str())
- : (is_pcm_s24le_not_supported(*oc_)
- ? avcodec_find_encoder(oc_->oformat->audio_codec)
- : avcodec_find_encoder_by_name("pcm_s24le"));
-
- if (!video_codec)
- CASPAR_THROW_EXCEPTION(user_error() << msg_info(
- "Failed to find video codec " + (video_codec_name
- ? *video_codec_name
- : "with id " + boost::lexical_cast<std::string>(
- oc_->oformat->video_codec))));
- if (!audio_codec)
- CASPAR_THROW_EXCEPTION(user_error() << msg_info(
- "Failed to find audio codec " + (audio_codec_name
- ? *audio_codec_name
- : "with id " + boost::lexical_cast<std::string>(
- oc_->oformat->audio_codec))));
-
- // Filters
-
- {
- configure_video_filters(
- *video_codec,
- try_remove_arg<std::string>(options_,
- boost::regex("vf|f:v|filter:v")).get_value_or(""));
-
- configure_audio_filters(
- *audio_codec,
- try_remove_arg<std::string>(options_,
- boost::regex("af|f:a|filter:a")).get_value_or(""));
- }
-
- // Encoders
-
- {
- auto video_options = options_;
- auto audio_options = options_;
-
- video_st_ = open_encoder(
- *video_codec,
- video_options,
- 0);
-
- for (int i = 0; i < audio_filter_->get_num_output_pads(); ++i)
- audio_sts_.push_back(open_encoder(
- *audio_codec,
- audio_options,
- i));
-
- auto it = options_.begin();
- while(it != options_.end())
- {
- if(video_options.find(it->first) == video_options.end() || audio_options.find(it->first) == audio_options.end())
- it = options_.erase(it);
- else
- ++it;
- }
- }
-
- // Output
- {
- AVDictionary* av_opts = nullptr;
-
- to_dict(
- &av_opts,
- std::move(options_));
-
- CASPAR_SCOPE_EXIT
- {
- av_dict_free(&av_opts);
- };
-
- if (!(oc_->oformat->flags & AVFMT_NOFILE))
- {
- FF(avio_open2(
- &oc_->pb,
- full_path_.string().c_str(),
- AVIO_FLAG_WRITE,
- &oc_->interrupt_callback,
- &av_opts));
- }
-
- FF(avformat_write_header(
- oc_.get(),
- &av_opts));
-
- options_ = to_map(av_opts);
- }
-
- // Dump Info
-
- av_dump_format(
- oc_.get(),
- 0,
- oc_->filename,
- 1);
-
- for (const auto& option : options_)
- {
- CASPAR_LOG(warning)
- << L"Invalid option: -"
- << u16(option.first)
- << L" "
- << u16(option.second);
- }
- }
- catch(...)
- {
- video_st_.reset();
- audio_sts_.clear();
- oc_.reset();
- throw;
- }
- }
-
- core::monitor::subject& monitor_output()
- {
- return subject_;
- }
-
- void send(core::const_frame frame)
- {
- CASPAR_VERIFY(in_video_format_.format != core::video_format::invalid);
-
- auto frame_timer = spl::make_shared<caspar::timer>();
-
- std::shared_ptr<void> token(
- nullptr,
- [this, frame, frame_timer](void*)
- {
- tokens_.release();
- current_encoding_delay_ = frame.get_age_millis();
- graph_->set_value("frame-time", frame_timer->elapsed() * in_video_format_.fps * 0.5);
- });
- tokens_.acquire();
-
- video_encoder_executor_.begin_invoke([=]() mutable
- {
- encode_video(
- frame,
- token);
- });
-
- audio_encoder_executor_.begin_invoke([=]() mutable
- {
- encode_audio(
- frame,
- token);
- });
- }
-
- bool ready_for_frame() const
- {
- return tokens_.permits() > 0;
- }
-
- void mark_dropped()
- {
- graph_->set_tag(diagnostics::tag_severity::WARNING, "dropped-frame");
- }
-
- std::wstring print() const
- {
- return L"ffmpeg_consumer[" + u16(path_) + L"]";
- }
-
- int64_t presentation_frame_age_millis() const
- {
- return current_encoding_delay_;
- }
-
-private:
-
- static int interrupt_cb(void* ctx)
- {
- CASPAR_ASSERT(ctx);
- return reinterpret_cast<ffmpeg_consumer*>(ctx)->abort_request_;
- }
-
- std::shared_ptr<AVStream> open_encoder(
- const AVCodec& codec,
- std::map<std::string,
- std::string>& options,
- int stream_number_for_media_type)
- {
- auto st =
- avformat_new_stream(
- oc_.get(),
- &codec);
-
- if (!st)
- CASPAR_THROW_EXCEPTION(caspar_exception() << msg_info("Could not allocate video-stream.") << boost::errinfo_api_function("avformat_new_stream"));
-
- auto enc = st->codec;
-
- CASPAR_VERIFY(enc);
-
- switch(enc->codec_type)
- {
- case AVMEDIA_TYPE_VIDEO:
- {
- enc->time_base = video_graph_out_->inputs[0]->time_base;
- enc->pix_fmt = static_cast<AVPixelFormat>(video_graph_out_->inputs[0]->format);
- enc->sample_aspect_ratio = st->sample_aspect_ratio = video_graph_out_->inputs[0]->sample_aspect_ratio;
- enc->width = video_graph_out_->inputs[0]->w;
- enc->height = video_graph_out_->inputs[0]->h;
- enc->bit_rate_tolerance = 400 * 1000000;
-
- break;
- }
- case AVMEDIA_TYPE_AUDIO:
- {
- enc->time_base = audio_filter_->get_output_pad_info(stream_number_for_media_type).time_base;
- enc->sample_fmt = static_cast<AVSampleFormat>(audio_filter_->get_output_pad_info(stream_number_for_media_type).format);
- enc->sample_rate = audio_filter_->get_output_pad_info(stream_number_for_media_type).sample_rate;
- enc->channel_layout = audio_filter_->get_output_pad_info(stream_number_for_media_type).channel_layout;
- enc->channels = audio_filter_->get_output_pad_info(stream_number_for_media_type).channels;
-
- break;
- }
- }
-
- setup_codec_defaults(*enc);
-
- if(oc_->oformat->flags & AVFMT_GLOBALHEADER)
- enc->flags |= CODEC_FLAG_GLOBAL_HEADER;
-
- static const std::array<std::string, 4> char_id_map = {{"v", "a", "d", "s"}};
-
- const auto char_id = char_id_map.at(enc->codec_type);
-
- const auto codec_opts =
- remove_options(
- options,
- boost::regex("^(" + char_id + "?[^:]+):" + char_id + "$"));
-
- AVDictionary* av_codec_opts = nullptr;
-
- to_dict(
- &av_codec_opts,
- options);
-
- to_dict(
- &av_codec_opts,
- codec_opts);
-
- options.clear();
-
- FF(avcodec_open2(
- enc,
- &codec,
- av_codec_opts ? &av_codec_opts : nullptr));
-
- if(av_codec_opts)
- {
- auto t =
- av_dict_get(
- av_codec_opts,
- "",
- nullptr,
- AV_DICT_IGNORE_SUFFIX);
-
- while(t)
- {
- options[t->key + (codec_opts.find(t->key) != codec_opts.end() ? ":" + char_id : "")] = t->value;
-
- t = av_dict_get(
- av_codec_opts,
- "",
- t,
- AV_DICT_IGNORE_SUFFIX);
- }
-
- av_dict_free(&av_codec_opts);
- }
-
- if(enc->codec_type == AVMEDIA_TYPE_AUDIO && !(codec.capabilities & CODEC_CAP_VARIABLE_FRAME_SIZE))
- {
- CASPAR_ASSERT(enc->frame_size > 0);
- audio_filter_->set_guaranteed_output_num_samples_per_frame(
- stream_number_for_media_type,
- enc->frame_size);
- }
-
- return std::shared_ptr<AVStream>(st, [this](AVStream* st)
- {
- avcodec_close(st->codec);
- });
- }
-
- void configure_video_filters(
- const AVCodec& codec,
- std::string filtergraph)
- {
- video_graph_.reset(
- avfilter_graph_alloc(),
- [](AVFilterGraph* p)
- {
- avfilter_graph_free(&p);
- });
-
- video_graph_->nb_threads = boost::thread::hardware_concurrency()/2;
- video_graph_->thread_type = AVFILTER_THREAD_SLICE;
-
- const auto sample_aspect_ratio =
- boost::rational<int>(
- in_video_format_.square_width,
- in_video_format_.square_height) /
- boost::rational<int>(
- in_video_format_.width,
- in_video_format_.height);
-
- const auto vsrc_options = (boost::format("video_size=%1%x%2%:pix_fmt=%3%:time_base=%4%/%5%:pixel_aspect=%6%/%7%:frame_rate=%8%/%9%")
- % in_video_format_.width % in_video_format_.height
- % AVPixelFormat::AV_PIX_FMT_BGRA
- % in_video_format_.duration % in_video_format_.time_scale
- % sample_aspect_ratio.numerator() % sample_aspect_ratio.denominator()
- % in_video_format_.time_scale % in_video_format_.duration).str();
-
- AVFilterContext* filt_vsrc = nullptr;
- FF(avfilter_graph_create_filter(
- &filt_vsrc,
- avfilter_get_by_name("buffer"),
- "ffmpeg_consumer_buffer",
- vsrc_options.c_str(),
- nullptr,
- video_graph_.get()));
-
- AVFilterContext* filt_vsink = nullptr;
- FF(avfilter_graph_create_filter(
- &filt_vsink,
- avfilter_get_by_name("buffersink"),
- "ffmpeg_consumer_buffersink",
- nullptr,
- nullptr,
- video_graph_.get()));
-
-#pragma warning (push)
-#pragma warning (disable : 4245)
-
- FF(av_opt_set_int_list(
- filt_vsink,
- "pix_fmts",
- codec.pix_fmts,
- -1,
- AV_OPT_SEARCH_CHILDREN));
-
-#pragma warning (pop)
-
- adjust_video_filter(codec, in_video_format_, filt_vsink, filtergraph);
-
- configure_filtergraph(
- *video_graph_,
- filtergraph,
- *filt_vsrc,
- *filt_vsink);
-
- video_graph_in_ = filt_vsrc;
- video_graph_out_ = filt_vsink;
-
- CASPAR_LOG(info)
- << u16(std::string("\n")
- + avfilter_graph_dump(
- video_graph_.get(),
- nullptr));
- }
-
- void configure_audio_filters(
- const AVCodec& codec,
- std::string filtergraph)
- {
- int num_output_pads = 1;
-
- if (mono_streams_)
- {
- num_output_pads = in_channel_layout_.num_channels;
- }
-
- if (num_output_pads > 1)
- {
- std::string splitfilter = "[a:0]channelsplit=channel_layout=";
-
- splitfilter += (boost::format("0x%|1$x|") % create_channel_layout_bitmask(in_channel_layout_.num_channels)).str();
-
- for (int i = 0; i < num_output_pads; ++i)
- splitfilter += "[aout:" + boost::lexical_cast<std::string>(i) + "]";
-
- filtergraph = u8(append_filter(u16(filtergraph), u16(splitfilter)));
- }
-
- std::vector<audio_output_pad> output_pads(
- num_output_pads,
- audio_output_pad(
- from_terminated_array<int>( codec.supported_samplerates, 0),
- from_terminated_array<AVSampleFormat>( codec.sample_fmts, AVSampleFormat::AV_SAMPLE_FMT_NONE),
- from_terminated_array<uint64_t>( codec.channel_layouts, 0ull)));
-
- audio_filter_.reset(new audio_filter(
- { audio_input_pad(
- boost::rational<int>(1, in_video_format_.audio_sample_rate),
- in_video_format_.audio_sample_rate,
- AVSampleFormat::AV_SAMPLE_FMT_S32,
- create_channel_layout_bitmask(in_channel_layout_.num_channels)) },
- output_pads,
- filtergraph));
- }
-
- void configure_filtergraph(
- AVFilterGraph& graph,
- const std::string& filtergraph,
- AVFilterContext& source_ctx,
- AVFilterContext& sink_ctx)
- {
- AVFilterInOut* outputs = nullptr;
- AVFilterInOut* inputs = nullptr;
-
- if(!filtergraph.empty())
- {
- outputs = avfilter_inout_alloc();
- inputs = avfilter_inout_alloc();
-
- try
- {
- CASPAR_VERIFY(outputs && inputs);
-
- outputs->name = av_strdup("in");
- outputs->filter_ctx = &source_ctx;
- outputs->pad_idx = 0;
- outputs->next = nullptr;
-
- inputs->name = av_strdup("out");
- inputs->filter_ctx = &sink_ctx;
- inputs->pad_idx = 0;
- inputs->next = nullptr;
- }
- catch (...)
- {
- avfilter_inout_free(&outputs);
- avfilter_inout_free(&inputs);
- throw;
- }
-
- FF(avfilter_graph_parse(
- &graph,
- filtergraph.c_str(),
- inputs,
- outputs,
- nullptr));
- }
- else
- {
- FF(avfilter_link(
- &source_ctx,
- 0,
- &sink_ctx,
- 0));
- }
-
- FF(avfilter_graph_config(
- &graph,
- nullptr));
- }
-
- void encode_video(core::const_frame frame_ptr, std::shared_ptr<void> token)
- {
- if(!video_st_)
- return;
-
- auto enc = video_st_->codec;
-
- if(frame_ptr != core::const_frame::empty())
- {
- auto src_av_frame = create_frame();
-
- const auto sample_aspect_ratio =
- boost::rational<int>(
- in_video_format_.square_width,
- in_video_format_.square_height) /
- boost::rational<int>(
- in_video_format_.width,
- in_video_format_.height);
-
- src_av_frame->format = AVPixelFormat::AV_PIX_FMT_BGRA;
- src_av_frame->width = in_video_format_.width;
- src_av_frame->height = in_video_format_.height;
- src_av_frame->sample_aspect_ratio.num = sample_aspect_ratio.numerator();
- src_av_frame->sample_aspect_ratio.den = sample_aspect_ratio.denominator();
- src_av_frame->pts = video_pts_;
-
- video_pts_ += 1;
-
- subject_
- << core::monitor::message("/frame") % video_pts_
- << core::monitor::message("/path") % path_
- << core::monitor::message("/fps") % in_video_format_.fps;
-
- FF(av_image_fill_arrays(
- src_av_frame->data,
- src_av_frame->linesize,
- frame_ptr.image_data().begin(),
- static_cast<AVPixelFormat>(src_av_frame->format),
- in_video_format_.width,
- in_video_format_.height,
- 1));
-
- FF(av_buffersrc_add_frame(
- video_graph_in_,
- src_av_frame.get()));
- }
-
- int ret = 0;
-
- while(ret >= 0)
- {
- auto filt_frame = create_frame();
-
- ret = av_buffersink_get_frame(
- video_graph_out_,
- filt_frame.get());
-
- video_encoder_executor_.begin_invoke([=]
- {
- if(ret == AVERROR_EOF)
- {
- if(enc->codec->capabilities & CODEC_CAP_DELAY)
- {
- while(encode_av_frame(
- *video_st_,
- avcodec_encode_video2,
- nullptr, token))
- {
- boost::this_thread::yield(); // TODO:
- }
- }
- }
- else if(ret != AVERROR(EAGAIN))
- {
- FF_RET(ret, "av_buffersink_get_frame");
-
- if (filt_frame->interlaced_frame)
- {
- if (enc->codec->id == AV_CODEC_ID_MJPEG)
- enc->field_order = filt_frame->top_field_first ? AV_FIELD_TT : AV_FIELD_BB;
- else
- enc->field_order = filt_frame->top_field_first ? AV_FIELD_TB : AV_FIELD_BT;
- }
- else
- enc->field_order = AV_FIELD_PROGRESSIVE;
-
- filt_frame->quality = enc->global_quality;
-
- if (!enc->me_threshold)
- filt_frame->pict_type = AV_PICTURE_TYPE_NONE;
-
- encode_av_frame(
- *video_st_,
- avcodec_encode_video2,
- filt_frame,
- token);
-
- boost::this_thread::yield(); // TODO:
- }
- });
- }
- }
-
- void encode_audio(core::const_frame frame_ptr, std::shared_ptr<void> token)
- {
- if(audio_sts_.empty())
- return;
-
- if(frame_ptr != core::const_frame::empty())
- {
- auto src_av_frame = create_frame();
-
- src_av_frame->channels = in_channel_layout_.num_channels;
- src_av_frame->channel_layout = create_channel_layout_bitmask(in_channel_layout_.num_channels);
- src_av_frame->sample_rate = in_video_format_.audio_sample_rate;
- src_av_frame->nb_samples = static_cast<int>(frame_ptr.audio_data().size()) / src_av_frame->channels;
- src_av_frame->format = AV_SAMPLE_FMT_S32;
- src_av_frame->pts = audio_pts_;
-
- audio_pts_ += src_av_frame->nb_samples;
-
- FF(av_samples_fill_arrays(
- src_av_frame->extended_data,
- src_av_frame->linesize,
- reinterpret_cast<const std::uint8_t*>(&*frame_ptr.audio_data().begin()),
- src_av_frame->channels,
- src_av_frame->nb_samples,
- static_cast<AVSampleFormat>(src_av_frame->format),
- 16));
-
- audio_filter_->push(0, src_av_frame);
- }
-
- for (int pad_id = 0; pad_id < audio_filter_->get_num_output_pads(); ++pad_id)
- {
- for (auto filt_frame : audio_filter_->poll_all(pad_id))
- {
- audio_encoder_executor_.begin_invoke([=]
- {
- encode_av_frame(
- *audio_sts_.at(pad_id),
- avcodec_encode_audio2,
- filt_frame,
- token);
-
- boost::this_thread::yield(); // TODO:
- });
- }
- }
-
- bool eof = frame_ptr == core::const_frame::empty();
-
- if (eof)
- {
- audio_encoder_executor_.begin_invoke([=]
- {
- for (int pad_id = 0; pad_id < audio_filter_->get_num_output_pads(); ++pad_id)
- {
- auto enc = audio_sts_.at(pad_id)->codec;
-
- if (enc->codec->capabilities & CODEC_CAP_DELAY)
- {
- while (encode_av_frame(
- *audio_sts_.at(pad_id),
- avcodec_encode_audio2,
- nullptr,
- token))
- {
- boost::this_thread::yield(); // TODO:
- }
- }
- }
- });
- }
- }
-
- template<typename F>
- bool encode_av_frame(
- AVStream& st,
- const F& func,
- const std::shared_ptr<AVFrame>& src_av_frame,
- std::shared_ptr<void> token)
- {
- AVPacket pkt = {};
- av_init_packet(&pkt);
-
- int got_packet = 0;
-
- FF(func(
- st.codec,
- &pkt,
- src_av_frame.get(),
- &got_packet));
-
- if(!got_packet || pkt.size <= 0)
- return false;
-
- pkt.stream_index = st.index;
-
- if (pkt.pts != AV_NOPTS_VALUE)
- {
- pkt.pts =
- av_rescale_q(
- pkt.pts,
- st.codec->time_base,
- st.time_base);
- }
-
- if (pkt.dts != AV_NOPTS_VALUE)
- {
- pkt.dts =
- av_rescale_q(
- pkt.dts,
- st.codec->time_base,
- st.time_base);
- }
-
- pkt.duration =
- static_cast<int>(
- av_rescale_q(
- pkt.duration,
- st.codec->time_base, st.time_base));
-
- write_packet(
- std::shared_ptr<AVPacket>(
- new AVPacket(pkt),
- [](AVPacket* p)
- {
- av_free_packet(p);
- delete p;
- }), token);
-
- return true;
- }
-
- void write_packet(
- const std::shared_ptr<AVPacket>& pkt_ptr,
- std::shared_ptr<void> token)
- {
- write_executor_.begin_invoke([this, pkt_ptr, token]() mutable
- {
- FF(av_interleaved_write_frame(
- oc_.get(),
- pkt_ptr.get()));
- });
- }
-
- template<typename T>
- static boost::optional<T> try_remove_arg(
- std::map<std::string, std::string>& options,
- const boost::regex& expr)
- {
- for(auto it = options.begin(); it != options.end(); ++it)
- {
- if(boost::regex_search(it->first, expr))
- {
- auto arg = it->second;
- options.erase(it);
- return boost::lexical_cast<T>(arg);
- }
- }
-
- return boost::optional<T>();
- }
-
- static std::map<std::string, std::string> remove_options(
- std::map<std::string, std::string>& options,
- const boost::regex& expr)
- {
- std::map<std::string, std::string> result;
-
- auto it = options.begin();
- while(it != options.end())
- {
- boost::smatch what;
- if(boost::regex_search(it->first, what, expr))
- {
- result[
- what.size() > 0 && what[1].matched
- ? what[1].str()
- : it->first] = it->second;
- it = options.erase(it);
- }
- else
- ++it;
- }
-
- return result;
- }
-
- static void to_dict(AVDictionary** dest, const std::map<std::string, std::string>& c)
- {
- for (const auto& entry : c)
- {
- av_dict_set(
- dest,
- entry.first.c_str(),
- entry.second.c_str(), 0);
- }
- }
-
- static std::map<std::string, std::string> to_map(AVDictionary* dict)
- {
- std::map<std::string, std::string> result;
-
- for(auto t = dict
- ? av_dict_get(
- dict,
- "",
- nullptr,
- AV_DICT_IGNORE_SUFFIX)
- : nullptr;
- t;
- t = av_dict_get(
- dict,
- "",
- t,
- AV_DICT_IGNORE_SUFFIX))
- {
- result[t->key] = t->value;
- }
-
- return result;
- }
-};
-
-int crc16(const std::string& str)
-{
- boost::crc_16_type result;
-
- result.process_bytes(str.data(), str.length());
-
- return result.checksum();
-}
-
-struct ffmpeg_consumer_proxy : public core::frame_consumer
-{
- const std::string path_;
- const std::string options_;
- const bool separate_key_;
- const bool mono_streams_;
- const bool compatibility_mode_;
- int consumer_index_offset_;
-
- std::unique_ptr<ffmpeg_consumer> consumer_;
- std::unique_ptr<ffmpeg_consumer> key_only_consumer_;
-
-public:
-
- ffmpeg_consumer_proxy(const std::string& path, const std::string& options, bool separate_key, bool mono_streams, bool compatibility_mode)
- : path_(path)
- , options_(options)
- , separate_key_(separate_key)
- , mono_streams_(mono_streams)
- , compatibility_mode_(compatibility_mode)
- , consumer_index_offset_(crc16(path))
- {
- }
-
- void initialize(const core::video_format_desc& format_desc, const core::audio_channel_layout& channel_layout, int) override
- {
- if (consumer_)
- CASPAR_THROW_EXCEPTION(invalid_operation() << msg_info("Cannot reinitialize ffmpeg-consumer."));
-
- consumer_.reset(new ffmpeg_consumer(path_, options_, mono_streams_));
- consumer_->initialize(format_desc, channel_layout);
-
- if (separate_key_)
- {
- boost::filesystem::path fill_file(path_);
- auto without_extension = u16(fill_file.parent_path().string() + "/" + fill_file.stem().string());
- auto key_file = without_extension + L"_A" + u16(fill_file.extension().string());
-
- key_only_consumer_.reset(new ffmpeg_consumer(u8(key_file), options_, mono_streams_));
- key_only_consumer_->initialize(format_desc, channel_layout);
- }
- }
-
- int64_t presentation_frame_age_millis() const override
- {
- return consumer_ ? static_cast<int64_t>(consumer_->presentation_frame_age_millis()) : 0;
- }
-
- std::future<bool> send(core::const_frame frame) override
- {
- bool ready_for_frame = consumer_->ready_for_frame();
-
- if (ready_for_frame && separate_key_)
- ready_for_frame = ready_for_frame && key_only_consumer_->ready_for_frame();
-
- if (ready_for_frame)
- {
- consumer_->send(frame);
-
- if (separate_key_)
- key_only_consumer_->send(frame.key_only());
- }
- else
- {
- consumer_->mark_dropped();
-
- if (separate_key_)
- key_only_consumer_->mark_dropped();
- }
-
- return make_ready_future(true);
- }
-
- std::wstring print() const override
- {
- return consumer_ ? consumer_->print() : L"[ffmpeg_consumer]";
- }
-
- std::wstring name() const override
- {
- return L"ffmpeg";
- }
-
- boost::property_tree::wptree info() const override
- {
- boost::property_tree::wptree info;
-
- info.add(L"type", L"ffmpeg");
- info.add(L"path", u16(path_));
- info.add(L"separate_key", separate_key_);
- info.add(L"mono_streams", mono_streams_);
-
- return info;
- }
-
- bool has_synchronization_clock() const override
- {
- return false;
- }
-
- int buffer_depth() const override
- {
- return -1;
- }
-
- int index() const override
- {
- return compatibility_mode_ ? 200 : 100000 + consumer_index_offset_;
- }
-
- core::monitor::subject& monitor_output() override
- {
- return consumer_->monitor_output();
- }
-};
-
-}
-
-void describe_streaming_consumer(core::help_sink& sink, const core::help_repository& repo)
-{
- sink.short_description(L"For streaming/recording the contents of a channel using FFmpeg.");
- sink.syntax(L"FILE,STREAM [filename:string],[url:string] {-[ffmpeg_param1:string] [value1:string] {-[ffmpeg_param2:string] [value2:string] {...}}} {[separate_key:SEPARATE_KEY]} {[mono_streams:MONO_STREAMS]}");
- sink.para()->text(L"For recording or streaming the contents of a channel using FFmpeg");
- sink.definitions()
- ->item(L"filename", L"The filename under the media folder including the extension (decides which kind of container format that will be used).")
- ->item(L"url", L"If the filename is given in the form of an URL a network stream will be created instead of a file on disk.")
- ->item(L"ffmpeg_paramX", L"A parameter supported by FFmpeg. For example vcodec or acodec etc.")
- ->item(L"separate_key", L"If defined will create two files simultaneously -- One for fill and one for key (_A will be appended).")
- ->item(L"mono_streams", L"If defined every audio channel will be written to its own audio stream.");
- sink.para()->text(L"Examples:");
- sink.example(L">> ADD 1 FILE output.mov -vcodec dnxhd");
- sink.example(L">> ADD 1 FILE output.mov -vcodec prores");
- sink.example(L">> ADD 1 FILE output.mov -vcodec dvvideo");
- sink.example(L">> ADD 1 FILE output.mov -vcodec libx264 -preset ultrafast -tune fastdecode -crf 25");
- sink.example(L">> ADD 1 FILE output.mov -vcodec dnxhd SEPARATE_KEY", L"for creating output.mov with fill and output_A.mov with key/alpha");
- sink.example(L">> ADD 1 FILE output.mxf -vcodec dnxhd MONO_STREAMS", L"for creating output.mxf with every audio channel encoded in its own mono stream.");
- sink.example(L">> ADD 1 STREAM udp://<client_ip_address>:9250 -format mpegts -vcodec libx264 -crf 25 -tune zerolatency -preset ultrafast",
- L"for streaming over UDP instead of creating a local file.");
-}
-
-spl::shared_ptr<core::frame_consumer> create_streaming_consumer(
- const std::vector<std::wstring>& params, core::interaction_sink*, std::vector<spl::shared_ptr<core::video_channel>> channels)
-{
- if (params.size() < 1 || (!boost::iequals(params.at(0), L"STREAM") && !boost::iequals(params.at(0), L"FILE")))
- return core::frame_consumer::empty();
-
- auto params2 = params;
- bool separate_key = get_and_consume_flag(L"SEPARATE_KEY", params2);
- bool mono_streams = get_and_consume_flag(L"MONO_STREAMS", params2);
- auto compatibility_mode = boost::iequals(params.at(0), L"FILE");
- auto path = u8(params2.size() > 1 ? params2.at(1) : L"");
- auto args = u8(boost::join(params2, L" "));
-
- return spl::make_shared<ffmpeg_consumer_proxy>(path, args, separate_key, mono_streams, compatibility_mode);
-}
-
-spl::shared_ptr<core::frame_consumer> create_preconfigured_streaming_consumer(
- const boost::property_tree::wptree& ptree, core::interaction_sink*, std::vector<spl::shared_ptr<core::video_channel>> channels)
-{
- return spl::make_shared<ffmpeg_consumer_proxy>(
- u8(ptree_get<std::wstring>(ptree, L"path")),
- u8(ptree.get<std::wstring>(L"args", L"")),
- ptree.get<bool>(L"separate-key", false),
- ptree.get<bool>(L"mono-streams", false),
- false);
-}
-
-}}