#include "../util/blue_velvet.h"\r
#include "../util/memory.h"\r
\r
+#include <core/parameters/parameters.h>\r
#include <core/video_format.h>\r
#include <core/mixer/read_frame.h>\r
\r
#include <common/memory/memcpy.h>\r
#include <common/memory/memshfl.h>\r
#include <common/utility/timer.h>\r
-#include <common/utility/param.h>\r
\r
#include <core/consumer/frame_consumer.h>\r
#include <core/mixer/audio/audio_util.h>\r
\r
#include <tbb/concurrent_queue.h>\r
+#include <tbb/atomic.h>\r
\r
#include <boost/timer.hpp>\r
#include <boost/range/algorithm.hpp>\r
\r
std::array<blue_dma_buffer_ptr, 4> reserved_frames_; \r
tbb::concurrent_bounded_queue<std::shared_ptr<core::read_frame>> frame_buffer_;\r
+ tbb::atomic<int64_t> presentation_delay_millis_;\r
+ std::shared_ptr<core::read_frame> previous_frame_;\r
\r
const bool embedded_audio_;\r
const bool key_only_;\r
, executor_(print())\r
{\r
executor_.set_capacity(1);\r
+ presentation_delay_millis_ = 0;\r
\r
graph_->set_color("tick-time", diagnostics::color(0.0f, 0.6f, 0.9f)); \r
graph_->set_color("sync-time", diagnostics::color(1.0f, 0.0f, 0.0f));\r
blue_->wait_output_video_synch(UPD_FMT_FRAME, n_field);\r
graph_->set_value("sync-time", sync_timer_.elapsed()*format_desc_.fps*0.5);\r
\r
- frame_timer_.restart(); \r
+ frame_timer_.restart();\r
+\r
+ if (previous_frame_)\r
+ presentation_delay_millis_ = previous_frame_->get_age_millis();\r
+\r
+ previous_frame_ = frame;\r
\r
// Copy to local buffers\r
\r
if(embedded_audio_)\r
{\r
auto src_view = frame->multichannel_view();\r
+ auto frame_audio = core::audio_32_to_24(\r
+ core::get_rearranged_and_mixed(\r
+ src_view,\r
+ channel_layout_,\r
+ channel_layout_.num_channels));\r
+ encode_hanc(\r
+ reinterpret_cast<BLUE_UINT32*>(reserved_frames_.front()->hanc_data()),\r
+ frame_audio.data(),\r
+ src_view.num_samples(),\r
+ channel_layout_.num_channels);\r
\r
- if (core::needs_rearranging(src_view, channel_layout_, channel_layout_.num_channels))\r
- {\r
- std::vector<int32_t> resulting_audio_data;\r
- resulting_audio_data.resize(src_view.num_samples() * channel_layout_.num_channels, 0);\r
-\r
- auto dest_view = core::make_multichannel_view<int32_t>(\r
- resulting_audio_data.begin(), \r
- resulting_audio_data.end(),\r
- channel_layout_);\r
-\r
- core::rearrange_or_rearrange_and_mix(\r
- src_view,\r
- dest_view,\r
- core::default_mix_config_repository());\r
-\r
- auto frame_audio = core::audio_32_to_24(resulting_audio_data);\r
- encode_hanc(\r
- reinterpret_cast<BLUE_UINT32*>(reserved_frames_.front()->hanc_data()),\r
- frame_audio.data(),\r
- src_view.num_samples(),\r
- channel_layout_.num_channels);\r
- }\r
- else\r
- {\r
- auto frame_audio = core::audio_32_to_24(frame->audio_data());\r
- encode_hanc(\r
- reinterpret_cast<BLUE_UINT32*>(reserved_frames_.front()->hanc_data()),\r
- frame_audio.data(),\r
- src_view.num_samples(),\r
- channel_layout_.num_channels);\r
- }\r
- \r
blue_->system_buffer_write_async(const_cast<uint8_t*>(reserved_frames_.front()->image_data()), \r
reserved_frames_.front()->image_size(), \r
nullptr, \r
return model_name_ + L" [" + boost::lexical_cast<std::wstring>(channel_index_) + L"-" + \r
boost::lexical_cast<std::wstring>(device_index_) + L"|" + format_desc_.name + L"]";\r
}\r
+\r
+ int64_t presentation_delay_millis() const\r
+ {\r
+ return presentation_delay_millis_;\r
+ }\r
};\r
\r
struct bluefish_consumer_proxy : public core::frame_consumer\r
\r
// frame_consumer\r
\r
- virtual void initialize(const core::video_format_desc& format_desc, int channel_index) override\r
+ virtual void initialize(\r
+ const core::video_format_desc& format_desc,\r
+ const core::channel_layout& audio_channel_layout,\r
+ int channel_index) override\r
{\r
consumer_.reset(new bluefish_consumer(\r
format_desc,\r
format_desc_ = format_desc;\r
CASPAR_LOG(info) << print() << L" Successfully Initialized."; \r
}\r
- \r
+\r
virtual boost::unique_future<bool> send(const safe_ptr<core::read_frame>& frame) override\r
{\r
CASPAR_VERIFY(audio_cadence_.front() * frame->num_channels() == static_cast<size_t>(frame->audio_data().size()));\r
info.add(L"key-only", key_only_);\r
info.add(L"device", device_index_);\r
info.add(L"embedded-audio", embedded_audio_);\r
+ info.add(L"presentation-frame-age", presentation_frame_age_millis());\r
return info;\r
}\r
\r
- size_t buffer_depth() const override\r
+ virtual size_t buffer_depth() const override\r
{\r
return 1;\r
}\r
{\r
return 400 + device_index_;\r
}\r
+\r
+ virtual int64_t presentation_frame_age_millis() const override\r
+ {\r
+ return consumer_ ? consumer_->presentation_delay_millis() : 0;\r
+ }\r
}; \r
\r
-safe_ptr<core::frame_consumer> create_consumer(const std::vector<std::wstring>& params)\r
+safe_ptr<core::frame_consumer> create_consumer(const core::parameters& params)\r
{\r
if(params.size() < 1 || params[0] != L"BLUEFISH")\r
return core::frame_consumer::empty();\r
const auto embedded_audio = std::find(params.begin(), params.end(), L"EMBEDDED_AUDIO") != params.end();\r
const auto key_only = std::find(params.begin(), params.end(), L"KEY_ONLY") != params.end();\r
const auto audio_layout = core::default_channel_layout_repository().get_by_name(\r
- get_param(L"CHANNEL_LAYOUT", params, L"STEREO"));\r
+ params.get(L"CHANNEL_LAYOUT", L"STEREO"));\r
\r
return make_safe<bluefish_consumer_proxy>(device_index, embedded_audio, key_only, audio_layout);\r
}\r