-/*\r
-* copyright (c) 2010 Sveriges Television AB <info@casparcg.com>\r
-*\r
-* This file is part of CasparCG.\r
-*\r
-* CasparCG is free software: you can redistribute it and/or modify\r
-* it under the terms of the GNU General Public License as published by\r
-* the Free Software Foundation, either version 3 of the License, or\r
-* (at your option) any later version.\r
-*\r
-* CasparCG is distributed in the hope that it will be useful,\r
-* but WITHOUT ANY WARRANTY; without even the implied warranty of\r
-* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\r
-* GNU General Public License for more details.\r
-\r
-* You should have received a copy of the GNU General Public License\r
-* along with CasparCG. If not, see <http://www.gnu.org/licenses/>.\r
-*\r
-*/\r
-#include "../../stdafx.h"\r
-\r
-#include "audio_mixer.h"\r
-\r
-#include <core/mixer/write_frame.h>\r
-#include <core/producer/frame/frame_transform.h>\r
-\r
-#include <tbb/parallel_for.h>\r
-\r
-#include <safeint.h>\r
-\r
-#include <stack>\r
-#include <deque>\r
-\r
-namespace caspar { namespace core {\r
-\r
-struct audio_item\r
-{\r
- const void* tag;\r
- frame_transform transform;\r
- std::vector<int32_t> audio_data;\r
-};\r
- \r
-struct audio_mixer::implementation\r
-{\r
- std::stack<core::frame_transform> transform_stack_;\r
- std::map<const void*, core::frame_transform> prev_frame_transforms_;\r
- const core::video_format_desc format_desc_;\r
- std::vector<audio_item> items;\r
-\r
-public:\r
- implementation(const core::video_format_desc& format_desc)\r
- : format_desc_(format_desc)\r
- {\r
- transform_stack_.push(core::frame_transform());\r
- }\r
- \r
- void begin(core::basic_frame& frame)\r
- {\r
- transform_stack_.push(transform_stack_.top()*frame.get_frame_transform());\r
- }\r
-\r
- void visit(const core::write_frame& frame)\r
- {\r
- // We only care about the last field.\r
- if(format_desc_.field_mode == field_mode::upper && transform_stack_.top().field_mode == field_mode::upper)\r
- return;\r
-\r
- if(format_desc_.field_mode == field_mode::lower && transform_stack_.top().field_mode == field_mode::lower)\r
- return;\r
-\r
- // Skip empty audio.\r
- if(transform_stack_.top().volume < 0.002 || frame.audio_data().empty())\r
- return;\r
-\r
- audio_item item;\r
- item.tag = frame.tag();\r
- item.transform = transform_stack_.top();\r
- item.audio_data = std::vector<int32_t>(frame.audio_data().begin(), frame.audio_data().end());\r
-\r
- items.push_back(item); \r
- }\r
-\r
- void begin(const core::frame_transform& transform)\r
- {\r
- transform_stack_.push(transform_stack_.top()*transform);\r
- }\r
- \r
- void end()\r
- {\r
- transform_stack_.pop();\r
- }\r
- \r
- std::vector<int32_t> mix()\r
- {\r
- auto result = std::vector<int32_t>(format_desc_.audio_samples_per_frame);\r
-\r
- std::map<const void*, core::frame_transform> next_frame_transforms;\r
-\r
- BOOST_FOREACH(auto& item, items)\r
- { \r
- const auto next = item.transform;\r
- auto prev = next;\r
-\r
- const auto it = prev_frame_transforms_.find(item.tag);\r
- if(it != prev_frame_transforms_.end())\r
- prev = it->second;\r
- \r
- next_frame_transforms[item.tag] = next; // Store all active tags, inactive tags will be removed at the end.\r
- \r
- if(next.volume < 0.001 && prev.volume < 0.001)\r
- continue;\r
- \r
- static const int BASE = 1<<31;\r
-\r
- const auto next_volume = static_cast<int64_t>(next.volume*BASE);\r
- const auto prev_volume = static_cast<int64_t>(prev.volume*BASE);\r
- \r
- const int n_samples = result.size();\r
- \r
- const auto in_size = static_cast<size_t>(item.audio_data.size());\r
- CASPAR_VERIFY(in_size == 0 || in_size == result.size());\r
-\r
- if(in_size > result.size())\r
- continue;\r
-\r
- tbb::parallel_for\r
- (\r
- tbb::blocked_range<size_t>(0, item.audio_data.size()),\r
- [&](const tbb::blocked_range<size_t>& r)\r
- {\r
- for(size_t n = r.begin(); n < r.end(); ++n)\r
- {\r
- const auto sample_volume = (prev_volume - (prev_volume * n)/n_samples) + (next_volume * n)/n_samples;\r
- const auto sample = static_cast<int32_t>((static_cast<int64_t>(item.audio_data[n])*sample_volume)/BASE);\r
- result[n] = result[n] + sample;\r
- }\r
- }\r
- );\r
- }\r
-\r
- items.clear();\r
- prev_frame_transforms_ = std::move(next_frame_transforms); \r
-\r
- return std::move(result);\r
- }\r
-};\r
-\r
-audio_mixer::audio_mixer(const core::video_format_desc& format_desc) : impl_(new implementation(format_desc)){}\r
-void audio_mixer::begin(core::basic_frame& frame){impl_->begin(frame);}\r
-void audio_mixer::visit(core::write_frame& frame){impl_->visit(frame);}\r
-void audio_mixer::end(){impl_->end();}\r
-std::vector<int32_t> audio_mixer::mix(){return impl_->mix();}\r
-audio_mixer& audio_mixer::operator=(audio_mixer&& other)\r
-{\r
- impl_ = std::move(other.impl_);\r
- return *this;\r
-}\r
-\r
-}}
\ No newline at end of file
+/*
+* Copyright (c) 2011 Sveriges Television AB <info@casparcg.com>
+*
+* This file is part of CasparCG (www.casparcg.com).
+*
+* CasparCG is free software: you can redistribute it and/or modify
+* it under the terms of the GNU General Public License as published by
+* the Free Software Foundation, either version 3 of the License, or
+* (at your option) any later version.
+*
+* CasparCG is distributed in the hope that it will be useful,
+* but WITHOUT ANY WARRANTY; without even the implied warranty of
+* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+* GNU General Public License for more details.
+*
+* You should have received a copy of the GNU General Public License
+* along with CasparCG. If not, see <http://www.gnu.org/licenses/>.
+*
+* Author: Robert Nagy, ronag89@gmail.com
+*/
+
+#include "../../StdAfx.h"
+
+#include "audio_mixer.h"
+
+#include <core/frame/frame.h>
+#include <core/frame/frame_transform.h>
+#include <common/diagnostics/graph.h>
+#include <common/linq.h>
+
+#include <boost/range/adaptors.hpp>
+#include <boost/range/distance.hpp>
+
+#include <map>
+#include <stack>
+#include <vector>
+
+namespace caspar { namespace core {
+
+struct audio_item
+{
+ const void* tag = nullptr;
+ audio_transform transform;
+ audio_buffer audio_data;
+
+ audio_item()
+ {
+ }
+
+ audio_item(audio_item&& other)
+ : tag(std::move(other.tag))
+ , transform(std::move(other.transform))
+ , audio_data(std::move(other.audio_data))
+ {
+ }
+};
+
+typedef cache_aligned_vector<float> audio_buffer_ps;
+
+struct audio_stream
+{
+ audio_transform prev_transform;
+ audio_buffer_ps audio_data;
+ bool is_still = false;
+};
+
+struct audio_mixer::impl : boost::noncopyable
+{
+ std::stack<core::audio_transform> transform_stack_;
+ std::map<const void*, audio_stream> audio_streams_;
+ std::vector<audio_item> items_;
+ std::vector<int> audio_cadence_;
+ video_format_desc format_desc_;
+ float master_volume_ = 1.0f;
+ float previous_master_volume_ = master_volume_;
+public:
+ impl()
+ {
+ transform_stack_.push(core::audio_transform());
+ }
+
+ void push(const frame_transform& transform)
+ {
+ transform_stack_.push(transform_stack_.top()*transform.audio_transform);
+ }
+
+ void visit(const const_frame& frame)
+ {
+ if(transform_stack_.top().volume < 0.002 || frame.audio_data().empty())
+ return;
+
+ audio_item item;
+ item.tag = frame.stream_tag();
+ item.transform = transform_stack_.top();
+ item.audio_data = frame.audio_data();
+
+ if(item.transform.is_still)
+ item.transform.volume = 0.0;
+
+ items_.push_back(std::move(item));
+ }
+
+ void begin(const core::audio_transform& transform)
+ {
+ transform_stack_.push(transform_stack_.top()*transform);
+ }
+
+ void pop()
+ {
+ transform_stack_.pop();
+ }
+
+ void set_master_volume(float volume)
+ {
+ master_volume_ = volume;
+ }
+
+ audio_buffer mix(const video_format_desc& format_desc)
+ {
+ if(format_desc_ != format_desc)
+ {
+ audio_streams_.clear();
+ audio_cadence_ = format_desc.audio_cadence;
+ format_desc_ = format_desc;
+ }
+
+ std::map<const void*, audio_stream> next_audio_streams;
+ std::vector<const void*> used_tags;
+
+ for (auto& item : items_)
+ {
+ audio_buffer_ps next_audio;
+
+ auto next_transform = item.transform;
+ auto prev_transform = next_transform;
+
+ auto tag = item.tag;
+
+ if(boost::range::find(used_tags, tag) != used_tags.end())
+ continue;
+
+ used_tags.push_back(tag);
+
+ const auto it = audio_streams_.find(tag);
+ if(it != audio_streams_.end())
+ {
+ prev_transform = it->second.prev_transform;
+ next_audio = std::move(it->second.audio_data);
+ }
+
+ // Skip it if there is no existing audio stream and item has no audio-data.
+ if(it == audio_streams_.end() && item.audio_data.empty())
+ continue;
+
+ const float prev_volume = static_cast<float>(prev_transform.volume) * previous_master_volume_;
+ const float next_volume = static_cast<float>(next_transform.volume) * master_volume_;
+
+ // TODO: Move volume mixing into code below, in order to support audio sample counts not corresponding to frame audio samples.
+ auto alpha = (next_volume-prev_volume)/static_cast<float>(item.audio_data.size()/format_desc.audio_channels);
+
+ for(size_t n = 0; n < item.audio_data.size(); ++n)
+ {
+ auto sample_multiplier = (prev_volume + (n/format_desc_.audio_channels) * alpha);
+ next_audio.push_back(item.audio_data[n] * sample_multiplier);
+ }
+
+ next_audio_streams[tag].prev_transform = std::move(next_transform); // Store all active tags, inactive tags will be removed at the end.
+ next_audio_streams[tag].audio_data = std::move(next_audio);
+ next_audio_streams[tag].is_still = item.transform.is_still;
+ }
+
+ previous_master_volume_ = master_volume_;
+ items_.clear();
+
+ audio_streams_ = std::move(next_audio_streams);
+
+ if(audio_streams_.empty())
+ audio_streams_[nullptr].audio_data = audio_buffer_ps(audio_size(audio_cadence_.front()), 0.0f);
+
+ { // sanity check
+
+ auto nb_invalid_streams = cpplinq::from(audio_streams_)
+ .select(values())
+ .where([&](const audio_stream& x) { return x.audio_data.size() < audio_size(audio_cadence_.front()); })
+ .count();
+
+ if(nb_invalid_streams > 0)
+ CASPAR_LOG(trace) << "[audio_mixer] Incorrect frame audio cadence detected.";
+ }
+
+ std::vector<float> result_ps(audio_size(audio_cadence_.front()), 0.0f);
+ for (auto& stream : audio_streams_ | boost::adaptors::map_values)
+ {
+ if(stream.audio_data.size() < result_ps.size())
+ stream.audio_data.resize(result_ps.size(), 0.0f);
+
+ auto out = boost::range::transform(result_ps, stream.audio_data, std::begin(result_ps), std::plus<float>());
+ stream.audio_data.erase(std::begin(stream.audio_data), std::begin(stream.audio_data) + std::distance(std::begin(result_ps), out));
+ }
+
+ boost::range::rotate(audio_cadence_, std::begin(audio_cadence_)+1);
+
+ audio_buffer result;
+ result.reserve(result_ps.size());
+ boost::range::transform(result_ps, std::back_inserter(result), [](float sample){return static_cast<int32_t>(sample);});
+
+ return result;
+ }
+
+ size_t audio_size(size_t num_samples) const
+ {
+ return num_samples * format_desc_.audio_channels;
+ }
+};
+
+audio_mixer::audio_mixer() : impl_(new impl()){}
+void audio_mixer::push(const frame_transform& transform){impl_->push(transform);}
+void audio_mixer::visit(const const_frame& frame){impl_->visit(frame);}
+void audio_mixer::pop(){impl_->pop();}
+void audio_mixer::set_master_volume(float volume) { impl_->set_master_volume(volume); }
+audio_buffer audio_mixer::operator()(const video_format_desc& format_desc){return impl_->mix(format_desc);}
+
+}}