]> git.sesse.net Git - casparcg/blob - core/mixer/audio/audio_mixer.cpp
2.0.0.2: ogl_device: Improved ogl GC.
[casparcg] / core / mixer / audio / audio_mixer.cpp
1 /*\r
2 * copyright (c) 2010 Sveriges Television AB <info@casparcg.com>\r
3 *\r
4 *  This file is part of CasparCG.\r
5 *\r
6 *    CasparCG is free software: you can redistribute it and/or modify\r
7 *    it under the terms of the GNU General Public License as published by\r
8 *    the Free Software Foundation, either version 3 of the License, or\r
9 *    (at your option) any later version.\r
10 *\r
11 *    CasparCG is distributed in the hope that it will be useful,\r
12 *    but WITHOUT ANY WARRANTY; without even the implied warranty of\r
13 *    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the\r
14 *    GNU General Public License for more details.\r
15 \r
16 *    You should have received a copy of the GNU General Public License\r
17 *    along with CasparCG.  If not, see <http://www.gnu.org/licenses/>.\r
18 *\r
19 */\r
20 #include "../../stdafx.h"\r
21 \r
22 #include "audio_mixer.h"\r
23 \r
24 #include <core/mixer/write_frame.h>\r
25 #include <core/producer/frame/audio_transform.h>\r
26 \r
27 namespace caspar { namespace core {\r
28         \r
29 struct audio_mixer::implementation\r
30 {\r
31         std::deque<std::vector<int16_t>> audio_data_;\r
32         std::stack<core::audio_transform> transform_stack_;\r
33 \r
34         std::map<int, core::audio_transform> prev_audio_transforms_;\r
35         std::map<int, core::audio_transform> next_audio_transforms_;\r
36 \r
37 public:\r
38         implementation()\r
39         {\r
40                 transform_stack_.push(core::audio_transform());\r
41                 audio_data_.push_back(std::vector<int16_t>()); // One frame delay\r
42         }\r
43         \r
44         void begin(const core::basic_frame& frame)\r
45         {\r
46                 transform_stack_.push(transform_stack_.top()*frame.get_audio_transform());\r
47         }\r
48 \r
49         void visit(const core::write_frame& frame)\r
50         {\r
51                 if(!transform_stack_.top().get_has_audio())\r
52                         return;\r
53 \r
54                 auto& audio_data = frame.audio_data();\r
55                 auto tag = frame.tag(); // Get the identifier for the audio-stream.\r
56 \r
57                 if(audio_data_.back().empty())\r
58                         audio_data_.back().resize(audio_data.size(), 0);\r
59                 \r
60                 auto next = transform_stack_.top();\r
61                 auto prev = next;\r
62 \r
63                 auto it = prev_audio_transforms_.find(tag);\r
64                 if(it != prev_audio_transforms_.end())\r
65                         prev = it->second;\r
66                                 \r
67                 next_audio_transforms_[tag] = next; // Store all active tags, inactive tags will be removed in end_pass.\r
68                                 \r
69                 if(next.get_gain() < 0.001 && prev.get_gain() < 0.001)\r
70                         return;\r
71                 \r
72                 static const int BASE = 1<<15;\r
73 \r
74                 auto next_gain = static_cast<int>(next.get_gain()*BASE);\r
75                 auto prev_gain = static_cast<int>(prev.get_gain()*BASE);\r
76                 \r
77                 int n_samples = audio_data_.back().size();\r
78 \r
79                 tbb::parallel_for\r
80                 (\r
81                         tbb::blocked_range<size_t>(0, audio_data.size()),\r
82                         [&](const tbb::blocked_range<size_t>& r)\r
83                         {\r
84                                 for(size_t n = r.begin(); n < r.end(); ++n)\r
85                                 {\r
86                                         int sample_gain = (prev_gain - (prev_gain * n)/n_samples) + (next_gain * n)/n_samples;\r
87                                         \r
88                                         int sample = (static_cast<int>(audio_data[n])*sample_gain)/BASE;\r
89                                         \r
90                                         audio_data_.back()[n] = static_cast<int16_t>((static_cast<int>(audio_data_.back()[n]) + sample) & 0xFFFF);\r
91                                 }\r
92                         }\r
93                 );\r
94         }\r
95 \r
96         void begin(const core::audio_transform& transform)\r
97         {\r
98                 transform_stack_.push(transform_stack_.top()*transform);\r
99         }\r
100                 \r
101         void end()\r
102         {\r
103                 transform_stack_.pop();\r
104         }\r
105         \r
106         std::vector<int16_t> mix()\r
107         {\r
108                 prev_audio_transforms_ = std::move(next_audio_transforms_);     \r
109                 auto result = std::move(audio_data_.front());\r
110                 audio_data_.pop_front();\r
111                 audio_data_.push_back(std::vector<int16_t>());\r
112                 return std::move(result);\r
113         }\r
114 };\r
115 \r
116 audio_mixer::audio_mixer() : impl_(new implementation()){}\r
117 void audio_mixer::begin(const core::basic_frame& frame){impl_->begin(frame);}\r
118 void audio_mixer::visit(core::write_frame& frame){impl_->visit(frame);}\r
119 void audio_mixer::end(){impl_->end();}\r
120 std::vector<int16_t> audio_mixer::mix(){return impl_->mix();}\r
121 audio_mixer& audio_mixer::operator=(audio_mixer&& other)\r
122 {\r
123         impl_ = std::move(other.impl_);\r
124         return *this;\r
125 }\r
126 \r
127 }}