2 * copyright (c) 2010 Sveriges Television AB <info@casparcg.com>
\r
4 * This file is part of CasparCG.
\r
6 * CasparCG is free software: you can redistribute it and/or modify
\r
7 * it under the terms of the GNU General Public License as published by
\r
8 * the Free Software Foundation, either version 3 of the License, or
\r
9 * (at your option) any later version.
\r
11 * CasparCG is distributed in the hope that it will be useful,
\r
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
\r
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
\r
14 * GNU General Public License for more details.
\r
16 * You should have received a copy of the GNU General Public License
\r
17 * along with CasparCG. If not, see <http://www.gnu.org/licenses/>.
\r
20 #include "../../stdafx.h"
\r
22 #include "image_mixer.h"
\r
23 #include "image_kernel.h"
\r
25 #include "../gpu/ogl_device.h"
\r
26 #include "../gpu/host_buffer.h"
\r
27 #include "../gpu/device_buffer.h"
\r
28 #include "../write_frame.h"
\r
30 #include "../../video_channel_context.h"
\r
32 #include <common/concurrency/executor.h>
\r
33 #include <common/exception/exceptions.h>
\r
34 #include <common/gl/gl_check.h>
\r
36 #include <core/producer/frame/image_transform.h>
\r
37 #include <core/producer/frame/pixel_format.h>
\r
38 #include <core/video_format.h>
\r
40 #include <boost/foreach.hpp>
\r
41 #include <boost/range.hpp>
\r
43 #include <algorithm>
\r
45 #include <unordered_map>
\r
47 namespace caspar { namespace core {
\r
49 struct image_mixer::implementation : boost::noncopyable
\r
51 typedef std::deque<render_item> layer;
\r
53 video_channel_context& channel_;
\r
55 std::vector<image_transform> transform_stack_;
\r
56 std::vector<video_mode::type> mode_stack_;
\r
58 std::queue<std::deque<render_item>> layers_; // layer/stream/items
\r
60 image_kernel kernel_;
\r
62 std::array<std::shared_ptr<device_buffer>,2> draw_buffer_;
\r
63 std::shared_ptr<device_buffer> write_buffer_;
\r
65 std::array<std::shared_ptr<device_buffer>,2> stream_key_buffer_;
\r
66 std::shared_ptr<device_buffer> layer_key_buffer_;
\r
69 implementation(video_channel_context& video_channel)
\r
70 : channel_(video_channel)
\r
71 , transform_stack_(1)
\r
72 , mode_stack_(1, video_mode::progressive)
\r
74 initialize_buffers();
\r
79 channel_.ogl().gc();
\r
82 void initialize_buffers()
\r
84 write_buffer_ = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 4);
\r
85 layer_key_buffer_ = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 1);
\r
86 draw_buffer_[0] = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 4);
\r
87 draw_buffer_[1] = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 4);
\r
88 stream_key_buffer_[0] = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 1);
\r
89 stream_key_buffer_[1] = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 1);
\r
90 channel_.ogl().gc();
\r
93 void begin(core::basic_frame& frame)
\r
95 transform_stack_.push_back(transform_stack_.back()*frame.get_image_transform());
\r
96 mode_stack_.push_back(frame.get_mode() == video_mode::progressive ? mode_stack_.back() : frame.get_mode());
\r
99 void visit(core::write_frame& frame)
\r
101 CASPAR_ASSERT(!layers_.empty());
\r
103 // Check if frame has been discarded by interlacing
\r
104 if(boost::range::find(mode_stack_, video_mode::upper) != mode_stack_.end() && boost::range::find(mode_stack_, video_mode::lower) != mode_stack_.end())
\r
107 core::render_item item = {frame.get_pixel_format_desc(), frame.get_textures(), transform_stack_.back(), mode_stack_.back(), frame.tag()};
\r
109 auto& layer = layers_.back();
\r
111 auto it = boost::range::find(layer, item);
\r
112 if(it == layer.end())
\r
113 layer.push_back(item);
\r
118 transform_stack_.pop_back();
\r
119 mode_stack_.pop_back();
\r
124 layers_.push(layer());
\r
131 boost::unique_future<safe_ptr<host_buffer>> render()
\r
133 auto layers = std::move(layers_);
\r
135 return channel_.ogl().begin_invoke([=]() mutable -> safe_ptr<host_buffer>
\r
137 if(channel_.get_format_desc().width != write_buffer_->width() || channel_.get_format_desc().height != write_buffer_->height())
\r
138 initialize_buffers();
\r
140 return do_render(std::move(layers));
\r
144 safe_ptr<host_buffer> do_render(std::queue<layer>&& layers)
\r
146 auto read_buffer = channel_.ogl().create_host_buffer(channel_.get_format_desc().size, host_buffer::read_only);
\r
148 layer_key_buffer_->clear();
\r
149 draw_buffer_[0]->clear();
\r
150 draw_buffer_[1]->clear();
\r
151 stream_key_buffer_[0]->clear();
\r
152 stream_key_buffer_[1]->clear();
\r
154 bool local_key = false;
\r
155 bool layer_key = false;
\r
157 while(!layers.empty())
\r
159 stream_key_buffer_[0]->clear();
\r
161 auto layer = std::move(layers.front());
\r
164 while(!layer.empty())
\r
166 auto item = std::move(layer.front());
\r
169 if(item.transform.get_is_key())
\r
171 render_item(stream_key_buffer_, std::move(item), nullptr, nullptr);
\r
176 render_item(draw_buffer_, std::move(item), local_key ? stream_key_buffer_[0] : nullptr, layer_key ? layer_key_buffer_ : nullptr);
\r
177 stream_key_buffer_[0]->clear();
\r
180 channel_.ogl().yield(); // Return resources to pool as early as possible.
\r
183 layer_key = local_key;
\r
185 std::swap(stream_key_buffer_[0], layer_key_buffer_);
\r
188 std::swap(draw_buffer_[0], write_buffer_);
\r
190 // device -> host.
\r
191 read_buffer->begin_read(*write_buffer_);
\r
193 return read_buffer;
\r
196 void render_item(std::array<std::shared_ptr<device_buffer>,2>& targets, render_item&& item, const std::shared_ptr<device_buffer>& local_key, const std::shared_ptr<device_buffer>& layer_key)
\r
198 if(!std::all_of(item.textures.begin(), item.textures.end(), std::mem_fn(&device_buffer::ready)))
\r
200 CASPAR_LOG(warning) << L"[image_mixer] Performance warning. Host to device transfer not complete, GPU will be stalled";
\r
201 channel_.ogl().yield(); // Try to give it some more time.
\r
204 targets[1]->attach();
\r
206 kernel_.draw(item, make_safe(targets[0]), local_key, layer_key);
\r
208 targets[0]->bind();
\r
210 glCopyTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 0, 0, targets[0]->width(), targets[0]->height());
\r
212 std::swap(targets[0], targets[1]);
\r
215 safe_ptr<write_frame> create_frame(const void* tag, const core::pixel_format_desc& desc)
\r
217 return make_safe<write_frame>(channel_.ogl(), tag, desc);
\r
221 image_mixer::image_mixer(video_channel_context& video_channel) : impl_(new implementation(video_channel)){}
\r
222 void image_mixer::begin(core::basic_frame& frame){impl_->begin(frame);}
\r
223 void image_mixer::visit(core::write_frame& frame){impl_->visit(frame);}
\r
224 void image_mixer::end(){impl_->end();}
\r
225 boost::unique_future<safe_ptr<host_buffer>> image_mixer::render(){return impl_->render();}
\r
226 safe_ptr<write_frame> image_mixer::create_frame(const void* tag, const core::pixel_format_desc& desc){return impl_->create_frame(tag, desc);}
\r
227 void image_mixer::begin_layer(){impl_->begin_layer();}
\r
228 void image_mixer::end_layer(){impl_->end_layer();}
\r
229 image_mixer& image_mixer::operator=(image_mixer&& other)
\r
231 impl_ = std::move(other.impl_);
\r