2 * Copyright (c) 2011 Sveriges Television AB <info@casparcg.com>
\r
4 * This file is part of CasparCG (www.casparcg.com).
\r
6 * CasparCG is free software: you can redistribute it and/or modify
\r
7 * it under the terms of the GNU General Public License as published by
\r
8 * the Free Software Foundation, either version 3 of the License, or
\r
9 * (at your option) any later version.
\r
11 * CasparCG is distributed in the hope that it will be useful,
\r
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
\r
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
\r
14 * GNU General Public License for more details.
\r
16 * You should have received a copy of the GNU General Public License
\r
17 * along with CasparCG. If not, see <http://www.gnu.org/licenses/>.
\r
19 * Author: Robert Nagy, ronag89@gmail.com
\r
24 #include "../../stdafx.h"
\r
29 #include "texture.h"
\r
32 #include <common/assert.h>
\r
33 #include <common/except.h>
\r
34 #include <common/future.h>
\r
35 #include <common/array.h>
\r
36 #include <common/gl/gl_check.h>
\r
37 #include <common/os/windows/windows.h>
\r
39 #include <boost/foreach.hpp>
\r
41 #include <gl/glew.h>
\r
43 #include <SFML/Window/Context.hpp>
\r
45 #include <tbb/concurrent_unordered_map.h>
\r
46 #include <tbb/concurrent_hash_map.h>
\r
47 #include <tbb/concurrent_queue.h>
\r
49 #include <boost/utility/declval.hpp>
\r
52 #include <unordered_map>
\r
55 #include <tbb/parallel_for.h>
\r
57 namespace caspar { namespace accelerator { namespace ogl {
\r
59 struct device::impl : public std::enable_shared_from_this<impl>
\r
61 static_assert(std::is_same<decltype(boost::declval<device>().impl_), spl::shared_ptr<impl>>::value, "impl_ must be shared_ptr");
\r
63 tbb::concurrent_hash_map<buffer*, std::shared_ptr<texture>> texture_cache_;
\r
65 std::unique_ptr<sf::Context> device_;
\r
66 std::unique_ptr<sf::Context> host_alloc_device_;
\r
68 std::array<tbb::concurrent_unordered_map<std::size_t, tbb::concurrent_bounded_queue<std::shared_ptr<texture>>>, 4> device_pools_;
\r
69 std::array<tbb::concurrent_unordered_map<std::size_t, tbb::concurrent_bounded_queue<std::shared_ptr<buffer>>>, 2> host_pools_;
\r
73 executor& render_executor_;
\r
74 executor alloc_executor_;
\r
76 impl(executor& executor)
\r
77 : render_executor_(executor)
\r
78 , alloc_executor_(L"OpenGL allocation context.")
\r
80 render_executor_.set_capacity(128);
\r
81 alloc_executor_.set_capacity(256);
\r
83 CASPAR_LOG(info) << L"Initializing OpenGL Device.";
\r
85 auto ctx1 = render_executor_.invoke([=]() -> HGLRC
\r
87 device_.reset(new sf::Context());
\r
88 device_->SetActive(true);
\r
90 if (glewInit() != GLEW_OK)
\r
91 CASPAR_THROW_EXCEPTION(gl::ogl_exception() << msg_info("Failed to initialize GLEW."));
\r
93 if(!GLEW_VERSION_3_0)
\r
94 CASPAR_THROW_EXCEPTION(not_supported() << msg_info("Your graphics card does not meet the minimum hardware requirements since it does not support OpenGL 3.0 or higher."));
\r
96 glGenFramebuffers(1, &fbo_);
\r
97 glBindFramebuffer(GL_FRAMEBUFFER, fbo_);
\r
99 auto ctx1 = wglGetCurrentContext();
\r
101 device_->SetActive(false);
\r
106 alloc_executor_.invoke([=]
\r
108 host_alloc_device_.reset(new sf::Context());
\r
109 host_alloc_device_->SetActive(true);
\r
110 auto ctx2 = wglGetCurrentContext();
\r
112 if(!wglShareLists(ctx1, ctx2))
\r
113 CASPAR_THROW_EXCEPTION(gl::ogl_exception() << msg_info("Failed to share OpenGL devices."));
\r
116 render_executor_.invoke([=]
\r
118 device_->SetActive(true);
\r
121 CASPAR_LOG(info) << L"Successfully initialized OpenGL " << version();
\r
126 alloc_executor_.invoke([=]
\r
128 host_alloc_device_.reset();
\r
129 BOOST_FOREACH(auto& pool, host_pools_)
\r
133 render_executor_.invoke([=]
\r
135 BOOST_FOREACH(auto& pool, device_pools_)
\r
137 glDeleteFramebuffers(1, &fbo_);
\r
143 std::wstring version()
\r
147 return alloc_executor_.invoke([]
\r
149 return u16(reinterpret_cast<const char*>(GL2(glGetString(GL_VERSION)))) + L" " + u16(reinterpret_cast<const char*>(GL2(glGetString(GL_VENDOR))));
\r
154 return L"Not found";;
\r
158 spl::shared_ptr<texture> create_texture(int width, int height, int stride, bool clear = false)
\r
160 CASPAR_VERIFY(stride > 0 && stride < 5);
\r
161 CASPAR_VERIFY(width > 0 && height > 0);
\r
163 if(!render_executor_.is_current())
\r
164 CASPAR_THROW_EXCEPTION(invalid_operation() << msg_info("Operation only valid in an OpenGL Context."));
\r
166 auto pool = &device_pools_[stride-1][((width << 16) & 0xFFFF0000) | (height & 0x0000FFFF)];
\r
168 std::shared_ptr<texture> tex;
\r
169 if(!pool->try_pop(tex))
\r
170 tex = spl::make_shared<texture>(width, height, stride);
\r
175 return spl::shared_ptr<texture>(tex.get(), [tex, pool](texture*) mutable
\r
181 spl::shared_ptr<buffer> create_buffer(std::size_t size, buffer::usage usage)
\r
183 CASPAR_VERIFY(size > 0);
\r
185 auto pool = &host_pools_[usage.value()][size];
\r
187 std::shared_ptr<buffer> buf;
\r
188 if(!pool->try_pop(buf))
\r
190 buf = alloc_executor_.invoke([&]
\r
192 return spl::make_shared<buffer>(size, usage);
\r
193 }, task_priority::high_priority);
\r
196 auto self = shared_from_this(); // buffers can leave the device context, take a hold on life-time.
\r
197 return spl::shared_ptr<buffer>(buf.get(), [=](buffer*) mutable
\r
199 texture_cache_.erase(buf.get());
\r
204 array<std::uint8_t> create_array(std::size_t size)
\r
206 auto buf = create_buffer(size, buffer::usage::write_only);
\r
207 return array<std::uint8_t>(buf->data(), buf->size(), false, buf);
\r
210 template<typename T>
\r
211 std::shared_ptr<buffer> copy_to_buf(const T& source)
\r
213 std::shared_ptr<buffer> buf;
\r
215 auto tmp = source.storage<spl::shared_ptr<buffer>>();
\r
220 buf = create_buffer(source.size(), buffer::usage::write_only);
\r
221 tbb::parallel_for(tbb::blocked_range<std::size_t>(0, source.size()), [&](const tbb::blocked_range<std::size_t>& r)
\r
223 A_memcpy(buf->data() + r.begin(), source.data() + r.begin(), r.size());
\r
230 // TODO: Since the returned texture is cached it SHOULD NOT be modified.
\r
231 boost::unique_future<spl::shared_ptr<texture>> copy_async(const array<const std::uint8_t>& source, int width, int height, int stride)
\r
233 std::shared_ptr<buffer> buf = copy_to_buf(source);
\r
235 return render_executor_.begin_invoke([=]() -> spl::shared_ptr<texture>
\r
237 tbb::concurrent_hash_map<buffer*, std::shared_ptr<texture>>::const_accessor a;
\r
238 if(texture_cache_.find(a, buf.get()))
\r
239 return spl::make_shared_ptr(a->second);
\r
241 auto texture = create_texture(width, height, stride);
\r
242 texture->copy_from(*buf);
\r
244 texture_cache_.insert(std::make_pair(buf.get(), texture));
\r
247 }, task_priority::high_priority);
\r
250 boost::unique_future<spl::shared_ptr<texture>> copy_async(const array<std::uint8_t>& source, int width, int height, int stride)
\r
252 std::shared_ptr<buffer> buf = copy_to_buf(source);
\r
254 return render_executor_.begin_invoke([=]() -> spl::shared_ptr<texture>
\r
256 auto texture = create_texture(width, height, stride, false);
\r
257 texture->copy_from(*buf);
\r
259 }, task_priority::high_priority);
\r
262 boost::unique_future<array<const std::uint8_t>> copy_async(const spl::shared_ptr<texture>& source)
\r
264 if(!render_executor_.is_current())
\r
265 CASPAR_THROW_EXCEPTION(invalid_operation() << msg_info("Operation only valid in an OpenGL Context."));
\r
267 auto buffer = create_buffer(source->size(), buffer::usage::read_only);
\r
268 source->copy_to(*buffer);
\r
270 auto self = shared_from_this();
\r
271 return async(launch::deferred, [self, buffer]() mutable -> array<const std::uint8_t>
\r
273 self->alloc_executor_.invoke(std::bind(&buffer::map, std::ref(buffer))); // Defer blocking "map" call until data is needed.
\r
274 return array<const std::uint8_t>(buffer->data(), buffer->size(), true, buffer);
\r
280 : executor_(L"OpenGL Rendering Context.")
\r
281 , impl_(new impl(executor_)){}
\r
282 device::~device(){}
\r
283 spl::shared_ptr<texture> device::create_texture(int width, int height, int stride){return impl_->create_texture(width, height, stride, true);}
\r
284 array<std::uint8_t> device::create_array(int size){return impl_->create_array(size);}
\r
285 boost::unique_future<spl::shared_ptr<texture>> device::copy_async(const array<const std::uint8_t>& source, int width, int height, int stride){return impl_->copy_async(source, width, height, stride);}
\r
286 boost::unique_future<spl::shared_ptr<texture>> device::copy_async(const array<std::uint8_t>& source, int width, int height, int stride){return impl_->copy_async(source, width, height, stride);}
\r
287 boost::unique_future<array<const std::uint8_t>> device::copy_async(const spl::shared_ptr<texture>& source){return impl_->copy_async(source);}
\r
288 std::wstring device::version() const{return impl_->version();}
\r