]> git.sesse.net Git - casparcg/blobdiff - core/mixer/image/image_mixer.cpp
2.0 image_mixer: Refactored, core: Fixed destruction proxy usage.
[casparcg] / core / mixer / image / image_mixer.cpp
index ad296cb4b7c4279bba5ee89a0b9c76ef3785c521..941e28d6c58fbbe18dd1c16af9671678bc07b915 100644 (file)
 #include "../../stdafx.h"\r
 \r
 #include "image_mixer.h"\r
-#include "image_kernel.h"\r
 \r
+#include "image_kernel.h"\r
+#include "../write_frame.h"\r
 #include "../gpu/ogl_device.h"\r
 #include "../gpu/host_buffer.h"\r
 #include "../gpu/device_buffer.h"\r
-#include "../write_frame.h"\r
-\r
 #include "../../video_channel_context.h"\r
 \r
-#include <common/concurrency/executor.h>\r
 #include <common/exception/exceptions.h>\r
 #include <common/gl/gl_check.h>\r
+#include <common/utility/move_on_copy.h>\r
 \r
-#include <core/producer/frame/image_transform.h>\r
+#include <core/producer/frame/frame_transform.h>\r
 #include <core/producer/frame/pixel_format.h>\r
 #include <core/video_format.h>\r
 \r
 #include <gl/glew.h>\r
 \r
 #include <boost/foreach.hpp>\r
-#include <boost/range.hpp>\r
-#include <boost/range/algorithm/find.hpp>\r
 \r
 #include <algorithm>\r
-#include <array>\r
 #include <deque>\r
-#include <unordered_map>\r
+\r
+using namespace boost::assign;\r
 \r
 namespace caspar { namespace core {\r
+       \r
+struct layer\r
+{\r
+       std::vector<render_item> items;\r
+       blend_mode::type                 blend_mode;\r
+\r
+       layer(blend_mode::type blend_mode) : blend_mode(blend_mode)\r
+       {\r
+       }\r
+};\r
+\r
+class image_renderer\r
+{\r
+       video_channel_context&                                  channel_;       \r
+       image_kernel                                                    kernel_;        \r
+       std::shared_ptr<device_buffer>                  active_buffer_;\r
+public:\r
+       image_renderer(video_channel_context& channel)\r
+               : channel_(channel)\r
+       {\r
+       }\r
+       \r
+       boost::unique_future<safe_ptr<host_buffer>> render(std::vector<layer>&& layers)\r
+       {               \r
+               auto layers2 = make_move_on_copy(std::move(layers));\r
+               return channel_.ogl().begin_invoke([=]\r
+               {\r
+                       return do_render(std::move(layers2.value));\r
+               });\r
+       }\r
+       \r
+private:\r
+       safe_ptr<host_buffer> do_render(std::vector<layer>&& layers)\r
+       {\r
+               std::shared_ptr<device_buffer> layer_key_buffer;\r
+\r
+               auto draw_buffer = create_device_buffer(4);\r
+                               \r
+               BOOST_FOREACH(auto& layer, layers)\r
+                       draw_layer(std::move(layer), draw_buffer, layer_key_buffer);\r
                \r
-struct image_mixer::implementation : boost::noncopyable\r
-{              \r
-       typedef std::deque<render_item>                 layer;\r
+               auto host_buffer = channel_.ogl().create_host_buffer(channel_.get_format_desc().size, host_buffer::read_only);\r
+               channel_.ogl().attach(*draw_buffer);\r
+               host_buffer->begin_read(draw_buffer->width(), draw_buffer->height(), format(draw_buffer->stride()));\r
+               \r
+               active_buffer_ = draw_buffer;\r
 \r
-       video_channel_context&                                  channel_;\r
+               channel_.ogl().flush(); // NOTE: This is important, otherwise fences will deadlock.\r
+                       \r
+               return host_buffer;\r
+       }\r
 \r
-       std::vector<image_transform>                    transform_stack_;\r
-       std::vector<video_mode::type>                   mode_stack_;\r
+       void draw_layer(layer&& layer, const safe_ptr<device_buffer>& draw_buffer, std::shared_ptr<device_buffer>& layer_key_buffer)\r
+       {                               \r
+               if(layer.items.empty())\r
+                       return;\r
 \r
-       std::deque<std::deque<render_item>>             layers_; // layer/stream/items\r
-       \r
-       image_kernel                                                    kernel_;\r
+               std::pair<int, std::shared_ptr<device_buffer>> local_key_buffer = std::make_pair(0, nullptr); // int is fields flag\r
+                               \r
+               if(layer.blend_mode != blend_mode::normal && has_overlapping_items(layer))\r
+               {\r
+                       auto layer_draw_buffer = create_device_buffer(4); // int is fields flag\r
+\r
+                       BOOST_FOREACH(auto& item, layer.items)\r
+                               draw_item(std::move(item), *layer_draw_buffer, local_key_buffer, layer_key_buffer);             \r
+                                               \r
+                       render_item item;\r
+                       item.pix_desc.pix_fmt   = pixel_format::bgra;\r
+                       item.pix_desc.planes    = list_of(pixel_format_desc::plane(channel_.get_format_desc().width, channel_.get_format_desc().height, 4));\r
+                       item.textures                   = list_of(layer_draw_buffer);\r
+                       item.transform                  = frame_transform();\r
+                       item.blend_mode                 = layer.blend_mode;\r
+\r
+                       kernel_.draw(channel_.ogl(), std::move(item), *draw_buffer, nullptr, nullptr);\r
+               }\r
+               else // fast path\r
+               {\r
+                       BOOST_FOREACH(auto& item, layer.items)          \r
+                               draw_item(std::move(item), *draw_buffer, local_key_buffer, layer_key_buffer);           \r
+               }                                       \r
+\r
+               CASPAR_ASSERT(local_key_buffer.first == 0 || local_key_buffer.first == core::field_mode::progressive);\r
+\r
+               std::swap(local_key_buffer.second, layer_key_buffer);\r
+       }\r
+\r
+       void draw_item(render_item&&                                                                    item, \r
+                                  device_buffer&                                                                       draw_buffer, \r
+                                  std::pair<int, std::shared_ptr<device_buffer>>&      local_key_buffer, \r
+                                  std::shared_ptr<device_buffer>&                                      layer_key_buffer)\r
+       {                                                                                       \r
+               if(item.transform.is_key)\r
+               {\r
+                       if(!local_key_buffer.second)\r
+                       {\r
+                               local_key_buffer.first = 0;\r
+                               local_key_buffer.second = create_device_buffer(1);\r
+                       }\r
+                       \r
+                       local_key_buffer.first |= item.transform.field_mode; // Add field to flag.\r
+                       kernel_.draw(channel_.ogl(), std::move(item), *local_key_buffer.second, nullptr, nullptr);\r
+               }\r
+               else\r
+               {\r
+                       kernel_.draw(channel_.ogl(), std::move(item), draw_buffer, local_key_buffer.second, layer_key_buffer);\r
+                       local_key_buffer.first ^= item.transform.field_mode; // Remove field from flag.\r
+                       \r
+                       if(local_key_buffer.first == 0) // If all fields from key has been used, reset it\r
+                       {\r
+                               local_key_buffer.first = 0;\r
+                               local_key_buffer.second.reset();\r
+                       }\r
+               }\r
+       }\r
+\r
+       bool has_overlapping_items(const layer& layer)\r
+       {               \r
+               auto upper_count = boost::range::count_if(layer.items, [&](const render_item& item)\r
+               {\r
+                       return !item.transform.is_key && (item.transform.field_mode & field_mode::upper);\r
+               });\r
+\r
+               auto lower_count = boost::range::count_if(layer.items, [&](const render_item& item)\r
+               {\r
+                       return  !item.transform.is_key && (item.transform.field_mode & field_mode::lower);\r
+               });\r
+\r
+               return upper_count > 1 || lower_count > 1;\r
+       }                       \r
                \r
-       std::shared_ptr<device_buffer>                  draw_buffer_;\r
+       safe_ptr<device_buffer> create_device_buffer(size_t stride)\r
+       {\r
+               auto buffer = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, stride);\r
+               channel_.ogl().clear(*buffer);\r
+               return buffer;\r
+       }\r
+};\r
 \r
-       std::shared_ptr<device_buffer>                  local_key_buffer_;\r
-       std::shared_ptr<device_buffer>                  layer_key_buffer_;\r
                \r
+struct image_mixer::implementation : boost::noncopyable\r
+{      \r
+       ogl_device&                                             ogl_;\r
+       image_renderer                                  renderer_;\r
+       std::vector<frame_transform>    transform_stack_;\r
+       std::vector<layer>                              layers_; // layer/stream/items\r
 public:\r
        implementation(video_channel_context& video_channel) \r
-               : channel_(video_channel)\r
-               , transform_stack_(1)\r
-               , mode_stack_(1, video_mode::progressive)\r
+               : ogl_(video_channel.ogl())\r
+               , renderer_(video_channel)\r
+               , transform_stack_(1)   \r
        {\r
        }\r
 \r
-       ~implementation()\r
+       void begin_layer(blend_mode::type blend_mode)\r
        {\r
-               channel_.ogl().gc();\r
+               layers_.push_back(layer(blend_mode));\r
        }\r
-       \r
+               \r
        void begin(core::basic_frame& frame)\r
        {\r
-               transform_stack_.push_back(transform_stack_.back()*frame.get_image_transform());\r
-               mode_stack_.push_back(frame.get_mode() == video_mode::progressive ? mode_stack_.back() : frame.get_mode());\r
+               transform_stack_.push_back(transform_stack_.back()*frame.get_frame_transform());\r
        }\r
                \r
        void visit(core::write_frame& frame)\r
        {       \r
-               CASPAR_ASSERT(!layers_.empty());\r
-\r
-               // Check if frame has been discarded by interlacing\r
-               if(boost::range::find(mode_stack_, video_mode::upper) != mode_stack_.end() && boost::range::find(mode_stack_, video_mode::lower) != mode_stack_.end())\r
+               if(transform_stack_.back().field_mode == field_mode::empty)\r
                        return;\r
                \r
-               core::render_item item(frame.get_pixel_format_desc(), frame.get_textures(), transform_stack_.back(), mode_stack_.back(), frame.tag());  \r
+               core::render_item item;\r
+               item.pix_desc   = frame.get_pixel_format_desc();\r
+               item.textures   = frame.get_textures();\r
+               item.transform  = transform_stack_.back();\r
 \r
-               auto& layer = layers_.back();\r
-\r
-               if(boost::range::find(layer, item) == layer.end())\r
-                       layer.push_back(item);\r
+               layers_.back().items.push_back(item);\r
        }\r
 \r
        void end()\r
        {\r
                transform_stack_.pop_back();\r
-               mode_stack_.pop_back();\r
-       }\r
-\r
-       void begin_layer()\r
-       {\r
-               layers_.push_back(layer());\r
        }\r
 \r
        void end_layer()\r
-       {\r
-       }\r
-       \r
-       boost::unique_future<safe_ptr<host_buffer>> render()\r
        {               \r
-               auto layers = std::move(layers_);\r
-               return channel_.ogl().begin_invoke([=]()mutable\r
-               {\r
-                       return render(std::move(layers));\r
-               });\r
        }\r
        \r
-       safe_ptr<host_buffer> render(std::deque<layer>&& layers)\r
+       boost::unique_future<safe_ptr<host_buffer>> render()\r
        {\r
-               draw_buffer_ = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 4);                             \r
-               channel_.ogl().clear(*draw_buffer_);\r
-                               \r
-               BOOST_FOREACH(auto& layer, layers)\r
-                       draw(std::move(layer));\r
-                               \r
-               auto host_buffer = channel_.ogl().create_host_buffer(channel_.get_format_desc().size, host_buffer::read_only);\r
-               channel_.ogl().attach(*draw_buffer_);\r
-               host_buffer->begin_read(draw_buffer_->width(), draw_buffer_->height(), format(draw_buffer_->stride()));\r
-               \r
-               GL(glFlush());\r
-               \r
-               return host_buffer;\r
+               return renderer_.render(std::move(layers_));\r
        }\r
 \r
-       void draw(layer&& layer)\r
-       {                                       \r
-               local_key_buffer_.reset();\r
-\r
-               BOOST_FOREACH(auto& item, layer)\r
-                       draw(std::move(item));\r
-               \r
-               std::swap(local_key_buffer_, layer_key_buffer_);\r
-       }\r
-\r
-       void draw(render_item&& item)\r
-       {                                                                                       \r
-               if(item.transform.get_is_key())\r
-               {\r
-                       if(!local_key_buffer_)\r
-                       {\r
-                               local_key_buffer_ = channel_.ogl().create_device_buffer(channel_.get_format_desc().width, channel_.get_format_desc().height, 1);\r
-                               channel_.ogl().clear(*local_key_buffer_);\r
-                       }\r
-\r
-                       draw(local_key_buffer_, std::move(item), nullptr, nullptr);\r
-               }\r
-               else\r
-               {\r
-                       draw(draw_buffer_, std::move(item), local_key_buffer_, layer_key_buffer_);      \r
-                       local_key_buffer_.reset();\r
-               }\r
-       }\r
-       \r
-       void draw(std::shared_ptr<device_buffer>& target, render_item&& item, const std::shared_ptr<device_buffer>& local_key, const std::shared_ptr<device_buffer>& layer_key)\r
-       {\r
-               if(!std::all_of(item.textures.begin(), item.textures.end(), std::mem_fn(&device_buffer::ready)))\r
-               {\r
-                       CASPAR_LOG(warning) << L"[image_mixer] Performance warning. Host to device transfer not complete, GPU will be stalled";\r
-                       channel_.ogl().yield(); // Try to give it some more time.\r
-               }               \r
-\r
-               kernel_.draw(channel_.ogl(), std::move(item), make_safe(target), local_key, layer_key);\r
-       }\r
-                               \r
        safe_ptr<write_frame> create_frame(const void* tag, const core::pixel_format_desc& desc)\r
        {\r
-               return make_safe<write_frame>(channel_.ogl(), tag, desc);\r
+               return make_safe<write_frame>(ogl_, tag, desc);\r
        }\r
 };\r
 \r
@@ -196,7 +248,7 @@ void image_mixer::visit(core::write_frame& frame){impl_->visit(frame);}
 void image_mixer::end(){impl_->end();}\r
 boost::unique_future<safe_ptr<host_buffer>> image_mixer::render(){return impl_->render();}\r
 safe_ptr<write_frame> image_mixer::create_frame(const void* tag, const core::pixel_format_desc& desc){return impl_->create_frame(tag, desc);}\r
-void image_mixer::begin_layer(){impl_->begin_layer();}\r
+void image_mixer::begin_layer(blend_mode::type blend_mode){impl_->begin_layer(blend_mode);}\r
 void image_mixer::end_layer(){impl_->end_layer();}\r
 image_mixer& image_mixer::operator=(image_mixer&& other)\r
 {\r