X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=src%2Fmodules%2Fcore%2Ftransition_composite.c;h=14e00575519a52f53c2bdc6c9c8ba34bae5088b1;hb=b698c58e57643e9df6941d91b07a40c997bd7e64;hp=b28283d812a8ecd2a16a8b64beacc9598cb8d28c;hpb=0ec09d9918d262619c480db16b5cb16768df2197;p=mlt diff --git a/src/modules/core/transition_composite.c b/src/modules/core/transition_composite.c index b28283d8..14e00575 100644 --- a/src/modules/core/transition_composite.c +++ b/src/modules/core/transition_composite.c @@ -3,19 +3,19 @@ * Copyright (C) 2003-2004 Ushodaya Enterprises Limited * Author: Dan Dennedy * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the GNU General Public License as published by - * the Free Software Foundation; either version 2 of the License, or - * (at your option) any later version. + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. * - * This program is distributed in the hope that it will be useful, + * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - * GNU General Public License for more details. + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. * - * You should have received a copy of the GNU General Public License - * along with this program; if not, write to the Free Software Foundation, - * Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. + * You should have received a copy of the GNU Lesser General Public + * License along with this library; if not, write to the Free Software + * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "transition_composite.h" @@ -27,7 +27,7 @@ #include #include -typedef void ( *composite_line_fn )( uint8_t *dest, uint8_t *src, int width_src, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness ); +typedef void ( *composite_line_fn )( uint8_t *dest, uint8_t *src, int width_src, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int softness, uint32_t step ); /** Geometry struct. */ @@ -41,6 +41,8 @@ struct geometry_s int sh; // scaled height, not including consumer scale based upon h/nh int halign; // horizontal alignment: 0=left, 1=center, 2=right int valign; // vertical alignment: 0=top, 1=middle, 2=bottom + int x_src; + int y_src; }; /** Parse the alignment properties into the geometry. @@ -199,7 +201,7 @@ static inline double delta_calculate( mlt_transition this, mlt_frame frame, mlt_ return length * ( y - x ) / 2.0; } -static int get_value( mlt_properties properties, char *preferred, char *fallback ) +static int get_value( mlt_properties properties, const char *preferred, const char *fallback ) { int value = mlt_properties_get_int( properties, preferred ); if ( value == 0 ) @@ -366,9 +368,9 @@ static void luma_read_yuv422( uint8_t *image, uint16_t **map, int width, int hei *p++ = ( image[ i ] - 16 ) * 299; // 299 = 65535 / 219 } -static inline int calculate_mix( uint16_t *luma, int j, int soft, int weight, int alpha ) +static inline int calculate_mix( uint16_t *luma, int j, int softness, int weight, int alpha, uint32_t step ) { - return ( ( ( luma == NULL ) ? weight : smoothstep( luma[ j ], luma[ j ] + soft, weight + soft ) ) * alpha ) >> 8; + return ( ( luma ? smoothstep( luma[ j ], luma[ j ] + softness, step ) : weight ) * alpha ) >> 8; } static inline uint8_t sample_mix( uint8_t dest, uint8_t src, int mix ) @@ -379,97 +381,118 @@ static inline uint8_t sample_mix( uint8_t dest, uint8_t src, int mix ) /** Composite a source line over a destination line */ -static void composite_line_yuv( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) +static void composite_line_yuv( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft, uint32_t step ) { register int j; register int mix; for ( j = 0; j < width; j ++ ) { - mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ ); + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++, step ); *dest = sample_mix( *dest, *src++, mix ); dest++; *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a = mix | *alpha_a; + *alpha_a = ( mix >> 8 ) | *alpha_a; alpha_a ++; } } -static void composite_line_yuv_or( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) +static void composite_line_yuv_or( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft, uint32_t step ) { register int j; register int mix; for ( j = 0; j < width; j ++ ) { - mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ | *alpha_a ); + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ | *alpha_a, step ); *dest = sample_mix( *dest, *src++, mix ); dest++; *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a ++ = mix; + *alpha_a ++ = mix >> 8; } } -static void composite_line_yuv_and( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) +static void composite_line_yuv_and( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft, uint32_t step ) { register int j; register int mix; for ( j = 0; j < width; j ++ ) { - mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ & *alpha_a ); + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ & *alpha_a, step ); *dest = sample_mix( *dest, *src++, mix ); dest++; *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a ++ = mix; + *alpha_a ++ = mix >> 8; } } -static void composite_line_yuv_xor( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft ) +static void composite_line_yuv_xor( uint8_t *dest, uint8_t *src, int width, uint8_t *alpha_b, uint8_t *alpha_a, int weight, uint16_t *luma, int soft, uint32_t step ) { register int j; register int mix; for ( j = 0; j < width; j ++ ) { - mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ ^ *alpha_a ); + mix = calculate_mix( luma, j, soft, weight, *alpha_b ++ ^ *alpha_a, step ); *dest = sample_mix( *dest, *src++, mix ); dest++; *dest = sample_mix( *dest, *src++, mix ); dest++; - *alpha_a ++ = mix; + *alpha_a ++ = mix >> 8; } } /** Composite function. */ -static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint8_t *p_src, int width_src, int height_src, uint8_t *alpha_b, uint8_t *alpha_a, struct geometry_s geometry, int field, uint16_t *p_luma, int32_t softness, composite_line_fn line_fn ) +static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint8_t *p_src, int width_src, int height_src, uint8_t *alpha_b, uint8_t *alpha_a, struct geometry_s geometry, int field, uint16_t *p_luma, double softness, composite_line_fn line_fn ) { int ret = 0; int i; - int x_src = 0, y_src = 0; - int32_t weight = ( ( 1 << 16 ) - 1 ) * ( geometry.item.mix / 100 ); + int x_src = -geometry.x_src, y_src = -geometry.y_src; + int uneven_x_src = ( x_src % 2 ); int step = ( field > -1 ) ? 2 : 1; int bpp = 2; - int stride_src = width_src * bpp; + int stride_src = geometry.sw * bpp; int stride_dest = width_dest * bpp; - + int i_softness = ( 1 << 16 ) * softness; + int weight = ( ( 1 << 16 ) - 1 ) * geometry.item.mix / 100; + uint32_t luma_step = ( ( 1 << 16 ) - 1 ) * geometry.item.mix / 100 * ( 1.0 + softness ); + // Adjust to consumer scale - int x = rint( 0.5 + geometry.item.x * width_dest / geometry.nw ); - int y = rint( 0.5 + geometry.item.y * height_dest / geometry.nh ); + int x = rint( geometry.item.x * width_dest / geometry.nw ); + int y = rint( geometry.item.y * height_dest / geometry.nh ); int uneven_x = ( x % 2 ); // optimization points - no work to do - if ( width_src <= 0 || height_src <= 0 ) + if ( width_src <= 0 || height_src <= 0 || y_src >= height_src || x_src >= width_src ) return ret; if ( ( x < 0 && -x >= width_src ) || ( y < 0 && -y >= height_src ) ) return ret; + // cropping affects the source width + if ( x_src > 0 ) + { + width_src -= x_src; + // and it implies cropping + if ( width_src > geometry.item.w ) + width_src = geometry.item.w; + } + + // cropping affects the source height + if ( y_src > 0 ) + { + height_src -= y_src; + // and it implies cropping + if ( height_src > geometry.item.h ) + height_src = geometry.item.h; + } + // crop overlay off the left edge of frame if ( x < 0 ) { @@ -477,7 +500,7 @@ static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint width_src -= x_src; x = 0; } - + // crop overlay beyond right edge of frame if ( x + width_src > width_dest ) width_src = width_dest - x; @@ -534,15 +557,18 @@ static int composite_yuv( uint8_t *p_dest, int width_dest, int height_dest, uint int alpha_b_stride = stride_src / bpp; int alpha_a_stride = stride_dest / bpp; - p_src += uneven_x * 2; - width_src -= 2 * uneven_x; - alpha_b += uneven_x; - uneven_x = 0; + // Align chroma of source and destination + if ( uneven_x != uneven_x_src ) + { + p_src += 2; + width_src -= 2; + alpha_b += 1; + } // now do the compositing only to cropped extents for ( i = 0; i < height_src; i += step ) { - line_fn( p_dest, p_src, width_src, alpha_b, alpha_a, weight, p_luma, softness ); + line_fn( p_dest, p_src, width_src, alpha_b, alpha_a, weight, p_luma, i_softness, luma_step ); p_src += stride_src; p_dest += stride_dest; @@ -581,7 +607,7 @@ scale_luma ( uint16_t *dest_buf, int dest_width, int dest_height, const uint16_t } } -static uint16_t* get_luma( mlt_properties properties, int width, int height ) +static uint16_t* get_luma( mlt_transition this, mlt_properties properties, int width, int height ) { // The cached luma map information int luma_width = mlt_properties_get_int( properties, "_luma.width" ); @@ -600,11 +626,11 @@ static uint16_t* get_luma( mlt_properties properties, int width, int height ) luma_height = height; } - if ( resource != NULL && strchr( resource, '%' ) ) + if ( resource && resource[0] && strchr( resource, '%' ) ) { // TODO: Clean up quick and dirty compressed/existence check FILE *test; - sprintf( temp, "%s/lumas/%s/%s", mlt_factory_prefix( ), mlt_environment( "MLT_NORMALISATION" ), strchr( resource, '%' ) + 1 ); + sprintf( temp, "%s/lumas/%s/%s", mlt_environment( "MLT_DATA" ), mlt_environment( "MLT_NORMALISATION" ), strchr( resource, '%' ) + 1 ); test = fopen( temp, "r" ); if ( test == NULL ) strcat( temp, ".png" ); @@ -613,7 +639,28 @@ static uint16_t* get_luma( mlt_properties properties, int width, int height ) resource = temp; } - if ( resource != NULL && ( luma_bitmap == NULL || luma_width != width || luma_height != height ) ) + if ( resource && resource[0] ) + { + char *old_luma = mlt_properties_get( properties, "_luma" ); + int old_invert = mlt_properties_get_int( properties, "_luma_invert" ); + + if ( invert != old_invert || ( old_luma && old_luma[0] && strcmp( resource, old_luma ) ) ) + { + mlt_properties_set_data( properties, "_luma.orig_bitmap", NULL, 0, NULL, NULL ); + luma_bitmap = NULL; + } + } + else { + char *old_luma = mlt_properties_get( properties, "_luma" ); + if ( old_luma && old_luma[0] ) + { + mlt_properties_set_data( properties, "_luma.orig_bitmap", NULL, 0, NULL, NULL ); + luma_bitmap = NULL; + mlt_properties_set( properties, "_luma", NULL); + } + } + + if ( resource && resource[0] && ( luma_bitmap == NULL || luma_width != width || luma_height != height ) ) { uint16_t *orig_bitmap = mlt_properties_get_data( properties, "_luma.orig_bitmap", NULL ); luma_width = mlt_properties_get_int( properties, "_luma.orig_width" ); @@ -647,7 +694,8 @@ static uint16_t* get_luma( mlt_properties properties, int width, int height ) char *factory = mlt_properties_get( properties, "factory" ); // Create the producer - mlt_producer producer = mlt_factory_producer( factory, resource ); + mlt_profile profile = mlt_service_profile( MLT_TRANSITION_SERVICE( this ) ); + mlt_producer producer = mlt_factory_producer( profile, factory, resource ); // If we have one if ( producer != NULL ) @@ -700,6 +748,8 @@ static uint16_t* get_luma( mlt_properties properties, int width, int height ) mlt_properties_set_int( properties, "_luma.width", width ); mlt_properties_set_int( properties, "_luma.height", height ); mlt_properties_set_data( properties, "_luma.bitmap", luma_bitmap, width * height * 2, mlt_pool_release, NULL ); + mlt_properties_set( properties, "_luma", resource ); + mlt_properties_set_int( properties, "_luma_invert", invert ); } return luma_bitmap; } @@ -715,8 +765,25 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** // Get the properties objects mlt_properties b_props = MLT_FRAME_PROPERTIES( b_frame ); mlt_properties properties = MLT_TRANSITION_PROPERTIES( this ); + uint8_t resize_alpha = mlt_properties_get_int( b_props, "resize_alpha" ); - if ( mlt_properties_get_int( properties, "distort" ) == 0 && mlt_properties_get_int( b_props, "distort" ) == 0 && geometry->item.distort == 0 ) + // Do not scale if we are cropping - the compositing rectangle can crop the b image + // TODO: Use the animatable w and h of the crop geometry to scale independently of crop rectangle + if ( mlt_properties_get( properties, "crop" ) ) + { + int real_width = get_value( b_props, "real_width", "width" ); + int real_height = get_value( b_props, "real_height", "height" ); + double input_ar = mlt_properties_get_double( b_props, "aspect_ratio" ); + double consumer_ar = mlt_properties_get_double( b_props, "consumer_aspect_ratio" ); + double background_ar = mlt_properties_get_double( b_props, "output_ratio" ); + double output_ar = background_ar != 0.0 ? background_ar : consumer_ar; + int scaled_width = rint( ( input_ar == 0.0 ? output_ar : input_ar ) / output_ar * real_width ); + int scaled_height = real_height; + geometry->sw = scaled_width; + geometry->sh = scaled_height; + } + // Normalise aspect ratios and scale preserving aspect ratio + else if ( mlt_properties_get_int( properties, "aligned" ) && mlt_properties_get_int( properties, "distort" ) == 0 && mlt_properties_get_int( b_props, "distort" ) == 0 && geometry->item.distort == 0 ) { // Adjust b_frame pixel aspect int normalised_width = geometry->item.w; @@ -727,18 +794,21 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** double consumer_ar = mlt_properties_get_double( b_props, "consumer_aspect_ratio" ); double background_ar = mlt_properties_get_double( b_props, "output_ratio" ); double output_ar = background_ar != 0.0 ? background_ar : consumer_ar; - int scaled_width = rint( 0.5 + ( input_ar == 0.0 ? output_ar : input_ar ) / output_ar * real_width ); + int scaled_width = rint( ( input_ar == 0.0 ? output_ar : input_ar ) / output_ar * real_width ); int scaled_height = real_height; +// fprintf(stderr, "%s: scaled %dx%d norm %dx%d real %dx%d output_ar %f => %f\n", __FILE__, +// scaled_width, scaled_height, normalised_width, normalised_height, real_width, real_height, +// background_ar, output_ar); // Now ensure that our images fit in the normalised frame if ( scaled_width > normalised_width ) { - scaled_height = rint( 0.5 + scaled_height * normalised_width / scaled_width ); + scaled_height = rint( scaled_height * normalised_width / scaled_width ); scaled_width = normalised_width; } if ( scaled_height > normalised_height ) { - scaled_width = rint( 0.5 + scaled_width * normalised_height / scaled_height ); + scaled_width = rint( scaled_width * normalised_height / scaled_height ); scaled_height = normalised_height; } @@ -746,14 +816,14 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** // ????: Shouln't this be the default behaviour? if ( mlt_properties_get_int( properties, "fill" ) && scaled_width > 0 && scaled_height > 0 ) { - if ( scaled_height < normalised_height && scaled_width * normalised_height / scaled_height < normalised_width ) + if ( scaled_height < normalised_height && scaled_width * normalised_height / scaled_height <= normalised_width ) { - scaled_width = rint( 0.5 + scaled_width * normalised_height / scaled_height ); + scaled_width = rint( scaled_width * normalised_height / scaled_height ); scaled_height = normalised_height; } else if ( scaled_width < normalised_width && scaled_height * normalised_width / scaled_width < normalised_height ) { - scaled_height = rint( 0.5 + scaled_height * normalised_width / scaled_width ); + scaled_height = rint( scaled_height * normalised_width / scaled_width ); scaled_width = normalised_width; } } @@ -769,21 +839,81 @@ static int get_b_frame_image( mlt_transition this, mlt_frame b_frame, uint8_t ** } // We want to ensure that we bypass resize now... - mlt_properties_set_int( b_props, "distort", mlt_properties_get_int( properties, "distort" ) ); + if ( resize_alpha == 0 ) + mlt_properties_set_int( b_props, "distort", mlt_properties_get_int( properties, "distort" ) ); + + // If we're not aligned, we want a non-transparent background + if ( mlt_properties_get_int( properties, "aligned" ) == 0 ) + mlt_properties_set_int( b_props, "resize_alpha", 255 ); // Take into consideration alignment for optimisation (titles are a special case) - if ( !mlt_properties_get_int( properties, "titles" ) ) + if ( !mlt_properties_get_int( properties, "titles" ) && + mlt_properties_get( properties, "crop" ) == NULL ) alignment_calculate( geometry ); // Adjust to consumer scale - *width = rint( 0.5 + geometry->sw * *width / geometry->nw ); - *height = rint( 0.5 + geometry->sh * *height / geometry->nh ); + *width = rint( geometry->sw * *width / geometry->nw ); + *height = rint( geometry->sh * *height / geometry->nh ); +// fprintf(stderr, "%s: scaled %dx%d norm %dx%d resize %dx%d\n", __FILE__, +// geometry->sw, geometry->sh, geometry->nw, geometry->nh, *width, *height); ret = mlt_frame_get_image( b_frame, image, &format, width, height, 1 ); + // composite_yuv uses geometry->sw to determine source stride, which + // should equal the image width if not using crop property. + if ( !mlt_properties_get( properties, "crop" ) ) + geometry->sw = *width; + + // Set the frame back + mlt_properties_set_int( b_props, "resize_alpha", resize_alpha ); + return ret && image != NULL; } +static void crop_calculate( mlt_transition this, mlt_properties properties, struct geometry_s *result, double position ) +{ + // Initialize panning info + result->x_src = 0; + result->y_src = 0; + if ( mlt_properties_get( properties, "crop" ) ) + { + mlt_geometry crop = mlt_properties_get_data( properties, "crop_geometry", NULL ); + if ( !crop ) + { + crop = mlt_geometry_init(); + mlt_position in = mlt_transition_get_in( this ); + mlt_position out = mlt_transition_get_out( this ); + int length = out - in + 1; + double cycle = mlt_properties_get_double( properties, "cycle" ); + + // Allow a geometry repeat cycle + if ( cycle >= 1 ) + length = cycle; + else if ( cycle > 0 ) + length *= cycle; + mlt_geometry_parse( crop, mlt_properties_get( properties, "crop" ), length, result->sw, result->sh ); + mlt_properties_set_data( properties, "crop_geometry", crop, 0, (mlt_destructor)mlt_geometry_close, NULL ); + } + + // Repeat processing + int length = mlt_geometry_get_length( crop ); + int mirror_off = mlt_properties_get_int( properties, "mirror_off" ); + int repeat_off = mlt_properties_get_int( properties, "repeat_off" ); + if ( !repeat_off && position >= length && length != 0 ) + { + int section = position / length; + position -= section * length; + if ( !mirror_off && section % 2 == 1 ) + position = length - position; + } + + // Compute the pan + struct mlt_geometry_item_s crop_item; + mlt_geometry_fetch( crop, &crop_item, position ); + result->x_src = rint( crop_item.x ); + result->y_src = rint( crop_item.y ); + } +} static mlt_geometry composite_calculate( mlt_transition this, struct geometry_s *result, mlt_frame a_frame, double position ) { @@ -842,13 +972,15 @@ static mlt_geometry composite_calculate( mlt_transition this, struct geometry_s result->halign = alignment_parse( mlt_properties_get( properties, "halign" ) ); result->valign = alignment_parse( mlt_properties_get( properties, "valign" ) ); + crop_calculate( this, properties, result, position ); + return start; } mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_position frame_position ) { // Create a frame to return - mlt_frame b_frame = mlt_frame_init( ); + mlt_frame b_frame = mlt_frame_init( MLT_TRANSITION_SERVICE( this ) ); // Get the properties of the a frame mlt_properties a_props = MLT_FRAME_PROPERTIES( a_frame ); @@ -891,10 +1023,10 @@ mlt_frame composite_copy_region( mlt_transition this, mlt_frame a_frame, mlt_pos composite_calculate( this, &result, a_frame, position ); // Need to scale down to actual dimensions - x = rint( 0.5 + result.item.x * width / result.nw ); - y = rint( 0.5 + result.item.y * height / result.nh ); - w = rint( 0.5 + result.item.w * width / result.nw ); - h = rint( 0.5 + result.item.h * height / result.nh ); + x = rint( result.item.x * width / result.nw ); + y = rint( result.item.y * height / result.nh ); + w = rint( result.item.w * width / result.nw ); + h = rint( result.item.h * height / result.nh ); if ( x % 2 ) { @@ -1013,7 +1145,7 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f uint8_t *alpha_b = NULL; // Composites always need scaling... defaulting to lowest - char *rescale = mlt_properties_get( a_props, "rescale.interp" ); + const char *rescale = mlt_properties_get( a_props, "rescale.interp" ); if ( rescale == NULL || !strcmp( rescale, "none" ) ) rescale = "nearest"; mlt_properties_set( a_props, "rescale.interp", rescale ); @@ -1021,13 +1153,14 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f // Do the calculation // NB: Locks needed here since the properties are being modified + int invert = mlt_properties_get_int( properties, "invert" ); mlt_service_lock( MLT_TRANSITION_SERVICE( this ) ); - composite_calculate( this, &result, a_frame, position ); + composite_calculate( this, &result, invert ? b_frame : a_frame, position ); mlt_service_unlock( MLT_TRANSITION_SERVICE( this ) ); // Since we are the consumer of the b_frame, we must pass along these // consumer properties from the a_frame - mlt_properties_set_double( b_props, "consumer_deinterlace", mlt_properties_get_double( a_props, "consumer_deinterlace" ) ); + mlt_properties_set_int( b_props, "consumer_deinterlace", mlt_properties_get_int( a_props, "consumer_deinterlace" ) || mlt_properties_get_int( properties, "deinterlace" ) ); mlt_properties_set( b_props, "consumer_deinterlace_method", mlt_properties_get( a_props, "consumer_deinterlace_method" ) ); mlt_properties_set_double( b_props, "consumer_aspect_ratio", mlt_properties_get_double( a_props, "consumer_aspect_ratio" ) ); @@ -1050,7 +1183,7 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f } // Get the image from the a frame - mlt_frame_get_image( a_frame, image, format, width, height, 1 ); + mlt_frame_get_image( a_frame, invert ? &image_b : image, format, width, height, 1 ); alpha_a = mlt_frame_get_alpha_mask( a_frame ); // Optimisation - no compositing required @@ -1080,7 +1213,7 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f height_b = mlt_properties_get_int( a_props, "dest_height" ); } - if ( *image != image_b && ( image_b != NULL || get_b_frame_image( this, b_frame, &image_b, &width_b, &height_b, &result ) == 0 ) ) + if ( *image != image_b && ( ( invert ? 0 : image_b ) || get_b_frame_image( this, b_frame, invert ? image : &image_b, &width_b, &height_b, &result ) == 0 ) ) { uint8_t *dest = *image; uint8_t *src = image_b; @@ -1089,8 +1222,10 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f mlt_properties_get_int( properties, "progressive" ); int field; - int32_t luma_softness = mlt_properties_get_double( properties, "softness" ) * ( 1 << 16 ); - uint16_t *luma_bitmap = get_luma( properties, width_b, height_b ); + double luma_softness = mlt_properties_get_double( properties, "softness" ); + mlt_service_lock( MLT_TRANSITION_SERVICE( this ) ); + uint16_t *luma_bitmap = get_luma( this, properties, width_b, height_b ); + mlt_service_unlock( MLT_TRANSITION_SERVICE( this ) ); char *operator = mlt_properties_get( properties, "operator" ); alpha_b = alpha_b == NULL ? mlt_frame_get_alpha_mask( b_frame ) : alpha_b; @@ -1123,24 +1258,38 @@ static int transition_get_image( mlt_frame a_frame, uint8_t **image, mlt_image_f // Do the calculation if we need to // NB: Locks needed here since the properties are being modified mlt_service_lock( MLT_TRANSITION_SERVICE( this ) ); - composite_calculate( this, &result, a_frame, field_position ); + composite_calculate( this, &result, invert ? b_frame : a_frame, field_position ); mlt_service_unlock( MLT_TRANSITION_SERVICE( this ) ); if ( mlt_properties_get_int( properties, "titles" ) ) { - result.item.w = rint( 0.5 + *width * ( result.item.w / result.nw ) ); + result.item.w = rint( *width * ( result.item.w / result.nw ) ); result.nw = result.item.w; - result.item.h = rint( 0.5 + *height * ( result.item.h / result.nh ) ); + result.item.h = rint( *height * ( result.item.h / result.nh ) ); result.nh = *height; result.sw = width_b; result.sh = height_b; } - // Align - alignment_calculate( &result ); + // Enforce cropping + if ( mlt_properties_get( properties, "crop" ) ) + { + if ( result.x_src == 0 ) + width_b = width_b > result.item.w ? result.item.w : width_b; + if ( result.y_src == 0 ) + height_b = height_b > result.item.h ? result.item.h : height_b; + } + else + { + // Otherwise, align + alignment_calculate( &result ); + } // Composite the b_frame on the a_frame - composite_yuv( dest, *width, *height, src, width_b, height_b, alpha_b, alpha_a, result, progressive ? -1 : field, luma_bitmap, luma_softness, line_fn ); + if ( invert ) + composite_yuv( dest, width_b, height_b, src, *width, *height, alpha_a, alpha_b, result, progressive ? -1 : field, luma_bitmap, luma_softness, line_fn ); + else + composite_yuv( dest, *width, *height, src, width_b, height_b, alpha_b, alpha_a, result, progressive ? -1 : field, luma_bitmap, luma_softness, line_fn ); } } } @@ -1181,7 +1330,7 @@ static mlt_frame composite_process( mlt_transition this, mlt_frame a_frame, mlt_ /** Constructor for the filter. */ -mlt_transition transition_composite_init( char *arg ) +mlt_transition transition_composite_init( mlt_profile profile, mlt_service_type type, const char *id, char *arg ) { mlt_transition this = calloc( sizeof( struct mlt_transition_s ), 1 ); if ( this != NULL && mlt_transition_init( this, NULL ) == 0 ) @@ -1194,8 +1343,14 @@ mlt_transition transition_composite_init( char *arg ) mlt_properties_set( properties, "start", arg != NULL ? arg : "0,0:100%x100%" ); // Default factory - mlt_properties_set( properties, "factory", "fezzik" ); + mlt_properties_set( properties, "factory", mlt_environment( "MLT_PRODUCER" ) ); + + // Use alignment (and hence alpha of b frame) + mlt_properties_set_int( properties, "aligned", 1 ); + // Default to progressive rendering + mlt_properties_set_int( properties, "progressive", 1 ); + // Inform apps and framework that this is a video only transition mlt_properties_set_int( properties, "_transition_type", 1 ); }