#include <string.h>
#include <pthread.h>
#include <limits.h>
-#include <dlfcn.h>
#if LIBAVUTIL_VERSION_INT < (50<<16)
#define PIX_FMT_RGB32 PIX_FMT_RGBA32
int max_frequency;
unsigned int invalid_pts_counter;
double resample_factor;
+ mlt_cache image_cache;
#ifdef VDPAU
struct
{
{
*format = mlt_image_rgb24a;
struct SwsContext *context = sws_getContext( width, height, pix_fmt,
- width, height, PIX_FMT_RGBA, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ width, height, PIX_FMT_RGBA, SWS_BILINEAR, NULL, NULL, NULL);
AVPicture output;
avpicture_fill( &output, buffer, PIX_FMT_RGBA, width, height );
sws_scale( context, frame->data, frame->linesize, 0, height,
else if ( *format == mlt_image_yuv420p )
{
struct SwsContext *context = sws_getContext( width, height, pix_fmt,
- width, height, PIX_FMT_YUV420P, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ width, height, PIX_FMT_YUV420P, SWS_BILINEAR, NULL, NULL, NULL);
AVPicture output;
output.data[0] = buffer;
output.data[1] = buffer + width * height;
else if ( *format == mlt_image_rgb24 )
{
struct SwsContext *context = sws_getContext( width, height, pix_fmt,
- width, height, PIX_FMT_RGB24, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ width, height, PIX_FMT_RGB24, SWS_BILINEAR, NULL, NULL, NULL);
AVPicture output;
avpicture_fill( &output, buffer, PIX_FMT_RGB24, width, height );
sws_scale( context, frame->data, frame->linesize, 0, height,
else if ( *format == mlt_image_rgb24a || *format == mlt_image_opengl )
{
struct SwsContext *context = sws_getContext( width, height, pix_fmt,
- width, height, PIX_FMT_RGBA, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ width, height, PIX_FMT_RGBA, SWS_BILINEAR, NULL, NULL, NULL);
AVPicture output;
avpicture_fill( &output, buffer, PIX_FMT_RGBA, width, height );
sws_scale( context, frame->data, frame->linesize, 0, height,
else
{
struct SwsContext *context = sws_getContext( width, height, pix_fmt,
- width, height, PIX_FMT_YUYV422, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ width, height, PIX_FMT_YUYV422, SWS_BILINEAR, NULL, NULL, NULL);
AVPicture output;
avpicture_fill( &output, buffer, PIX_FMT_YUYV422, width, height );
sws_scale( context, frame->data, frame->linesize, 0, height,
// Get the producer properties
mlt_properties properties = MLT_PRODUCER_PROPERTIES( producer );
- avformat_lock();
-
// Fetch the video format context
AVFormatContext *context = this->video_format;
// Get codec context
AVCodecContext *codec_context = stream->codec;
+ // Get the image cache
+ if ( ! this->image_cache && ! mlt_properties_get_int( properties, "noimagecache" ) )
+ this->image_cache = mlt_cache_init();
+ if ( this->image_cache )
+ {
+ mlt_cache_item item = mlt_cache_get( this->image_cache, (void*) position );
+ *buffer = mlt_cache_item_data( item, format );
+ if ( *buffer )
+ {
+ // Set the resolution
+ *width = codec_context->width;
+ *height = codec_context->height;
+ mlt_properties_set_int( frame_properties, "width", *width );
+ mlt_properties_set_int( frame_properties, "height", *height );
+
+ // Cache hit
+ int size;
+ switch ( *format )
+ {
+ case mlt_image_yuv420p:
+ size = *width * 3 * ( *height + 1 ) / 2;
+ break;
+ case mlt_image_rgb24:
+ size = *width * ( *height + 1 ) * 3;
+ break;
+ case mlt_image_rgb24a:
+ case mlt_image_opengl:
+ size = *width * ( *height + 1 ) * 4;
+ break;
+ default:
+ *format = mlt_image_yuv422;
+ size = *width * ( *height + 1 ) * 2;
+ break;
+ }
+ mlt_properties_set_data( frame_properties, "avformat.image_cache", item, 0, ( mlt_destructor )mlt_cache_item_close, NULL );
+ mlt_properties_set_data( frame_properties, "image", *buffer, size, NULL, NULL );
+ this->top_field_first = mlt_properties_get_int( frame_properties, "top_field_first" );
+ this->got_picture = 1;
+
+ goto exit_get_image;
+ }
+ }
+ // Cache miss
+ int image_size = 0;
+
+ avformat_lock();
+
// Packet
AVPacket pkt;
}
// Duplicate the last image if necessary (see comment on rawvideo below)
- if ( this->av_frame && this->got_picture && this->seekable
+ if ( this->av_frame && this->av_frame->linesize[0] && this->got_picture && this->seekable
&& ( paused
|| this->current_position == req_position
|| ( !use_new_seek && this->current_position > req_position ) ) )
{
// Duplicate it
- if ( allocate_buffer( frame_properties, codec_context, buffer, format, width, height ) )
+ if ( ( image_size = allocate_buffer( frame_properties, codec_context, buffer, format, width, height ) ) )
#ifdef VDPAU
if ( this->vdpau && this->vdpau->buffer )
{
else if ( context->start_time != AV_NOPTS_VALUE )
pts -= context->start_time;
int_position = ( int )( av_q2d( stream->time_base) * pts * source_fps + 0.1 );
- mlt_log_verbose( MLT_PRODUCER_SERVICE(producer), "got frame %d, key %d\n", int_position, this->av_frame->key_frame );
+ mlt_log_debug( MLT_PRODUCER_SERVICE(producer), "got frame %d, key %d\n", int_position, this->av_frame->key_frame );
}
// Handle ignore
if ( int_position < req_position )
// Now handle the picture if we have one
if ( got_picture )
{
- if ( allocate_buffer( frame_properties, codec_context, buffer, format, width, height ) )
+ if ( ( image_size = allocate_buffer( frame_properties, codec_context, buffer, format, width, height ) ) )
{
#ifdef VDPAU
if ( this->vdpau )
}
}
}
- if ( !got_picture )
- mlt_frame_get_image( frame, buffer, format, width, height, writable );
}
- // Very untidy - for rawvideo, the packet contains the frame, hence the free packet
- // above will break the pause behaviour - so we wipe the frame now
- if ( !strcmp( codec_context->codec->name, "rawvideo" ) )
- av_freep( &this->av_frame );
-
avformat_unlock();
+ if ( this->got_picture && image_size > 0 && this->image_cache )
+ {
+ // Copy buffer to image cache
+ uint8_t *image = mlt_pool_alloc( image_size );
+ memcpy( image, *buffer, image_size );
+ mlt_cache_put( this->image_cache, (void*) position, image, *format, mlt_pool_release );
+ }
+
+exit_get_image:
// Set the progressive flag
- mlt_properties_set_int( frame_properties, "progressive",
- !this->av_frame->interlaced_frame || !!mlt_properties_get_int( properties, "force_progressive" ) );
+ if ( mlt_properties_get( properties, "force_progressive" ) )
+ mlt_properties_set_int( frame_properties, "progressive", !!mlt_properties_get_int( properties, "force_progressive" ) );
+ else if ( this->av_frame )
+ mlt_properties_set_int( frame_properties, "progressive", !this->av_frame->interlaced_frame );
// Set the field order property for this frame
mlt_properties_set_int( frame_properties, "top_field_first", this->top_field_first );
// Regardless of speed, we expect to get the next frame (cos we ain't too bright)
this->video_expected = position + 1;
- return 0;
+ return !this->got_picture;
}
/** Process properties as AVOptions and apply to AV context obj
{
const char *opt_name = mlt_properties_get_name( properties, i );
const AVOption *opt = av_find_opt( obj, opt_name, NULL, flags, flags );
- if ( opt )
+ if ( opt_name && mlt_properties_get( properties, opt_name ) )
+ {
+ if ( opt )
#if LIBAVCODEC_VERSION_INT >= ((52<<16)+(7<<8)+0)
- av_set_string3( obj, opt_name, mlt_properties_get( properties, opt_name), 0, NULL );
+ av_set_string3( obj, opt_name, mlt_properties_get( properties, opt_name), 0, NULL );
#elif LIBAVCODEC_VERSION_INT >= ((51<<16)+(59<<8)+0)
- av_set_string2( obj, opt_name, mlt_properties_get( properties, opt_name), 0 );
+ av_set_string2( obj, opt_name, mlt_properties_get( properties, opt_name), 0 );
#else
- av_set_string( obj, opt_name, mlt_properties_get( properties, opt_name) );
+ av_set_string( obj, opt_name, mlt_properties_get( properties, opt_name) );
#endif
+ }
}
}
AVPacket pkt;
av_init_packet( &pkt );
+
+ // If not resampling, give consumer more than requested.
+ // It requested number samples based on requested frame rate.
+ // Do not clean this up with a samples *= ...!
+ if ( this->audio_index != INT_MAX && ! this->audio_resample[ this->audio_index ] )
+ *samples = *samples * this->audio_codec[ this->audio_index ]->sample_rate / *frequency;
while ( ret >= 0 && !got_audio )
{
int index = mlt_properties_get_int( properties, "audio_index" );
// Handle all audio tracks
- if ( mlt_properties_get( properties, "audio_index" ) &&
- !strcmp( mlt_properties_get( properties, "audio_index" ), "all" ) )
+ if ( this->audio_index > -1 &&
+ mlt_properties_get( properties, "audio_index" ) &&
+ !strcmp( mlt_properties_get( properties, "audio_index" ), "all" ) )
index = INT_MAX;
// Reopen the file if necessary
- if ( !context && index > -1 )
+ if ( !context && this->audio_index > -1 && index > -1 )
{
mlt_events_block( properties, producer );
producer_open( this, mlt_service_profile( MLT_PRODUCER_SERVICE(producer) ),
if ( context && index > -1 && index < INT_MAX &&
context->streams[ index ]->codec->codec_type != CODEC_TYPE_AUDIO )
{
- index = -1;
+ index = this->audio_index;
mlt_properties_set_int( properties, "audio_index", index );
}
// Update the audio properties if the index changed
- if ( index > -1 && index != this->audio_index )
+ if ( context && index > -1 && index != this->audio_index )
{
if ( this->audio_codec[ this->audio_index ] )
{
}
this->audio_codec[ this->audio_index ] = NULL;
}
- this->audio_index = index;
+ if ( this->audio_index != -1 )
+ this->audio_index = index;
+ else
+ index = -1;
// Get the codec(s)
if ( context && index == INT_MAX )
static int producer_get_frame( mlt_producer producer, mlt_frame_ptr frame, int index )
{
// Access the private data
- mlt_cache_item cache_item = mlt_service_cache_get( MLT_PRODUCER_SERVICE(producer), "producer_avformat" );
+ mlt_service service = MLT_PRODUCER_SERVICE( producer );
+ mlt_cache_item cache_item = mlt_service_cache_get( service, "producer_avformat" );
producer_avformat this = mlt_cache_item_data( cache_item, NULL );
+ // If cache miss
+ if ( !this )
+ {
+ this = calloc( 1, sizeof( struct producer_avformat_s ) );
+ producer->child = this;
+ this->parent = producer;
+ mlt_service_cache_put( service, "producer_avformat", this, 0, (mlt_destructor) producer_avformat_close );
+ cache_item = mlt_service_cache_get( service, "producer_avformat" );
+ }
+
// Create an empty frame
- *frame = mlt_frame_init( MLT_PRODUCER_SERVICE( producer ) );
+ *frame = mlt_frame_init( service);
if ( *frame )
+ {
mlt_properties_set_data( MLT_FRAME_PROPERTIES(*frame), "avformat_cache", cache_item, 0, (mlt_destructor) mlt_cache_item_close, NULL );
+ }
else
+ {
mlt_cache_item_close( cache_item );
+ return 1;
+ }
// Update timecode on the frame we're creating
mlt_frame_set_position( *frame, mlt_producer_position( producer ) );
// Set the position of this producer
mlt_properties_set_position( MLT_FRAME_PROPERTIES( *frame ), "avformat_position", mlt_producer_frame( producer ) );
- // If cache miss
- if ( !this )
- {
- this = calloc( 1, sizeof( struct producer_avformat_s ) );
- producer->child = this;
- this->parent = producer;
- mlt_service_cache_put( MLT_PRODUCER_SERVICE(producer), "producer_avformat", this, 0, (mlt_destructor) producer_avformat_close );
- }
-
// Set up the video
producer_set_up_video( this, *frame );
#ifdef VDPAU
vdpau_producer_close( this );
#endif
+ if ( this->image_cache )
+ mlt_cache_close( this->image_cache );
free( this );
}