X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=src%2Fmodules%2Favformat%2Fproducer_avformat.c;h=c0b226ef433880db0f34385ad98bfbd7425ed1c3;hb=02a0ab41399ce94753e94d1a2229e43d09370463;hp=85cf346a4e91fa4444ccb09fa5d12f3dda723aab;hpb=9b64e669afff2b5212f0002702a4d58e3239d25c;p=mlt diff --git a/src/modules/avformat/producer_avformat.c b/src/modules/avformat/producer_avformat.c index 85cf346a..c0b226ef 100644 --- a/src/modules/avformat/producer_avformat.c +++ b/src/modules/avformat/producer_avformat.c @@ -1,6 +1,6 @@ /* * producer_avformat.c -- avformat producer - * Copyright (C) 2003-2012 Ushodaya Enterprises Limited + * Copyright (C) 2003-2014 Ushodaya Enterprises Limited * Author: Charles Yates * Author: Dan Dennedy * Much code borrowed from ffmpeg.c: Copyright (c) 2000-2003 Fabrice Bellard @@ -31,27 +31,13 @@ // ffmpeg Header files #include -#ifdef SWSCALE -# include -#endif - -#if LIBAVUTIL_VERSION_INT >= ((50<<16)+(38<<8)+0) -# include -#else -# define AV_SAMPLE_FMT_S16 SAMPLE_FMT_S16 -# define AV_SAMPLE_FMT_S32 SAMPLE_FMT_S32 -# define AV_SAMPLE_FMT_FLT SAMPLE_FMT_FLT -# if (LIBAVCODEC_VERSION_INT >= ((51<<16)+(71<<8)+0)) -const char *avcodec_get_sample_fmt_name(int sample_fmt); -# endif -#endif +#include +#include +#include #ifdef VDPAU # include #endif -#if (LIBAVUTIL_VERSION_INT > ((50<<16)+(7<<8)+0)) -# include -#endif #if (LIBAVUTIL_VERSION_INT >= ((51<<16)+(8<<8)+0)) # include #endif @@ -62,11 +48,6 @@ const char *avcodec_get_sample_fmt_name(int sample_fmt); #include #include -#if LIBAVUTIL_VERSION_INT < (50<<16) -#define PIX_FMT_RGB32 PIX_FMT_RGBA32 -#define PIX_FMT_YUYV422 PIX_FMT_YUV422 -#endif - #if LIBAVCODEC_VERSION_MAJOR >= 53 #include #define CODEC_TYPE_VIDEO AVMEDIA_TYPE_VIDEO @@ -76,11 +57,17 @@ const char *avcodec_get_sample_fmt_name(int sample_fmt); #include #endif +#if LIBAVCODEC_VERSION_MAJOR < 55 +#define AV_CODEC_ID_DVVIDEO CODEC_ID_DVVIDEO +#define AV_CODEC_ID_H264 CODEC_ID_H264 +#endif + #define POSITION_INITIAL (-2) #define POSITION_INVALID (-1) -#define MAX_AUDIO_STREAMS (10) +#define MAX_AUDIO_STREAMS (32) #define MAX_VDPAU_SURFACES (10) +#define MAX_AUDIO_FRAME_SIZE (192000) // 1 second of 48khz 32bit audio struct producer_avformat_s { @@ -90,8 +77,9 @@ struct producer_avformat_s AVFormatContext *video_format; AVCodecContext *audio_codec[ MAX_AUDIO_STREAMS ]; AVCodecContext *video_codec; - AVFrame *av_frame; - ReSampleContext *audio_resample[ MAX_AUDIO_STREAMS ]; + AVFrame *video_frame; + AVFrame *audio_frame; + AVPacket pkt; mlt_position audio_expected; mlt_position video_expected; int audio_index; @@ -112,15 +100,18 @@ struct producer_avformat_s int max_channel; int max_frequency; unsigned int invalid_pts_counter; - double resample_factor; + unsigned int invalid_dts_counter; mlt_cache image_cache; - int colorspace; + int yuv_colorspace, color_primaries; + int full_luma; pthread_mutex_t video_mutex; pthread_mutex_t audio_mutex; mlt_deque apackets; mlt_deque vpackets; pthread_mutex_t packets_mutex; pthread_mutex_t open_mutex; + int is_mutex_init; + AVRational video_time_base; #ifdef VDPAU struct { @@ -152,6 +143,7 @@ static void producer_set_up_audio( producer_avformat self, mlt_frame frame ); static void apply_properties( void *obj, mlt_properties properties, int flags ); static int video_codec_init( producer_avformat self, int index, mlt_properties properties ); static void get_audio_streams_info( producer_avformat self ); +static mlt_audio_format pick_audio_format( int sample_fmt ); #ifdef VDPAU #include "vdpau.c" @@ -191,16 +183,6 @@ mlt_producer producer_avformat_init( mlt_profile profile, const char *service, c // Register our get_frame implementation producer->get_frame = producer_get_frame; - // init mutexes - pthread_mutex_init( &self->audio_mutex, NULL ); - pthread_mutex_init( &self->video_mutex, NULL ); - pthread_mutex_init( &self->packets_mutex, NULL ); - pthread_mutex_init( &self->open_mutex, NULL ); - - // init queues - self->apackets = mlt_deque_init(); - self->vpackets = mlt_deque_init(); - if ( strcmp( service, "avformat-novalidate" ) ) { // Open the file @@ -282,6 +264,23 @@ int list_components( char* file ) return skip; } +static int first_video_index( producer_avformat self ) +{ + AVFormatContext *context = self->video_format? self->video_format : self->audio_format; + int i = -1; // not found + + if ( context ) { + for ( i = 0; i < context->nb_streams; i++ ) { + if ( context->streams[i]->codec && + context->streams[i]->codec->codec_type == CODEC_TYPE_VIDEO ) + break; + } + if ( i == context->nb_streams ) + i = -1; + } + return i; +} + /** Find the default streams. */ @@ -325,26 +324,22 @@ static mlt_properties find_default_streams( producer_avformat self ) self->video_index = i; mlt_properties_set( meta_media, key, "video" ); snprintf( key, sizeof(key), "meta.media.%d.stream.frame_rate", i ); -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(42<<8)+0) double ffmpeg_fps = av_q2d( context->streams[ i ]->avg_frame_rate ); +#if LIBAVFORMAT_VERSION_MAJOR < 55 if ( isnan( ffmpeg_fps ) || ffmpeg_fps == 0 ) ffmpeg_fps = av_q2d( context->streams[ i ]->r_frame_rate ); - mlt_properties_set_double( meta_media, key, ffmpeg_fps ); -#else - mlt_properties_set_double( meta_media, key, av_q2d( context->streams[ i ]->r_frame_rate ) ); #endif + mlt_properties_set_double( meta_media, key, ffmpeg_fps ); -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(21<<8)+0) snprintf( key, sizeof(key), "meta.media.%d.stream.sample_aspect_ratio", i ); mlt_properties_set_double( meta_media, key, av_q2d( context->streams[ i ]->sample_aspect_ratio ) ); -#endif snprintf( key, sizeof(key), "meta.media.%d.codec.width", i ); mlt_properties_set_int( meta_media, key, codec_context->width ); snprintf( key, sizeof(key), "meta.media.%d.codec.height", i ); mlt_properties_set_int( meta_media, key, codec_context->height ); snprintf( key, sizeof(key), "meta.media.%d.codec.frame_rate", i ); - mlt_properties_set_double( meta_media, key, (double) codec_context->time_base.den / - ( codec_context->time_base.num == 0 ? 1 : codec_context->time_base.num ) ); + AVRational frame_rate = { codec_context->time_base.den, codec_context->time_base.num * codec_context->ticks_per_frame }; + mlt_properties_set_double( meta_media, key, av_q2d( frame_rate ) ); snprintf( key, sizeof(key), "meta.media.%d.codec.pix_fmt", i ); #if LIBAVUTIL_VERSION_INT >= ((51<<16)+(3<<8)+0) mlt_properties_set( meta_media, key, av_get_pix_fmt_name( codec_context->pix_fmt ) ); @@ -353,7 +348,6 @@ static mlt_properties find_default_streams( producer_avformat self ) #endif snprintf( key, sizeof(key), "meta.media.%d.codec.sample_aspect_ratio", i ); mlt_properties_set_double( meta_media, key, av_q2d( codec_context->sample_aspect_ratio ) ); -#if LIBAVCODEC_VERSION_INT > ((52<<16)+(28<<8)+0) snprintf( key, sizeof(key), "meta.media.%d.codec.colorspace", i ); switch ( codec_context->colorspace ) { @@ -372,22 +366,17 @@ static mlt_properties find_default_streams( producer_avformat self ) mlt_properties_set_int( meta_media, key, codec_context->width * codec_context->height > 750000 ? 709 : 601 ); break; } -#endif break; case CODEC_TYPE_AUDIO: if ( !codec_context->channels ) break; // Use first audio stream - if ( self->audio_index < 0 ) + if ( self->audio_index < 0 && pick_audio_format( codec_context->sample_fmt ) != mlt_audio_none ) self->audio_index = i; + mlt_properties_set( meta_media, key, "audio" ); -#if LIBAVUTIL_VERSION_INT >= ((50<<16)+(38<<8)+0) snprintf( key, sizeof(key), "meta.media.%d.codec.sample_fmt", i ); mlt_properties_set( meta_media, key, av_get_sample_fmt_name( codec_context->sample_fmt ) ); -#elif (LIBAVCODEC_VERSION_INT >= ((51<<16)+(71<<8)+0)) - snprintf( key, sizeof(key), "meta.media.%d.codec.sample_fmt", i ); - mlt_properties_set( meta_media, key, avcodec_get_sample_fmt_name( codec_context->sample_fmt ) ); -#endif snprintf( key, sizeof(key), "meta.media.%d.codec.sample_rate", i ); mlt_properties_set_int( meta_media, key, codec_context->sample_rate ); snprintf( key, sizeof(key), "meta.media.%d.codec.channels", i ); @@ -400,10 +389,8 @@ static mlt_properties find_default_streams( producer_avformat self ) // mlt_properties_set_double( meta_media, key, av_q2d( context->streams[ i ]->time_base ) ); snprintf( key, sizeof(key), "meta.media.%d.codec.name", i ); mlt_properties_set( meta_media, key, codec->name ); -#if (LIBAVCODEC_VERSION_INT >= ((51<<16)+(55<<8)+0)) snprintf( key, sizeof(key), "meta.media.%d.codec.long_name", i ); mlt_properties_set( meta_media, key, codec->long_name ); -#endif snprintf( key, sizeof(key), "meta.media.%d.codec.bit_rate", i ); mlt_properties_set_int( meta_media, key, codec_context->bit_rate ); // snprintf( key, sizeof(key), "meta.media.%d.codec.time_base", i ); @@ -414,7 +401,6 @@ static mlt_properties find_default_streams( producer_avformat self ) // mlt_properties_set_int( meta_media, key, codec_context->level ); // Read Metadata -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(31<<8)+0) #if LIBAVUTIL_VERSION_INT >= ((51<<16)+(8<<8)+0) while ( ( tag = av_dict_get( stream->metadata, "", tag, AV_DICT_IGNORE_SUFFIX ) ) ) #else @@ -427,9 +413,7 @@ static mlt_properties find_default_streams( producer_avformat self ) mlt_properties_set( meta_media, key, tag->value ); } } -#endif } -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(31<<8)+0) #if LIBAVUTIL_VERSION_INT >= ((51<<16)+(8<<8)+0) while ( ( tag = av_dict_get( context->metadata, "", tag, AV_DICT_IGNORE_SUFFIX ) ) ) #else @@ -442,151 +426,20 @@ static mlt_properties find_default_streams( producer_avformat self ) mlt_properties_set( meta_media, key, tag->value ); } } -#else - if ( context->title && strcmp( context->title, "" ) ) - mlt_properties_set(properties, "meta.attr.title.markup", context->title ); - if ( context->author && strcmp( context->author, "" ) ) - mlt_properties_set(properties, "meta.attr.author.markup", context->author ); - if ( context->copyright && strcmp( context->copyright, "" ) ) - mlt_properties_set(properties, "meta.attr.copyright.markup", context->copyright ); - if ( context->comment ) - mlt_properties_set(properties, "meta.attr.comment.markup", context->comment ); - if ( context->album ) - mlt_properties_set(properties, "meta.attr.album.markup", context->album ); - if ( context->year ) - mlt_properties_set_int(properties, "meta.attr.year.markup", context->year ); - if ( context->track ) - mlt_properties_set_int(properties, "meta.attr.track.markup", context->track ); -#endif return meta_media; } -static inline int dv_is_pal( AVPacket *pkt ) +static void get_aspect_ratio( mlt_properties properties, AVStream *stream, AVCodecContext *codec_context ) { - return pkt->data[3] & 0x80; -} - -static int dv_is_wide( AVPacket *pkt ) -{ - int i = 80 /* block size */ *3 /* VAUX starts at block 3 */ +3 /* skip block header */; - - for ( ; i < pkt->size; i += 5 /* packet size */ ) - { - if ( pkt->data[ i ] == 0x61 ) - { - uint8_t x = pkt->data[ i + 2 ] & 0x7; - return ( x == 2 ) || ( x == 7 ); - } - } - return 0; -} - -static double get_aspect_ratio( mlt_properties properties, AVStream *stream, AVCodecContext *codec_context, AVPacket *pkt ) -{ - double aspect_ratio = 1.0; - - if ( codec_context->codec_id == CODEC_ID_DVVIDEO ) - { - if ( pkt ) - { - if ( dv_is_pal( pkt ) ) - { - if ( dv_is_wide( pkt ) ) - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 64 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 45 ); - } - else - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 16 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 15 ); - } - } - else - { - if ( dv_is_wide( pkt ) ) - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 32 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 27 ); - } - else - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 8 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 9 ); - } - } - } - else - { - AVRational ar = -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(21<<8)+0) - stream->sample_aspect_ratio; -#else - codec_context->sample_aspect_ratio; -#endif - // Override FFmpeg's notion of DV aspect ratios, which are - // based upon a width of 704. Since we do not have a normaliser - // that crops (nor is cropping 720 wide ITU-R 601 video always desirable) - // we just coerce the values to facilitate a passive behaviour through - // the rescale normaliser when using equivalent producers and consumers. - // = display_aspect / (width * height) - if ( ar.num == 10 && ar.den == 11 ) - { - // 4:3 NTSC - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 8 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 9 ); - } - else if ( ar.num == 59 && ar.den == 54 ) - { - // 4:3 PAL - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 16 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 15 ); - } - else if ( ar.num == 40 && ar.den == 33 ) - { - // 16:9 NTSC - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 32 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 27 ); - } - else if ( ar.num == 118 && ar.den == 81 ) - { - // 16:9 PAL - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 64 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 45 ); - } - } - } - else - { - AVRational codec_sar = codec_context->sample_aspect_ratio; - AVRational stream_sar = -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(21<<8)+0) - stream->sample_aspect_ratio; -#else - { 0, 1 }; -#endif - if ( codec_sar.num > 0 ) - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", codec_sar.num ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", codec_sar.den ); - } - else if ( stream_sar.num > 0 ) - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", stream_sar.num ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", stream_sar.den ); - } - else - { - mlt_properties_set_int( properties, "meta.media.sample_aspect_num", 1 ); - mlt_properties_set_int( properties, "meta.media.sample_aspect_den", 1 ); - } - } - AVRational ar = { mlt_properties_get_double( properties, "meta.media.sample_aspect_num" ), mlt_properties_get_double( properties, "meta.media.sample_aspect_den" ) }; - aspect_ratio = av_q2d( ar ); - mlt_properties_set_double( properties, "aspect_ratio", aspect_ratio ); - - return aspect_ratio; + AVRational sar = stream->sample_aspect_ratio; + if ( sar.num <= 0 || sar.den <= 0 ) + sar = codec_context->sample_aspect_ratio; + if ( sar.num <= 0 || sar.den <= 0 ) + sar.num = sar.den = 1; + mlt_properties_set_int( properties, "meta.media.sample_aspect_num", sar.num ); + mlt_properties_set_int( properties, "meta.media.sample_aspect_den", sar.den ); + mlt_properties_set_double( properties, "aspect_ratio", av_q2d( sar ) ); } #if LIBAVFORMAT_VERSION_INT > ((53<<16)+(6<<8)+0) @@ -675,10 +528,8 @@ static char* parse_url( mlt_profile profile, const char* URL, AVInputFormat **fo params->channel = atoi( value ); else if ( !strcmp( name, "channels" ) ) params->channels = atoi( value ); -#if (LIBAVUTIL_VERSION_INT > ((50<<16)+(7<<8)+0)) else if ( !strcmp( name, "pix_fmt" ) ) params->pix_fmt = av_get_pix_fmt( value ); -#endif else if ( !strcmp( name, "width" ) ) params->width = atoi( value ); else if ( !strcmp( name, "height" ) ) @@ -724,10 +575,6 @@ static int get_basic_info( producer_avformat self, mlt_profile profile, const ch AVFormatContext *format = self->video_format; - // We will treat everything with the producer fps. - // TODO: make this more flexible. - double fps = mlt_profile_fps( profile ); - // Get the duration if ( !mlt_properties_get_int( properties, "_length_computed" ) ) { @@ -736,7 +583,9 @@ static int get_basic_info( producer_avformat self, mlt_profile profile, const ch if ( format->duration != AV_NOPTS_VALUE ) { // This isn't going to be accurate for all formats - mlt_position frames = ( mlt_position )( ( ( double )format->duration / ( double )AV_TIME_BASE ) * fps ); + // We will treat everything with the producer fps. + mlt_position frames = ( mlt_position )( int )( format->duration * + profile->frame_rate_num / profile->frame_rate_den / AV_TIME_BASE); mlt_properties_set_position( properties, "out", frames - 1 ); mlt_properties_set_position( properties, "length", frames ); mlt_properties_set_int( properties, "_length_computed", 1 ); @@ -788,31 +637,8 @@ static int get_basic_info( producer_avformat self, mlt_profile profile, const ch AVCodecContext *codec_context = format->streams[ self->video_index ]->codec; mlt_properties_set_int( properties, "width", codec_context->width ); mlt_properties_set_int( properties, "height", codec_context->height ); + get_aspect_ratio( properties, format->streams[ self->video_index ], codec_context ); - if ( codec_context->codec_id == CODEC_ID_DVVIDEO ) - { - // Fetch the first frame of DV so we can read it directly - AVPacket pkt; - int ret = 0; - while ( ret >= 0 ) - { - ret = av_read_frame( format, &pkt ); - if ( ret >= 0 && pkt.stream_index == self->video_index && pkt.size > 0 ) - { - get_aspect_ratio( properties, format->streams[ self->video_index ], codec_context, &pkt ); - av_free_packet(&pkt); - break; - } - if ( ret >= 0 ) - av_free_packet(&pkt); - } - } - else - { - get_aspect_ratio( properties, format->streams[ self->video_index ], codec_context, NULL ); - } - -#ifdef SWSCALE // Verify that we can convert this to YUV 4:2:2 // TODO: we can now also return RGB and RGBA and quite possibly more in the future. struct SwsContext *context = sws_getContext( codec_context->width, codec_context->height, codec_context->pix_fmt, @@ -821,7 +647,6 @@ static int get_basic_info( producer_avformat self, mlt_profile profile, const ch sws_freeContext( context ); else error = 1; -#endif } return error; } @@ -835,6 +660,15 @@ static int producer_open( producer_avformat self, mlt_profile profile, const cha int error = 0; mlt_properties properties = MLT_PRODUCER_PROPERTIES( self->parent ); + if ( !self->is_mutex_init ) + { + pthread_mutex_init( &self->audio_mutex, NULL ); + pthread_mutex_init( &self->video_mutex, NULL ); + pthread_mutex_init( &self->packets_mutex, NULL ); + pthread_mutex_init( &self->open_mutex, NULL ); + self->is_mutex_init = 1; + } + // Lock the service if ( take_lock ) { @@ -870,10 +704,8 @@ static int producer_open( producer_avformat self, mlt_profile profile, const cha if ( !error && self->video_format ) { apply_properties( self->video_format, properties, AV_OPT_FLAG_DECODING_PARAM ); -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(110<<8)+0) if ( self->video_format->iformat && self->video_format->iformat->priv_class && self->video_format->priv_data ) apply_properties( self->video_format->priv_data, properties, AV_OPT_FLAG_DECODING_PARAM ); -#endif } #if LIBAVFORMAT_VERSION_INT > ((53<<16)+(6<<8)+0) @@ -954,6 +786,11 @@ static int producer_open( producer_avformat self, mlt_profile profile, const cha } if ( filename ) free( filename ); + if ( !error ) + { + self->apackets = mlt_deque_init(); + self->vpackets = mlt_deque_init(); + } if ( self->dummy_context ) { @@ -978,50 +815,106 @@ static int producer_open( producer_avformat self, mlt_profile profile, const cha return error; } -static void reopen_video( producer_avformat self, mlt_producer producer ) +static void prepare_reopen( producer_avformat self ) { - mlt_properties properties = MLT_PRODUCER_PROPERTIES( producer ); - mlt_service_lock( MLT_PRODUCER_SERVICE( producer ) ); + mlt_service_lock( MLT_PRODUCER_SERVICE( self->parent ) ); pthread_mutex_lock( &self->audio_mutex ); pthread_mutex_lock( &self->open_mutex ); + int i; + for ( i = 0; i < MAX_AUDIO_STREAMS; i++ ) + { + mlt_pool_release( self->audio_buffer[i] ); + self->audio_buffer[i] = NULL; + av_free( self->decode_buffer[i] ); + self->decode_buffer[i] = NULL; + if ( self->audio_codec[i] ) + avcodec_close( self->audio_codec[i] ); + self->audio_codec[i] = NULL; + } if ( self->video_codec ) avcodec_close( self->video_codec ); self->video_codec = NULL; + #if LIBAVFORMAT_VERSION_INT >= ((53<<16)+(17<<8)+0) - if ( self->dummy_context ) - avformat_close_input( &self->dummy_context ); + if ( self->seekable && self->audio_format ) + avformat_close_input( &self->audio_format ); if ( self->video_format ) avformat_close_input( &self->video_format ); #else - if ( self->dummy_context ) - av_close_input_file( self->dummy_context ); + if ( self->seekable && self->audio_format ) + av_close_input_file( self->audio_format ); if ( self->video_format ) av_close_input_file( self->video_format ); #endif - self->dummy_context = NULL; + self->audio_format = NULL; self->video_format = NULL; pthread_mutex_unlock( &self->open_mutex ); - int audio_index = self->audio_index; - int video_index = self->video_index; + // Cleanup the packet queues + AVPacket *pkt; + if ( self->apackets ) + { + while ( ( pkt = mlt_deque_pop_back( self->apackets ) ) ) + { + av_free_packet( pkt ); + free( pkt ); + } + mlt_deque_close( self->apackets ); + self->apackets = NULL; + } + if ( self->vpackets ) + { + while ( ( pkt = mlt_deque_pop_back( self->vpackets ) ) ) + { + av_free_packet( pkt ); + free( pkt ); + } + mlt_deque_close( self->vpackets ); + self->vpackets = NULL; + } + pthread_mutex_unlock( &self->audio_mutex ); + mlt_service_unlock( MLT_PRODUCER_SERVICE( self->parent ) ); +} + +static int64_t best_pts( producer_avformat self, int64_t pts, int64_t dts ) +{ + self->invalid_pts_counter += pts == AV_NOPTS_VALUE; + self->invalid_dts_counter += dts == AV_NOPTS_VALUE; + if ( ( self->invalid_pts_counter <= self->invalid_dts_counter + || dts == AV_NOPTS_VALUE ) && pts != AV_NOPTS_VALUE ) + return pts; + else + return dts; +} - producer_open( self, mlt_service_profile( MLT_PRODUCER_SERVICE(producer) ), - mlt_properties_get( properties, "resource" ), 0 ); +static void find_first_pts( producer_avformat self, int video_index ) +{ + // find initial PTS + AVFormatContext *context = self->video_format? self->video_format : self->audio_format; + int ret = 0; + int toscan = 500; + AVPacket pkt; - self->audio_index = audio_index; - if ( self->video_format && video_index > -1 ) + while ( ret >= 0 && toscan-- > 0 ) { - self->video_index = video_index; - video_codec_init( self, video_index, properties ); + ret = av_read_frame( context, &pkt ); + if ( ret >= 0 && pkt.stream_index == video_index && ( pkt.flags & PKT_FLAG_KEY ) ) + { + mlt_log_debug( MLT_PRODUCER_SERVICE(self->parent), + "first_pts %"PRId64" dts %"PRId64" pts_dts_delta %d\n", + pkt.pts, pkt.dts, (int)(pkt.pts - pkt.dts) ); + self->first_pts = best_pts( self, pkt.pts, pkt.dts ); + if ( self->first_pts != AV_NOPTS_VALUE ) + toscan = 0; + } + av_free_packet( &pkt ); } - - pthread_mutex_unlock( &self->audio_mutex ); - mlt_service_unlock( MLT_PRODUCER_SERVICE( producer ) ); + av_seek_frame( context, -1, 0, AVSEEK_FLAG_BACKWARD ); } static int seek_video( producer_avformat self, mlt_position position, - int64_t req_position, int preseek, int use_pts ) + int64_t req_position, int preseek ) { mlt_producer producer = self->parent; int paused = 0; @@ -1043,31 +936,10 @@ static int seek_video( producer_avformat self, mlt_position position, double source_fps = mlt_properties_get_double( properties, "meta.media.frame_rate_num" ) / mlt_properties_get_double( properties, "meta.media.frame_rate_den" ); - // find initial PTS if ( self->last_position == POSITION_INITIAL ) - { - int ret = 0; - int toscan = 100; - AVPacket pkt; - - while ( ret >= 0 && toscan-- > 0 ) - { - ret = av_read_frame( context, &pkt ); - if ( ret >= 0 && pkt.stream_index == self->video_index ) - { - mlt_log_debug( MLT_PRODUCER_SERVICE(producer), - "first_pts %"PRId64" dts %"PRId64" pts_dts_delta %d\n", - pkt.pts, pkt.dts, (int)(pkt.pts - pkt.dts) ); - self->first_pts = use_pts? pkt.pts : pkt.dts; - if ( self->first_pts != AV_NOPTS_VALUE ) - toscan = 0; - } - av_free_packet( &pkt ); - } - av_seek_frame( context, -1, 0, AVSEEK_FLAG_BACKWARD ); - } + find_first_pts( self, self->video_index ); - if ( self->av_frame && position + 1 == self->video_expected ) + if ( self->video_frame && position + 1 == self->video_expected ) { // We're paused - use last image paused = 1; @@ -1075,42 +947,31 @@ static int seek_video( producer_avformat self, mlt_position position, else if ( self->seekable && ( position < self->video_expected || position - self->video_expected >= 12 || self->last_position < 0 ) ) { // Calculate the timestamp for the requested frame - int64_t timestamp = req_position / ( av_q2d( stream->time_base ) * source_fps ); + int64_t timestamp = req_position / ( av_q2d( self->video_time_base ) * source_fps ); if ( req_position <= 0 ) timestamp = 0; else if ( self->first_pts != AV_NOPTS_VALUE ) timestamp += self->first_pts; else if ( context->start_time != AV_NOPTS_VALUE ) timestamp += context->start_time; - if ( preseek && av_q2d( stream->time_base ) != 0 ) - timestamp -= 2 / av_q2d( stream->time_base ); + if ( preseek && av_q2d( self->video_time_base ) != 0 ) + timestamp -= 2 / av_q2d( self->video_time_base ); if ( timestamp < 0 ) timestamp = 0; - mlt_log_debug( MLT_PRODUCER_SERVICE(producer), "seeking timestamp %"PRId64" position %d expected %d last_pos %"PRId64"\n", + mlt_log_debug( MLT_PRODUCER_SERVICE(producer), "seeking timestamp %"PRId64" position " MLT_POSITION_FMT " expected "MLT_POSITION_FMT" last_pos %"PRId64"\n", timestamp, position, self->video_expected, self->last_position ); // Seek to the timestamp - // NOTE: reopen_video is disabled at this time because it is causing trouble with A/V sync. - if ( 1 || use_pts || req_position > 0 || self->last_position <= 0 ) - { - codec_context->skip_loop_filter = AVDISCARD_NONREF; - av_seek_frame( context, self->video_index, timestamp, AVSEEK_FLAG_BACKWARD ); + codec_context->skip_loop_filter = AVDISCARD_NONREF; + av_seek_frame( context, self->video_index, timestamp, AVSEEK_FLAG_BACKWARD ); - // flush any pictures still in decode buffer - avcodec_flush_buffers( codec_context ); - } - else - { - // Re-open video stream when rewinding to beginning from somewhere else. - // This is rather ugly, and I prefer not to do it this way, but ffmpeg is - // not reliably seeking to the first frame across formats. - reopen_video( self, producer ); - } + // flush any pictures still in decode buffer + avcodec_flush_buffers( codec_context ); // Remove the cached info relating to the previous position self->current_position = POSITION_INVALID; self->last_position = POSITION_INVALID; - av_freep( &self->av_frame ); + av_freep( &self->video_frame ); } } return paused; @@ -1163,50 +1024,55 @@ static void get_audio_streams_info( producer_avformat self ) } mlt_log_verbose( NULL, "[producer avformat] audio: total_streams %d max_stream %d total_channels %d max_channels %d\n", self->audio_streams, self->audio_max_stream, self->total_channels, self->max_channel ); - - // Other audio-specific initializations - self->resample_factor = 1.0; } -static void set_luma_transfer( struct SwsContext *context, int colorspace, int use_full_range ) +static int set_luma_transfer( struct SwsContext *context, int src_colorspace, int dst_colorspace, int full_range ) { -#if defined(SWSCALE) && (LIBSWSCALE_VERSION_INT >= ((0<<16)+(7<<8)+2)) - int *coefficients; - const int *new_coefficients; - int full_range; - int brightness, contrast, saturation; - - if ( sws_getColorspaceDetails( context, &coefficients, &full_range, &coefficients, &full_range, - &brightness, &contrast, &saturation ) != -1 ) - { - // Don't change these from defaults unless explicitly told to. - if ( use_full_range >= 0 ) - full_range = use_full_range; - switch ( colorspace ) - { - case 170: - case 470: - case 601: - case 624: - new_coefficients = sws_getCoefficients( SWS_CS_ITU601 ); - break; - case 240: - new_coefficients = sws_getCoefficients( SWS_CS_SMPTE240M ); - break; - case 709: - new_coefficients = sws_getCoefficients( SWS_CS_ITU709 ); - break; - default: - new_coefficients = coefficients; - break; - } - sws_setColorspaceDetails( context, new_coefficients, full_range, new_coefficients, full_range, - brightness, contrast, saturation ); + const int *src_coefficients = sws_getCoefficients( SWS_CS_DEFAULT ); + const int *dst_coefficients = sws_getCoefficients( SWS_CS_DEFAULT ); + int brightness = 0; + int contrast = 1 << 16; + int saturation = 1 << 16; + + switch ( src_colorspace ) + { + case 170: + case 470: + case 601: + case 624: + src_coefficients = sws_getCoefficients( SWS_CS_ITU601 ); + break; + case 240: + src_coefficients = sws_getCoefficients( SWS_CS_SMPTE240M ); + break; + case 709: + src_coefficients = sws_getCoefficients( SWS_CS_ITU709 ); + break; + default: + break; + } + switch ( dst_colorspace ) + { + case 170: + case 470: + case 601: + case 624: + src_coefficients = sws_getCoefficients( SWS_CS_ITU601 ); + break; + case 240: + src_coefficients = sws_getCoefficients( SWS_CS_SMPTE240M ); + break; + case 709: + src_coefficients = sws_getCoefficients( SWS_CS_ITU709 ); + break; + default: + break; } -#endif + return sws_setColorspaceDetails( context, src_coefficients, full_range, dst_coefficients, full_range, + brightness, contrast, saturation ); } -static mlt_image_format pick_format( enum PixelFormat pix_fmt ) +static mlt_image_format pick_pix_format( enum PixelFormat pix_fmt ) { switch ( pix_fmt ) { @@ -1232,12 +1098,42 @@ static mlt_image_format pick_format( enum PixelFormat pix_fmt ) } } -static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, - mlt_image_format *format, int width, int height, int colorspace, uint8_t **alpha ) +static mlt_audio_format pick_audio_format( int sample_fmt ) +{ + switch ( sample_fmt ) + { + // interleaved + case AV_SAMPLE_FMT_U8: + return mlt_audio_u8; + case AV_SAMPLE_FMT_S16: + return mlt_audio_s16; + case AV_SAMPLE_FMT_S32: + return mlt_audio_s32le; + case AV_SAMPLE_FMT_FLT: + return mlt_audio_f32le; + // planar - this producer converts planar to interleaved +#if LIBAVUTIL_VERSION_INT >= ((51<<16)+(17<<8)+0) + case AV_SAMPLE_FMT_U8P: + return mlt_audio_u8; + case AV_SAMPLE_FMT_S16P: + return mlt_audio_s16; + case AV_SAMPLE_FMT_S32P: + return mlt_audio_s32le; + case AV_SAMPLE_FMT_FLTP: + return mlt_audio_f32le; +#endif + default: + return mlt_audio_none; + } +} + +// returns resulting YUV colorspace +static int convert_image( producer_avformat self, AVFrame *frame, uint8_t *buffer, int pix_fmt, + mlt_image_format *format, int width, int height, uint8_t **alpha ) { -#ifdef SWSCALE - int full_range = -1; int flags = SWS_BICUBIC | SWS_ACCURATE_RND; + mlt_profile profile = mlt_service_profile( MLT_PRODUCER_SERVICE( self->parent ) ); + int result = self->yuv_colorspace; #ifdef USE_MMX flags |= SWS_CPU_CAPS_MMX; @@ -1246,9 +1142,13 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, flags |= SWS_CPU_CAPS_MMX2; #endif + mlt_log_debug( MLT_PRODUCER_SERVICE(self->parent), "%s @ %dx%d space %d->%d\n", + mlt_image_format_name( *format ), + width, height, self->yuv_colorspace, profile->colorspace ); + // extract alpha from planar formats if ( ( pix_fmt == PIX_FMT_YUVA420P -#if LIBAVUTIL_VERSION_INT >= ((51<<16)+(35<<8)+101) +#if defined(FFUDIV) && LIBAVUTIL_VERSION_INT >= ((51<<16)+(35<<8)+101) || pix_fmt == PIX_FMT_YUVA444P #endif ) && @@ -1268,7 +1168,8 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, if ( *format == mlt_image_yuv420p ) { struct SwsContext *context = sws_getContext( width, height, pix_fmt, - width, height, PIX_FMT_YUV420P, flags, NULL, NULL, NULL); + width, height, self->full_luma ? PIX_FMT_YUVJ420P : PIX_FMT_YUV420P, + flags, NULL, NULL, NULL); AVPicture output; output.data[0] = buffer; output.data[1] = buffer + width * height; @@ -1276,7 +1177,8 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, output.linesize[0] = width; output.linesize[1] = width >> 1; output.linesize[2] = width >> 1; - set_luma_transfer( context, colorspace, full_range ); + if ( !set_luma_transfer( context, self->yuv_colorspace, profile->colorspace, self->full_luma ) ) + result = profile->colorspace; sws_scale( context, (const uint8_t* const*) frame->data, frame->linesize, 0, height, output.data, output.linesize); sws_freeContext( context ); @@ -1287,7 +1189,8 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, width, height, PIX_FMT_RGB24, flags | SWS_FULL_CHR_H_INT, NULL, NULL, NULL); AVPicture output; avpicture_fill( &output, buffer, PIX_FMT_RGB24, width, height ); - set_luma_transfer( context, colorspace, full_range ); + // libswscale wants the RGB colorspace to be SWS_CS_DEFAULT, which is = SWS_CS_ITU601. + set_luma_transfer( context, self->yuv_colorspace, 601, self->full_luma ); sws_scale( context, (const uint8_t* const*) frame->data, frame->linesize, 0, height, output.data, output.linesize); sws_freeContext( context ); @@ -1298,7 +1201,8 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, width, height, PIX_FMT_RGBA, flags | SWS_FULL_CHR_H_INT, NULL, NULL, NULL); AVPicture output; avpicture_fill( &output, buffer, PIX_FMT_RGBA, width, height ); - set_luma_transfer( context, colorspace, full_range ); + // libswscale wants the RGB colorspace to be SWS_CS_DEFAULT, which is = SWS_CS_ITU601. + set_luma_transfer( context, self->yuv_colorspace, 601, self->full_luma ); sws_scale( context, (const uint8_t* const*) frame->data, frame->linesize, 0, height, output.data, output.linesize); sws_freeContext( context ); @@ -1309,42 +1213,13 @@ static void convert_image( AVFrame *frame, uint8_t *buffer, int pix_fmt, width, height, PIX_FMT_YUYV422, flags | SWS_FULL_CHR_H_INP, NULL, NULL, NULL); AVPicture output; avpicture_fill( &output, buffer, PIX_FMT_YUYV422, width, height ); - set_luma_transfer( context, colorspace, full_range ); + if ( !set_luma_transfer( context, self->yuv_colorspace, profile->colorspace, self->full_luma ) ) + result = profile->colorspace; sws_scale( context, (const uint8_t* const*) frame->data, frame->linesize, 0, height, output.data, output.linesize); sws_freeContext( context ); } -#else - if ( *format == mlt_image_yuv420p ) - { - AVPicture pict; - pict.data[0] = buffer; - pict.data[1] = buffer + width * height; - pict.data[2] = buffer + ( 5 * width * height ) / 4; - pict.linesize[0] = width; - pict.linesize[1] = width >> 1; - pict.linesize[2] = width >> 1; - img_convert( &pict, PIX_FMT_YUV420P, (AVPicture *)frame, pix_fmt, width, height ); - } - else if ( *format == mlt_image_rgb24 ) - { - AVPicture output; - avpicture_fill( &output, buffer, PIX_FMT_RGB24, width, height ); - img_convert( &output, PIX_FMT_RGB24, (AVPicture *)frame, pix_fmt, width, height ); - } - else if ( format == mlt_image_rgb24a || format == mlt_image_opengl ) - { - AVPicture output; - avpicture_fill( &output, buffer, PIX_FMT_RGB32, width, height ); - img_convert( &output, PIX_FMT_RGB32, (AVPicture *)frame, pix_fmt, width, height ); - } - else - { - AVPicture output; - avpicture_fill( &output, buffer, PIX_FMT_YUYV422, width, height ); - img_convert( &output, PIX_FMT_YUYV422, (AVPicture *)frame, pix_fmt, width, height ); - } -#endif + return result; } /** Allocate the image buffer and set it on the frame. @@ -1356,6 +1231,10 @@ static int allocate_buffer( mlt_frame frame, AVCodecContext *codec_context, uint if ( codec_context->width == 0 || codec_context->height == 0 ) return size; + + if ( *format == mlt_image_glsl ) + *format = pick_pix_format( codec_context->pix_fmt ); + *width = codec_context->width; *height = codec_context->height; size = mlt_image_format_size( *format, *width, *height, NULL ); @@ -1381,15 +1260,21 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form mlt_properties frame_properties = MLT_FRAME_PROPERTIES( frame ); // Obtain the frame number of this frame - mlt_position position = mlt_properties_get_position( frame_properties, "avformat_position" ); + mlt_position position = mlt_frame_original_position( frame ); // Get the producer properties mlt_properties properties = MLT_PRODUCER_PROPERTIES( producer ); pthread_mutex_lock( &self->video_mutex ); + uint8_t *alpha = NULL; + int got_picture = 0; + int image_size = 0; + // Fetch the video format context AVFormatContext *context = self->video_format; + if ( !context ) + goto exit_get_image; // Get the video stream AVStream *stream = context->streams[ self->video_index ]; @@ -1397,10 +1282,6 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form // Get codec context AVCodecContext *codec_context = stream->codec; - uint8_t *alpha = NULL; - int got_picture = 0; - int image_size = 0; - // Get the image cache if ( ! self->image_cache ) { @@ -1425,7 +1306,7 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form } if ( self->image_cache ) { - mlt_frame original = mlt_cache_get_frame( self->image_cache, mlt_frame_get_position( frame ) ); + mlt_frame original = mlt_cache_get_frame( self->image_cache, position ); if ( original ) { mlt_properties orig_props = MLT_FRAME_PROPERTIES( original ); @@ -1453,9 +1334,6 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form } // Cache miss - // Packet - AVPacket pkt; - // We may want to use the source fps if available double source_fps = mlt_properties_get_double( properties, "meta.media.frame_rate_num" ) / mlt_properties_get_double( properties, "meta.media.frame_rate_den" ); @@ -1472,21 +1350,16 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form strcmp( codec_context->codec->name, "mjpeg" ) && strcmp( codec_context->codec->name, "rawvideo" ) ); - // Turn on usage of new seek API and PTS for seeking - int use_pts = self->seekable && - codec_context->codec_id == CODEC_ID_H264 && codec_context->has_b_frames; - if ( mlt_properties_get( properties, "use_pts" ) ) - use_pts = mlt_properties_get_int( properties, "use_pts" ); double delay = mlt_properties_get_double( properties, "video_delay" ); // Seek if necessary const char *interp = mlt_properties_get( frame_properties, "rescale.interp" ); int preseek = must_decode #if defined(FFUDIV) && LIBAVFORMAT_VERSION_INT >= ((53<<16)+(24<<8)+2) - && ( !use_pts || ( interp && strcmp( interp, "nearest" ) ) ) + && ( interp && strcmp( interp, "nearest" ) ) #endif && codec_context->has_b_frames; - int paused = seek_video( self, position, req_position, preseek, use_pts ); + int paused = seek_video( self, position, req_position, preseek ); // Seek might have reopened the file context = self->video_format; @@ -1497,15 +1370,16 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form codec_context->pix_fmt == PIX_FMT_RGBA || codec_context->pix_fmt == PIX_FMT_ABGR || codec_context->pix_fmt == PIX_FMT_BGRA ) - *format = pick_format( codec_context->pix_fmt ); + *format = pick_pix_format( codec_context->pix_fmt ); // Duplicate the last image if necessary - if ( self->av_frame && self->av_frame->linesize[0] + if ( self->video_frame && self->video_frame->linesize[0] && ( paused || self->current_position >= req_position ) ) { // Duplicate it if ( ( image_size = allocate_buffer( frame, codec_context, buffer, format, width, height ) ) ) { + int yuv_colorspace; // Workaround 1088 encodings missing cropping info. if ( *height == 1088 && mlt_profile_dar( mlt_service_profile( MLT_PRODUCER_SERVICE( producer ) ) ) == 16.0/9.0 ) *height = 1080; @@ -1519,13 +1393,14 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form picture.linesize[0] = codec_context->width; picture.linesize[1] = codec_context->width / 2; picture.linesize[2] = codec_context->width / 2; - convert_image( (AVFrame*) &picture, *buffer, - PIX_FMT_YUV420P, format, *width, *height, self->colorspace, &alpha ); + yuv_colorspace = convert_image( self, (AVFrame*) &picture, *buffer, + PIX_FMT_YUV420P, format, *width, *height, &alpha ); } else #endif - convert_image( self->av_frame, *buffer, codec_context->pix_fmt, - format, *width, *height, self->colorspace, &alpha ); + yuv_colorspace = convert_image( self, self->video_frame, *buffer, codec_context->pix_fmt, + format, *width, *height, &alpha ); + mlt_properties_set_int( frame_properties, "colorspace", yuv_colorspace ); got_picture = 1; } } @@ -1535,56 +1410,59 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form int64_t int_position = 0; int decode_errors = 0; - av_init_packet( &pkt ); - // Construct an AVFrame for YUV422 conversion - if ( !self->av_frame ) - self->av_frame = avcodec_alloc_frame( ); + if ( !self->video_frame ) + self->video_frame = avcodec_alloc_frame( ); while( ret >= 0 && !got_picture ) { // Read a packet + if ( self->pkt.stream_index == self->video_index ) + av_free_packet( &self->pkt ); + av_init_packet( &self->pkt ); pthread_mutex_lock( &self->packets_mutex ); if ( mlt_deque_count( self->vpackets ) ) { AVPacket *tmp = (AVPacket*) mlt_deque_pop_front( self->vpackets ); - pkt = *tmp; + self->pkt = *tmp; free( tmp ); } else { - ret = av_read_frame( context, &pkt ); - if ( ret >= 0 && !self->seekable && pkt.stream_index == self->audio_index ) + ret = av_read_frame( context, &self->pkt ); + if ( ret >= 0 && !self->seekable && self->pkt.stream_index == self->audio_index ) { - if ( !av_dup_packet( &pkt ) ) + if ( !av_dup_packet( &self->pkt ) ) { AVPacket *tmp = malloc( sizeof(AVPacket) ); - *tmp = pkt; + *tmp = self->pkt; mlt_deque_push_back( self->apackets, tmp ); } } + else if ( ret < 0 ) + { + mlt_log_verbose( MLT_PRODUCER_SERVICE(producer), "av_read_frame returned error %d inside get_image\n", ret ); + if ( !self->seekable && mlt_properties_get_int( properties, "reconnect" ) ) + { + // Try to reconnect to live sources by closing context and codecs, + // and letting next call to get_frame() reopen. + prepare_reopen( self ); + pthread_mutex_unlock( &self->packets_mutex ); + goto exit_get_image; + } + if ( !self->seekable && mlt_properties_get_int( properties, "exit_on_disconnect" ) ) + { + mlt_log_fatal( MLT_PRODUCER_SERVICE(producer), "Exiting with error due to disconnected source.\n" ); + exit( EXIT_FAILURE ); + } + } } pthread_mutex_unlock( &self->packets_mutex ); // We only deal with video from the selected video_index - if ( ret >= 0 && pkt.stream_index == self->video_index && pkt.size > 0 ) + if ( ret >= 0 && self->pkt.stream_index == self->video_index && self->pkt.size > 0 ) { - // Determine time code of the packet - if ( use_pts && pkt.pts == AV_NOPTS_VALUE ) - { - self->invalid_pts_counter++; - if ( self->invalid_pts_counter > 20 ) - { - mlt_log_warning( MLT_PRODUCER_SERVICE(producer), "PTS invalid; using DTS instead\n" ); - mlt_properties_set_int( properties, "use_pts", 0 ); - use_pts = 0; - } - } - else - { - self->invalid_pts_counter = 0; - } - int64_t pts = ( use_pts && pkt.pts != AV_NOPTS_VALUE )? pkt.pts : pkt.dts; + int64_t pts = best_pts( self, self->pkt.pts, self->pkt.dts ); if ( pts != AV_NOPTS_VALUE ) { if ( !self->seekable && self->first_pts == AV_NOPTS_VALUE ) @@ -1593,13 +1471,13 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form pts -= self->first_pts; else if ( context->start_time != AV_NOPTS_VALUE ) pts -= context->start_time; - int_position = ( int64_t )( ( av_q2d( stream->time_base ) * pts + delay ) * source_fps + 0.5 ); + int_position = ( int64_t )( ( av_q2d( self->video_time_base ) * pts + delay ) * source_fps + 0.5 ); if ( int_position == self->last_position ) int_position = self->last_position + 1; } mlt_log_debug( MLT_PRODUCER_SERVICE(producer), "V pkt.pts %"PRId64" pkt.dts %"PRId64" req_pos %"PRId64" cur_pos %"PRId64" pkt_pos %"PRId64"\n", - pkt.pts, pkt.dts, req_position, self->current_position, int_position ); + self->pkt.pts, self->pkt.dts, req_position, self->current_position, int_position ); // Make a dumb assumption on streams that contain wild timestamps if ( abs( req_position - int_position ) > 999 ) @@ -1626,15 +1504,17 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form if ( int_position >= req_position ) codec_context->skip_loop_filter = AVDISCARD_NONE; #if (LIBAVCODEC_VERSION_INT >= ((52<<16)+(26<<8)+0)) - ret = avcodec_decode_video2( codec_context, self->av_frame, &got_picture, &pkt ); + ret = avcodec_decode_video2( codec_context, self->video_frame, &got_picture, &self->pkt ); #else - ret = avcodec_decode_video( codec_context, self->av_frame, &got_picture, pkt.data, pkt.size ); + ret = avcodec_decode_video( codec_context, self->video_frame, &got_picture, self->pkt.data, self->pkt.size ); #endif // Note: decode may fail at the beginning of MPEGfile (B-frames referencing before first I-frame), so allow a few errors. if ( ret < 0 ) { if ( ++decode_errors <= 10 ) ret = 0; + else + mlt_log_warning( MLT_PRODUCER_SERVICE(producer), "video decoding error %d\n", ret ); } else { @@ -1644,16 +1524,26 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form if ( got_picture ) { - if ( use_pts ) - // Get position of reordered frame - int_position = self->av_frame->reordered_opaque; + // Get position of reordered frame + int_position = self->video_frame->reordered_opaque; +#if (LIBAVCODEC_VERSION_INT >= ((52<<16)+(106<<8)+0)) + pts = best_pts( self, self->video_frame->pkt_pts, self->video_frame->pkt_dts ); + if ( pts != AV_NOPTS_VALUE ) + { + if ( self->first_pts != AV_NOPTS_VALUE ) + pts -= self->first_pts; + else if ( context->start_time != AV_NOPTS_VALUE ) + pts -= context->start_time; + int_position = ( int64_t )( ( av_q2d( self->video_time_base ) * pts + delay ) * source_fps + 0.5 ); + } +#endif if ( int_position < req_position ) got_picture = 0; else if ( int_position >= req_position ) codec_context->skip_loop_filter = AVDISCARD_NONE; } - mlt_log_debug( MLT_PRODUCER_SERVICE(producer), " got_pic %d key %d\n", got_picture, pkt.flags & PKT_FLAG_KEY ); + mlt_log_debug( MLT_PRODUCER_SERVICE(producer), " got_pic %d key %d\n", got_picture, self->pkt.flags & PKT_FLAG_KEY ); } // Now handle the picture if we have one @@ -1661,6 +1551,7 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form { if ( ( image_size = allocate_buffer( frame, codec_context, buffer, format, width, height ) ) ) { + int yuv_colorspace; // Workaround 1088 encodings missing cropping info. if ( *height == 1088 && mlt_profile_dar( mlt_service_profile( MLT_PRODUCER_SERVICE( producer ) ) ) == 16.0/9.0 ) *height = 1080; @@ -1669,25 +1560,26 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form { if ( self->vdpau->is_decoded ) { - struct vdpau_render_state *render = (struct vdpau_render_state*) self->av_frame->data[0]; + struct vdpau_render_state *render = (struct vdpau_render_state*) self->video_frame->data[0]; void *planes[3]; uint32_t pitches[3]; VdpYCbCrFormat dest_format = VDP_YCBCR_FORMAT_YV12; if ( !self->vdpau->buffer ) self->vdpau->buffer = mlt_pool_alloc( codec_context->width * codec_context->height * 3 / 2 ); - self->av_frame->data[0] = planes[0] = self->vdpau->buffer; - self->av_frame->data[2] = planes[1] = self->vdpau->buffer + codec_context->width * codec_context->height; - self->av_frame->data[1] = planes[2] = self->vdpau->buffer + codec_context->width * codec_context->height * 5 / 4; - self->av_frame->linesize[0] = pitches[0] = codec_context->width; - self->av_frame->linesize[1] = pitches[1] = codec_context->width / 2; - self->av_frame->linesize[2] = pitches[2] = codec_context->width / 2; + self->video_frame->data[0] = planes[0] = self->vdpau->buffer; + self->video_frame->data[2] = planes[1] = self->vdpau->buffer + codec_context->width * codec_context->height; + self->video_frame->data[1] = planes[2] = self->vdpau->buffer + codec_context->width * codec_context->height * 5 / 4; + self->video_frame->linesize[0] = pitches[0] = codec_context->width; + self->video_frame->linesize[1] = pitches[1] = codec_context->width / 2; + self->video_frame->linesize[2] = pitches[2] = codec_context->width / 2; VdpStatus status = vdp_surface_get_bits( render->surface, dest_format, planes, pitches ); if ( status == VDP_STATUS_OK ) { - convert_image( self->av_frame, *buffer, PIX_FMT_YUV420P, - format, *width, *height, self->colorspace, &alpha ); + yuv_colorspace = convert_image( self, self->video_frame, *buffer, PIX_FMT_YUV420P, + format, *width, *height, &alpha ); + mlt_properties_set_int( frame_properties, "colorspace", yuv_colorspace ); } else { @@ -1703,9 +1595,10 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form } else #endif - convert_image( self->av_frame, *buffer, codec_context->pix_fmt, - format, *width, *height, self->colorspace, &alpha ); - self->top_field_first |= self->av_frame->top_field_first; + yuv_colorspace = convert_image( self, self->video_frame, *buffer, codec_context->pix_fmt, + format, *width, *height, &alpha ); + mlt_properties_set_int( frame_properties, "colorspace", yuv_colorspace ); + self->top_field_first |= self->video_frame->top_field_first; self->current_position = int_position; } else @@ -1713,8 +1606,11 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form got_picture = 0; } } - if ( self->seekable || pkt.stream_index != self->audio_index ) - av_free_packet( &pkt ); + + // Free packet data if not video and not live audio packet + if ( self->pkt.stream_index != self->video_index && + !( !self->seekable && self->pkt.stream_index == self->audio_index ) ) + av_free_packet( &self->pkt ); } } @@ -1731,11 +1627,12 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form // Try to duplicate last image if there was a decoding failure // TODO: with multithread decoding a partial frame decoding resulting // in failure also resets av_frame making test below fail. - if ( !image_size && self->av_frame && self->av_frame->linesize[0] ) + if ( !image_size && self->video_frame && self->video_frame->linesize[0] ) { // Duplicate it if ( ( image_size = allocate_buffer( frame, codec_context, buffer, format, width, height ) ) ) { + int yuv_colorspace; // Workaround 1088 encodings missing cropping info. if ( *height == 1088 && mlt_profile_dar( mlt_service_profile( MLT_PRODUCER_SERVICE( producer ) ) ) == 16.0/9.0 ) *height = 1080; @@ -1749,13 +1646,15 @@ static int producer_get_image( mlt_frame frame, uint8_t **buffer, mlt_image_form picture.linesize[0] = codec_context->width; picture.linesize[1] = codec_context->width / 2; picture.linesize[2] = codec_context->width / 2; - convert_image( (AVFrame*) &picture, *buffer, - PIX_FMT_YUV420P, format, *width, *height, self->colorspace, &alpha ); + yuv_colorspace = convert_image( self, (AVFrame*) &picture, *buffer, + PIX_FMT_YUV420P, format, *width, *height, &alpha ); + mlt_properties_set_int( frame_properties, "colorspace", yuv_colorspace ); } else #endif - convert_image( self->av_frame, *buffer, codec_context->pix_fmt, - format, *width, *height, self->colorspace, &alpha ); + yuv_colorspace = convert_image( self, self->video_frame, *buffer, codec_context->pix_fmt, + format, *width, *height, &alpha ); + mlt_properties_set_int( frame_properties, "colorspace", yuv_colorspace ); got_picture = 1; } } @@ -1770,8 +1669,8 @@ exit_get_image: // Set the progressive flag if ( mlt_properties_get( properties, "force_progressive" ) ) mlt_properties_set_int( frame_properties, "progressive", !!mlt_properties_get_int( properties, "force_progressive" ) ); - else if ( self->av_frame ) - mlt_properties_set_int( frame_properties, "progressive", !self->av_frame->interlaced_frame ); + else if ( self->video_frame ) + mlt_properties_set_int( frame_properties, "progressive", !self->video_frame->interlaced_frame ); // Set the field order property for this frame if ( mlt_properties_get( properties, "force_tff" ) ) @@ -1785,6 +1684,11 @@ exit_get_image: mlt_properties_set_int( properties, "meta.media.progressive", mlt_properties_get_int( frame_properties, "progressive" ) ); mlt_service_unlock( MLT_PRODUCER_SERVICE( producer ) ); + // If we already have RGB, then the full range processing either happened already + // or does not apply (RGB source). + if ( *format == mlt_image_rgb24 || *format == mlt_image_rgb24a || *format == mlt_image_opengl ) + mlt_properties_set( frame_properties, "force_full_luma", NULL ); + return !got_picture; } @@ -1836,7 +1740,7 @@ static int video_codec_init( producer_avformat self, int index, mlt_properties p // Find the codec AVCodec *codec = avcodec_find_decoder( codec_context->codec_id ); #ifdef VDPAU - if ( codec_context->codec_id == CODEC_ID_H264 ) + if ( codec_context->codec_id == AV_CODEC_ID_H264 ) { if ( ( codec = avcodec_find_decoder_by_name( "h264_vdpau" ) ) ) { @@ -1887,90 +1791,118 @@ static int video_codec_init( producer_avformat self, int index, mlt_properties p #endif // Reset some image properties - mlt_properties_set_int( properties, "width", self->video_codec->width ); - mlt_properties_set_int( properties, "height", self->video_codec->height ); - // For DV, we'll just use the saved aspect ratio - if ( codec_context->codec_id != CODEC_ID_DVVIDEO ) - get_aspect_ratio( properties, stream, self->video_codec, NULL ); - - // Determine the fps first from the codec - double source_fps = (double) self->video_codec->time_base.den / - ( self->video_codec->time_base.num == 0 ? 1 : self->video_codec->time_base.num ); - - if ( mlt_properties_get( properties, "force_fps" ) ) + if ( self->video_codec ) { - source_fps = mlt_properties_get_double( properties, "force_fps" ); - stream->time_base = av_d2q( source_fps, 1024 ); - mlt_properties_set_int( properties, "meta.media.frame_rate_num", stream->time_base.num ); - mlt_properties_set_int( properties, "meta.media.frame_rate_den", stream->time_base.den ); + mlt_properties_set_int( properties, "width", self->video_codec->width ); + mlt_properties_set_int( properties, "height", self->video_codec->height ); + get_aspect_ratio( properties, stream, self->video_codec ); } - else - { - // If the muxer reports a frame rate different than the codec + + // Start with the muxer frame rate. #if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(42<<8)+0) - double muxer_fps = av_q2d( stream->avg_frame_rate ); - if ( isnan( muxer_fps ) || muxer_fps == 0 ) - muxer_fps = av_q2d( stream->r_frame_rate ); + AVRational frame_rate = stream->avg_frame_rate; #else - double muxer_fps = av_q2d( stream->r_frame_rate ); + AVRational frame_rate = stream->r_frame_rate; #endif - // Choose the lesser - the wrong tends to be off by some multiple of 10 - source_fps = FFMIN( source_fps, muxer_fps ); - if ( source_fps >= 1.0 && ( source_fps < muxer_fps || isnan( muxer_fps ) ) ) - { - mlt_properties_set_int( properties, "meta.media.frame_rate_num", self->video_codec->time_base.den ); - mlt_properties_set_int( properties, "meta.media.frame_rate_den", self->video_codec->time_base.num == 0 ? 1 : self->video_codec->time_base.num ); - } - else if ( muxer_fps > 0 ) - { - AVRational frame_rate = stream->r_frame_rate; - // With my samples when r_frame_rate != 1000 but avg_frame_rate is valid, - // avg_frame_rate gives some approximate value that does not well match the media. - // Also, on my sample where r_frame_rate = 1000, using avg_frame_rate directly - // results in some very choppy output, but some value slightly different works - // great. -#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(42<<8)+0) - if ( av_q2d( stream->r_frame_rate ) >= 1000 && av_q2d( stream->avg_frame_rate ) > 0 ) - frame_rate = av_d2q( av_q2d( stream->avg_frame_rate ), 1024 ); + double fps = av_q2d( frame_rate ); + +#if LIBAVFORMAT_VERSION_MAJOR < 55 || defined(FFUDIV) + // Verify and sanitize the muxer frame rate. + if ( isnan( fps ) || isinf( fps ) || fps == 0 ) + { + frame_rate = stream->r_frame_rate; + fps = av_q2d( frame_rate ); + } #endif - mlt_properties_set_int( properties, "meta.media.frame_rate_num", frame_rate.num ); - mlt_properties_set_int( properties, "meta.media.frame_rate_den", frame_rate.den ); - } - else - { - source_fps = mlt_producer_get_fps( self->parent ); - AVRational frame_rate = av_d2q( source_fps, 255 ); - mlt_properties_set_int( properties, "meta.media.frame_rate_num", frame_rate.num ); - mlt_properties_set_int( properties, "meta.media.frame_rate_den", frame_rate.den ); - } +#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(42<<8)+0) && ( LIBAVFORMAT_VERSION_MAJOR < 55 || defined(FFUDIV) ) + // With my samples when r_frame_rate != 1000 but avg_frame_rate is valid, + // avg_frame_rate gives some approximate value that does not well match the media. + // Also, on my sample where r_frame_rate = 1000, using avg_frame_rate directly + // results in some very choppy output, but some value slightly different works + // great. + if ( av_q2d( stream->r_frame_rate ) >= 1000 && av_q2d( stream->avg_frame_rate ) > 0 ) + { + frame_rate = av_d2q( av_q2d( stream->avg_frame_rate ), 1024 ); + fps = av_q2d( frame_rate ); + } +#endif + // XXX frame rates less than 1 fps are not considered sane + if ( isnan( fps ) || isinf( fps ) || fps < 1.0 ) + { + // Get the frame rate from the codec. + frame_rate.num = self->video_codec->time_base.den; + frame_rate.den = self->video_codec->time_base.num * self->video_codec->ticks_per_frame; + fps = av_q2d( frame_rate ); + } + if ( isnan( fps ) || isinf( fps ) || fps < 1.0 ) + { + // Use the profile frame rate if all else fails. + mlt_profile profile = mlt_service_profile( MLT_PRODUCER_SERVICE( self->parent ) ); + frame_rate.num = profile->frame_rate_num; + frame_rate.den = profile->frame_rate_den; } + self->video_time_base = stream->time_base; + if ( mlt_properties_get( properties, "force_fps" ) ) + { + AVRational force_fps = av_d2q( mlt_properties_get_double( properties, "force_fps" ), 1024 ); + self->video_time_base = av_mul_q( stream->time_base, av_div_q( frame_rate, force_fps ) ); + frame_rate = force_fps; + } + mlt_properties_set_int( properties, "meta.media.frame_rate_num", frame_rate.num ); + mlt_properties_set_int( properties, "meta.media.frame_rate_den", frame_rate.den ); + // Set the YUV colorspace from override or detect - self->colorspace = mlt_properties_get_int( properties, "force_colorspace" ); -#if LIBAVCODEC_VERSION_INT > ((52<<16)+(28<<8)+0) - if ( ! self->colorspace ) + self->yuv_colorspace = mlt_properties_get_int( properties, "force_colorspace" ); +#if LIBAVCODEC_VERSION_INT > ((52<<16)+(28<<8)+0) + if ( ! self->yuv_colorspace ) { switch ( self->video_codec->colorspace ) { case AVCOL_SPC_SMPTE240M: - self->colorspace = 240; + self->yuv_colorspace = 240; break; case AVCOL_SPC_BT470BG: case AVCOL_SPC_SMPTE170M: - self->colorspace = 601; + self->yuv_colorspace = 601; break; case AVCOL_SPC_BT709: - self->colorspace = 709; + self->yuv_colorspace = 709; break; default: // This is a heuristic Charles Poynton suggests in "Digital Video and HDTV" - self->colorspace = self->video_codec->width * self->video_codec->height > 750000 ? 709 : 601; + self->yuv_colorspace = self->video_codec->width * self->video_codec->height > 750000 ? 709 : 601; break; } } #endif // Let apps get chosen colorspace - mlt_properties_set_int( properties, "meta.media.colorspace", self->colorspace ); + mlt_properties_set_int( properties, "meta.media.colorspace", self->yuv_colorspace ); + + switch ( self->video_codec->color_primaries ) + { + case AVCOL_PRI_BT470BG: + self->color_primaries = 601625; + break; + case AVCOL_PRI_SMPTE170M: + case AVCOL_PRI_SMPTE240M: + self->color_primaries = 601525; + break; + case AVCOL_PRI_BT709: + case AVCOL_PRI_UNSPECIFIED: + default: + self->color_primaries = 709; + break; + } + + self->full_luma = 0; +#if LIBAVCODEC_VERSION_INT >= ((52<<16)+(72<<8)+2) + mlt_log_debug( MLT_PRODUCER_SERVICE(self->parent), "color_range %d\n", codec_context->color_range ); + if ( codec_context->color_range == AVCOL_RANGE_JPEG ) + self->full_luma = 1; +#endif + if ( mlt_properties_get( properties, "set.force_full_luma" ) ) + self->full_luma = mlt_properties_get_int( properties, "set.force_full_luma" ); } return self->video_codec && self->video_index > -1; } @@ -1992,11 +1924,15 @@ static void producer_set_up_video( producer_avformat self, mlt_frame frame ) // Get the video_index int index = mlt_properties_get_int( properties, "video_index" ); + int unlock_needed = 0; + // Reopen the file if necessary if ( !context && index > -1 ) { + unlock_needed = 1; + pthread_mutex_lock( &self->video_mutex ); producer_open( self, mlt_service_profile( MLT_PRODUCER_SERVICE(producer) ), - mlt_properties_get( properties, "resource" ), 1 ); + mlt_properties_get( properties, "resource" ), 0 ); context = self->video_format; } @@ -2017,7 +1953,7 @@ static void producer_set_up_video( producer_avformat self, mlt_frame frame ) } // Update the video properties if the index changed - if ( index != self->video_index ) + if ( context && index > -1 && index != self->video_index ) { // Reset the video properties if the index changed self->video_index = index; @@ -2045,7 +1981,9 @@ static void producer_set_up_video( producer_avformat self, mlt_frame frame ) mlt_properties_set_int( properties, "meta.media.width", self->video_codec->width ); mlt_properties_set_int( properties, "meta.media.height", self->video_codec->height ); mlt_properties_set_double( frame_properties, "aspect_ratio", aspect_ratio ); - mlt_properties_set_int( frame_properties, "colorspace", self->colorspace ); + mlt_properties_set_int( frame_properties, "colorspace", self->yuv_colorspace ); + mlt_properties_set_int( frame_properties, "color_primaries", self->color_primaries ); + mlt_properties_set_int( frame_properties, "full_luma", self->full_luma ); // Workaround 1088 encodings missing cropping info. if ( self->video_codec->height == 1088 && mlt_profile_dar( mlt_service_profile( MLT_PRODUCER_SERVICE( producer ) ) ) == 16.0/9.0 ) @@ -2062,6 +2000,8 @@ static void producer_set_up_video( producer_avformat self, mlt_frame frame ) // If something failed, use test card image mlt_properties_set_int( frame_properties, "test_image", 1 ); } + if ( unlock_needed ) + pthread_mutex_unlock( &self->video_mutex ); } static int seek_audio( producer_avformat self, mlt_position position, double timecode ) @@ -2069,8 +2009,17 @@ static int seek_audio( producer_avformat self, mlt_position position, double tim int paused = 0; // Seek if necessary - if ( self->seekable && position != self->audio_expected ) + if ( self->seekable && ( position != self->audio_expected || self->last_position < 0 ) ) { + if ( self->last_position == POSITION_INITIAL ) + { + int video_index = self->video_index; + if ( video_index == -1 ) + video_index = first_video_index( self ); + if ( video_index >= 0 ) + find_first_pts( self, video_index ); + } + if ( position + 1 == self->audio_expected ) { // We're paused - silence required @@ -2109,6 +2058,34 @@ static int sample_bytes( AVCodecContext *context ) #endif } +#if LIBAVCODEC_VERSION_MAJOR >= 55 +static void planar_to_interleaved( uint8_t *dest, AVFrame *src, int samples, int channels, int bytes_per_sample ) +{ + int s, c; + for ( s = 0; s < samples; s++ ) + { + for ( c = 0; c < channels; c++ ) + { + memcpy( dest, &src->data[c][s * bytes_per_sample], bytes_per_sample ); + dest += bytes_per_sample; + } + } +} +#else +static void planar_to_interleaved( uint8_t *dest, uint8_t *src, int samples, int channels, int bytes_per_sample ) +{ + int s, c; + for ( s = 0; s < samples; s++ ) + { + for ( c = 0; c < channels; c++ ) + { + memcpy( dest, src + ( c * samples + s ) * bytes_per_sample, bytes_per_sample ); + dest += bytes_per_sample; + } + } +} +#endif + static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int channels, int samples, double timecode, double fps ) { // Fetch the audio_format @@ -2120,9 +2097,6 @@ static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int // Get codec context AVCodecContext *codec_context = self->audio_codec[ index ]; - // Obtain the resample context if it exists (not always needed) - ReSampleContext *resample = self->audio_resample[ index ]; - // Obtain the audio buffers uint8_t *audio_buffer = self->audio_buffer[ index ]; uint8_t *decode_buffer = self->decode_buffer[ index ]; @@ -2134,16 +2108,23 @@ static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int while ( ptr && ret >= 0 && len > 0 ) { - int sizeof_sample = resample? sizeof( int16_t ) : sample_bytes( codec_context ); + int sizeof_sample = sample_bytes( codec_context ); int data_size = self->audio_buffer_size[ index ]; // Decode the audio -#if (LIBAVCODEC_VERSION_INT >= ((52<<16)+(26<<8)+0)) - ret = avcodec_decode_audio3( codec_context, (int16_t*) decode_buffer, &data_size, &pkt ); -#elif (LIBAVCODEC_VERSION_INT >= ((51<<16)+(29<<8)+0)) - ret = avcodec_decode_audio2( codec_context, decode_buffer, &data_size, ptr, len ); +#if LIBAVCODEC_VERSION_MAJOR >= 55 + if ( !self->audio_frame ) + self->audio_frame = avcodec_alloc_frame(); + else + avcodec_get_frame_defaults( self->audio_frame ); + ret = avcodec_decode_audio4( codec_context, self->audio_frame, &data_size, &pkt ); + if ( data_size ) { + data_size = av_samples_get_buffer_size( NULL, codec_context->channels, + self->audio_frame->nb_samples, codec_context->sample_fmt, 1 ); + decode_buffer = self->audio_frame->data[0]; + } #else - ret = avcodec_decode_audio( codec_context, decode_buffer, &data_size, ptr, len ); + ret = avcodec_decode_audio3( codec_context, (int16_t*) decode_buffer, &data_size, &pkt ); #endif if ( ret < 0 ) { @@ -2159,34 +2140,40 @@ static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int { // Figure out how many samples will be needed after resampling int convert_samples = data_size / codec_context->channels / sample_bytes( codec_context ); - int samples_needed = self->resample_factor * convert_samples; // Resize audio buffer to prevent overflow - if ( ( audio_used + samples_needed ) * channels * sizeof_sample > self->audio_buffer_size[ index ] ) + if ( ( audio_used + convert_samples ) * channels * sizeof_sample > self->audio_buffer_size[ index ] ) { - self->audio_buffer_size[ index ] = ( audio_used + samples_needed * 2 ) * channels * sizeof_sample; + self->audio_buffer_size[ index ] = ( audio_used + convert_samples * 2 ) * channels * sizeof_sample; audio_buffer = self->audio_buffer[ index ] = mlt_pool_realloc( audio_buffer, self->audio_buffer_size[ index ] ); } - if ( resample ) - { - // Copy to audio buffer while resampling - uint8_t *source = decode_buffer; - uint8_t *dest = &audio_buffer[ audio_used * channels * sizeof_sample ]; - audio_used += audio_resample( resample, (short*) dest, (short*) source, convert_samples ); - } - else + uint8_t *dest = &audio_buffer[ audio_used * codec_context->channels * sizeof_sample ]; + switch ( codec_context->sample_fmt ) { +#if LIBAVUTIL_VERSION_INT >= ((51<<16)+(17<<8)+0) + case AV_SAMPLE_FMT_U8P: + case AV_SAMPLE_FMT_S16P: + case AV_SAMPLE_FMT_S32P: + case AV_SAMPLE_FMT_FLTP: +#if LIBAVCODEC_VERSION_MAJOR >= 55 + planar_to_interleaved( dest, self->audio_frame, convert_samples, codec_context->channels, sizeof_sample ); +#else + planar_to_interleaved( dest, decode_buffer, convert_samples, codec_context->channels, sizeof_sample ); +#endif + break; +#endif + default: // Straight copy to audio buffer - memcpy( &audio_buffer[ audio_used * codec_context->channels * sizeof_sample ], decode_buffer, data_size ); - audio_used += convert_samples; + memcpy( dest, decode_buffer, data_size ); } + audio_used += convert_samples; // Handle ignore while ( *ignore && audio_used ) { *ignore -= 1; audio_used -= audio_used > samples ? samples : audio_used; - memmove( audio_buffer, &audio_buffer[ samples * (resample? channels : codec_context->channels) * sizeof_sample ], + memmove( audio_buffer, &audio_buffer[ samples * codec_context->channels * sizeof_sample ], audio_used * sizeof_sample ); } } @@ -2197,7 +2184,7 @@ static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int if ( pkt.pts >= 0 && ( self->seekable || self->video_format ) && *ignore == 0 && audio_used > samples / 2 ) { int64_t pts = pkt.pts; - if ( self->first_pts != 0 ) + if ( self->first_pts != AV_NOPTS_VALUE ) pts -= self->first_pts; else if ( context->start_time != AV_NOPTS_VALUE ) pts -= context->start_time; @@ -2218,6 +2205,9 @@ static int decode_audio( producer_avformat self, int *ignore, AVPacket pkt, int // We are ahead, so seek backwards some more seek_audio( self, req_position, timecode - 1.0 ); } + // Cancel the find_first_pts() in seek_audio() + if ( self->video_index == -1 && self->last_position == POSITION_INITIAL ) + self->last_position = int_position; } self->audio_used[ index ] = audio_used; @@ -2235,7 +2225,7 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format pthread_mutex_lock( &self->audio_mutex ); // Obtain the frame number of this frame - mlt_position position = mlt_properties_get_position( MLT_FRAME_PROPERTIES( frame ), "avformat_position" ); + mlt_position position = mlt_frame_original_position( frame ); // Calculate the real time code double real_timecode = producer_time_of_frame( self->parent, position ); @@ -2256,6 +2246,8 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format // Fetch the audio_format AVFormatContext *context = self->audio_format; + if ( !context ) + goto exit_get_audio; int sizeof_sample = sizeof( int16_t ); @@ -2265,13 +2257,13 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format if ( self->audio_index == INT_MAX ) { index = 0; - index_max = context->nb_streams; + index_max = FFMIN( MAX_AUDIO_STREAMS, context->nb_streams ); *channels = self->total_channels; - *samples = *samples * FFMAX( self->max_frequency, *frequency ) / *frequency; - *frequency = FFMAX( self->max_frequency, *frequency ); + *samples = mlt_sample_calculator( fps, self->max_frequency, position ); + *frequency = self->max_frequency; } - // Initialize the resamplers and buffers + // Initialize the buffers for ( ; index < index_max && index < MAX_AUDIO_STREAMS; index++ ) { // Get codec context @@ -2279,35 +2271,11 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format if ( codec_context && !self->audio_buffer[ index ] ) { - // Check for resample and create if necessary - if ( codec_context->channels <= 2 ) - { - // Determine by how much resampling will increase number of samples - double resample_factor = self->audio_index == INT_MAX ? 1 : (double) *channels / codec_context->channels; - resample_factor *= (double) *frequency / codec_context->sample_rate; - if ( resample_factor > self->resample_factor ) - self->resample_factor = resample_factor; - - // Create the resampler -#if (LIBAVCODEC_VERSION_INT >= ((52<<16)+(15<<8)+0)) - self->audio_resample[ index ] = av_audio_resample_init( - self->audio_index == INT_MAX ? codec_context->channels : *channels, - codec_context->channels, *frequency, codec_context->sample_rate, - AV_SAMPLE_FMT_S16, codec_context->sample_fmt, 16, 10, 0, 0.8 ); -#else - self->audio_resample[ index ] = audio_resample_init( - self->audio_index == INT_MAX ? codec_context->channels : *channels, - codec_context->channels, *frequency, codec_context->sample_rate ); -#endif - } - else - { - codec_context->request_channels = self->audio_index == INT_MAX ? codec_context->channels : *channels; - sizeof_sample = sample_bytes( codec_context ); - } + codec_context->request_channels = self->audio_index == INT_MAX ? codec_context->channels : *channels; + sizeof_sample = sample_bytes( codec_context ); // Check for audio buffer and create if necessary - self->audio_buffer_size[ index ] = AVCODEC_MAX_AUDIO_FRAME_SIZE * sizeof_sample; + self->audio_buffer_size[ index ] = MAX_AUDIO_FRAME_SIZE * sizeof_sample; self->audio_buffer[ index ] = mlt_pool_alloc( self->audio_buffer_size[ index ] ); // Check for decoder buffer and create if necessary @@ -2316,7 +2284,7 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format } // Get the audio if required - if ( !paused ) + if ( !paused && *frequency > 0 ) { int ret = 0; int got_audio = 0; @@ -2324,11 +2292,9 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format av_init_packet( &pkt ); - // If not resampling, give consumer more than requested. - // It requested number samples based on requested frame rate. - // Do not clean this up with a samples *= ...! - if ( self->audio_index != INT_MAX && ! self->audio_resample[ self->audio_index ] ) - *samples = *samples * self->audio_codec[ self->audio_index ]->sample_rate / *frequency; + // Caller requested number samples based on requested sample rate. + if ( self->audio_index != INT_MAX ) + *samples = mlt_sample_calculator( fps, self->audio_codec[ self->audio_index ]->sample_rate, position ); while ( ret >= 0 && !got_audio ) { @@ -2344,7 +2310,7 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format { // Check if there is enough audio for all streams got_audio = 1; - for ( index = 0; got_audio && index < context->nb_streams; index++ ) + for ( index = 0; got_audio && index < index_max; index++ ) if ( ( self->audio_codec[ index ] && self->audio_used[ index ] < *samples ) || ignore[ index ] ) got_audio = 0; if ( got_audio ) @@ -2371,6 +2337,25 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format mlt_deque_push_back( self->vpackets, tmp ); } } + else if ( ret < 0 ) + { + mlt_producer producer = self->parent; + mlt_properties properties = MLT_PRODUCER_PROPERTIES( producer ); + mlt_log_verbose( MLT_PRODUCER_SERVICE(producer), "av_read_frame returned error %d inside get_audio\n", ret ); + if ( !self->seekable && mlt_properties_get_int( properties, "reconnect" ) ) + { + // Try to reconnect to live sources by closing context and codecs, + // and letting next call to get_frame() reopen. + prepare_reopen( self ); + pthread_mutex_unlock( &self->packets_mutex ); + goto exit_get_audio; + } + if ( !self->seekable && mlt_properties_get_int( properties, "exit_on_disconnect" ) ) + { + mlt_log_fatal( MLT_PRODUCER_SERVICE(producer), "Exiting with error due to disconnected source.\n" ); + exit( EXIT_FAILURE ); + } + } } pthread_mutex_unlock( &self->packets_mutex ); @@ -2379,8 +2364,7 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format if ( index < MAX_AUDIO_STREAMS && ret >= 0 && pkt.data && pkt.size > 0 && ( index == self->audio_index || ( self->audio_index == INT_MAX && context->streams[ index ]->codec->codec_type == CODEC_TYPE_AUDIO ) ) ) { - int channels2 = ( self->audio_index == INT_MAX || !self->audio_resample[index] ) ? - self->audio_codec[index]->channels : *channels; + int channels2 = self->audio_codec[index]->channels; ret = decode_audio( self, &ignore[index], pkt, channels2, *samples, real_timecode, fps ); } @@ -2391,25 +2375,21 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format // Set some additional return values *format = mlt_audio_s16; - if ( self->audio_index != INT_MAX && !self->audio_resample[ self->audio_index ] ) + if ( self->audio_index != INT_MAX ) { index = self->audio_index; *channels = self->audio_codec[ index ]->channels; *frequency = self->audio_codec[ index ]->sample_rate; - *format = self->audio_codec[ index ]->sample_fmt == AV_SAMPLE_FMT_S32 ? mlt_audio_s32le - : self->audio_codec[ index ]->sample_fmt == AV_SAMPLE_FMT_FLT ? mlt_audio_f32le - : mlt_audio_s16; + *format = pick_audio_format( self->audio_codec[ index ]->sample_fmt ); sizeof_sample = sample_bytes( self->audio_codec[ index ] ); } else if ( self->audio_index == INT_MAX ) { - // This only works if all audio tracks have the same sample format. for ( index = 0; index < index_max; index++ ) - if ( self->audio_codec[ index ] && !self->audio_resample[ index ] ) + if ( self->audio_codec[ index ] ) { - *format = self->audio_codec[ index ]->sample_fmt == AV_SAMPLE_FMT_S32 ? mlt_audio_s32le - : self->audio_codec[ index ]->sample_fmt == AV_SAMPLE_FMT_FLT ? mlt_audio_f32le - : mlt_audio_s16; + // XXX: This only works if all audio tracks have the same sample format. + *format = pick_audio_format( self->audio_codec[ index ]->sample_fmt ); sizeof_sample = sample_bytes( self->audio_codec[ index ] ); break; } @@ -2473,6 +2453,7 @@ static int producer_get_audio( mlt_frame frame, void **buffer, mlt_audio_format } else { +exit_get_audio: // Get silence and don't touch the context mlt_frame_get_audio( frame, buffer, format, frequency, channels, samples ); } @@ -2522,10 +2503,8 @@ static int audio_codec_init( producer_avformat self, int index, mlt_properties p // Process properties as AVOptions apply_properties( codec_context, properties, AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_DECODING_PARAM ); -#if LIBAVCODEC_VERSION_INT >= ((52<<16)+(122<<8)+0) if ( codec && codec->priv_class && codec_context->priv_data ) apply_properties( codec_context->priv_data, properties, AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_DECODING_PARAM ); -#endif } return self->audio_codec[ index ] && self->audio_index > -1; } @@ -2577,6 +2556,11 @@ static void producer_set_up_audio( producer_avformat self, mlt_frame frame ) index = self->audio_index; mlt_properties_set_int( properties, "audio_index", index ); } + if ( context && index > -1 && index < INT_MAX && + pick_audio_format( context->streams[ index ]->codec->sample_fmt ) == mlt_audio_none ) + { + index = -1; + } // Update the audio properties if the index changed if ( context && index > -1 && index != self->audio_index ) @@ -2603,14 +2587,11 @@ static void producer_set_up_audio( producer_avformat self, mlt_frame frame ) audio_codec_init( self, index, properties ); } } - else if ( context && index > -1 && audio_codec_init( self, index, properties ) ) + else if ( context && index > -1 && index < MAX_AUDIO_STREAMS && + audio_codec_init( self, index, properties ) ) { - // Set the frame properties - if ( index < MAX_AUDIO_STREAMS ) - { - mlt_properties_set_int( frame_properties, "frequency", self->audio_codec[ index ]->sample_rate ); - mlt_properties_set_int( frame_properties, "channels", self->audio_codec[ index ]->channels ); - } + mlt_properties_set_int( frame_properties, "audio_frequency", self->audio_codec[ index ]->sample_rate ); + mlt_properties_set_int( frame_properties, "audio_channels", self->audio_codec[ index ]->channels ); } if ( context && index > -1 ) { @@ -2664,7 +2645,7 @@ static int producer_get_frame( mlt_producer producer, mlt_frame_ptr frame, int i // Set the position of this producer mlt_position position = self->seekable ? mlt_producer_frame( producer ) : self->nonseek_position++; - mlt_properties_set_position( MLT_FRAME_PROPERTIES( *frame ), "avformat_position", position ); + mlt_properties_set_position( MLT_FRAME_PROPERTIES( *frame ), "original_position", position ); // Calculate the next timecode mlt_producer_prepare_next( producer ); @@ -2677,13 +2658,14 @@ static void producer_avformat_close( producer_avformat self ) mlt_log_debug( NULL, "producer_avformat_close\n" ); // Cleanup av contexts - av_free( self->av_frame ); - pthread_mutex_lock( &self->open_mutex ); + av_free_packet( &self->pkt ); + av_free( self->video_frame ); + av_free( self->audio_frame ); + if ( self->is_mutex_init ) + pthread_mutex_lock( &self->open_mutex ); int i; for ( i = 0; i < MAX_AUDIO_STREAMS; i++ ) { - if ( self->audio_resample[i] ) - audio_resample_close( self->audio_resample[i] ); mlt_pool_release( self->audio_buffer[i] ); av_free( self->decode_buffer[i] ); if ( self->audio_codec[i] ) @@ -2709,7 +2691,8 @@ static void producer_avformat_close( producer_avformat self ) if ( self->video_format ) av_close_input_file( self->video_format ); #endif - pthread_mutex_unlock( &self->open_mutex ); + if ( self->is_mutex_init ) + pthread_mutex_unlock( &self->open_mutex ); #ifdef VDPAU vdpau_producer_close( self ); #endif @@ -2717,10 +2700,13 @@ static void producer_avformat_close( producer_avformat self ) mlt_cache_close( self->image_cache ); // Cleanup the mutexes - pthread_mutex_destroy( &self->audio_mutex ); - pthread_mutex_destroy( &self->video_mutex ); - pthread_mutex_destroy( &self->packets_mutex ); - pthread_mutex_destroy( &self->open_mutex ); + if ( self->is_mutex_init ) + { + pthread_mutex_destroy( &self->audio_mutex ); + pthread_mutex_destroy( &self->video_mutex ); + pthread_mutex_destroy( &self->packets_mutex ); + pthread_mutex_destroy( &self->open_mutex ); + } // Cleanup the packet queues AVPacket *pkt;