* consumer_avformat.c -- an encoder based on avformat
* Copyright (C) 2003-2004 Ushodaya Enterprises Limited
* Author: Charles Yates <charles.yates@pandora.be>
+ * Much code borrowed from ffmpeg.c: Copyright (c) 2000-2003 Fabrice Bellard
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
// mlt Header files
#include <framework/mlt_consumer.h>
#include <framework/mlt_frame.h>
+#include <framework/mlt_profile.h>
+#include <framework/mlt_log.h>
+#include <framework/mlt_events.h>
// System header files
#include <stdio.h>
#include <limits.h>
#include <pthread.h>
#include <sys/time.h>
-#include <math.h>
+#include <unistd.h>
// avformat header files
-#include <avformat.h>
+#include <libavformat/avformat.h>
#ifdef SWSCALE
-#include <swscale.h>
+#include <libswscale/swscale.h>
#endif
+#include <libavcodec/opt.h>
+#if LIBAVUTIL_VERSION_INT >= ((50<<16)+(8<<8)+0)
+#include <libavutil/pixdesc.h>
+#endif
+
+#if LIBAVUTIL_VERSION_INT < (50<<16)
+#define PIX_FMT_RGB32 PIX_FMT_RGBA32
+#define PIX_FMT_YUYV422 PIX_FMT_YUV422
+#endif
+
+#define MAX_AUDIO_STREAMS (8)
+#define AUDIO_ENCODE_BUFFER_SIZE (48000 * 2 * MAX_AUDIO_STREAMS)
+#define AUDIO_BUFFER_SIZE (1024 * 42)
+#define VIDEO_BUFFER_SIZE (2048 * 1024)
+
+void avformat_lock( );
+void avformat_unlock( );
//
// This structure should be extended and made globally available in mlt
sample_fifo sample_fifo_init( int frequency, int channels )
{
- sample_fifo this = calloc( 1, sizeof( sample_fifo_s ) );
- this->frequency = frequency;
- this->channels = channels;
- return this;
+ sample_fifo fifo = calloc( 1, sizeof( sample_fifo_s ) );
+ fifo->frequency = frequency;
+ fifo->channels = channels;
+ return fifo;
}
// sample_fifo_clear and check are temporarily aborted (not working as intended)
-void sample_fifo_clear( sample_fifo this, double time )
+void sample_fifo_clear( sample_fifo fifo, double time )
{
- int words = ( float )( time - this->time ) * this->frequency * this->channels;
- if ( ( int )( ( float )time * 100 ) < ( int )( ( float )this->time * 100 ) && this->used > words && words > 0 )
+ int words = ( float )( time - fifo->time ) * fifo->frequency * fifo->channels;
+ if ( ( int )( ( float )time * 100 ) < ( int )( ( float )fifo->time * 100 ) && fifo->used > words && words > 0 )
{
- memmove( this->buffer, &this->buffer[ words ], ( this->used - words ) * sizeof( int16_t ) );
- this->used -= words;
- this->time = time;
+ memmove( fifo->buffer, &fifo->buffer[ words ], ( fifo->used - words ) * sizeof( int16_t ) );
+ fifo->used -= words;
+ fifo->time = time;
}
- else if ( ( int )( ( float )time * 100 ) != ( int )( ( float )this->time * 100 ) )
+ else if ( ( int )( ( float )time * 100 ) != ( int )( ( float )fifo->time * 100 ) )
{
- this->used = 0;
- this->time = time;
+ fifo->used = 0;
+ fifo->time = time;
}
}
-void sample_fifo_check( sample_fifo this, double time )
+void sample_fifo_check( sample_fifo fifo, double time )
{
- if ( this->used == 0 )
+ if ( fifo->used == 0 )
{
- if ( ( int )( ( float )time * 100 ) < ( int )( ( float )this->time * 100 ) )
- this->time = time;
+ if ( ( int )( ( float )time * 100 ) < ( int )( ( float )fifo->time * 100 ) )
+ fifo->time = time;
}
}
-void sample_fifo_append( sample_fifo this, int16_t *samples, int count )
+void sample_fifo_append( sample_fifo fifo, int16_t *samples, int count )
{
- if ( ( this->size - this->used ) < count )
+ if ( ( fifo->size - fifo->used ) < count )
{
- this->size += count * 5;
- this->buffer = realloc( this->buffer, this->size * sizeof( int16_t ) );
+ fifo->size += count * 5;
+ fifo->buffer = realloc( fifo->buffer, fifo->size * sizeof( int16_t ) );
}
- memcpy( &this->buffer[ this->used ], samples, count * sizeof( int16_t ) );
- this->used += count;
+ memcpy( &fifo->buffer[ fifo->used ], samples, count * sizeof( int16_t ) );
+ fifo->used += count;
}
-int sample_fifo_used( sample_fifo this )
+int sample_fifo_used( sample_fifo fifo )
{
- return this->used;
+ return fifo->used;
}
-int sample_fifo_fetch( sample_fifo this, int16_t *samples, int count )
+int sample_fifo_fetch( sample_fifo fifo, int16_t *samples, int count )
{
- if ( count > this->used )
- count = this->used;
+ if ( count > fifo->used )
+ count = fifo->used;
- memcpy( samples, this->buffer, count * sizeof( int16_t ) );
- this->used -= count;
- memmove( this->buffer, &this->buffer[ count ], this->used * sizeof( int16_t ) );
+ memcpy( samples, fifo->buffer, count * sizeof( int16_t ) );
+ fifo->used -= count;
+ memmove( fifo->buffer, &fifo->buffer[ count ], fifo->used * sizeof( int16_t ) );
- this->time += ( double )count / this->channels / this->frequency;
+ fifo->time += ( double )count / fifo->channels / fifo->frequency;
return count;
}
-void sample_fifo_close( sample_fifo this )
+void sample_fifo_close( sample_fifo fifo )
{
- free( this->buffer );
- free( this );
+ free( fifo->buffer );
+ free( fifo );
}
// Forward references.
-static int consumer_start( mlt_consumer this );
-static int consumer_stop( mlt_consumer this );
-static int consumer_is_stopped( mlt_consumer this );
+static int consumer_start( mlt_consumer consumer );
+static int consumer_stop( mlt_consumer consumer );
+static int consumer_is_stopped( mlt_consumer consumer );
static void *consumer_thread( void *arg );
-static void consumer_close( mlt_consumer this );
+static void consumer_close( mlt_consumer consumer );
-/** Initialise the dv consumer.
+/** Initialise the consumer.
*/
mlt_consumer consumer_avformat_init( mlt_profile profile, char *arg )
{
// Allocate the consumer
- mlt_consumer this = mlt_consumer_new( profile );
+ mlt_consumer consumer = mlt_consumer_new( profile );
// If memory allocated and initialises without error
- if ( this != NULL )
+ if ( consumer != NULL )
{
// Get properties from the consumer
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
// Assign close callback
- this->close = consumer_close;
+ consumer->close = consumer_close;
// Interpret the argument
if ( arg != NULL )
// sample and frame queue
mlt_properties_set_data( properties, "frame_queue", mlt_deque_init( ), 0, ( mlt_destructor )mlt_deque_close, NULL );
- // Set avformat defaults (all lifted from ffmpeg.c)
- mlt_properties_set_int( properties, "audio_bit_rate", 128000 );
- mlt_properties_set_int( properties, "video_bit_rate", 200 * 1000 );
- mlt_properties_set_int( properties, "video_bit_rate_tolerance", 4000 * 1000 );
- mlt_properties_set_int( properties, "gop_size", 12 );
- mlt_properties_set_int( properties, "b_frames", 0 );
- mlt_properties_set_int( properties, "mb_decision", FF_MB_DECISION_SIMPLE );
- mlt_properties_set_double( properties, "qscale", 0 );
- mlt_properties_set_int( properties, "me_method", ME_EPZS );
- mlt_properties_set_int( properties, "mb_cmp", FF_CMP_SAD );
- mlt_properties_set_int( properties, "ildct_cmp", FF_CMP_VSAD );
- mlt_properties_set_int( properties, "sub_cmp", FF_CMP_SAD );
- mlt_properties_set_int( properties, "cmp", FF_CMP_SAD );
- mlt_properties_set_int( properties, "pre_cmp", FF_CMP_SAD );
- mlt_properties_set_int( properties, "pre_me", 0 );
- mlt_properties_set_double( properties, "lumi_mask", 0 );
- mlt_properties_set_double( properties, "dark_mask", 0 );
- mlt_properties_set_double( properties, "scplx_mask", 0 );
- mlt_properties_set_double( properties, "tcplx_mask", 0 );
- mlt_properties_set_double( properties, "p_mask", 0 );
- mlt_properties_set_int( properties, "qns", 0 );
- mlt_properties_set_int( properties, "video_qmin", 2 );
- mlt_properties_set_int( properties, "video_qmax", 31 );
- mlt_properties_set_int( properties, "video_lmin", 2*FF_QP2LAMBDA );
- mlt_properties_set_int( properties, "video_lmax", 31*FF_QP2LAMBDA );
- mlt_properties_set_int( properties, "video_mb_qmin", 2 );
- mlt_properties_set_int( properties, "video_mb_qmax", 31 );
- mlt_properties_set_int( properties, "video_qdiff", 3 );
- mlt_properties_set_double( properties, "video_qblur", 0.5 );
- mlt_properties_set_double( properties, "video_qcomp", 0.5 );
- mlt_properties_set_int( properties, "video_rc_max_rate", 0 );
- mlt_properties_set_int( properties, "video_rc_min_rate", 0 );
- mlt_properties_set_int( properties, "video_rc_buffer_size", 0 );
- mlt_properties_set_double( properties, "video_rc_buffer_aggressivity", 1.0 );
- mlt_properties_set_double( properties, "video_rc_initial_cplx", 0 );
- mlt_properties_set_double( properties, "video_i_qfactor", -0.8 );
- mlt_properties_set_double( properties, "video_b_qfactor", 1.25 );
- mlt_properties_set_double( properties, "video_i_qoffset", 0 );
- mlt_properties_set_double( properties, "video_b_qoffset", 1.25 );
- mlt_properties_set_int( properties, "video_intra_quant_bias", FF_DEFAULT_QUANT_BIAS );
- mlt_properties_set_int( properties, "video_inter_quant_bias", FF_DEFAULT_QUANT_BIAS );
- mlt_properties_set_int( properties, "dct_algo", 0 );
- mlt_properties_set_int( properties, "idct_algo", 0 );
- mlt_properties_set_int( properties, "me_threshold", 0 );
- mlt_properties_set_int( properties, "mb_threshold", 0 );
- mlt_properties_set_int( properties, "intra_dc_precision", 0 );
- mlt_properties_set_int( properties, "strict", 0 );
- mlt_properties_set_int( properties, "error_rate", 0 );
- mlt_properties_set_int( properties, "noise_reduction", 0 );
- mlt_properties_set_int( properties, "sc_threshold", 0 );
- mlt_properties_set_int( properties, "me_range", 0 );
- mlt_properties_set_int( properties, "coder", 0 );
- mlt_properties_set_int( properties, "context", 0 );
- mlt_properties_set_int( properties, "predictor", 0 );
- mlt_properties_set_int( properties, "ildct", 0 );
- mlt_properties_set_int( properties, "ilme", 0 );
+ // Audio options not fully handled by AVOptions
+#define QSCALE_NONE (-99999)
+ mlt_properties_set_int( properties, "aq", QSCALE_NONE );
+
+ // Video options not fully handled by AVOptions
+ mlt_properties_set_int( properties, "dc", 8 );
+
+ // Muxer options not fully handled by AVOptions
+ mlt_properties_set_double( properties, "muxdelay", 0.7 );
+ mlt_properties_set_double( properties, "muxpreload", 0.5 );
// Ensure termination at end of the stream
mlt_properties_set_int( properties, "terminate_on_pause", 1 );
+
+ // Default to separate processing threads for producer and consumer with no frame dropping!
+ mlt_properties_set_int( properties, "real_time", -1 );
+ mlt_properties_set_int( properties, "prefill", 1 );
// Set up start/stop/terminated callbacks
- this->start = consumer_start;
- this->stop = consumer_stop;
- this->is_stopped = consumer_is_stopped;
+ consumer->start = consumer_start;
+ consumer->stop = consumer_stop;
+ consumer->is_stopped = consumer_is_stopped;
+
+ mlt_events_register( properties, "consumer-fatal-error", NULL );
}
- // Return this
- return this;
+ // Return consumer
+ return consumer;
}
/** Start the consumer.
*/
-static int consumer_start( mlt_consumer this )
+static int consumer_start( mlt_consumer consumer )
{
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
+ int error = 0;
+
+ // Report information about available muxers and codecs as YAML Tiny
+ char *s = mlt_properties_get( properties, "f" );
+ if ( s && strcmp( s, "list" ) == 0 )
+ {
+ mlt_properties doc = mlt_properties_new();
+ mlt_properties formats = mlt_properties_new();
+ char key[20];
+ AVOutputFormat *format = NULL;
+
+ mlt_properties_set_data( properties, "f", formats, 0, (mlt_destructor) mlt_properties_close, NULL );
+ mlt_properties_set_data( doc, "formats", formats, 0, NULL, NULL );
+ while ( ( format = av_oformat_next( format ) ) )
+ {
+ snprintf( key, sizeof(key), "%d", mlt_properties_count( formats ) );
+ mlt_properties_set( formats, key, format->name );
+ }
+ fprintf( stderr, "%s", mlt_properties_serialise_yaml( doc ) );
+ mlt_properties_close( doc );
+ error = 1;
+ }
+ s = mlt_properties_get( properties, "acodec" );
+ if ( s && strcmp( s, "list" ) == 0 )
+ {
+ mlt_properties doc = mlt_properties_new();
+ mlt_properties codecs = mlt_properties_new();
+ char key[20];
+ AVCodec *codec = NULL;
+
+ mlt_properties_set_data( properties, "acodec", codecs, 0, (mlt_destructor) mlt_properties_close, NULL );
+ mlt_properties_set_data( doc, "audio_codecs", codecs, 0, NULL, NULL );
+ while ( ( codec = av_codec_next( codec ) ) )
+ if ( codec->encode && codec->type == CODEC_TYPE_AUDIO )
+ {
+ snprintf( key, sizeof(key), "%d", mlt_properties_count( codecs ) );
+ mlt_properties_set( codecs, key, codec->name );
+ }
+ fprintf( stderr, "%s", mlt_properties_serialise_yaml( doc ) );
+ mlt_properties_close( doc );
+ error = 1;
+ }
+ s = mlt_properties_get( properties, "vcodec" );
+ if ( s && strcmp( s, "list" ) == 0 )
+ {
+ mlt_properties doc = mlt_properties_new();
+ mlt_properties codecs = mlt_properties_new();
+ char key[20];
+ AVCodec *codec = NULL;
+
+ mlt_properties_set_data( properties, "vcodec", codecs, 0, (mlt_destructor) mlt_properties_close, NULL );
+ mlt_properties_set_data( doc, "video_codecs", codecs, 0, NULL, NULL );
+ while ( ( codec = av_codec_next( codec ) ) )
+ if ( codec->encode && codec->type == CODEC_TYPE_VIDEO )
+ {
+ snprintf( key, sizeof(key), "%d", mlt_properties_count( codecs ) );
+ mlt_properties_set( codecs, key, codec->name );
+ }
+ fprintf( stderr, "%s", mlt_properties_serialise_yaml( doc ) );
+ mlt_properties_close( doc );
+ error = 1;
+ }
// Check that we're not already running
- if ( !mlt_properties_get_int( properties, "running" ) )
+ if ( !error && !mlt_properties_get_int( properties, "running" ) )
{
// Allocate a thread
pthread_t *thread = calloc( 1, sizeof( pthread_t ) );
int height = mlt_properties_get_int( properties, "height" );
// Obtain the size property
- char *size = mlt_properties_get( properties, "size" );
+ char *size = mlt_properties_get( properties, "s" );
// Interpret it
if ( size != NULL )
}
else
{
- fprintf( stderr, "consumer_avformat: Invalid size property %s - ignoring.\n", size );
+ mlt_log_warning( MLT_CONSUMER_SERVICE( consumer ), "Invalid size property %s - ignoring.\n", size );
}
}
// Now ensure we honour the multiple of two requested by libavformat
- mlt_properties_set_int( properties, "width", ( width / 2 ) * 2 );
- mlt_properties_set_int( properties, "height", ( height / 2 ) * 2 );
+ width = ( width / 2 ) * 2;
+ height = ( height / 2 ) * 2;
+ mlt_properties_set_int( properties, "width", width );
+ mlt_properties_set_int( properties, "height", height );
+
+ // We need to set these on the profile as well because the s property is
+ // an alias to mlt properties that correspond to profile settings.
+ mlt_profile profile = mlt_service_profile( MLT_CONSUMER_SERVICE( consumer ) );
+ if ( profile )
+ {
+ profile->width = width;
+ profile->height = height;
+ }
+
+ // Handle the ffmpeg command line "-r" property for frame rate
+ if ( mlt_properties_get( properties, "r" ) )
+ {
+ double frame_rate = mlt_properties_get_double( properties, "r" );
+ AVRational rational = av_d2q( frame_rate, 255 );
+ mlt_properties_set_int( properties, "frame_rate_num", rational.num );
+ mlt_properties_set_int( properties, "frame_rate_den", rational.den );
+ if ( profile )
+ {
+ profile->frame_rate_num = rational.num;
+ profile->frame_rate_den = rational.den;
+ mlt_properties_set_double( properties, "fps", mlt_profile_fps( profile ) );
+ }
+ }
+
+ // Apply AVOptions that are synonyms for standard mlt_consumer options
+ if ( mlt_properties_get( properties, "ac" ) )
+ mlt_properties_set_int( properties, "channels", mlt_properties_get_int( properties, "ac" ) );
+ if ( mlt_properties_get( properties, "ar" ) )
+ mlt_properties_set_int( properties, "frequency", mlt_properties_get_int( properties, "ar" ) );
// Assign the thread to properties
mlt_properties_set_data( properties, "thread", thread, sizeof( pthread_t ), free, NULL );
mlt_properties_set_int( properties, "running", 1 );
// Create the thread
- pthread_create( thread, NULL, consumer_thread, this );
+ pthread_create( thread, NULL, consumer_thread, consumer );
}
- return 0;
+ return error;
}
/** Stop the consumer.
*/
-static int consumer_stop( mlt_consumer this )
+static int consumer_stop( mlt_consumer consumer )
{
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
// Check that we're running
if ( mlt_properties_get_int( properties, "running" ) )
/** Determine if the consumer is stopped.
*/
-static int consumer_is_stopped( mlt_consumer this )
+static int consumer_is_stopped( mlt_consumer consumer )
{
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
return !mlt_properties_get_int( properties, "running" );
}
+/** Process properties as AVOptions and apply to AV context obj
+*/
+
+static void apply_properties( void *obj, mlt_properties properties, int flags, int alloc )
+{
+ int i;
+ int count = mlt_properties_count( properties );
+ for ( i = 0; i < count; i++ )
+ {
+ const char *opt_name = mlt_properties_get_name( properties, i );
+ const AVOption *opt = av_find_opt( obj, opt_name, NULL, flags, flags );
+ if ( opt != NULL )
+#if LIBAVCODEC_VERSION_INT >= ((52<<16)+(7<<8)+0)
+ av_set_string3( obj, opt_name, mlt_properties_get( properties, opt_name), alloc, NULL );
+#elif LIBAVCODEC_VERSION_INT >= ((51<<16)+(59<<8)+0)
+ av_set_string2( obj, opt_name, mlt_properties_get( properties, opt_name), alloc );
+#else
+ av_set_string( obj, opt_name, mlt_properties_get( properties, opt_name) );
+#endif
+ }
+}
+
/** Add an audio output stream
*/
-static AVStream *add_audio_stream( mlt_consumer this, AVFormatContext *oc, int codec_id )
+static AVStream *add_audio_stream( mlt_consumer consumer, AVFormatContext *oc, int codec_id, int channels )
{
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
// Create a new stream
- AVStream *st = av_new_stream( oc, 1 );
+ AVStream *st = av_new_stream( oc, oc->nb_streams );
// If created, then initialise from properties
if ( st != NULL )
{
AVCodecContext *c = st->codec;
- int thread_count = mlt_properties_get_int( properties, "threads" );
- if ( thread_count == 0 && getenv( "MLT_AVFORMAT_THREADS" ) )
- thread_count = atoi( getenv( "MLT_AVFORMAT_THREADS" ) );
+
+ // Establish defaults from AVOptions
+ avcodec_get_context_defaults2( c, CODEC_TYPE_AUDIO );
c->codec_id = codec_id;
c->codec_type = CODEC_TYPE_AUDIO;
+ c->sample_fmt = SAMPLE_FMT_S16;
- // Put sample parameters
- c->bit_rate = mlt_properties_get_int( properties, "audio_bit_rate" );
- c->sample_rate = mlt_properties_get_int( properties, "frequency" );
- c->channels = mlt_properties_get_int( properties, "channels" );
-
- if (oc->oformat->flags & AVFMT_GLOBALHEADER)
- c->flags |= CODEC_FLAG_GLOBAL_HEADER;
-
+#if 0 // disabled until some audio codecs are multi-threaded
+ // Setup multi-threading
+ int thread_count = mlt_properties_get_int( properties, "threads" );
+ if ( thread_count == 0 && getenv( "MLT_AVFORMAT_THREADS" ) )
+ thread_count = atoi( getenv( "MLT_AVFORMAT_THREADS" ) );
+ if ( thread_count > 1 )
+ avcodec_thread_init( c, thread_count );
+#endif
+
+ if (oc->oformat->flags & AVFMT_GLOBALHEADER)
+ c->flags |= CODEC_FLAG_GLOBAL_HEADER;
+
// Allow the user to override the audio fourcc
- if ( mlt_properties_get( properties, "afourcc" ) )
+ if ( mlt_properties_get( properties, "atag" ) )
{
char *tail = NULL;
- char *arg = mlt_properties_get( properties, "afourcc" );
- int tag = strtol( arg, &tail, 0);
- if( !tail || *tail )
- tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
+ char *arg = mlt_properties_get( properties, "atag" );
+ int tag = strtol( arg, &tail, 0);
+ if( !tail || *tail )
+ tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
c->codec_tag = tag;
}
- if ( thread_count > 1 )
+
+ // Process properties as AVOptions
+ char *apre = mlt_properties_get( properties, "apre" );
+ if ( apre )
{
- avcodec_thread_init( c, thread_count );
- c->thread_count = thread_count;
+ mlt_properties p = mlt_properties_load( apre );
+ apply_properties( c, p, AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_ENCODING_PARAM, 1 );
+ mlt_properties_close( p );
}
+ apply_properties( c, properties, AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_ENCODING_PARAM, 0 );
+
+ int audio_qscale = mlt_properties_get_int( properties, "aq" );
+ if ( audio_qscale > QSCALE_NONE )
+ {
+ c->flags |= CODEC_FLAG_QSCALE;
+ c->global_quality = st->quality = FF_QP2LAMBDA * audio_qscale;
+ }
+
+ // Set parameters controlled by MLT
+ c->sample_rate = mlt_properties_get_int( properties, "frequency" );
+ c->time_base = ( AVRational ){ 1, c->sample_rate };
+ c->channels = channels;
+
+ if ( mlt_properties_get( properties, "alang" ) != NULL )
+ strncpy( st->language, mlt_properties_get( properties, "alang" ), sizeof( st->language ) );
}
else
{
- fprintf( stderr, "Could not allocate a stream for audio\n" );
+ mlt_log_error( MLT_CONSUMER_SERVICE( consumer ), "Could not allocate a stream for audio\n" );
}
return st;
}
-static int open_audio( AVFormatContext *oc, AVStream *st, int audio_outbuf_size )
+static int open_audio( AVFormatContext *oc, AVStream *st, int audio_outbuf_size, const char *codec_name )
{
// We will return the audio input size from here
int audio_input_frame_size = 0;
AVCodecContext *c = st->codec;
// Find the encoder
- AVCodec *codec = avcodec_find_encoder( c->codec_id );
+ AVCodec *codec;
+ if ( codec_name )
+ codec = avcodec_find_encoder_by_name( codec_name );
+ else
+ codec = avcodec_find_encoder( c->codec_id );
+ avformat_lock();
+
// Continue if codec found and we can open it
- if ( codec != NULL && avcodec_open(c, codec) >= 0 )
+ if ( codec != NULL && avcodec_open( c, codec ) >= 0 )
{
// ugly hack for PCM codecs (will be removed ASAP with new PCM
// support to compute the input frame size in samples
}
// Some formats want stream headers to be seperate (hmm)
- if( !strcmp( oc->oformat->name, "mp4" ) ||
- !strcmp( oc->oformat->name, "mov" ) ||
- !strcmp( oc->oformat->name, "3gp" ) )
+ if ( !strcmp( oc->oformat->name, "mp4" ) ||
+ !strcmp( oc->oformat->name, "mov" ) ||
+ !strcmp( oc->oformat->name, "3gp" ) )
c->flags |= CODEC_FLAG_GLOBAL_HEADER;
}
else
{
- fprintf( stderr, "Unable to encode audio - disabling audio output.\n" );
+ mlt_log_warning( NULL, "%s: Unable to encode audio - disabling audio output.\n", __FILE__ );
}
+
+ avformat_unlock();
return audio_input_frame_size;
}
static void close_audio( AVFormatContext *oc, AVStream *st )
{
- avcodec_close( st->codec );
+ if ( st && st->codec )
+ {
+ avformat_lock();
+ avcodec_close( st->codec );
+ avformat_unlock();
+ }
}
/** Add a video output stream
*/
-static AVStream *add_video_stream( mlt_consumer this, AVFormatContext *oc, int codec_id )
+static AVStream *add_video_stream( mlt_consumer consumer, AVFormatContext *oc, int codec_id )
{
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
// Create a new stream
- AVStream *st = av_new_stream( oc, 0 );
+ AVStream *st = av_new_stream( oc, oc->nb_streams );
if ( st != NULL )
{
char *pix_fmt = mlt_properties_get( properties, "pix_fmt" );
- double ar = mlt_properties_get_double( properties, "display_ratio" );
AVCodecContext *c = st->codec;
- int thread_count = mlt_properties_get_int( properties, "threads" );
- if ( thread_count == 0 && getenv( "MLT_AVFORMAT_THREADS" ) )
- thread_count = atoi( getenv( "MLT_AVFORMAT_THREADS" ) );
+
+ // Establish defaults from AVOptions
+ avcodec_get_context_defaults2( c, CODEC_TYPE_VIDEO );
c->codec_id = codec_id;
c->codec_type = CODEC_TYPE_VIDEO;
+
+ // Setup multi-threading
+ int thread_count = mlt_properties_get_int( properties, "threads" );
+ if ( thread_count == 0 && getenv( "MLT_AVFORMAT_THREADS" ) )
+ thread_count = atoi( getenv( "MLT_AVFORMAT_THREADS" ) );
+ if ( thread_count > 1 )
+ avcodec_thread_init( c, thread_count );
+
+ // Process properties as AVOptions
+ char *vpre = mlt_properties_get( properties, "vpre" );
+ if ( vpre )
+ {
+ mlt_properties p = mlt_properties_load( vpre );
+#ifdef AVDATADIR
+ if ( mlt_properties_count( p ) < 1 )
+ {
+ AVCodec *codec = avcodec_find_encoder( c->codec_id );
+ if ( codec )
+ {
+ char *path = malloc( strlen(AVDATADIR) + strlen(codec->name) + strlen(vpre) + strlen(".ffpreset") + 2 );
+ strcpy( path, AVDATADIR );
+ strcat( path, codec->name );
+ strcat( path, "-" );
+ strcat( path, vpre );
+ strcat( path, ".ffpreset" );
+
+ mlt_properties_close( p );
+ p = mlt_properties_load( path );
+ mlt_properties_debug( p, path, stderr );
+ free( path );
+ }
+ }
+ else
+ {
+ mlt_properties_debug( p, vpre, stderr );
+ }
+#endif
+ apply_properties( c, p, AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM, 1 );
+ mlt_properties_close( p );
+ }
+ int colorspace = mlt_properties_get_int( properties, "colorspace" );
+ mlt_properties_set( properties, "colorspace", NULL );
+ apply_properties( c, properties, AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM, 0 );
+ mlt_properties_set_int( properties, "colorspace", colorspace );
- // put sample parameters
- c->bit_rate = mlt_properties_get_int( properties, "video_bit_rate" );
- c->bit_rate_tolerance = mlt_properties_get_int( properties, "video_bit_rate_tolerance" );
+ // Set options controlled by MLT
c->width = mlt_properties_get_int( properties, "width" );
c->height = mlt_properties_get_int( properties, "height" );
c->time_base.num = mlt_properties_get_int( properties, "frame_rate_den" );
c->time_base.den = mlt_properties_get_int( properties, "frame_rate_num" );
- c->gop_size = mlt_properties_get_int( properties, "gop_size" );
+ if ( st->time_base.den == 0 )
+ st->time_base = c->time_base;
+#if LIBAVUTIL_VERSION_INT >= ((50<<16)+(8<<8)+0)
+ c->pix_fmt = pix_fmt ? av_get_pix_fmt( pix_fmt ) : PIX_FMT_YUV420P;
+#else
c->pix_fmt = pix_fmt ? avcodec_get_pix_fmt( pix_fmt ) : PIX_FMT_YUV420P;
-
- if ( mlt_properties_get_int( properties, "b_frames" ) )
+#endif
+
+#if LIBAVCODEC_VERSION_INT > ((52<<16)+(28<<8)+0)
+ switch ( colorspace )
{
- c->max_b_frames = mlt_properties_get_int( properties, "b_frames" );
- c->b_frame_strategy = 0;
- c->b_quant_factor = 2.0;
+ case 170:
+ c->colorspace = AVCOL_SPC_SMPTE170M;
+ break;
+ case 240:
+ c->colorspace = AVCOL_SPC_SMPTE240M;
+ break;
+ case 470:
+ c->colorspace = AVCOL_SPC_BT470BG;
+ break;
+ case 601:
+ c->colorspace = ( 576 % c->height ) ? AVCOL_SPC_SMPTE170M : AVCOL_SPC_BT470BG;
+ break;
+ case 709:
+ c->colorspace = AVCOL_SPC_BT709;
+ break;
}
+#endif
- c->mb_decision = mlt_properties_get_int( properties, "mb_decision" );
- c->sample_aspect_ratio = av_d2q( ar * c->height / c->width , 255);
- c->mb_cmp = mlt_properties_get_int( properties, "mb_cmp" );
- c->ildct_cmp = mlt_properties_get_int( properties, "ildct_cmp" );
- c->me_sub_cmp = mlt_properties_get_int( properties, "sub_cmp" );
- c->me_cmp = mlt_properties_get_int( properties, "cmp" );
- c->me_pre_cmp = mlt_properties_get_int( properties, "pre_cmp" );
- c->pre_me = mlt_properties_get_int( properties, "pre_me" );
- c->lumi_masking = mlt_properties_get_double( properties, "lumi_mask" );
- c->dark_masking = mlt_properties_get_double( properties, "dark_mask" );
- c->spatial_cplx_masking = mlt_properties_get_double( properties, "scplx_mask" );
- c->temporal_cplx_masking = mlt_properties_get_double( properties, "tcplx_mask" );
- c->p_masking = mlt_properties_get_double( properties, "p_mask" );
- c->quantizer_noise_shaping= mlt_properties_get_int( properties, "qns" );
- c->qmin = mlt_properties_get_int( properties, "video_qmin" );
- c->qmax = mlt_properties_get_int( properties, "video_qmax" );
- c->lmin = mlt_properties_get_int( properties, "video_lmin" );
- c->lmax = mlt_properties_get_int( properties, "video_lmax" );
- c->mb_qmin = mlt_properties_get_int( properties, "video_mb_qmin" );
- c->mb_qmax = mlt_properties_get_int( properties, "video_mb_qmax" );
- c->max_qdiff = mlt_properties_get_int( properties, "video_qdiff" );
- c->qblur = mlt_properties_get_double( properties, "video_qblur" );
- c->qcompress = mlt_properties_get_double( properties, "video_qcomp" );
+ if ( mlt_properties_get( properties, "aspect" ) )
+ {
+ // "-aspect" on ffmpeg command line is display aspect ratio
+ double ar = mlt_properties_get_double( properties, "aspect" );
+ AVRational rational = av_d2q( ar, 255 );
+
+ // Update the profile and properties as well since this is an alias
+ // for mlt properties that correspond to profile settings
+ mlt_properties_set_int( properties, "display_aspect_num", rational.num );
+ mlt_properties_set_int( properties, "display_aspect_den", rational.den );
+ mlt_profile profile = mlt_service_profile( MLT_CONSUMER_SERVICE( consumer ) );
+ if ( profile )
+ {
+ profile->display_aspect_num = rational.num;
+ profile->display_aspect_den = rational.den;
+ mlt_properties_set_double( properties, "display_ratio", mlt_profile_dar( profile ) );
+ }
+
+ // Now compute the sample aspect ratio
+ rational = av_d2q( ar * c->height / c->width, 255 );
+ c->sample_aspect_ratio = rational;
+ // Update the profile and properties as well since this is an alias
+ // for mlt properties that correspond to profile settings
+ mlt_properties_set_int( properties, "sample_aspect_num", rational.num );
+ mlt_properties_set_int( properties, "sample_aspect_den", rational.den );
+ if ( profile )
+ {
+ profile->sample_aspect_num = rational.num;
+ profile->sample_aspect_den = rational.den;
+ mlt_properties_set_double( properties, "aspect_ratio", mlt_profile_sar( profile ) );
+ }
+ }
+ else
+ {
+ c->sample_aspect_ratio.num = mlt_properties_get_int( properties, "sample_aspect_num" );
+ c->sample_aspect_ratio.den = mlt_properties_get_int( properties, "sample_aspect_den" );
+ }
+#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(21<<8)+0)
+ st->sample_aspect_ratio = c->sample_aspect_ratio;
+#endif
if ( mlt_properties_get_double( properties, "qscale" ) > 0 )
{
}
// Allow the user to override the video fourcc
- if ( mlt_properties_get( properties, "vfourcc" ) )
+ if ( mlt_properties_get( properties, "vtag" ) )
{
char *tail = NULL;
- const char *arg = mlt_properties_get( properties, "vfourcc" );
- int tag = strtol( arg, &tail, 0);
- if( !tail || *tail )
- tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
+ const char *arg = mlt_properties_get( properties, "vtag" );
+ int tag = strtol( arg, &tail, 0);
+ if( !tail || *tail )
+ tag = arg[ 0 ] + ( arg[ 1 ] << 8 ) + ( arg[ 2 ] << 16 ) + ( arg[ 3 ] << 24 );
c->codec_tag = tag;
}
if ( oc->oformat->flags & AVFMT_GLOBALHEADER )
c->flags |= CODEC_FLAG_GLOBAL_HEADER;
- c->rc_max_rate = mlt_properties_get_int( properties, "video_rc_max_rate" );
- c->rc_min_rate = mlt_properties_get_int( properties, "video_rc_min_rate" );
- c->rc_buffer_size = mlt_properties_get_int( properties, "video_rc_buffer_size" );
- c->rc_initial_buffer_occupancy = c->rc_buffer_size*3/4;
- c->rc_buffer_aggressivity= mlt_properties_get_double( properties, "video_rc_buffer_aggressivity" );
- c->rc_initial_cplx= mlt_properties_get_double( properties, "video_rc_initial_cplx" );
- c->i_quant_factor = mlt_properties_get_double( properties, "video_i_qfactor" );
- c->b_quant_factor = mlt_properties_get_double( properties, "video_b_qfactor" );
- c->i_quant_offset = mlt_properties_get_double( properties, "video_i_qoffset" );
- c->b_quant_offset = mlt_properties_get_double( properties, "video_b_qoffset" );
- c->intra_quant_bias = mlt_properties_get_int( properties, "video_intra_quant_bias" );
- c->inter_quant_bias = mlt_properties_get_int( properties, "video_inter_quant_bias" );
- c->dct_algo = mlt_properties_get_int( properties, "dct_algo" );
- c->idct_algo = mlt_properties_get_int( properties, "idct_algo" );
- c->me_threshold= mlt_properties_get_int( properties, "me_threshold" );
- c->mb_threshold= mlt_properties_get_int( properties, "mb_threshold" );
- c->intra_dc_precision= mlt_properties_get_int( properties, "intra_dc_precision" );
- c->strict_std_compliance = mlt_properties_get_int( properties, "strict" );
- c->error_rate = mlt_properties_get_int( properties, "error_rate" );
- c->noise_reduction= mlt_properties_get_int( properties, "noise_reduction" );
- c->scenechange_threshold= mlt_properties_get_int( properties, "sc_threshold" );
- c->me_range = mlt_properties_get_int( properties, "me_range" );
- c->coder_type= mlt_properties_get_int( properties, "coder" );
- c->context_model= mlt_properties_get_int( properties, "context" );
- c->prediction_method= mlt_properties_get_int( properties, "predictor" );
- c->me_method = mlt_properties_get_int( properties, "me_method" );
+ // Translate these standard mlt consumer properties to ffmpeg
if ( mlt_properties_get_int( properties, "progressive" ) == 0 &&
mlt_properties_get_int( properties, "deinterlace" ) == 0 )
{
- if ( mlt_properties_get_int( properties, "ildct" ) )
+ if ( ! mlt_properties_get( properties, "ildct" ) || mlt_properties_get_int( properties, "ildct" ) )
c->flags |= CODEC_FLAG_INTERLACED_DCT;
- if ( mlt_properties_get_int( properties, "ilme" ) )
+ if ( ! mlt_properties_get( properties, "ilme" ) || mlt_properties_get_int( properties, "ilme" ) )
c->flags |= CODEC_FLAG_INTERLACED_ME;
}
- if ( thread_count > 1 )
+
+ // parse the ratecontrol override string
+ int i;
+ char *rc_override = mlt_properties_get( properties, "rc_override" );
+ for ( i = 0; rc_override; i++ )
{
- avcodec_thread_init( c, thread_count );
- c->thread_count = thread_count;
+ int start, end, q;
+ int e = sscanf( rc_override, "%d,%d,%d", &start, &end, &q );
+ if ( e != 3 )
+ mlt_log_warning( MLT_CONSUMER_SERVICE( consumer ), "Error parsing rc_override\n" );
+ c->rc_override = av_realloc( c->rc_override, sizeof( RcOverride ) * ( i + 1 ) );
+ c->rc_override[i].start_frame = start;
+ c->rc_override[i].end_frame = end;
+ if ( q > 0 )
+ {
+ c->rc_override[i].qscale = q;
+ c->rc_override[i].quality_factor = 1.0;
+ }
+ else
+ {
+ c->rc_override[i].qscale = 0;
+ c->rc_override[i].quality_factor = -q / 100.0;
+ }
+ rc_override = strchr( rc_override, '/' );
+ if ( rc_override )
+ rc_override++;
+ }
+ c->rc_override_count = i;
+ if ( !c->rc_initial_buffer_occupancy )
+ c->rc_initial_buffer_occupancy = c->rc_buffer_size * 3/4;
+ c->intra_dc_precision = mlt_properties_get_int( properties, "dc" ) - 8;
+
+ // Setup dual-pass
+ i = mlt_properties_get_int( properties, "pass" );
+ if ( i == 1 )
+ c->flags |= CODEC_FLAG_PASS1;
+ else if ( i == 2 )
+ c->flags |= CODEC_FLAG_PASS2;
+ if ( codec_id != CODEC_ID_H264 && ( c->flags & ( CODEC_FLAG_PASS1 | CODEC_FLAG_PASS2 ) ) )
+ {
+ char logfilename[1024];
+ FILE *f;
+ int size;
+ char *logbuffer;
+
+ snprintf( logfilename, sizeof(logfilename), "%s_2pass.log",
+ mlt_properties_get( properties, "passlogfile" ) ? mlt_properties_get( properties, "passlogfile" ) : mlt_properties_get( properties, "target" ) );
+ if ( c->flags & CODEC_FLAG_PASS1 )
+ {
+ f = fopen( logfilename, "w" );
+ if ( !f )
+ perror( logfilename );
+ else
+ mlt_properties_set_data( properties, "_logfile", f, 0, ( mlt_destructor )fclose, NULL );
+ }
+ else
+ {
+ /* read the log file */
+ f = fopen( logfilename, "r" );
+ if ( !f )
+ {
+ perror(logfilename);
+ }
+ else
+ {
+ mlt_properties_set( properties, "_logfilename", logfilename );
+ fseek( f, 0, SEEK_END );
+ size = ftell( f );
+ fseek( f, 0, SEEK_SET );
+ logbuffer = av_malloc( size + 1 );
+ if ( !logbuffer )
+ mlt_log_fatal( MLT_CONSUMER_SERVICE( consumer ), "Could not allocate log buffer\n" );
+ else
+ {
+ size = fread( logbuffer, 1, size, f );
+ fclose( f );
+ logbuffer[size] = '\0';
+ c->stats_in = logbuffer;
+ mlt_properties_set_data( properties, "_logbuffer", logbuffer, 0, ( mlt_destructor )av_free, NULL );
+ }
+ }
+ }
}
}
else
{
- fprintf( stderr, "Could not allocate a stream for video\n" );
+ mlt_log_error( MLT_CONSUMER_SERVICE( consumer ), "Could not allocate a stream for video\n" );
}
return st;
AVFrame *picture = avcodec_alloc_frame();
// Determine size of the
- int size = avpicture_get_size(pix_fmt, width, height);
+ int size = avpicture_get_size(pix_fmt, width, height);
// Allocate the picture buf
- uint8_t *picture_buf = av_malloc(size);
+ uint8_t *picture_buf = av_malloc(size);
// If we have both, then fill the image
if ( picture != NULL && picture_buf != NULL )
return picture;
}
-static int open_video(AVFormatContext *oc, AVStream *st)
+static int open_video(AVFormatContext *oc, AVStream *st, const char *codec_name)
{
// Get the codec
AVCodecContext *video_enc = st->codec;
// find the video encoder
- AVCodec *codec = avcodec_find_encoder( video_enc->codec_id );
+ AVCodec *codec;
+ if ( codec_name )
+ codec = avcodec_find_encoder_by_name( codec_name );
+ else
+ codec = avcodec_find_encoder( video_enc->codec_id );
if( codec && codec->pix_fmts )
{
}
// Open the codec safely
- return codec != NULL && avcodec_open( video_enc, codec ) >= 0;
+ avformat_lock();
+ int result = codec != NULL && avcodec_open( video_enc, codec ) >= 0;
+ avformat_unlock();
+
+ return result;
}
void close_video(AVFormatContext *oc, AVStream *st)
{
- avcodec_close(st->codec);
+ if ( st && st->codec )
+ {
+ avformat_lock();
+ avcodec_close(st->codec);
+ avformat_unlock();
+ }
}
static inline long time_difference( struct timeval *time1 )
static void *consumer_thread( void *arg )
{
// Map the argument to the object
- mlt_consumer this = arg;
+ mlt_consumer consumer = arg;
// Get the properties
- mlt_properties properties = MLT_CONSUMER_PROPERTIES( this );
+ mlt_properties properties = MLT_CONSUMER_PROPERTIES( consumer );
// Get the terminate on pause property
int terminate_on_pause = mlt_properties_get_int( properties, "terminate_on_pause" );
int img_height = height;
// Get default audio properties
- mlt_audio_format aud_fmt = mlt_audio_pcm;
+ mlt_audio_format aud_fmt = mlt_audio_s16;
int channels = mlt_properties_get_int( properties, "channels" );
+ int total_channels = channels;
int frequency = mlt_properties_get_int( properties, "frequency" );
int16_t *pcm = NULL;
int samples = 0;
// AVFormat audio buffer and frame size
- int audio_outbuf_size = 10000;
- uint8_t *audio_outbuf = av_malloc( audio_outbuf_size );
+ int audio_outbuf_size = AUDIO_BUFFER_SIZE;
+ uint8_t *audio_outbuf = av_malloc( audio_outbuf_size );
int audio_input_frame_size = 0;
// AVFormat video buffer and frame count
int frame_count = 0;
- int video_outbuf_size = ( 1024 * 1024 );
+ int video_outbuf_size = VIDEO_BUFFER_SIZE;
uint8_t *video_outbuf = av_malloc( video_outbuf_size );
// Used for the frame properties
// Need two av pictures for converting
AVFrame *output = NULL;
- AVFrame *input = alloc_picture( PIX_FMT_YUV422, width, height );
+ AVFrame *input = alloc_picture( PIX_FMT_YUYV422, width, height );
// For receiving images from an mlt_frame
uint8_t *image;
mlt_image_format img_fmt = mlt_image_yuv422;
// For receiving audio samples back from the fifo
- int16_t *buffer = av_malloc( 48000 * 2 );
+ int16_t *audio_buf_1 = av_malloc( AUDIO_ENCODE_BUFFER_SIZE );
+ int16_t *audio_buf_2 = NULL;
int count = 0;
// Allocate the context
+#if (LIBAVFORMAT_VERSION_INT >= ((52<<16)+(26<<8)+0))
+ AVFormatContext *oc = avformat_alloc_context( );
+#else
AVFormatContext *oc = av_alloc_format_context( );
+#endif
// Streams
- AVStream *audio_st = NULL;
AVStream *video_st = NULL;
+ AVStream *audio_st[ MAX_AUDIO_STREAMS ];
// Time stamps
double audio_pts = 0;
double video_pts = 0;
- // Loop variable
- int i;
-
- // Frames despatched
+ // Frames dispatched
long int frames = 0;
long int total_time = 0;
// Determine the format
AVOutputFormat *fmt = NULL;
- char *filename = mlt_properties_get( properties, "target" );
- char *format = mlt_properties_get( properties, "format" );
+ const char *filename = mlt_properties_get( properties, "target" );
+ char *format = mlt_properties_get( properties, "f" );
char *vcodec = mlt_properties_get( properties, "vcodec" );
char *acodec = mlt_properties_get( properties, "acodec" );
-
+
// Used to store and override codec ids
int audio_codec_id;
int video_codec_id;
+ // Misc
+ char key[27];
+ mlt_properties frame_meta_properties = mlt_properties_new();
+
+ // Initialize audio_st
+ int i = MAX_AUDIO_STREAMS;
+ while ( i-- )
+ audio_st[i] = NULL;
+
// Check for user selected format first
if ( format != NULL )
+#if LIBAVFORMAT_VERSION_INT < ((52<<16)+(45<<8)+0)
fmt = guess_format( format, NULL, NULL );
+#else
+ fmt = av_guess_format( format, NULL, NULL );
+#endif
// Otherwise check on the filename
if ( fmt == NULL && filename != NULL )
+#if LIBAVFORMAT_VERSION_INT < ((52<<16)+(45<<8)+0)
fmt = guess_format( NULL, filename, NULL );
+#else
+ fmt = av_guess_format( NULL, filename, NULL );
+#endif
// Otherwise default to mpeg
if ( fmt == NULL )
+#if LIBAVFORMAT_VERSION_INT < ((52<<16)+(45<<8)+0)
fmt = guess_format( "mpeg", NULL, NULL );
+#else
+ fmt = av_guess_format( "mpeg", NULL, NULL );
+#endif
// We need a filename - default to stdout?
if ( filename == NULL || !strcmp( filename, "" ) )
video_codec_id = fmt->video_codec;
// Check for audio codec overides
- if ( acodec != NULL )
+ if ( ( acodec && strcmp( acodec, "none" ) == 0 ) || mlt_properties_get_int( properties, "an" ) )
+ audio_codec_id = CODEC_ID_NONE;
+ else if ( acodec )
{
- AVCodec *p = first_avcodec;
- while( p != NULL )
- {
- if ( !strcmp( p->name, acodec ) && p->type == CODEC_TYPE_AUDIO )
- break;
- p = p->next;
- }
+ AVCodec *p = avcodec_find_encoder_by_name( acodec );
if ( p != NULL )
+ {
audio_codec_id = p->id;
+ if ( audio_codec_id == CODEC_ID_AC3 && avcodec_find_encoder_by_name( "ac3_fixed" ) )
+ {
+ mlt_properties_set( properties, "_acodec", "ac3_fixed" );
+ acodec = mlt_properties_get( properties, "_acodec" );
+ }
+ }
else
- fprintf( stderr, "consumer_avcodec: audio codec %s unrecognised - ignoring\n", acodec );
+ mlt_log_warning( MLT_CONSUMER_SERVICE( consumer ), "audio codec %s unrecognised - ignoring\n", acodec );
}
// Check for video codec overides
- if ( vcodec != NULL )
+ if ( ( vcodec && strcmp( vcodec, "none" ) == 0 ) || mlt_properties_get_int( properties, "vn" ) )
+ video_codec_id = CODEC_ID_NONE;
+ else if ( vcodec )
{
- AVCodec *p = first_avcodec;
- while( p != NULL )
- {
- if ( !strcmp( p->name, vcodec ) && p->type == CODEC_TYPE_VIDEO )
- break;
- p = p->next;
- }
+ AVCodec *p = avcodec_find_encoder_by_name( vcodec );
if ( p != NULL )
video_codec_id = p->id;
else
- fprintf( stderr, "consumer_avcodec: video codec %s unrecognised - ignoring\n", vcodec );
+ mlt_log_warning( MLT_CONSUMER_SERVICE( consumer ), "video codec %s unrecognised - ignoring\n", vcodec );
}
- // Update the output context
-
// Write metadata
+#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(31<<8)+0)
+ for ( i = 0; i < mlt_properties_count( properties ); i++ )
+ {
+ char *name = mlt_properties_get_name( properties, i );
+ if ( name && !strncmp( name, "meta.attr.", 10 ) )
+ {
+ char *key = strdup( name + 10 );
+ char *markup = strrchr( key, '.' );
+ if ( markup && !strcmp( markup, ".markup") )
+ {
+ markup[0] = '\0';
+ if ( !strstr( key, ".stream." ) )
+#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(43<<8)+0)
+ av_metadata_set2( &oc->metadata, key, mlt_properties_get_value( properties, i ), 0 );
+#else
+ av_metadata_set( &oc->metadata, key, mlt_properties_get_value( properties, i ) );
+#endif
+ }
+ free( key );
+ }
+ }
+#else
char *tmp = NULL;
int metavalue;
metavalue = mlt_properties_get_int( properties, "meta.attr.track.markup");
if (metavalue != 0) oc->track = metavalue;
+#endif
oc->oformat = fmt;
snprintf( oc->filename, sizeof(oc->filename), "%s", filename );
- // Add audio and video streams
- if ( fmt->video_codec != CODEC_ID_NONE )
- video_st = add_video_stream( this, oc, video_codec_id );
- if ( fmt->audio_codec != CODEC_ID_NONE )
- audio_st = add_audio_stream( this, oc, audio_codec_id );
+ // Add audio and video streams
+ if ( video_codec_id != CODEC_ID_NONE )
+ video_st = add_video_stream( consumer, oc, video_codec_id );
+ if ( audio_codec_id != CODEC_ID_NONE )
+ {
+ int is_multi = 0;
+
+ total_channels = 0;
+ // multitrack audio
+ for ( i = 0; i < MAX_AUDIO_STREAMS; i++ )
+ {
+ sprintf( key, "channels.%d", i );
+ int j = mlt_properties_get_int( properties, key );
+ if ( j )
+ {
+ is_multi = 1;
+ total_channels += j;
+ audio_st[i] = add_audio_stream( consumer, oc, audio_codec_id, j );
+ }
+ }
+ // single track
+ if ( !is_multi )
+ {
+ audio_st[0] = add_audio_stream( consumer, oc, audio_codec_id, channels );
+ total_channels = channels;
+ }
+ }
// Set the parameters (even though we have none...)
if ( av_set_parameters(oc, NULL) >= 0 )
{
- if ( video_st && !open_video( oc, video_st ) )
+ oc->preload = ( int )( mlt_properties_get_double( properties, "muxpreload" ) * AV_TIME_BASE );
+ oc->max_delay= ( int )( mlt_properties_get_double( properties, "muxdelay" ) * AV_TIME_BASE );
+
+ // Process properties as AVOptions
+ char *fpre = mlt_properties_get( properties, "fpre" );
+ if ( fpre )
+ {
+ mlt_properties p = mlt_properties_load( fpre );
+ apply_properties( oc, p, AV_OPT_FLAG_ENCODING_PARAM, 1 );
+ mlt_properties_close( p );
+ }
+ apply_properties( oc, properties, AV_OPT_FLAG_ENCODING_PARAM, 0 );
+
+ if ( video_st && !open_video( oc, video_st, vcodec? vcodec : NULL ) )
video_st = NULL;
- if ( audio_st )
- audio_input_frame_size = open_audio( oc, audio_st, audio_outbuf_size );
+ for ( i = 0; i < MAX_AUDIO_STREAMS && audio_st[i]; i++ )
+ {
+ audio_input_frame_size = open_audio( oc, audio_st[i], audio_outbuf_size,
+ acodec? acodec : NULL );
+ if ( !audio_input_frame_size )
+ audio_st[i] = NULL;
+ }
// Open the output file, if needed
if ( !( fmt->flags & AVFMT_NOFILE ) )
{
- if (url_fopen(&oc->pb, filename, URL_WRONLY) < 0)
+ if ( url_fopen( &oc->pb, filename, URL_WRONLY ) < 0 )
{
- fprintf(stderr, "Could not open '%s'\n", filename);
+ mlt_log_error( MLT_CONSUMER_SERVICE( consumer ), "Could not open '%s'\n", filename );
mlt_properties_set_int( properties, "running", 0 );
}
}
- // Write the stream header, if any
+ // Write the stream header.
if ( mlt_properties_get_int( properties, "running" ) )
av_write_header( oc );
}
else
{
- fprintf(stderr, "Invalid output format parameters\n");
+ mlt_log_error( MLT_CONSUMER_SERVICE( consumer ), "Invalid output format parameters\n" );
mlt_properties_set_int( properties, "running", 0 );
}
output = alloc_picture( video_st->codec->pix_fmt, width, height );
// Last check - need at least one stream
- if ( audio_st == NULL && video_st == NULL )
+ if ( !audio_st[0] && !video_st )
mlt_properties_set_int( properties, "running", 0 );
// Get the starting time (can ignore the times above)
gettimeofday( &ante, NULL );
// Loop while running
- while( mlt_properties_get_int( properties, "running" ) && !terminated )
+ while( mlt_properties_get_int( properties, "running" ) &&
+ ( !terminated || ( video_st && mlt_deque_count( queue ) ) ) )
{
- // Get the frame
- frame = mlt_consumer_rt_frame( this );
+ frame = mlt_consumer_rt_frame( consumer );
// Check that we have a frame to work with
if ( frame != NULL )
{
- // Increment frames despatched
+ // Increment frames dispatched
frames ++;
// Default audio args
terminated = terminate_on_pause && mlt_properties_get_double( frame_properties, "_speed" ) == 0.0;
// Get audio and append to the fifo
- if ( !terminated && audio_st )
+ if ( !terminated && audio_st[0] )
{
samples = mlt_sample_calculator( fps, frequency, count ++ );
- mlt_frame_get_audio( frame, &pcm, &aud_fmt, &frequency, &channels, &samples );
+ mlt_frame_get_audio( frame, (void**) &pcm, &aud_fmt, &frequency, &channels, &samples );
+
+ // Save the audio channel remap properties for later
+ mlt_properties_pass( frame_meta_properties, frame_properties, "meta.map.audio." );
// Create the fifo if we don't have one
if ( fifo == NULL )
mlt_properties_set_data( properties, "sample_fifo", fifo, 0, ( mlt_destructor )sample_fifo_close, NULL );
}
+ // Silence if not normal forward speed
if ( mlt_properties_get_double( frame_properties, "_speed" ) != 1.0 )
memset( pcm, 0, samples * channels * 2 );
// Append the samples
sample_fifo_append( fifo, pcm, samples * channels );
total_time += ( samples * 1000000 ) / frequency;
+
+ if ( !video_st )
+ mlt_events_fire( properties, "consumer-frame-show", frame, NULL );
}
// Encode the image
// While we have stuff to process, process...
while ( 1 )
{
- if (audio_st)
- audio_pts = (double)audio_st->pts.val * audio_st->time_base.num / audio_st->time_base.den;
- else
- audio_pts = 0.0;
-
- if (video_st)
- video_pts = (double)video_st->pts.val * video_st->time_base.num / video_st->time_base.den;
- else
- video_pts = 0.0;
-
- // Write interleaved audio and video frames
- if ( !video_st || ( video_st && audio_st && audio_pts < video_pts ) )
+ // Write interleaved audio and video frames
+ if ( !video_st || ( video_st && audio_st[0] && audio_pts < video_pts ) )
{
- if ( channels * audio_input_frame_size < sample_fifo_used( fifo ) )
+ // Write audio
+ if ( ( video_st && terminated ) || ( channels * audio_input_frame_size ) < sample_fifo_used( fifo ) )
{
- AVCodecContext *c;
- AVPacket pkt;
- av_init_packet( &pkt );
+ int j = 0; // channel offset into interleaved source buffer
+ int n = FFMIN( FFMIN( channels * audio_input_frame_size, sample_fifo_used( fifo ) ), AUDIO_ENCODE_BUFFER_SIZE );
+
+ // Get the audio samples
+ if ( n > 0 )
+ {
+ sample_fifo_fetch( fifo, audio_buf_1, n );
+ }
+ else if ( audio_codec_id == CODEC_ID_VORBIS && terminated )
+ {
+ // This prevents an infinite loop when some versions of vorbis do not
+ // increment pts when encoding silence.
+ audio_pts = video_pts;
+ break;
+ }
+ else
+ {
+ memset( audio_buf_1, 0, AUDIO_ENCODE_BUFFER_SIZE );
+ }
+ samples = n / channels;
+
+ // For each output stream
+ for ( i = 0; i < MAX_AUDIO_STREAMS && audio_st[i] && j < total_channels; i++ )
+ {
+ AVStream *stream = audio_st[i];
+ AVCodecContext *codec = stream->codec;
+ AVPacket pkt;
+
+ av_init_packet( &pkt );
+
+ // Optimized for single track and no channel remap
+ if ( !audio_st[1] && !mlt_properties_count( frame_meta_properties ) )
+ {
+ pkt.size = avcodec_encode_audio( codec, audio_outbuf, audio_outbuf_size, audio_buf_1 );
+ }
+ else
+ {
+ // Extract the audio channels according to channel mapping
+ int dest_offset = 0; // channel offset into interleaved dest buffer
+
+ // Get the number of channels for this stream
+ sprintf( key, "channels.%d", i );
+ int current_channels = mlt_properties_get_int( properties, key );
- c = audio_st->codec;
+ // Clear the destination audio buffer.
+ if ( !audio_buf_2 )
+ audio_buf_2 = av_mallocz( AUDIO_ENCODE_BUFFER_SIZE );
+ else
+ memset( audio_buf_2, 0, AUDIO_ENCODE_BUFFER_SIZE );
+
+ // For each output channel
+ while ( dest_offset < current_channels && j < total_channels )
+ {
+ int map_start = -1, map_channels = 0;
+ int source_offset = 0;
+ int k;
+
+ // Look for a mapping that starts at j
+ for ( k = 0; k < (MAX_AUDIO_STREAMS * 2) && map_start != j; k++ )
+ {
+ sprintf( key, "%d.channels", k );
+ map_channels = mlt_properties_get_int( frame_meta_properties, key );
+ sprintf( key, "%d.start", k );
+ if ( mlt_properties_get( frame_meta_properties, key ) )
+ map_start = mlt_properties_get_int( frame_meta_properties, key );
+ if ( map_start != j )
+ source_offset += map_channels;
+ }
- sample_fifo_fetch( fifo, buffer, channels * audio_input_frame_size );
+ // If no mapping
+ if ( map_start != j )
+ {
+ map_channels = current_channels;
+ source_offset = j;
+ }
- pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, buffer );
- // Write the compressed frame in the media file
- if ( c->coded_frame && c->coded_frame->pts != AV_NOPTS_VALUE )
- pkt.pts = av_rescale_q( c->coded_frame->pts, c->time_base, audio_st->time_base );
- pkt.flags |= PKT_FLAG_KEY;
- pkt.stream_index= audio_st->index;
- pkt.data= audio_outbuf;
+ // Copy samples if source offset valid
+ if ( source_offset < channels )
+ {
+ // Interleave the audio buffer with the # channels for this stream/mapping.
+ for ( k = 0; k < map_channels; k++, j++, source_offset++, dest_offset++ )
+ {
+ int16_t *src = audio_buf_1 + source_offset;
+ int16_t *dest = audio_buf_2 + dest_offset;
+ int s = samples + 1;
+
+ while ( --s ) {
+ *dest = *src;
+ dest += current_channels;
+ src += channels;
+ }
+ }
+ }
+ // Otherwise silence
+ else
+ {
+ j += current_channels;
+ dest_offset += current_channels;
+ }
+ }
+ pkt.size = avcodec_encode_audio( codec, audio_outbuf, audio_outbuf_size, audio_buf_2 );
+ }
- if ( pkt.size )
- if ( av_interleaved_write_frame( oc, &pkt ) != 0)
- fprintf(stderr, "Error while writing audio frame\n");
+ // Write the compressed frame in the media file
+ if ( codec->coded_frame && codec->coded_frame->pts != AV_NOPTS_VALUE )
+ {
+ pkt.pts = av_rescale_q( codec->coded_frame->pts, codec->time_base, stream->time_base );
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "audio stream %d pkt pts %lld frame pts %lld",
+ stream->index, pkt.pts, codec->coded_frame->pts );
+ }
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index = stream->index;
+ pkt.data = audio_outbuf;
- audio_pts += c->frame_size;
+ if ( pkt.size > 0 )
+ {
+ if ( av_interleaved_write_frame( oc, &pkt ) )
+ {
+ mlt_log_fatal( MLT_CONSUMER_SERVICE( consumer ), "error writing audio frame\n" );
+ mlt_events_fire( properties, "consumer-fatal-error", NULL );
+ goto on_fatal_error;
+ }
+ }
+
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), " frame_size %d\n", codec->frame_size );
+ if ( i == 0 )
+ {
+ audio_pts = (double)stream->pts.val * av_q2d( stream->time_base );
+ }
+ }
}
else
{
break;
}
}
- else if ( video_st )
+ else if ( video_st )
{
+ // Write video
if ( mlt_deque_count( queue ) )
{
- int out_size, ret;
- AVCodecContext *c;
+ int out_size, ret = 0;
+ AVCodecContext *c;
frame = mlt_deque_pop_front( queue );
frame_properties = MLT_FRAME_PROPERTIES( frame );
c = video_st->codec;
-
+
if ( mlt_properties_get_int( frame_properties, "rendered" ) )
{
int i = 0;
- int j = 0;
uint8_t *p;
uint8_t *q;
- mlt_events_fire( properties, "consumer-frame-show", frame, NULL );
-
mlt_frame_get_image( frame, &image, &img_fmt, &img_width, &img_height, 0 );
q = image;
for ( i = 0; i < height; i ++ )
{
p = input->data[ 0 ] + i * input->linesize[ 0 ];
- j = width;
- while( j -- )
- {
- *p ++ = *q ++;
- *p ++ = *q ++;
- }
+ memcpy( p, q, width * 2 );
+ q += width * 2;
}
// Do the colour space conversion
#ifdef SWSCALE
- struct SwsContext *context = sws_getContext( width, height, PIX_FMT_YUV422,
- width, height, video_st->codec->pix_fmt, SWS_FAST_BILINEAR, NULL, NULL, NULL);
+ int flags = SWS_BILINEAR;
+#ifdef USE_MMX
+ flags |= SWS_CPU_CAPS_MMX;
+#endif
+#ifdef USE_SSE
+ flags |= SWS_CPU_CAPS_MMX2;
+#endif
+ struct SwsContext *context = sws_getContext( width, height, PIX_FMT_YUYV422,
+ width, height, video_st->codec->pix_fmt, flags, NULL, NULL, NULL);
sws_scale( context, input->data, input->linesize, 0, height,
output->data, output->linesize);
sws_freeContext( context );
#else
- img_convert( ( AVPicture * )output, video_st->codec->pix_fmt, ( AVPicture * )input, PIX_FMT_YUV422, width, height );
+ img_convert( ( AVPicture * )output, video_st->codec->pix_fmt, ( AVPicture * )input, PIX_FMT_YUYV422, width, height );
#endif
+ mlt_events_fire( properties, "consumer-frame-show", frame, NULL );
+
// Apply the alpha if applicable
- if ( video_st->codec->pix_fmt == PIX_FMT_RGBA32 )
+ if ( video_st->codec->pix_fmt == PIX_FMT_RGB32 )
{
uint8_t *alpha = mlt_frame_get_alpha_mask( frame );
register int n;
for ( i = 0; i < height; i ++ )
{
n = ( width + 7 ) / 8;
- p = output->data[ 0 ] + i * output->linesize[ 0 ];
-
- #ifndef __DARWIN__
- p += 3;
- #endif
+ p = output->data[ 0 ] + i * output->linesize[ 0 ] + 3;
switch( width % 8 )
{
}
}
}
-
- if (oc->oformat->flags & AVFMT_RAWPICTURE)
+
+ if (oc->oformat->flags & AVFMT_RAWPICTURE)
{
// raw video case. The API will change slightly in the near future for that
AVPacket pkt;
av_init_packet(&pkt);
-
+
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index= video_st->index;
pkt.data= (uint8_t *)output;
ret = av_write_frame(oc, &pkt);
video_pts += c->frame_size;
- }
+ }
else
{
// Set the quality
// Set frame interlace hints
output->interlaced_frame = !mlt_properties_get_int( frame_properties, "progressive" );
output->top_field_first = mlt_properties_get_int( frame_properties, "top_field_first" );
+ output->pts = frame_count;
// Encode the image
out_size = avcodec_encode_video(c, video_outbuf, video_outbuf_size, output );
// If zero size, it means the image was buffered
- if (out_size > 0)
+ if ( out_size > 0 )
{
AVPacket pkt;
av_init_packet( &pkt );
if ( c->coded_frame && c->coded_frame->pts != AV_NOPTS_VALUE )
pkt.pts= av_rescale_q( c->coded_frame->pts, c->time_base, video_st->time_base );
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "video pkt pts %lld frame pts %lld", pkt.pts, c->coded_frame->pts );
if( c->coded_frame && c->coded_frame->key_frame )
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index= video_st->index;
pkt.data= video_outbuf;
pkt.size= out_size;
- // write the compressed frame in the media file
+ // write the compressed frame in the media file
ret = av_interleaved_write_frame(oc, &pkt);
- video_pts += c->frame_size;
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), " frame_size %d\n", c->frame_size );
+ video_pts = (double)video_st->pts.val * av_q2d( video_st->time_base );
+
+ // Dual pass logging
+ if ( mlt_properties_get_data( properties, "_logfile", NULL ) && c->stats_out )
+ fprintf( mlt_properties_get_data( properties, "_logfile", NULL ), "%s", c->stats_out );
}
- else
+ else if ( out_size < 0 )
{
- fprintf( stderr, "Error with video encode\n" );
+ mlt_log_warning( MLT_CONSUMER_SERVICE( consumer ), "error with video encode %d\n", frame_count );
}
}
frame_count++;
+ if ( ret )
+ {
+ mlt_log_fatal( MLT_CONSUMER_SERVICE( consumer ), "error writing video frame\n" );
+ mlt_events_fire( properties, "consumer-fatal-error", NULL );
+ goto on_fatal_error;
+ }
mlt_frame_close( frame );
}
else
break;
}
}
+ if ( audio_st[0] )
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "audio pts %lld (%f) ", audio_st[0]->pts.val, audio_pts );
+ if ( video_st )
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "video pts %lld (%f) ", video_st->pts.val, video_pts );
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "\n" );
}
- if ( real_time_output && frames % 12 == 0 )
+ if ( real_time_output == 1 && frames % 2 == 0 )
{
long passed = time_difference( &ante );
if ( fifo != NULL )
}
}
-#ifdef FLUSH
- if ( ! real_time_output )
+ // Flush the encoder buffers
+ if ( real_time_output <= 0 )
{
// Flush audio fifo
- if ( audio_st && audio_st->codec->frame_size > 1 ) for (;;)
+ // TODO: flush all audio streams
+ if ( audio_st[0] && audio_st[0]->codec->frame_size > 1 ) for (;;)
{
- AVCodecContext *c = audio_st->codec;
+ AVCodecContext *c = audio_st[0]->codec;
AVPacket pkt;
av_init_packet( &pkt );
pkt.size = 0;
if ( /*( c->capabilities & CODEC_CAP_SMALL_LAST_FRAME ) &&*/
( channels * audio_input_frame_size < sample_fifo_used( fifo ) ) )
{
- sample_fifo_fetch( fifo, buffer, channels * audio_input_frame_size );
- pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, buffer );
+ sample_fifo_fetch( fifo, audio_buf_1, channels * audio_input_frame_size );
+ pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, audio_buf_1 );
}
if ( pkt.size <= 0 )
pkt.size = avcodec_encode_audio( c, audio_outbuf, audio_outbuf_size, NULL );
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "flushing audio size %d\n", pkt.size );
if ( pkt.size <= 0 )
break;
// Write the compressed frame in the media file
if ( c->coded_frame && c->coded_frame->pts != AV_NOPTS_VALUE )
- pkt.pts = av_rescale_q( c->coded_frame->pts, c->time_base, audio_st->time_base );
+ pkt.pts = av_rescale_q( c->coded_frame->pts, c->time_base, audio_st[0]->time_base );
pkt.flags |= PKT_FLAG_KEY;
- pkt.stream_index = audio_st->index;
+ pkt.stream_index = audio_st[0]->index;
pkt.data = audio_outbuf;
if ( av_interleaved_write_frame( oc, &pkt ) != 0 )
{
- fprintf(stderr, "Error while writing flushed audio frame\n");
- break;
+ mlt_log_fatal( MLT_CONSUMER_SERVICE( consumer ), "error writing flushed audio frame\n" );
+ mlt_events_fire( properties, "consumer-fatal-error", NULL );
+ goto on_fatal_error;
}
}
// Encode the image
pkt.size = avcodec_encode_video( c, video_outbuf, video_outbuf_size, NULL );
+ mlt_log_debug( MLT_CONSUMER_SERVICE( consumer ), "flushing video size %d\n", pkt.size );
if ( pkt.size <= 0 )
break;
// write the compressed frame in the media file
if ( av_interleaved_write_frame( oc, &pkt ) != 0 )
{
- fprintf(stderr, "Error while writing flushed video frame\n");
- break;
+ mlt_log_fatal( MLT_CONSUMER_SERVICE(consumer), "error writing flushed video frame\n" );
+ mlt_events_fire( properties, "consumer-fatal-error", NULL );
+ goto on_fatal_error;
}
+ // Dual pass logging
+ if ( mlt_properties_get_data( properties, "_logfile", NULL ) && c->stats_out )
+ fprintf( mlt_properties_get_data( properties, "_logfile", NULL ), "%s", c->stats_out );
}
}
-#endif
-
- // close each codec
- if (video_st)
- close_video(oc, video_st);
- if (audio_st)
- close_audio(oc, audio_st);
+on_fatal_error:
+
// Write the trailer, if any
- av_write_trailer(oc);
+ av_write_trailer( oc );
+
+ // close each codec
+ if ( video_st )
+ close_video(oc, video_st);
+ for ( i = 0; i < MAX_AUDIO_STREAMS && audio_st[i]; i++ )
+ close_audio( oc, audio_st[i] );
// Free the streams
- for(i = 0; i < oc->nb_streams; i++)
- av_freep(&oc->streams[i]);
+ for ( i = 0; i < oc->nb_streams; i++ )
+ av_freep( &oc->streams[i] );
// Close the output file
- if (!(fmt->flags & AVFMT_NOFILE))
+ if ( !( fmt->flags & AVFMT_NOFILE ) )
#if LIBAVFORMAT_VERSION_INT >= ((52<<16)+(0<<8)+0)
- url_fclose(oc->pb);
+ url_fclose( oc->pb );
#else
- url_fclose(&oc->pb);
+ url_fclose( &oc->pb );
#endif
// Clean up input and output frames
av_free( input->data[0] );
av_free( input );
av_free( video_outbuf );
- av_free( buffer );
+ av_free( audio_buf_1 );
+ av_free( audio_buf_2 );
// Free the stream
- av_free(oc);
+ av_free( oc );
// Just in case we terminated on pause
mlt_properties_set_int( properties, "running", 0 );
- mlt_consumer_stopped( this );
+ mlt_consumer_stopped( consumer );
+ mlt_properties_close( frame_meta_properties );
+
+ if ( mlt_properties_get_int( properties, "pass" ) > 1 )
+ {
+ // Remove the dual pass log file
+ if ( mlt_properties_get( properties, "_logfilename" ) )
+ remove( mlt_properties_get( properties, "_logfilename" ) );
+
+ // Remove the x264 dual pass logs
+ char *cwd = getcwd( NULL, 0 );
+ const char *file = "x264_2pass.log";
+ char *full = malloc( strlen( cwd ) + strlen( file ) + 2 );
+ sprintf( full, "%s/%s", cwd, file );
+ remove( full );
+ free( full );
+ file = "x264_2pass.log.temp";
+ full = malloc( strlen( cwd ) + strlen( file ) + 2 );
+ sprintf( full, "%s/%s", cwd, file );
+ remove( full );
+ free( full );
+ file = "x264_2pass.log.mbtree";
+ full = malloc( strlen( cwd ) + strlen( file ) + 2 );
+ sprintf( full, "%s/%s", cwd, file );
+ remove( full );
+ free( full );
+ free( cwd );
+ remove( "x264_2pass.log.temp" );
+ }
return NULL;
}
/** Close the consumer.
*/
-static void consumer_close( mlt_consumer this )
+static void consumer_close( mlt_consumer consumer )
{
// Stop the consumer
- mlt_consumer_stop( this );
+ mlt_consumer_stop( consumer );
// Close the parent
- mlt_consumer_close( this );
+ mlt_consumer_close( consumer );
// Free the memory
- free( this );
+ free( consumer );
}