]> git.sesse.net Git - ffmpeg/blob - avconv.c
2bc9d82378e428debe37db0581b95e9d3edee51d
[ffmpeg] / avconv.c
1 /*
2  * avconv main
3  * Copyright (c) 2000-2011 The libav developers.
4  *
5  * This file is part of Libav.
6  *
7  * Libav is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * Libav is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with Libav; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21
22 #include "config.h"
23 #include <ctype.h>
24 #include <string.h>
25 #include <math.h>
26 #include <stdlib.h>
27 #include <errno.h>
28 #include <signal.h>
29 #include <limits.h>
30 #include "libavformat/avformat.h"
31 #include "libavdevice/avdevice.h"
32 #include "libswscale/swscale.h"
33 #include "libavresample/avresample.h"
34 #include "libavutil/opt.h"
35 #include "libavutil/audioconvert.h"
36 #include "libavutil/parseutils.h"
37 #include "libavutil/samplefmt.h"
38 #include "libavutil/colorspace.h"
39 #include "libavutil/fifo.h"
40 #include "libavutil/intreadwrite.h"
41 #include "libavutil/dict.h"
42 #include "libavutil/mathematics.h"
43 #include "libavutil/pixdesc.h"
44 #include "libavutil/avstring.h"
45 #include "libavutil/libm.h"
46 #include "libavutil/imgutils.h"
47 #include "libavutil/time.h"
48 #include "libavformat/os_support.h"
49
50 # include "libavfilter/avfilter.h"
51 # include "libavfilter/avfiltergraph.h"
52 # include "libavfilter/buffersrc.h"
53 # include "libavfilter/buffersink.h"
54
55 #if HAVE_SYS_RESOURCE_H
56 #include <sys/types.h>
57 #include <sys/resource.h>
58 #elif HAVE_GETPROCESSTIMES
59 #include <windows.h>
60 #endif
61 #if HAVE_GETPROCESSMEMORYINFO
62 #include <windows.h>
63 #include <psapi.h>
64 #endif
65
66 #if HAVE_SYS_SELECT_H
67 #include <sys/select.h>
68 #endif
69
70 #if HAVE_PTHREADS
71 #include <pthread.h>
72 #endif
73
74 #include <time.h>
75
76 #include "avconv.h"
77 #include "cmdutils.h"
78
79 #include "libavutil/avassert.h"
80
81 const char program_name[] = "avconv";
82 const int program_birth_year = 2000;
83
84 static FILE *vstats_file;
85
86 static int64_t video_size = 0;
87 static int64_t audio_size = 0;
88 static int64_t extra_size = 0;
89 static int nb_frames_dup = 0;
90 static int nb_frames_drop = 0;
91
92
93
94 #if HAVE_PTHREADS
95 /* signal to input threads that they should exit; set by the main thread */
96 static int transcoding_finished;
97 #endif
98
99 #define DEFAULT_PASS_LOGFILENAME_PREFIX "av2pass"
100
101 InputStream **input_streams = NULL;
102 int        nb_input_streams = 0;
103 InputFile   **input_files   = NULL;
104 int        nb_input_files   = 0;
105
106 OutputStream **output_streams = NULL;
107 int         nb_output_streams = 0;
108 OutputFile   **output_files   = NULL;
109 int         nb_output_files   = 0;
110
111 FilterGraph **filtergraphs;
112 int        nb_filtergraphs;
113
114 static void term_exit(void)
115 {
116     av_log(NULL, AV_LOG_QUIET, "");
117 }
118
119 static volatile int received_sigterm = 0;
120 static volatile int received_nb_signals = 0;
121
122 static void
123 sigterm_handler(int sig)
124 {
125     received_sigterm = sig;
126     received_nb_signals++;
127     term_exit();
128 }
129
130 static void term_init(void)
131 {
132     signal(SIGINT , sigterm_handler); /* Interrupt (ANSI).    */
133     signal(SIGTERM, sigterm_handler); /* Termination (ANSI).  */
134 #ifdef SIGXCPU
135     signal(SIGXCPU, sigterm_handler);
136 #endif
137 }
138
139 static int decode_interrupt_cb(void *ctx)
140 {
141     return received_nb_signals > 1;
142 }
143
144 const AVIOInterruptCB int_cb = { decode_interrupt_cb, NULL };
145
146 void exit_program(int ret)
147 {
148     int i, j;
149
150     for (i = 0; i < nb_filtergraphs; i++) {
151         avfilter_graph_free(&filtergraphs[i]->graph);
152         for (j = 0; j < filtergraphs[i]->nb_inputs; j++) {
153             av_freep(&filtergraphs[i]->inputs[j]->name);
154             av_freep(&filtergraphs[i]->inputs[j]);
155         }
156         av_freep(&filtergraphs[i]->inputs);
157         for (j = 0; j < filtergraphs[i]->nb_outputs; j++) {
158             av_freep(&filtergraphs[i]->outputs[j]->name);
159             av_freep(&filtergraphs[i]->outputs[j]);
160         }
161         av_freep(&filtergraphs[i]->outputs);
162         av_freep(&filtergraphs[i]);
163     }
164     av_freep(&filtergraphs);
165
166     /* close files */
167     for (i = 0; i < nb_output_files; i++) {
168         AVFormatContext *s = output_files[i]->ctx;
169         if (!(s->oformat->flags & AVFMT_NOFILE) && s->pb)
170             avio_close(s->pb);
171         avformat_free_context(s);
172         av_dict_free(&output_files[i]->opts);
173         av_freep(&output_files[i]);
174     }
175     for (i = 0; i < nb_output_streams; i++) {
176         AVBitStreamFilterContext *bsfc = output_streams[i]->bitstream_filters;
177         while (bsfc) {
178             AVBitStreamFilterContext *next = bsfc->next;
179             av_bitstream_filter_close(bsfc);
180             bsfc = next;
181         }
182         output_streams[i]->bitstream_filters = NULL;
183
184         av_freep(&output_streams[i]->forced_keyframes);
185         av_freep(&output_streams[i]->avfilter);
186         av_freep(&output_streams[i]->filtered_frame);
187         av_freep(&output_streams[i]);
188     }
189     for (i = 0; i < nb_input_files; i++) {
190         avformat_close_input(&input_files[i]->ctx);
191         av_freep(&input_files[i]);
192     }
193     for (i = 0; i < nb_input_streams; i++) {
194         av_freep(&input_streams[i]->decoded_frame);
195         av_dict_free(&input_streams[i]->opts);
196         free_buffer_pool(&input_streams[i]->buffer_pool);
197         av_freep(&input_streams[i]->filters);
198         av_freep(&input_streams[i]);
199     }
200
201     if (vstats_file)
202         fclose(vstats_file);
203     av_free(vstats_filename);
204
205     av_freep(&input_streams);
206     av_freep(&input_files);
207     av_freep(&output_streams);
208     av_freep(&output_files);
209
210     uninit_opts();
211
212     avfilter_uninit();
213     avformat_network_deinit();
214
215     if (received_sigterm) {
216         av_log(NULL, AV_LOG_INFO, "Received signal %d: terminating.\n",
217                (int) received_sigterm);
218         exit (255);
219     }
220
221     exit(ret);
222 }
223
224 void assert_avoptions(AVDictionary *m)
225 {
226     AVDictionaryEntry *t;
227     if ((t = av_dict_get(m, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
228         av_log(NULL, AV_LOG_FATAL, "Option %s not found.\n", t->key);
229         exit_program(1);
230     }
231 }
232
233 static void assert_codec_experimental(AVCodecContext *c, int encoder)
234 {
235     const char *codec_string = encoder ? "encoder" : "decoder";
236     AVCodec *codec;
237     if (c->codec->capabilities & CODEC_CAP_EXPERIMENTAL &&
238         c->strict_std_compliance > FF_COMPLIANCE_EXPERIMENTAL) {
239         av_log(NULL, AV_LOG_FATAL, "%s '%s' is experimental and might produce bad "
240                 "results.\nAdd '-strict experimental' if you want to use it.\n",
241                 codec_string, c->codec->name);
242         codec = encoder ? avcodec_find_encoder(c->codec->id) : avcodec_find_decoder(c->codec->id);
243         if (!(codec->capabilities & CODEC_CAP_EXPERIMENTAL))
244             av_log(NULL, AV_LOG_FATAL, "Or use the non experimental %s '%s'.\n",
245                    codec_string, codec->name);
246         exit_program(1);
247     }
248 }
249
250 /**
251  * Update the requested input sample format based on the output sample format.
252  * This is currently only used to request float output from decoders which
253  * support multiple sample formats, one of which is AV_SAMPLE_FMT_FLT.
254  * Ideally this will be removed in the future when decoders do not do format
255  * conversion and only output in their native format.
256  */
257 static void update_sample_fmt(AVCodecContext *dec, AVCodec *dec_codec,
258                               AVCodecContext *enc)
259 {
260     /* if sample formats match or a decoder sample format has already been
261        requested, just return */
262     if (enc->sample_fmt == dec->sample_fmt ||
263         dec->request_sample_fmt > AV_SAMPLE_FMT_NONE)
264         return;
265
266     /* if decoder supports more than one output format */
267     if (dec_codec && dec_codec->sample_fmts &&
268         dec_codec->sample_fmts[0] != AV_SAMPLE_FMT_NONE &&
269         dec_codec->sample_fmts[1] != AV_SAMPLE_FMT_NONE) {
270         const enum AVSampleFormat *p;
271         int min_dec = -1, min_inc = -1;
272
273         /* find a matching sample format in the encoder */
274         for (p = dec_codec->sample_fmts; *p != AV_SAMPLE_FMT_NONE; p++) {
275             if (*p == enc->sample_fmt) {
276                 dec->request_sample_fmt = *p;
277                 return;
278             } else if (*p > enc->sample_fmt) {
279                 min_inc = FFMIN(min_inc, *p - enc->sample_fmt);
280             } else
281                 min_dec = FFMIN(min_dec, enc->sample_fmt - *p);
282         }
283
284         /* if none match, provide the one that matches quality closest */
285         dec->request_sample_fmt = min_inc > 0 ? enc->sample_fmt + min_inc :
286                                   enc->sample_fmt - min_dec;
287     }
288 }
289
290 static void write_frame(AVFormatContext *s, AVPacket *pkt, OutputStream *ost)
291 {
292     AVBitStreamFilterContext *bsfc = ost->bitstream_filters;
293     AVCodecContext          *avctx = ost->st->codec;
294     int ret;
295
296     /*
297      * Audio encoders may split the packets --  #frames in != #packets out.
298      * But there is no reordering, so we can limit the number of output packets
299      * by simply dropping them here.
300      * Counting encoded video frames needs to be done separately because of
301      * reordering, see do_video_out()
302      */
303     if (!(avctx->codec_type == AVMEDIA_TYPE_VIDEO && avctx->codec)) {
304         if (ost->frame_number >= ost->max_frames) {
305             av_free_packet(pkt);
306             return;
307         }
308         ost->frame_number++;
309     }
310
311     while (bsfc) {
312         AVPacket new_pkt = *pkt;
313         int a = av_bitstream_filter_filter(bsfc, avctx, NULL,
314                                            &new_pkt.data, &new_pkt.size,
315                                            pkt->data, pkt->size,
316                                            pkt->flags & AV_PKT_FLAG_KEY);
317         if (a > 0) {
318             av_free_packet(pkt);
319             new_pkt.destruct = av_destruct_packet;
320         } else if (a < 0) {
321             av_log(NULL, AV_LOG_ERROR, "%s failed for stream %d, codec %s",
322                    bsfc->filter->name, pkt->stream_index,
323                    avctx->codec ? avctx->codec->name : "copy");
324             print_error("", a);
325             if (exit_on_error)
326                 exit_program(1);
327         }
328         *pkt = new_pkt;
329
330         bsfc = bsfc->next;
331     }
332
333     pkt->stream_index = ost->index;
334     ret = av_interleaved_write_frame(s, pkt);
335     if (ret < 0) {
336         print_error("av_interleaved_write_frame()", ret);
337         exit_program(1);
338     }
339 }
340
341 static int check_recording_time(OutputStream *ost)
342 {
343     OutputFile *of = output_files[ost->file_index];
344
345     if (of->recording_time != INT64_MAX &&
346         av_compare_ts(ost->sync_opts - ost->first_pts, ost->st->codec->time_base, of->recording_time,
347                       AV_TIME_BASE_Q) >= 0) {
348         ost->finished = 1;
349         return 0;
350     }
351     return 1;
352 }
353
354 static void do_audio_out(AVFormatContext *s, OutputStream *ost,
355                          AVFrame *frame)
356 {
357     AVCodecContext *enc = ost->st->codec;
358     AVPacket pkt;
359     int got_packet = 0;
360
361     av_init_packet(&pkt);
362     pkt.data = NULL;
363     pkt.size = 0;
364
365     if (!check_recording_time(ost))
366         return;
367
368     if (frame->pts == AV_NOPTS_VALUE || audio_sync_method < 0)
369         frame->pts = ost->sync_opts;
370     ost->sync_opts = frame->pts + frame->nb_samples;
371
372     if (avcodec_encode_audio2(enc, &pkt, frame, &got_packet) < 0) {
373         av_log(NULL, AV_LOG_FATAL, "Audio encoding failed\n");
374         exit_program(1);
375     }
376
377     if (got_packet) {
378         if (pkt.pts != AV_NOPTS_VALUE)
379             pkt.pts      = av_rescale_q(pkt.pts,      enc->time_base, ost->st->time_base);
380         if (pkt.dts != AV_NOPTS_VALUE)
381             pkt.dts      = av_rescale_q(pkt.dts,      enc->time_base, ost->st->time_base);
382         if (pkt.duration > 0)
383             pkt.duration = av_rescale_q(pkt.duration, enc->time_base, ost->st->time_base);
384
385         write_frame(s, &pkt, ost);
386
387         audio_size += pkt.size;
388     }
389 }
390
391 static void pre_process_video_frame(InputStream *ist, AVPicture *picture, void **bufp)
392 {
393     AVCodecContext *dec;
394     AVPicture *picture2;
395     AVPicture picture_tmp;
396     uint8_t *buf = 0;
397
398     dec = ist->st->codec;
399
400     /* deinterlace : must be done before any resize */
401     if (do_deinterlace) {
402         int size;
403
404         /* create temporary picture */
405         size = avpicture_get_size(dec->pix_fmt, dec->width, dec->height);
406         buf  = av_malloc(size);
407         if (!buf)
408             return;
409
410         picture2 = &picture_tmp;
411         avpicture_fill(picture2, buf, dec->pix_fmt, dec->width, dec->height);
412
413         if (avpicture_deinterlace(picture2, picture,
414                                  dec->pix_fmt, dec->width, dec->height) < 0) {
415             /* if error, do not deinterlace */
416             av_log(NULL, AV_LOG_WARNING, "Deinterlacing failed\n");
417             av_free(buf);
418             buf = NULL;
419             picture2 = picture;
420         }
421     } else {
422         picture2 = picture;
423     }
424
425     if (picture != picture2)
426         *picture = *picture2;
427     *bufp = buf;
428 }
429
430 static void do_subtitle_out(AVFormatContext *s,
431                             OutputStream *ost,
432                             InputStream *ist,
433                             AVSubtitle *sub,
434                             int64_t pts)
435 {
436     static uint8_t *subtitle_out = NULL;
437     int subtitle_out_max_size = 1024 * 1024;
438     int subtitle_out_size, nb, i;
439     AVCodecContext *enc;
440     AVPacket pkt;
441
442     if (pts == AV_NOPTS_VALUE) {
443         av_log(NULL, AV_LOG_ERROR, "Subtitle packets must have a pts\n");
444         if (exit_on_error)
445             exit_program(1);
446         return;
447     }
448
449     enc = ost->st->codec;
450
451     if (!subtitle_out) {
452         subtitle_out = av_malloc(subtitle_out_max_size);
453     }
454
455     /* Note: DVB subtitle need one packet to draw them and one other
456        packet to clear them */
457     /* XXX: signal it in the codec context ? */
458     if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE)
459         nb = 2;
460     else
461         nb = 1;
462
463     for (i = 0; i < nb; i++) {
464         ost->sync_opts = av_rescale_q(pts, ist->st->time_base, enc->time_base);
465         if (!check_recording_time(ost))
466             return;
467
468         sub->pts = av_rescale_q(pts, ist->st->time_base, AV_TIME_BASE_Q);
469         // start_display_time is required to be 0
470         sub->pts               += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, AV_TIME_BASE_Q);
471         sub->end_display_time  -= sub->start_display_time;
472         sub->start_display_time = 0;
473         subtitle_out_size = avcodec_encode_subtitle(enc, subtitle_out,
474                                                     subtitle_out_max_size, sub);
475         if (subtitle_out_size < 0) {
476             av_log(NULL, AV_LOG_FATAL, "Subtitle encoding failed\n");
477             exit_program(1);
478         }
479
480         av_init_packet(&pkt);
481         pkt.data = subtitle_out;
482         pkt.size = subtitle_out_size;
483         pkt.pts  = av_rescale_q(sub->pts, AV_TIME_BASE_Q, ost->st->time_base);
484         if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE) {
485             /* XXX: the pts correction is handled here. Maybe handling
486                it in the codec would be better */
487             if (i == 0)
488                 pkt.pts += 90 * sub->start_display_time;
489             else
490                 pkt.pts += 90 * sub->end_display_time;
491         }
492         write_frame(s, &pkt, ost);
493     }
494 }
495
496 static void do_video_out(AVFormatContext *s,
497                          OutputStream *ost,
498                          AVFrame *in_picture,
499                          int *frame_size, float quality)
500 {
501     int ret, format_video_sync;
502     AVPacket pkt;
503     AVCodecContext *enc = ost->st->codec;
504
505     *frame_size = 0;
506
507     format_video_sync = video_sync_method;
508     if (format_video_sync == VSYNC_AUTO)
509         format_video_sync = (s->oformat->flags & AVFMT_NOTIMESTAMPS) ? VSYNC_PASSTHROUGH :
510                             (s->oformat->flags & AVFMT_VARIABLE_FPS) ? VSYNC_VFR : VSYNC_CFR;
511     if (format_video_sync != VSYNC_PASSTHROUGH &&
512         ost->frame_number &&
513         in_picture->pts != AV_NOPTS_VALUE &&
514         in_picture->pts < ost->sync_opts) {
515         nb_frames_drop++;
516         av_log(NULL, AV_LOG_VERBOSE, "*** drop!\n");
517         return;
518     }
519
520     if (in_picture->pts == AV_NOPTS_VALUE)
521         in_picture->pts = ost->sync_opts;
522     ost->sync_opts = in_picture->pts;
523
524
525     if (!ost->frame_number)
526         ost->first_pts = in_picture->pts;
527
528     av_init_packet(&pkt);
529     pkt.data = NULL;
530     pkt.size = 0;
531
532     if (!check_recording_time(ost) ||
533         ost->frame_number >= ost->max_frames)
534         return;
535
536     if (s->oformat->flags & AVFMT_RAWPICTURE &&
537         enc->codec->id == AV_CODEC_ID_RAWVIDEO) {
538         /* raw pictures are written as AVPicture structure to
539            avoid any copies. We support temporarily the older
540            method. */
541         enc->coded_frame->interlaced_frame = in_picture->interlaced_frame;
542         enc->coded_frame->top_field_first  = in_picture->top_field_first;
543         pkt.data   = (uint8_t *)in_picture;
544         pkt.size   =  sizeof(AVPicture);
545         pkt.pts    = av_rescale_q(in_picture->pts, enc->time_base, ost->st->time_base);
546         pkt.flags |= AV_PKT_FLAG_KEY;
547
548         write_frame(s, &pkt, ost);
549     } else {
550         int got_packet;
551         AVFrame big_picture;
552
553         big_picture = *in_picture;
554         /* better than nothing: use input picture interlaced
555            settings */
556         big_picture.interlaced_frame = in_picture->interlaced_frame;
557         if (ost->st->codec->flags & (CODEC_FLAG_INTERLACED_DCT|CODEC_FLAG_INTERLACED_ME)) {
558             if (ost->top_field_first == -1)
559                 big_picture.top_field_first = in_picture->top_field_first;
560             else
561                 big_picture.top_field_first = !!ost->top_field_first;
562         }
563
564         /* handles same_quant here. This is not correct because it may
565            not be a global option */
566         big_picture.quality = quality;
567         if (!enc->me_threshold)
568             big_picture.pict_type = 0;
569         if (ost->forced_kf_index < ost->forced_kf_count &&
570             big_picture.pts >= ost->forced_kf_pts[ost->forced_kf_index]) {
571             big_picture.pict_type = AV_PICTURE_TYPE_I;
572             ost->forced_kf_index++;
573         }
574         ret = avcodec_encode_video2(enc, &pkt, &big_picture, &got_packet);
575         if (ret < 0) {
576             av_log(NULL, AV_LOG_FATAL, "Video encoding failed\n");
577             exit_program(1);
578         }
579
580         if (got_packet) {
581             if (pkt.pts != AV_NOPTS_VALUE)
582                 pkt.pts = av_rescale_q(pkt.pts, enc->time_base, ost->st->time_base);
583             if (pkt.dts != AV_NOPTS_VALUE)
584                 pkt.dts = av_rescale_q(pkt.dts, enc->time_base, ost->st->time_base);
585
586             write_frame(s, &pkt, ost);
587             *frame_size = pkt.size;
588             video_size += pkt.size;
589
590             /* if two pass, output log */
591             if (ost->logfile && enc->stats_out) {
592                 fprintf(ost->logfile, "%s", enc->stats_out);
593             }
594         }
595     }
596     ost->sync_opts++;
597     /*
598      * For video, number of frames in == number of packets out.
599      * But there may be reordering, so we can't throw away frames on encoder
600      * flush, we need to limit them here, before they go into encoder.
601      */
602     ost->frame_number++;
603 }
604
605 static double psnr(double d)
606 {
607     return -10.0 * log(d) / log(10.0);
608 }
609
610 static void do_video_stats(AVFormatContext *os, OutputStream *ost,
611                            int frame_size)
612 {
613     AVCodecContext *enc;
614     int frame_number;
615     double ti1, bitrate, avg_bitrate;
616
617     /* this is executed just the first time do_video_stats is called */
618     if (!vstats_file) {
619         vstats_file = fopen(vstats_filename, "w");
620         if (!vstats_file) {
621             perror("fopen");
622             exit_program(1);
623         }
624     }
625
626     enc = ost->st->codec;
627     if (enc->codec_type == AVMEDIA_TYPE_VIDEO) {
628         frame_number = ost->frame_number;
629         fprintf(vstats_file, "frame= %5d q= %2.1f ", frame_number, enc->coded_frame->quality / (float)FF_QP2LAMBDA);
630         if (enc->flags&CODEC_FLAG_PSNR)
631             fprintf(vstats_file, "PSNR= %6.2f ", psnr(enc->coded_frame->error[0] / (enc->width * enc->height * 255.0 * 255.0)));
632
633         fprintf(vstats_file,"f_size= %6d ", frame_size);
634         /* compute pts value */
635         ti1 = ost->sync_opts * av_q2d(enc->time_base);
636         if (ti1 < 0.01)
637             ti1 = 0.01;
638
639         bitrate     = (frame_size * 8) / av_q2d(enc->time_base) / 1000.0;
640         avg_bitrate = (double)(video_size * 8) / ti1 / 1000.0;
641         fprintf(vstats_file, "s_size= %8.0fkB time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ",
642                (double)video_size / 1024, ti1, bitrate, avg_bitrate);
643         fprintf(vstats_file, "type= %c\n", av_get_picture_type_char(enc->coded_frame->pict_type));
644     }
645 }
646
647 /**
648  * Read one frame for lavfi output for ost and encode it.
649  */
650 static int poll_filter(OutputStream *ost)
651 {
652     OutputFile    *of = output_files[ost->file_index];
653     AVFilterBufferRef *picref;
654     AVFrame *filtered_frame = NULL;
655     int frame_size, ret;
656
657     if (!ost->filtered_frame && !(ost->filtered_frame = avcodec_alloc_frame())) {
658         return AVERROR(ENOMEM);
659     } else
660         avcodec_get_frame_defaults(ost->filtered_frame);
661     filtered_frame = ost->filtered_frame;
662
663     if (ost->enc->type == AVMEDIA_TYPE_AUDIO &&
664         !(ost->enc->capabilities & CODEC_CAP_VARIABLE_FRAME_SIZE))
665         ret = av_buffersink_read_samples(ost->filter->filter, &picref,
666                                          ost->st->codec->frame_size);
667     else
668         ret = av_buffersink_read(ost->filter->filter, &picref);
669
670     if (ret < 0)
671         return ret;
672
673     avfilter_copy_buf_props(filtered_frame, picref);
674     if (picref->pts != AV_NOPTS_VALUE) {
675         filtered_frame->pts = av_rescale_q(picref->pts,
676                                            ost->filter->filter->inputs[0]->time_base,
677                                            ost->st->codec->time_base) -
678                               av_rescale_q(of->start_time,
679                                            AV_TIME_BASE_Q,
680                                            ost->st->codec->time_base);
681
682         if (of->start_time && filtered_frame->pts < 0) {
683             avfilter_unref_buffer(picref);
684             return 0;
685         }
686     }
687
688     switch (ost->filter->filter->inputs[0]->type) {
689     case AVMEDIA_TYPE_VIDEO:
690         if (!ost->frame_aspect_ratio)
691             ost->st->codec->sample_aspect_ratio = picref->video->pixel_aspect;
692
693         do_video_out(of->ctx, ost, filtered_frame, &frame_size,
694                      same_quant ? ost->last_quality :
695                                   ost->st->codec->global_quality);
696         if (vstats_filename && frame_size)
697             do_video_stats(of->ctx, ost, frame_size);
698         break;
699     case AVMEDIA_TYPE_AUDIO:
700         do_audio_out(of->ctx, ost, filtered_frame);
701         break;
702     default:
703         // TODO support subtitle filters
704         av_assert0(0);
705     }
706
707     avfilter_unref_buffer(picref);
708
709     return 0;
710 }
711
712 /**
713  * Read as many frames from possible from lavfi and encode them.
714  *
715  * Always read from the active stream with the lowest timestamp. If no frames
716  * are available for it then return EAGAIN and wait for more input. This way we
717  * can use lavfi sources that generate unlimited amount of frames without memory
718  * usage exploding.
719  */
720 static int poll_filters(void)
721 {
722     int i, ret = 0;
723
724     while (ret >= 0 && !received_sigterm) {
725         OutputStream *ost = NULL;
726         int64_t min_pts = INT64_MAX;
727
728         /* choose output stream with the lowest timestamp */
729         for (i = 0; i < nb_output_streams; i++) {
730             int64_t pts = output_streams[i]->sync_opts;
731
732             if (!output_streams[i]->filter || output_streams[i]->finished)
733                 continue;
734
735             pts = av_rescale_q(pts, output_streams[i]->st->codec->time_base,
736                                AV_TIME_BASE_Q);
737             if (pts < min_pts) {
738                 min_pts = pts;
739                 ost = output_streams[i];
740             }
741         }
742
743         if (!ost)
744             break;
745
746         ret = poll_filter(ost);
747
748         if (ret == AVERROR_EOF) {
749             ost->finished = 1;
750
751             if (opt_shortest)
752                 return ret;
753
754             ret = 0;
755         } else if (ret == AVERROR(EAGAIN))
756             return 0;
757     }
758
759     return ret;
760 }
761
762 static void print_report(int is_last_report, int64_t timer_start)
763 {
764     char buf[1024];
765     OutputStream *ost;
766     AVFormatContext *oc;
767     int64_t total_size;
768     AVCodecContext *enc;
769     int frame_number, vid, i;
770     double bitrate, ti1, pts;
771     static int64_t last_time = -1;
772     static int qp_histogram[52];
773
774     if (!print_stats && !is_last_report)
775         return;
776
777     if (!is_last_report) {
778         int64_t cur_time;
779         /* display the report every 0.5 seconds */
780         cur_time = av_gettime();
781         if (last_time == -1) {
782             last_time = cur_time;
783             return;
784         }
785         if ((cur_time - last_time) < 500000)
786             return;
787         last_time = cur_time;
788     }
789
790
791     oc = output_files[0]->ctx;
792
793     total_size = avio_size(oc->pb);
794     if (total_size < 0) // FIXME improve avio_size() so it works with non seekable output too
795         total_size = avio_tell(oc->pb);
796
797     buf[0] = '\0';
798     ti1 = 1e10;
799     vid = 0;
800     for (i = 0; i < nb_output_streams; i++) {
801         float q = -1;
802         ost = output_streams[i];
803         enc = ost->st->codec;
804         if (!ost->stream_copy && enc->coded_frame)
805             q = enc->coded_frame->quality / (float)FF_QP2LAMBDA;
806         if (vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {
807             snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "q=%2.1f ", q);
808         }
809         if (!vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {
810             float t = (av_gettime() - timer_start) / 1000000.0;
811
812             frame_number = ost->frame_number;
813             snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "frame=%5d fps=%3d q=%3.1f ",
814                      frame_number, (t > 1) ? (int)(frame_number / t + 0.5) : 0, q);
815             if (is_last_report)
816                 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "L");
817             if (qp_hist) {
818                 int j;
819                 int qp = lrintf(q);
820                 if (qp >= 0 && qp < FF_ARRAY_ELEMS(qp_histogram))
821                     qp_histogram[qp]++;
822                 for (j = 0; j < 32; j++)
823                     snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%X", (int)lrintf(log(qp_histogram[j] + 1) / log(2)));
824             }
825             if (enc->flags&CODEC_FLAG_PSNR) {
826                 int j;
827                 double error, error_sum = 0;
828                 double scale, scale_sum = 0;
829                 char type[3] = { 'Y','U','V' };
830                 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "PSNR=");
831                 for (j = 0; j < 3; j++) {
832                     if (is_last_report) {
833                         error = enc->error[j];
834                         scale = enc->width * enc->height * 255.0 * 255.0 * frame_number;
835                     } else {
836                         error = enc->coded_frame->error[j];
837                         scale = enc->width * enc->height * 255.0 * 255.0;
838                     }
839                     if (j)
840                         scale /= 4;
841                     error_sum += error;
842                     scale_sum += scale;
843                     snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%c:%2.2f ", type[j], psnr(error / scale));
844                 }
845                 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "*:%2.2f ", psnr(error_sum / scale_sum));
846             }
847             vid = 1;
848         }
849         /* compute min output value */
850         pts = (double)ost->st->pts.val * av_q2d(ost->st->time_base);
851         if ((pts < ti1) && (pts > 0))
852             ti1 = pts;
853     }
854     if (ti1 < 0.01)
855         ti1 = 0.01;
856
857     bitrate = (double)(total_size * 8) / ti1 / 1000.0;
858
859     snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),
860             "size=%8.0fkB time=%0.2f bitrate=%6.1fkbits/s",
861             (double)total_size / 1024, ti1, bitrate);
862
863     if (nb_frames_dup || nb_frames_drop)
864         snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), " dup=%d drop=%d",
865                 nb_frames_dup, nb_frames_drop);
866
867     av_log(NULL, AV_LOG_INFO, "%s    \r", buf);
868
869     fflush(stderr);
870
871     if (is_last_report) {
872         int64_t raw= audio_size + video_size + extra_size;
873         av_log(NULL, AV_LOG_INFO, "\n");
874         av_log(NULL, AV_LOG_INFO, "video:%1.0fkB audio:%1.0fkB global headers:%1.0fkB muxing overhead %f%%\n",
875                video_size / 1024.0,
876                audio_size / 1024.0,
877                extra_size / 1024.0,
878                100.0 * (total_size - raw) / raw
879         );
880     }
881 }
882
883 static void flush_encoders(void)
884 {
885     int i, ret;
886
887     for (i = 0; i < nb_output_streams; i++) {
888         OutputStream   *ost = output_streams[i];
889         AVCodecContext *enc = ost->st->codec;
890         AVFormatContext *os = output_files[ost->file_index]->ctx;
891         int stop_encoding = 0;
892
893         if (!ost->encoding_needed)
894             continue;
895
896         if (ost->st->codec->codec_type == AVMEDIA_TYPE_AUDIO && enc->frame_size <= 1)
897             continue;
898         if (ost->st->codec->codec_type == AVMEDIA_TYPE_VIDEO && (os->oformat->flags & AVFMT_RAWPICTURE) && enc->codec->id == AV_CODEC_ID_RAWVIDEO)
899             continue;
900
901         for (;;) {
902             int (*encode)(AVCodecContext*, AVPacket*, const AVFrame*, int*) = NULL;
903             const char *desc;
904             int64_t *size;
905
906             switch (ost->st->codec->codec_type) {
907             case AVMEDIA_TYPE_AUDIO:
908                 encode = avcodec_encode_audio2;
909                 desc   = "Audio";
910                 size   = &audio_size;
911                 break;
912             case AVMEDIA_TYPE_VIDEO:
913                 encode = avcodec_encode_video2;
914                 desc   = "Video";
915                 size   = &video_size;
916                 break;
917             default:
918                 stop_encoding = 1;
919             }
920
921             if (encode) {
922                 AVPacket pkt;
923                 int got_packet;
924                 av_init_packet(&pkt);
925                 pkt.data = NULL;
926                 pkt.size = 0;
927
928                 ret = encode(enc, &pkt, NULL, &got_packet);
929                 if (ret < 0) {
930                     av_log(NULL, AV_LOG_FATAL, "%s encoding failed\n", desc);
931                     exit_program(1);
932                 }
933                 *size += ret;
934                 if (ost->logfile && enc->stats_out) {
935                     fprintf(ost->logfile, "%s", enc->stats_out);
936                 }
937                 if (!got_packet) {
938                     stop_encoding = 1;
939                     break;
940                 }
941                 if (pkt.pts != AV_NOPTS_VALUE)
942                     pkt.pts = av_rescale_q(pkt.pts, enc->time_base, ost->st->time_base);
943                 if (pkt.dts != AV_NOPTS_VALUE)
944                     pkt.dts = av_rescale_q(pkt.dts, enc->time_base, ost->st->time_base);
945                 write_frame(os, &pkt, ost);
946             }
947
948             if (stop_encoding)
949                 break;
950         }
951     }
952 }
953
954 /*
955  * Check whether a packet from ist should be written into ost at this time
956  */
957 static int check_output_constraints(InputStream *ist, OutputStream *ost)
958 {
959     OutputFile *of = output_files[ost->file_index];
960     int ist_index  = input_files[ist->file_index]->ist_index + ist->st->index;
961
962     if (ost->source_index != ist_index)
963         return 0;
964
965     if (of->start_time && ist->last_dts < of->start_time)
966         return 0;
967
968     return 1;
969 }
970
971 static void do_streamcopy(InputStream *ist, OutputStream *ost, const AVPacket *pkt)
972 {
973     OutputFile *of = output_files[ost->file_index];
974     int64_t ost_tb_start_time = av_rescale_q(of->start_time, AV_TIME_BASE_Q, ost->st->time_base);
975     AVPacket opkt;
976
977     av_init_packet(&opkt);
978
979     if ((!ost->frame_number && !(pkt->flags & AV_PKT_FLAG_KEY)) &&
980         !ost->copy_initial_nonkeyframes)
981         return;
982
983     if (of->recording_time != INT64_MAX &&
984         ist->last_dts >= of->recording_time + of->start_time) {
985         ost->finished = 1;
986         return;
987     }
988
989     /* force the input stream PTS */
990     if (ost->st->codec->codec_type == AVMEDIA_TYPE_AUDIO)
991         audio_size += pkt->size;
992     else if (ost->st->codec->codec_type == AVMEDIA_TYPE_VIDEO) {
993         video_size += pkt->size;
994         ost->sync_opts++;
995     }
996
997     if (pkt->pts != AV_NOPTS_VALUE)
998         opkt.pts = av_rescale_q(pkt->pts, ist->st->time_base, ost->st->time_base) - ost_tb_start_time;
999     else
1000         opkt.pts = AV_NOPTS_VALUE;
1001
1002     if (pkt->dts == AV_NOPTS_VALUE)
1003         opkt.dts = av_rescale_q(ist->last_dts, AV_TIME_BASE_Q, ost->st->time_base);
1004     else
1005         opkt.dts = av_rescale_q(pkt->dts, ist->st->time_base, ost->st->time_base);
1006     opkt.dts -= ost_tb_start_time;
1007
1008     opkt.duration = av_rescale_q(pkt->duration, ist->st->time_base, ost->st->time_base);
1009     opkt.flags    = pkt->flags;
1010
1011     // FIXME remove the following 2 lines they shall be replaced by the bitstream filters
1012     if (  ost->st->codec->codec_id != AV_CODEC_ID_H264
1013        && ost->st->codec->codec_id != AV_CODEC_ID_MPEG1VIDEO
1014        && ost->st->codec->codec_id != AV_CODEC_ID_MPEG2VIDEO
1015        && ost->st->codec->codec_id != AV_CODEC_ID_VC1
1016        ) {
1017         if (av_parser_change(ist->st->parser, ost->st->codec, &opkt.data, &opkt.size, pkt->data, pkt->size, pkt->flags & AV_PKT_FLAG_KEY))
1018             opkt.destruct = av_destruct_packet;
1019     } else {
1020         opkt.data = pkt->data;
1021         opkt.size = pkt->size;
1022     }
1023
1024     write_frame(of->ctx, &opkt, ost);
1025     ost->st->codec->frame_number++;
1026     av_free_packet(&opkt);
1027 }
1028
1029 static void rate_emu_sleep(InputStream *ist)
1030 {
1031     if (input_files[ist->file_index]->rate_emu) {
1032         int64_t pts = av_rescale(ist->last_dts, 1000000, AV_TIME_BASE);
1033         int64_t now = av_gettime() - ist->start;
1034         if (pts > now)
1035             av_usleep(pts - now);
1036     }
1037 }
1038
1039 int guess_input_channel_layout(InputStream *ist)
1040 {
1041     AVCodecContext *dec = ist->st->codec;
1042
1043     if (!dec->channel_layout) {
1044         char layout_name[256];
1045
1046         dec->channel_layout = av_get_default_channel_layout(dec->channels);
1047         if (!dec->channel_layout)
1048             return 0;
1049         av_get_channel_layout_string(layout_name, sizeof(layout_name),
1050                                      dec->channels, dec->channel_layout);
1051         av_log(NULL, AV_LOG_WARNING, "Guessed Channel Layout for  Input Stream "
1052                "#%d.%d : %s\n", ist->file_index, ist->st->index, layout_name);
1053     }
1054     return 1;
1055 }
1056
1057 static int decode_audio(InputStream *ist, AVPacket *pkt, int *got_output)
1058 {
1059     AVFrame *decoded_frame;
1060     AVCodecContext *avctx = ist->st->codec;
1061     int bps = av_get_bytes_per_sample(ist->st->codec->sample_fmt);
1062     int i, ret, resample_changed;
1063
1064     if (!ist->decoded_frame && !(ist->decoded_frame = avcodec_alloc_frame()))
1065         return AVERROR(ENOMEM);
1066     else
1067         avcodec_get_frame_defaults(ist->decoded_frame);
1068     decoded_frame = ist->decoded_frame;
1069
1070     ret = avcodec_decode_audio4(avctx, decoded_frame, got_output, pkt);
1071     if (!*got_output || ret < 0) {
1072         if (!pkt->size) {
1073             for (i = 0; i < ist->nb_filters; i++)
1074                 av_buffersrc_buffer(ist->filters[i]->filter, NULL);
1075         }
1076         return ret;
1077     }
1078
1079     /* if the decoder provides a pts, use it instead of the last packet pts.
1080        the decoder could be delaying output by a packet or more. */
1081     if (decoded_frame->pts != AV_NOPTS_VALUE)
1082         ist->next_dts = decoded_frame->pts;
1083     else if (pkt->pts != AV_NOPTS_VALUE) {
1084         decoded_frame->pts = pkt->pts;
1085         pkt->pts           = AV_NOPTS_VALUE;
1086     }
1087
1088     // preprocess audio (volume)
1089     if (audio_volume != 256) {
1090         int decoded_data_size = decoded_frame->nb_samples * avctx->channels * bps;
1091         void *samples = decoded_frame->data[0];
1092         switch (avctx->sample_fmt) {
1093         case AV_SAMPLE_FMT_U8:
1094         {
1095             uint8_t *volp = samples;
1096             for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1097                 int v = (((*volp - 128) * audio_volume + 128) >> 8) + 128;
1098                 *volp++ = av_clip_uint8(v);
1099             }
1100             break;
1101         }
1102         case AV_SAMPLE_FMT_S16:
1103         {
1104             int16_t *volp = samples;
1105             for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1106                 int v = ((*volp) * audio_volume + 128) >> 8;
1107                 *volp++ = av_clip_int16(v);
1108             }
1109             break;
1110         }
1111         case AV_SAMPLE_FMT_S32:
1112         {
1113             int32_t *volp = samples;
1114             for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1115                 int64_t v = (((int64_t)*volp * audio_volume + 128) >> 8);
1116                 *volp++ = av_clipl_int32(v);
1117             }
1118             break;
1119         }
1120         case AV_SAMPLE_FMT_FLT:
1121         {
1122             float *volp = samples;
1123             float scale = audio_volume / 256.f;
1124             for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1125                 *volp++ *= scale;
1126             }
1127             break;
1128         }
1129         case AV_SAMPLE_FMT_DBL:
1130         {
1131             double *volp = samples;
1132             double scale = audio_volume / 256.;
1133             for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1134                 *volp++ *= scale;
1135             }
1136             break;
1137         }
1138         default:
1139             av_log(NULL, AV_LOG_FATAL,
1140                    "Audio volume adjustment on sample format %s is not supported.\n",
1141                    av_get_sample_fmt_name(ist->st->codec->sample_fmt));
1142             exit_program(1);
1143         }
1144     }
1145
1146     rate_emu_sleep(ist);
1147
1148     resample_changed = ist->resample_sample_fmt     != decoded_frame->format         ||
1149                        ist->resample_channels       != avctx->channels               ||
1150                        ist->resample_channel_layout != decoded_frame->channel_layout ||
1151                        ist->resample_sample_rate    != decoded_frame->sample_rate;
1152     if (resample_changed) {
1153         char layout1[64], layout2[64];
1154
1155         if (!guess_input_channel_layout(ist)) {
1156             av_log(NULL, AV_LOG_FATAL, "Unable to find default channel "
1157                    "layout for Input Stream #%d.%d\n", ist->file_index,
1158                    ist->st->index);
1159             exit_program(1);
1160         }
1161         decoded_frame->channel_layout = avctx->channel_layout;
1162
1163         av_get_channel_layout_string(layout1, sizeof(layout1), ist->resample_channels,
1164                                      ist->resample_channel_layout);
1165         av_get_channel_layout_string(layout2, sizeof(layout2), avctx->channels,
1166                                      decoded_frame->channel_layout);
1167
1168         av_log(NULL, AV_LOG_INFO,
1169                "Input stream #%d:%d frame changed from rate:%d fmt:%s ch:%d chl:%s to rate:%d fmt:%s ch:%d chl:%s\n",
1170                ist->file_index, ist->st->index,
1171                ist->resample_sample_rate,  av_get_sample_fmt_name(ist->resample_sample_fmt),
1172                ist->resample_channels, layout1,
1173                decoded_frame->sample_rate, av_get_sample_fmt_name(decoded_frame->format),
1174                avctx->channels, layout2);
1175
1176         ist->resample_sample_fmt     = decoded_frame->format;
1177         ist->resample_sample_rate    = decoded_frame->sample_rate;
1178         ist->resample_channel_layout = decoded_frame->channel_layout;
1179         ist->resample_channels       = avctx->channels;
1180
1181         for (i = 0; i < nb_filtergraphs; i++)
1182             if (ist_in_filtergraph(filtergraphs[i], ist) &&
1183                 configure_filtergraph(filtergraphs[i]) < 0) {
1184                 av_log(NULL, AV_LOG_FATAL, "Error reinitializing filters!\n");
1185                 exit_program(1);
1186             }
1187     }
1188
1189     if (decoded_frame->pts != AV_NOPTS_VALUE)
1190         decoded_frame->pts = av_rescale_q(decoded_frame->pts,
1191                                           ist->st->time_base,
1192                                           (AVRational){1, ist->st->codec->sample_rate});
1193     for (i = 0; i < ist->nb_filters; i++)
1194         av_buffersrc_write_frame(ist->filters[i]->filter, decoded_frame);
1195
1196     return ret;
1197 }
1198
1199 static int decode_video(InputStream *ist, AVPacket *pkt, int *got_output)
1200 {
1201     AVFrame *decoded_frame;
1202     void *buffer_to_free = NULL;
1203     int i, ret = 0, resample_changed;
1204     float quality;
1205
1206     if (!ist->decoded_frame && !(ist->decoded_frame = avcodec_alloc_frame()))
1207         return AVERROR(ENOMEM);
1208     else
1209         avcodec_get_frame_defaults(ist->decoded_frame);
1210     decoded_frame = ist->decoded_frame;
1211
1212     ret = avcodec_decode_video2(ist->st->codec,
1213                                 decoded_frame, got_output, pkt);
1214     if (!*got_output || ret < 0) {
1215         if (!pkt->size) {
1216             for (i = 0; i < ist->nb_filters; i++)
1217                 av_buffersrc_buffer(ist->filters[i]->filter, NULL);
1218         }
1219         return ret;
1220     }
1221
1222     quality = same_quant ? decoded_frame->quality : 0;
1223     decoded_frame->pts = guess_correct_pts(&ist->pts_ctx, decoded_frame->pkt_pts,
1224                                            decoded_frame->pkt_dts);
1225     pkt->size = 0;
1226     pre_process_video_frame(ist, (AVPicture *)decoded_frame, &buffer_to_free);
1227
1228     rate_emu_sleep(ist);
1229
1230     if (ist->st->sample_aspect_ratio.num)
1231         decoded_frame->sample_aspect_ratio = ist->st->sample_aspect_ratio;
1232
1233     resample_changed = ist->resample_width   != decoded_frame->width  ||
1234                        ist->resample_height  != decoded_frame->height ||
1235                        ist->resample_pix_fmt != decoded_frame->format;
1236     if (resample_changed) {
1237         av_log(NULL, AV_LOG_INFO,
1238                "Input stream #%d:%d frame changed from size:%dx%d fmt:%s to size:%dx%d fmt:%s\n",
1239                ist->file_index, ist->st->index,
1240                ist->resample_width,  ist->resample_height,  av_get_pix_fmt_name(ist->resample_pix_fmt),
1241                decoded_frame->width, decoded_frame->height, av_get_pix_fmt_name(decoded_frame->format));
1242
1243         ist->resample_width   = decoded_frame->width;
1244         ist->resample_height  = decoded_frame->height;
1245         ist->resample_pix_fmt = decoded_frame->format;
1246
1247         for (i = 0; i < nb_filtergraphs; i++)
1248             if (ist_in_filtergraph(filtergraphs[i], ist) &&
1249                 configure_filtergraph(filtergraphs[i]) < 0) {
1250                 av_log(NULL, AV_LOG_FATAL, "Error reinitializing filters!\n");
1251                 exit_program(1);
1252             }
1253     }
1254
1255     for (i = 0; i < ist->nb_filters; i++) {
1256         // XXX what an ugly hack
1257         if (ist->filters[i]->graph->nb_outputs == 1)
1258             ist->filters[i]->graph->outputs[0]->ost->last_quality = quality;
1259
1260         if (ist->st->codec->codec->capabilities & CODEC_CAP_DR1) {
1261             FrameBuffer      *buf = decoded_frame->opaque;
1262             AVFilterBufferRef *fb = avfilter_get_video_buffer_ref_from_arrays(
1263                                         decoded_frame->data, decoded_frame->linesize,
1264                                         AV_PERM_READ | AV_PERM_PRESERVE,
1265                                         ist->st->codec->width, ist->st->codec->height,
1266                                         ist->st->codec->pix_fmt);
1267
1268             avfilter_copy_frame_props(fb, decoded_frame);
1269             fb->buf->priv           = buf;
1270             fb->buf->free           = filter_release_buffer;
1271
1272             buf->refcount++;
1273             av_buffersrc_buffer(ist->filters[i]->filter, fb);
1274         } else
1275             av_buffersrc_write_frame(ist->filters[i]->filter, decoded_frame);
1276     }
1277
1278     av_free(buffer_to_free);
1279     return ret;
1280 }
1281
1282 static int transcode_subtitles(InputStream *ist, AVPacket *pkt, int *got_output)
1283 {
1284     AVSubtitle subtitle;
1285     int i, ret = avcodec_decode_subtitle2(ist->st->codec,
1286                                           &subtitle, got_output, pkt);
1287     if (ret < 0)
1288         return ret;
1289     if (!*got_output)
1290         return ret;
1291
1292     rate_emu_sleep(ist);
1293
1294     for (i = 0; i < nb_output_streams; i++) {
1295         OutputStream *ost = output_streams[i];
1296
1297         if (!check_output_constraints(ist, ost) || !ost->encoding_needed)
1298             continue;
1299
1300         do_subtitle_out(output_files[ost->file_index]->ctx, ost, ist, &subtitle, pkt->pts);
1301     }
1302
1303     avsubtitle_free(&subtitle);
1304     return ret;
1305 }
1306
1307 /* pkt = NULL means EOF (needed to flush decoder buffers) */
1308 static int output_packet(InputStream *ist, const AVPacket *pkt)
1309 {
1310     int i;
1311     int got_output;
1312     AVPacket avpkt;
1313
1314     if (ist->next_dts == AV_NOPTS_VALUE)
1315         ist->next_dts = ist->last_dts;
1316
1317     if (pkt == NULL) {
1318         /* EOF handling */
1319         av_init_packet(&avpkt);
1320         avpkt.data = NULL;
1321         avpkt.size = 0;
1322         goto handle_eof;
1323     } else {
1324         avpkt = *pkt;
1325     }
1326
1327     if (pkt->dts != AV_NOPTS_VALUE)
1328         ist->next_dts = ist->last_dts = av_rescale_q(pkt->dts, ist->st->time_base, AV_TIME_BASE_Q);
1329
1330     // while we have more to decode or while the decoder did output something on EOF
1331     while (ist->decoding_needed && (avpkt.size > 0 || (!pkt && got_output))) {
1332         int ret = 0;
1333     handle_eof:
1334
1335         ist->last_dts = ist->next_dts;
1336
1337         if (avpkt.size && avpkt.size != pkt->size) {
1338             av_log(NULL, ist->showed_multi_packet_warning ? AV_LOG_VERBOSE : AV_LOG_WARNING,
1339                    "Multiple frames in a packet from stream %d\n", pkt->stream_index);
1340             ist->showed_multi_packet_warning = 1;
1341         }
1342
1343         switch (ist->st->codec->codec_type) {
1344         case AVMEDIA_TYPE_AUDIO:
1345             ret = decode_audio    (ist, &avpkt, &got_output);
1346             break;
1347         case AVMEDIA_TYPE_VIDEO:
1348             ret = decode_video    (ist, &avpkt, &got_output);
1349             if (avpkt.duration)
1350                 ist->next_dts += av_rescale_q(avpkt.duration, ist->st->time_base, AV_TIME_BASE_Q);
1351             else if (ist->st->avg_frame_rate.num)
1352                 ist->next_dts += av_rescale_q(1, av_inv_q(ist->st->avg_frame_rate),
1353                                               AV_TIME_BASE_Q);
1354             else if (ist->st->codec->time_base.num != 0) {
1355                 int ticks      = ist->st->parser ? ist->st->parser->repeat_pict + 1 :
1356                                                    ist->st->codec->ticks_per_frame;
1357                 ist->next_dts += av_rescale_q(ticks, ist->st->codec->time_base, AV_TIME_BASE_Q);
1358             }
1359             break;
1360         case AVMEDIA_TYPE_SUBTITLE:
1361             ret = transcode_subtitles(ist, &avpkt, &got_output);
1362             break;
1363         default:
1364             return -1;
1365         }
1366
1367         if (ret < 0)
1368             return ret;
1369         // touch data and size only if not EOF
1370         if (pkt) {
1371             avpkt.data += ret;
1372             avpkt.size -= ret;
1373         }
1374         if (!got_output) {
1375             continue;
1376         }
1377     }
1378
1379     /* handle stream copy */
1380     if (!ist->decoding_needed) {
1381         rate_emu_sleep(ist);
1382         ist->last_dts = ist->next_dts;
1383         switch (ist->st->codec->codec_type) {
1384         case AVMEDIA_TYPE_AUDIO:
1385             ist->next_dts += ((int64_t)AV_TIME_BASE * ist->st->codec->frame_size) /
1386                              ist->st->codec->sample_rate;
1387             break;
1388         case AVMEDIA_TYPE_VIDEO:
1389             if (ist->st->codec->time_base.num != 0) {
1390                 int ticks = ist->st->parser ? ist->st->parser->repeat_pict + 1 : ist->st->codec->ticks_per_frame;
1391                 ist->next_dts += ((int64_t)AV_TIME_BASE *
1392                                   ist->st->codec->time_base.num * ticks) /
1393                                   ist->st->codec->time_base.den;
1394             }
1395             break;
1396         }
1397     }
1398     for (i = 0; pkt && i < nb_output_streams; i++) {
1399         OutputStream *ost = output_streams[i];
1400
1401         if (!check_output_constraints(ist, ost) || ost->encoding_needed)
1402             continue;
1403
1404         do_streamcopy(ist, ost, pkt);
1405     }
1406
1407     return 0;
1408 }
1409
1410 static void print_sdp(void)
1411 {
1412     char sdp[2048];
1413     int i;
1414     AVFormatContext **avc = av_malloc(sizeof(*avc) * nb_output_files);
1415
1416     if (!avc)
1417         exit_program(1);
1418     for (i = 0; i < nb_output_files; i++)
1419         avc[i] = output_files[i]->ctx;
1420
1421     av_sdp_create(avc, nb_output_files, sdp, sizeof(sdp));
1422     printf("SDP:\n%s\n", sdp);
1423     fflush(stdout);
1424     av_freep(&avc);
1425 }
1426
1427 static int init_input_stream(int ist_index, char *error, int error_len)
1428 {
1429     int i;
1430     InputStream *ist = input_streams[ist_index];
1431     if (ist->decoding_needed) {
1432         AVCodec *codec = ist->dec;
1433         if (!codec) {
1434             snprintf(error, error_len, "Decoder (codec id %d) not found for input stream #%d:%d",
1435                     ist->st->codec->codec_id, ist->file_index, ist->st->index);
1436             return AVERROR(EINVAL);
1437         }
1438
1439         /* update requested sample format for the decoder based on the
1440            corresponding encoder sample format */
1441         for (i = 0; i < nb_output_streams; i++) {
1442             OutputStream *ost = output_streams[i];
1443             if (ost->source_index == ist_index) {
1444                 update_sample_fmt(ist->st->codec, codec, ost->st->codec);
1445                 break;
1446             }
1447         }
1448
1449         if (codec->type == AVMEDIA_TYPE_VIDEO && codec->capabilities & CODEC_CAP_DR1) {
1450             ist->st->codec->get_buffer     = codec_get_buffer;
1451             ist->st->codec->release_buffer = codec_release_buffer;
1452             ist->st->codec->opaque         = &ist->buffer_pool;
1453         }
1454
1455         if (!av_dict_get(ist->opts, "threads", NULL, 0))
1456             av_dict_set(&ist->opts, "threads", "auto", 0);
1457         if (avcodec_open2(ist->st->codec, codec, &ist->opts) < 0) {
1458             snprintf(error, error_len, "Error while opening decoder for input stream #%d:%d",
1459                     ist->file_index, ist->st->index);
1460             return AVERROR(EINVAL);
1461         }
1462         assert_codec_experimental(ist->st->codec, 0);
1463         assert_avoptions(ist->opts);
1464     }
1465
1466     ist->last_dts = ist->st->avg_frame_rate.num ? - ist->st->codec->has_b_frames * AV_TIME_BASE / av_q2d(ist->st->avg_frame_rate) : 0;
1467     ist->next_dts = AV_NOPTS_VALUE;
1468     init_pts_correction(&ist->pts_ctx);
1469     ist->is_start = 1;
1470
1471     return 0;
1472 }
1473
1474 static InputStream *get_input_stream(OutputStream *ost)
1475 {
1476     if (ost->source_index >= 0)
1477         return input_streams[ost->source_index];
1478
1479     if (ost->filter) {
1480         FilterGraph *fg = ost->filter->graph;
1481         int i;
1482
1483         for (i = 0; i < fg->nb_inputs; i++)
1484             if (fg->inputs[i]->ist->st->codec->codec_type == ost->st->codec->codec_type)
1485                 return fg->inputs[i]->ist;
1486     }
1487
1488     return NULL;
1489 }
1490
1491 static void parse_forced_key_frames(char *kf, OutputStream *ost,
1492                                     AVCodecContext *avctx)
1493 {
1494     char *p;
1495     int n = 1, i;
1496     int64_t t;
1497
1498     for (p = kf; *p; p++)
1499         if (*p == ',')
1500             n++;
1501     ost->forced_kf_count = n;
1502     ost->forced_kf_pts   = av_malloc(sizeof(*ost->forced_kf_pts) * n);
1503     if (!ost->forced_kf_pts) {
1504         av_log(NULL, AV_LOG_FATAL, "Could not allocate forced key frames array.\n");
1505         exit_program(1);
1506     }
1507
1508     p = kf;
1509     for (i = 0; i < n; i++) {
1510         char *next = strchr(p, ',');
1511
1512         if (next)
1513             *next++ = 0;
1514
1515         t = parse_time_or_die("force_key_frames", p, 1);
1516         ost->forced_kf_pts[i] = av_rescale_q(t, AV_TIME_BASE_Q, avctx->time_base);
1517
1518         p = next;
1519     }
1520 }
1521
1522 static int transcode_init(void)
1523 {
1524     int ret = 0, i, j, k;
1525     AVFormatContext *oc;
1526     AVCodecContext *codec;
1527     OutputStream *ost;
1528     InputStream *ist;
1529     char error[1024];
1530     int want_sdp = 1;
1531
1532     /* init framerate emulation */
1533     for (i = 0; i < nb_input_files; i++) {
1534         InputFile *ifile = input_files[i];
1535         if (ifile->rate_emu)
1536             for (j = 0; j < ifile->nb_streams; j++)
1537                 input_streams[j + ifile->ist_index]->start = av_gettime();
1538     }
1539
1540     /* output stream init */
1541     for (i = 0; i < nb_output_files; i++) {
1542         oc = output_files[i]->ctx;
1543         if (!oc->nb_streams && !(oc->oformat->flags & AVFMT_NOSTREAMS)) {
1544             av_dump_format(oc, i, oc->filename, 1);
1545             av_log(NULL, AV_LOG_ERROR, "Output file #%d does not contain any stream\n", i);
1546             return AVERROR(EINVAL);
1547         }
1548     }
1549
1550     /* init complex filtergraphs */
1551     for (i = 0; i < nb_filtergraphs; i++)
1552         if ((ret = avfilter_graph_config(filtergraphs[i]->graph, NULL)) < 0)
1553             return ret;
1554
1555     /* for each output stream, we compute the right encoding parameters */
1556     for (i = 0; i < nb_output_streams; i++) {
1557         AVCodecContext *icodec = NULL;
1558         ost = output_streams[i];
1559         oc  = output_files[ost->file_index]->ctx;
1560         ist = get_input_stream(ost);
1561
1562         if (ost->attachment_filename)
1563             continue;
1564
1565         codec  = ost->st->codec;
1566
1567         if (ist) {
1568             icodec = ist->st->codec;
1569
1570             ost->st->disposition          = ist->st->disposition;
1571             codec->bits_per_raw_sample    = icodec->bits_per_raw_sample;
1572             codec->chroma_sample_location = icodec->chroma_sample_location;
1573         }
1574
1575         if (ost->stream_copy) {
1576             uint64_t extra_size;
1577
1578             av_assert0(ist && !ost->filter);
1579
1580             extra_size = (uint64_t)icodec->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE;
1581
1582             if (extra_size > INT_MAX) {
1583                 return AVERROR(EINVAL);
1584             }
1585
1586             /* if stream_copy is selected, no need to decode or encode */
1587             codec->codec_id   = icodec->codec_id;
1588             codec->codec_type = icodec->codec_type;
1589
1590             if (!codec->codec_tag) {
1591                 if (!oc->oformat->codec_tag ||
1592                      av_codec_get_id (oc->oformat->codec_tag, icodec->codec_tag) == codec->codec_id ||
1593                      av_codec_get_tag(oc->oformat->codec_tag, icodec->codec_id) <= 0)
1594                     codec->codec_tag = icodec->codec_tag;
1595             }
1596
1597             codec->bit_rate       = icodec->bit_rate;
1598             codec->rc_max_rate    = icodec->rc_max_rate;
1599             codec->rc_buffer_size = icodec->rc_buffer_size;
1600             codec->field_order    = icodec->field_order;
1601             codec->extradata      = av_mallocz(extra_size);
1602             if (!codec->extradata) {
1603                 return AVERROR(ENOMEM);
1604             }
1605             memcpy(codec->extradata, icodec->extradata, icodec->extradata_size);
1606             codec->extradata_size = icodec->extradata_size;
1607             if (!copy_tb) {
1608                 codec->time_base      = icodec->time_base;
1609                 codec->time_base.num *= icodec->ticks_per_frame;
1610                 av_reduce(&codec->time_base.num, &codec->time_base.den,
1611                           codec->time_base.num, codec->time_base.den, INT_MAX);
1612             } else
1613                 codec->time_base = ist->st->time_base;
1614
1615             switch (codec->codec_type) {
1616             case AVMEDIA_TYPE_AUDIO:
1617                 if (audio_volume != 256) {
1618                     av_log(NULL, AV_LOG_FATAL, "-acodec copy and -vol are incompatible (frames are not decoded)\n");
1619                     exit_program(1);
1620                 }
1621                 codec->channel_layout     = icodec->channel_layout;
1622                 codec->sample_rate        = icodec->sample_rate;
1623                 codec->channels           = icodec->channels;
1624                 codec->frame_size         = icodec->frame_size;
1625                 codec->audio_service_type = icodec->audio_service_type;
1626                 codec->block_align        = icodec->block_align;
1627                 break;
1628             case AVMEDIA_TYPE_VIDEO:
1629                 codec->pix_fmt            = icodec->pix_fmt;
1630                 codec->width              = icodec->width;
1631                 codec->height             = icodec->height;
1632                 codec->has_b_frames       = icodec->has_b_frames;
1633                 if (!codec->sample_aspect_ratio.num) {
1634                     codec->sample_aspect_ratio   =
1635                     ost->st->sample_aspect_ratio =
1636                         ist->st->sample_aspect_ratio.num ? ist->st->sample_aspect_ratio :
1637                         ist->st->codec->sample_aspect_ratio.num ?
1638                         ist->st->codec->sample_aspect_ratio : (AVRational){0, 1};
1639                 }
1640                 break;
1641             case AVMEDIA_TYPE_SUBTITLE:
1642                 codec->width  = icodec->width;
1643                 codec->height = icodec->height;
1644                 break;
1645             case AVMEDIA_TYPE_DATA:
1646             case AVMEDIA_TYPE_ATTACHMENT:
1647                 break;
1648             default:
1649                 abort();
1650             }
1651         } else {
1652             if (!ost->enc) {
1653                 /* should only happen when a default codec is not present. */
1654                 snprintf(error, sizeof(error), "Automatic encoder selection "
1655                          "failed for output stream #%d:%d. Default encoder for "
1656                          "format %s is probably disabled. Please choose an "
1657                          "encoder manually.\n", ost->file_index, ost->index,
1658                          oc->oformat->name);
1659                 ret = AVERROR(EINVAL);
1660                 goto dump_format;
1661             }
1662
1663             if (ist)
1664                 ist->decoding_needed = 1;
1665             ost->encoding_needed = 1;
1666
1667             /*
1668              * We want CFR output if and only if one of those is true:
1669              * 1) user specified output framerate with -r
1670              * 2) user specified -vsync cfr
1671              * 3) output format is CFR and the user didn't force vsync to
1672              *    something else than CFR
1673              *
1674              * in such a case, set ost->frame_rate
1675              */
1676             if (codec->codec_type == AVMEDIA_TYPE_VIDEO &&
1677                 !ost->frame_rate.num && ist &&
1678                 (video_sync_method ==  VSYNC_CFR ||
1679                  (video_sync_method ==  VSYNC_AUTO &&
1680                   !(oc->oformat->flags & (AVFMT_NOTIMESTAMPS | AVFMT_VARIABLE_FPS))))) {
1681                 ost->frame_rate = ist->st->avg_frame_rate.num ? ist->st->avg_frame_rate : (AVRational){25, 1};
1682                 if (ost->enc && ost->enc->supported_framerates && !ost->force_fps) {
1683                     int idx = av_find_nearest_q_idx(ost->frame_rate, ost->enc->supported_framerates);
1684                     ost->frame_rate = ost->enc->supported_framerates[idx];
1685                 }
1686             }
1687
1688             if (!ost->filter &&
1689                 (codec->codec_type == AVMEDIA_TYPE_VIDEO ||
1690                  codec->codec_type == AVMEDIA_TYPE_AUDIO)) {
1691                     FilterGraph *fg;
1692                     fg = init_simple_filtergraph(ist, ost);
1693                     if (configure_filtergraph(fg)) {
1694                         av_log(NULL, AV_LOG_FATAL, "Error opening filters!\n");
1695                         exit(1);
1696                     }
1697             }
1698
1699             switch (codec->codec_type) {
1700             case AVMEDIA_TYPE_AUDIO:
1701                 codec->sample_fmt     = ost->filter->filter->inputs[0]->format;
1702                 codec->sample_rate    = ost->filter->filter->inputs[0]->sample_rate;
1703                 codec->channel_layout = ost->filter->filter->inputs[0]->channel_layout;
1704                 codec->channels       = av_get_channel_layout_nb_channels(codec->channel_layout);
1705                 codec->time_base      = (AVRational){ 1, codec->sample_rate };
1706                 break;
1707             case AVMEDIA_TYPE_VIDEO:
1708                 codec->time_base = ost->filter->filter->inputs[0]->time_base;
1709
1710                 codec->width  = ost->filter->filter->inputs[0]->w;
1711                 codec->height = ost->filter->filter->inputs[0]->h;
1712                 codec->sample_aspect_ratio = ost->st->sample_aspect_ratio =
1713                     ost->frame_aspect_ratio ? // overridden by the -aspect cli option
1714                     av_d2q(ost->frame_aspect_ratio * codec->height/codec->width, 255) :
1715                     ost->filter->filter->inputs[0]->sample_aspect_ratio;
1716                 codec->pix_fmt = ost->filter->filter->inputs[0]->format;
1717
1718                 if (icodec &&
1719                     (codec->width   != icodec->width  ||
1720                      codec->height  != icodec->height ||
1721                      codec->pix_fmt != icodec->pix_fmt)) {
1722                     codec->bits_per_raw_sample = 0;
1723                 }
1724
1725                 if (ost->forced_keyframes)
1726                     parse_forced_key_frames(ost->forced_keyframes, ost,
1727                                             ost->st->codec);
1728                 break;
1729             case AVMEDIA_TYPE_SUBTITLE:
1730                 codec->time_base = (AVRational){1, 1000};
1731                 break;
1732             default:
1733                 abort();
1734                 break;
1735             }
1736             /* two pass mode */
1737             if ((codec->flags & (CODEC_FLAG_PASS1 | CODEC_FLAG_PASS2))) {
1738                 char logfilename[1024];
1739                 FILE *f;
1740
1741                 snprintf(logfilename, sizeof(logfilename), "%s-%d.log",
1742                          pass_logfilename_prefix ? pass_logfilename_prefix : DEFAULT_PASS_LOGFILENAME_PREFIX,
1743                          i);
1744                 if (!strcmp(ost->enc->name, "libx264")) {
1745                     av_dict_set(&ost->opts, "stats", logfilename, AV_DICT_DONT_OVERWRITE);
1746                 } else {
1747                     if (codec->flags & CODEC_FLAG_PASS1) {
1748                         f = fopen(logfilename, "wb");
1749                         if (!f) {
1750                             av_log(NULL, AV_LOG_FATAL, "Cannot write log file '%s' for pass-1 encoding: %s\n",
1751                                    logfilename, strerror(errno));
1752                             exit_program(1);
1753                         }
1754                         ost->logfile = f;
1755                     } else {
1756                         char  *logbuffer;
1757                         size_t logbuffer_size;
1758                         if (cmdutils_read_file(logfilename, &logbuffer, &logbuffer_size) < 0) {
1759                             av_log(NULL, AV_LOG_FATAL, "Error reading log file '%s' for pass-2 encoding\n",
1760                                    logfilename);
1761                             exit_program(1);
1762                         }
1763                         codec->stats_in = logbuffer;
1764                     }
1765                 }
1766             }
1767         }
1768     }
1769
1770     /* open each encoder */
1771     for (i = 0; i < nb_output_streams; i++) {
1772         ost = output_streams[i];
1773         if (ost->encoding_needed) {
1774             AVCodec      *codec = ost->enc;
1775             AVCodecContext *dec = NULL;
1776
1777             if ((ist = get_input_stream(ost)))
1778                 dec = ist->st->codec;
1779             if (dec && dec->subtitle_header) {
1780                 ost->st->codec->subtitle_header = av_malloc(dec->subtitle_header_size);
1781                 if (!ost->st->codec->subtitle_header) {
1782                     ret = AVERROR(ENOMEM);
1783                     goto dump_format;
1784                 }
1785                 memcpy(ost->st->codec->subtitle_header, dec->subtitle_header, dec->subtitle_header_size);
1786                 ost->st->codec->subtitle_header_size = dec->subtitle_header_size;
1787             }
1788             if (!av_dict_get(ost->opts, "threads", NULL, 0))
1789                 av_dict_set(&ost->opts, "threads", "auto", 0);
1790             if (avcodec_open2(ost->st->codec, codec, &ost->opts) < 0) {
1791                 snprintf(error, sizeof(error), "Error while opening encoder for output stream #%d:%d - maybe incorrect parameters such as bit_rate, rate, width or height",
1792                         ost->file_index, ost->index);
1793                 ret = AVERROR(EINVAL);
1794                 goto dump_format;
1795             }
1796             assert_codec_experimental(ost->st->codec, 1);
1797             assert_avoptions(ost->opts);
1798             if (ost->st->codec->bit_rate && ost->st->codec->bit_rate < 1000)
1799                 av_log(NULL, AV_LOG_WARNING, "The bitrate parameter is set too low."
1800                                              "It takes bits/s as argument, not kbits/s\n");
1801             extra_size += ost->st->codec->extradata_size;
1802
1803             if (ost->st->codec->me_threshold)
1804                 input_streams[ost->source_index]->st->codec->debug |= FF_DEBUG_MV;
1805         }
1806     }
1807
1808     /* init input streams */
1809     for (i = 0; i < nb_input_streams; i++)
1810         if ((ret = init_input_stream(i, error, sizeof(error))) < 0)
1811             goto dump_format;
1812
1813     /* discard unused programs */
1814     for (i = 0; i < nb_input_files; i++) {
1815         InputFile *ifile = input_files[i];
1816         for (j = 0; j < ifile->ctx->nb_programs; j++) {
1817             AVProgram *p = ifile->ctx->programs[j];
1818             int discard  = AVDISCARD_ALL;
1819
1820             for (k = 0; k < p->nb_stream_indexes; k++)
1821                 if (!input_streams[ifile->ist_index + p->stream_index[k]]->discard) {
1822                     discard = AVDISCARD_DEFAULT;
1823                     break;
1824                 }
1825             p->discard = discard;
1826         }
1827     }
1828
1829     /* open files and write file headers */
1830     for (i = 0; i < nb_output_files; i++) {
1831         oc = output_files[i]->ctx;
1832         oc->interrupt_callback = int_cb;
1833         if ((ret = avformat_write_header(oc, &output_files[i]->opts)) < 0) {
1834             char errbuf[128];
1835             const char *errbuf_ptr = errbuf;
1836             if (av_strerror(ret, errbuf, sizeof(errbuf)) < 0)
1837                 errbuf_ptr = strerror(AVUNERROR(ret));
1838             snprintf(error, sizeof(error), "Could not write header for output file #%d (incorrect codec parameters ?): %s", i, errbuf_ptr);
1839             ret = AVERROR(EINVAL);
1840             goto dump_format;
1841         }
1842         assert_avoptions(output_files[i]->opts);
1843         if (strcmp(oc->oformat->name, "rtp")) {
1844             want_sdp = 0;
1845         }
1846     }
1847
1848  dump_format:
1849     /* dump the file output parameters - cannot be done before in case
1850        of stream copy */
1851     for (i = 0; i < nb_output_files; i++) {
1852         av_dump_format(output_files[i]->ctx, i, output_files[i]->ctx->filename, 1);
1853     }
1854
1855     /* dump the stream mapping */
1856     av_log(NULL, AV_LOG_INFO, "Stream mapping:\n");
1857     for (i = 0; i < nb_input_streams; i++) {
1858         ist = input_streams[i];
1859
1860         for (j = 0; j < ist->nb_filters; j++) {
1861             if (ist->filters[j]->graph->graph_desc) {
1862                 av_log(NULL, AV_LOG_INFO, "  Stream #%d:%d (%s) -> %s",
1863                        ist->file_index, ist->st->index, ist->dec ? ist->dec->name : "?",
1864                        ist->filters[j]->name);
1865                 if (nb_filtergraphs > 1)
1866                     av_log(NULL, AV_LOG_INFO, " (graph %d)", ist->filters[j]->graph->index);
1867                 av_log(NULL, AV_LOG_INFO, "\n");
1868             }
1869         }
1870     }
1871
1872     for (i = 0; i < nb_output_streams; i++) {
1873         ost = output_streams[i];
1874
1875         if (ost->attachment_filename) {
1876             /* an attached file */
1877             av_log(NULL, AV_LOG_INFO, "  File %s -> Stream #%d:%d\n",
1878                    ost->attachment_filename, ost->file_index, ost->index);
1879             continue;
1880         }
1881
1882         if (ost->filter && ost->filter->graph->graph_desc) {
1883             /* output from a complex graph */
1884             av_log(NULL, AV_LOG_INFO, "  %s", ost->filter->name);
1885             if (nb_filtergraphs > 1)
1886                 av_log(NULL, AV_LOG_INFO, " (graph %d)", ost->filter->graph->index);
1887
1888             av_log(NULL, AV_LOG_INFO, " -> Stream #%d:%d (%s)\n", ost->file_index,
1889                    ost->index, ost->enc ? ost->enc->name : "?");
1890             continue;
1891         }
1892
1893         av_log(NULL, AV_LOG_INFO, "  Stream #%d:%d -> #%d:%d",
1894                input_streams[ost->source_index]->file_index,
1895                input_streams[ost->source_index]->st->index,
1896                ost->file_index,
1897                ost->index);
1898         if (ost->sync_ist != input_streams[ost->source_index])
1899             av_log(NULL, AV_LOG_INFO, " [sync #%d:%d]",
1900                    ost->sync_ist->file_index,
1901                    ost->sync_ist->st->index);
1902         if (ost->stream_copy)
1903             av_log(NULL, AV_LOG_INFO, " (copy)");
1904         else
1905             av_log(NULL, AV_LOG_INFO, " (%s -> %s)", input_streams[ost->source_index]->dec ?
1906                    input_streams[ost->source_index]->dec->name : "?",
1907                    ost->enc ? ost->enc->name : "?");
1908         av_log(NULL, AV_LOG_INFO, "\n");
1909     }
1910
1911     if (ret) {
1912         av_log(NULL, AV_LOG_ERROR, "%s\n", error);
1913         return ret;
1914     }
1915
1916     if (want_sdp) {
1917         print_sdp();
1918     }
1919
1920     return 0;
1921 }
1922
1923 /**
1924  * @return 1 if there are still streams where more output is wanted,
1925  *         0 otherwise
1926  */
1927 static int need_output(void)
1928 {
1929     int i;
1930
1931     for (i = 0; i < nb_output_streams; i++) {
1932         OutputStream *ost    = output_streams[i];
1933         OutputFile *of       = output_files[ost->file_index];
1934         AVFormatContext *os  = output_files[ost->file_index]->ctx;
1935
1936         if (ost->finished ||
1937             (os->pb && avio_tell(os->pb) >= of->limit_filesize))
1938             continue;
1939         if (ost->frame_number >= ost->max_frames) {
1940             int j;
1941             for (j = 0; j < of->ctx->nb_streams; j++)
1942                 output_streams[of->ost_index + j]->finished = 1;
1943             continue;
1944         }
1945
1946         return 1;
1947     }
1948
1949     return 0;
1950 }
1951
1952 static InputFile *select_input_file(void)
1953 {
1954     InputFile *ifile = NULL;
1955     int64_t ipts_min = INT64_MAX;
1956     int i;
1957
1958     for (i = 0; i < nb_input_streams; i++) {
1959         InputStream *ist = input_streams[i];
1960         int64_t ipts     = ist->last_dts;
1961
1962         if (ist->discard || input_files[ist->file_index]->eagain)
1963             continue;
1964         if (!input_files[ist->file_index]->eof_reached) {
1965             if (ipts < ipts_min) {
1966                 ipts_min = ipts;
1967                 ifile    = input_files[ist->file_index];
1968             }
1969         }
1970     }
1971
1972     return ifile;
1973 }
1974
1975 #if HAVE_PTHREADS
1976 static void *input_thread(void *arg)
1977 {
1978     InputFile *f = arg;
1979     int ret = 0;
1980
1981     while (!transcoding_finished && ret >= 0) {
1982         AVPacket pkt;
1983         ret = av_read_frame(f->ctx, &pkt);
1984
1985         if (ret == AVERROR(EAGAIN)) {
1986             av_usleep(10000);
1987             ret = 0;
1988             continue;
1989         } else if (ret < 0)
1990             break;
1991
1992         pthread_mutex_lock(&f->fifo_lock);
1993         while (!av_fifo_space(f->fifo))
1994             pthread_cond_wait(&f->fifo_cond, &f->fifo_lock);
1995
1996         av_dup_packet(&pkt);
1997         av_fifo_generic_write(f->fifo, &pkt, sizeof(pkt), NULL);
1998
1999         pthread_mutex_unlock(&f->fifo_lock);
2000     }
2001
2002     f->finished = 1;
2003     return NULL;
2004 }
2005
2006 static void free_input_threads(void)
2007 {
2008     int i;
2009
2010     if (nb_input_files == 1)
2011         return;
2012
2013     transcoding_finished = 1;
2014
2015     for (i = 0; i < nb_input_files; i++) {
2016         InputFile *f = input_files[i];
2017         AVPacket pkt;
2018
2019         if (!f->fifo || f->joined)
2020             continue;
2021
2022         pthread_mutex_lock(&f->fifo_lock);
2023         while (av_fifo_size(f->fifo)) {
2024             av_fifo_generic_read(f->fifo, &pkt, sizeof(pkt), NULL);
2025             av_free_packet(&pkt);
2026         }
2027         pthread_cond_signal(&f->fifo_cond);
2028         pthread_mutex_unlock(&f->fifo_lock);
2029
2030         pthread_join(f->thread, NULL);
2031         f->joined = 1;
2032
2033         while (av_fifo_size(f->fifo)) {
2034             av_fifo_generic_read(f->fifo, &pkt, sizeof(pkt), NULL);
2035             av_free_packet(&pkt);
2036         }
2037         av_fifo_free(f->fifo);
2038     }
2039 }
2040
2041 static int init_input_threads(void)
2042 {
2043     int i, ret;
2044
2045     if (nb_input_files == 1)
2046         return 0;
2047
2048     for (i = 0; i < nb_input_files; i++) {
2049         InputFile *f = input_files[i];
2050
2051         if (!(f->fifo = av_fifo_alloc(8*sizeof(AVPacket))))
2052             return AVERROR(ENOMEM);
2053
2054         pthread_mutex_init(&f->fifo_lock, NULL);
2055         pthread_cond_init (&f->fifo_cond, NULL);
2056
2057         if ((ret = pthread_create(&f->thread, NULL, input_thread, f)))
2058             return AVERROR(ret);
2059     }
2060     return 0;
2061 }
2062
2063 static int get_input_packet_mt(InputFile *f, AVPacket *pkt)
2064 {
2065     int ret = 0;
2066
2067     pthread_mutex_lock(&f->fifo_lock);
2068
2069     if (av_fifo_size(f->fifo)) {
2070         av_fifo_generic_read(f->fifo, pkt, sizeof(*pkt), NULL);
2071         pthread_cond_signal(&f->fifo_cond);
2072     } else {
2073         if (f->finished)
2074             ret = AVERROR_EOF;
2075         else
2076             ret = AVERROR(EAGAIN);
2077     }
2078
2079     pthread_mutex_unlock(&f->fifo_lock);
2080
2081     return ret;
2082 }
2083 #endif
2084
2085 static int get_input_packet(InputFile *f, AVPacket *pkt)
2086 {
2087 #if HAVE_PTHREADS
2088     if (nb_input_files > 1)
2089         return get_input_packet_mt(f, pkt);
2090 #endif
2091     return av_read_frame(f->ctx, pkt);
2092 }
2093
2094 static int got_eagain(void)
2095 {
2096     int i;
2097     for (i = 0; i < nb_input_files; i++)
2098         if (input_files[i]->eagain)
2099             return 1;
2100     return 0;
2101 }
2102
2103 static void reset_eagain(void)
2104 {
2105     int i;
2106     for (i = 0; i < nb_input_files; i++)
2107         input_files[i]->eagain = 0;
2108 }
2109
2110 /**
2111  * Read one packet from an input file and send it for
2112  * - decoding -> lavfi (audio/video)
2113  * - decoding -> encoding -> muxing (subtitles)
2114  * - muxing (streamcopy)
2115  *
2116  * @return
2117  * - 0 -- one packet was read and processed
2118  * - AVERROR(EAGAIN) -- no packets were available for selected file,
2119  *   this function should be called again
2120  * - AVERROR_EOF -- this function should not be called again
2121  */
2122 static int process_input(void)
2123 {
2124     InputFile *ifile;
2125     AVFormatContext *is;
2126     InputStream *ist;
2127     AVPacket pkt;
2128     int ret, i, j;
2129
2130     /* select the stream that we must read now */
2131     ifile = select_input_file();
2132     /* if none, if is finished */
2133     if (!ifile) {
2134         if (got_eagain()) {
2135             reset_eagain();
2136             av_usleep(10000);
2137             return AVERROR(EAGAIN);
2138         }
2139         av_log(NULL, AV_LOG_VERBOSE, "No more inputs to read from.\n");
2140         return AVERROR_EOF;
2141     }
2142
2143     is  = ifile->ctx;
2144     ret = get_input_packet(ifile, &pkt);
2145
2146     if (ret == AVERROR(EAGAIN)) {
2147         ifile->eagain = 1;
2148         return ret;
2149     }
2150     if (ret < 0) {
2151         if (ret != AVERROR_EOF) {
2152             print_error(is->filename, ret);
2153             if (exit_on_error)
2154                 exit_program(1);
2155         }
2156         ifile->eof_reached = 1;
2157
2158         for (i = 0; i < ifile->nb_streams; i++) {
2159             ist = input_streams[ifile->ist_index + i];
2160             if (ist->decoding_needed)
2161                 output_packet(ist, NULL);
2162
2163             /* mark all outputs that don't go through lavfi as finished */
2164             for (j = 0; j < nb_output_streams; j++) {
2165                 OutputStream *ost = output_streams[j];
2166
2167                 if (ost->source_index == ifile->ist_index + i &&
2168                     (ost->stream_copy || ost->enc->type == AVMEDIA_TYPE_SUBTITLE))
2169                     ost->finished= 1;
2170             }
2171         }
2172
2173         if (opt_shortest)
2174             return AVERROR_EOF;
2175         else
2176             return AVERROR(EAGAIN);
2177     }
2178
2179     reset_eagain();
2180
2181     if (do_pkt_dump) {
2182         av_pkt_dump_log2(NULL, AV_LOG_DEBUG, &pkt, do_hex_dump,
2183                          is->streams[pkt.stream_index]);
2184     }
2185     /* the following test is needed in case new streams appear
2186        dynamically in stream : we ignore them */
2187     if (pkt.stream_index >= ifile->nb_streams)
2188         goto discard_packet;
2189
2190     ist = input_streams[ifile->ist_index + pkt.stream_index];
2191     if (ist->discard)
2192         goto discard_packet;
2193
2194     if (pkt.dts != AV_NOPTS_VALUE)
2195         pkt.dts += av_rescale_q(ifile->ts_offset, AV_TIME_BASE_Q, ist->st->time_base);
2196     if (pkt.pts != AV_NOPTS_VALUE)
2197         pkt.pts += av_rescale_q(ifile->ts_offset, AV_TIME_BASE_Q, ist->st->time_base);
2198
2199     if (pkt.pts != AV_NOPTS_VALUE)
2200         pkt.pts *= ist->ts_scale;
2201     if (pkt.dts != AV_NOPTS_VALUE)
2202         pkt.dts *= ist->ts_scale;
2203
2204     if (pkt.dts != AV_NOPTS_VALUE && ist->next_dts != AV_NOPTS_VALUE &&
2205         (is->iformat->flags & AVFMT_TS_DISCONT)) {
2206         int64_t pkt_dts = av_rescale_q(pkt.dts, ist->st->time_base, AV_TIME_BASE_Q);
2207         int64_t delta   = pkt_dts - ist->next_dts;
2208
2209         if ((FFABS(delta) > 1LL * dts_delta_threshold * AV_TIME_BASE || pkt_dts + 1 < ist->last_dts) && !copy_ts) {
2210             ifile->ts_offset -= delta;
2211             av_log(NULL, AV_LOG_DEBUG,
2212                    "timestamp discontinuity %"PRId64", new offset= %"PRId64"\n",
2213                    delta, ifile->ts_offset);
2214             pkt.dts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);
2215             if (pkt.pts != AV_NOPTS_VALUE)
2216                 pkt.pts -= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);
2217         }
2218     }
2219
2220     ret = output_packet(ist, &pkt);
2221     if (ret < 0) {
2222         av_log(NULL, AV_LOG_ERROR, "Error while decoding stream #%d:%d\n",
2223                ist->file_index, ist->st->index);
2224         if (exit_on_error)
2225             exit_program(1);
2226     }
2227
2228 discard_packet:
2229     av_free_packet(&pkt);
2230
2231     return 0;
2232 }
2233
2234 /*
2235  * The following code is the main loop of the file converter
2236  */
2237 static int transcode(void)
2238 {
2239     int ret, i, need_input = 1;
2240     AVFormatContext *os;
2241     OutputStream *ost;
2242     InputStream *ist;
2243     int64_t timer_start;
2244
2245     ret = transcode_init();
2246     if (ret < 0)
2247         goto fail;
2248
2249     av_log(NULL, AV_LOG_INFO, "Press ctrl-c to stop encoding\n");
2250     term_init();
2251
2252     timer_start = av_gettime();
2253
2254 #if HAVE_PTHREADS
2255     if ((ret = init_input_threads()) < 0)
2256         goto fail;
2257 #endif
2258
2259     while (!received_sigterm) {
2260         /* check if there's any stream where output is still needed */
2261         if (!need_output()) {
2262             av_log(NULL, AV_LOG_VERBOSE, "No more output streams to write to, finishing.\n");
2263             break;
2264         }
2265
2266         /* read and process one input packet if needed */
2267         if (need_input) {
2268             ret = process_input();
2269             if (ret == AVERROR_EOF)
2270                 need_input = 0;
2271         }
2272
2273         ret = poll_filters();
2274         if (ret < 0) {
2275             if (ret == AVERROR_EOF || ret == AVERROR(EAGAIN))
2276                 continue;
2277
2278             av_log(NULL, AV_LOG_ERROR, "Error while filtering.\n");
2279             break;
2280         }
2281
2282         /* dump report by using the output first video and audio streams */
2283         print_report(0, timer_start);
2284     }
2285 #if HAVE_PTHREADS
2286     free_input_threads();
2287 #endif
2288
2289     /* at the end of stream, we must flush the decoder buffers */
2290     for (i = 0; i < nb_input_streams; i++) {
2291         ist = input_streams[i];
2292         if (!input_files[ist->file_index]->eof_reached && ist->decoding_needed) {
2293             output_packet(ist, NULL);
2294         }
2295     }
2296     poll_filters();
2297     flush_encoders();
2298
2299     term_exit();
2300
2301     /* write the trailer if needed and close file */
2302     for (i = 0; i < nb_output_files; i++) {
2303         os = output_files[i]->ctx;
2304         av_write_trailer(os);
2305     }
2306
2307     /* dump report by using the first video and audio streams */
2308     print_report(1, timer_start);
2309
2310     /* close each encoder */
2311     for (i = 0; i < nb_output_streams; i++) {
2312         ost = output_streams[i];
2313         if (ost->encoding_needed) {
2314             av_freep(&ost->st->codec->stats_in);
2315             avcodec_close(ost->st->codec);
2316         }
2317     }
2318
2319     /* close each decoder */
2320     for (i = 0; i < nb_input_streams; i++) {
2321         ist = input_streams[i];
2322         if (ist->decoding_needed) {
2323             avcodec_close(ist->st->codec);
2324         }
2325     }
2326
2327     /* finished ! */
2328     ret = 0;
2329
2330  fail:
2331 #if HAVE_PTHREADS
2332     free_input_threads();
2333 #endif
2334
2335     if (output_streams) {
2336         for (i = 0; i < nb_output_streams; i++) {
2337             ost = output_streams[i];
2338             if (ost) {
2339                 if (ost->stream_copy)
2340                     av_freep(&ost->st->codec->extradata);
2341                 if (ost->logfile) {
2342                     fclose(ost->logfile);
2343                     ost->logfile = NULL;
2344                 }
2345                 av_freep(&ost->st->codec->subtitle_header);
2346                 av_free(ost->forced_kf_pts);
2347                 av_dict_free(&ost->opts);
2348             }
2349         }
2350     }
2351     return ret;
2352 }
2353
2354 static int64_t getutime(void)
2355 {
2356 #if HAVE_GETRUSAGE
2357     struct rusage rusage;
2358
2359     getrusage(RUSAGE_SELF, &rusage);
2360     return (rusage.ru_utime.tv_sec * 1000000LL) + rusage.ru_utime.tv_usec;
2361 #elif HAVE_GETPROCESSTIMES
2362     HANDLE proc;
2363     FILETIME c, e, k, u;
2364     proc = GetCurrentProcess();
2365     GetProcessTimes(proc, &c, &e, &k, &u);
2366     return ((int64_t) u.dwHighDateTime << 32 | u.dwLowDateTime) / 10;
2367 #else
2368     return av_gettime();
2369 #endif
2370 }
2371
2372 static int64_t getmaxrss(void)
2373 {
2374 #if HAVE_GETRUSAGE && HAVE_STRUCT_RUSAGE_RU_MAXRSS
2375     struct rusage rusage;
2376     getrusage(RUSAGE_SELF, &rusage);
2377     return (int64_t)rusage.ru_maxrss * 1024;
2378 #elif HAVE_GETPROCESSMEMORYINFO
2379     HANDLE proc;
2380     PROCESS_MEMORY_COUNTERS memcounters;
2381     proc = GetCurrentProcess();
2382     memcounters.cb = sizeof(memcounters);
2383     GetProcessMemoryInfo(proc, &memcounters, sizeof(memcounters));
2384     return memcounters.PeakPagefileUsage;
2385 #else
2386     return 0;
2387 #endif
2388 }
2389
2390 static void parse_cpuflags(int argc, char **argv, const OptionDef *options)
2391 {
2392     int idx = locate_option(argc, argv, options, "cpuflags");
2393     if (idx && argv[idx + 1])
2394         opt_cpuflags("cpuflags", argv[idx + 1]);
2395 }
2396
2397 int main(int argc, char **argv)
2398 {
2399     OptionsContext o = { 0 };
2400     int64_t ti;
2401
2402     reset_options(&o);
2403
2404     av_log_set_flags(AV_LOG_SKIP_REPEATED);
2405     parse_loglevel(argc, argv, options);
2406
2407     avcodec_register_all();
2408 #if CONFIG_AVDEVICE
2409     avdevice_register_all();
2410 #endif
2411     avfilter_register_all();
2412     av_register_all();
2413     avformat_network_init();
2414
2415     show_banner();
2416
2417     parse_cpuflags(argc, argv, options);
2418
2419     /* parse options */
2420     parse_options(&o, argc, argv, options, opt_output_file);
2421
2422     if (nb_output_files <= 0 && nb_input_files == 0) {
2423         show_usage();
2424         av_log(NULL, AV_LOG_WARNING, "Use -h to get full help or, even better, run 'man %s'\n", program_name);
2425         exit_program(1);
2426     }
2427
2428     /* file converter / grab */
2429     if (nb_output_files <= 0) {
2430         fprintf(stderr, "At least one output file must be specified\n");
2431         exit_program(1);
2432     }
2433
2434     ti = getutime();
2435     if (transcode() < 0)
2436         exit_program(1);
2437     ti = getutime() - ti;
2438     if (do_benchmark) {
2439         int maxrss = getmaxrss() / 1024;
2440         printf("bench: utime=%0.3fs maxrss=%ikB\n", ti / 1000000.0, maxrss);
2441     }
2442
2443     exit_program(0);
2444     return 0;
2445 }