]> git.sesse.net Git - ffmpeg/blob - ffplay.c
Merge remote-tracking branch 'qatar/master'
[ffmpeg] / ffplay.c
1 /*
2  * Copyright (c) 2003 Fabrice Bellard
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20
21 /**
22  * @file
23  * simple media player based on the FFmpeg libraries
24  */
25
26 #include "config.h"
27 #include <inttypes.h>
28 #include <math.h>
29 #include <limits.h>
30 #include <signal.h>
31 #include "libavutil/avstring.h"
32 #include "libavutil/colorspace.h"
33 #include "libavutil/mathematics.h"
34 #include "libavutil/pixdesc.h"
35 #include "libavutil/imgutils.h"
36 #include "libavutil/dict.h"
37 #include "libavutil/parseutils.h"
38 #include "libavutil/samplefmt.h"
39 #include "libavutil/avassert.h"
40 #include "libavutil/time.h"
41 #include "libavformat/avformat.h"
42 #include "libavdevice/avdevice.h"
43 #include "libswscale/swscale.h"
44 #include "libavutil/opt.h"
45 #include "libavcodec/avfft.h"
46 #include "libswresample/swresample.h"
47
48 #if CONFIG_AVFILTER
49 # include "libavfilter/avcodec.h"
50 # include "libavfilter/avfilter.h"
51 # include "libavfilter/avfiltergraph.h"
52 # include "libavfilter/buffersink.h"
53 # include "libavfilter/buffersrc.h"
54 #endif
55
56 #include <SDL.h>
57 #include <SDL_thread.h>
58
59 #include "cmdutils.h"
60
61 #include <assert.h>
62
63 const char program_name[] = "ffplay";
64 const int program_birth_year = 2003;
65
66 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
67 #define MIN_FRAMES 5
68
69 /* SDL audio buffer size, in samples. Should be small to have precise
70    A/V sync as SDL does not have hardware buffer fullness info. */
71 #define SDL_AUDIO_BUFFER_SIZE 1024
72
73 /* no AV sync correction is done if below the AV sync threshold */
74 #define AV_SYNC_THRESHOLD 0.01
75 /* no AV correction is done if too big error */
76 #define AV_NOSYNC_THRESHOLD 10.0
77
78 /* maximum audio speed change to get correct sync */
79 #define SAMPLE_CORRECTION_PERCENT_MAX 10
80
81 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
82 #define AUDIO_DIFF_AVG_NB   20
83
84 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
85 #define SAMPLE_ARRAY_SIZE (2 * 65536)
86
87 static int sws_flags = SWS_BICUBIC;
88
89 typedef struct PacketQueue {
90     AVPacketList *first_pkt, *last_pkt;
91     int nb_packets;
92     int size;
93     int abort_request;
94     SDL_mutex *mutex;
95     SDL_cond *cond;
96 } PacketQueue;
97
98 #define VIDEO_PICTURE_QUEUE_SIZE 2
99 #define SUBPICTURE_QUEUE_SIZE 4
100
101 typedef struct VideoPicture {
102     double pts;                                  ///< presentation time stamp for this picture
103     int64_t pos;                                 ///< byte position in file
104     int skip;
105     SDL_Overlay *bmp;
106     int width, height; /* source height & width */
107     AVRational sample_aspect_ratio;
108     int allocated;
109     int reallocate;
110
111 #if CONFIG_AVFILTER
112     AVFilterBufferRef *picref;
113 #endif
114 } VideoPicture;
115
116 typedef struct SubPicture {
117     double pts; /* presentation time stamp for this picture */
118     AVSubtitle sub;
119 } SubPicture;
120
121 typedef struct AudioParams {
122     int freq;
123     int channels;
124     int channel_layout;
125     enum AVSampleFormat fmt;
126 } AudioParams;
127
128 enum {
129     AV_SYNC_AUDIO_MASTER, /* default choice */
130     AV_SYNC_VIDEO_MASTER,
131     AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
132 };
133
134 typedef struct VideoState {
135     SDL_Thread *read_tid;
136     SDL_Thread *video_tid;
137     SDL_Thread *refresh_tid;
138     AVInputFormat *iformat;
139     int no_background;
140     int abort_request;
141     int force_refresh;
142     int paused;
143     int last_paused;
144     int que_attachments_req;
145     int seek_req;
146     int seek_flags;
147     int64_t seek_pos;
148     int64_t seek_rel;
149     int read_pause_return;
150     AVFormatContext *ic;
151
152     int audio_stream;
153
154     int av_sync_type;
155     double external_clock; /* external clock base */
156     int64_t external_clock_time;
157
158     double audio_clock;
159     double audio_diff_cum; /* used for AV difference average computation */
160     double audio_diff_avg_coef;
161     double audio_diff_threshold;
162     int audio_diff_avg_count;
163     AVStream *audio_st;
164     PacketQueue audioq;
165     int audio_hw_buf_size;
166     DECLARE_ALIGNED(16,uint8_t,audio_buf2)[AVCODEC_MAX_AUDIO_FRAME_SIZE * 4];
167     uint8_t silence_buf[SDL_AUDIO_BUFFER_SIZE];
168     uint8_t *audio_buf;
169     uint8_t *audio_buf1;
170     unsigned int audio_buf_size; /* in bytes */
171     int audio_buf_index; /* in bytes */
172     int audio_write_buf_size;
173     AVPacket audio_pkt_temp;
174     AVPacket audio_pkt;
175     struct AudioParams audio_src;
176     struct AudioParams audio_tgt;
177     struct SwrContext *swr_ctx;
178     double audio_current_pts;
179     double audio_current_pts_drift;
180     int frame_drops_early;
181     int frame_drops_late;
182     AVFrame *frame;
183
184     enum ShowMode {
185         SHOW_MODE_NONE = -1, SHOW_MODE_VIDEO = 0, SHOW_MODE_WAVES, SHOW_MODE_RDFT, SHOW_MODE_NB
186     } show_mode;
187     int16_t sample_array[SAMPLE_ARRAY_SIZE];
188     int sample_array_index;
189     int last_i_start;
190     RDFTContext *rdft;
191     int rdft_bits;
192     FFTSample *rdft_data;
193     int xpos;
194
195     SDL_Thread *subtitle_tid;
196     int subtitle_stream;
197     int subtitle_stream_changed;
198     AVStream *subtitle_st;
199     PacketQueue subtitleq;
200     SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
201     int subpq_size, subpq_rindex, subpq_windex;
202     SDL_mutex *subpq_mutex;
203     SDL_cond *subpq_cond;
204
205     double frame_timer;
206     double frame_last_pts;
207     double frame_last_duration;
208     double frame_last_dropped_pts;
209     double frame_last_returned_time;
210     double frame_last_filter_delay;
211     int64_t frame_last_dropped_pos;
212     double video_clock;                          ///< pts of last decoded frame / predicted pts of next decoded frame
213     int video_stream;
214     AVStream *video_st;
215     PacketQueue videoq;
216     double video_current_pts;                    ///< current displayed pts (different from video_clock if frame fifos are used)
217     double video_current_pts_drift;              ///< video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
218     int64_t video_current_pos;                   ///< current displayed file pos
219     VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
220     int pictq_size, pictq_rindex, pictq_windex;
221     SDL_mutex *pictq_mutex;
222     SDL_cond *pictq_cond;
223 #if !CONFIG_AVFILTER
224     struct SwsContext *img_convert_ctx;
225 #endif
226
227     char filename[1024];
228     int width, height, xleft, ytop;
229     int step;
230
231 #if CONFIG_AVFILTER
232     AVFilterContext *in_video_filter;           ///< the first filter in the video chain
233     AVFilterContext *out_video_filter;          ///< the last filter in the video chain
234     int use_dr1;
235     FrameBuffer *buffer_pool;
236 #endif
237
238     int refresh;
239     int last_video_stream, last_audio_stream, last_subtitle_stream;
240 } VideoState;
241
242 typedef struct AllocEventProps {
243     VideoState *is;
244     AVFrame *frame;
245 } AllocEventProps;
246
247 /* options specified by the user */
248 static AVInputFormat *file_iformat;
249 static const char *input_filename;
250 static const char *window_title;
251 static int fs_screen_width;
252 static int fs_screen_height;
253 static int screen_width  = 0;
254 static int screen_height = 0;
255 static int audio_disable;
256 static int video_disable;
257 static int wanted_stream[AVMEDIA_TYPE_NB] = {
258     [AVMEDIA_TYPE_AUDIO]    = -1,
259     [AVMEDIA_TYPE_VIDEO]    = -1,
260     [AVMEDIA_TYPE_SUBTITLE] = -1,
261 };
262 static int seek_by_bytes = -1;
263 static int display_disable;
264 static int show_status = 1;
265 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
266 static int64_t start_time = AV_NOPTS_VALUE;
267 static int64_t duration = AV_NOPTS_VALUE;
268 static int workaround_bugs = 1;
269 static int fast = 0;
270 static int genpts = 0;
271 static int lowres = 0;
272 static int idct = FF_IDCT_AUTO;
273 static enum AVDiscard skip_frame       = AVDISCARD_DEFAULT;
274 static enum AVDiscard skip_idct        = AVDISCARD_DEFAULT;
275 static enum AVDiscard skip_loop_filter = AVDISCARD_DEFAULT;
276 static int error_concealment = 3;
277 static int decoder_reorder_pts = -1;
278 static int autoexit;
279 static int exit_on_keydown;
280 static int exit_on_mousedown;
281 static int loop = 1;
282 static int framedrop = -1;
283 static int infinite_buffer = 0;
284 static enum ShowMode show_mode = SHOW_MODE_NONE;
285 static const char *audio_codec_name;
286 static const char *subtitle_codec_name;
287 static const char *video_codec_name;
288 static int rdftspeed = 20;
289 #if CONFIG_AVFILTER
290 static char *vfilters = NULL;
291 #endif
292
293 /* current context */
294 static int is_full_screen;
295 static int64_t audio_callback_time;
296
297 static AVPacket flush_pkt;
298
299 #define FF_ALLOC_EVENT   (SDL_USEREVENT)
300 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
301 #define FF_QUIT_EVENT    (SDL_USEREVENT + 2)
302
303 static SDL_Surface *screen;
304
305 void av_noreturn exit_program(int ret)
306 {
307     exit(ret);
308 }
309
310 static int packet_queue_put_private(PacketQueue *q, AVPacket *pkt)
311 {
312     AVPacketList *pkt1;
313
314     if (q->abort_request)
315        return -1;
316
317     pkt1 = av_malloc(sizeof(AVPacketList));
318     if (!pkt1)
319         return -1;
320     pkt1->pkt = *pkt;
321     pkt1->next = NULL;
322
323     if (!q->last_pkt)
324         q->first_pkt = pkt1;
325     else
326         q->last_pkt->next = pkt1;
327     q->last_pkt = pkt1;
328     q->nb_packets++;
329     q->size += pkt1->pkt.size + sizeof(*pkt1);
330     /* XXX: should duplicate packet data in DV case */
331     SDL_CondSignal(q->cond);
332     return 0;
333 }
334
335 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
336 {
337     int ret;
338
339     /* duplicate the packet */
340     if (pkt != &flush_pkt && av_dup_packet(pkt) < 0)
341         return -1;
342
343     SDL_LockMutex(q->mutex);
344     ret = packet_queue_put_private(q, pkt);
345     SDL_UnlockMutex(q->mutex);
346
347     if (pkt != &flush_pkt && ret < 0)
348         av_free_packet(pkt);
349
350     return ret;
351 }
352
353 /* packet queue handling */
354 static void packet_queue_init(PacketQueue *q)
355 {
356     memset(q, 0, sizeof(PacketQueue));
357     q->mutex = SDL_CreateMutex();
358     q->cond = SDL_CreateCond();
359     q->abort_request = 1;
360 }
361
362 static void packet_queue_flush(PacketQueue *q)
363 {
364     AVPacketList *pkt, *pkt1;
365
366     SDL_LockMutex(q->mutex);
367     for (pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
368         pkt1 = pkt->next;
369         av_free_packet(&pkt->pkt);
370         av_freep(&pkt);
371     }
372     q->last_pkt = NULL;
373     q->first_pkt = NULL;
374     q->nb_packets = 0;
375     q->size = 0;
376     SDL_UnlockMutex(q->mutex);
377 }
378
379 static void packet_queue_destroy(PacketQueue *q)
380 {
381     packet_queue_flush(q);
382     SDL_DestroyMutex(q->mutex);
383     SDL_DestroyCond(q->cond);
384 }
385
386 static void packet_queue_abort(PacketQueue *q)
387 {
388     SDL_LockMutex(q->mutex);
389
390     q->abort_request = 1;
391
392     SDL_CondSignal(q->cond);
393
394     SDL_UnlockMutex(q->mutex);
395 }
396
397 static void packet_queue_start(PacketQueue *q)
398 {
399     SDL_LockMutex(q->mutex);
400     q->abort_request = 0;
401     packet_queue_put_private(q, &flush_pkt);
402     SDL_UnlockMutex(q->mutex);
403 }
404
405 /* return < 0 if aborted, 0 if no packet and > 0 if packet.  */
406 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
407 {
408     AVPacketList *pkt1;
409     int ret;
410
411     SDL_LockMutex(q->mutex);
412
413     for (;;) {
414         if (q->abort_request) {
415             ret = -1;
416             break;
417         }
418
419         pkt1 = q->first_pkt;
420         if (pkt1) {
421             q->first_pkt = pkt1->next;
422             if (!q->first_pkt)
423                 q->last_pkt = NULL;
424             q->nb_packets--;
425             q->size -= pkt1->pkt.size + sizeof(*pkt1);
426             *pkt = pkt1->pkt;
427             av_free(pkt1);
428             ret = 1;
429             break;
430         } else if (!block) {
431             ret = 0;
432             break;
433         } else {
434             SDL_CondWait(q->cond, q->mutex);
435         }
436     }
437     SDL_UnlockMutex(q->mutex);
438     return ret;
439 }
440
441 static inline void fill_rectangle(SDL_Surface *screen,
442                                   int x, int y, int w, int h, int color)
443 {
444     SDL_Rect rect;
445     rect.x = x;
446     rect.y = y;
447     rect.w = w;
448     rect.h = h;
449     SDL_FillRect(screen, &rect, color);
450 }
451
452 #define ALPHA_BLEND(a, oldp, newp, s)\
453 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
454
455 #define RGBA_IN(r, g, b, a, s)\
456 {\
457     unsigned int v = ((const uint32_t *)(s))[0];\
458     a = (v >> 24) & 0xff;\
459     r = (v >> 16) & 0xff;\
460     g = (v >> 8) & 0xff;\
461     b = v & 0xff;\
462 }
463
464 #define YUVA_IN(y, u, v, a, s, pal)\
465 {\
466     unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
467     a = (val >> 24) & 0xff;\
468     y = (val >> 16) & 0xff;\
469     u = (val >> 8) & 0xff;\
470     v = val & 0xff;\
471 }
472
473 #define YUVA_OUT(d, y, u, v, a)\
474 {\
475     ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
476 }
477
478
479 #define BPP 1
480
481 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
482 {
483     int wrap, wrap3, width2, skip2;
484     int y, u, v, a, u1, v1, a1, w, h;
485     uint8_t *lum, *cb, *cr;
486     const uint8_t *p;
487     const uint32_t *pal;
488     int dstx, dsty, dstw, dsth;
489
490     dstw = av_clip(rect->w, 0, imgw);
491     dsth = av_clip(rect->h, 0, imgh);
492     dstx = av_clip(rect->x, 0, imgw - dstw);
493     dsty = av_clip(rect->y, 0, imgh - dsth);
494     lum = dst->data[0] + dsty * dst->linesize[0];
495     cb  = dst->data[1] + (dsty >> 1) * dst->linesize[1];
496     cr  = dst->data[2] + (dsty >> 1) * dst->linesize[2];
497
498     width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
499     skip2 = dstx >> 1;
500     wrap = dst->linesize[0];
501     wrap3 = rect->pict.linesize[0];
502     p = rect->pict.data[0];
503     pal = (const uint32_t *)rect->pict.data[1];  /* Now in YCrCb! */
504
505     if (dsty & 1) {
506         lum += dstx;
507         cb += skip2;
508         cr += skip2;
509
510         if (dstx & 1) {
511             YUVA_IN(y, u, v, a, p, pal);
512             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
513             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
514             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
515             cb++;
516             cr++;
517             lum++;
518             p += BPP;
519         }
520         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
521             YUVA_IN(y, u, v, a, p, pal);
522             u1 = u;
523             v1 = v;
524             a1 = a;
525             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
526
527             YUVA_IN(y, u, v, a, p + BPP, pal);
528             u1 += u;
529             v1 += v;
530             a1 += a;
531             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
532             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
533             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
534             cb++;
535             cr++;
536             p += 2 * BPP;
537             lum += 2;
538         }
539         if (w) {
540             YUVA_IN(y, u, v, a, p, pal);
541             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
542             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
543             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
544             p++;
545             lum++;
546         }
547         p += wrap3 - dstw * BPP;
548         lum += wrap - dstw - dstx;
549         cb += dst->linesize[1] - width2 - skip2;
550         cr += dst->linesize[2] - width2 - skip2;
551     }
552     for (h = dsth - (dsty & 1); h >= 2; h -= 2) {
553         lum += dstx;
554         cb += skip2;
555         cr += skip2;
556
557         if (dstx & 1) {
558             YUVA_IN(y, u, v, a, p, pal);
559             u1 = u;
560             v1 = v;
561             a1 = a;
562             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
563             p += wrap3;
564             lum += wrap;
565             YUVA_IN(y, u, v, a, p, pal);
566             u1 += u;
567             v1 += v;
568             a1 += a;
569             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
570             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
571             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
572             cb++;
573             cr++;
574             p += -wrap3 + BPP;
575             lum += -wrap + 1;
576         }
577         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
578             YUVA_IN(y, u, v, a, p, pal);
579             u1 = u;
580             v1 = v;
581             a1 = a;
582             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
583
584             YUVA_IN(y, u, v, a, p + BPP, pal);
585             u1 += u;
586             v1 += v;
587             a1 += a;
588             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
589             p += wrap3;
590             lum += wrap;
591
592             YUVA_IN(y, u, v, a, p, pal);
593             u1 += u;
594             v1 += v;
595             a1 += a;
596             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
597
598             YUVA_IN(y, u, v, a, p + BPP, pal);
599             u1 += u;
600             v1 += v;
601             a1 += a;
602             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
603
604             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
605             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
606
607             cb++;
608             cr++;
609             p += -wrap3 + 2 * BPP;
610             lum += -wrap + 2;
611         }
612         if (w) {
613             YUVA_IN(y, u, v, a, p, pal);
614             u1 = u;
615             v1 = v;
616             a1 = a;
617             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
618             p += wrap3;
619             lum += wrap;
620             YUVA_IN(y, u, v, a, p, pal);
621             u1 += u;
622             v1 += v;
623             a1 += a;
624             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
625             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
626             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
627             cb++;
628             cr++;
629             p += -wrap3 + BPP;
630             lum += -wrap + 1;
631         }
632         p += wrap3 + (wrap3 - dstw * BPP);
633         lum += wrap + (wrap - dstw - dstx);
634         cb += dst->linesize[1] - width2 - skip2;
635         cr += dst->linesize[2] - width2 - skip2;
636     }
637     /* handle odd height */
638     if (h) {
639         lum += dstx;
640         cb += skip2;
641         cr += skip2;
642
643         if (dstx & 1) {
644             YUVA_IN(y, u, v, a, p, pal);
645             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
646             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
647             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
648             cb++;
649             cr++;
650             lum++;
651             p += BPP;
652         }
653         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
654             YUVA_IN(y, u, v, a, p, pal);
655             u1 = u;
656             v1 = v;
657             a1 = a;
658             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
659
660             YUVA_IN(y, u, v, a, p + BPP, pal);
661             u1 += u;
662             v1 += v;
663             a1 += a;
664             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
665             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
666             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
667             cb++;
668             cr++;
669             p += 2 * BPP;
670             lum += 2;
671         }
672         if (w) {
673             YUVA_IN(y, u, v, a, p, pal);
674             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
675             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
676             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
677         }
678     }
679 }
680
681 static void free_subpicture(SubPicture *sp)
682 {
683     avsubtitle_free(&sp->sub);
684 }
685
686 static void video_image_display(VideoState *is)
687 {
688     VideoPicture *vp;
689     SubPicture *sp;
690     AVPicture pict;
691     float aspect_ratio;
692     int width, height, x, y;
693     SDL_Rect rect;
694     int i;
695
696     vp = &is->pictq[is->pictq_rindex];
697     if (vp->bmp) {
698         if (vp->sample_aspect_ratio.num == 0)
699             aspect_ratio = 0;
700         else
701             aspect_ratio = av_q2d(vp->sample_aspect_ratio);
702
703         if (aspect_ratio <= 0.0)
704             aspect_ratio = 1.0;
705         aspect_ratio *= (float)vp->width / (float)vp->height;
706
707         if (is->subtitle_st) {
708             if (is->subpq_size > 0) {
709                 sp = &is->subpq[is->subpq_rindex];
710
711                 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000)) {
712                     SDL_LockYUVOverlay (vp->bmp);
713
714                     pict.data[0] = vp->bmp->pixels[0];
715                     pict.data[1] = vp->bmp->pixels[2];
716                     pict.data[2] = vp->bmp->pixels[1];
717
718                     pict.linesize[0] = vp->bmp->pitches[0];
719                     pict.linesize[1] = vp->bmp->pitches[2];
720                     pict.linesize[2] = vp->bmp->pitches[1];
721
722                     for (i = 0; i < sp->sub.num_rects; i++)
723                         blend_subrect(&pict, sp->sub.rects[i],
724                                       vp->bmp->w, vp->bmp->h);
725
726                     SDL_UnlockYUVOverlay (vp->bmp);
727                 }
728             }
729         }
730
731
732         /* XXX: we suppose the screen has a 1.0 pixel ratio */
733         height = is->height;
734         width = ((int)rint(height * aspect_ratio)) & ~1;
735         if (width > is->width) {
736             width = is->width;
737             height = ((int)rint(width / aspect_ratio)) & ~1;
738         }
739         x = (is->width - width) / 2;
740         y = (is->height - height) / 2;
741         is->no_background = 0;
742         rect.x = is->xleft + x;
743         rect.y = is->ytop  + y;
744         rect.w = FFMAX(width,  1);
745         rect.h = FFMAX(height, 1);
746         SDL_DisplayYUVOverlay(vp->bmp, &rect);
747     }
748 }
749
750 static inline int compute_mod(int a, int b)
751 {
752     return a < 0 ? a%b + b : a%b;
753 }
754
755 static void video_audio_display(VideoState *s)
756 {
757     int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
758     int ch, channels, h, h2, bgcolor, fgcolor;
759     int16_t time_diff;
760     int rdft_bits, nb_freq;
761
762     for (rdft_bits = 1; (1 << rdft_bits) < 2 * s->height; rdft_bits++)
763         ;
764     nb_freq = 1 << (rdft_bits - 1);
765
766     /* compute display index : center on currently output samples */
767     channels = s->audio_tgt.channels;
768     nb_display_channels = channels;
769     if (!s->paused) {
770         int data_used= s->show_mode == SHOW_MODE_WAVES ? s->width : (2*nb_freq);
771         n = 2 * channels;
772         delay = s->audio_write_buf_size;
773         delay /= n;
774
775         /* to be more precise, we take into account the time spent since
776            the last buffer computation */
777         if (audio_callback_time) {
778             time_diff = av_gettime() - audio_callback_time;
779             delay -= (time_diff * s->audio_tgt.freq) / 1000000;
780         }
781
782         delay += 2 * data_used;
783         if (delay < data_used)
784             delay = data_used;
785
786         i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
787         if (s->show_mode == SHOW_MODE_WAVES) {
788             h = INT_MIN;
789             for (i = 0; i < 1000; i += channels) {
790                 int idx = (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
791                 int a = s->sample_array[idx];
792                 int b = s->sample_array[(idx + 4 * channels) % SAMPLE_ARRAY_SIZE];
793                 int c = s->sample_array[(idx + 5 * channels) % SAMPLE_ARRAY_SIZE];
794                 int d = s->sample_array[(idx + 9 * channels) % SAMPLE_ARRAY_SIZE];
795                 int score = a - d;
796                 if (h < score && (b ^ c) < 0) {
797                     h = score;
798                     i_start = idx;
799                 }
800             }
801         }
802
803         s->last_i_start = i_start;
804     } else {
805         i_start = s->last_i_start;
806     }
807
808     bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
809     if (s->show_mode == SHOW_MODE_WAVES) {
810         fill_rectangle(screen,
811                        s->xleft, s->ytop, s->width, s->height,
812                        bgcolor);
813
814         fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
815
816         /* total height for one channel */
817         h = s->height / nb_display_channels;
818         /* graph height / 2 */
819         h2 = (h * 9) / 20;
820         for (ch = 0; ch < nb_display_channels; ch++) {
821             i = i_start + ch;
822             y1 = s->ytop + ch * h + (h / 2); /* position of center line */
823             for (x = 0; x < s->width; x++) {
824                 y = (s->sample_array[i] * h2) >> 15;
825                 if (y < 0) {
826                     y = -y;
827                     ys = y1 - y;
828                 } else {
829                     ys = y1;
830                 }
831                 fill_rectangle(screen,
832                                s->xleft + x, ys, 1, y,
833                                fgcolor);
834                 i += channels;
835                 if (i >= SAMPLE_ARRAY_SIZE)
836                     i -= SAMPLE_ARRAY_SIZE;
837             }
838         }
839
840         fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
841
842         for (ch = 1; ch < nb_display_channels; ch++) {
843             y = s->ytop + ch * h;
844             fill_rectangle(screen,
845                            s->xleft, y, s->width, 1,
846                            fgcolor);
847         }
848         SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
849     } else {
850         nb_display_channels= FFMIN(nb_display_channels, 2);
851         if (rdft_bits != s->rdft_bits) {
852             av_rdft_end(s->rdft);
853             av_free(s->rdft_data);
854             s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
855             s->rdft_bits = rdft_bits;
856             s->rdft_data = av_malloc(4 * nb_freq * sizeof(*s->rdft_data));
857         }
858         {
859             FFTSample *data[2];
860             for (ch = 0; ch < nb_display_channels; ch++) {
861                 data[ch] = s->rdft_data + 2 * nb_freq * ch;
862                 i = i_start + ch;
863                 for (x = 0; x < 2 * nb_freq; x++) {
864                     double w = (x-nb_freq) * (1.0 / nb_freq);
865                     data[ch][x] = s->sample_array[i] * (1.0 - w * w);
866                     i += channels;
867                     if (i >= SAMPLE_ARRAY_SIZE)
868                         i -= SAMPLE_ARRAY_SIZE;
869                 }
870                 av_rdft_calc(s->rdft, data[ch]);
871             }
872             // least efficient way to do this, we should of course directly access it but its more than fast enough
873             for (y = 0; y < s->height; y++) {
874                 double w = 1 / sqrt(nb_freq);
875                 int a = sqrt(w * sqrt(data[0][2 * y + 0] * data[0][2 * y + 0] + data[0][2 * y + 1] * data[0][2 * y + 1]));
876                 int b = (nb_display_channels == 2 ) ? sqrt(w * sqrt(data[1][2 * y + 0] * data[1][2 * y + 0]
877                        + data[1][2 * y + 1] * data[1][2 * y + 1])) : a;
878                 a = FFMIN(a, 255);
879                 b = FFMIN(b, 255);
880                 fgcolor = SDL_MapRGB(screen->format, a, b, (a + b) / 2);
881
882                 fill_rectangle(screen,
883                             s->xpos, s->height-y, 1, 1,
884                             fgcolor);
885             }
886         }
887         SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
888         if (!s->paused)
889             s->xpos++;
890         if (s->xpos >= s->width)
891             s->xpos= s->xleft;
892     }
893 }
894
895 static void stream_close(VideoState *is)
896 {
897     VideoPicture *vp;
898     int i;
899     /* XXX: use a special url_shutdown call to abort parse cleanly */
900     is->abort_request = 1;
901     SDL_WaitThread(is->read_tid, NULL);
902     SDL_WaitThread(is->refresh_tid, NULL);
903     packet_queue_destroy(&is->videoq);
904     packet_queue_destroy(&is->audioq);
905     packet_queue_destroy(&is->subtitleq);
906
907     /* free all pictures */
908     for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
909         vp = &is->pictq[i];
910 #if CONFIG_AVFILTER
911         avfilter_unref_bufferp(&vp->picref);
912 #endif
913         if (vp->bmp) {
914             SDL_FreeYUVOverlay(vp->bmp);
915             vp->bmp = NULL;
916         }
917     }
918     SDL_DestroyMutex(is->pictq_mutex);
919     SDL_DestroyCond(is->pictq_cond);
920     SDL_DestroyMutex(is->subpq_mutex);
921     SDL_DestroyCond(is->subpq_cond);
922 #if !CONFIG_AVFILTER
923     if (is->img_convert_ctx)
924         sws_freeContext(is->img_convert_ctx);
925 #endif
926     av_free(is);
927 }
928
929 static void do_exit(VideoState *is)
930 {
931     if (is) {
932         stream_close(is);
933     }
934     av_lockmgr_register(NULL);
935     uninit_opts();
936 #if CONFIG_AVFILTER
937     avfilter_uninit();
938 #endif
939     avformat_network_deinit();
940     if (show_status)
941         printf("\n");
942     SDL_Quit();
943     av_log(NULL, AV_LOG_QUIET, "%s", "");
944     exit(0);
945 }
946
947 static void sigterm_handler(int sig)
948 {
949     exit(123);
950 }
951
952 static int video_open(VideoState *is, int force_set_video_mode)
953 {
954     int flags = SDL_HWSURFACE | SDL_ASYNCBLIT | SDL_HWACCEL;
955     int w,h;
956     VideoPicture *vp = &is->pictq[is->pictq_rindex];
957
958     if (is_full_screen) flags |= SDL_FULLSCREEN;
959     else                flags |= SDL_RESIZABLE;
960
961     if (is_full_screen && fs_screen_width) {
962         w = fs_screen_width;
963         h = fs_screen_height;
964     } else if (!is_full_screen && screen_width) {
965         w = screen_width;
966         h = screen_height;
967     } else if (vp->width) {
968         w = vp->width;
969         h = vp->height;
970     } else {
971         w = 640;
972         h = 480;
973     }
974     if (screen && is->width == screen->w && screen->w == w
975        && is->height== screen->h && screen->h == h && !force_set_video_mode)
976         return 0;
977     screen = SDL_SetVideoMode(w, h, 0, flags);
978     if (!screen) {
979         fprintf(stderr, "SDL: could not set video mode - exiting\n");
980         do_exit(is);
981     }
982     if (!window_title)
983         window_title = input_filename;
984     SDL_WM_SetCaption(window_title, window_title);
985
986     is->width  = screen->w;
987     is->height = screen->h;
988
989     return 0;
990 }
991
992 /* display the current picture, if any */
993 static void video_display(VideoState *is)
994 {
995     if (!screen)
996         video_open(is, 0);
997     if (is->audio_st && is->show_mode != SHOW_MODE_VIDEO)
998         video_audio_display(is);
999     else if (is->video_st)
1000         video_image_display(is);
1001 }
1002
1003 static int refresh_thread(void *opaque)
1004 {
1005     VideoState *is= opaque;
1006     while (!is->abort_request) {
1007         SDL_Event event;
1008         event.type = FF_REFRESH_EVENT;
1009         event.user.data1 = opaque;
1010         if (!is->refresh && (!is->paused || is->force_refresh)) {
1011             is->refresh = 1;
1012             SDL_PushEvent(&event);
1013         }
1014         //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
1015         av_usleep(is->audio_st && is->show_mode != SHOW_MODE_VIDEO ? rdftspeed*1000 : 5000);
1016     }
1017     return 0;
1018 }
1019
1020 /* get the current audio clock value */
1021 static double get_audio_clock(VideoState *is)
1022 {
1023     if (is->paused) {
1024         return is->audio_current_pts;
1025     } else {
1026         return is->audio_current_pts_drift + av_gettime() / 1000000.0;
1027     }
1028 }
1029
1030 /* get the current video clock value */
1031 static double get_video_clock(VideoState *is)
1032 {
1033     if (is->paused) {
1034         return is->video_current_pts;
1035     } else {
1036         return is->video_current_pts_drift + av_gettime() / 1000000.0;
1037     }
1038 }
1039
1040 /* get the current external clock value */
1041 static double get_external_clock(VideoState *is)
1042 {
1043     int64_t ti;
1044     ti = av_gettime();
1045     return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1046 }
1047
1048 /* get the current master clock value */
1049 static double get_master_clock(VideoState *is)
1050 {
1051     double val;
1052
1053     if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1054         if (is->video_st)
1055             val = get_video_clock(is);
1056         else
1057             val = get_audio_clock(is);
1058     } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1059         if (is->audio_st)
1060             val = get_audio_clock(is);
1061         else
1062             val = get_video_clock(is);
1063     } else {
1064         val = get_external_clock(is);
1065     }
1066     return val;
1067 }
1068
1069 /* seek in the stream */
1070 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1071 {
1072     if (!is->seek_req) {
1073         is->seek_pos = pos;
1074         is->seek_rel = rel;
1075         is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1076         if (seek_by_bytes)
1077             is->seek_flags |= AVSEEK_FLAG_BYTE;
1078         is->seek_req = 1;
1079     }
1080 }
1081
1082 /* pause or resume the video */
1083 static void stream_toggle_pause(VideoState *is)
1084 {
1085     if (is->paused) {
1086         is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1087         if (is->read_pause_return != AVERROR(ENOSYS)) {
1088             is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1089         }
1090         is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1091     }
1092     is->paused = !is->paused;
1093 }
1094
1095 static double compute_target_delay(double delay, VideoState *is)
1096 {
1097     double sync_threshold, diff;
1098
1099     /* update delay to follow master synchronisation source */
1100     if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1101          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1102         /* if video is slave, we try to correct big delays by
1103            duplicating or deleting a frame */
1104         diff = get_video_clock(is) - get_master_clock(is);
1105
1106         /* skip or repeat frame. We take into account the
1107            delay to compute the threshold. I still don't know
1108            if it is the best guess */
1109         sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1110         if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1111             if (diff <= -sync_threshold)
1112                 delay = 0;
1113             else if (diff >= sync_threshold)
1114                 delay = 2 * delay;
1115         }
1116     }
1117
1118     av_dlog(NULL, "video: delay=%0.3f A-V=%f\n",
1119             delay, -diff);
1120
1121     return delay;
1122 }
1123
1124 static void pictq_next_picture(VideoState *is) {
1125     /* update queue size and signal for next picture */
1126     if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1127         is->pictq_rindex = 0;
1128
1129     SDL_LockMutex(is->pictq_mutex);
1130     is->pictq_size--;
1131     SDL_CondSignal(is->pictq_cond);
1132     SDL_UnlockMutex(is->pictq_mutex);
1133 }
1134
1135 static void update_video_pts(VideoState *is, double pts, int64_t pos) {
1136     double time = av_gettime() / 1000000.0;
1137     /* update current video pts */
1138     is->video_current_pts = pts;
1139     is->video_current_pts_drift = is->video_current_pts - time;
1140     is->video_current_pos = pos;
1141     is->frame_last_pts = pts;
1142 }
1143
1144 /* called to display each frame */
1145 static void video_refresh(void *opaque)
1146 {
1147     VideoState *is = opaque;
1148     VideoPicture *vp;
1149     double time;
1150
1151     SubPicture *sp, *sp2;
1152
1153     if (is->video_st) {
1154 retry:
1155         if (is->pictq_size == 0) {
1156             SDL_LockMutex(is->pictq_mutex);
1157             if (is->frame_last_dropped_pts != AV_NOPTS_VALUE && is->frame_last_dropped_pts > is->frame_last_pts) {
1158                 update_video_pts(is, is->frame_last_dropped_pts, is->frame_last_dropped_pos);
1159                 is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1160             }
1161             SDL_UnlockMutex(is->pictq_mutex);
1162             // nothing to do, no picture to display in the que
1163         } else {
1164             double last_duration, duration, delay;
1165             /* dequeue the picture */
1166             vp = &is->pictq[is->pictq_rindex];
1167
1168             if (vp->skip) {
1169                 pictq_next_picture(is);
1170                 goto retry;
1171             }
1172
1173             if (is->paused)
1174                 goto display;
1175
1176             /* compute nominal last_duration */
1177             last_duration = vp->pts - is->frame_last_pts;
1178             if (last_duration > 0 && last_duration < 10.0) {
1179                 /* if duration of the last frame was sane, update last_duration in video state */
1180                 is->frame_last_duration = last_duration;
1181             }
1182             delay = compute_target_delay(is->frame_last_duration, is);
1183
1184             time= av_gettime()/1000000.0;
1185             if (time < is->frame_timer + delay)
1186                 return;
1187
1188             if (delay > 0)
1189                 is->frame_timer += delay * FFMAX(1, floor((time-is->frame_timer) / delay));
1190
1191             SDL_LockMutex(is->pictq_mutex);
1192             update_video_pts(is, vp->pts, vp->pos);
1193             SDL_UnlockMutex(is->pictq_mutex);
1194
1195             if (is->pictq_size > 1) {
1196                 VideoPicture *nextvp = &is->pictq[(is->pictq_rindex + 1) % VIDEO_PICTURE_QUEUE_SIZE];
1197                 duration = nextvp->pts - vp->pts;
1198                 if((framedrop>0 || (framedrop && is->audio_st)) && time > is->frame_timer + duration){
1199                     is->frame_drops_late++;
1200                     pictq_next_picture(is);
1201                     goto retry;
1202                 }
1203             }
1204
1205             if (is->subtitle_st) {
1206                 if (is->subtitle_stream_changed) {
1207                     SDL_LockMutex(is->subpq_mutex);
1208
1209                     while (is->subpq_size) {
1210                         free_subpicture(&is->subpq[is->subpq_rindex]);
1211
1212                         /* update queue size and signal for next picture */
1213                         if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1214                             is->subpq_rindex = 0;
1215
1216                         is->subpq_size--;
1217                     }
1218                     is->subtitle_stream_changed = 0;
1219
1220                     SDL_CondSignal(is->subpq_cond);
1221                     SDL_UnlockMutex(is->subpq_mutex);
1222                 } else {
1223                     if (is->subpq_size > 0) {
1224                         sp = &is->subpq[is->subpq_rindex];
1225
1226                         if (is->subpq_size > 1)
1227                             sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1228                         else
1229                             sp2 = NULL;
1230
1231                         if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1232                                 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1233                         {
1234                             free_subpicture(sp);
1235
1236                             /* update queue size and signal for next picture */
1237                             if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1238                                 is->subpq_rindex = 0;
1239
1240                             SDL_LockMutex(is->subpq_mutex);
1241                             is->subpq_size--;
1242                             SDL_CondSignal(is->subpq_cond);
1243                             SDL_UnlockMutex(is->subpq_mutex);
1244                         }
1245                     }
1246                 }
1247             }
1248
1249 display:
1250             /* display picture */
1251             if (!display_disable)
1252                 video_display(is);
1253
1254             if (!is->paused)
1255                 pictq_next_picture(is);
1256         }
1257     } else if (is->audio_st) {
1258         /* draw the next audio frame */
1259
1260         /* if only audio stream, then display the audio bars (better
1261            than nothing, just to test the implementation */
1262
1263         /* display picture */
1264         if (!display_disable)
1265             video_display(is);
1266     }
1267     is->force_refresh = 0;
1268     if (show_status) {
1269         static int64_t last_time;
1270         int64_t cur_time;
1271         int aqsize, vqsize, sqsize;
1272         double av_diff;
1273
1274         cur_time = av_gettime();
1275         if (!last_time || (cur_time - last_time) >= 30000) {
1276             aqsize = 0;
1277             vqsize = 0;
1278             sqsize = 0;
1279             if (is->audio_st)
1280                 aqsize = is->audioq.size;
1281             if (is->video_st)
1282                 vqsize = is->videoq.size;
1283             if (is->subtitle_st)
1284                 sqsize = is->subtitleq.size;
1285             av_diff = 0;
1286             if (is->audio_st && is->video_st)
1287                 av_diff = get_audio_clock(is) - get_video_clock(is);
1288             printf("%7.2f A-V:%7.3f fd=%4d aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64"   \r",
1289                    get_master_clock(is),
1290                    av_diff,
1291                    is->frame_drops_early + is->frame_drops_late,
1292                    aqsize / 1024,
1293                    vqsize / 1024,
1294                    sqsize,
1295                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_dts : 0,
1296                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_pts : 0);
1297             fflush(stdout);
1298             last_time = cur_time;
1299         }
1300     }
1301 }
1302
1303 /* allocate a picture (needs to do that in main thread to avoid
1304    potential locking problems */
1305 static void alloc_picture(AllocEventProps *event_props)
1306 {
1307     VideoState *is = event_props->is;
1308     AVFrame *frame = event_props->frame;
1309     VideoPicture *vp;
1310
1311     vp = &is->pictq[is->pictq_windex];
1312
1313     if (vp->bmp)
1314         SDL_FreeYUVOverlay(vp->bmp);
1315
1316 #if CONFIG_AVFILTER
1317     avfilter_unref_bufferp(&vp->picref);
1318 #endif
1319
1320     vp->width   = frame->width;
1321     vp->height  = frame->height;
1322
1323     video_open(event_props->is, 0);
1324
1325     vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1326                                    SDL_YV12_OVERLAY,
1327                                    screen);
1328     if (!vp->bmp || vp->bmp->pitches[0] < vp->width) {
1329         /* SDL allocates a buffer smaller than requested if the video
1330          * overlay hardware is unable to support the requested size. */
1331         fprintf(stderr, "Error: the video system does not support an image\n"
1332                         "size of %dx%d pixels. Try using -lowres or -vf \"scale=w:h\"\n"
1333                         "to reduce the image size.\n", vp->width, vp->height );
1334         do_exit(is);
1335     }
1336
1337     SDL_LockMutex(is->pictq_mutex);
1338     vp->allocated = 1;
1339     SDL_CondSignal(is->pictq_cond);
1340     SDL_UnlockMutex(is->pictq_mutex);
1341 }
1342
1343 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1344 {
1345     VideoPicture *vp;
1346     double frame_delay, pts = pts1;
1347
1348     /* compute the exact PTS for the picture if it is omitted in the stream
1349      * pts1 is the dts of the pkt / pts of the frame */
1350     if (pts != 0) {
1351         /* update video clock with pts, if present */
1352         is->video_clock = pts;
1353     } else {
1354         pts = is->video_clock;
1355     }
1356     /* update video clock for next frame */
1357     frame_delay = av_q2d(is->video_st->codec->time_base);
1358     /* for MPEG2, the frame can be repeated, so we update the
1359        clock accordingly */
1360     frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1361     is->video_clock += frame_delay;
1362
1363 #if defined(DEBUG_SYNC) && 0
1364     printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1365            av_get_picture_type_char(src_frame->pict_type), pts, pts1);
1366 #endif
1367
1368     /* wait until we have space to put a new picture */
1369     SDL_LockMutex(is->pictq_mutex);
1370
1371     while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1372            !is->videoq.abort_request) {
1373         SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1374     }
1375     SDL_UnlockMutex(is->pictq_mutex);
1376
1377     if (is->videoq.abort_request)
1378         return -1;
1379
1380     vp = &is->pictq[is->pictq_windex];
1381
1382     /* alloc or resize hardware picture buffer */
1383     if (!vp->bmp || vp->reallocate ||
1384         vp->width  != src_frame->width ||
1385         vp->height != src_frame->height) {
1386         SDL_Event event;
1387         AllocEventProps event_props;
1388
1389         event_props.frame = src_frame;
1390         event_props.is = is;
1391
1392         vp->allocated  = 0;
1393         vp->reallocate = 0;
1394
1395         /* the allocation must be done in the main thread to avoid
1396            locking problems. We wait in this block for the event to complete,
1397            so we can pass a pointer to event_props to it. */
1398         event.type = FF_ALLOC_EVENT;
1399         event.user.data1 = &event_props;
1400         SDL_PushEvent(&event);
1401
1402         /* wait until the picture is allocated */
1403         SDL_LockMutex(is->pictq_mutex);
1404         while (!vp->allocated && !is->videoq.abort_request) {
1405             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1406         }
1407         /* if the queue is aborted, we have to pop the pending ALLOC event or wait for the allocation to complete */
1408         if (is->videoq.abort_request && SDL_PeepEvents(&event, 1, SDL_GETEVENT, SDL_EVENTMASK(FF_ALLOC_EVENT)) != 1) {
1409             while (!vp->allocated) {
1410                 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1411             }
1412         }
1413         SDL_UnlockMutex(is->pictq_mutex);
1414
1415         if (is->videoq.abort_request)
1416             return -1;
1417     }
1418
1419     /* if the frame is not skipped, then display it */
1420     if (vp->bmp) {
1421         AVPicture pict = { { 0 } };
1422 #if CONFIG_AVFILTER
1423         avfilter_unref_bufferp(&vp->picref);
1424         vp->picref = src_frame->opaque;
1425 #endif
1426
1427         /* get a pointer on the bitmap */
1428         SDL_LockYUVOverlay (vp->bmp);
1429
1430         pict.data[0] = vp->bmp->pixels[0];
1431         pict.data[1] = vp->bmp->pixels[2];
1432         pict.data[2] = vp->bmp->pixels[1];
1433
1434         pict.linesize[0] = vp->bmp->pitches[0];
1435         pict.linesize[1] = vp->bmp->pitches[2];
1436         pict.linesize[2] = vp->bmp->pitches[1];
1437
1438 #if CONFIG_AVFILTER
1439         // FIXME use direct rendering
1440         av_picture_copy(&pict, (AVPicture *)src_frame,
1441                         src_frame->format, vp->width, vp->height);
1442         vp->sample_aspect_ratio = vp->picref->video->sample_aspect_ratio;
1443 #else
1444         sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1445         is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1446             vp->width, vp->height, src_frame->format, vp->width, vp->height,
1447             PIX_FMT_YUV420P, sws_flags, NULL, NULL, NULL);
1448         if (is->img_convert_ctx == NULL) {
1449             fprintf(stderr, "Cannot initialize the conversion context\n");
1450             exit(1);
1451         }
1452         sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1453                   0, vp->height, pict.data, pict.linesize);
1454         vp->sample_aspect_ratio = av_guess_sample_aspect_ratio(is->ic, is->video_st, src_frame);
1455 #endif
1456         /* update the bitmap content */
1457         SDL_UnlockYUVOverlay(vp->bmp);
1458
1459         vp->pts = pts;
1460         vp->pos = pos;
1461         vp->skip = 0;
1462
1463         /* now we can update the picture count */
1464         if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1465             is->pictq_windex = 0;
1466         SDL_LockMutex(is->pictq_mutex);
1467         is->pictq_size++;
1468         SDL_UnlockMutex(is->pictq_mutex);
1469     }
1470     return 0;
1471 }
1472
1473 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1474 {
1475     int got_picture, i;
1476
1477     if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1478         return -1;
1479
1480     if (pkt->data == flush_pkt.data) {
1481         avcodec_flush_buffers(is->video_st->codec);
1482
1483         SDL_LockMutex(is->pictq_mutex);
1484         // Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1485         for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
1486             is->pictq[i].skip = 1;
1487         }
1488         while (is->pictq_size && !is->videoq.abort_request) {
1489             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1490         }
1491         is->video_current_pos = -1;
1492         is->frame_last_pts = AV_NOPTS_VALUE;
1493         is->frame_last_duration = 0;
1494         is->frame_timer = (double)av_gettime() / 1000000.0;
1495         is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1496         SDL_UnlockMutex(is->pictq_mutex);
1497
1498         return 0;
1499     }
1500
1501     if(avcodec_decode_video2(is->video_st->codec, frame, &got_picture, pkt) < 0)
1502         return 0;
1503
1504     if (got_picture) {
1505         int ret = 1;
1506
1507         if (decoder_reorder_pts == -1) {
1508             *pts = av_frame_get_best_effort_timestamp(frame);
1509         } else if (decoder_reorder_pts) {
1510             *pts = frame->pkt_pts;
1511         } else {
1512             *pts = frame->pkt_dts;
1513         }
1514
1515         if (*pts == AV_NOPTS_VALUE) {
1516             *pts = 0;
1517         }
1518
1519         if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) || is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK) &&
1520              (framedrop>0 || (framedrop && is->audio_st))) {
1521             SDL_LockMutex(is->pictq_mutex);
1522             if (is->frame_last_pts != AV_NOPTS_VALUE && *pts) {
1523                 double clockdiff = get_video_clock(is) - get_master_clock(is);
1524                 double dpts = av_q2d(is->video_st->time_base) * *pts;
1525                 double ptsdiff = dpts - is->frame_last_pts;
1526                 if (fabs(clockdiff) < AV_NOSYNC_THRESHOLD &&
1527                      ptsdiff > 0 && ptsdiff < AV_NOSYNC_THRESHOLD &&
1528                      clockdiff + ptsdiff - is->frame_last_filter_delay < 0) {
1529                     is->frame_last_dropped_pos = pkt->pos;
1530                     is->frame_last_dropped_pts = dpts;
1531                     is->frame_drops_early++;
1532                     ret = 0;
1533                 }
1534             }
1535             SDL_UnlockMutex(is->pictq_mutex);
1536         }
1537
1538         return ret;
1539     }
1540     return 0;
1541 }
1542
1543 #if CONFIG_AVFILTER
1544 static int configure_filtergraph(AVFilterGraph *graph, const char *filtergraph,
1545                                  AVFilterContext *source_ctx, AVFilterContext *sink_ctx)
1546 {
1547     int ret;
1548     AVFilterInOut *outputs = NULL, *inputs = NULL;
1549
1550     if (filtergraph) {
1551         outputs = avfilter_inout_alloc();
1552         inputs  = avfilter_inout_alloc();
1553         if (!outputs || !inputs) {
1554             ret = AVERROR(ENOMEM);
1555             goto fail;
1556         }
1557
1558         outputs->name       = av_strdup("in");
1559         outputs->filter_ctx = source_ctx;
1560         outputs->pad_idx    = 0;
1561         outputs->next       = NULL;
1562
1563         inputs->name        = av_strdup("out");
1564         inputs->filter_ctx  = sink_ctx;
1565         inputs->pad_idx     = 0;
1566         inputs->next        = NULL;
1567
1568         if ((ret = avfilter_graph_parse(graph, filtergraph, &inputs, &outputs, NULL)) < 0)
1569             goto fail;
1570     } else {
1571         if ((ret = avfilter_link(source_ctx, 0, sink_ctx, 0)) < 0)
1572             goto fail;
1573     }
1574
1575     return avfilter_graph_config(graph, NULL);
1576 fail:
1577     avfilter_inout_free(&outputs);
1578     avfilter_inout_free(&inputs);
1579     return ret;
1580 }
1581
1582 static int configure_video_filters(AVFilterGraph *graph, VideoState *is, const char *vfilters)
1583 {
1584     static const enum PixelFormat pix_fmts[] = { PIX_FMT_YUV420P, PIX_FMT_NONE };
1585     char sws_flags_str[128];
1586     char buffersrc_args[256];
1587     int ret;
1588     AVBufferSinkParams *buffersink_params = av_buffersink_params_alloc();
1589     AVFilterContext *filt_src = NULL, *filt_out = NULL, *filt_format;
1590     AVCodecContext *codec = is->video_st->codec;
1591
1592     snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1593     graph->scale_sws_opts = av_strdup(sws_flags_str);
1594
1595     snprintf(buffersrc_args, sizeof(buffersrc_args),
1596              "video_size=%dx%d:pix_fmt=%d:time_base=%d/%d:pixel_aspect=%d/%d",
1597              codec->width, codec->height, codec->pix_fmt,
1598              is->video_st->time_base.num, is->video_st->time_base.den,
1599              codec->sample_aspect_ratio.num, codec->sample_aspect_ratio.den);
1600
1601     if ((ret = avfilter_graph_create_filter(&filt_src,
1602                                             avfilter_get_by_name("buffer"),
1603                                             "ffplay_buffer", buffersrc_args, NULL,
1604                                             graph)) < 0)
1605         return ret;
1606
1607     buffersink_params->pixel_fmts = pix_fmts;
1608     ret = avfilter_graph_create_filter(&filt_out,
1609                                        avfilter_get_by_name("buffersink"),
1610                                        "ffplay_buffersink", NULL, buffersink_params, graph);
1611     av_freep(&buffersink_params);
1612     if (ret < 0)
1613         return ret;
1614
1615     if ((ret = avfilter_graph_create_filter(&filt_format,
1616                                             avfilter_get_by_name("format"),
1617                                             "format", "yuv420p", NULL, graph)) < 0)
1618         return ret;
1619     if ((ret = avfilter_link(filt_format, 0, filt_out, 0)) < 0)
1620         return ret;
1621
1622     if ((ret = configure_filtergraph(graph, vfilters, filt_src, filt_format)) < 0)
1623         return ret;
1624
1625     is->in_video_filter  = filt_src;
1626     is->out_video_filter = filt_out;
1627
1628     return ret;
1629 }
1630
1631 #endif  /* CONFIG_AVFILTER */
1632
1633 static int video_thread(void *arg)
1634 {
1635     AVPacket pkt = { 0 };
1636     VideoState *is = arg;
1637     AVFrame *frame = avcodec_alloc_frame();
1638     int64_t pts_int = AV_NOPTS_VALUE, pos = -1;
1639     double pts;
1640     int ret;
1641
1642 #if CONFIG_AVFILTER
1643     AVCodecContext *codec = is->video_st->codec;
1644     AVFilterGraph *graph = avfilter_graph_alloc();
1645     AVFilterContext *filt_out = NULL, *filt_in = NULL;
1646     int last_w = 0;
1647     int last_h = 0;
1648     enum PixelFormat last_format = -2;
1649
1650     if (codec->codec->capabilities & CODEC_CAP_DR1) {
1651         is->use_dr1 = 1;
1652         codec->get_buffer     = codec_get_buffer;
1653         codec->release_buffer = codec_release_buffer;
1654         codec->opaque         = &is->buffer_pool;
1655     }
1656 #endif
1657
1658     for (;;) {
1659 #if CONFIG_AVFILTER
1660         AVFilterBufferRef *picref;
1661         AVRational tb;
1662 #endif
1663         while (is->paused && !is->videoq.abort_request)
1664             SDL_Delay(10);
1665
1666         avcodec_get_frame_defaults(frame);
1667         av_free_packet(&pkt);
1668
1669         ret = get_video_frame(is, frame, &pts_int, &pkt);
1670         if (ret < 0)
1671             goto the_end;
1672
1673         if (!ret)
1674             continue;
1675
1676 #if CONFIG_AVFILTER
1677         if (   last_w != is->video_st->codec->width
1678             || last_h != is->video_st->codec->height
1679             || last_format != is->video_st->codec->pix_fmt) {
1680             av_log(NULL, AV_LOG_INFO, "Frame changed from size:%dx%d to size:%dx%d\n",
1681                    last_w, last_h, is->video_st->codec->width, is->video_st->codec->height);
1682             avfilter_graph_free(&graph);
1683             graph = avfilter_graph_alloc();
1684             if ((ret = configure_video_filters(graph, is, vfilters)) < 0) {
1685                 SDL_Event event;
1686                 event.type = FF_QUIT_EVENT;
1687                 event.user.data1 = is;
1688                 SDL_PushEvent(&event);
1689                 av_free_packet(&pkt);
1690                 goto the_end;
1691             }
1692             filt_in  = is->in_video_filter;
1693             filt_out = is->out_video_filter;
1694             last_w = is->video_st->codec->width;
1695             last_h = is->video_st->codec->height;
1696             last_format = is->video_st->codec->pix_fmt;
1697         }
1698
1699         frame->pts = pts_int;
1700         frame->sample_aspect_ratio = av_guess_sample_aspect_ratio(is->ic, is->video_st, frame);
1701         if (is->use_dr1 && frame->opaque) {
1702             FrameBuffer      *buf = frame->opaque;
1703             AVFilterBufferRef *fb = avfilter_get_video_buffer_ref_from_arrays(
1704                                         frame->data, frame->linesize,
1705                                         AV_PERM_READ | AV_PERM_PRESERVE,
1706                                         frame->width, frame->height,
1707                                         frame->format);
1708
1709             avfilter_copy_frame_props(fb, frame);
1710             fb->buf->priv           = buf;
1711             fb->buf->free           = filter_release_buffer;
1712
1713             buf->refcount++;
1714             av_buffersrc_add_ref(filt_in, fb, AV_BUFFERSRC_FLAG_NO_COPY);
1715
1716         } else
1717             av_buffersrc_write_frame(filt_in, frame);
1718
1719         av_free_packet(&pkt);
1720
1721         while (ret >= 0) {
1722             is->frame_last_returned_time = av_gettime() / 1000000.0;
1723
1724             ret = av_buffersink_get_buffer_ref(filt_out, &picref, 0);
1725             if (ret < 0) {
1726                 ret = 0;
1727                 break;
1728             }
1729
1730             is->frame_last_filter_delay = av_gettime() / 1000000.0 - is->frame_last_returned_time;
1731             if (fabs(is->frame_last_filter_delay) > AV_NOSYNC_THRESHOLD / 10.0)
1732                 is->frame_last_filter_delay = 0;
1733
1734             avfilter_copy_buf_props(frame, picref);
1735
1736             pts_int = picref->pts;
1737             tb      = filt_out->inputs[0]->time_base;
1738             pos     = picref->pos;
1739             frame->opaque = picref;
1740
1741             if (av_cmp_q(tb, is->video_st->time_base)) {
1742                 av_unused int64_t pts1 = pts_int;
1743                 pts_int = av_rescale_q(pts_int, tb, is->video_st->time_base);
1744                 av_dlog(NULL, "video_thread(): "
1745                         "tb:%d/%d pts:%"PRId64" -> tb:%d/%d pts:%"PRId64"\n",
1746                         tb.num, tb.den, pts1,
1747                         is->video_st->time_base.num, is->video_st->time_base.den, pts_int);
1748             }
1749             pts = pts_int * av_q2d(is->video_st->time_base);
1750             ret = queue_picture(is, frame, pts, pos);
1751         }
1752 #else
1753         pts = pts_int * av_q2d(is->video_st->time_base);
1754         ret = queue_picture(is, frame, pts, pkt.pos);
1755 #endif
1756
1757         if (ret < 0)
1758             goto the_end;
1759
1760         if (is->step)
1761             stream_toggle_pause(is);
1762     }
1763  the_end:
1764     avcodec_flush_buffers(is->video_st->codec);
1765 #if CONFIG_AVFILTER
1766     av_freep(&vfilters);
1767     avfilter_graph_free(&graph);
1768 #endif
1769     av_free_packet(&pkt);
1770     av_free(frame);
1771     return 0;
1772 }
1773
1774 static int subtitle_thread(void *arg)
1775 {
1776     VideoState *is = arg;
1777     SubPicture *sp;
1778     AVPacket pkt1, *pkt = &pkt1;
1779     int got_subtitle;
1780     double pts;
1781     int i, j;
1782     int r, g, b, y, u, v, a;
1783
1784     for (;;) {
1785         while (is->paused && !is->subtitleq.abort_request) {
1786             SDL_Delay(10);
1787         }
1788         if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1789             break;
1790
1791         if (pkt->data == flush_pkt.data) {
1792             avcodec_flush_buffers(is->subtitle_st->codec);
1793             continue;
1794         }
1795         SDL_LockMutex(is->subpq_mutex);
1796         while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1797                !is->subtitleq.abort_request) {
1798             SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1799         }
1800         SDL_UnlockMutex(is->subpq_mutex);
1801
1802         if (is->subtitleq.abort_request)
1803             return 0;
1804
1805         sp = &is->subpq[is->subpq_windex];
1806
1807        /* NOTE: ipts is the PTS of the _first_ picture beginning in
1808            this packet, if any */
1809         pts = 0;
1810         if (pkt->pts != AV_NOPTS_VALUE)
1811             pts = av_q2d(is->subtitle_st->time_base) * pkt->pts;
1812
1813         avcodec_decode_subtitle2(is->subtitle_st->codec, &sp->sub,
1814                                  &got_subtitle, pkt);
1815
1816         if (got_subtitle && sp->sub.format == 0) {
1817             sp->pts = pts;
1818
1819             for (i = 0; i < sp->sub.num_rects; i++)
1820             {
1821                 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1822                 {
1823                     RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1824                     y = RGB_TO_Y_CCIR(r, g, b);
1825                     u = RGB_TO_U_CCIR(r, g, b, 0);
1826                     v = RGB_TO_V_CCIR(r, g, b, 0);
1827                     YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1828                 }
1829             }
1830
1831             /* now we can update the picture count */
1832             if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1833                 is->subpq_windex = 0;
1834             SDL_LockMutex(is->subpq_mutex);
1835             is->subpq_size++;
1836             SDL_UnlockMutex(is->subpq_mutex);
1837         }
1838         av_free_packet(pkt);
1839     }
1840     return 0;
1841 }
1842
1843 /* copy samples for viewing in editor window */
1844 static void update_sample_display(VideoState *is, short *samples, int samples_size)
1845 {
1846     int size, len;
1847
1848     size = samples_size / sizeof(short);
1849     while (size > 0) {
1850         len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
1851         if (len > size)
1852             len = size;
1853         memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
1854         samples += len;
1855         is->sample_array_index += len;
1856         if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
1857             is->sample_array_index = 0;
1858         size -= len;
1859     }
1860 }
1861
1862 /* return the wanted number of samples to get better sync if sync_type is video
1863  * or external master clock */
1864 static int synchronize_audio(VideoState *is, int nb_samples)
1865 {
1866     int wanted_nb_samples = nb_samples;
1867
1868     /* if not master, then we try to remove or add samples to correct the clock */
1869     if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
1870          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1871         double diff, avg_diff;
1872         int min_nb_samples, max_nb_samples;
1873
1874         diff = get_audio_clock(is) - get_master_clock(is);
1875
1876         if (diff < AV_NOSYNC_THRESHOLD) {
1877             is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
1878             if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
1879                 /* not enough measures to have a correct estimate */
1880                 is->audio_diff_avg_count++;
1881             } else {
1882                 /* estimate the A-V difference */
1883                 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
1884
1885                 if (fabs(avg_diff) >= is->audio_diff_threshold) {
1886                     wanted_nb_samples = nb_samples + (int)(diff * is->audio_src.freq);
1887                     min_nb_samples = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX) / 100));
1888                     max_nb_samples = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX) / 100));
1889                     wanted_nb_samples = FFMIN(FFMAX(wanted_nb_samples, min_nb_samples), max_nb_samples);
1890                 }
1891                 av_dlog(NULL, "diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
1892                         diff, avg_diff, wanted_nb_samples - nb_samples,
1893                         is->audio_clock, is->video_clock, is->audio_diff_threshold);
1894             }
1895         } else {
1896             /* too big difference : may be initial PTS errors, so
1897                reset A-V filter */
1898             is->audio_diff_avg_count = 0;
1899             is->audio_diff_cum       = 0;
1900         }
1901     }
1902
1903     return wanted_nb_samples;
1904 }
1905
1906 /* decode one audio frame and returns its uncompressed size */
1907 static int audio_decode_frame(VideoState *is, double *pts_ptr)
1908 {
1909     AVPacket *pkt_temp = &is->audio_pkt_temp;
1910     AVPacket *pkt = &is->audio_pkt;
1911     AVCodecContext *dec = is->audio_st->codec;
1912     int len1, len2, data_size, resampled_data_size;
1913     int64_t dec_channel_layout;
1914     int got_frame;
1915     double pts;
1916     int new_packet = 0;
1917     int flush_complete = 0;
1918     int wanted_nb_samples;
1919
1920     for (;;) {
1921         /* NOTE: the audio packet can contain several frames */
1922         while (pkt_temp->size > 0 || (!pkt_temp->data && new_packet)) {
1923             if (!is->frame) {
1924                 if (!(is->frame = avcodec_alloc_frame()))
1925                     return AVERROR(ENOMEM);
1926             } else
1927                 avcodec_get_frame_defaults(is->frame);
1928
1929             if (is->paused)
1930                 return -1;
1931
1932             if (flush_complete)
1933                 break;
1934             new_packet = 0;
1935             len1 = avcodec_decode_audio4(dec, is->frame, &got_frame, pkt_temp);
1936             if (len1 < 0) {
1937                 /* if error, we skip the frame */
1938                 pkt_temp->size = 0;
1939                 break;
1940             }
1941
1942             pkt_temp->data += len1;
1943             pkt_temp->size -= len1;
1944
1945             if (!got_frame) {
1946                 /* stop sending empty packets if the decoder is finished */
1947                 if (!pkt_temp->data && dec->codec->capabilities & CODEC_CAP_DELAY)
1948                     flush_complete = 1;
1949                 continue;
1950             }
1951             data_size = av_samples_get_buffer_size(NULL, dec->channels,
1952                                                    is->frame->nb_samples,
1953                                                    dec->sample_fmt, 1);
1954
1955             dec_channel_layout =
1956                 (dec->channel_layout && dec->channels == av_get_channel_layout_nb_channels(dec->channel_layout)) ?
1957                 dec->channel_layout : av_get_default_channel_layout(dec->channels);
1958             wanted_nb_samples = synchronize_audio(is, is->frame->nb_samples);
1959
1960             if (dec->sample_fmt    != is->audio_src.fmt            ||
1961                 dec_channel_layout != is->audio_src.channel_layout ||
1962                 dec->sample_rate   != is->audio_src.freq           ||
1963                 (wanted_nb_samples != is->frame->nb_samples && !is->swr_ctx)) {
1964                 swr_free(&is->swr_ctx);
1965                 is->swr_ctx = swr_alloc_set_opts(NULL,
1966                                                  is->audio_tgt.channel_layout, is->audio_tgt.fmt, is->audio_tgt.freq,
1967                                                  dec_channel_layout,           dec->sample_fmt,   dec->sample_rate,
1968                                                  0, NULL);
1969                 if (!is->swr_ctx || swr_init(is->swr_ctx) < 0) {
1970                     fprintf(stderr, "Cannot create sample rate converter for conversion of %d Hz %s %d channels to %d Hz %s %d channels!\n",
1971                         dec->sample_rate,   av_get_sample_fmt_name(dec->sample_fmt),   dec->channels,
1972                         is->audio_tgt.freq, av_get_sample_fmt_name(is->audio_tgt.fmt), is->audio_tgt.channels);
1973                     break;
1974                 }
1975                 is->audio_src.channel_layout = dec_channel_layout;
1976                 is->audio_src.channels = dec->channels;
1977                 is->audio_src.freq = dec->sample_rate;
1978                 is->audio_src.fmt = dec->sample_fmt;
1979             }
1980
1981             if (is->swr_ctx) {
1982                 const uint8_t **in = (const uint8_t **)is->frame->extended_data;
1983                 uint8_t *out[] = {is->audio_buf2};
1984                 int out_count = sizeof(is->audio_buf2) / is->audio_tgt.channels / av_get_bytes_per_sample(is->audio_tgt.fmt);
1985                 if (wanted_nb_samples != is->frame->nb_samples) {
1986                     if (swr_set_compensation(is->swr_ctx, (wanted_nb_samples - is->frame->nb_samples) * is->audio_tgt.freq / dec->sample_rate,
1987                                                 wanted_nb_samples * is->audio_tgt.freq / dec->sample_rate) < 0) {
1988                         fprintf(stderr, "swr_set_compensation() failed\n");
1989                         break;
1990                     }
1991                 }
1992                 len2 = swr_convert(is->swr_ctx, out, out_count, in, is->frame->nb_samples);
1993                 if (len2 < 0) {
1994                     fprintf(stderr, "swr_convert() failed\n");
1995                     break;
1996                 }
1997                 if (len2 == out_count) {
1998                     fprintf(stderr, "warning: audio buffer is probably too small\n");
1999                     swr_init(is->swr_ctx);
2000                 }
2001                 is->audio_buf = is->audio_buf2;
2002                 resampled_data_size = len2 * is->audio_tgt.channels * av_get_bytes_per_sample(is->audio_tgt.fmt);
2003             } else {
2004                 is->audio_buf = is->frame->data[0];
2005                 resampled_data_size = data_size;
2006             }
2007
2008             /* if no pts, then compute it */
2009             pts = is->audio_clock;
2010             *pts_ptr = pts;
2011             is->audio_clock += (double)data_size /
2012                 (dec->channels * dec->sample_rate * av_get_bytes_per_sample(dec->sample_fmt));
2013 #ifdef DEBUG
2014             {
2015                 static double last_clock;
2016                 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2017                        is->audio_clock - last_clock,
2018                        is->audio_clock, pts);
2019                 last_clock = is->audio_clock;
2020             }
2021 #endif
2022             return resampled_data_size;
2023         }
2024
2025         /* free the current packet */
2026         if (pkt->data)
2027             av_free_packet(pkt);
2028         memset(pkt_temp, 0, sizeof(*pkt_temp));
2029
2030         if (is->paused || is->audioq.abort_request) {
2031             return -1;
2032         }
2033
2034         /* read next packet */
2035         if ((new_packet = packet_queue_get(&is->audioq, pkt, 1)) < 0)
2036             return -1;
2037
2038         if (pkt->data == flush_pkt.data) {
2039             avcodec_flush_buffers(dec);
2040             flush_complete = 0;
2041         }
2042
2043         *pkt_temp = *pkt;
2044
2045         /* if update the audio clock with the pts */
2046         if (pkt->pts != AV_NOPTS_VALUE) {
2047             is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2048         }
2049     }
2050 }
2051
2052 /* prepare a new audio buffer */
2053 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2054 {
2055     VideoState *is = opaque;
2056     int audio_size, len1;
2057     int bytes_per_sec;
2058     int frame_size = av_samples_get_buffer_size(NULL, is->audio_tgt.channels, 1, is->audio_tgt.fmt, 1);
2059     double pts;
2060
2061     audio_callback_time = av_gettime();
2062
2063     while (len > 0) {
2064         if (is->audio_buf_index >= is->audio_buf_size) {
2065            audio_size = audio_decode_frame(is, &pts);
2066            if (audio_size < 0) {
2067                 /* if error, just output silence */
2068                is->audio_buf      = is->silence_buf;
2069                is->audio_buf_size = sizeof(is->silence_buf) / frame_size * frame_size;
2070            } else {
2071                if (is->show_mode != SHOW_MODE_VIDEO)
2072                    update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2073                is->audio_buf_size = audio_size;
2074            }
2075            is->audio_buf_index = 0;
2076         }
2077         len1 = is->audio_buf_size - is->audio_buf_index;
2078         if (len1 > len)
2079             len1 = len;
2080         memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2081         len -= len1;
2082         stream += len1;
2083         is->audio_buf_index += len1;
2084     }
2085     bytes_per_sec = is->audio_tgt.freq * is->audio_tgt.channels * av_get_bytes_per_sample(is->audio_tgt.fmt);
2086     is->audio_write_buf_size = is->audio_buf_size - is->audio_buf_index;
2087     /* Let's assume the audio driver that is used by SDL has two periods. */
2088     is->audio_current_pts = is->audio_clock - (double)(2 * is->audio_hw_buf_size + is->audio_write_buf_size) / bytes_per_sec;
2089     is->audio_current_pts_drift = is->audio_current_pts - audio_callback_time / 1000000.0;
2090 }
2091
2092 static int audio_open(void *opaque, int64_t wanted_channel_layout, int wanted_nb_channels, int wanted_sample_rate, struct AudioParams *audio_hw_params)
2093 {
2094     SDL_AudioSpec wanted_spec, spec;
2095     const char *env;
2096     const int next_nb_channels[] = {0, 0, 1, 6, 2, 6, 4, 6};
2097
2098     env = SDL_getenv("SDL_AUDIO_CHANNELS");
2099     if (env) {
2100         wanted_nb_channels = atoi(env);
2101         wanted_channel_layout = av_get_default_channel_layout(wanted_nb_channels);
2102     }
2103     if (!wanted_channel_layout || wanted_nb_channels != av_get_channel_layout_nb_channels(wanted_channel_layout)) {
2104         wanted_channel_layout = av_get_default_channel_layout(wanted_nb_channels);
2105         wanted_channel_layout &= ~AV_CH_LAYOUT_STEREO_DOWNMIX;
2106     }
2107     wanted_spec.channels = av_get_channel_layout_nb_channels(wanted_channel_layout);
2108     wanted_spec.freq = wanted_sample_rate;
2109     if (wanted_spec.freq <= 0 || wanted_spec.channels <= 0) {
2110         fprintf(stderr, "Invalid sample rate or channel count!\n");
2111         return -1;
2112     }
2113     wanted_spec.format = AUDIO_S16SYS;
2114     wanted_spec.silence = 0;
2115     wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2116     wanted_spec.callback = sdl_audio_callback;
2117     wanted_spec.userdata = opaque;
2118     while (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2119         fprintf(stderr, "SDL_OpenAudio (%d channels): %s\n", wanted_spec.channels, SDL_GetError());
2120         wanted_spec.channels = next_nb_channels[FFMIN(7, wanted_spec.channels)];
2121         if (!wanted_spec.channels) {
2122             fprintf(stderr, "No more channel combinations to try, audio open failed\n");
2123             return -1;
2124         }
2125         wanted_channel_layout = av_get_default_channel_layout(wanted_spec.channels);
2126     }
2127     if (spec.format != AUDIO_S16SYS) {
2128         fprintf(stderr, "SDL advised audio format %d is not supported!\n", spec.format);
2129         return -1;
2130     }
2131     if (spec.channels != wanted_spec.channels) {
2132         wanted_channel_layout = av_get_default_channel_layout(spec.channels);
2133         if (!wanted_channel_layout) {
2134             fprintf(stderr, "SDL advised channel count %d is not supported!\n", spec.channels);
2135             return -1;
2136         }
2137     }
2138
2139     audio_hw_params->fmt = AV_SAMPLE_FMT_S16;
2140     audio_hw_params->freq = spec.freq;
2141     audio_hw_params->channel_layout = wanted_channel_layout;
2142     audio_hw_params->channels =  spec.channels;
2143     return spec.size;
2144 }
2145
2146 /* open a given stream. Return 0 if OK */
2147 static int stream_component_open(VideoState *is, int stream_index)
2148 {
2149     AVFormatContext *ic = is->ic;
2150     AVCodecContext *avctx;
2151     AVCodec *codec;
2152     AVDictionary *opts;
2153     AVDictionaryEntry *t = NULL;
2154
2155     if (stream_index < 0 || stream_index >= ic->nb_streams)
2156         return -1;
2157     avctx = ic->streams[stream_index]->codec;
2158
2159     codec = avcodec_find_decoder(avctx->codec_id);
2160     opts = filter_codec_opts(codec_opts, avctx->codec_id, ic, ic->streams[stream_index], codec);
2161
2162     switch(avctx->codec_type){
2163         case AVMEDIA_TYPE_AUDIO   : is->last_audio_stream    = stream_index; if(audio_codec_name   ) codec= avcodec_find_decoder_by_name(   audio_codec_name); break;
2164         case AVMEDIA_TYPE_SUBTITLE: is->last_subtitle_stream = stream_index; if(subtitle_codec_name) codec= avcodec_find_decoder_by_name(subtitle_codec_name); break;
2165         case AVMEDIA_TYPE_VIDEO   : is->last_video_stream    = stream_index; if(video_codec_name   ) codec= avcodec_find_decoder_by_name(   video_codec_name); break;
2166     }
2167     if (!codec)
2168         return -1;
2169
2170     avctx->workaround_bugs   = workaround_bugs;
2171     avctx->lowres            = lowres;
2172     if(avctx->lowres > codec->max_lowres){
2173         av_log(avctx, AV_LOG_WARNING, "The maximum value for lowres supported by the decoder is %d\n",
2174                 codec->max_lowres);
2175         avctx->lowres= codec->max_lowres;
2176     }
2177     avctx->idct_algo         = idct;
2178     avctx->skip_frame        = skip_frame;
2179     avctx->skip_idct         = skip_idct;
2180     avctx->skip_loop_filter  = skip_loop_filter;
2181     avctx->error_concealment = error_concealment;
2182
2183     if(avctx->lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2184     if (fast)   avctx->flags2 |= CODEC_FLAG2_FAST;
2185     if(codec->capabilities & CODEC_CAP_DR1)
2186         avctx->flags |= CODEC_FLAG_EMU_EDGE;
2187
2188     if (!av_dict_get(opts, "threads", NULL, 0))
2189         av_dict_set(&opts, "threads", "auto", 0);
2190     if (!codec ||
2191         avcodec_open2(avctx, codec, &opts) < 0)
2192         return -1;
2193     if ((t = av_dict_get(opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2194         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2195         return AVERROR_OPTION_NOT_FOUND;
2196     }
2197
2198     /* prepare audio output */
2199     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2200         int audio_hw_buf_size = audio_open(is, avctx->channel_layout, avctx->channels, avctx->sample_rate, &is->audio_src);
2201         if (audio_hw_buf_size < 0)
2202             return -1;
2203         is->audio_hw_buf_size = audio_hw_buf_size;
2204         is->audio_tgt = is->audio_src;
2205     }
2206
2207     ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2208     switch (avctx->codec_type) {
2209     case AVMEDIA_TYPE_AUDIO:
2210         is->audio_stream = stream_index;
2211         is->audio_st = ic->streams[stream_index];
2212         is->audio_buf_size  = 0;
2213         is->audio_buf_index = 0;
2214
2215         /* init averaging filter */
2216         is->audio_diff_avg_coef  = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2217         is->audio_diff_avg_count = 0;
2218         /* since we do not have a precise anough audio fifo fullness,
2219            we correct audio sync only if larger than this threshold */
2220         is->audio_diff_threshold = 2.0 * is->audio_hw_buf_size / av_samples_get_buffer_size(NULL, is->audio_tgt.channels, is->audio_tgt.freq, is->audio_tgt.fmt, 1);
2221
2222         memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2223         memset(&is->audio_pkt_temp, 0, sizeof(is->audio_pkt_temp));
2224         packet_queue_start(&is->audioq);
2225         SDL_PauseAudio(0);
2226         break;
2227     case AVMEDIA_TYPE_VIDEO:
2228         is->video_stream = stream_index;
2229         is->video_st = ic->streams[stream_index];
2230
2231         packet_queue_start(&is->videoq);
2232         is->video_tid = SDL_CreateThread(video_thread, is);
2233         break;
2234     case AVMEDIA_TYPE_SUBTITLE:
2235         is->subtitle_stream = stream_index;
2236         is->subtitle_st = ic->streams[stream_index];
2237         packet_queue_start(&is->subtitleq);
2238
2239         is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2240         break;
2241     default:
2242         break;
2243     }
2244     return 0;
2245 }
2246
2247 static void stream_component_close(VideoState *is, int stream_index)
2248 {
2249     AVFormatContext *ic = is->ic;
2250     AVCodecContext *avctx;
2251
2252     if (stream_index < 0 || stream_index >= ic->nb_streams)
2253         return;
2254     avctx = ic->streams[stream_index]->codec;
2255
2256     switch (avctx->codec_type) {
2257     case AVMEDIA_TYPE_AUDIO:
2258         packet_queue_abort(&is->audioq);
2259
2260         SDL_CloseAudio();
2261
2262         packet_queue_flush(&is->audioq);
2263         av_free_packet(&is->audio_pkt);
2264         swr_free(&is->swr_ctx);
2265         av_freep(&is->audio_buf1);
2266         is->audio_buf = NULL;
2267         av_freep(&is->frame);
2268
2269         if (is->rdft) {
2270             av_rdft_end(is->rdft);
2271             av_freep(&is->rdft_data);
2272             is->rdft = NULL;
2273             is->rdft_bits = 0;
2274         }
2275         break;
2276     case AVMEDIA_TYPE_VIDEO:
2277         packet_queue_abort(&is->videoq);
2278
2279         /* note: we also signal this mutex to make sure we deblock the
2280            video thread in all cases */
2281         SDL_LockMutex(is->pictq_mutex);
2282         SDL_CondSignal(is->pictq_cond);
2283         SDL_UnlockMutex(is->pictq_mutex);
2284
2285         SDL_WaitThread(is->video_tid, NULL);
2286
2287         packet_queue_flush(&is->videoq);
2288         break;
2289     case AVMEDIA_TYPE_SUBTITLE:
2290         packet_queue_abort(&is->subtitleq);
2291
2292         /* note: we also signal this mutex to make sure we deblock the
2293            video thread in all cases */
2294         SDL_LockMutex(is->subpq_mutex);
2295         is->subtitle_stream_changed = 1;
2296
2297         SDL_CondSignal(is->subpq_cond);
2298         SDL_UnlockMutex(is->subpq_mutex);
2299
2300         SDL_WaitThread(is->subtitle_tid, NULL);
2301
2302         packet_queue_flush(&is->subtitleq);
2303         break;
2304     default:
2305         break;
2306     }
2307
2308     ic->streams[stream_index]->discard = AVDISCARD_ALL;
2309     avcodec_close(avctx);
2310 #if CONFIG_AVFILTER
2311     free_buffer_pool(&is->buffer_pool);
2312 #endif
2313     switch (avctx->codec_type) {
2314     case AVMEDIA_TYPE_AUDIO:
2315         is->audio_st = NULL;
2316         is->audio_stream = -1;
2317         break;
2318     case AVMEDIA_TYPE_VIDEO:
2319         is->video_st = NULL;
2320         is->video_stream = -1;
2321         break;
2322     case AVMEDIA_TYPE_SUBTITLE:
2323         is->subtitle_st = NULL;
2324         is->subtitle_stream = -1;
2325         break;
2326     default:
2327         break;
2328     }
2329 }
2330
2331 static int decode_interrupt_cb(void *ctx)
2332 {
2333     VideoState *is = ctx;
2334     return is->abort_request;
2335 }
2336
2337 /* this thread gets the stream from the disk or the network */
2338 static int read_thread(void *arg)
2339 {
2340     VideoState *is = arg;
2341     AVFormatContext *ic = NULL;
2342     int err, i, ret;
2343     int st_index[AVMEDIA_TYPE_NB];
2344     AVPacket pkt1, *pkt = &pkt1;
2345     int eof = 0;
2346     int pkt_in_play_range = 0;
2347     AVDictionaryEntry *t;
2348     AVDictionary **opts;
2349     int orig_nb_streams;
2350
2351     memset(st_index, -1, sizeof(st_index));
2352     is->last_video_stream = is->video_stream = -1;
2353     is->last_audio_stream = is->audio_stream = -1;
2354     is->last_subtitle_stream = is->subtitle_stream = -1;
2355
2356     ic = avformat_alloc_context();
2357     ic->interrupt_callback.callback = decode_interrupt_cb;
2358     ic->interrupt_callback.opaque = is;
2359     err = avformat_open_input(&ic, is->filename, is->iformat, &format_opts);
2360     if (err < 0) {
2361         print_error(is->filename, err);
2362         ret = -1;
2363         goto fail;
2364     }
2365     if ((t = av_dict_get(format_opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2366         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2367         ret = AVERROR_OPTION_NOT_FOUND;
2368         goto fail;
2369     }
2370     is->ic = ic;
2371
2372     if (genpts)
2373         ic->flags |= AVFMT_FLAG_GENPTS;
2374
2375     opts = setup_find_stream_info_opts(ic, codec_opts);
2376     orig_nb_streams = ic->nb_streams;
2377
2378     err = avformat_find_stream_info(ic, opts);
2379     if (err < 0) {
2380         fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2381         ret = -1;
2382         goto fail;
2383     }
2384     for (i = 0; i < orig_nb_streams; i++)
2385         av_dict_free(&opts[i]);
2386     av_freep(&opts);
2387
2388     if (ic->pb)
2389         ic->pb->eof_reached = 0; // FIXME hack, ffplay maybe should not use url_feof() to test for the end
2390
2391     if (seek_by_bytes < 0)
2392         seek_by_bytes = !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2393
2394     /* if seeking requested, we execute it */
2395     if (start_time != AV_NOPTS_VALUE) {
2396         int64_t timestamp;
2397
2398         timestamp = start_time;
2399         /* add the stream start time */
2400         if (ic->start_time != AV_NOPTS_VALUE)
2401             timestamp += ic->start_time;
2402         ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2403         if (ret < 0) {
2404             fprintf(stderr, "%s: could not seek to position %0.3f\n",
2405                     is->filename, (double)timestamp / AV_TIME_BASE);
2406         }
2407     }
2408
2409     for (i = 0; i < ic->nb_streams; i++)
2410         ic->streams[i]->discard = AVDISCARD_ALL;
2411     if (!video_disable)
2412         st_index[AVMEDIA_TYPE_VIDEO] =
2413             av_find_best_stream(ic, AVMEDIA_TYPE_VIDEO,
2414                                 wanted_stream[AVMEDIA_TYPE_VIDEO], -1, NULL, 0);
2415     if (!audio_disable)
2416         st_index[AVMEDIA_TYPE_AUDIO] =
2417             av_find_best_stream(ic, AVMEDIA_TYPE_AUDIO,
2418                                 wanted_stream[AVMEDIA_TYPE_AUDIO],
2419                                 st_index[AVMEDIA_TYPE_VIDEO],
2420                                 NULL, 0);
2421     if (!video_disable)
2422         st_index[AVMEDIA_TYPE_SUBTITLE] =
2423             av_find_best_stream(ic, AVMEDIA_TYPE_SUBTITLE,
2424                                 wanted_stream[AVMEDIA_TYPE_SUBTITLE],
2425                                 (st_index[AVMEDIA_TYPE_AUDIO] >= 0 ?
2426                                  st_index[AVMEDIA_TYPE_AUDIO] :
2427                                  st_index[AVMEDIA_TYPE_VIDEO]),
2428                                 NULL, 0);
2429     if (show_status) {
2430         av_dump_format(ic, 0, is->filename, 0);
2431     }
2432
2433     is->show_mode = show_mode;
2434
2435     /* open the streams */
2436     if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2437         stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2438     }
2439
2440     ret = -1;
2441     if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2442         ret = stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2443     }
2444     is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2445     if (is->show_mode == SHOW_MODE_NONE)
2446         is->show_mode = ret >= 0 ? SHOW_MODE_VIDEO : SHOW_MODE_RDFT;
2447
2448     if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2449         stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2450     }
2451
2452     if (is->video_stream < 0 && is->audio_stream < 0) {
2453         fprintf(stderr, "%s: could not open codecs\n", is->filename);
2454         ret = -1;
2455         goto fail;
2456     }
2457
2458     for (;;) {
2459         if (is->abort_request)
2460             break;
2461         if (is->paused != is->last_paused) {
2462             is->last_paused = is->paused;
2463             if (is->paused)
2464                 is->read_pause_return = av_read_pause(ic);
2465             else
2466                 av_read_play(ic);
2467         }
2468 #if CONFIG_RTSP_DEMUXER || CONFIG_MMSH_PROTOCOL
2469         if (is->paused &&
2470                 (!strcmp(ic->iformat->name, "rtsp") ||
2471                  (ic->pb && !strncmp(input_filename, "mmsh:", 5)))) {
2472             /* wait 10 ms to avoid trying to get another packet */
2473             /* XXX: horrible */
2474             SDL_Delay(10);
2475             continue;
2476         }
2477 #endif
2478         if (is->seek_req) {
2479             int64_t seek_target = is->seek_pos;
2480             int64_t seek_min    = is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2481             int64_t seek_max    = is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2482 // FIXME the +-2 is due to rounding being not done in the correct direction in generation
2483 //      of the seek_pos/seek_rel variables
2484
2485             ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2486             if (ret < 0) {
2487                 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2488             } else {
2489                 if (is->audio_stream >= 0) {
2490                     packet_queue_flush(&is->audioq);
2491                     packet_queue_put(&is->audioq, &flush_pkt);
2492                 }
2493                 if (is->subtitle_stream >= 0) {
2494                     packet_queue_flush(&is->subtitleq);
2495                     packet_queue_put(&is->subtitleq, &flush_pkt);
2496                 }
2497                 if (is->video_stream >= 0) {
2498                     packet_queue_flush(&is->videoq);
2499                     packet_queue_put(&is->videoq, &flush_pkt);
2500                 }
2501             }
2502             is->seek_req = 0;
2503             eof = 0;
2504         }
2505         if (is->que_attachments_req) {
2506             avformat_queue_attached_pictures(ic);
2507             is->que_attachments_req = 0;
2508         }
2509
2510         /* if the queue are full, no need to read more */
2511         if (!infinite_buffer &&
2512               (is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2513             || (   (is->audioq   .nb_packets > MIN_FRAMES || is->audio_stream < 0 || is->audioq.abort_request)
2514                 && (is->videoq   .nb_packets > MIN_FRAMES || is->video_stream < 0 || is->videoq.abort_request)
2515                 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream < 0 || is->subtitleq.abort_request)))) {
2516             /* wait 10 ms */
2517             SDL_Delay(10);
2518             continue;
2519         }
2520         if (eof) {
2521             if (is->video_stream >= 0) {
2522                 av_init_packet(pkt);
2523                 pkt->data = NULL;
2524                 pkt->size = 0;
2525                 pkt->stream_index = is->video_stream;
2526                 packet_queue_put(&is->videoq, pkt);
2527             }
2528             if (is->audio_stream >= 0 &&
2529                 is->audio_st->codec->codec->capabilities & CODEC_CAP_DELAY) {
2530                 av_init_packet(pkt);
2531                 pkt->data = NULL;
2532                 pkt->size = 0;
2533                 pkt->stream_index = is->audio_stream;
2534                 packet_queue_put(&is->audioq, pkt);
2535             }
2536             SDL_Delay(10);
2537             if (is->audioq.size + is->videoq.size + is->subtitleq.size == 0) {
2538                 if (loop != 1 && (!loop || --loop)) {
2539                     stream_seek(is, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2540                 } else if (autoexit) {
2541                     ret = AVERROR_EOF;
2542                     goto fail;
2543                 }
2544             }
2545             eof=0;
2546             continue;
2547         }
2548         ret = av_read_frame(ic, pkt);
2549         if (ret < 0) {
2550             if (ret == AVERROR_EOF || url_feof(ic->pb))
2551                 eof = 1;
2552             if (ic->pb && ic->pb->error)
2553                 break;
2554             SDL_Delay(100); /* wait for user event */
2555             continue;
2556         }
2557         /* check if packet is in play range specified by user, then queue, otherwise discard */
2558         pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2559                 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2560                 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2561                 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0) / 1000000
2562                 <= ((double)duration / 1000000);
2563         if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2564             packet_queue_put(&is->audioq, pkt);
2565         } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2566             packet_queue_put(&is->videoq, pkt);
2567         } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2568             packet_queue_put(&is->subtitleq, pkt);
2569         } else {
2570             av_free_packet(pkt);
2571         }
2572     }
2573     /* wait until the end */
2574     while (!is->abort_request) {
2575         SDL_Delay(100);
2576     }
2577
2578     ret = 0;
2579  fail:
2580     /* close each stream */
2581     if (is->audio_stream >= 0)
2582         stream_component_close(is, is->audio_stream);
2583     if (is->video_stream >= 0)
2584         stream_component_close(is, is->video_stream);
2585     if (is->subtitle_stream >= 0)
2586         stream_component_close(is, is->subtitle_stream);
2587     if (is->ic) {
2588         avformat_close_input(&is->ic);
2589     }
2590
2591     if (ret != 0) {
2592         SDL_Event event;
2593
2594         event.type = FF_QUIT_EVENT;
2595         event.user.data1 = is;
2596         SDL_PushEvent(&event);
2597     }
2598     return 0;
2599 }
2600
2601 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2602 {
2603     VideoState *is;
2604
2605     is = av_mallocz(sizeof(VideoState));
2606     if (!is)
2607         return NULL;
2608     av_strlcpy(is->filename, filename, sizeof(is->filename));
2609     is->iformat = iformat;
2610     is->ytop    = 0;
2611     is->xleft   = 0;
2612
2613     /* start video display */
2614     is->pictq_mutex = SDL_CreateMutex();
2615     is->pictq_cond  = SDL_CreateCond();
2616
2617     is->subpq_mutex = SDL_CreateMutex();
2618     is->subpq_cond  = SDL_CreateCond();
2619
2620     packet_queue_init(&is->videoq);
2621     packet_queue_init(&is->audioq);
2622     packet_queue_init(&is->subtitleq);
2623
2624     is->av_sync_type = av_sync_type;
2625     is->read_tid     = SDL_CreateThread(read_thread, is);
2626     if (!is->read_tid) {
2627         av_free(is);
2628         return NULL;
2629     }
2630     return is;
2631 }
2632
2633 static void stream_cycle_channel(VideoState *is, int codec_type)
2634 {
2635     AVFormatContext *ic = is->ic;
2636     int start_index, stream_index;
2637     int old_index;
2638     AVStream *st;
2639
2640     if (codec_type == AVMEDIA_TYPE_VIDEO) {
2641         start_index = is->last_video_stream;
2642         old_index = is->video_stream;
2643     } else if (codec_type == AVMEDIA_TYPE_AUDIO) {
2644         start_index = is->last_audio_stream;
2645         old_index = is->audio_stream;
2646     } else {
2647         start_index = is->last_subtitle_stream;
2648         old_index = is->subtitle_stream;
2649     }
2650     stream_index = start_index;
2651     for (;;) {
2652         if (++stream_index >= is->ic->nb_streams)
2653         {
2654             if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2655             {
2656                 stream_index = -1;
2657                 is->last_subtitle_stream = -1;
2658                 goto the_end;
2659             }
2660             if (start_index == -1)
2661                 return;
2662             stream_index = 0;
2663         }
2664         if (stream_index == start_index)
2665             return;
2666         st = ic->streams[stream_index];
2667         if (st->codec->codec_type == codec_type) {
2668             /* check that parameters are OK */
2669             switch (codec_type) {
2670             case AVMEDIA_TYPE_AUDIO:
2671                 if (st->codec->sample_rate != 0 &&
2672                     st->codec->channels != 0)
2673                     goto the_end;
2674                 break;
2675             case AVMEDIA_TYPE_VIDEO:
2676             case AVMEDIA_TYPE_SUBTITLE:
2677                 goto the_end;
2678             default:
2679                 break;
2680             }
2681         }
2682     }
2683  the_end:
2684     stream_component_close(is, old_index);
2685     stream_component_open(is, stream_index);
2686     if (codec_type == AVMEDIA_TYPE_VIDEO)
2687         is->que_attachments_req = 1;
2688 }
2689
2690
2691 static void toggle_full_screen(VideoState *is)
2692 {
2693 #if defined(__APPLE__) && SDL_VERSION_ATLEAST(1, 2, 14)
2694     /* OS X needs to reallocate the SDL overlays */
2695     int i;
2696     for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++)
2697         is->pictq[i].reallocate = 1;
2698 #endif
2699     is_full_screen = !is_full_screen;
2700     video_open(is, 1);
2701 }
2702
2703 static void toggle_pause(VideoState *is)
2704 {
2705     stream_toggle_pause(is);
2706     is->step = 0;
2707 }
2708
2709 static void step_to_next_frame(VideoState *is)
2710 {
2711     /* if the stream is paused unpause it, then step */
2712     if (is->paused)
2713         stream_toggle_pause(is);
2714     is->step = 1;
2715 }
2716
2717 static void toggle_audio_display(VideoState *is)
2718 {
2719     int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2720     is->show_mode = (is->show_mode + 1) % SHOW_MODE_NB;
2721     fill_rectangle(screen,
2722                 is->xleft, is->ytop, is->width, is->height,
2723                 bgcolor);
2724     SDL_UpdateRect(screen, is->xleft, is->ytop, is->width, is->height);
2725 }
2726
2727 /* handle an event sent by the GUI */
2728 static void event_loop(VideoState *cur_stream)
2729 {
2730     SDL_Event event;
2731     double incr, pos, frac;
2732
2733     for (;;) {
2734         double x;
2735         SDL_WaitEvent(&event);
2736         switch (event.type) {
2737         case SDL_KEYDOWN:
2738             if (exit_on_keydown) {
2739                 do_exit(cur_stream);
2740                 break;
2741             }
2742             switch (event.key.keysym.sym) {
2743             case SDLK_ESCAPE:
2744             case SDLK_q:
2745                 do_exit(cur_stream);
2746                 break;
2747             case SDLK_f:
2748                 toggle_full_screen(cur_stream);
2749                 cur_stream->force_refresh = 1;
2750                 break;
2751             case SDLK_p:
2752             case SDLK_SPACE:
2753                 toggle_pause(cur_stream);
2754                 break;
2755             case SDLK_s: // S: Step to next frame
2756                 step_to_next_frame(cur_stream);
2757                 break;
2758             case SDLK_a:
2759                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2760                 break;
2761             case SDLK_v:
2762                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2763                 break;
2764             case SDLK_t:
2765                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2766                 break;
2767             case SDLK_w:
2768                 toggle_audio_display(cur_stream);
2769                 cur_stream->force_refresh = 1;
2770                 break;
2771             case SDLK_PAGEUP:
2772                 incr = 600.0;
2773                 goto do_seek;
2774             case SDLK_PAGEDOWN:
2775                 incr = -600.0;
2776                 goto do_seek;
2777             case SDLK_LEFT:
2778                 incr = -10.0;
2779                 goto do_seek;
2780             case SDLK_RIGHT:
2781                 incr = 10.0;
2782                 goto do_seek;
2783             case SDLK_UP:
2784                 incr = 60.0;
2785                 goto do_seek;
2786             case SDLK_DOWN:
2787                 incr = -60.0;
2788             do_seek:
2789                     if (seek_by_bytes) {
2790                         if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos >= 0) {
2791                             pos = cur_stream->video_current_pos;
2792                         } else if (cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos >= 0) {
2793                             pos = cur_stream->audio_pkt.pos;
2794                         } else
2795                             pos = avio_tell(cur_stream->ic->pb);
2796                         if (cur_stream->ic->bit_rate)
2797                             incr *= cur_stream->ic->bit_rate / 8.0;
2798                         else
2799                             incr *= 180000.0;
2800                         pos += incr;
2801                         stream_seek(cur_stream, pos, incr, 1);
2802                     } else {
2803                         pos = get_master_clock(cur_stream);
2804                         pos += incr;
2805                         stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2806                     }
2807                 break;
2808             default:
2809                 break;
2810             }
2811             break;
2812         case SDL_VIDEOEXPOSE:
2813             cur_stream->force_refresh = 1;
2814             break;
2815         case SDL_MOUSEBUTTONDOWN:
2816             if (exit_on_mousedown) {
2817                 do_exit(cur_stream);
2818                 break;
2819             }
2820         case SDL_MOUSEMOTION:
2821             if (event.type == SDL_MOUSEBUTTONDOWN) {
2822                 x = event.button.x;
2823             } else {
2824                 if (event.motion.state != SDL_PRESSED)
2825                     break;
2826                 x = event.motion.x;
2827             }
2828                 if (seek_by_bytes || cur_stream->ic->duration <= 0) {
2829                     uint64_t size =  avio_size(cur_stream->ic->pb);
2830                     stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2831                 } else {
2832                     int64_t ts;
2833                     int ns, hh, mm, ss;
2834                     int tns, thh, tmm, tss;
2835                     tns  = cur_stream->ic->duration / 1000000LL;
2836                     thh  = tns / 3600;
2837                     tmm  = (tns % 3600) / 60;
2838                     tss  = (tns % 60);
2839                     frac = x / cur_stream->width;
2840                     ns   = frac * tns;
2841                     hh   = ns / 3600;
2842                     mm   = (ns % 3600) / 60;
2843                     ss   = (ns % 60);
2844                     fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d)       \n", frac*100,
2845                             hh, mm, ss, thh, tmm, tss);
2846                     ts = frac * cur_stream->ic->duration;
2847                     if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2848                         ts += cur_stream->ic->start_time;
2849                     stream_seek(cur_stream, ts, 0, 0);
2850                 }
2851             break;
2852         case SDL_VIDEORESIZE:
2853                 screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2854                                           SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2855                 screen_width  = cur_stream->width  = event.resize.w;
2856                 screen_height = cur_stream->height = event.resize.h;
2857                 cur_stream->force_refresh = 1;
2858             break;
2859         case SDL_QUIT:
2860         case FF_QUIT_EVENT:
2861             do_exit(cur_stream);
2862             break;
2863         case FF_ALLOC_EVENT:
2864             alloc_picture(event.user.data1);
2865             break;
2866         case FF_REFRESH_EVENT:
2867             video_refresh(event.user.data1);
2868             cur_stream->refresh = 0;
2869             break;
2870         default:
2871             break;
2872         }
2873     }
2874 }
2875
2876 static int opt_frame_size(const char *opt, const char *arg)
2877 {
2878     av_log(NULL, AV_LOG_WARNING, "Option -s is deprecated, use -video_size.\n");
2879     return opt_default("video_size", arg);
2880 }
2881
2882 static int opt_width(const char *opt, const char *arg)
2883 {
2884     screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2885     return 0;
2886 }
2887
2888 static int opt_height(const char *opt, const char *arg)
2889 {
2890     screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2891     return 0;
2892 }
2893
2894 static int opt_format(const char *opt, const char *arg)
2895 {
2896     file_iformat = av_find_input_format(arg);
2897     if (!file_iformat) {
2898         fprintf(stderr, "Unknown input format: %s\n", arg);
2899         return AVERROR(EINVAL);
2900     }
2901     return 0;
2902 }
2903
2904 static int opt_frame_pix_fmt(const char *opt, const char *arg)
2905 {
2906     av_log(NULL, AV_LOG_WARNING, "Option -pix_fmt is deprecated, use -pixel_format.\n");
2907     return opt_default("pixel_format", arg);
2908 }
2909
2910 static int opt_sync(const char *opt, const char *arg)
2911 {
2912     if (!strcmp(arg, "audio"))
2913         av_sync_type = AV_SYNC_AUDIO_MASTER;
2914     else if (!strcmp(arg, "video"))
2915         av_sync_type = AV_SYNC_VIDEO_MASTER;
2916     else if (!strcmp(arg, "ext"))
2917         av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
2918     else {
2919         fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
2920         exit(1);
2921     }
2922     return 0;
2923 }
2924
2925 static int opt_seek(const char *opt, const char *arg)
2926 {
2927     start_time = parse_time_or_die(opt, arg, 1);
2928     return 0;
2929 }
2930
2931 static int opt_duration(const char *opt, const char *arg)
2932 {
2933     duration = parse_time_or_die(opt, arg, 1);
2934     return 0;
2935 }
2936
2937 static int opt_show_mode(const char *opt, const char *arg)
2938 {
2939     show_mode = !strcmp(arg, "video") ? SHOW_MODE_VIDEO :
2940                 !strcmp(arg, "waves") ? SHOW_MODE_WAVES :
2941                 !strcmp(arg, "rdft" ) ? SHOW_MODE_RDFT  :
2942                 parse_number_or_die(opt, arg, OPT_INT, 0, SHOW_MODE_NB-1);
2943     return 0;
2944 }
2945
2946 static void opt_input_file(void *optctx, const char *filename)
2947 {
2948     if (input_filename) {
2949         fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
2950                 filename, input_filename);
2951         exit_program(1);
2952     }
2953     if (!strcmp(filename, "-"))
2954         filename = "pipe:";
2955     input_filename = filename;
2956 }
2957
2958 static int opt_codec(void *o, const char *opt, const char *arg)
2959 {
2960     switch(opt[strlen(opt)-1]){
2961     case 'a' :    audio_codec_name = arg; break;
2962     case 's' : subtitle_codec_name = arg; break;
2963     case 'v' :    video_codec_name = arg; break;
2964     }
2965     return 0;
2966 }
2967
2968 static int dummy;
2969
2970 static const OptionDef options[] = {
2971 #include "cmdutils_common_opts.h"
2972     { "x", HAS_ARG, { (void*)opt_width }, "force displayed width", "width" },
2973     { "y", HAS_ARG, { (void*)opt_height }, "force displayed height", "height" },
2974     { "s", HAS_ARG | OPT_VIDEO, { (void*)opt_frame_size }, "set frame size (WxH or abbreviation)", "size" },
2975     { "fs", OPT_BOOL, { (void*)&is_full_screen }, "force full screen" },
2976     { "an", OPT_BOOL, { (void*)&audio_disable }, "disable audio" },
2977     { "vn", OPT_BOOL, { (void*)&video_disable }, "disable video" },
2978     { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_AUDIO] }, "select desired audio stream", "stream_number" },
2979     { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_VIDEO] }, "select desired video stream", "stream_number" },
2980     { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE] }, "select desired subtitle stream", "stream_number" },
2981     { "ss", HAS_ARG, { (void*)&opt_seek }, "seek to a given position in seconds", "pos" },
2982     { "t", HAS_ARG, { (void*)&opt_duration }, "play  \"duration\" seconds of audio/video", "duration" },
2983     { "bytes", OPT_INT | HAS_ARG, { (void*)&seek_by_bytes }, "seek by bytes 0=off 1=on -1=auto", "val" },
2984     { "nodisp", OPT_BOOL, { (void*)&display_disable }, "disable graphical display" },
2985     { "f", HAS_ARG, { (void*)opt_format }, "force format", "fmt" },
2986     { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, { (void*)opt_frame_pix_fmt }, "set pixel format", "format" },
2987     { "stats", OPT_BOOL | OPT_EXPERT, { (void*)&show_status }, "show status", "" },
2988     { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&workaround_bugs }, "workaround bugs", "" },
2989     { "fast", OPT_BOOL | OPT_EXPERT, { (void*)&fast }, "non spec compliant optimizations", "" },
2990     { "genpts", OPT_BOOL | OPT_EXPERT, { (void*)&genpts }, "generate pts", "" },
2991     { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&decoder_reorder_pts }, "let decoder reorder pts 0=off 1=on -1=auto", ""},
2992     { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&lowres }, "", "" },
2993     { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_loop_filter }, "", "" },
2994     { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_frame }, "", "" },
2995     { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_idct }, "", "" },
2996     { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&idct }, "set idct algo",  "algo" },
2997     { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&error_concealment }, "set error concealment options",  "bit_mask" },
2998     { "sync", HAS_ARG | OPT_EXPERT, { (void*)opt_sync }, "set audio-video sync. type (type=audio/video/ext)", "type" },
2999     { "autoexit", OPT_BOOL | OPT_EXPERT, { (void*)&autoexit }, "exit at the end", "" },
3000     { "exitonkeydown", OPT_BOOL | OPT_EXPERT, { (void*)&exit_on_keydown }, "exit on key down", "" },
3001     { "exitonmousedown", OPT_BOOL | OPT_EXPERT, { (void*)&exit_on_mousedown }, "exit on mouse down", "" },
3002     { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&loop }, "set number of times the playback shall be looped", "loop count" },
3003     { "framedrop", OPT_BOOL | OPT_EXPERT, { (void*)&framedrop }, "drop frames when cpu is too slow", "" },
3004     { "infbuf", OPT_BOOL | OPT_EXPERT, { (void*)&infinite_buffer }, "don't limit the input buffer size (useful with realtime streams)", "" },
3005     { "window_title", OPT_STRING | HAS_ARG, { (void*)&window_title }, "set window title", "window title" },
3006 #if CONFIG_AVFILTER
3007     { "vf", OPT_STRING | HAS_ARG, { (void*)&vfilters }, "video filters", "filter list" },
3008 #endif
3009     { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, { (void*)&rdftspeed }, "rdft speed", "msecs" },
3010     { "showmode", HAS_ARG, {(void*)opt_show_mode}, "select show mode (0 = video, 1 = waves, 2 = RDFT)", "mode" },
3011     { "default", HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, { (void*)opt_default }, "generic catch all option", "" },
3012     { "i", OPT_BOOL, {(void *)&dummy}, "read specified file", "input_file"},
3013     { "codec", HAS_ARG | OPT_FUNC2, {(void*)opt_codec}, "force decoder", "decoder" },
3014     { NULL, },
3015 };
3016
3017 static void show_usage(void)
3018 {
3019     av_log(NULL, AV_LOG_INFO, "Simple media player\n");
3020     av_log(NULL, AV_LOG_INFO, "usage: %s [options] input_file\n", program_name);
3021     av_log(NULL, AV_LOG_INFO, "\n");
3022 }
3023
3024 void show_help_default(const char *opt, const char *arg)
3025 {
3026     av_log_set_callback(log_callback_help);
3027     show_usage();
3028     show_help_options(options, "Main options:", 0, OPT_EXPERT, 0);
3029     show_help_options(options, "Advanced options:", OPT_EXPERT, 0, 0);
3030     printf("\n");
3031     show_help_children(avcodec_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3032     show_help_children(avformat_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3033 #if !CONFIG_AVFILTER
3034     show_help_children(sws_get_class(), AV_OPT_FLAG_ENCODING_PARAM);
3035 #else
3036     show_help_children(avfilter_get_class(), AV_OPT_FLAG_FILTERING_PARAM);
3037 #endif
3038     printf("\nWhile playing:\n"
3039            "q, ESC              quit\n"
3040            "f                   toggle full screen\n"
3041            "p, SPC              pause\n"
3042            "a                   cycle audio channel\n"
3043            "v                   cycle video channel\n"
3044            "t                   cycle subtitle channel\n"
3045            "w                   show audio waves\n"
3046            "s                   activate frame-step mode\n"
3047            "left/right          seek backward/forward 10 seconds\n"
3048            "down/up             seek backward/forward 1 minute\n"
3049            "page down/page up   seek backward/forward 10 minutes\n"
3050            "mouse click         seek to percentage in file corresponding to fraction of width\n"
3051            );
3052 }
3053
3054 static int lockmgr(void **mtx, enum AVLockOp op)
3055 {
3056    switch(op) {
3057       case AV_LOCK_CREATE:
3058           *mtx = SDL_CreateMutex();
3059           if(!*mtx)
3060               return 1;
3061           return 0;
3062       case AV_LOCK_OBTAIN:
3063           return !!SDL_LockMutex(*mtx);
3064       case AV_LOCK_RELEASE:
3065           return !!SDL_UnlockMutex(*mtx);
3066       case AV_LOCK_DESTROY:
3067           SDL_DestroyMutex(*mtx);
3068           return 0;
3069    }
3070    return 1;
3071 }
3072
3073 /* Called from the main */
3074 int main(int argc, char **argv)
3075 {
3076     int flags;
3077     VideoState *is;
3078     char dummy_videodriver[] = "SDL_VIDEODRIVER=dummy";
3079
3080     av_log_set_flags(AV_LOG_SKIP_REPEATED);
3081     parse_loglevel(argc, argv, options);
3082
3083     /* register all codecs, demux and protocols */
3084     avcodec_register_all();
3085 #if CONFIG_AVDEVICE
3086     avdevice_register_all();
3087 #endif
3088 #if CONFIG_AVFILTER
3089     avfilter_register_all();
3090 #endif
3091     av_register_all();
3092     avformat_network_init();
3093
3094     init_opts();
3095
3096     signal(SIGINT , sigterm_handler); /* Interrupt (ANSI).    */
3097     signal(SIGTERM, sigterm_handler); /* Termination (ANSI).  */
3098
3099     show_banner(argc, argv, options);
3100
3101     parse_options(NULL, argc, argv, options, opt_input_file);
3102
3103     if (!input_filename) {
3104         show_usage();
3105         fprintf(stderr, "An input file must be specified\n");
3106         fprintf(stderr, "Use -h to get full help or, even better, run 'man %s'\n", program_name);
3107         exit(1);
3108     }
3109
3110     if (display_disable) {
3111         video_disable = 1;
3112     }
3113     flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3114     if (audio_disable)
3115         flags &= ~SDL_INIT_AUDIO;
3116     if (display_disable)
3117         SDL_putenv(dummy_videodriver); /* For the event queue, we always need a video driver. */
3118 #if !defined(__MINGW32__) && !defined(__APPLE__)
3119     flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3120 #endif
3121     if (SDL_Init (flags)) {
3122         fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3123         fprintf(stderr, "(Did you set the DISPLAY variable?)\n");
3124         exit(1);
3125     }
3126
3127     if (!display_disable) {
3128 #if HAVE_SDL_VIDEO_SIZE
3129         const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3130         fs_screen_width = vi->current_w;
3131         fs_screen_height = vi->current_h;
3132 #endif
3133     }
3134
3135     SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3136     SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3137     SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3138
3139     if (av_lockmgr_register(lockmgr)) {
3140         fprintf(stderr, "Could not initialize lock manager!\n");
3141         do_exit(NULL);
3142     }
3143
3144     av_init_packet(&flush_pkt);
3145     flush_pkt.data = (char *)(intptr_t)"FLUSH";
3146
3147     is = stream_open(input_filename, file_iformat);
3148     if (!is) {
3149         fprintf(stderr, "Failed to initialize VideoState!\n");
3150         do_exit(NULL);
3151     }
3152
3153     event_loop(is);
3154
3155     /* never returns */
3156
3157     return 0;
3158 }