]> git.sesse.net Git - ffmpeg/blob - ffplay.c
Merge remote-tracking branch 'qatar/master'
[ffmpeg] / ffplay.c
1 /*
2  * Copyright (c) 2003 Fabrice Bellard
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20
21 /**
22  * @file
23  * simple media player based on the FFmpeg libraries
24  */
25
26 #include "config.h"
27 #include <inttypes.h>
28 #include <math.h>
29 #include <limits.h>
30 #include <signal.h>
31 #include "libavutil/avstring.h"
32 #include "libavutil/colorspace.h"
33 #include "libavutil/mathematics.h"
34 #include "libavutil/pixdesc.h"
35 #include "libavutil/imgutils.h"
36 #include "libavutil/dict.h"
37 #include "libavutil/parseutils.h"
38 #include "libavutil/samplefmt.h"
39 #include "libavutil/avassert.h"
40 #include "libavformat/avformat.h"
41 #include "libavdevice/avdevice.h"
42 #include "libswscale/swscale.h"
43 #include "libavutil/opt.h"
44 #include "libavcodec/avfft.h"
45 #include "libswresample/swresample.h"
46
47 #if CONFIG_AVFILTER
48 # include "libavfilter/avcodec.h"
49 # include "libavfilter/avfilter.h"
50 # include "libavfilter/avfiltergraph.h"
51 # include "libavfilter/buffersink.h"
52 #endif
53
54 #include <SDL.h>
55 #include <SDL_thread.h>
56
57 #include "cmdutils.h"
58
59 #include <unistd.h>
60 #include <assert.h>
61
62 const char program_name[] = "ffplay";
63 const int program_birth_year = 2003;
64
65 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
66 #define MIN_FRAMES 5
67
68 /* SDL audio buffer size, in samples. Should be small to have precise
69    A/V sync as SDL does not have hardware buffer fullness info. */
70 #define SDL_AUDIO_BUFFER_SIZE 1024
71
72 /* no AV sync correction is done if below the AV sync threshold */
73 #define AV_SYNC_THRESHOLD 0.01
74 /* no AV correction is done if too big error */
75 #define AV_NOSYNC_THRESHOLD 10.0
76
77 /* maximum audio speed change to get correct sync */
78 #define SAMPLE_CORRECTION_PERCENT_MAX 10
79
80 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
81 #define AUDIO_DIFF_AVG_NB   20
82
83 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
84 #define SAMPLE_ARRAY_SIZE (2 * 65536)
85
86 static int sws_flags = SWS_BICUBIC;
87
88 typedef struct PacketQueue {
89     AVPacketList *first_pkt, *last_pkt;
90     int nb_packets;
91     int size;
92     int abort_request;
93     SDL_mutex *mutex;
94     SDL_cond *cond;
95 } PacketQueue;
96
97 #define VIDEO_PICTURE_QUEUE_SIZE 2
98 #define SUBPICTURE_QUEUE_SIZE 4
99
100 typedef struct VideoPicture {
101     double pts;                                  ///< presentation time stamp for this picture
102     double duration;                             ///< expected duration of the frame
103     int64_t pos;                                 ///< byte position in file
104     int skip;
105     SDL_Overlay *bmp;
106     int width, height; /* source height & width */
107     int allocated;
108     int reallocate;
109     enum PixelFormat pix_fmt;
110
111 #if CONFIG_AVFILTER
112     AVFilterBufferRef *picref;
113 #endif
114 } VideoPicture;
115
116 typedef struct SubPicture {
117     double pts; /* presentation time stamp for this picture */
118     AVSubtitle sub;
119 } SubPicture;
120
121 enum {
122     AV_SYNC_AUDIO_MASTER, /* default choice */
123     AV_SYNC_VIDEO_MASTER,
124     AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
125 };
126
127 typedef struct VideoState {
128     SDL_Thread *read_tid;
129     SDL_Thread *video_tid;
130     SDL_Thread *refresh_tid;
131     AVInputFormat *iformat;
132     int no_background;
133     int abort_request;
134     int force_refresh;
135     int paused;
136     int last_paused;
137     int seek_req;
138     int seek_flags;
139     int64_t seek_pos;
140     int64_t seek_rel;
141     int read_pause_return;
142     AVFormatContext *ic;
143
144     int audio_stream;
145
146     int av_sync_type;
147     double external_clock; /* external clock base */
148     int64_t external_clock_time;
149
150     double audio_clock;
151     double audio_diff_cum; /* used for AV difference average computation */
152     double audio_diff_avg_coef;
153     double audio_diff_threshold;
154     int audio_diff_avg_count;
155     AVStream *audio_st;
156     PacketQueue audioq;
157     int audio_hw_buf_size;
158     DECLARE_ALIGNED(16,uint8_t,audio_buf2)[AVCODEC_MAX_AUDIO_FRAME_SIZE * 4];
159     uint8_t silence_buf[SDL_AUDIO_BUFFER_SIZE];
160     uint8_t *audio_buf;
161     uint8_t *audio_buf1;
162     unsigned int audio_buf_size; /* in bytes */
163     int audio_buf_index; /* in bytes */
164     int audio_write_buf_size;
165     AVPacket audio_pkt_temp;
166     AVPacket audio_pkt;
167     enum AVSampleFormat audio_src_fmt;
168     enum AVSampleFormat audio_tgt_fmt;
169     int audio_src_channels;
170     int audio_tgt_channels;
171     int64_t audio_src_channel_layout;
172     int64_t audio_tgt_channel_layout;
173     int audio_src_freq;
174     int audio_tgt_freq;
175     struct SwrContext *swr_ctx;
176     double audio_current_pts;
177     double audio_current_pts_drift;
178     int frame_drops_early;
179     int frame_drops_late;
180     AVFrame *frame;
181
182     enum ShowMode {
183         SHOW_MODE_NONE = -1, SHOW_MODE_VIDEO = 0, SHOW_MODE_WAVES, SHOW_MODE_RDFT, SHOW_MODE_NB
184     } show_mode;
185     int16_t sample_array[SAMPLE_ARRAY_SIZE];
186     int sample_array_index;
187     int last_i_start;
188     RDFTContext *rdft;
189     int rdft_bits;
190     FFTSample *rdft_data;
191     int xpos;
192
193     SDL_Thread *subtitle_tid;
194     int subtitle_stream;
195     int subtitle_stream_changed;
196     AVStream *subtitle_st;
197     PacketQueue subtitleq;
198     SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
199     int subpq_size, subpq_rindex, subpq_windex;
200     SDL_mutex *subpq_mutex;
201     SDL_cond *subpq_cond;
202
203     double frame_timer;
204     double frame_last_pts;
205     double frame_last_duration;
206     double frame_last_dropped_pts;
207     double frame_last_returned_time;
208     double frame_last_filter_delay;
209     int64_t frame_last_dropped_pos;
210     double video_clock;                          ///< pts of last decoded frame / predicted pts of next decoded frame
211     int video_stream;
212     AVStream *video_st;
213     PacketQueue videoq;
214     double video_current_pts;                    ///< current displayed pts (different from video_clock if frame fifos are used)
215     double video_current_pts_drift;              ///< video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
216     int64_t video_current_pos;                   ///< current displayed file pos
217     VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
218     int pictq_size, pictq_rindex, pictq_windex;
219     SDL_mutex *pictq_mutex;
220     SDL_cond *pictq_cond;
221 #if !CONFIG_AVFILTER
222     struct SwsContext *img_convert_ctx;
223 #endif
224
225     char filename[1024];
226     int width, height, xleft, ytop;
227     int step;
228
229 #if CONFIG_AVFILTER
230     AVFilterContext *out_video_filter;          ///< the last filter in the video chain
231 #endif
232
233     int refresh;
234 } VideoState;
235
236 static int opt_help(const char *opt, const char *arg);
237
238 /* options specified by the user */
239 static AVInputFormat *file_iformat;
240 static const char *input_filename;
241 static const char *window_title;
242 static int fs_screen_width;
243 static int fs_screen_height;
244 static int screen_width  = 0;
245 static int screen_height = 0;
246 static int audio_disable;
247 static int video_disable;
248 static int wanted_stream[AVMEDIA_TYPE_NB] = {
249     [AVMEDIA_TYPE_AUDIO]    = -1,
250     [AVMEDIA_TYPE_VIDEO]    = -1,
251     [AVMEDIA_TYPE_SUBTITLE] = -1,
252 };
253 static int seek_by_bytes = -1;
254 static int display_disable;
255 static int show_status = 1;
256 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
257 static int64_t start_time = AV_NOPTS_VALUE;
258 static int64_t duration = AV_NOPTS_VALUE;
259 static int workaround_bugs = 1;
260 static int fast = 0;
261 static int genpts = 0;
262 static int lowres = 0;
263 static int idct = FF_IDCT_AUTO;
264 static enum AVDiscard skip_frame       = AVDISCARD_DEFAULT;
265 static enum AVDiscard skip_idct        = AVDISCARD_DEFAULT;
266 static enum AVDiscard skip_loop_filter = AVDISCARD_DEFAULT;
267 static int error_concealment = 3;
268 static int decoder_reorder_pts = -1;
269 static int autoexit;
270 static int exit_on_keydown;
271 static int exit_on_mousedown;
272 static int loop = 1;
273 static int framedrop = -1;
274 static enum ShowMode show_mode = SHOW_MODE_NONE;
275 static const char *audio_codec_name;
276 static const char *subtitle_codec_name;
277 static const char *video_codec_name;
278 static int rdftspeed = 20;
279 #if CONFIG_AVFILTER
280 static char *vfilters = NULL;
281 #endif
282
283 /* current context */
284 static int is_full_screen;
285 static int64_t audio_callback_time;
286
287 static AVPacket flush_pkt;
288
289 #define FF_ALLOC_EVENT   (SDL_USEREVENT)
290 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
291 #define FF_QUIT_EVENT    (SDL_USEREVENT + 2)
292
293 static SDL_Surface *screen;
294
295 void av_noreturn exit_program(int ret)
296 {
297     exit(ret);
298 }
299
300 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
301 {
302     AVPacketList *pkt1;
303
304     /* duplicate the packet */
305     if (pkt != &flush_pkt && av_dup_packet(pkt) < 0)
306         return -1;
307
308     pkt1 = av_malloc(sizeof(AVPacketList));
309     if (!pkt1)
310         return -1;
311     pkt1->pkt = *pkt;
312     pkt1->next = NULL;
313
314
315     SDL_LockMutex(q->mutex);
316
317     if (!q->last_pkt)
318
319         q->first_pkt = pkt1;
320     else
321         q->last_pkt->next = pkt1;
322     q->last_pkt = pkt1;
323     q->nb_packets++;
324     q->size += pkt1->pkt.size + sizeof(*pkt1);
325     /* XXX: should duplicate packet data in DV case */
326     SDL_CondSignal(q->cond);
327
328     SDL_UnlockMutex(q->mutex);
329     return 0;
330 }
331
332 /* packet queue handling */
333 static void packet_queue_init(PacketQueue *q)
334 {
335     memset(q, 0, sizeof(PacketQueue));
336     q->mutex = SDL_CreateMutex();
337     q->cond = SDL_CreateCond();
338     packet_queue_put(q, &flush_pkt);
339 }
340
341 static void packet_queue_flush(PacketQueue *q)
342 {
343     AVPacketList *pkt, *pkt1;
344
345     SDL_LockMutex(q->mutex);
346     for (pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
347         pkt1 = pkt->next;
348         av_free_packet(&pkt->pkt);
349         av_freep(&pkt);
350     }
351     q->last_pkt = NULL;
352     q->first_pkt = NULL;
353     q->nb_packets = 0;
354     q->size = 0;
355     SDL_UnlockMutex(q->mutex);
356 }
357
358 static void packet_queue_end(PacketQueue *q)
359 {
360     packet_queue_flush(q);
361     SDL_DestroyMutex(q->mutex);
362     SDL_DestroyCond(q->cond);
363 }
364
365 static void packet_queue_abort(PacketQueue *q)
366 {
367     SDL_LockMutex(q->mutex);
368
369     q->abort_request = 1;
370
371     SDL_CondSignal(q->cond);
372
373     SDL_UnlockMutex(q->mutex);
374 }
375
376 /* return < 0 if aborted, 0 if no packet and > 0 if packet.  */
377 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
378 {
379     AVPacketList *pkt1;
380     int ret;
381
382     SDL_LockMutex(q->mutex);
383
384     for (;;) {
385         if (q->abort_request) {
386             ret = -1;
387             break;
388         }
389
390         pkt1 = q->first_pkt;
391         if (pkt1) {
392             q->first_pkt = pkt1->next;
393             if (!q->first_pkt)
394                 q->last_pkt = NULL;
395             q->nb_packets--;
396             q->size -= pkt1->pkt.size + sizeof(*pkt1);
397             *pkt = pkt1->pkt;
398             av_free(pkt1);
399             ret = 1;
400             break;
401         } else if (!block) {
402             ret = 0;
403             break;
404         } else {
405             SDL_CondWait(q->cond, q->mutex);
406         }
407     }
408     SDL_UnlockMutex(q->mutex);
409     return ret;
410 }
411
412 static inline void fill_rectangle(SDL_Surface *screen,
413                                   int x, int y, int w, int h, int color)
414 {
415     SDL_Rect rect;
416     rect.x = x;
417     rect.y = y;
418     rect.w = w;
419     rect.h = h;
420     SDL_FillRect(screen, &rect, color);
421 }
422
423 #define ALPHA_BLEND(a, oldp, newp, s)\
424 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
425
426 #define RGBA_IN(r, g, b, a, s)\
427 {\
428     unsigned int v = ((const uint32_t *)(s))[0];\
429     a = (v >> 24) & 0xff;\
430     r = (v >> 16) & 0xff;\
431     g = (v >> 8) & 0xff;\
432     b = v & 0xff;\
433 }
434
435 #define YUVA_IN(y, u, v, a, s, pal)\
436 {\
437     unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
438     a = (val >> 24) & 0xff;\
439     y = (val >> 16) & 0xff;\
440     u = (val >> 8) & 0xff;\
441     v = val & 0xff;\
442 }
443
444 #define YUVA_OUT(d, y, u, v, a)\
445 {\
446     ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
447 }
448
449
450 #define BPP 1
451
452 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
453 {
454     int wrap, wrap3, width2, skip2;
455     int y, u, v, a, u1, v1, a1, w, h;
456     uint8_t *lum, *cb, *cr;
457     const uint8_t *p;
458     const uint32_t *pal;
459     int dstx, dsty, dstw, dsth;
460
461     dstw = av_clip(rect->w, 0, imgw);
462     dsth = av_clip(rect->h, 0, imgh);
463     dstx = av_clip(rect->x, 0, imgw - dstw);
464     dsty = av_clip(rect->y, 0, imgh - dsth);
465     lum = dst->data[0] + dsty * dst->linesize[0];
466     cb  = dst->data[1] + (dsty >> 1) * dst->linesize[1];
467     cr  = dst->data[2] + (dsty >> 1) * dst->linesize[2];
468
469     width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
470     skip2 = dstx >> 1;
471     wrap = dst->linesize[0];
472     wrap3 = rect->pict.linesize[0];
473     p = rect->pict.data[0];
474     pal = (const uint32_t *)rect->pict.data[1];  /* Now in YCrCb! */
475
476     if (dsty & 1) {
477         lum += dstx;
478         cb += skip2;
479         cr += skip2;
480
481         if (dstx & 1) {
482             YUVA_IN(y, u, v, a, p, pal);
483             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
484             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
485             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
486             cb++;
487             cr++;
488             lum++;
489             p += BPP;
490         }
491         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
492             YUVA_IN(y, u, v, a, p, pal);
493             u1 = u;
494             v1 = v;
495             a1 = a;
496             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
497
498             YUVA_IN(y, u, v, a, p + BPP, pal);
499             u1 += u;
500             v1 += v;
501             a1 += a;
502             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
503             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
504             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
505             cb++;
506             cr++;
507             p += 2 * BPP;
508             lum += 2;
509         }
510         if (w) {
511             YUVA_IN(y, u, v, a, p, pal);
512             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
513             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
514             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
515             p++;
516             lum++;
517         }
518         p += wrap3 - dstw * BPP;
519         lum += wrap - dstw - dstx;
520         cb += dst->linesize[1] - width2 - skip2;
521         cr += dst->linesize[2] - width2 - skip2;
522     }
523     for (h = dsth - (dsty & 1); h >= 2; h -= 2) {
524         lum += dstx;
525         cb += skip2;
526         cr += skip2;
527
528         if (dstx & 1) {
529             YUVA_IN(y, u, v, a, p, pal);
530             u1 = u;
531             v1 = v;
532             a1 = a;
533             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
534             p += wrap3;
535             lum += wrap;
536             YUVA_IN(y, u, v, a, p, pal);
537             u1 += u;
538             v1 += v;
539             a1 += a;
540             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
541             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
542             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
543             cb++;
544             cr++;
545             p += -wrap3 + BPP;
546             lum += -wrap + 1;
547         }
548         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
549             YUVA_IN(y, u, v, a, p, pal);
550             u1 = u;
551             v1 = v;
552             a1 = a;
553             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
554
555             YUVA_IN(y, u, v, a, p + BPP, pal);
556             u1 += u;
557             v1 += v;
558             a1 += a;
559             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
560             p += wrap3;
561             lum += wrap;
562
563             YUVA_IN(y, u, v, a, p, pal);
564             u1 += u;
565             v1 += v;
566             a1 += a;
567             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
568
569             YUVA_IN(y, u, v, a, p + BPP, pal);
570             u1 += u;
571             v1 += v;
572             a1 += a;
573             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
574
575             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
576             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
577
578             cb++;
579             cr++;
580             p += -wrap3 + 2 * BPP;
581             lum += -wrap + 2;
582         }
583         if (w) {
584             YUVA_IN(y, u, v, a, p, pal);
585             u1 = u;
586             v1 = v;
587             a1 = a;
588             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
589             p += wrap3;
590             lum += wrap;
591             YUVA_IN(y, u, v, a, p, pal);
592             u1 += u;
593             v1 += v;
594             a1 += a;
595             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
596             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
597             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
598             cb++;
599             cr++;
600             p += -wrap3 + BPP;
601             lum += -wrap + 1;
602         }
603         p += wrap3 + (wrap3 - dstw * BPP);
604         lum += wrap + (wrap - dstw - dstx);
605         cb += dst->linesize[1] - width2 - skip2;
606         cr += dst->linesize[2] - width2 - skip2;
607     }
608     /* handle odd height */
609     if (h) {
610         lum += dstx;
611         cb += skip2;
612         cr += skip2;
613
614         if (dstx & 1) {
615             YUVA_IN(y, u, v, a, p, pal);
616             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
617             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
618             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
619             cb++;
620             cr++;
621             lum++;
622             p += BPP;
623         }
624         for (w = dstw - (dstx & 1); w >= 2; w -= 2) {
625             YUVA_IN(y, u, v, a, p, pal);
626             u1 = u;
627             v1 = v;
628             a1 = a;
629             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
630
631             YUVA_IN(y, u, v, a, p + BPP, pal);
632             u1 += u;
633             v1 += v;
634             a1 += a;
635             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
636             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
637             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
638             cb++;
639             cr++;
640             p += 2 * BPP;
641             lum += 2;
642         }
643         if (w) {
644             YUVA_IN(y, u, v, a, p, pal);
645             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
646             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
647             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
648         }
649     }
650 }
651
652 static void free_subpicture(SubPicture *sp)
653 {
654     avsubtitle_free(&sp->sub);
655 }
656
657 static void video_image_display(VideoState *is)
658 {
659     VideoPicture *vp;
660     SubPicture *sp;
661     AVPicture pict;
662     float aspect_ratio;
663     int width, height, x, y;
664     SDL_Rect rect;
665     int i;
666
667     vp = &is->pictq[is->pictq_rindex];
668     if (vp->bmp) {
669 #if CONFIG_AVFILTER
670          if (vp->picref->video->sample_aspect_ratio.num == 0)
671              aspect_ratio = 0;
672          else
673              aspect_ratio = av_q2d(vp->picref->video->sample_aspect_ratio);
674 #else
675
676         /* XXX: use variable in the frame */
677         if (is->video_st->sample_aspect_ratio.num)
678             aspect_ratio = av_q2d(is->video_st->sample_aspect_ratio);
679         else if (is->video_st->codec->sample_aspect_ratio.num)
680             aspect_ratio = av_q2d(is->video_st->codec->sample_aspect_ratio);
681         else
682             aspect_ratio = 0;
683 #endif
684         if (aspect_ratio <= 0.0)
685             aspect_ratio = 1.0;
686         aspect_ratio *= (float)vp->width / (float)vp->height;
687
688         if (is->subtitle_st) {
689             if (is->subpq_size > 0) {
690                 sp = &is->subpq[is->subpq_rindex];
691
692                 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000)) {
693                     SDL_LockYUVOverlay (vp->bmp);
694
695                     pict.data[0] = vp->bmp->pixels[0];
696                     pict.data[1] = vp->bmp->pixels[2];
697                     pict.data[2] = vp->bmp->pixels[1];
698
699                     pict.linesize[0] = vp->bmp->pitches[0];
700                     pict.linesize[1] = vp->bmp->pitches[2];
701                     pict.linesize[2] = vp->bmp->pitches[1];
702
703                     for (i = 0; i < sp->sub.num_rects; i++)
704                         blend_subrect(&pict, sp->sub.rects[i],
705                                       vp->bmp->w, vp->bmp->h);
706
707                     SDL_UnlockYUVOverlay (vp->bmp);
708                 }
709             }
710         }
711
712
713         /* XXX: we suppose the screen has a 1.0 pixel ratio */
714         height = is->height;
715         width = ((int)rint(height * aspect_ratio)) & ~1;
716         if (width > is->width) {
717             width = is->width;
718             height = ((int)rint(width / aspect_ratio)) & ~1;
719         }
720         x = (is->width - width) / 2;
721         y = (is->height - height) / 2;
722         is->no_background = 0;
723         rect.x = is->xleft + x;
724         rect.y = is->ytop  + y;
725         rect.w = FFMAX(width,  1);
726         rect.h = FFMAX(height, 1);
727         SDL_DisplayYUVOverlay(vp->bmp, &rect);
728     }
729 }
730
731 static inline int compute_mod(int a, int b)
732 {
733     return a < 0 ? a%b + b : a%b;
734 }
735
736 static void video_audio_display(VideoState *s)
737 {
738     int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
739     int ch, channels, h, h2, bgcolor, fgcolor;
740     int16_t time_diff;
741     int rdft_bits, nb_freq;
742
743     for (rdft_bits = 1; (1 << rdft_bits) < 2 * s->height; rdft_bits++)
744         ;
745     nb_freq = 1 << (rdft_bits - 1);
746
747     /* compute display index : center on currently output samples */
748     channels = s->audio_tgt_channels;
749     nb_display_channels = channels;
750     if (!s->paused) {
751         int data_used= s->show_mode == SHOW_MODE_WAVES ? s->width : (2*nb_freq);
752         n = 2 * channels;
753         delay = s->audio_write_buf_size;
754         delay /= n;
755
756         /* to be more precise, we take into account the time spent since
757            the last buffer computation */
758         if (audio_callback_time) {
759             time_diff = av_gettime() - audio_callback_time;
760             delay -= (time_diff * s->audio_tgt_freq) / 1000000;
761         }
762
763         delay += 2 * data_used;
764         if (delay < data_used)
765             delay = data_used;
766
767         i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
768         if (s->show_mode == SHOW_MODE_WAVES) {
769             h = INT_MIN;
770             for (i = 0; i < 1000; i += channels) {
771                 int idx = (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
772                 int a = s->sample_array[idx];
773                 int b = s->sample_array[(idx + 4 * channels) % SAMPLE_ARRAY_SIZE];
774                 int c = s->sample_array[(idx + 5 * channels) % SAMPLE_ARRAY_SIZE];
775                 int d = s->sample_array[(idx + 9 * channels) % SAMPLE_ARRAY_SIZE];
776                 int score = a - d;
777                 if (h < score && (b ^ c) < 0) {
778                     h = score;
779                     i_start = idx;
780                 }
781             }
782         }
783
784         s->last_i_start = i_start;
785     } else {
786         i_start = s->last_i_start;
787     }
788
789     bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
790     if (s->show_mode == SHOW_MODE_WAVES) {
791         fill_rectangle(screen,
792                        s->xleft, s->ytop, s->width, s->height,
793                        bgcolor);
794
795         fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
796
797         /* total height for one channel */
798         h = s->height / nb_display_channels;
799         /* graph height / 2 */
800         h2 = (h * 9) / 20;
801         for (ch = 0; ch < nb_display_channels; ch++) {
802             i = i_start + ch;
803             y1 = s->ytop + ch * h + (h / 2); /* position of center line */
804             for (x = 0; x < s->width; x++) {
805                 y = (s->sample_array[i] * h2) >> 15;
806                 if (y < 0) {
807                     y = -y;
808                     ys = y1 - y;
809                 } else {
810                     ys = y1;
811                 }
812                 fill_rectangle(screen,
813                                s->xleft + x, ys, 1, y,
814                                fgcolor);
815                 i += channels;
816                 if (i >= SAMPLE_ARRAY_SIZE)
817                     i -= SAMPLE_ARRAY_SIZE;
818             }
819         }
820
821         fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
822
823         for (ch = 1; ch < nb_display_channels; ch++) {
824             y = s->ytop + ch * h;
825             fill_rectangle(screen,
826                            s->xleft, y, s->width, 1,
827                            fgcolor);
828         }
829         SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
830     } else {
831         nb_display_channels= FFMIN(nb_display_channels, 2);
832         if (rdft_bits != s->rdft_bits) {
833             av_rdft_end(s->rdft);
834             av_free(s->rdft_data);
835             s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
836             s->rdft_bits = rdft_bits;
837             s->rdft_data = av_malloc(4 * nb_freq * sizeof(*s->rdft_data));
838         }
839         {
840             FFTSample *data[2];
841             for (ch = 0; ch < nb_display_channels; ch++) {
842                 data[ch] = s->rdft_data + 2 * nb_freq * ch;
843                 i = i_start + ch;
844                 for (x = 0; x < 2 * nb_freq; x++) {
845                     double w = (x-nb_freq) * (1.0 / nb_freq);
846                     data[ch][x] = s->sample_array[i] * (1.0 - w * w);
847                     i += channels;
848                     if (i >= SAMPLE_ARRAY_SIZE)
849                         i -= SAMPLE_ARRAY_SIZE;
850                 }
851                 av_rdft_calc(s->rdft, data[ch]);
852             }
853             // least efficient way to do this, we should of course directly access it but its more than fast enough
854             for (y = 0; y < s->height; y++) {
855                 double w = 1 / sqrt(nb_freq);
856                 int a = sqrt(w * sqrt(data[0][2 * y + 0] * data[0][2 * y + 0] + data[0][2 * y + 1] * data[0][2 * y + 1]));
857                 int b = (nb_display_channels == 2 ) ? sqrt(w * sqrt(data[1][2 * y + 0] * data[1][2 * y + 0]
858                        + data[1][2 * y + 1] * data[1][2 * y + 1])) : a;
859                 a = FFMIN(a, 255);
860                 b = FFMIN(b, 255);
861                 fgcolor = SDL_MapRGB(screen->format, a, b, (a + b) / 2);
862
863                 fill_rectangle(screen,
864                             s->xpos, s->height-y, 1, 1,
865                             fgcolor);
866             }
867         }
868         SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
869         if (!s->paused)
870             s->xpos++;
871         if (s->xpos >= s->width)
872             s->xpos= s->xleft;
873     }
874 }
875
876 static void stream_close(VideoState *is)
877 {
878     VideoPicture *vp;
879     int i;
880     /* XXX: use a special url_shutdown call to abort parse cleanly */
881     is->abort_request = 1;
882     SDL_WaitThread(is->read_tid, NULL);
883     SDL_WaitThread(is->refresh_tid, NULL);
884
885     /* free all pictures */
886     for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
887         vp = &is->pictq[i];
888 #if CONFIG_AVFILTER
889         if (vp->picref) {
890             avfilter_unref_buffer(vp->picref);
891             vp->picref = NULL;
892         }
893 #endif
894         if (vp->bmp) {
895             SDL_FreeYUVOverlay(vp->bmp);
896             vp->bmp = NULL;
897         }
898     }
899     SDL_DestroyMutex(is->pictq_mutex);
900     SDL_DestroyCond(is->pictq_cond);
901     SDL_DestroyMutex(is->subpq_mutex);
902     SDL_DestroyCond(is->subpq_cond);
903 #if !CONFIG_AVFILTER
904     if (is->img_convert_ctx)
905         sws_freeContext(is->img_convert_ctx);
906 #endif
907     av_free(is);
908 }
909
910 static void do_exit(VideoState *is)
911 {
912     if (is) {
913         stream_close(is);
914     }
915     av_lockmgr_register(NULL);
916     uninit_opts();
917 #if CONFIG_AVFILTER
918     avfilter_uninit();
919 #endif
920     avformat_network_deinit();
921     if (show_status)
922         printf("\n");
923     SDL_Quit();
924     av_log(NULL, AV_LOG_QUIET, "%s", "");
925     exit(0);
926 }
927
928 static void sigterm_handler(int sig)
929 {
930     exit(123);
931 }
932
933 static int video_open(VideoState *is, int force_set_video_mode)
934 {
935     int flags = SDL_HWSURFACE | SDL_ASYNCBLIT | SDL_HWACCEL;
936     int w,h;
937
938     if (is_full_screen) flags |= SDL_FULLSCREEN;
939     else                flags |= SDL_RESIZABLE;
940
941     if (is_full_screen && fs_screen_width) {
942         w = fs_screen_width;
943         h = fs_screen_height;
944     } else if (!is_full_screen && screen_width) {
945         w = screen_width;
946         h = screen_height;
947 #if CONFIG_AVFILTER
948     } else if (is->out_video_filter && is->out_video_filter->inputs[0]) {
949         w = is->out_video_filter->inputs[0]->w;
950         h = is->out_video_filter->inputs[0]->h;
951 #else
952     } else if (is->video_st && is->video_st->codec->width) {
953         w = is->video_st->codec->width;
954         h = is->video_st->codec->height;
955 #endif
956     } else {
957         w = 640;
958         h = 480;
959     }
960     if (screen && is->width == screen->w && screen->w == w
961        && is->height== screen->h && screen->h == h && !force_set_video_mode)
962         return 0;
963     screen = SDL_SetVideoMode(w, h, 0, flags);
964     if (!screen) {
965         fprintf(stderr, "SDL: could not set video mode - exiting\n");
966         do_exit(is);
967     }
968     if (!window_title)
969         window_title = input_filename;
970     SDL_WM_SetCaption(window_title, window_title);
971
972     is->width  = screen->w;
973     is->height = screen->h;
974
975     return 0;
976 }
977
978 /* display the current picture, if any */
979 static void video_display(VideoState *is)
980 {
981     if (!screen)
982         video_open(is, 0);
983     if (is->audio_st && is->show_mode != SHOW_MODE_VIDEO)
984         video_audio_display(is);
985     else if (is->video_st)
986         video_image_display(is);
987 }
988
989 static int refresh_thread(void *opaque)
990 {
991     VideoState *is= opaque;
992     while (!is->abort_request) {
993         SDL_Event event;
994         event.type = FF_REFRESH_EVENT;
995         event.user.data1 = opaque;
996         if (!is->refresh && (!is->paused || is->force_refresh)) {
997             is->refresh = 1;
998             SDL_PushEvent(&event);
999         }
1000         //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
1001         usleep(is->audio_st && is->show_mode != SHOW_MODE_VIDEO ? rdftspeed*1000 : 5000);
1002     }
1003     return 0;
1004 }
1005
1006 /* get the current audio clock value */
1007 static double get_audio_clock(VideoState *is)
1008 {
1009     if (is->paused) {
1010         return is->audio_current_pts;
1011     } else {
1012         return is->audio_current_pts_drift + av_gettime() / 1000000.0;
1013     }
1014 }
1015
1016 /* get the current video clock value */
1017 static double get_video_clock(VideoState *is)
1018 {
1019     if (is->paused) {
1020         return is->video_current_pts;
1021     } else {
1022         return is->video_current_pts_drift + av_gettime() / 1000000.0;
1023     }
1024 }
1025
1026 /* get the current external clock value */
1027 static double get_external_clock(VideoState *is)
1028 {
1029     int64_t ti;
1030     ti = av_gettime();
1031     return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1032 }
1033
1034 /* get the current master clock value */
1035 static double get_master_clock(VideoState *is)
1036 {
1037     double val;
1038
1039     if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1040         if (is->video_st)
1041             val = get_video_clock(is);
1042         else
1043             val = get_audio_clock(is);
1044     } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1045         if (is->audio_st)
1046             val = get_audio_clock(is);
1047         else
1048             val = get_video_clock(is);
1049     } else {
1050         val = get_external_clock(is);
1051     }
1052     return val;
1053 }
1054
1055 /* seek in the stream */
1056 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1057 {
1058     if (!is->seek_req) {
1059         is->seek_pos = pos;
1060         is->seek_rel = rel;
1061         is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1062         if (seek_by_bytes)
1063             is->seek_flags |= AVSEEK_FLAG_BYTE;
1064         is->seek_req = 1;
1065     }
1066 }
1067
1068 /* pause or resume the video */
1069 static void stream_toggle_pause(VideoState *is)
1070 {
1071     if (is->paused) {
1072         is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1073         if (is->read_pause_return != AVERROR(ENOSYS)) {
1074             is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1075         }
1076         is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1077     }
1078     is->paused = !is->paused;
1079 }
1080
1081 static double compute_target_delay(double delay, VideoState *is)
1082 {
1083     double sync_threshold, diff;
1084
1085     /* update delay to follow master synchronisation source */
1086     if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1087          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1088         /* if video is slave, we try to correct big delays by
1089            duplicating or deleting a frame */
1090         diff = get_video_clock(is) - get_master_clock(is);
1091
1092         /* skip or repeat frame. We take into account the
1093            delay to compute the threshold. I still don't know
1094            if it is the best guess */
1095         sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1096         if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1097             if (diff <= -sync_threshold)
1098                 delay = 0;
1099             else if (diff >= sync_threshold)
1100                 delay = 2 * delay;
1101         }
1102     }
1103
1104     av_dlog(NULL, "video: delay=%0.3f A-V=%f\n",
1105             delay, -diff);
1106
1107     return delay;
1108 }
1109
1110 static void pictq_next_picture(VideoState *is) {
1111     /* update queue size and signal for next picture */
1112     if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1113         is->pictq_rindex = 0;
1114
1115     SDL_LockMutex(is->pictq_mutex);
1116     is->pictq_size--;
1117     SDL_CondSignal(is->pictq_cond);
1118     SDL_UnlockMutex(is->pictq_mutex);
1119 }
1120
1121 static void update_video_pts(VideoState *is, double pts, int64_t pos) {
1122     double time = av_gettime() / 1000000.0;
1123     /* update current video pts */
1124     is->video_current_pts = pts;
1125     is->video_current_pts_drift = is->video_current_pts - time;
1126     is->video_current_pos = pos;
1127     is->frame_last_pts = pts;
1128 }
1129
1130 /* called to display each frame */
1131 static void video_refresh(void *opaque)
1132 {
1133     VideoState *is = opaque;
1134     VideoPicture *vp;
1135     double time;
1136
1137     SubPicture *sp, *sp2;
1138
1139     if (is->video_st) {
1140 retry:
1141         if (is->pictq_size == 0) {
1142             SDL_LockMutex(is->pictq_mutex);
1143             if (is->frame_last_dropped_pts != AV_NOPTS_VALUE && is->frame_last_dropped_pts > is->frame_last_pts) {
1144                 update_video_pts(is, is->frame_last_dropped_pts, is->frame_last_dropped_pos);
1145                 is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1146             }
1147             SDL_UnlockMutex(is->pictq_mutex);
1148             // nothing to do, no picture to display in the que
1149         } else {
1150             double last_duration, duration, delay;
1151             /* dequeue the picture */
1152             vp = &is->pictq[is->pictq_rindex];
1153
1154             if (vp->skip) {
1155                 pictq_next_picture(is);
1156                 goto retry;
1157             }
1158
1159             if (is->paused)
1160                 goto display;
1161
1162             /* compute nominal last_duration */
1163             last_duration = vp->pts - is->frame_last_pts;
1164             if (last_duration > 0 && last_duration < 10.0) {
1165                 /* if duration of the last frame was sane, update last_duration in video state */
1166                 is->frame_last_duration = last_duration;
1167             }
1168             delay = compute_target_delay(is->frame_last_duration, is);
1169
1170             time= av_gettime()/1000000.0;
1171             if (time < is->frame_timer + delay)
1172                 return;
1173
1174             if (delay > 0)
1175                 is->frame_timer += delay * FFMAX(1, floor((time-is->frame_timer) / delay));
1176
1177             SDL_LockMutex(is->pictq_mutex);
1178             update_video_pts(is, vp->pts, vp->pos);
1179             SDL_UnlockMutex(is->pictq_mutex);
1180
1181             if (is->pictq_size > 1) {
1182                 VideoPicture *nextvp = &is->pictq[(is->pictq_rindex + 1) % VIDEO_PICTURE_QUEUE_SIZE];
1183                 duration = nextvp->pts - vp->pts; // More accurate this way, 1/time_base is often not reflecting FPS
1184             } else {
1185                 duration = vp->duration;
1186             }
1187
1188             if((framedrop>0 || (framedrop && is->audio_st)) && time > is->frame_timer + duration){
1189                 if(is->pictq_size > 1){
1190                     is->frame_drops_late++;
1191                     pictq_next_picture(is);
1192                     goto retry;
1193                 }
1194             }
1195
1196             if (is->subtitle_st) {
1197                 if (is->subtitle_stream_changed) {
1198                     SDL_LockMutex(is->subpq_mutex);
1199
1200                     while (is->subpq_size) {
1201                         free_subpicture(&is->subpq[is->subpq_rindex]);
1202
1203                         /* update queue size and signal for next picture */
1204                         if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1205                             is->subpq_rindex = 0;
1206
1207                         is->subpq_size--;
1208                     }
1209                     is->subtitle_stream_changed = 0;
1210
1211                     SDL_CondSignal(is->subpq_cond);
1212                     SDL_UnlockMutex(is->subpq_mutex);
1213                 } else {
1214                     if (is->subpq_size > 0) {
1215                         sp = &is->subpq[is->subpq_rindex];
1216
1217                         if (is->subpq_size > 1)
1218                             sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1219                         else
1220                             sp2 = NULL;
1221
1222                         if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1223                                 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1224                         {
1225                             free_subpicture(sp);
1226
1227                             /* update queue size and signal for next picture */
1228                             if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1229                                 is->subpq_rindex = 0;
1230
1231                             SDL_LockMutex(is->subpq_mutex);
1232                             is->subpq_size--;
1233                             SDL_CondSignal(is->subpq_cond);
1234                             SDL_UnlockMutex(is->subpq_mutex);
1235                         }
1236                     }
1237                 }
1238             }
1239
1240 display:
1241             /* display picture */
1242             if (!display_disable)
1243                 video_display(is);
1244
1245             if (!is->paused)
1246                 pictq_next_picture(is);
1247         }
1248     } else if (is->audio_st) {
1249         /* draw the next audio frame */
1250
1251         /* if only audio stream, then display the audio bars (better
1252            than nothing, just to test the implementation */
1253
1254         /* display picture */
1255         if (!display_disable)
1256             video_display(is);
1257     }
1258     is->force_refresh = 0;
1259     if (show_status) {
1260         static int64_t last_time;
1261         int64_t cur_time;
1262         int aqsize, vqsize, sqsize;
1263         double av_diff;
1264
1265         cur_time = av_gettime();
1266         if (!last_time || (cur_time - last_time) >= 30000) {
1267             aqsize = 0;
1268             vqsize = 0;
1269             sqsize = 0;
1270             if (is->audio_st)
1271                 aqsize = is->audioq.size;
1272             if (is->video_st)
1273                 vqsize = is->videoq.size;
1274             if (is->subtitle_st)
1275                 sqsize = is->subtitleq.size;
1276             av_diff = 0;
1277             if (is->audio_st && is->video_st)
1278                 av_diff = get_audio_clock(is) - get_video_clock(is);
1279             printf("%7.2f A-V:%7.3f fd=%4d aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64"   \r",
1280                    get_master_clock(is),
1281                    av_diff,
1282                    is->frame_drops_early + is->frame_drops_late,
1283                    aqsize / 1024,
1284                    vqsize / 1024,
1285                    sqsize,
1286                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_dts : 0,
1287                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_pts : 0);
1288             fflush(stdout);
1289             last_time = cur_time;
1290         }
1291     }
1292 }
1293
1294 /* allocate a picture (needs to do that in main thread to avoid
1295    potential locking problems */
1296 static void alloc_picture(void *opaque)
1297 {
1298     VideoState *is = opaque;
1299     VideoPicture *vp;
1300
1301     vp = &is->pictq[is->pictq_windex];
1302
1303     if (vp->bmp)
1304         SDL_FreeYUVOverlay(vp->bmp);
1305
1306 #if CONFIG_AVFILTER
1307     if (vp->picref)
1308         avfilter_unref_buffer(vp->picref);
1309     vp->picref = NULL;
1310
1311     vp->width   = is->out_video_filter->inputs[0]->w;
1312     vp->height  = is->out_video_filter->inputs[0]->h;
1313     vp->pix_fmt = is->out_video_filter->inputs[0]->format;
1314 #else
1315     vp->width   = is->video_st->codec->width;
1316     vp->height  = is->video_st->codec->height;
1317     vp->pix_fmt = is->video_st->codec->pix_fmt;
1318 #endif
1319
1320     vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1321                                    SDL_YV12_OVERLAY,
1322                                    screen);
1323     if (!vp->bmp || vp->bmp->pitches[0] < vp->width) {
1324         /* SDL allocates a buffer smaller than requested if the video
1325          * overlay hardware is unable to support the requested size. */
1326         fprintf(stderr, "Error: the video system does not support an image\n"
1327                         "size of %dx%d pixels. Try using -lowres or -vf \"scale=w:h\"\n"
1328                         "to reduce the image size.\n", vp->width, vp->height );
1329         do_exit(is);
1330     }
1331
1332     SDL_LockMutex(is->pictq_mutex);
1333     vp->allocated = 1;
1334     SDL_CondSignal(is->pictq_cond);
1335     SDL_UnlockMutex(is->pictq_mutex);
1336 }
1337
1338 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1339 {
1340     VideoPicture *vp;
1341     double frame_delay, pts = pts1;
1342
1343     /* compute the exact PTS for the picture if it is omitted in the stream
1344      * pts1 is the dts of the pkt / pts of the frame */
1345     if (pts != 0) {
1346         /* update video clock with pts, if present */
1347         is->video_clock = pts;
1348     } else {
1349         pts = is->video_clock;
1350     }
1351     /* update video clock for next frame */
1352     frame_delay = av_q2d(is->video_st->codec->time_base);
1353     /* for MPEG2, the frame can be repeated, so we update the
1354        clock accordingly */
1355     frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1356     is->video_clock += frame_delay;
1357
1358 #if defined(DEBUG_SYNC) && 0
1359     printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1360            av_get_picture_type_char(src_frame->pict_type), pts, pts1);
1361 #endif
1362
1363     /* wait until we have space to put a new picture */
1364     SDL_LockMutex(is->pictq_mutex);
1365
1366     while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1367            !is->videoq.abort_request) {
1368         SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1369     }
1370     SDL_UnlockMutex(is->pictq_mutex);
1371
1372     if (is->videoq.abort_request)
1373         return -1;
1374
1375     vp = &is->pictq[is->pictq_windex];
1376
1377     vp->duration = frame_delay;
1378
1379     /* alloc or resize hardware picture buffer */
1380     if (!vp->bmp || vp->reallocate ||
1381 #if CONFIG_AVFILTER
1382         vp->width  != is->out_video_filter->inputs[0]->w ||
1383         vp->height != is->out_video_filter->inputs[0]->h) {
1384 #else
1385         vp->width != is->video_st->codec->width ||
1386         vp->height != is->video_st->codec->height) {
1387 #endif
1388         SDL_Event event;
1389
1390         vp->allocated  = 0;
1391         vp->reallocate = 0;
1392
1393         /* the allocation must be done in the main thread to avoid
1394            locking problems */
1395         event.type = FF_ALLOC_EVENT;
1396         event.user.data1 = is;
1397         SDL_PushEvent(&event);
1398
1399         /* wait until the picture is allocated */
1400         SDL_LockMutex(is->pictq_mutex);
1401         while (!vp->allocated && !is->videoq.abort_request) {
1402             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1403         }
1404         /* if the queue is aborted, we have to pop the pending ALLOC event or wait for the allocation to complete */
1405         if (is->videoq.abort_request && SDL_PeepEvents(&event, 1, SDL_GETEVENT, SDL_EVENTMASK(FF_ALLOC_EVENT)) != 1) {
1406             while (!vp->allocated) {
1407                 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1408             }
1409         }
1410         SDL_UnlockMutex(is->pictq_mutex);
1411
1412         if (is->videoq.abort_request)
1413             return -1;
1414     }
1415
1416     /* if the frame is not skipped, then display it */
1417     if (vp->bmp) {
1418         AVPicture pict = { { 0 } };
1419 #if CONFIG_AVFILTER
1420         if (vp->picref)
1421             avfilter_unref_buffer(vp->picref);
1422         vp->picref = src_frame->opaque;
1423 #endif
1424
1425         /* get a pointer on the bitmap */
1426         SDL_LockYUVOverlay (vp->bmp);
1427
1428         pict.data[0] = vp->bmp->pixels[0];
1429         pict.data[1] = vp->bmp->pixels[2];
1430         pict.data[2] = vp->bmp->pixels[1];
1431
1432         pict.linesize[0] = vp->bmp->pitches[0];
1433         pict.linesize[1] = vp->bmp->pitches[2];
1434         pict.linesize[2] = vp->bmp->pitches[1];
1435
1436 #if CONFIG_AVFILTER
1437         // FIXME use direct rendering
1438         av_picture_copy(&pict, (AVPicture *)src_frame,
1439                         vp->pix_fmt, vp->width, vp->height);
1440 #else
1441         sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1442         is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1443             vp->width, vp->height, vp->pix_fmt, vp->width, vp->height,
1444             PIX_FMT_YUV420P, sws_flags, NULL, NULL, NULL);
1445         if (is->img_convert_ctx == NULL) {
1446             fprintf(stderr, "Cannot initialize the conversion context\n");
1447             exit(1);
1448         }
1449         sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1450                   0, vp->height, pict.data, pict.linesize);
1451 #endif
1452         /* update the bitmap content */
1453         SDL_UnlockYUVOverlay(vp->bmp);
1454
1455         vp->pts = pts;
1456         vp->pos = pos;
1457         vp->skip = 0;
1458
1459         /* now we can update the picture count */
1460         if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1461             is->pictq_windex = 0;
1462         SDL_LockMutex(is->pictq_mutex);
1463         is->pictq_size++;
1464         SDL_UnlockMutex(is->pictq_mutex);
1465     }
1466     return 0;
1467 }
1468
1469 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1470 {
1471     int got_picture, i;
1472
1473     if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1474         return -1;
1475
1476     if (pkt->data == flush_pkt.data) {
1477         avcodec_flush_buffers(is->video_st->codec);
1478
1479         SDL_LockMutex(is->pictq_mutex);
1480         // Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1481         for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
1482             is->pictq[i].skip = 1;
1483         }
1484         while (is->pictq_size && !is->videoq.abort_request) {
1485             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1486         }
1487         is->video_current_pos = -1;
1488         is->frame_last_pts = AV_NOPTS_VALUE;
1489         is->frame_last_duration = 0;
1490         is->frame_timer = (double)av_gettime() / 1000000.0;
1491         is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1492         SDL_UnlockMutex(is->pictq_mutex);
1493
1494         return 0;
1495     }
1496
1497     avcodec_decode_video2(is->video_st->codec, frame, &got_picture, pkt);
1498
1499     if (got_picture) {
1500         int ret = 1;
1501
1502         if (decoder_reorder_pts == -1) {
1503             *pts = av_frame_get_best_effort_timestamp(frame);
1504         } else if (decoder_reorder_pts) {
1505             *pts = frame->pkt_pts;
1506         } else {
1507             *pts = frame->pkt_dts;
1508         }
1509
1510         if (*pts == AV_NOPTS_VALUE) {
1511             *pts = 0;
1512         }
1513
1514         if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) || is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK) &&
1515              (framedrop>0 || (framedrop && is->audio_st))) {
1516             SDL_LockMutex(is->pictq_mutex);
1517             if (is->frame_last_pts != AV_NOPTS_VALUE && *pts) {
1518                 double clockdiff = get_video_clock(is) - get_master_clock(is);
1519                 double dpts = av_q2d(is->video_st->time_base) * *pts;
1520                 double ptsdiff = dpts - is->frame_last_pts;
1521                 if (fabs(clockdiff) < AV_NOSYNC_THRESHOLD &&
1522                      ptsdiff > 0 && ptsdiff < AV_NOSYNC_THRESHOLD &&
1523                      clockdiff + ptsdiff - is->frame_last_filter_delay < 0) {
1524                     is->frame_last_dropped_pos = pkt->pos;
1525                     is->frame_last_dropped_pts = dpts;
1526                     is->frame_drops_early++;
1527                     ret = 0;
1528                 }
1529             }
1530             SDL_UnlockMutex(is->pictq_mutex);
1531         }
1532
1533         if (ret)
1534             is->frame_last_returned_time = av_gettime() / 1000000.0;
1535
1536         return ret;
1537     }
1538     return 0;
1539 }
1540
1541 #if CONFIG_AVFILTER
1542 typedef struct {
1543     VideoState *is;
1544     AVFrame *frame;
1545     int use_dr1;
1546 } FilterPriv;
1547
1548 static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
1549 {
1550     AVFilterContext *ctx = codec->opaque;
1551     AVFilterBufferRef  *ref;
1552     int perms = AV_PERM_WRITE;
1553     int i, w, h, stride[AV_NUM_DATA_POINTERS];
1554     unsigned edge;
1555     int pixel_size;
1556
1557     av_assert0(codec->flags & CODEC_FLAG_EMU_EDGE);
1558
1559     if (codec->codec->capabilities & CODEC_CAP_NEG_LINESIZES)
1560         perms |= AV_PERM_NEG_LINESIZES;
1561
1562     if (pic->buffer_hints & FF_BUFFER_HINTS_VALID) {
1563         if (pic->buffer_hints & FF_BUFFER_HINTS_READABLE) perms |= AV_PERM_READ;
1564         if (pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) perms |= AV_PERM_PRESERVE;
1565         if (pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) perms |= AV_PERM_REUSE2;
1566     }
1567     if (pic->reference) perms |= AV_PERM_READ | AV_PERM_PRESERVE;
1568
1569     w = codec->width;
1570     h = codec->height;
1571
1572     if(av_image_check_size(w, h, 0, codec) || codec->pix_fmt<0)
1573         return -1;
1574
1575     avcodec_align_dimensions2(codec, &w, &h, stride);
1576     edge = codec->flags & CODEC_FLAG_EMU_EDGE ? 0 : avcodec_get_edge_width();
1577     w += edge << 1;
1578     h += edge << 1;
1579     if (codec->pix_fmt != ctx->outputs[0]->format) {
1580         av_log(codec, AV_LOG_ERROR, "Pixel format mismatches %d %d\n", codec->pix_fmt, ctx->outputs[0]->format);
1581         return -1;
1582     }
1583     if (!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
1584         return -1;
1585
1586     pixel_size = av_pix_fmt_descriptors[ref->format].comp[0].step_minus1 + 1;
1587     ref->video->w = codec->width;
1588     ref->video->h = codec->height;
1589     for (i = 0; i < 4; i ++) {
1590         unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_w : 0;
1591         unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_h : 0;
1592
1593         if (ref->data[i]) {
1594             ref->data[i]    += ((edge * pixel_size) >> hshift) + ((edge * ref->linesize[i]) >> vshift);
1595         }
1596         pic->data[i]     = ref->data[i];
1597         pic->linesize[i] = ref->linesize[i];
1598     }
1599     pic->opaque = ref;
1600     pic->type   = FF_BUFFER_TYPE_USER;
1601     pic->reordered_opaque = codec->reordered_opaque;
1602     pic->width               = codec->width;
1603     pic->height              = codec->height;
1604     pic->format              = codec->pix_fmt;
1605     pic->sample_aspect_ratio = codec->sample_aspect_ratio;
1606     if (codec->pkt) pic->pkt_pts = codec->pkt->pts;
1607     else            pic->pkt_pts = AV_NOPTS_VALUE;
1608     return 0;
1609 }
1610
1611 static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
1612 {
1613     memset(pic->data, 0, sizeof(pic->data));
1614     avfilter_unref_buffer(pic->opaque);
1615 }
1616
1617 static int input_reget_buffer(AVCodecContext *codec, AVFrame *pic)
1618 {
1619     AVFilterBufferRef *ref = pic->opaque;
1620
1621     if (pic->data[0] == NULL) {
1622         pic->buffer_hints |= FF_BUFFER_HINTS_READABLE;
1623         return codec->get_buffer(codec, pic);
1624     }
1625
1626     if ((codec->width != ref->video->w) || (codec->height != ref->video->h) ||
1627         (codec->pix_fmt != ref->format)) {
1628         av_log(codec, AV_LOG_ERROR, "Picture properties changed.\n");
1629         return -1;
1630     }
1631
1632     pic->reordered_opaque = codec->reordered_opaque;
1633     if (codec->pkt) pic->pkt_pts = codec->pkt->pts;
1634     else            pic->pkt_pts = AV_NOPTS_VALUE;
1635     return 0;
1636 }
1637
1638 static int input_init(AVFilterContext *ctx, const char *args, void *opaque)
1639 {
1640     FilterPriv *priv = ctx->priv;
1641     AVCodecContext *codec;
1642     if (!opaque) return -1;
1643
1644     priv->is = opaque;
1645     codec    = priv->is->video_st->codec;
1646     codec->opaque = ctx;
1647     if (codec->codec->capabilities & CODEC_CAP_DR1) {
1648         av_assert0(codec->flags & CODEC_FLAG_EMU_EDGE);
1649         priv->use_dr1 = 1;
1650         codec->get_buffer     = input_get_buffer;
1651         codec->release_buffer = input_release_buffer;
1652         codec->reget_buffer   = input_reget_buffer;
1653         codec->thread_safe_callbacks = 1;
1654     }
1655
1656     priv->frame = avcodec_alloc_frame();
1657
1658     return 0;
1659 }
1660
1661 static void input_uninit(AVFilterContext *ctx)
1662 {
1663     FilterPriv *priv = ctx->priv;
1664     av_free(priv->frame);
1665 }
1666
1667 static int input_request_frame(AVFilterLink *link)
1668 {
1669     FilterPriv *priv = link->src->priv;
1670     AVFilterBufferRef *picref;
1671     int64_t pts = 0;
1672     AVPacket pkt;
1673     int ret;
1674
1675     while (!(ret = get_video_frame(priv->is, priv->frame, &pts, &pkt)))
1676         av_free_packet(&pkt);
1677     if (ret < 0)
1678         return -1;
1679
1680     if (priv->use_dr1 && priv->frame->opaque) {
1681         picref = avfilter_ref_buffer(priv->frame->opaque, ~0);
1682     } else {
1683         picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
1684         av_image_copy(picref->data, picref->linesize,
1685                       (const uint8_t **)(void **)priv->frame->data, priv->frame->linesize,
1686                       picref->format, link->w, link->h);
1687     }
1688     av_free_packet(&pkt);
1689
1690     avfilter_copy_frame_props(picref, priv->frame);
1691     picref->pts = pts;
1692
1693     avfilter_start_frame(link, picref);
1694     avfilter_draw_slice(link, 0, link->h, 1);
1695     avfilter_end_frame(link);
1696
1697     return 0;
1698 }
1699
1700 static int input_query_formats(AVFilterContext *ctx)
1701 {
1702     FilterPriv *priv = ctx->priv;
1703     enum PixelFormat pix_fmts[] = {
1704         priv->is->video_st->codec->pix_fmt, PIX_FMT_NONE
1705     };
1706
1707     avfilter_set_common_pixel_formats(ctx, avfilter_make_format_list(pix_fmts));
1708     return 0;
1709 }
1710
1711 static int input_config_props(AVFilterLink *link)
1712 {
1713     FilterPriv *priv  = link->src->priv;
1714     AVStream *s = priv->is->video_st;
1715
1716     link->w = s->codec->width;
1717     link->h = s->codec->height;
1718     link->sample_aspect_ratio = s->sample_aspect_ratio.num ?
1719         s->sample_aspect_ratio : s->codec->sample_aspect_ratio;
1720     link->time_base = s->time_base;
1721
1722     return 0;
1723 }
1724
1725 static AVFilter input_filter =
1726 {
1727     .name      = "ffplay_input",
1728
1729     .priv_size = sizeof(FilterPriv),
1730
1731     .init      = input_init,
1732     .uninit    = input_uninit,
1733
1734     .query_formats = input_query_formats,
1735
1736     .inputs    = (AVFilterPad[]) {{ .name = NULL }},
1737     .outputs   = (AVFilterPad[]) {{ .name = "default",
1738                                     .type = AVMEDIA_TYPE_VIDEO,
1739                                     .request_frame = input_request_frame,
1740                                     .config_props  = input_config_props, },
1741                                   { .name = NULL }},
1742 };
1743
1744 static int configure_video_filters(AVFilterGraph *graph, VideoState *is, const char *vfilters)
1745 {
1746     static const enum PixelFormat pix_fmts[] = { PIX_FMT_YUV420P, PIX_FMT_NONE };
1747     char sws_flags_str[128];
1748     int ret;
1749     AVBufferSinkParams *buffersink_params = av_buffersink_params_alloc();
1750     AVFilterContext *filt_src = NULL, *filt_out = NULL;
1751     snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1752     graph->scale_sws_opts = av_strdup(sws_flags_str);
1753
1754     if ((ret = avfilter_graph_create_filter(&filt_src, &input_filter, "src",
1755                                             NULL, is, graph)) < 0)
1756         return ret;
1757
1758 #if FF_API_OLD_VSINK_API
1759     ret = avfilter_graph_create_filter(&filt_out, avfilter_get_by_name("buffersink"), "out",
1760                                        NULL, pix_fmts, graph);
1761 #else
1762     buffersink_params->pixel_fmts = pix_fmts;
1763     ret = avfilter_graph_create_filter(&filt_out, avfilter_get_by_name("buffersink"), "out",
1764                                        NULL, buffersink_params, graph);
1765 #endif
1766     av_freep(&buffersink_params);
1767     if (ret < 0)
1768         return ret;
1769
1770     if (vfilters) {
1771         AVFilterInOut *outputs = avfilter_inout_alloc();
1772         AVFilterInOut *inputs  = avfilter_inout_alloc();
1773
1774         outputs->name    = av_strdup("in");
1775         outputs->filter_ctx = filt_src;
1776         outputs->pad_idx = 0;
1777         outputs->next    = NULL;
1778
1779         inputs->name    = av_strdup("out");
1780         inputs->filter_ctx = filt_out;
1781         inputs->pad_idx = 0;
1782         inputs->next    = NULL;
1783
1784         if ((ret = avfilter_graph_parse(graph, vfilters, &inputs, &outputs, NULL)) < 0)
1785             return ret;
1786     } else {
1787         if ((ret = avfilter_link(filt_src, 0, filt_out, 0)) < 0)
1788             return ret;
1789     }
1790
1791     if ((ret = avfilter_graph_config(graph, NULL)) < 0)
1792         return ret;
1793
1794     is->out_video_filter = filt_out;
1795
1796     return ret;
1797 }
1798
1799 #endif  /* CONFIG_AVFILTER */
1800
1801 static int video_thread(void *arg)
1802 {
1803     VideoState *is = arg;
1804     AVFrame *frame = avcodec_alloc_frame();
1805     int64_t pts_int = AV_NOPTS_VALUE, pos = -1;
1806     double pts;
1807     int ret;
1808
1809 #if CONFIG_AVFILTER
1810     AVFilterGraph *graph = avfilter_graph_alloc();
1811     AVFilterContext *filt_out = NULL;
1812     int last_w = is->video_st->codec->width;
1813     int last_h = is->video_st->codec->height;
1814
1815     if ((ret = configure_video_filters(graph, is, vfilters)) < 0)
1816         goto the_end;
1817     filt_out = is->out_video_filter;
1818 #endif
1819
1820     for (;;) {
1821 #if !CONFIG_AVFILTER
1822         AVPacket pkt;
1823 #else
1824         AVFilterBufferRef *picref;
1825         AVRational tb = filt_out->inputs[0]->time_base;
1826 #endif
1827         while (is->paused && !is->videoq.abort_request)
1828             SDL_Delay(10);
1829 #if CONFIG_AVFILTER
1830         if (   last_w != is->video_st->codec->width
1831             || last_h != is->video_st->codec->height) {
1832             av_log(NULL, AV_LOG_INFO, "Frame changed from size:%dx%d to size:%dx%d\n",
1833                    last_w, last_h, is->video_st->codec->width, is->video_st->codec->height);
1834             avfilter_graph_free(&graph);
1835             graph = avfilter_graph_alloc();
1836             if ((ret = configure_video_filters(graph, is, vfilters)) < 0)
1837                 goto the_end;
1838             filt_out = is->out_video_filter;
1839             last_w = is->video_st->codec->width;
1840             last_h = is->video_st->codec->height;
1841         }
1842         ret = av_buffersink_get_buffer_ref(filt_out, &picref, 0);
1843         if (picref) {
1844             avfilter_fill_frame_from_video_buffer_ref(frame, picref);
1845             pts_int = picref->pts;
1846             pos     = picref->pos;
1847             frame->opaque = picref;
1848         }
1849
1850         if (ret >= 0 && av_cmp_q(tb, is->video_st->time_base)) {
1851             av_unused int64_t pts1 = pts_int;
1852             pts_int = av_rescale_q(pts_int, tb, is->video_st->time_base);
1853             av_dlog(NULL, "video_thread(): "
1854                     "tb:%d/%d pts:%"PRId64" -> tb:%d/%d pts:%"PRId64"\n",
1855                     tb.num, tb.den, pts1,
1856                     is->video_st->time_base.num, is->video_st->time_base.den, pts_int);
1857         }
1858 #else
1859         ret = get_video_frame(is, frame, &pts_int, &pkt);
1860         pos = pkt.pos;
1861         av_free_packet(&pkt);
1862 #endif
1863
1864         if (ret < 0)
1865             goto the_end;
1866
1867         is->frame_last_filter_delay = av_gettime() / 1000000.0 - is->frame_last_returned_time;
1868         if (fabs(is->frame_last_filter_delay) > AV_NOSYNC_THRESHOLD / 10.0)
1869             is->frame_last_filter_delay = 0;
1870
1871 #if CONFIG_AVFILTER
1872         if (!picref)
1873             continue;
1874 #endif
1875
1876         pts = pts_int * av_q2d(is->video_st->time_base);
1877
1878         ret = queue_picture(is, frame, pts, pos);
1879
1880         if (ret < 0)
1881             goto the_end;
1882
1883         if (is->step)
1884             stream_toggle_pause(is);
1885     }
1886  the_end:
1887 #if CONFIG_AVFILTER
1888     av_freep(&vfilters);
1889     avfilter_graph_free(&graph);
1890 #endif
1891     av_free(frame);
1892     return 0;
1893 }
1894
1895 static int subtitle_thread(void *arg)
1896 {
1897     VideoState *is = arg;
1898     SubPicture *sp;
1899     AVPacket pkt1, *pkt = &pkt1;
1900     int got_subtitle;
1901     double pts;
1902     int i, j;
1903     int r, g, b, y, u, v, a;
1904
1905     for (;;) {
1906         while (is->paused && !is->subtitleq.abort_request) {
1907             SDL_Delay(10);
1908         }
1909         if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1910             break;
1911
1912         if (pkt->data == flush_pkt.data) {
1913             avcodec_flush_buffers(is->subtitle_st->codec);
1914             continue;
1915         }
1916         SDL_LockMutex(is->subpq_mutex);
1917         while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1918                !is->subtitleq.abort_request) {
1919             SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1920         }
1921         SDL_UnlockMutex(is->subpq_mutex);
1922
1923         if (is->subtitleq.abort_request)
1924             return 0;
1925
1926         sp = &is->subpq[is->subpq_windex];
1927
1928        /* NOTE: ipts is the PTS of the _first_ picture beginning in
1929            this packet, if any */
1930         pts = 0;
1931         if (pkt->pts != AV_NOPTS_VALUE)
1932             pts = av_q2d(is->subtitle_st->time_base) * pkt->pts;
1933
1934         avcodec_decode_subtitle2(is->subtitle_st->codec, &sp->sub,
1935                                  &got_subtitle, pkt);
1936
1937         if (got_subtitle && sp->sub.format == 0) {
1938             sp->pts = pts;
1939
1940             for (i = 0; i < sp->sub.num_rects; i++)
1941             {
1942                 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1943                 {
1944                     RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1945                     y = RGB_TO_Y_CCIR(r, g, b);
1946                     u = RGB_TO_U_CCIR(r, g, b, 0);
1947                     v = RGB_TO_V_CCIR(r, g, b, 0);
1948                     YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1949                 }
1950             }
1951
1952             /* now we can update the picture count */
1953             if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1954                 is->subpq_windex = 0;
1955             SDL_LockMutex(is->subpq_mutex);
1956             is->subpq_size++;
1957             SDL_UnlockMutex(is->subpq_mutex);
1958         }
1959         av_free_packet(pkt);
1960     }
1961     return 0;
1962 }
1963
1964 /* copy samples for viewing in editor window */
1965 static void update_sample_display(VideoState *is, short *samples, int samples_size)
1966 {
1967     int size, len;
1968
1969     size = samples_size / sizeof(short);
1970     while (size > 0) {
1971         len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
1972         if (len > size)
1973             len = size;
1974         memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
1975         samples += len;
1976         is->sample_array_index += len;
1977         if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
1978             is->sample_array_index = 0;
1979         size -= len;
1980     }
1981 }
1982
1983 /* return the wanted number of samples to get better sync if sync_type is video
1984  * or external master clock */
1985 static int synchronize_audio(VideoState *is, int nb_samples)
1986 {
1987     int wanted_nb_samples = nb_samples;
1988
1989     /* if not master, then we try to remove or add samples to correct the clock */
1990     if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
1991          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1992         double diff, avg_diff;
1993         int min_nb_samples, max_nb_samples;
1994
1995         diff = get_audio_clock(is) - get_master_clock(is);
1996
1997         if (diff < AV_NOSYNC_THRESHOLD) {
1998             is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
1999             if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
2000                 /* not enough measures to have a correct estimate */
2001                 is->audio_diff_avg_count++;
2002             } else {
2003                 /* estimate the A-V difference */
2004                 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
2005
2006                 if (fabs(avg_diff) >= is->audio_diff_threshold) {
2007                     wanted_nb_samples = nb_samples + (int)(diff * is->audio_src_freq);
2008                     min_nb_samples = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX) / 100));
2009                     max_nb_samples = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX) / 100));
2010                     wanted_nb_samples = FFMIN(FFMAX(wanted_nb_samples, min_nb_samples), max_nb_samples);
2011                 }
2012                 av_dlog(NULL, "diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
2013                         diff, avg_diff, wanted_nb_samples - nb_samples,
2014                         is->audio_clock, is->video_clock, is->audio_diff_threshold);
2015             }
2016         } else {
2017             /* too big difference : may be initial PTS errors, so
2018                reset A-V filter */
2019             is->audio_diff_avg_count = 0;
2020             is->audio_diff_cum       = 0;
2021         }
2022     }
2023
2024     return wanted_nb_samples;
2025 }
2026
2027 /* decode one audio frame and returns its uncompressed size */
2028 static int audio_decode_frame(VideoState *is, double *pts_ptr)
2029 {
2030     AVPacket *pkt_temp = &is->audio_pkt_temp;
2031     AVPacket *pkt = &is->audio_pkt;
2032     AVCodecContext *dec = is->audio_st->codec;
2033     int len1, len2, data_size, resampled_data_size;
2034     int64_t dec_channel_layout;
2035     int got_frame;
2036     double pts;
2037     int new_packet = 0;
2038     int flush_complete = 0;
2039     int wanted_nb_samples;
2040
2041     for (;;) {
2042         /* NOTE: the audio packet can contain several frames */
2043         while (pkt_temp->size > 0 || (!pkt_temp->data && new_packet)) {
2044             if (!is->frame) {
2045                 if (!(is->frame = avcodec_alloc_frame()))
2046                     return AVERROR(ENOMEM);
2047             } else
2048                 avcodec_get_frame_defaults(is->frame);
2049
2050             if (flush_complete)
2051                 break;
2052             new_packet = 0;
2053             len1 = avcodec_decode_audio4(dec, is->frame, &got_frame, pkt_temp);
2054             if (len1 < 0) {
2055                 /* if error, we skip the frame */
2056                 pkt_temp->size = 0;
2057                 break;
2058             }
2059
2060             pkt_temp->data += len1;
2061             pkt_temp->size -= len1;
2062
2063             if (!got_frame) {
2064                 /* stop sending empty packets if the decoder is finished */
2065                 if (!pkt_temp->data && dec->codec->capabilities & CODEC_CAP_DELAY)
2066                     flush_complete = 1;
2067                 continue;
2068             }
2069             data_size = av_samples_get_buffer_size(NULL, dec->channels,
2070                                                    is->frame->nb_samples,
2071                                                    dec->sample_fmt, 1);
2072
2073             dec_channel_layout = (dec->channel_layout && dec->channels == av_get_channel_layout_nb_channels(dec->channel_layout)) ? dec->channel_layout : av_get_default_channel_layout(dec->channels);
2074             wanted_nb_samples = synchronize_audio(is, is->frame->nb_samples);
2075
2076             if (dec->sample_fmt != is->audio_src_fmt ||
2077                 dec_channel_layout != is->audio_src_channel_layout ||
2078                 dec->sample_rate != is->audio_src_freq ||
2079                 (wanted_nb_samples != is->frame->nb_samples && !is->swr_ctx)) {
2080                 if (is->swr_ctx)
2081                     swr_free(&is->swr_ctx);
2082                 is->swr_ctx = swr_alloc_set_opts(NULL,
2083                                                  is->audio_tgt_channel_layout, is->audio_tgt_fmt, is->audio_tgt_freq,
2084                                                  dec_channel_layout,           dec->sample_fmt,   dec->sample_rate,
2085                                                  0, NULL);
2086                 if (!is->swr_ctx || swr_init(is->swr_ctx) < 0) {
2087                     fprintf(stderr, "Cannot create sample rate converter for conversion of %d Hz %s %d channels to %d Hz %s %d channels!\n",
2088                         dec->sample_rate,
2089                         av_get_sample_fmt_name(dec->sample_fmt),
2090                         dec->channels,
2091                         is->audio_tgt_freq,
2092                         av_get_sample_fmt_name(is->audio_tgt_fmt),
2093                         is->audio_tgt_channels);
2094                     break;
2095                 }
2096                 is->audio_src_channel_layout = dec_channel_layout;
2097                 is->audio_src_channels = dec->channels;
2098                 is->audio_src_freq = dec->sample_rate;
2099                 is->audio_src_fmt = dec->sample_fmt;
2100             }
2101
2102             resampled_data_size = data_size;
2103             if (is->swr_ctx) {
2104                 const uint8_t *in[] = { is->frame->data[0] };
2105                 uint8_t *out[] = {is->audio_buf2};
2106                 if (wanted_nb_samples != is->frame->nb_samples) {
2107                     if (swr_set_compensation(is->swr_ctx, (wanted_nb_samples - is->frame->nb_samples) * is->audio_tgt_freq / dec->sample_rate,
2108                                                 wanted_nb_samples * is->audio_tgt_freq / dec->sample_rate) < 0) {
2109                         fprintf(stderr, "swr_set_compensation() failed\n");
2110                         break;
2111                     }
2112                 }
2113                 len2 = swr_convert(is->swr_ctx, out, sizeof(is->audio_buf2) / is->audio_tgt_channels / av_get_bytes_per_sample(is->audio_tgt_fmt),
2114                                                 in, is->frame->nb_samples);
2115                 if (len2 < 0) {
2116                     fprintf(stderr, "audio_resample() failed\n");
2117                     break;
2118                 }
2119                 if (len2 == sizeof(is->audio_buf2) / is->audio_tgt_channels / av_get_bytes_per_sample(is->audio_tgt_fmt)) {
2120                     fprintf(stderr, "warning: audio buffer is probably too small\n");
2121                     swr_init(is->swr_ctx);
2122                 }
2123                 is->audio_buf = is->audio_buf2;
2124                 resampled_data_size = len2 * is->audio_tgt_channels * av_get_bytes_per_sample(is->audio_tgt_fmt);
2125             } else {
2126                 is->audio_buf = is->frame->data[0];
2127             }
2128
2129             /* if no pts, then compute it */
2130             pts = is->audio_clock;
2131             *pts_ptr = pts;
2132             is->audio_clock += (double)data_size /
2133                 (dec->channels * dec->sample_rate * av_get_bytes_per_sample(dec->sample_fmt));
2134 #ifdef DEBUG
2135             {
2136                 static double last_clock;
2137                 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2138                        is->audio_clock - last_clock,
2139                        is->audio_clock, pts);
2140                 last_clock = is->audio_clock;
2141             }
2142 #endif
2143             return resampled_data_size;
2144         }
2145
2146         /* free the current packet */
2147         if (pkt->data)
2148             av_free_packet(pkt);
2149         memset(pkt_temp, 0, sizeof(*pkt_temp));
2150
2151         if (is->paused || is->audioq.abort_request) {
2152             return -1;
2153         }
2154
2155         /* read next packet */
2156         if ((new_packet = packet_queue_get(&is->audioq, pkt, 1)) < 0)
2157             return -1;
2158
2159         if (pkt->data == flush_pkt.data) {
2160             avcodec_flush_buffers(dec);
2161             flush_complete = 0;
2162         }
2163
2164         *pkt_temp = *pkt;
2165
2166         /* if update the audio clock with the pts */
2167         if (pkt->pts != AV_NOPTS_VALUE) {
2168             is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2169         }
2170     }
2171 }
2172
2173 /* prepare a new audio buffer */
2174 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2175 {
2176     VideoState *is = opaque;
2177     int audio_size, len1;
2178     int bytes_per_sec;
2179     int frame_size = av_samples_get_buffer_size(NULL, is->audio_tgt_channels, 1, is->audio_tgt_fmt, 1);
2180     double pts;
2181
2182     audio_callback_time = av_gettime();
2183
2184     while (len > 0) {
2185         if (is->audio_buf_index >= is->audio_buf_size) {
2186            audio_size = audio_decode_frame(is, &pts);
2187            if (audio_size < 0) {
2188                 /* if error, just output silence */
2189                is->audio_buf      = is->silence_buf;
2190                is->audio_buf_size = sizeof(is->silence_buf) / frame_size * frame_size;
2191            } else {
2192                if (is->show_mode != SHOW_MODE_VIDEO)
2193                    update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2194                is->audio_buf_size = audio_size;
2195            }
2196            is->audio_buf_index = 0;
2197         }
2198         len1 = is->audio_buf_size - is->audio_buf_index;
2199         if (len1 > len)
2200             len1 = len;
2201         memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2202         len -= len1;
2203         stream += len1;
2204         is->audio_buf_index += len1;
2205     }
2206     bytes_per_sec = is->audio_tgt_freq * is->audio_tgt_channels * av_get_bytes_per_sample(is->audio_tgt_fmt);
2207     is->audio_write_buf_size = is->audio_buf_size - is->audio_buf_index;
2208     /* Let's assume the audio driver that is used by SDL has two periods. */
2209     is->audio_current_pts = is->audio_clock - (double)(2 * is->audio_hw_buf_size + is->audio_write_buf_size) / bytes_per_sec;
2210     is->audio_current_pts_drift = is->audio_current_pts - audio_callback_time / 1000000.0;
2211 }
2212
2213 /* open a given stream. Return 0 if OK */
2214 static int stream_component_open(VideoState *is, int stream_index)
2215 {
2216     AVFormatContext *ic = is->ic;
2217     AVCodecContext *avctx;
2218     AVCodec *codec;
2219     SDL_AudioSpec wanted_spec, spec;
2220     AVDictionary *opts;
2221     AVDictionaryEntry *t = NULL;
2222     int64_t wanted_channel_layout = 0;
2223     int wanted_nb_channels;
2224     const char *env;
2225
2226     if (stream_index < 0 || stream_index >= ic->nb_streams)
2227         return -1;
2228     avctx = ic->streams[stream_index]->codec;
2229
2230     codec = avcodec_find_decoder(avctx->codec_id);
2231     opts = filter_codec_opts(codec_opts, codec, ic, ic->streams[stream_index]);
2232
2233     switch(avctx->codec_type){
2234         case AVMEDIA_TYPE_AUDIO   : if(audio_codec_name   ) codec= avcodec_find_decoder_by_name(   audio_codec_name); break;
2235         case AVMEDIA_TYPE_SUBTITLE: if(subtitle_codec_name) codec= avcodec_find_decoder_by_name(subtitle_codec_name); break;
2236         case AVMEDIA_TYPE_VIDEO   : if(video_codec_name   ) codec= avcodec_find_decoder_by_name(   video_codec_name); break;
2237     }
2238     if (!codec)
2239         return -1;
2240
2241     avctx->workaround_bugs   = workaround_bugs;
2242     avctx->lowres            = lowres;
2243     if(avctx->lowres > codec->max_lowres){
2244         av_log(avctx, AV_LOG_WARNING, "The maximum value for lowres supported by the decoder is %d\n",
2245                 codec->max_lowres);
2246         avctx->lowres= codec->max_lowres;
2247     }
2248     avctx->idct_algo         = idct;
2249     avctx->skip_frame        = skip_frame;
2250     avctx->skip_idct         = skip_idct;
2251     avctx->skip_loop_filter  = skip_loop_filter;
2252     avctx->error_concealment = error_concealment;
2253
2254     if(avctx->lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2255     if (fast)   avctx->flags2 |= CODEC_FLAG2_FAST;
2256     if(codec->capabilities & CODEC_CAP_DR1)
2257         avctx->flags |= CODEC_FLAG_EMU_EDGE;
2258
2259     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2260         memset(&is->audio_pkt_temp, 0, sizeof(is->audio_pkt_temp));
2261         env = SDL_getenv("SDL_AUDIO_CHANNELS");
2262         if (env)
2263             wanted_channel_layout = av_get_default_channel_layout(SDL_atoi(env));
2264         if (!wanted_channel_layout) {
2265             wanted_channel_layout = (avctx->channel_layout && avctx->channels == av_get_channel_layout_nb_channels(avctx->channel_layout)) ? avctx->channel_layout : av_get_default_channel_layout(avctx->channels);
2266             wanted_channel_layout &= ~AV_CH_LAYOUT_STEREO_DOWNMIX;
2267             wanted_nb_channels = av_get_channel_layout_nb_channels(wanted_channel_layout);
2268             /* SDL only supports 1, 2, 4 or 6 channels at the moment, so we have to make sure not to request anything else. */
2269             while (wanted_nb_channels > 0 && (wanted_nb_channels == 3 || wanted_nb_channels == 5 || wanted_nb_channels > 6)) {
2270                 wanted_nb_channels--;
2271                 wanted_channel_layout = av_get_default_channel_layout(wanted_nb_channels);
2272             }
2273         }
2274         wanted_spec.channels = av_get_channel_layout_nb_channels(wanted_channel_layout);
2275         wanted_spec.freq = avctx->sample_rate;
2276         if (wanted_spec.freq <= 0 || wanted_spec.channels <= 0) {
2277             fprintf(stderr, "Invalid sample rate or channel count!\n");
2278             return -1;
2279         }
2280     }
2281
2282     if (!av_dict_get(opts, "threads", NULL, 0))
2283         av_dict_set(&opts, "threads", "auto", 0);
2284     if (!codec ||
2285         avcodec_open2(avctx, codec, &opts) < 0)
2286         return -1;
2287     if ((t = av_dict_get(opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2288         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2289         return AVERROR_OPTION_NOT_FOUND;
2290     }
2291
2292     /* prepare audio output */
2293     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2294         wanted_spec.format = AUDIO_S16SYS;
2295         wanted_spec.silence = 0;
2296         wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2297         wanted_spec.callback = sdl_audio_callback;
2298         wanted_spec.userdata = is;
2299         if (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2300             fprintf(stderr, "SDL_OpenAudio: %s\n", SDL_GetError());
2301             return -1;
2302         }
2303         is->audio_hw_buf_size = spec.size;
2304         if (spec.format != AUDIO_S16SYS) {
2305             fprintf(stderr, "SDL advised audio format %d is not supported!\n", spec.format);
2306             return -1;
2307         }
2308         if (spec.channels != wanted_spec.channels) {
2309             wanted_channel_layout = av_get_default_channel_layout(spec.channels);
2310             if (!wanted_channel_layout) {
2311                 fprintf(stderr, "SDL advised channel count %d is not supported!\n", spec.channels);
2312                 return -1;
2313             }
2314         }
2315         is->audio_src_fmt = is->audio_tgt_fmt = AV_SAMPLE_FMT_S16;
2316         is->audio_src_freq = is->audio_tgt_freq = spec.freq;
2317         is->audio_src_channel_layout = is->audio_tgt_channel_layout = wanted_channel_layout;
2318         is->audio_src_channels = is->audio_tgt_channels = spec.channels;
2319     }
2320
2321     ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2322     switch (avctx->codec_type) {
2323     case AVMEDIA_TYPE_AUDIO:
2324         is->audio_stream = stream_index;
2325         is->audio_st = ic->streams[stream_index];
2326         is->audio_buf_size  = 0;
2327         is->audio_buf_index = 0;
2328
2329         /* init averaging filter */
2330         is->audio_diff_avg_coef  = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2331         is->audio_diff_avg_count = 0;
2332         /* since we do not have a precise anough audio fifo fullness,
2333            we correct audio sync only if larger than this threshold */
2334         is->audio_diff_threshold = 2.0 * SDL_AUDIO_BUFFER_SIZE / wanted_spec.freq;
2335
2336         memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2337         packet_queue_init(&is->audioq);
2338         SDL_PauseAudio(0);
2339         break;
2340     case AVMEDIA_TYPE_VIDEO:
2341         is->video_stream = stream_index;
2342         is->video_st = ic->streams[stream_index];
2343
2344         packet_queue_init(&is->videoq);
2345         is->video_tid = SDL_CreateThread(video_thread, is);
2346         break;
2347     case AVMEDIA_TYPE_SUBTITLE:
2348         is->subtitle_stream = stream_index;
2349         is->subtitle_st = ic->streams[stream_index];
2350         packet_queue_init(&is->subtitleq);
2351
2352         is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2353         break;
2354     default:
2355         break;
2356     }
2357     return 0;
2358 }
2359
2360 static void stream_component_close(VideoState *is, int stream_index)
2361 {
2362     AVFormatContext *ic = is->ic;
2363     AVCodecContext *avctx;
2364
2365     if (stream_index < 0 || stream_index >= ic->nb_streams)
2366         return;
2367     avctx = ic->streams[stream_index]->codec;
2368
2369     switch (avctx->codec_type) {
2370     case AVMEDIA_TYPE_AUDIO:
2371         packet_queue_abort(&is->audioq);
2372
2373         SDL_CloseAudio();
2374
2375         packet_queue_end(&is->audioq);
2376         av_free_packet(&is->audio_pkt);
2377         if (is->swr_ctx)
2378             swr_free(&is->swr_ctx);
2379         av_freep(&is->audio_buf1);
2380         is->audio_buf = NULL;
2381         av_freep(&is->frame);
2382
2383         if (is->rdft) {
2384             av_rdft_end(is->rdft);
2385             av_freep(&is->rdft_data);
2386             is->rdft = NULL;
2387             is->rdft_bits = 0;
2388         }
2389         break;
2390     case AVMEDIA_TYPE_VIDEO:
2391         packet_queue_abort(&is->videoq);
2392
2393         /* note: we also signal this mutex to make sure we deblock the
2394            video thread in all cases */
2395         SDL_LockMutex(is->pictq_mutex);
2396         SDL_CondSignal(is->pictq_cond);
2397         SDL_UnlockMutex(is->pictq_mutex);
2398
2399         SDL_WaitThread(is->video_tid, NULL);
2400
2401         packet_queue_end(&is->videoq);
2402         break;
2403     case AVMEDIA_TYPE_SUBTITLE:
2404         packet_queue_abort(&is->subtitleq);
2405
2406         /* note: we also signal this mutex to make sure we deblock the
2407            video thread in all cases */
2408         SDL_LockMutex(is->subpq_mutex);
2409         is->subtitle_stream_changed = 1;
2410
2411         SDL_CondSignal(is->subpq_cond);
2412         SDL_UnlockMutex(is->subpq_mutex);
2413
2414         SDL_WaitThread(is->subtitle_tid, NULL);
2415
2416         packet_queue_end(&is->subtitleq);
2417         break;
2418     default:
2419         break;
2420     }
2421
2422     ic->streams[stream_index]->discard = AVDISCARD_ALL;
2423     avcodec_close(avctx);
2424     switch (avctx->codec_type) {
2425     case AVMEDIA_TYPE_AUDIO:
2426         is->audio_st = NULL;
2427         is->audio_stream = -1;
2428         break;
2429     case AVMEDIA_TYPE_VIDEO:
2430         is->video_st = NULL;
2431         is->video_stream = -1;
2432         break;
2433     case AVMEDIA_TYPE_SUBTITLE:
2434         is->subtitle_st = NULL;
2435         is->subtitle_stream = -1;
2436         break;
2437     default:
2438         break;
2439     }
2440 }
2441
2442 static int decode_interrupt_cb(void *ctx)
2443 {
2444     VideoState *is = ctx;
2445     return is->abort_request;
2446 }
2447
2448 /* this thread gets the stream from the disk or the network */
2449 static int read_thread(void *arg)
2450 {
2451     VideoState *is = arg;
2452     AVFormatContext *ic = NULL;
2453     int err, i, ret;
2454     int st_index[AVMEDIA_TYPE_NB];
2455     AVPacket pkt1, *pkt = &pkt1;
2456     int eof = 0;
2457     int pkt_in_play_range = 0;
2458     AVDictionaryEntry *t;
2459     AVDictionary **opts;
2460     int orig_nb_streams;
2461
2462     memset(st_index, -1, sizeof(st_index));
2463     is->video_stream = -1;
2464     is->audio_stream = -1;
2465     is->subtitle_stream = -1;
2466
2467     ic = avformat_alloc_context();
2468     ic->interrupt_callback.callback = decode_interrupt_cb;
2469     ic->interrupt_callback.opaque = is;
2470     err = avformat_open_input(&ic, is->filename, is->iformat, &format_opts);
2471     if (err < 0) {
2472         print_error(is->filename, err);
2473         ret = -1;
2474         goto fail;
2475     }
2476     if ((t = av_dict_get(format_opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2477         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2478         ret = AVERROR_OPTION_NOT_FOUND;
2479         goto fail;
2480     }
2481     is->ic = ic;
2482
2483     if (genpts)
2484         ic->flags |= AVFMT_FLAG_GENPTS;
2485
2486     opts = setup_find_stream_info_opts(ic, codec_opts);
2487     orig_nb_streams = ic->nb_streams;
2488
2489     err = avformat_find_stream_info(ic, opts);
2490     if (err < 0) {
2491         fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2492         ret = -1;
2493         goto fail;
2494     }
2495     for (i = 0; i < orig_nb_streams; i++)
2496         av_dict_free(&opts[i]);
2497     av_freep(&opts);
2498
2499     if (ic->pb)
2500         ic->pb->eof_reached = 0; // FIXME hack, ffplay maybe should not use url_feof() to test for the end
2501
2502     if (seek_by_bytes < 0)
2503         seek_by_bytes = !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2504
2505     /* if seeking requested, we execute it */
2506     if (start_time != AV_NOPTS_VALUE) {
2507         int64_t timestamp;
2508
2509         timestamp = start_time;
2510         /* add the stream start time */
2511         if (ic->start_time != AV_NOPTS_VALUE)
2512             timestamp += ic->start_time;
2513         ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2514         if (ret < 0) {
2515             fprintf(stderr, "%s: could not seek to position %0.3f\n",
2516                     is->filename, (double)timestamp / AV_TIME_BASE);
2517         }
2518     }
2519
2520     for (i = 0; i < ic->nb_streams; i++)
2521         ic->streams[i]->discard = AVDISCARD_ALL;
2522     if (!video_disable)
2523         st_index[AVMEDIA_TYPE_VIDEO] =
2524             av_find_best_stream(ic, AVMEDIA_TYPE_VIDEO,
2525                                 wanted_stream[AVMEDIA_TYPE_VIDEO], -1, NULL, 0);
2526     if (!audio_disable)
2527         st_index[AVMEDIA_TYPE_AUDIO] =
2528             av_find_best_stream(ic, AVMEDIA_TYPE_AUDIO,
2529                                 wanted_stream[AVMEDIA_TYPE_AUDIO],
2530                                 st_index[AVMEDIA_TYPE_VIDEO],
2531                                 NULL, 0);
2532     if (!video_disable)
2533         st_index[AVMEDIA_TYPE_SUBTITLE] =
2534             av_find_best_stream(ic, AVMEDIA_TYPE_SUBTITLE,
2535                                 wanted_stream[AVMEDIA_TYPE_SUBTITLE],
2536                                 (st_index[AVMEDIA_TYPE_AUDIO] >= 0 ?
2537                                  st_index[AVMEDIA_TYPE_AUDIO] :
2538                                  st_index[AVMEDIA_TYPE_VIDEO]),
2539                                 NULL, 0);
2540     if (show_status) {
2541         av_dump_format(ic, 0, is->filename, 0);
2542     }
2543
2544     is->show_mode = show_mode;
2545
2546     /* open the streams */
2547     if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2548         stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2549     }
2550
2551     ret = -1;
2552     if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2553         ret = stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2554     }
2555     is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2556     if (is->show_mode == SHOW_MODE_NONE)
2557         is->show_mode = ret >= 0 ? SHOW_MODE_VIDEO : SHOW_MODE_RDFT;
2558
2559     if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2560         stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2561     }
2562
2563     if (is->video_stream < 0 && is->audio_stream < 0) {
2564         fprintf(stderr, "%s: could not open codecs\n", is->filename);
2565         ret = -1;
2566         goto fail;
2567     }
2568
2569     for (;;) {
2570         if (is->abort_request)
2571             break;
2572         if (is->paused != is->last_paused) {
2573             is->last_paused = is->paused;
2574             if (is->paused)
2575                 is->read_pause_return = av_read_pause(ic);
2576             else
2577                 av_read_play(ic);
2578         }
2579 #if CONFIG_RTSP_DEMUXER || CONFIG_MMSH_PROTOCOL
2580         if (is->paused &&
2581                 (!strcmp(ic->iformat->name, "rtsp") ||
2582                  (ic->pb && !strncmp(input_filename, "mmsh:", 5)))) {
2583             /* wait 10 ms to avoid trying to get another packet */
2584             /* XXX: horrible */
2585             SDL_Delay(10);
2586             continue;
2587         }
2588 #endif
2589         if (is->seek_req) {
2590             int64_t seek_target = is->seek_pos;
2591             int64_t seek_min    = is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2592             int64_t seek_max    = is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2593 // FIXME the +-2 is due to rounding being not done in the correct direction in generation
2594 //      of the seek_pos/seek_rel variables
2595
2596             ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2597             if (ret < 0) {
2598                 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2599             } else {
2600                 if (is->audio_stream >= 0) {
2601                     packet_queue_flush(&is->audioq);
2602                     packet_queue_put(&is->audioq, &flush_pkt);
2603                 }
2604                 if (is->subtitle_stream >= 0) {
2605                     packet_queue_flush(&is->subtitleq);
2606                     packet_queue_put(&is->subtitleq, &flush_pkt);
2607                 }
2608                 if (is->video_stream >= 0) {
2609                     packet_queue_flush(&is->videoq);
2610                     packet_queue_put(&is->videoq, &flush_pkt);
2611                 }
2612             }
2613             is->seek_req = 0;
2614             eof = 0;
2615         }
2616
2617         /* if the queue are full, no need to read more */
2618         if (   is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2619             || (   (is->audioq   .nb_packets > MIN_FRAMES || is->audio_stream < 0)
2620                 && (is->videoq   .nb_packets > MIN_FRAMES || is->video_stream < 0)
2621                 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream < 0))) {
2622             /* wait 10 ms */
2623             SDL_Delay(10);
2624             continue;
2625         }
2626         if (eof) {
2627             if (is->video_stream >= 0) {
2628                 av_init_packet(pkt);
2629                 pkt->data = NULL;
2630                 pkt->size = 0;
2631                 pkt->stream_index = is->video_stream;
2632                 packet_queue_put(&is->videoq, pkt);
2633             }
2634             if (is->audio_stream >= 0 &&
2635                 is->audio_st->codec->codec->capabilities & CODEC_CAP_DELAY) {
2636                 av_init_packet(pkt);
2637                 pkt->data = NULL;
2638                 pkt->size = 0;
2639                 pkt->stream_index = is->audio_stream;
2640                 packet_queue_put(&is->audioq, pkt);
2641             }
2642             SDL_Delay(10);
2643             if (is->audioq.size + is->videoq.size + is->subtitleq.size == 0) {
2644                 if (loop != 1 && (!loop || --loop)) {
2645                     stream_seek(is, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2646                 } else if (autoexit) {
2647                     ret = AVERROR_EOF;
2648                     goto fail;
2649                 }
2650             }
2651             eof=0;
2652             continue;
2653         }
2654         ret = av_read_frame(ic, pkt);
2655         if (ret < 0) {
2656             if (ret == AVERROR_EOF || url_feof(ic->pb))
2657                 eof = 1;
2658             if (ic->pb && ic->pb->error)
2659                 break;
2660             SDL_Delay(100); /* wait for user event */
2661             continue;
2662         }
2663         /* check if packet is in play range specified by user, then queue, otherwise discard */
2664         pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2665                 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2666                 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2667                 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0) / 1000000
2668                 <= ((double)duration / 1000000);
2669         if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2670             packet_queue_put(&is->audioq, pkt);
2671         } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2672             packet_queue_put(&is->videoq, pkt);
2673         } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2674             packet_queue_put(&is->subtitleq, pkt);
2675         } else {
2676             av_free_packet(pkt);
2677         }
2678     }
2679     /* wait until the end */
2680     while (!is->abort_request) {
2681         SDL_Delay(100);
2682     }
2683
2684     ret = 0;
2685  fail:
2686     /* close each stream */
2687     if (is->audio_stream >= 0)
2688         stream_component_close(is, is->audio_stream);
2689     if (is->video_stream >= 0)
2690         stream_component_close(is, is->video_stream);
2691     if (is->subtitle_stream >= 0)
2692         stream_component_close(is, is->subtitle_stream);
2693     if (is->ic) {
2694         avformat_close_input(&is->ic);
2695     }
2696
2697     if (ret != 0) {
2698         SDL_Event event;
2699
2700         event.type = FF_QUIT_EVENT;
2701         event.user.data1 = is;
2702         SDL_PushEvent(&event);
2703     }
2704     return 0;
2705 }
2706
2707 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2708 {
2709     VideoState *is;
2710
2711     is = av_mallocz(sizeof(VideoState));
2712     if (!is)
2713         return NULL;
2714     av_strlcpy(is->filename, filename, sizeof(is->filename));
2715     is->iformat = iformat;
2716     is->ytop    = 0;
2717     is->xleft   = 0;
2718
2719     /* start video display */
2720     is->pictq_mutex = SDL_CreateMutex();
2721     is->pictq_cond  = SDL_CreateCond();
2722
2723     is->subpq_mutex = SDL_CreateMutex();
2724     is->subpq_cond  = SDL_CreateCond();
2725
2726     is->av_sync_type = av_sync_type;
2727     is->read_tid     = SDL_CreateThread(read_thread, is);
2728     if (!is->read_tid) {
2729         av_free(is);
2730         return NULL;
2731     }
2732     return is;
2733 }
2734
2735 static void stream_cycle_channel(VideoState *is, int codec_type)
2736 {
2737     AVFormatContext *ic = is->ic;
2738     int start_index, stream_index;
2739     AVStream *st;
2740
2741     if (codec_type == AVMEDIA_TYPE_VIDEO)
2742         start_index = is->video_stream;
2743     else if (codec_type == AVMEDIA_TYPE_AUDIO)
2744         start_index = is->audio_stream;
2745     else
2746         start_index = is->subtitle_stream;
2747     if (start_index < (codec_type == AVMEDIA_TYPE_SUBTITLE ? -1 : 0))
2748         return;
2749     stream_index = start_index;
2750     for (;;) {
2751         if (++stream_index >= is->ic->nb_streams)
2752         {
2753             if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2754             {
2755                 stream_index = -1;
2756                 goto the_end;
2757             } else
2758                 stream_index = 0;
2759         }
2760         if (stream_index == start_index)
2761             return;
2762         st = ic->streams[stream_index];
2763         if (st->codec->codec_type == codec_type) {
2764             /* check that parameters are OK */
2765             switch (codec_type) {
2766             case AVMEDIA_TYPE_AUDIO:
2767                 if (st->codec->sample_rate != 0 &&
2768                     st->codec->channels != 0)
2769                     goto the_end;
2770                 break;
2771             case AVMEDIA_TYPE_VIDEO:
2772             case AVMEDIA_TYPE_SUBTITLE:
2773                 goto the_end;
2774             default:
2775                 break;
2776             }
2777         }
2778     }
2779  the_end:
2780     stream_component_close(is, start_index);
2781     stream_component_open(is, stream_index);
2782 }
2783
2784
2785 static void toggle_full_screen(VideoState *is)
2786 {
2787     av_unused int i;
2788     is_full_screen = !is_full_screen;
2789 #if defined(__APPLE__) && SDL_VERSION_ATLEAST(1, 2, 14)
2790     /* OS X needs to reallocate the SDL overlays */
2791     for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
2792         is->pictq[i].reallocate = 1;
2793     }
2794 #endif
2795     video_open(is, 1);
2796 }
2797
2798 static void toggle_pause(VideoState *is)
2799 {
2800     stream_toggle_pause(is);
2801     is->step = 0;
2802 }
2803
2804 static void step_to_next_frame(VideoState *is)
2805 {
2806     /* if the stream is paused unpause it, then step */
2807     if (is->paused)
2808         stream_toggle_pause(is);
2809     is->step = 1;
2810 }
2811
2812 static void toggle_audio_display(VideoState *is)
2813 {
2814     int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2815     is->show_mode = (is->show_mode + 1) % SHOW_MODE_NB;
2816     fill_rectangle(screen,
2817                 is->xleft, is->ytop, is->width, is->height,
2818                 bgcolor);
2819     SDL_UpdateRect(screen, is->xleft, is->ytop, is->width, is->height);
2820 }
2821
2822 /* handle an event sent by the GUI */
2823 static void event_loop(VideoState *cur_stream)
2824 {
2825     SDL_Event event;
2826     double incr, pos, frac;
2827
2828     for (;;) {
2829         double x;
2830         SDL_WaitEvent(&event);
2831         switch (event.type) {
2832         case SDL_KEYDOWN:
2833             if (exit_on_keydown) {
2834                 do_exit(cur_stream);
2835                 break;
2836             }
2837             switch (event.key.keysym.sym) {
2838             case SDLK_ESCAPE:
2839             case SDLK_q:
2840                 do_exit(cur_stream);
2841                 break;
2842             case SDLK_f:
2843                 toggle_full_screen(cur_stream);
2844                 cur_stream->force_refresh = 1;
2845                 break;
2846             case SDLK_p:
2847             case SDLK_SPACE:
2848                 toggle_pause(cur_stream);
2849                 break;
2850             case SDLK_s: // S: Step to next frame
2851                 step_to_next_frame(cur_stream);
2852                 break;
2853             case SDLK_a:
2854                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2855                 break;
2856             case SDLK_v:
2857                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2858                 break;
2859             case SDLK_t:
2860                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2861                 break;
2862             case SDLK_w:
2863                 toggle_audio_display(cur_stream);
2864                 cur_stream->force_refresh = 1;
2865                 break;
2866             case SDLK_PAGEUP:
2867                 incr = 600.0;
2868                 goto do_seek;
2869             case SDLK_PAGEDOWN:
2870                 incr = -600.0;
2871                 goto do_seek;
2872             case SDLK_LEFT:
2873                 incr = -10.0;
2874                 goto do_seek;
2875             case SDLK_RIGHT:
2876                 incr = 10.0;
2877                 goto do_seek;
2878             case SDLK_UP:
2879                 incr = 60.0;
2880                 goto do_seek;
2881             case SDLK_DOWN:
2882                 incr = -60.0;
2883             do_seek:
2884                     if (seek_by_bytes) {
2885                         if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos >= 0) {
2886                             pos = cur_stream->video_current_pos;
2887                         } else if (cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos >= 0) {
2888                             pos = cur_stream->audio_pkt.pos;
2889                         } else
2890                             pos = avio_tell(cur_stream->ic->pb);
2891                         if (cur_stream->ic->bit_rate)
2892                             incr *= cur_stream->ic->bit_rate / 8.0;
2893                         else
2894                             incr *= 180000.0;
2895                         pos += incr;
2896                         stream_seek(cur_stream, pos, incr, 1);
2897                     } else {
2898                         pos = get_master_clock(cur_stream);
2899                         pos += incr;
2900                         stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2901                     }
2902                 break;
2903             default:
2904                 break;
2905             }
2906             break;
2907         case SDL_VIDEOEXPOSE:
2908             cur_stream->force_refresh = 1;
2909             break;
2910         case SDL_MOUSEBUTTONDOWN:
2911             if (exit_on_mousedown) {
2912                 do_exit(cur_stream);
2913                 break;
2914             }
2915         case SDL_MOUSEMOTION:
2916             if (event.type == SDL_MOUSEBUTTONDOWN) {
2917                 x = event.button.x;
2918             } else {
2919                 if (event.motion.state != SDL_PRESSED)
2920                     break;
2921                 x = event.motion.x;
2922             }
2923                 if (seek_by_bytes || cur_stream->ic->duration <= 0) {
2924                     uint64_t size =  avio_size(cur_stream->ic->pb);
2925                     stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2926                 } else {
2927                     int64_t ts;
2928                     int ns, hh, mm, ss;
2929                     int tns, thh, tmm, tss;
2930                     tns  = cur_stream->ic->duration / 1000000LL;
2931                     thh  = tns / 3600;
2932                     tmm  = (tns % 3600) / 60;
2933                     tss  = (tns % 60);
2934                     frac = x / cur_stream->width;
2935                     ns   = frac * tns;
2936                     hh   = ns / 3600;
2937                     mm   = (ns % 3600) / 60;
2938                     ss   = (ns % 60);
2939                     fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d)       \n", frac*100,
2940                             hh, mm, ss, thh, tmm, tss);
2941                     ts = frac * cur_stream->ic->duration;
2942                     if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2943                         ts += cur_stream->ic->start_time;
2944                     stream_seek(cur_stream, ts, 0, 0);
2945                 }
2946             break;
2947         case SDL_VIDEORESIZE:
2948                 screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2949                                           SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2950                 screen_width  = cur_stream->width  = event.resize.w;
2951                 screen_height = cur_stream->height = event.resize.h;
2952                 cur_stream->force_refresh = 1;
2953             break;
2954         case SDL_QUIT:
2955         case FF_QUIT_EVENT:
2956             do_exit(cur_stream);
2957             break;
2958         case FF_ALLOC_EVENT:
2959             video_open(event.user.data1, 0);
2960             alloc_picture(event.user.data1);
2961             break;
2962         case FF_REFRESH_EVENT:
2963             video_refresh(event.user.data1);
2964             cur_stream->refresh = 0;
2965             break;
2966         default:
2967             break;
2968         }
2969     }
2970 }
2971
2972 static int opt_frame_size(const char *opt, const char *arg)
2973 {
2974     av_log(NULL, AV_LOG_WARNING, "Option -s is deprecated, use -video_size.\n");
2975     return opt_default("video_size", arg);
2976 }
2977
2978 static int opt_width(const char *opt, const char *arg)
2979 {
2980     screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2981     return 0;
2982 }
2983
2984 static int opt_height(const char *opt, const char *arg)
2985 {
2986     screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2987     return 0;
2988 }
2989
2990 static int opt_format(const char *opt, const char *arg)
2991 {
2992     file_iformat = av_find_input_format(arg);
2993     if (!file_iformat) {
2994         fprintf(stderr, "Unknown input format: %s\n", arg);
2995         return AVERROR(EINVAL);
2996     }
2997     return 0;
2998 }
2999
3000 static int opt_frame_pix_fmt(const char *opt, const char *arg)
3001 {
3002     av_log(NULL, AV_LOG_WARNING, "Option -pix_fmt is deprecated, use -pixel_format.\n");
3003     return opt_default("pixel_format", arg);
3004 }
3005
3006 static int opt_sync(const char *opt, const char *arg)
3007 {
3008     if (!strcmp(arg, "audio"))
3009         av_sync_type = AV_SYNC_AUDIO_MASTER;
3010     else if (!strcmp(arg, "video"))
3011         av_sync_type = AV_SYNC_VIDEO_MASTER;
3012     else if (!strcmp(arg, "ext"))
3013         av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
3014     else {
3015         fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
3016         exit(1);
3017     }
3018     return 0;
3019 }
3020
3021 static int opt_seek(const char *opt, const char *arg)
3022 {
3023     start_time = parse_time_or_die(opt, arg, 1);
3024     return 0;
3025 }
3026
3027 static int opt_duration(const char *opt, const char *arg)
3028 {
3029     duration = parse_time_or_die(opt, arg, 1);
3030     return 0;
3031 }
3032
3033 static int opt_show_mode(const char *opt, const char *arg)
3034 {
3035     show_mode = !strcmp(arg, "video") ? SHOW_MODE_VIDEO :
3036                 !strcmp(arg, "waves") ? SHOW_MODE_WAVES :
3037                 !strcmp(arg, "rdft" ) ? SHOW_MODE_RDFT  :
3038                 parse_number_or_die(opt, arg, OPT_INT, 0, SHOW_MODE_NB-1);
3039     return 0;
3040 }
3041
3042 static void opt_input_file(void *optctx, const char *filename)
3043 {
3044     if (input_filename) {
3045         fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
3046                 filename, input_filename);
3047         exit_program(1);
3048     }
3049     if (!strcmp(filename, "-"))
3050         filename = "pipe:";
3051     input_filename = filename;
3052 }
3053
3054 static int opt_codec(void *o, const char *opt, const char *arg)
3055 {
3056     switch(opt[strlen(opt)-1]){
3057     case 'a' :    audio_codec_name = arg; break;
3058     case 's' : subtitle_codec_name = arg; break;
3059     case 'v' :    video_codec_name = arg; break;
3060     }
3061     return 0;
3062 }
3063
3064 static int dummy;
3065
3066 static const OptionDef options[] = {
3067 #include "cmdutils_common_opts.h"
3068     { "x", HAS_ARG, { (void*)opt_width }, "force displayed width", "width" },
3069     { "y", HAS_ARG, { (void*)opt_height }, "force displayed height", "height" },
3070     { "s", HAS_ARG | OPT_VIDEO, { (void*)opt_frame_size }, "set frame size (WxH or abbreviation)", "size" },
3071     { "fs", OPT_BOOL, { (void*)&is_full_screen }, "force full screen" },
3072     { "an", OPT_BOOL, { (void*)&audio_disable }, "disable audio" },
3073     { "vn", OPT_BOOL, { (void*)&video_disable }, "disable video" },
3074     { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_AUDIO] }, "select desired audio stream", "stream_number" },
3075     { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_VIDEO] }, "select desired video stream", "stream_number" },
3076     { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE] }, "select desired subtitle stream", "stream_number" },
3077     { "ss", HAS_ARG, { (void*)&opt_seek }, "seek to a given position in seconds", "pos" },
3078     { "t", HAS_ARG, { (void*)&opt_duration }, "play  \"duration\" seconds of audio/video", "duration" },
3079     { "bytes", OPT_INT | HAS_ARG, { (void*)&seek_by_bytes }, "seek by bytes 0=off 1=on -1=auto", "val" },
3080     { "nodisp", OPT_BOOL, { (void*)&display_disable }, "disable graphical display" },
3081     { "f", HAS_ARG, { (void*)opt_format }, "force format", "fmt" },
3082     { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, { (void*)opt_frame_pix_fmt }, "set pixel format", "format" },
3083     { "stats", OPT_BOOL | OPT_EXPERT, { (void*)&show_status }, "show status", "" },
3084     { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&workaround_bugs }, "workaround bugs", "" },
3085     { "fast", OPT_BOOL | OPT_EXPERT, { (void*)&fast }, "non spec compliant optimizations", "" },
3086     { "genpts", OPT_BOOL | OPT_EXPERT, { (void*)&genpts }, "generate pts", "" },
3087     { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&decoder_reorder_pts }, "let decoder reorder pts 0=off 1=on -1=auto", ""},
3088     { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&lowres }, "", "" },
3089     { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_loop_filter }, "", "" },
3090     { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_frame }, "", "" },
3091     { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&skip_idct }, "", "" },
3092     { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&idct }, "set idct algo",  "algo" },
3093     { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&error_concealment }, "set error concealment options",  "bit_mask" },
3094     { "sync", HAS_ARG | OPT_EXPERT, { (void*)opt_sync }, "set audio-video sync. type (type=audio/video/ext)", "type" },
3095     { "autoexit", OPT_BOOL | OPT_EXPERT, { (void*)&autoexit }, "exit at the end", "" },
3096     { "exitonkeydown", OPT_BOOL | OPT_EXPERT, { (void*)&exit_on_keydown }, "exit on key down", "" },
3097     { "exitonmousedown", OPT_BOOL | OPT_EXPERT, { (void*)&exit_on_mousedown }, "exit on mouse down", "" },
3098     { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, { (void*)&loop }, "set number of times the playback shall be looped", "loop count" },
3099     { "framedrop", OPT_BOOL | OPT_EXPERT, { (void*)&framedrop }, "drop frames when cpu is too slow", "" },
3100     { "window_title", OPT_STRING | HAS_ARG, { (void*)&window_title }, "set window title", "window title" },
3101 #if CONFIG_AVFILTER
3102     { "vf", OPT_STRING | HAS_ARG, { (void*)&vfilters }, "video filters", "filter list" },
3103 #endif
3104     { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, { (void*)&rdftspeed }, "rdft speed", "msecs" },
3105     { "showmode", HAS_ARG, {(void*)opt_show_mode}, "select show mode (0 = video, 1 = waves, 2 = RDFT)", "mode" },
3106     { "default", HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, { (void*)opt_default }, "generic catch all option", "" },
3107     { "i", OPT_BOOL, {(void *)&dummy}, "read specified file", "input_file"},
3108     { "codec", HAS_ARG | OPT_FUNC2, {(void*)opt_codec}, "force decoder", "decoder" },
3109     { NULL, },
3110 };
3111
3112 static void show_usage(void)
3113 {
3114     av_log(NULL, AV_LOG_INFO, "Simple media player\n");
3115     av_log(NULL, AV_LOG_INFO, "usage: %s [options] input_file\n", program_name);
3116     av_log(NULL, AV_LOG_INFO, "\n");
3117 }
3118
3119 static int opt_help(const char *opt, const char *arg)
3120 {
3121     av_log_set_callback(log_callback_help);
3122     show_usage();
3123     show_help_options(options, "Main options:\n",
3124                       OPT_EXPERT, 0);
3125     show_help_options(options, "\nAdvanced options:\n",
3126                       OPT_EXPERT, OPT_EXPERT);
3127     printf("\n");
3128     show_help_children(avcodec_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3129     show_help_children(avformat_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3130 #if !CONFIG_AVFILTER
3131     show_help_children(sws_get_class(), AV_OPT_FLAG_ENCODING_PARAM);
3132 #endif
3133     printf("\nWhile playing:\n"
3134            "q, ESC              quit\n"
3135            "f                   toggle full screen\n"
3136            "p, SPC              pause\n"
3137            "a                   cycle audio channel\n"
3138            "v                   cycle video channel\n"
3139            "t                   cycle subtitle channel\n"
3140            "w                   show audio waves\n"
3141            "s                   activate frame-step mode\n"
3142            "left/right          seek backward/forward 10 seconds\n"
3143            "down/up             seek backward/forward 1 minute\n"
3144            "page down/page up   seek backward/forward 10 minutes\n"
3145            "mouse click         seek to percentage in file corresponding to fraction of width\n"
3146            );
3147     return 0;
3148 }
3149
3150 static int lockmgr(void **mtx, enum AVLockOp op)
3151 {
3152    switch(op) {
3153       case AV_LOCK_CREATE:
3154           *mtx = SDL_CreateMutex();
3155           if(!*mtx)
3156               return 1;
3157           return 0;
3158       case AV_LOCK_OBTAIN:
3159           return !!SDL_LockMutex(*mtx);
3160       case AV_LOCK_RELEASE:
3161           return !!SDL_UnlockMutex(*mtx);
3162       case AV_LOCK_DESTROY:
3163           SDL_DestroyMutex(*mtx);
3164           return 0;
3165    }
3166    return 1;
3167 }
3168
3169 /* Called from the main */
3170 int main(int argc, char **argv)
3171 {
3172     int flags;
3173     VideoState *is;
3174
3175     av_log_set_flags(AV_LOG_SKIP_REPEATED);
3176     parse_loglevel(argc, argv, options);
3177
3178     /* register all codecs, demux and protocols */
3179     avcodec_register_all();
3180 #if CONFIG_AVDEVICE
3181     avdevice_register_all();
3182 #endif
3183 #if CONFIG_AVFILTER
3184     avfilter_register_all();
3185 #endif
3186     av_register_all();
3187     avformat_network_init();
3188
3189     init_opts();
3190
3191     signal(SIGINT , sigterm_handler); /* Interrupt (ANSI).    */
3192     signal(SIGTERM, sigterm_handler); /* Termination (ANSI).  */
3193
3194     show_banner(argc, argv, options);
3195
3196     parse_options(NULL, argc, argv, options, opt_input_file);
3197
3198     if (!input_filename) {
3199         show_usage();
3200         fprintf(stderr, "An input file must be specified\n");
3201         fprintf(stderr, "Use -h to get full help or, even better, run 'man %s'\n", program_name);
3202         exit(1);
3203     }
3204
3205     if (display_disable) {
3206         video_disable = 1;
3207     }
3208     flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3209     if (audio_disable)
3210         flags &= ~SDL_INIT_AUDIO;
3211 #if !defined(__MINGW32__) && !defined(__APPLE__)
3212     flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3213 #endif
3214     if (SDL_Init (flags)) {
3215         fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3216         fprintf(stderr, "(Did you set the DISPLAY variable?)\n");
3217         exit(1);
3218     }
3219
3220     if (!display_disable) {
3221 #if HAVE_SDL_VIDEO_SIZE
3222         const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3223         fs_screen_width = vi->current_w;
3224         fs_screen_height = vi->current_h;
3225 #endif
3226     }
3227
3228     SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3229     SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3230     SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3231
3232     if (av_lockmgr_register(lockmgr)) {
3233         fprintf(stderr, "Could not initialize lock manager!\n");
3234         do_exit(NULL);
3235     }
3236
3237     av_init_packet(&flush_pkt);
3238     flush_pkt.data = (char *)(intptr_t)"FLUSH";
3239
3240     is = stream_open(input_filename, file_iformat);
3241     if (!is) {
3242         fprintf(stderr, "Failed to initialize VideoState!\n");
3243         do_exit(NULL);
3244     }
3245
3246     event_loop(is);
3247
3248     /* never returns */
3249
3250     return 0;
3251 }