]> git.sesse.net Git - ffmpeg/blob - ffplay.c
Add init and uninit functions to cmdutils, reduces code duplication
[ffmpeg] / ffplay.c
1 /*
2  * FFplay : Simple Media Player based on the FFmpeg libraries
3  * Copyright (c) 2003 Fabrice Bellard
4  *
5  * This file is part of FFmpeg.
6  *
7  * FFmpeg is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * FFmpeg is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with FFmpeg; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21
22 #define _XOPEN_SOURCE 600
23
24 #include "config.h"
25 #include <inttypes.h>
26 #include <math.h>
27 #include <limits.h>
28 #include "libavutil/avstring.h"
29 #include "libavutil/colorspace.h"
30 #include "libavutil/pixdesc.h"
31 #include "libavcore/imgutils.h"
32 #include "libavcore/parseutils.h"
33 #include "libavformat/avformat.h"
34 #include "libavdevice/avdevice.h"
35 #include "libswscale/swscale.h"
36 #include "libavcodec/audioconvert.h"
37 #include "libavcodec/opt.h"
38 #include "libavcodec/avfft.h"
39
40 #if CONFIG_AVFILTER
41 # include "libavfilter/avfilter.h"
42 # include "libavfilter/avfiltergraph.h"
43 # include "libavfilter/graphparser.h"
44 #endif
45
46 #include "cmdutils.h"
47
48 #include <SDL.h>
49 #include <SDL_thread.h>
50
51 #ifdef __MINGW32__
52 #undef main /* We don't want SDL to override our main() */
53 #endif
54
55 #include <unistd.h>
56 #include <assert.h>
57
58 const char program_name[] = "FFplay";
59 const int program_birth_year = 2003;
60
61 //#define DEBUG_SYNC
62
63 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
64 #define MIN_AUDIOQ_SIZE (20 * 16 * 1024)
65 #define MIN_FRAMES 5
66
67 /* SDL audio buffer size, in samples. Should be small to have precise
68    A/V sync as SDL does not have hardware buffer fullness info. */
69 #define SDL_AUDIO_BUFFER_SIZE 1024
70
71 /* no AV sync correction is done if below the AV sync threshold */
72 #define AV_SYNC_THRESHOLD 0.01
73 /* no AV correction is done if too big error */
74 #define AV_NOSYNC_THRESHOLD 10.0
75
76 #define FRAME_SKIP_FACTOR 0.05
77
78 /* maximum audio speed change to get correct sync */
79 #define SAMPLE_CORRECTION_PERCENT_MAX 10
80
81 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
82 #define AUDIO_DIFF_AVG_NB   20
83
84 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
85 #define SAMPLE_ARRAY_SIZE (2*65536)
86
87 static int sws_flags = SWS_BICUBIC;
88
89 typedef struct PacketQueue {
90     AVPacketList *first_pkt, *last_pkt;
91     int nb_packets;
92     int size;
93     int abort_request;
94     SDL_mutex *mutex;
95     SDL_cond *cond;
96 } PacketQueue;
97
98 #define VIDEO_PICTURE_QUEUE_SIZE 2
99 #define SUBPICTURE_QUEUE_SIZE 4
100
101 typedef struct VideoPicture {
102     double pts;                                  ///<presentation time stamp for this picture
103     double target_clock;                         ///<av_gettime() time at which this should be displayed ideally
104     int64_t pos;                                 ///<byte position in file
105     SDL_Overlay *bmp;
106     int width, height; /* source height & width */
107     int allocated;
108     enum PixelFormat pix_fmt;
109
110 #if CONFIG_AVFILTER
111     AVFilterBufferRef *picref;
112 #endif
113 } VideoPicture;
114
115 typedef struct SubPicture {
116     double pts; /* presentation time stamp for this picture */
117     AVSubtitle sub;
118 } SubPicture;
119
120 enum {
121     AV_SYNC_AUDIO_MASTER, /* default choice */
122     AV_SYNC_VIDEO_MASTER,
123     AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
124 };
125
126 typedef struct VideoState {
127     SDL_Thread *parse_tid;
128     SDL_Thread *video_tid;
129     SDL_Thread *refresh_tid;
130     AVInputFormat *iformat;
131     int no_background;
132     int abort_request;
133     int paused;
134     int last_paused;
135     int seek_req;
136     int seek_flags;
137     int64_t seek_pos;
138     int64_t seek_rel;
139     int read_pause_return;
140     AVFormatContext *ic;
141     int dtg_active_format;
142
143     int audio_stream;
144
145     int av_sync_type;
146     double external_clock; /* external clock base */
147     int64_t external_clock_time;
148
149     double audio_clock;
150     double audio_diff_cum; /* used for AV difference average computation */
151     double audio_diff_avg_coef;
152     double audio_diff_threshold;
153     int audio_diff_avg_count;
154     AVStream *audio_st;
155     PacketQueue audioq;
156     int audio_hw_buf_size;
157     /* samples output by the codec. we reserve more space for avsync
158        compensation */
159     DECLARE_ALIGNED(16,uint8_t,audio_buf1)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
160     DECLARE_ALIGNED(16,uint8_t,audio_buf2)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
161     uint8_t *audio_buf;
162     unsigned int audio_buf_size; /* in bytes */
163     int audio_buf_index; /* in bytes */
164     AVPacket audio_pkt_temp;
165     AVPacket audio_pkt;
166     enum SampleFormat audio_src_fmt;
167     AVAudioConvert *reformat_ctx;
168
169     int show_audio; /* if true, display audio samples */
170     int16_t sample_array[SAMPLE_ARRAY_SIZE];
171     int sample_array_index;
172     int last_i_start;
173     RDFTContext *rdft;
174     int rdft_bits;
175     FFTSample *rdft_data;
176     int xpos;
177
178     SDL_Thread *subtitle_tid;
179     int subtitle_stream;
180     int subtitle_stream_changed;
181     AVStream *subtitle_st;
182     PacketQueue subtitleq;
183     SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
184     int subpq_size, subpq_rindex, subpq_windex;
185     SDL_mutex *subpq_mutex;
186     SDL_cond *subpq_cond;
187
188     double frame_timer;
189     double frame_last_pts;
190     double frame_last_delay;
191     double video_clock;                          ///<pts of last decoded frame / predicted pts of next decoded frame
192     int video_stream;
193     AVStream *video_st;
194     PacketQueue videoq;
195     double video_current_pts;                    ///<current displayed pts (different from video_clock if frame fifos are used)
196     double video_current_pts_drift;              ///<video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
197     int64_t video_current_pos;                   ///<current displayed file pos
198     VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
199     int pictq_size, pictq_rindex, pictq_windex;
200     SDL_mutex *pictq_mutex;
201     SDL_cond *pictq_cond;
202 #if !CONFIG_AVFILTER
203     struct SwsContext *img_convert_ctx;
204 #endif
205
206     //    QETimer *video_timer;
207     char filename[1024];
208     int width, height, xleft, ytop;
209
210     PtsCorrectionContext pts_ctx;
211
212 #if CONFIG_AVFILTER
213     AVFilterContext *out_video_filter;          ///<the last filter in the video chain
214 #endif
215
216     float skip_frames;
217     float skip_frames_index;
218     int refresh;
219 } VideoState;
220
221 static void show_help(void);
222 static int audio_write_get_buf_size(VideoState *is);
223
224 /* options specified by the user */
225 static AVInputFormat *file_iformat;
226 static const char *input_filename;
227 static const char *window_title;
228 static int fs_screen_width;
229 static int fs_screen_height;
230 static int screen_width = 0;
231 static int screen_height = 0;
232 static int frame_width = 0;
233 static int frame_height = 0;
234 static enum PixelFormat frame_pix_fmt = PIX_FMT_NONE;
235 static int audio_disable;
236 static int video_disable;
237 static int wanted_stream[AVMEDIA_TYPE_NB]={
238     [AVMEDIA_TYPE_AUDIO]=-1,
239     [AVMEDIA_TYPE_VIDEO]=-1,
240     [AVMEDIA_TYPE_SUBTITLE]=-1,
241 };
242 static int seek_by_bytes=-1;
243 static int display_disable;
244 static int show_status = 1;
245 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
246 static int64_t start_time = AV_NOPTS_VALUE;
247 static int64_t duration = AV_NOPTS_VALUE;
248 static int debug = 0;
249 static int debug_mv = 0;
250 static int step = 0;
251 static int thread_count = 1;
252 static int workaround_bugs = 1;
253 static int fast = 0;
254 static int genpts = 0;
255 static int lowres = 0;
256 static int idct = FF_IDCT_AUTO;
257 static enum AVDiscard skip_frame= AVDISCARD_DEFAULT;
258 static enum AVDiscard skip_idct= AVDISCARD_DEFAULT;
259 static enum AVDiscard skip_loop_filter= AVDISCARD_DEFAULT;
260 static int error_recognition = FF_ER_CAREFUL;
261 static int error_concealment = 3;
262 static int decoder_reorder_pts= -1;
263 static int autoexit;
264 static int exit_on_keydown;
265 static int exit_on_mousedown;
266 static int loop=1;
267 static int framedrop=1;
268
269 static int rdftspeed=20;
270 #if CONFIG_AVFILTER
271 static char *vfilters = NULL;
272 #endif
273
274 /* current context */
275 static int is_full_screen;
276 static VideoState *cur_stream;
277 static int64_t audio_callback_time;
278
279 static AVPacket flush_pkt;
280
281 #define FF_ALLOC_EVENT   (SDL_USEREVENT)
282 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
283 #define FF_QUIT_EVENT    (SDL_USEREVENT + 2)
284
285 static SDL_Surface *screen;
286
287 static int packet_queue_put(PacketQueue *q, AVPacket *pkt);
288
289 /* packet queue handling */
290 static void packet_queue_init(PacketQueue *q)
291 {
292     memset(q, 0, sizeof(PacketQueue));
293     q->mutex = SDL_CreateMutex();
294     q->cond = SDL_CreateCond();
295     packet_queue_put(q, &flush_pkt);
296 }
297
298 static void packet_queue_flush(PacketQueue *q)
299 {
300     AVPacketList *pkt, *pkt1;
301
302     SDL_LockMutex(q->mutex);
303     for(pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
304         pkt1 = pkt->next;
305         av_free_packet(&pkt->pkt);
306         av_freep(&pkt);
307     }
308     q->last_pkt = NULL;
309     q->first_pkt = NULL;
310     q->nb_packets = 0;
311     q->size = 0;
312     SDL_UnlockMutex(q->mutex);
313 }
314
315 static void packet_queue_end(PacketQueue *q)
316 {
317     packet_queue_flush(q);
318     SDL_DestroyMutex(q->mutex);
319     SDL_DestroyCond(q->cond);
320 }
321
322 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
323 {
324     AVPacketList *pkt1;
325
326     /* duplicate the packet */
327     if (pkt!=&flush_pkt && av_dup_packet(pkt) < 0)
328         return -1;
329
330     pkt1 = av_malloc(sizeof(AVPacketList));
331     if (!pkt1)
332         return -1;
333     pkt1->pkt = *pkt;
334     pkt1->next = NULL;
335
336
337     SDL_LockMutex(q->mutex);
338
339     if (!q->last_pkt)
340
341         q->first_pkt = pkt1;
342     else
343         q->last_pkt->next = pkt1;
344     q->last_pkt = pkt1;
345     q->nb_packets++;
346     q->size += pkt1->pkt.size + sizeof(*pkt1);
347     /* XXX: should duplicate packet data in DV case */
348     SDL_CondSignal(q->cond);
349
350     SDL_UnlockMutex(q->mutex);
351     return 0;
352 }
353
354 static void packet_queue_abort(PacketQueue *q)
355 {
356     SDL_LockMutex(q->mutex);
357
358     q->abort_request = 1;
359
360     SDL_CondSignal(q->cond);
361
362     SDL_UnlockMutex(q->mutex);
363 }
364
365 /* return < 0 if aborted, 0 if no packet and > 0 if packet.  */
366 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
367 {
368     AVPacketList *pkt1;
369     int ret;
370
371     SDL_LockMutex(q->mutex);
372
373     for(;;) {
374         if (q->abort_request) {
375             ret = -1;
376             break;
377         }
378
379         pkt1 = q->first_pkt;
380         if (pkt1) {
381             q->first_pkt = pkt1->next;
382             if (!q->first_pkt)
383                 q->last_pkt = NULL;
384             q->nb_packets--;
385             q->size -= pkt1->pkt.size + sizeof(*pkt1);
386             *pkt = pkt1->pkt;
387             av_free(pkt1);
388             ret = 1;
389             break;
390         } else if (!block) {
391             ret = 0;
392             break;
393         } else {
394             SDL_CondWait(q->cond, q->mutex);
395         }
396     }
397     SDL_UnlockMutex(q->mutex);
398     return ret;
399 }
400
401 static inline void fill_rectangle(SDL_Surface *screen,
402                                   int x, int y, int w, int h, int color)
403 {
404     SDL_Rect rect;
405     rect.x = x;
406     rect.y = y;
407     rect.w = w;
408     rect.h = h;
409     SDL_FillRect(screen, &rect, color);
410 }
411
412 #if 0
413 /* draw only the border of a rectangle */
414 void fill_border(VideoState *s, int x, int y, int w, int h, int color)
415 {
416     int w1, w2, h1, h2;
417
418     /* fill the background */
419     w1 = x;
420     if (w1 < 0)
421         w1 = 0;
422     w2 = s->width - (x + w);
423     if (w2 < 0)
424         w2 = 0;
425     h1 = y;
426     if (h1 < 0)
427         h1 = 0;
428     h2 = s->height - (y + h);
429     if (h2 < 0)
430         h2 = 0;
431     fill_rectangle(screen,
432                    s->xleft, s->ytop,
433                    w1, s->height,
434                    color);
435     fill_rectangle(screen,
436                    s->xleft + s->width - w2, s->ytop,
437                    w2, s->height,
438                    color);
439     fill_rectangle(screen,
440                    s->xleft + w1, s->ytop,
441                    s->width - w1 - w2, h1,
442                    color);
443     fill_rectangle(screen,
444                    s->xleft + w1, s->ytop + s->height - h2,
445                    s->width - w1 - w2, h2,
446                    color);
447 }
448 #endif
449
450 #define ALPHA_BLEND(a, oldp, newp, s)\
451 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
452
453 #define RGBA_IN(r, g, b, a, s)\
454 {\
455     unsigned int v = ((const uint32_t *)(s))[0];\
456     a = (v >> 24) & 0xff;\
457     r = (v >> 16) & 0xff;\
458     g = (v >> 8) & 0xff;\
459     b = v & 0xff;\
460 }
461
462 #define YUVA_IN(y, u, v, a, s, pal)\
463 {\
464     unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
465     a = (val >> 24) & 0xff;\
466     y = (val >> 16) & 0xff;\
467     u = (val >> 8) & 0xff;\
468     v = val & 0xff;\
469 }
470
471 #define YUVA_OUT(d, y, u, v, a)\
472 {\
473     ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
474 }
475
476
477 #define BPP 1
478
479 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
480 {
481     int wrap, wrap3, width2, skip2;
482     int y, u, v, a, u1, v1, a1, w, h;
483     uint8_t *lum, *cb, *cr;
484     const uint8_t *p;
485     const uint32_t *pal;
486     int dstx, dsty, dstw, dsth;
487
488     dstw = av_clip(rect->w, 0, imgw);
489     dsth = av_clip(rect->h, 0, imgh);
490     dstx = av_clip(rect->x, 0, imgw - dstw);
491     dsty = av_clip(rect->y, 0, imgh - dsth);
492     lum = dst->data[0] + dsty * dst->linesize[0];
493     cb = dst->data[1] + (dsty >> 1) * dst->linesize[1];
494     cr = dst->data[2] + (dsty >> 1) * dst->linesize[2];
495
496     width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
497     skip2 = dstx >> 1;
498     wrap = dst->linesize[0];
499     wrap3 = rect->pict.linesize[0];
500     p = rect->pict.data[0];
501     pal = (const uint32_t *)rect->pict.data[1];  /* Now in YCrCb! */
502
503     if (dsty & 1) {
504         lum += dstx;
505         cb += skip2;
506         cr += skip2;
507
508         if (dstx & 1) {
509             YUVA_IN(y, u, v, a, p, pal);
510             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
511             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
512             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
513             cb++;
514             cr++;
515             lum++;
516             p += BPP;
517         }
518         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
519             YUVA_IN(y, u, v, a, p, pal);
520             u1 = u;
521             v1 = v;
522             a1 = a;
523             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
524
525             YUVA_IN(y, u, v, a, p + BPP, pal);
526             u1 += u;
527             v1 += v;
528             a1 += a;
529             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
530             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
531             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
532             cb++;
533             cr++;
534             p += 2 * BPP;
535             lum += 2;
536         }
537         if (w) {
538             YUVA_IN(y, u, v, a, p, pal);
539             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
540             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
541             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
542             p++;
543             lum++;
544         }
545         p += wrap3 - dstw * BPP;
546         lum += wrap - dstw - dstx;
547         cb += dst->linesize[1] - width2 - skip2;
548         cr += dst->linesize[2] - width2 - skip2;
549     }
550     for(h = dsth - (dsty & 1); h >= 2; h -= 2) {
551         lum += dstx;
552         cb += skip2;
553         cr += skip2;
554
555         if (dstx & 1) {
556             YUVA_IN(y, u, v, a, p, pal);
557             u1 = u;
558             v1 = v;
559             a1 = a;
560             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
561             p += wrap3;
562             lum += wrap;
563             YUVA_IN(y, u, v, a, p, pal);
564             u1 += u;
565             v1 += v;
566             a1 += a;
567             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
568             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
569             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
570             cb++;
571             cr++;
572             p += -wrap3 + BPP;
573             lum += -wrap + 1;
574         }
575         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
576             YUVA_IN(y, u, v, a, p, pal);
577             u1 = u;
578             v1 = v;
579             a1 = a;
580             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
581
582             YUVA_IN(y, u, v, a, p + BPP, pal);
583             u1 += u;
584             v1 += v;
585             a1 += a;
586             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
587             p += wrap3;
588             lum += wrap;
589
590             YUVA_IN(y, u, v, a, p, pal);
591             u1 += u;
592             v1 += v;
593             a1 += a;
594             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
595
596             YUVA_IN(y, u, v, a, p + BPP, pal);
597             u1 += u;
598             v1 += v;
599             a1 += a;
600             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
601
602             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
603             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
604
605             cb++;
606             cr++;
607             p += -wrap3 + 2 * BPP;
608             lum += -wrap + 2;
609         }
610         if (w) {
611             YUVA_IN(y, u, v, a, p, pal);
612             u1 = u;
613             v1 = v;
614             a1 = a;
615             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
616             p += wrap3;
617             lum += wrap;
618             YUVA_IN(y, u, v, a, p, pal);
619             u1 += u;
620             v1 += v;
621             a1 += a;
622             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
623             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
624             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
625             cb++;
626             cr++;
627             p += -wrap3 + BPP;
628             lum += -wrap + 1;
629         }
630         p += wrap3 + (wrap3 - dstw * BPP);
631         lum += wrap + (wrap - dstw - dstx);
632         cb += dst->linesize[1] - width2 - skip2;
633         cr += dst->linesize[2] - width2 - skip2;
634     }
635     /* handle odd height */
636     if (h) {
637         lum += dstx;
638         cb += skip2;
639         cr += skip2;
640
641         if (dstx & 1) {
642             YUVA_IN(y, u, v, a, p, pal);
643             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
644             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
645             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
646             cb++;
647             cr++;
648             lum++;
649             p += BPP;
650         }
651         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
652             YUVA_IN(y, u, v, a, p, pal);
653             u1 = u;
654             v1 = v;
655             a1 = a;
656             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
657
658             YUVA_IN(y, u, v, a, p + BPP, pal);
659             u1 += u;
660             v1 += v;
661             a1 += a;
662             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
663             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
664             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
665             cb++;
666             cr++;
667             p += 2 * BPP;
668             lum += 2;
669         }
670         if (w) {
671             YUVA_IN(y, u, v, a, p, pal);
672             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
673             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
674             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
675         }
676     }
677 }
678
679 static void free_subpicture(SubPicture *sp)
680 {
681     avsubtitle_free(&sp->sub);
682 }
683
684 static void video_image_display(VideoState *is)
685 {
686     VideoPicture *vp;
687     SubPicture *sp;
688     AVPicture pict;
689     float aspect_ratio;
690     int width, height, x, y;
691     SDL_Rect rect;
692     int i;
693
694     vp = &is->pictq[is->pictq_rindex];
695     if (vp->bmp) {
696 #if CONFIG_AVFILTER
697          if (vp->picref->video->pixel_aspect.num == 0)
698              aspect_ratio = 0;
699          else
700              aspect_ratio = av_q2d(vp->picref->video->pixel_aspect);
701 #else
702
703         /* XXX: use variable in the frame */
704         if (is->video_st->sample_aspect_ratio.num)
705             aspect_ratio = av_q2d(is->video_st->sample_aspect_ratio);
706         else if (is->video_st->codec->sample_aspect_ratio.num)
707             aspect_ratio = av_q2d(is->video_st->codec->sample_aspect_ratio);
708         else
709             aspect_ratio = 0;
710 #endif
711         if (aspect_ratio <= 0.0)
712             aspect_ratio = 1.0;
713         aspect_ratio *= (float)vp->width / (float)vp->height;
714         /* if an active format is indicated, then it overrides the
715            mpeg format */
716 #if 0
717         if (is->video_st->codec->dtg_active_format != is->dtg_active_format) {
718             is->dtg_active_format = is->video_st->codec->dtg_active_format;
719             printf("dtg_active_format=%d\n", is->dtg_active_format);
720         }
721 #endif
722 #if 0
723         switch(is->video_st->codec->dtg_active_format) {
724         case FF_DTG_AFD_SAME:
725         default:
726             /* nothing to do */
727             break;
728         case FF_DTG_AFD_4_3:
729             aspect_ratio = 4.0 / 3.0;
730             break;
731         case FF_DTG_AFD_16_9:
732             aspect_ratio = 16.0 / 9.0;
733             break;
734         case FF_DTG_AFD_14_9:
735             aspect_ratio = 14.0 / 9.0;
736             break;
737         case FF_DTG_AFD_4_3_SP_14_9:
738             aspect_ratio = 14.0 / 9.0;
739             break;
740         case FF_DTG_AFD_16_9_SP_14_9:
741             aspect_ratio = 14.0 / 9.0;
742             break;
743         case FF_DTG_AFD_SP_4_3:
744             aspect_ratio = 4.0 / 3.0;
745             break;
746         }
747 #endif
748
749         if (is->subtitle_st)
750         {
751             if (is->subpq_size > 0)
752             {
753                 sp = &is->subpq[is->subpq_rindex];
754
755                 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000))
756                 {
757                     SDL_LockYUVOverlay (vp->bmp);
758
759                     pict.data[0] = vp->bmp->pixels[0];
760                     pict.data[1] = vp->bmp->pixels[2];
761                     pict.data[2] = vp->bmp->pixels[1];
762
763                     pict.linesize[0] = vp->bmp->pitches[0];
764                     pict.linesize[1] = vp->bmp->pitches[2];
765                     pict.linesize[2] = vp->bmp->pitches[1];
766
767                     for (i = 0; i < sp->sub.num_rects; i++)
768                         blend_subrect(&pict, sp->sub.rects[i],
769                                       vp->bmp->w, vp->bmp->h);
770
771                     SDL_UnlockYUVOverlay (vp->bmp);
772                 }
773             }
774         }
775
776
777         /* XXX: we suppose the screen has a 1.0 pixel ratio */
778         height = is->height;
779         width = ((int)rint(height * aspect_ratio)) & ~1;
780         if (width > is->width) {
781             width = is->width;
782             height = ((int)rint(width / aspect_ratio)) & ~1;
783         }
784         x = (is->width - width) / 2;
785         y = (is->height - height) / 2;
786         if (!is->no_background) {
787             /* fill the background */
788             //            fill_border(is, x, y, width, height, QERGB(0x00, 0x00, 0x00));
789         } else {
790             is->no_background = 0;
791         }
792         rect.x = is->xleft + x;
793         rect.y = is->ytop  + y;
794         rect.w = width;
795         rect.h = height;
796         SDL_DisplayYUVOverlay(vp->bmp, &rect);
797     } else {
798 #if 0
799         fill_rectangle(screen,
800                        is->xleft, is->ytop, is->width, is->height,
801                        QERGB(0x00, 0x00, 0x00));
802 #endif
803     }
804 }
805
806 static inline int compute_mod(int a, int b)
807 {
808     a = a % b;
809     if (a >= 0)
810         return a;
811     else
812         return a + b;
813 }
814
815 static void video_audio_display(VideoState *s)
816 {
817     int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
818     int ch, channels, h, h2, bgcolor, fgcolor;
819     int16_t time_diff;
820     int rdft_bits, nb_freq;
821
822     for(rdft_bits=1; (1<<rdft_bits)<2*s->height; rdft_bits++)
823         ;
824     nb_freq= 1<<(rdft_bits-1);
825
826     /* compute display index : center on currently output samples */
827     channels = s->audio_st->codec->channels;
828     nb_display_channels = channels;
829     if (!s->paused) {
830         int data_used= s->show_audio==1 ? s->width : (2*nb_freq);
831         n = 2 * channels;
832         delay = audio_write_get_buf_size(s);
833         delay /= n;
834
835         /* to be more precise, we take into account the time spent since
836            the last buffer computation */
837         if (audio_callback_time) {
838             time_diff = av_gettime() - audio_callback_time;
839             delay -= (time_diff * s->audio_st->codec->sample_rate) / 1000000;
840         }
841
842         delay += 2*data_used;
843         if (delay < data_used)
844             delay = data_used;
845
846         i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
847         if(s->show_audio==1){
848             h= INT_MIN;
849             for(i=0; i<1000; i+=channels){
850                 int idx= (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
851                 int a= s->sample_array[idx];
852                 int b= s->sample_array[(idx + 4*channels)%SAMPLE_ARRAY_SIZE];
853                 int c= s->sample_array[(idx + 5*channels)%SAMPLE_ARRAY_SIZE];
854                 int d= s->sample_array[(idx + 9*channels)%SAMPLE_ARRAY_SIZE];
855                 int score= a-d;
856                 if(h<score && (b^c)<0){
857                     h= score;
858                     i_start= idx;
859                 }
860             }
861         }
862
863         s->last_i_start = i_start;
864     } else {
865         i_start = s->last_i_start;
866     }
867
868     bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
869     if(s->show_audio==1){
870         fill_rectangle(screen,
871                        s->xleft, s->ytop, s->width, s->height,
872                        bgcolor);
873
874         fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
875
876         /* total height for one channel */
877         h = s->height / nb_display_channels;
878         /* graph height / 2 */
879         h2 = (h * 9) / 20;
880         for(ch = 0;ch < nb_display_channels; ch++) {
881             i = i_start + ch;
882             y1 = s->ytop + ch * h + (h / 2); /* position of center line */
883             for(x = 0; x < s->width; x++) {
884                 y = (s->sample_array[i] * h2) >> 15;
885                 if (y < 0) {
886                     y = -y;
887                     ys = y1 - y;
888                 } else {
889                     ys = y1;
890                 }
891                 fill_rectangle(screen,
892                                s->xleft + x, ys, 1, y,
893                                fgcolor);
894                 i += channels;
895                 if (i >= SAMPLE_ARRAY_SIZE)
896                     i -= SAMPLE_ARRAY_SIZE;
897             }
898         }
899
900         fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
901
902         for(ch = 1;ch < nb_display_channels; ch++) {
903             y = s->ytop + ch * h;
904             fill_rectangle(screen,
905                            s->xleft, y, s->width, 1,
906                            fgcolor);
907         }
908         SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
909     }else{
910         nb_display_channels= FFMIN(nb_display_channels, 2);
911         if(rdft_bits != s->rdft_bits){
912             av_rdft_end(s->rdft);
913             av_free(s->rdft_data);
914             s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
915             s->rdft_bits= rdft_bits;
916             s->rdft_data= av_malloc(4*nb_freq*sizeof(*s->rdft_data));
917         }
918         {
919             FFTSample *data[2];
920             for(ch = 0;ch < nb_display_channels; ch++) {
921                 data[ch] = s->rdft_data + 2*nb_freq*ch;
922                 i = i_start + ch;
923                 for(x = 0; x < 2*nb_freq; x++) {
924                     double w= (x-nb_freq)*(1.0/nb_freq);
925                     data[ch][x]= s->sample_array[i]*(1.0-w*w);
926                     i += channels;
927                     if (i >= SAMPLE_ARRAY_SIZE)
928                         i -= SAMPLE_ARRAY_SIZE;
929                 }
930                 av_rdft_calc(s->rdft, data[ch]);
931             }
932             //least efficient way to do this, we should of course directly access it but its more than fast enough
933             for(y=0; y<s->height; y++){
934                 double w= 1/sqrt(nb_freq);
935                 int a= sqrt(w*sqrt(data[0][2*y+0]*data[0][2*y+0] + data[0][2*y+1]*data[0][2*y+1]));
936                 int b= (nb_display_channels == 2 ) ? sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0]
937                        + data[1][2*y+1]*data[1][2*y+1])) : a;
938                 a= FFMIN(a,255);
939                 b= FFMIN(b,255);
940                 fgcolor = SDL_MapRGB(screen->format, a, b, (a+b)/2);
941
942                 fill_rectangle(screen,
943                             s->xpos, s->height-y, 1, 1,
944                             fgcolor);
945             }
946         }
947         SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
948         s->xpos++;
949         if(s->xpos >= s->width)
950             s->xpos= s->xleft;
951     }
952 }
953
954 static int video_open(VideoState *is){
955     int flags = SDL_HWSURFACE|SDL_ASYNCBLIT|SDL_HWACCEL;
956     int w,h;
957
958     if(is_full_screen) flags |= SDL_FULLSCREEN;
959     else               flags |= SDL_RESIZABLE;
960
961     if (is_full_screen && fs_screen_width) {
962         w = fs_screen_width;
963         h = fs_screen_height;
964     } else if(!is_full_screen && screen_width){
965         w = screen_width;
966         h = screen_height;
967 #if CONFIG_AVFILTER
968     }else if (is->out_video_filter && is->out_video_filter->inputs[0]){
969         w = is->out_video_filter->inputs[0]->w;
970         h = is->out_video_filter->inputs[0]->h;
971 #else
972     }else if (is->video_st && is->video_st->codec->width){
973         w = is->video_st->codec->width;
974         h = is->video_st->codec->height;
975 #endif
976     } else {
977         w = 640;
978         h = 480;
979     }
980     if(screen && is->width == screen->w && screen->w == w
981        && is->height== screen->h && screen->h == h)
982         return 0;
983
984 #ifndef __APPLE__
985     screen = SDL_SetVideoMode(w, h, 0, flags);
986 #else
987     /* setting bits_per_pixel = 0 or 32 causes blank video on OS X */
988     screen = SDL_SetVideoMode(w, h, 24, flags);
989 #endif
990     if (!screen) {
991         fprintf(stderr, "SDL: could not set video mode - exiting\n");
992         return -1;
993     }
994     if (!window_title)
995         window_title = input_filename;
996     SDL_WM_SetCaption(window_title, window_title);
997
998     is->width = screen->w;
999     is->height = screen->h;
1000
1001     return 0;
1002 }
1003
1004 /* display the current picture, if any */
1005 static void video_display(VideoState *is)
1006 {
1007     if(!screen)
1008         video_open(cur_stream);
1009     if (is->audio_st && is->show_audio)
1010         video_audio_display(is);
1011     else if (is->video_st)
1012         video_image_display(is);
1013 }
1014
1015 static int refresh_thread(void *opaque)
1016 {
1017     VideoState *is= opaque;
1018     while(!is->abort_request){
1019     SDL_Event event;
1020     event.type = FF_REFRESH_EVENT;
1021     event.user.data1 = opaque;
1022         if(!is->refresh){
1023             is->refresh=1;
1024     SDL_PushEvent(&event);
1025         }
1026         usleep(is->audio_st && is->show_audio ? rdftspeed*1000 : 5000); //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
1027     }
1028     return 0;
1029 }
1030
1031 /* get the current audio clock value */
1032 static double get_audio_clock(VideoState *is)
1033 {
1034     double pts;
1035     int hw_buf_size, bytes_per_sec;
1036     pts = is->audio_clock;
1037     hw_buf_size = audio_write_get_buf_size(is);
1038     bytes_per_sec = 0;
1039     if (is->audio_st) {
1040         bytes_per_sec = is->audio_st->codec->sample_rate *
1041             2 * is->audio_st->codec->channels;
1042     }
1043     if (bytes_per_sec)
1044         pts -= (double)hw_buf_size / bytes_per_sec;
1045     return pts;
1046 }
1047
1048 /* get the current video clock value */
1049 static double get_video_clock(VideoState *is)
1050 {
1051     if (is->paused) {
1052         return is->video_current_pts;
1053     } else {
1054         return is->video_current_pts_drift + av_gettime() / 1000000.0;
1055     }
1056 }
1057
1058 /* get the current external clock value */
1059 static double get_external_clock(VideoState *is)
1060 {
1061     int64_t ti;
1062     ti = av_gettime();
1063     return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1064 }
1065
1066 /* get the current master clock value */
1067 static double get_master_clock(VideoState *is)
1068 {
1069     double val;
1070
1071     if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1072         if (is->video_st)
1073             val = get_video_clock(is);
1074         else
1075             val = get_audio_clock(is);
1076     } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1077         if (is->audio_st)
1078             val = get_audio_clock(is);
1079         else
1080             val = get_video_clock(is);
1081     } else {
1082         val = get_external_clock(is);
1083     }
1084     return val;
1085 }
1086
1087 /* seek in the stream */
1088 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1089 {
1090     if (!is->seek_req) {
1091         is->seek_pos = pos;
1092         is->seek_rel = rel;
1093         is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1094         if (seek_by_bytes)
1095             is->seek_flags |= AVSEEK_FLAG_BYTE;
1096         is->seek_req = 1;
1097     }
1098 }
1099
1100 /* pause or resume the video */
1101 static void stream_pause(VideoState *is)
1102 {
1103     if (is->paused) {
1104         is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1105         if(is->read_pause_return != AVERROR(ENOSYS)){
1106             is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1107         }
1108         is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1109     }
1110     is->paused = !is->paused;
1111 }
1112
1113 static double compute_target_time(double frame_current_pts, VideoState *is)
1114 {
1115     double delay, sync_threshold, diff;
1116
1117     /* compute nominal delay */
1118     delay = frame_current_pts - is->frame_last_pts;
1119     if (delay <= 0 || delay >= 10.0) {
1120         /* if incorrect delay, use previous one */
1121         delay = is->frame_last_delay;
1122     } else {
1123         is->frame_last_delay = delay;
1124     }
1125     is->frame_last_pts = frame_current_pts;
1126
1127     /* update delay to follow master synchronisation source */
1128     if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1129          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1130         /* if video is slave, we try to correct big delays by
1131            duplicating or deleting a frame */
1132         diff = get_video_clock(is) - get_master_clock(is);
1133
1134         /* skip or repeat frame. We take into account the
1135            delay to compute the threshold. I still don't know
1136            if it is the best guess */
1137         sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1138         if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1139             if (diff <= -sync_threshold)
1140                 delay = 0;
1141             else if (diff >= sync_threshold)
1142                 delay = 2 * delay;
1143         }
1144     }
1145     is->frame_timer += delay;
1146 #if defined(DEBUG_SYNC)
1147     printf("video: delay=%0.3f actual_delay=%0.3f pts=%0.3f A-V=%f\n",
1148             delay, actual_delay, frame_current_pts, -diff);
1149 #endif
1150
1151     return is->frame_timer;
1152 }
1153
1154 /* called to display each frame */
1155 static void video_refresh_timer(void *opaque)
1156 {
1157     VideoState *is = opaque;
1158     VideoPicture *vp;
1159
1160     SubPicture *sp, *sp2;
1161
1162     if (is->video_st) {
1163 retry:
1164         if (is->pictq_size == 0) {
1165             //nothing to do, no picture to display in the que
1166         } else {
1167             double time= av_gettime()/1000000.0;
1168             double next_target;
1169             /* dequeue the picture */
1170             vp = &is->pictq[is->pictq_rindex];
1171
1172             if(time < vp->target_clock)
1173                 return;
1174             /* update current video pts */
1175             is->video_current_pts = vp->pts;
1176             is->video_current_pts_drift = is->video_current_pts - time;
1177             is->video_current_pos = vp->pos;
1178             if(is->pictq_size > 1){
1179                 VideoPicture *nextvp= &is->pictq[(is->pictq_rindex+1)%VIDEO_PICTURE_QUEUE_SIZE];
1180                 assert(nextvp->target_clock >= vp->target_clock);
1181                 next_target= nextvp->target_clock;
1182             }else{
1183                 next_target= vp->target_clock + is->video_clock - vp->pts; //FIXME pass durations cleanly
1184             }
1185             if(framedrop && time > next_target){
1186                 is->skip_frames *= 1.0 + FRAME_SKIP_FACTOR;
1187                 if(is->pictq_size > 1 || time > next_target + 0.5){
1188                     /* update queue size and signal for next picture */
1189                     if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1190                         is->pictq_rindex = 0;
1191
1192                     SDL_LockMutex(is->pictq_mutex);
1193                     is->pictq_size--;
1194                     SDL_CondSignal(is->pictq_cond);
1195                     SDL_UnlockMutex(is->pictq_mutex);
1196                     goto retry;
1197                 }
1198             }
1199
1200             if(is->subtitle_st) {
1201                 if (is->subtitle_stream_changed) {
1202                     SDL_LockMutex(is->subpq_mutex);
1203
1204                     while (is->subpq_size) {
1205                         free_subpicture(&is->subpq[is->subpq_rindex]);
1206
1207                         /* update queue size and signal for next picture */
1208                         if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1209                             is->subpq_rindex = 0;
1210
1211                         is->subpq_size--;
1212                     }
1213                     is->subtitle_stream_changed = 0;
1214
1215                     SDL_CondSignal(is->subpq_cond);
1216                     SDL_UnlockMutex(is->subpq_mutex);
1217                 } else {
1218                     if (is->subpq_size > 0) {
1219                         sp = &is->subpq[is->subpq_rindex];
1220
1221                         if (is->subpq_size > 1)
1222                             sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1223                         else
1224                             sp2 = NULL;
1225
1226                         if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1227                                 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1228                         {
1229                             free_subpicture(sp);
1230
1231                             /* update queue size and signal for next picture */
1232                             if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1233                                 is->subpq_rindex = 0;
1234
1235                             SDL_LockMutex(is->subpq_mutex);
1236                             is->subpq_size--;
1237                             SDL_CondSignal(is->subpq_cond);
1238                             SDL_UnlockMutex(is->subpq_mutex);
1239                         }
1240                     }
1241                 }
1242             }
1243
1244             /* display picture */
1245             video_display(is);
1246
1247             /* update queue size and signal for next picture */
1248             if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1249                 is->pictq_rindex = 0;
1250
1251             SDL_LockMutex(is->pictq_mutex);
1252             is->pictq_size--;
1253             SDL_CondSignal(is->pictq_cond);
1254             SDL_UnlockMutex(is->pictq_mutex);
1255         }
1256     } else if (is->audio_st) {
1257         /* draw the next audio frame */
1258
1259         /* if only audio stream, then display the audio bars (better
1260            than nothing, just to test the implementation */
1261
1262         /* display picture */
1263         video_display(is);
1264     }
1265     if (show_status) {
1266         static int64_t last_time;
1267         int64_t cur_time;
1268         int aqsize, vqsize, sqsize;
1269         double av_diff;
1270
1271         cur_time = av_gettime();
1272         if (!last_time || (cur_time - last_time) >= 30000) {
1273             aqsize = 0;
1274             vqsize = 0;
1275             sqsize = 0;
1276             if (is->audio_st)
1277                 aqsize = is->audioq.size;
1278             if (is->video_st)
1279                 vqsize = is->videoq.size;
1280             if (is->subtitle_st)
1281                 sqsize = is->subtitleq.size;
1282             av_diff = 0;
1283             if (is->audio_st && is->video_st)
1284                 av_diff = get_audio_clock(is) - get_video_clock(is);
1285             printf("%7.2f A-V:%7.3f s:%3.1f aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64"   \r",
1286                    get_master_clock(is), av_diff, FFMAX(is->skip_frames-1, 0), aqsize / 1024, vqsize / 1024, sqsize, is->pts_ctx.num_faulty_dts, is->pts_ctx.num_faulty_pts);
1287             fflush(stdout);
1288             last_time = cur_time;
1289         }
1290     }
1291 }
1292
1293 static void stream_close(VideoState *is)
1294 {
1295     VideoPicture *vp;
1296     int i;
1297     /* XXX: use a special url_shutdown call to abort parse cleanly */
1298     is->abort_request = 1;
1299     SDL_WaitThread(is->parse_tid, NULL);
1300     SDL_WaitThread(is->refresh_tid, NULL);
1301
1302     /* free all pictures */
1303     for(i=0;i<VIDEO_PICTURE_QUEUE_SIZE; i++) {
1304         vp = &is->pictq[i];
1305 #if CONFIG_AVFILTER
1306         if (vp->picref) {
1307             avfilter_unref_buffer(vp->picref);
1308             vp->picref = NULL;
1309         }
1310 #endif
1311         if (vp->bmp) {
1312             SDL_FreeYUVOverlay(vp->bmp);
1313             vp->bmp = NULL;
1314         }
1315     }
1316     SDL_DestroyMutex(is->pictq_mutex);
1317     SDL_DestroyCond(is->pictq_cond);
1318     SDL_DestroyMutex(is->subpq_mutex);
1319     SDL_DestroyCond(is->subpq_cond);
1320 #if !CONFIG_AVFILTER
1321     if (is->img_convert_ctx)
1322         sws_freeContext(is->img_convert_ctx);
1323 #endif
1324     av_free(is);
1325 }
1326
1327 static void do_exit(void)
1328 {
1329     if (cur_stream) {
1330         stream_close(cur_stream);
1331         cur_stream = NULL;
1332     }
1333     uninit_opts();
1334 #if CONFIG_AVFILTER
1335     avfilter_uninit();
1336 #endif
1337     if (show_status)
1338         printf("\n");
1339     SDL_Quit();
1340     av_log(NULL, AV_LOG_QUIET, "");
1341     exit(0);
1342 }
1343
1344 /* allocate a picture (needs to do that in main thread to avoid
1345    potential locking problems */
1346 static void alloc_picture(void *opaque)
1347 {
1348     VideoState *is = opaque;
1349     VideoPicture *vp;
1350
1351     vp = &is->pictq[is->pictq_windex];
1352
1353     if (vp->bmp)
1354         SDL_FreeYUVOverlay(vp->bmp);
1355
1356 #if CONFIG_AVFILTER
1357     if (vp->picref)
1358         avfilter_unref_buffer(vp->picref);
1359     vp->picref = NULL;
1360
1361     vp->width   = is->out_video_filter->inputs[0]->w;
1362     vp->height  = is->out_video_filter->inputs[0]->h;
1363     vp->pix_fmt = is->out_video_filter->inputs[0]->format;
1364 #else
1365     vp->width   = is->video_st->codec->width;
1366     vp->height  = is->video_st->codec->height;
1367     vp->pix_fmt = is->video_st->codec->pix_fmt;
1368 #endif
1369
1370     vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1371                                    SDL_YV12_OVERLAY,
1372                                    screen);
1373     if (!vp->bmp || vp->bmp->pitches[0] < vp->width) {
1374         /* SDL allocates a buffer smaller than requested if the video
1375          * overlay hardware is unable to support the requested size. */
1376         fprintf(stderr, "Error: the video system does not support an image\n"
1377                         "size of %dx%d pixels. Try using -lowres or -vf \"scale=w:h\"\n"
1378                         "to reduce the image size.\n", vp->width, vp->height );
1379         do_exit();
1380     }
1381
1382     SDL_LockMutex(is->pictq_mutex);
1383     vp->allocated = 1;
1384     SDL_CondSignal(is->pictq_cond);
1385     SDL_UnlockMutex(is->pictq_mutex);
1386 }
1387
1388 /**
1389  *
1390  * @param pts the dts of the pkt / pts of the frame and guessed if not known
1391  */
1392 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts, int64_t pos)
1393 {
1394     VideoPicture *vp;
1395     int dst_pix_fmt;
1396 #if CONFIG_AVFILTER
1397     AVPicture pict_src;
1398 #endif
1399     /* wait until we have space to put a new picture */
1400     SDL_LockMutex(is->pictq_mutex);
1401
1402     if(is->pictq_size>=VIDEO_PICTURE_QUEUE_SIZE && !is->refresh)
1403         is->skip_frames= FFMAX(1.0 - FRAME_SKIP_FACTOR, is->skip_frames * (1.0-FRAME_SKIP_FACTOR));
1404
1405     while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1406            !is->videoq.abort_request) {
1407         SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1408     }
1409     SDL_UnlockMutex(is->pictq_mutex);
1410
1411     if (is->videoq.abort_request)
1412         return -1;
1413
1414     vp = &is->pictq[is->pictq_windex];
1415
1416     /* alloc or resize hardware picture buffer */
1417     if (!vp->bmp ||
1418 #if CONFIG_AVFILTER
1419         vp->width  != is->out_video_filter->inputs[0]->w ||
1420         vp->height != is->out_video_filter->inputs[0]->h) {
1421 #else
1422         vp->width != is->video_st->codec->width ||
1423         vp->height != is->video_st->codec->height) {
1424 #endif
1425         SDL_Event event;
1426
1427         vp->allocated = 0;
1428
1429         /* the allocation must be done in the main thread to avoid
1430            locking problems */
1431         event.type = FF_ALLOC_EVENT;
1432         event.user.data1 = is;
1433         SDL_PushEvent(&event);
1434
1435         /* wait until the picture is allocated */
1436         SDL_LockMutex(is->pictq_mutex);
1437         while (!vp->allocated && !is->videoq.abort_request) {
1438             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1439         }
1440         SDL_UnlockMutex(is->pictq_mutex);
1441
1442         if (is->videoq.abort_request)
1443             return -1;
1444     }
1445
1446     /* if the frame is not skipped, then display it */
1447     if (vp->bmp) {
1448         AVPicture pict;
1449 #if CONFIG_AVFILTER
1450         if(vp->picref)
1451             avfilter_unref_buffer(vp->picref);
1452         vp->picref = src_frame->opaque;
1453 #endif
1454
1455         /* get a pointer on the bitmap */
1456         SDL_LockYUVOverlay (vp->bmp);
1457
1458         dst_pix_fmt = PIX_FMT_YUV420P;
1459         memset(&pict,0,sizeof(AVPicture));
1460         pict.data[0] = vp->bmp->pixels[0];
1461         pict.data[1] = vp->bmp->pixels[2];
1462         pict.data[2] = vp->bmp->pixels[1];
1463
1464         pict.linesize[0] = vp->bmp->pitches[0];
1465         pict.linesize[1] = vp->bmp->pitches[2];
1466         pict.linesize[2] = vp->bmp->pitches[1];
1467
1468 #if CONFIG_AVFILTER
1469         pict_src.data[0] = src_frame->data[0];
1470         pict_src.data[1] = src_frame->data[1];
1471         pict_src.data[2] = src_frame->data[2];
1472
1473         pict_src.linesize[0] = src_frame->linesize[0];
1474         pict_src.linesize[1] = src_frame->linesize[1];
1475         pict_src.linesize[2] = src_frame->linesize[2];
1476
1477         //FIXME use direct rendering
1478         av_picture_copy(&pict, &pict_src,
1479                         vp->pix_fmt, vp->width, vp->height);
1480 #else
1481         sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1482         is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1483             vp->width, vp->height, vp->pix_fmt, vp->width, vp->height,
1484             dst_pix_fmt, sws_flags, NULL, NULL, NULL);
1485         if (is->img_convert_ctx == NULL) {
1486             fprintf(stderr, "Cannot initialize the conversion context\n");
1487             exit(1);
1488         }
1489         sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1490                   0, vp->height, pict.data, pict.linesize);
1491 #endif
1492         /* update the bitmap content */
1493         SDL_UnlockYUVOverlay(vp->bmp);
1494
1495         vp->pts = pts;
1496         vp->pos = pos;
1497
1498         /* now we can update the picture count */
1499         if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1500             is->pictq_windex = 0;
1501         SDL_LockMutex(is->pictq_mutex);
1502         vp->target_clock= compute_target_time(vp->pts, is);
1503
1504         is->pictq_size++;
1505         SDL_UnlockMutex(is->pictq_mutex);
1506     }
1507     return 0;
1508 }
1509
1510 /**
1511  * compute the exact PTS for the picture if it is omitted in the stream
1512  * @param pts1 the dts of the pkt / pts of the frame
1513  */
1514 static int output_picture2(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1515 {
1516     double frame_delay, pts;
1517
1518     pts = pts1;
1519
1520     if (pts != 0) {
1521         /* update video clock with pts, if present */
1522         is->video_clock = pts;
1523     } else {
1524         pts = is->video_clock;
1525     }
1526     /* update video clock for next frame */
1527     frame_delay = av_q2d(is->video_st->codec->time_base);
1528     /* for MPEG2, the frame can be repeated, so we update the
1529        clock accordingly */
1530     frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1531     is->video_clock += frame_delay;
1532
1533 #if defined(DEBUG_SYNC) && 0
1534     printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1535            av_get_pict_type_char(src_frame->pict_type), pts, pts1);
1536 #endif
1537     return queue_picture(is, src_frame, pts, pos);
1538 }
1539
1540 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1541 {
1542     int len1, got_picture, i;
1543
1544         if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1545             return -1;
1546
1547         if(pkt->data == flush_pkt.data){
1548             avcodec_flush_buffers(is->video_st->codec);
1549
1550             SDL_LockMutex(is->pictq_mutex);
1551             //Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1552             for(i=0; i<VIDEO_PICTURE_QUEUE_SIZE; i++){
1553                 is->pictq[i].target_clock= 0;
1554             }
1555             while (is->pictq_size && !is->videoq.abort_request) {
1556                 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1557             }
1558             is->video_current_pos= -1;
1559             SDL_UnlockMutex(is->pictq_mutex);
1560
1561             init_pts_correction(&is->pts_ctx);
1562             is->frame_last_pts= AV_NOPTS_VALUE;
1563             is->frame_last_delay = 0;
1564             is->frame_timer = (double)av_gettime() / 1000000.0;
1565             is->skip_frames= 1;
1566             is->skip_frames_index= 0;
1567             return 0;
1568         }
1569
1570         /* NOTE: ipts is the PTS of the _first_ picture beginning in
1571            this packet, if any */
1572         is->video_st->codec->reordered_opaque= pkt->pts;
1573         len1 = avcodec_decode_video2(is->video_st->codec,
1574                                     frame, &got_picture,
1575                                     pkt);
1576
1577         if (got_picture) {
1578             if (decoder_reorder_pts == -1) {
1579                 *pts = guess_correct_pts(&is->pts_ctx, frame->reordered_opaque, pkt->dts);
1580             } else if (decoder_reorder_pts) {
1581                 *pts = frame->reordered_opaque;
1582             } else {
1583                 *pts = pkt->dts;
1584             }
1585
1586             if (*pts == AV_NOPTS_VALUE) {
1587                 *pts = 0;
1588             }
1589         }
1590
1591 //            if (len1 < 0)
1592 //                break;
1593     if (got_picture){
1594         is->skip_frames_index += 1;
1595         if(is->skip_frames_index >= is->skip_frames){
1596             is->skip_frames_index -= FFMAX(is->skip_frames, 1.0);
1597             return 1;
1598         }
1599
1600     }
1601     return 0;
1602 }
1603
1604 #if CONFIG_AVFILTER
1605 typedef struct {
1606     VideoState *is;
1607     AVFrame *frame;
1608     int use_dr1;
1609 } FilterPriv;
1610
1611 static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
1612 {
1613     AVFilterContext *ctx = codec->opaque;
1614     AVFilterBufferRef  *ref;
1615     int perms = AV_PERM_WRITE;
1616     int i, w, h, stride[4];
1617     unsigned edge;
1618
1619     if(pic->buffer_hints & FF_BUFFER_HINTS_VALID) {
1620         if(pic->buffer_hints & FF_BUFFER_HINTS_READABLE) perms |= AV_PERM_READ;
1621         if(pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) perms |= AV_PERM_PRESERVE;
1622         if(pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) perms |= AV_PERM_REUSE2;
1623     }
1624     if(pic->reference) perms |= AV_PERM_READ | AV_PERM_PRESERVE;
1625
1626     w = codec->width;
1627     h = codec->height;
1628     avcodec_align_dimensions2(codec, &w, &h, stride);
1629     edge = codec->flags & CODEC_FLAG_EMU_EDGE ? 0 : avcodec_get_edge_width();
1630     w += edge << 1;
1631     h += edge << 1;
1632
1633     if(!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
1634         return -1;
1635
1636     ref->video->w = codec->width;
1637     ref->video->h = codec->height;
1638     for(i = 0; i < 4; i ++) {
1639         unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_w : 0;
1640         unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_h : 0;
1641
1642         if (ref->data[i]) {
1643             ref->data[i]    += (edge >> hshift) + ((edge * ref->linesize[i]) >> vshift);
1644         }
1645         pic->data[i]     = ref->data[i];
1646         pic->linesize[i] = ref->linesize[i];
1647     }
1648     pic->opaque = ref;
1649     pic->age    = INT_MAX;
1650     pic->type   = FF_BUFFER_TYPE_USER;
1651     pic->reordered_opaque = codec->reordered_opaque;
1652     return 0;
1653 }
1654
1655 static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
1656 {
1657     memset(pic->data, 0, sizeof(pic->data));
1658     avfilter_unref_buffer(pic->opaque);
1659 }
1660
1661 static int input_reget_buffer(AVCodecContext *codec, AVFrame *pic)
1662 {
1663     AVFilterBufferRef *ref = pic->opaque;
1664
1665     if (pic->data[0] == NULL) {
1666         pic->buffer_hints |= FF_BUFFER_HINTS_READABLE;
1667         return codec->get_buffer(codec, pic);
1668     }
1669
1670     if ((codec->width != ref->video->w) || (codec->height != ref->video->h) ||
1671         (codec->pix_fmt != ref->format)) {
1672         av_log(codec, AV_LOG_ERROR, "Picture properties changed.\n");
1673         return -1;
1674     }
1675
1676     pic->reordered_opaque = codec->reordered_opaque;
1677     return 0;
1678 }
1679
1680 static int input_init(AVFilterContext *ctx, const char *args, void *opaque)
1681 {
1682     FilterPriv *priv = ctx->priv;
1683     AVCodecContext *codec;
1684     if(!opaque) return -1;
1685
1686     priv->is = opaque;
1687     codec    = priv->is->video_st->codec;
1688     codec->opaque = ctx;
1689     if(codec->codec->capabilities & CODEC_CAP_DR1) {
1690         priv->use_dr1 = 1;
1691         codec->get_buffer     = input_get_buffer;
1692         codec->release_buffer = input_release_buffer;
1693         codec->reget_buffer   = input_reget_buffer;
1694     }
1695
1696     priv->frame = avcodec_alloc_frame();
1697
1698     return 0;
1699 }
1700
1701 static void input_uninit(AVFilterContext *ctx)
1702 {
1703     FilterPriv *priv = ctx->priv;
1704     av_free(priv->frame);
1705 }
1706
1707 static int input_request_frame(AVFilterLink *link)
1708 {
1709     FilterPriv *priv = link->src->priv;
1710     AVFilterBufferRef *picref;
1711     int64_t pts = 0;
1712     AVPacket pkt;
1713     int ret;
1714
1715     while (!(ret = get_video_frame(priv->is, priv->frame, &pts, &pkt)))
1716         av_free_packet(&pkt);
1717     if (ret < 0)
1718         return -1;
1719
1720     if(priv->use_dr1) {
1721         picref = avfilter_ref_buffer(priv->frame->opaque, ~0);
1722     } else {
1723         picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
1724         av_image_copy(picref->data, picref->linesize,
1725                       priv->frame->data, priv->frame->linesize,
1726                       picref->format, link->w, link->h);
1727     }
1728     av_free_packet(&pkt);
1729
1730     picref->pts = pts;
1731     picref->pos = pkt.pos;
1732     picref->video->pixel_aspect = priv->is->video_st->codec->sample_aspect_ratio;
1733     avfilter_start_frame(link, picref);
1734     avfilter_draw_slice(link, 0, link->h, 1);
1735     avfilter_end_frame(link);
1736
1737     return 0;
1738 }
1739
1740 static int input_query_formats(AVFilterContext *ctx)
1741 {
1742     FilterPriv *priv = ctx->priv;
1743     enum PixelFormat pix_fmts[] = {
1744         priv->is->video_st->codec->pix_fmt, PIX_FMT_NONE
1745     };
1746
1747     avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
1748     return 0;
1749 }
1750
1751 static int input_config_props(AVFilterLink *link)
1752 {
1753     FilterPriv *priv  = link->src->priv;
1754     AVCodecContext *c = priv->is->video_st->codec;
1755
1756     link->w = c->width;
1757     link->h = c->height;
1758
1759     return 0;
1760 }
1761
1762 static AVFilter input_filter =
1763 {
1764     .name      = "ffplay_input",
1765
1766     .priv_size = sizeof(FilterPriv),
1767
1768     .init      = input_init,
1769     .uninit    = input_uninit,
1770
1771     .query_formats = input_query_formats,
1772
1773     .inputs    = (AVFilterPad[]) {{ .name = NULL }},
1774     .outputs   = (AVFilterPad[]) {{ .name = "default",
1775                                     .type = AVMEDIA_TYPE_VIDEO,
1776                                     .request_frame = input_request_frame,
1777                                     .config_props  = input_config_props, },
1778                                   { .name = NULL }},
1779 };
1780
1781 static void output_end_frame(AVFilterLink *link)
1782 {
1783 }
1784
1785 static int output_query_formats(AVFilterContext *ctx)
1786 {
1787     enum PixelFormat pix_fmts[] = { PIX_FMT_YUV420P, PIX_FMT_NONE };
1788
1789     avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
1790     return 0;
1791 }
1792
1793 static int get_filtered_video_frame(AVFilterContext *ctx, AVFrame *frame,
1794                                     int64_t *pts, int64_t *pos)
1795 {
1796     AVFilterBufferRef *pic;
1797
1798     if(avfilter_request_frame(ctx->inputs[0]))
1799         return -1;
1800     if(!(pic = ctx->inputs[0]->cur_buf))
1801         return -1;
1802     ctx->inputs[0]->cur_buf = NULL;
1803
1804     frame->opaque = pic;
1805     *pts          = pic->pts;
1806     *pos          = pic->pos;
1807
1808     memcpy(frame->data,     pic->data,     sizeof(frame->data));
1809     memcpy(frame->linesize, pic->linesize, sizeof(frame->linesize));
1810
1811     return 1;
1812 }
1813
1814 static AVFilter output_filter =
1815 {
1816     .name      = "ffplay_output",
1817
1818     .query_formats = output_query_formats,
1819
1820     .inputs    = (AVFilterPad[]) {{ .name          = "default",
1821                                     .type          = AVMEDIA_TYPE_VIDEO,
1822                                     .end_frame     = output_end_frame,
1823                                     .min_perms     = AV_PERM_READ, },
1824                                   { .name = NULL }},
1825     .outputs   = (AVFilterPad[]) {{ .name = NULL }},
1826 };
1827 #endif  /* CONFIG_AVFILTER */
1828
1829 static int video_thread(void *arg)
1830 {
1831     VideoState *is = arg;
1832     AVFrame *frame= avcodec_alloc_frame();
1833     int64_t pts_int;
1834     double pts;
1835     int ret;
1836
1837 #if CONFIG_AVFILTER
1838     int64_t pos;
1839     char sws_flags_str[128];
1840     AVFilterContext *filt_src = NULL, *filt_out = NULL;
1841     AVFilterGraph *graph = av_mallocz(sizeof(AVFilterGraph));
1842     snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1843     graph->scale_sws_opts = av_strdup(sws_flags_str);
1844
1845     if (avfilter_open(&filt_src, &input_filter,  "src") < 0) goto the_end;
1846     if (avfilter_open(&filt_out, &output_filter, "out") < 0) goto the_end;
1847
1848     if(avfilter_init_filter(filt_src, NULL, is))             goto the_end;
1849     if(avfilter_init_filter(filt_out, NULL, frame))          goto the_end;
1850
1851
1852     if(vfilters) {
1853         AVFilterInOut *outputs = av_malloc(sizeof(AVFilterInOut));
1854         AVFilterInOut *inputs  = av_malloc(sizeof(AVFilterInOut));
1855
1856         outputs->name    = av_strdup("in");
1857         outputs->filter  = filt_src;
1858         outputs->pad_idx = 0;
1859         outputs->next    = NULL;
1860
1861         inputs->name    = av_strdup("out");
1862         inputs->filter  = filt_out;
1863         inputs->pad_idx = 0;
1864         inputs->next    = NULL;
1865
1866         if (avfilter_graph_parse(graph, vfilters, inputs, outputs, NULL) < 0)
1867             goto the_end;
1868         av_freep(&vfilters);
1869     } else {
1870         if(avfilter_link(filt_src, 0, filt_out, 0) < 0)          goto the_end;
1871     }
1872     avfilter_graph_add_filter(graph, filt_src);
1873     avfilter_graph_add_filter(graph, filt_out);
1874
1875     if(avfilter_graph_check_validity(graph, NULL))           goto the_end;
1876     if(avfilter_graph_config_formats(graph, NULL))           goto the_end;
1877     if(avfilter_graph_config_links(graph, NULL))             goto the_end;
1878
1879     is->out_video_filter = filt_out;
1880 #endif
1881
1882     for(;;) {
1883 #if !CONFIG_AVFILTER
1884         AVPacket pkt;
1885 #endif
1886         while (is->paused && !is->videoq.abort_request)
1887             SDL_Delay(10);
1888 #if CONFIG_AVFILTER
1889         ret = get_filtered_video_frame(filt_out, frame, &pts_int, &pos);
1890 #else
1891         ret = get_video_frame(is, frame, &pts_int, &pkt);
1892 #endif
1893
1894         if (ret < 0) goto the_end;
1895
1896         if (!ret)
1897             continue;
1898
1899         pts = pts_int*av_q2d(is->video_st->time_base);
1900
1901 #if CONFIG_AVFILTER
1902         ret = output_picture2(is, frame, pts, pos);
1903 #else
1904         ret = output_picture2(is, frame, pts,  pkt.pos);
1905         av_free_packet(&pkt);
1906 #endif
1907         if (ret < 0)
1908             goto the_end;
1909
1910         if (step)
1911             if (cur_stream)
1912                 stream_pause(cur_stream);
1913     }
1914  the_end:
1915 #if CONFIG_AVFILTER
1916     avfilter_graph_destroy(graph);
1917     av_freep(&graph);
1918 #endif
1919     av_free(frame);
1920     return 0;
1921 }
1922
1923 static int subtitle_thread(void *arg)
1924 {
1925     VideoState *is = arg;
1926     SubPicture *sp;
1927     AVPacket pkt1, *pkt = &pkt1;
1928     int len1, got_subtitle;
1929     double pts;
1930     int i, j;
1931     int r, g, b, y, u, v, a;
1932
1933     for(;;) {
1934         while (is->paused && !is->subtitleq.abort_request) {
1935             SDL_Delay(10);
1936         }
1937         if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1938             break;
1939
1940         if(pkt->data == flush_pkt.data){
1941             avcodec_flush_buffers(is->subtitle_st->codec);
1942             continue;
1943         }
1944         SDL_LockMutex(is->subpq_mutex);
1945         while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1946                !is->subtitleq.abort_request) {
1947             SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1948         }
1949         SDL_UnlockMutex(is->subpq_mutex);
1950
1951         if (is->subtitleq.abort_request)
1952             goto the_end;
1953
1954         sp = &is->subpq[is->subpq_windex];
1955
1956        /* NOTE: ipts is the PTS of the _first_ picture beginning in
1957            this packet, if any */
1958         pts = 0;
1959         if (pkt->pts != AV_NOPTS_VALUE)
1960             pts = av_q2d(is->subtitle_st->time_base)*pkt->pts;
1961
1962         len1 = avcodec_decode_subtitle2(is->subtitle_st->codec,
1963                                     &sp->sub, &got_subtitle,
1964                                     pkt);
1965 //            if (len1 < 0)
1966 //                break;
1967         if (got_subtitle && sp->sub.format == 0) {
1968             sp->pts = pts;
1969
1970             for (i = 0; i < sp->sub.num_rects; i++)
1971             {
1972                 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1973                 {
1974                     RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1975                     y = RGB_TO_Y_CCIR(r, g, b);
1976                     u = RGB_TO_U_CCIR(r, g, b, 0);
1977                     v = RGB_TO_V_CCIR(r, g, b, 0);
1978                     YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1979                 }
1980             }
1981
1982             /* now we can update the picture count */
1983             if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1984                 is->subpq_windex = 0;
1985             SDL_LockMutex(is->subpq_mutex);
1986             is->subpq_size++;
1987             SDL_UnlockMutex(is->subpq_mutex);
1988         }
1989         av_free_packet(pkt);
1990 //        if (step)
1991 //            if (cur_stream)
1992 //                stream_pause(cur_stream);
1993     }
1994  the_end:
1995     return 0;
1996 }
1997
1998 /* copy samples for viewing in editor window */
1999 static void update_sample_display(VideoState *is, short *samples, int samples_size)
2000 {
2001     int size, len, channels;
2002
2003     channels = is->audio_st->codec->channels;
2004
2005     size = samples_size / sizeof(short);
2006     while (size > 0) {
2007         len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
2008         if (len > size)
2009             len = size;
2010         memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
2011         samples += len;
2012         is->sample_array_index += len;
2013         if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
2014             is->sample_array_index = 0;
2015         size -= len;
2016     }
2017 }
2018
2019 /* return the new audio buffer size (samples can be added or deleted
2020    to get better sync if video or external master clock) */
2021 static int synchronize_audio(VideoState *is, short *samples,
2022                              int samples_size1, double pts)
2023 {
2024     int n, samples_size;
2025     double ref_clock;
2026
2027     n = 2 * is->audio_st->codec->channels;
2028     samples_size = samples_size1;
2029
2030     /* if not master, then we try to remove or add samples to correct the clock */
2031     if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
2032          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
2033         double diff, avg_diff;
2034         int wanted_size, min_size, max_size, nb_samples;
2035
2036         ref_clock = get_master_clock(is);
2037         diff = get_audio_clock(is) - ref_clock;
2038
2039         if (diff < AV_NOSYNC_THRESHOLD) {
2040             is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
2041             if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
2042                 /* not enough measures to have a correct estimate */
2043                 is->audio_diff_avg_count++;
2044             } else {
2045                 /* estimate the A-V difference */
2046                 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
2047
2048                 if (fabs(avg_diff) >= is->audio_diff_threshold) {
2049                     wanted_size = samples_size + ((int)(diff * is->audio_st->codec->sample_rate) * n);
2050                     nb_samples = samples_size / n;
2051
2052                     min_size = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2053                     max_size = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2054                     if (wanted_size < min_size)
2055                         wanted_size = min_size;
2056                     else if (wanted_size > max_size)
2057                         wanted_size = max_size;
2058
2059                     /* add or remove samples to correction the synchro */
2060                     if (wanted_size < samples_size) {
2061                         /* remove samples */
2062                         samples_size = wanted_size;
2063                     } else if (wanted_size > samples_size) {
2064                         uint8_t *samples_end, *q;
2065                         int nb;
2066
2067                         /* add samples */
2068                         nb = (samples_size - wanted_size);
2069                         samples_end = (uint8_t *)samples + samples_size - n;
2070                         q = samples_end + n;
2071                         while (nb > 0) {
2072                             memcpy(q, samples_end, n);
2073                             q += n;
2074                             nb -= n;
2075                         }
2076                         samples_size = wanted_size;
2077                     }
2078                 }
2079 #if 0
2080                 printf("diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
2081                        diff, avg_diff, samples_size - samples_size1,
2082                        is->audio_clock, is->video_clock, is->audio_diff_threshold);
2083 #endif
2084             }
2085         } else {
2086             /* too big difference : may be initial PTS errors, so
2087                reset A-V filter */
2088             is->audio_diff_avg_count = 0;
2089             is->audio_diff_cum = 0;
2090         }
2091     }
2092
2093     return samples_size;
2094 }
2095
2096 /* decode one audio frame and returns its uncompressed size */
2097 static int audio_decode_frame(VideoState *is, double *pts_ptr)
2098 {
2099     AVPacket *pkt_temp = &is->audio_pkt_temp;
2100     AVPacket *pkt = &is->audio_pkt;
2101     AVCodecContext *dec= is->audio_st->codec;
2102     int n, len1, data_size;
2103     double pts;
2104
2105     for(;;) {
2106         /* NOTE: the audio packet can contain several frames */
2107         while (pkt_temp->size > 0) {
2108             data_size = sizeof(is->audio_buf1);
2109             len1 = avcodec_decode_audio3(dec,
2110                                         (int16_t *)is->audio_buf1, &data_size,
2111                                         pkt_temp);
2112             if (len1 < 0) {
2113                 /* if error, we skip the frame */
2114                 pkt_temp->size = 0;
2115                 break;
2116             }
2117
2118             pkt_temp->data += len1;
2119             pkt_temp->size -= len1;
2120             if (data_size <= 0)
2121                 continue;
2122
2123             if (dec->sample_fmt != is->audio_src_fmt) {
2124                 if (is->reformat_ctx)
2125                     av_audio_convert_free(is->reformat_ctx);
2126                 is->reformat_ctx= av_audio_convert_alloc(SAMPLE_FMT_S16, 1,
2127                                                          dec->sample_fmt, 1, NULL, 0);
2128                 if (!is->reformat_ctx) {
2129                     fprintf(stderr, "Cannot convert %s sample format to %s sample format\n",
2130                         avcodec_get_sample_fmt_name(dec->sample_fmt),
2131                         avcodec_get_sample_fmt_name(SAMPLE_FMT_S16));
2132                         break;
2133                 }
2134                 is->audio_src_fmt= dec->sample_fmt;
2135             }
2136
2137             if (is->reformat_ctx) {
2138                 const void *ibuf[6]= {is->audio_buf1};
2139                 void *obuf[6]= {is->audio_buf2};
2140                 int istride[6]= {av_get_bits_per_sample_format(dec->sample_fmt)/8};
2141                 int ostride[6]= {2};
2142                 int len= data_size/istride[0];
2143                 if (av_audio_convert(is->reformat_ctx, obuf, ostride, ibuf, istride, len)<0) {
2144                     printf("av_audio_convert() failed\n");
2145                     break;
2146                 }
2147                 is->audio_buf= is->audio_buf2;
2148                 /* FIXME: existing code assume that data_size equals framesize*channels*2
2149                           remove this legacy cruft */
2150                 data_size= len*2;
2151             }else{
2152                 is->audio_buf= is->audio_buf1;
2153             }
2154
2155             /* if no pts, then compute it */
2156             pts = is->audio_clock;
2157             *pts_ptr = pts;
2158             n = 2 * dec->channels;
2159             is->audio_clock += (double)data_size /
2160                 (double)(n * dec->sample_rate);
2161 #if defined(DEBUG_SYNC)
2162             {
2163                 static double last_clock;
2164                 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2165                        is->audio_clock - last_clock,
2166                        is->audio_clock, pts);
2167                 last_clock = is->audio_clock;
2168             }
2169 #endif
2170             return data_size;
2171         }
2172
2173         /* free the current packet */
2174         if (pkt->data)
2175             av_free_packet(pkt);
2176
2177         if (is->paused || is->audioq.abort_request) {
2178             return -1;
2179         }
2180
2181         /* read next packet */
2182         if (packet_queue_get(&is->audioq, pkt, 1) < 0)
2183             return -1;
2184         if(pkt->data == flush_pkt.data){
2185             avcodec_flush_buffers(dec);
2186             continue;
2187         }
2188
2189         pkt_temp->data = pkt->data;
2190         pkt_temp->size = pkt->size;
2191
2192         /* if update the audio clock with the pts */
2193         if (pkt->pts != AV_NOPTS_VALUE) {
2194             is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2195         }
2196     }
2197 }
2198
2199 /* get the current audio output buffer size, in samples. With SDL, we
2200    cannot have a precise information */
2201 static int audio_write_get_buf_size(VideoState *is)
2202 {
2203     return is->audio_buf_size - is->audio_buf_index;
2204 }
2205
2206
2207 /* prepare a new audio buffer */
2208 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2209 {
2210     VideoState *is = opaque;
2211     int audio_size, len1;
2212     double pts;
2213
2214     audio_callback_time = av_gettime();
2215
2216     while (len > 0) {
2217         if (is->audio_buf_index >= is->audio_buf_size) {
2218            audio_size = audio_decode_frame(is, &pts);
2219            if (audio_size < 0) {
2220                 /* if error, just output silence */
2221                is->audio_buf = is->audio_buf1;
2222                is->audio_buf_size = 1024;
2223                memset(is->audio_buf, 0, is->audio_buf_size);
2224            } else {
2225                if (is->show_audio)
2226                    update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2227                audio_size = synchronize_audio(is, (int16_t *)is->audio_buf, audio_size,
2228                                               pts);
2229                is->audio_buf_size = audio_size;
2230            }
2231            is->audio_buf_index = 0;
2232         }
2233         len1 = is->audio_buf_size - is->audio_buf_index;
2234         if (len1 > len)
2235             len1 = len;
2236         memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2237         len -= len1;
2238         stream += len1;
2239         is->audio_buf_index += len1;
2240     }
2241 }
2242
2243 /* open a given stream. Return 0 if OK */
2244 static int stream_component_open(VideoState *is, int stream_index)
2245 {
2246     AVFormatContext *ic = is->ic;
2247     AVCodecContext *avctx;
2248     AVCodec *codec;
2249     SDL_AudioSpec wanted_spec, spec;
2250
2251     if (stream_index < 0 || stream_index >= ic->nb_streams)
2252         return -1;
2253     avctx = ic->streams[stream_index]->codec;
2254
2255     /* prepare audio output */
2256     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2257         if (avctx->channels > 0) {
2258             avctx->request_channels = FFMIN(2, avctx->channels);
2259         } else {
2260             avctx->request_channels = 2;
2261         }
2262     }
2263
2264     codec = avcodec_find_decoder(avctx->codec_id);
2265     avctx->debug_mv = debug_mv;
2266     avctx->debug = debug;
2267     avctx->workaround_bugs = workaround_bugs;
2268     avctx->lowres = lowres;
2269     if(lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2270     avctx->idct_algo= idct;
2271     if(fast) avctx->flags2 |= CODEC_FLAG2_FAST;
2272     avctx->skip_frame= skip_frame;
2273     avctx->skip_idct= skip_idct;
2274     avctx->skip_loop_filter= skip_loop_filter;
2275     avctx->error_recognition= error_recognition;
2276     avctx->error_concealment= error_concealment;
2277     avcodec_thread_init(avctx, thread_count);
2278
2279     set_context_opts(avctx, avcodec_opts[avctx->codec_type], 0, codec);
2280
2281     if (!codec ||
2282         avcodec_open(avctx, codec) < 0)
2283         return -1;
2284
2285     /* prepare audio output */
2286     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2287         wanted_spec.freq = avctx->sample_rate;
2288         wanted_spec.format = AUDIO_S16SYS;
2289         wanted_spec.channels = avctx->channels;
2290         wanted_spec.silence = 0;
2291         wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2292         wanted_spec.callback = sdl_audio_callback;
2293         wanted_spec.userdata = is;
2294         if (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2295             fprintf(stderr, "SDL_OpenAudio: %s\n", SDL_GetError());
2296             return -1;
2297         }
2298         is->audio_hw_buf_size = spec.size;
2299         is->audio_src_fmt= SAMPLE_FMT_S16;
2300     }
2301
2302     ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2303     switch(avctx->codec_type) {
2304     case AVMEDIA_TYPE_AUDIO:
2305         is->audio_stream = stream_index;
2306         is->audio_st = ic->streams[stream_index];
2307         is->audio_buf_size = 0;
2308         is->audio_buf_index = 0;
2309
2310         /* init averaging filter */
2311         is->audio_diff_avg_coef = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2312         is->audio_diff_avg_count = 0;
2313         /* since we do not have a precise anough audio fifo fullness,
2314            we correct audio sync only if larger than this threshold */
2315         is->audio_diff_threshold = 2.0 * SDL_AUDIO_BUFFER_SIZE / avctx->sample_rate;
2316
2317         memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2318         packet_queue_init(&is->audioq);
2319         SDL_PauseAudio(0);
2320         break;
2321     case AVMEDIA_TYPE_VIDEO:
2322         is->video_stream = stream_index;
2323         is->video_st = ic->streams[stream_index];
2324
2325 //        is->video_current_pts_time = av_gettime();
2326
2327         packet_queue_init(&is->videoq);
2328         is->video_tid = SDL_CreateThread(video_thread, is);
2329         break;
2330     case AVMEDIA_TYPE_SUBTITLE:
2331         is->subtitle_stream = stream_index;
2332         is->subtitle_st = ic->streams[stream_index];
2333         packet_queue_init(&is->subtitleq);
2334
2335         is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2336         break;
2337     default:
2338         break;
2339     }
2340     return 0;
2341 }
2342
2343 static void stream_component_close(VideoState *is, int stream_index)
2344 {
2345     AVFormatContext *ic = is->ic;
2346     AVCodecContext *avctx;
2347
2348     if (stream_index < 0 || stream_index >= ic->nb_streams)
2349         return;
2350     avctx = ic->streams[stream_index]->codec;
2351
2352     switch(avctx->codec_type) {
2353     case AVMEDIA_TYPE_AUDIO:
2354         packet_queue_abort(&is->audioq);
2355
2356         SDL_CloseAudio();
2357
2358         packet_queue_end(&is->audioq);
2359         if (is->reformat_ctx)
2360             av_audio_convert_free(is->reformat_ctx);
2361         is->reformat_ctx = NULL;
2362         break;
2363     case AVMEDIA_TYPE_VIDEO:
2364         packet_queue_abort(&is->videoq);
2365
2366         /* note: we also signal this mutex to make sure we deblock the
2367            video thread in all cases */
2368         SDL_LockMutex(is->pictq_mutex);
2369         SDL_CondSignal(is->pictq_cond);
2370         SDL_UnlockMutex(is->pictq_mutex);
2371
2372         SDL_WaitThread(is->video_tid, NULL);
2373
2374         packet_queue_end(&is->videoq);
2375         break;
2376     case AVMEDIA_TYPE_SUBTITLE:
2377         packet_queue_abort(&is->subtitleq);
2378
2379         /* note: we also signal this mutex to make sure we deblock the
2380            video thread in all cases */
2381         SDL_LockMutex(is->subpq_mutex);
2382         is->subtitle_stream_changed = 1;
2383
2384         SDL_CondSignal(is->subpq_cond);
2385         SDL_UnlockMutex(is->subpq_mutex);
2386
2387         SDL_WaitThread(is->subtitle_tid, NULL);
2388
2389         packet_queue_end(&is->subtitleq);
2390         break;
2391     default:
2392         break;
2393     }
2394
2395     ic->streams[stream_index]->discard = AVDISCARD_ALL;
2396     avcodec_close(avctx);
2397     switch(avctx->codec_type) {
2398     case AVMEDIA_TYPE_AUDIO:
2399         is->audio_st = NULL;
2400         is->audio_stream = -1;
2401         break;
2402     case AVMEDIA_TYPE_VIDEO:
2403         is->video_st = NULL;
2404         is->video_stream = -1;
2405         break;
2406     case AVMEDIA_TYPE_SUBTITLE:
2407         is->subtitle_st = NULL;
2408         is->subtitle_stream = -1;
2409         break;
2410     default:
2411         break;
2412     }
2413 }
2414
2415 /* since we have only one decoding thread, we can use a global
2416    variable instead of a thread local variable */
2417 static VideoState *global_video_state;
2418
2419 static int decode_interrupt_cb(void)
2420 {
2421     return (global_video_state && global_video_state->abort_request);
2422 }
2423
2424 /* this thread gets the stream from the disk or the network */
2425 static int decode_thread(void *arg)
2426 {
2427     VideoState *is = arg;
2428     AVFormatContext *ic;
2429     int err, i, ret;
2430     int st_index[AVMEDIA_TYPE_NB];
2431     int st_count[AVMEDIA_TYPE_NB]={0};
2432     int st_best_packet_count[AVMEDIA_TYPE_NB];
2433     AVPacket pkt1, *pkt = &pkt1;
2434     AVFormatParameters params, *ap = &params;
2435     int eof=0;
2436     int pkt_in_play_range = 0;
2437
2438     ic = avformat_alloc_context();
2439
2440     memset(st_index, -1, sizeof(st_index));
2441     memset(st_best_packet_count, -1, sizeof(st_best_packet_count));
2442     is->video_stream = -1;
2443     is->audio_stream = -1;
2444     is->subtitle_stream = -1;
2445
2446     global_video_state = is;
2447     url_set_interrupt_cb(decode_interrupt_cb);
2448
2449     memset(ap, 0, sizeof(*ap));
2450
2451     ap->prealloced_context = 1;
2452     ap->width = frame_width;
2453     ap->height= frame_height;
2454     ap->time_base= (AVRational){1, 25};
2455     ap->pix_fmt = frame_pix_fmt;
2456
2457     set_context_opts(ic, avformat_opts, AV_OPT_FLAG_DECODING_PARAM, NULL);
2458
2459     err = av_open_input_file(&ic, is->filename, is->iformat, 0, ap);
2460     if (err < 0) {
2461         print_error(is->filename, err);
2462         ret = -1;
2463         goto fail;
2464     }
2465     is->ic = ic;
2466
2467     if(genpts)
2468         ic->flags |= AVFMT_FLAG_GENPTS;
2469
2470     err = av_find_stream_info(ic);
2471     if (err < 0) {
2472         fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2473         ret = -1;
2474         goto fail;
2475     }
2476     if(ic->pb)
2477         ic->pb->eof_reached= 0; //FIXME hack, ffplay maybe should not use url_feof() to test for the end
2478
2479     if(seek_by_bytes<0)
2480         seek_by_bytes= !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2481
2482     /* if seeking requested, we execute it */
2483     if (start_time != AV_NOPTS_VALUE) {
2484         int64_t timestamp;
2485
2486         timestamp = start_time;
2487         /* add the stream start time */
2488         if (ic->start_time != AV_NOPTS_VALUE)
2489             timestamp += ic->start_time;
2490         ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2491         if (ret < 0) {
2492             fprintf(stderr, "%s: could not seek to position %0.3f\n",
2493                     is->filename, (double)timestamp / AV_TIME_BASE);
2494         }
2495     }
2496
2497     for(i = 0; i < ic->nb_streams; i++) {
2498         AVStream *st= ic->streams[i];
2499         AVCodecContext *avctx = st->codec;
2500         ic->streams[i]->discard = AVDISCARD_ALL;
2501         if(avctx->codec_type >= (unsigned)AVMEDIA_TYPE_NB)
2502             continue;
2503         if(st_count[avctx->codec_type]++ != wanted_stream[avctx->codec_type] && wanted_stream[avctx->codec_type] >= 0)
2504             continue;
2505
2506         if(st_best_packet_count[avctx->codec_type] >= st->codec_info_nb_frames)
2507             continue;
2508         st_best_packet_count[avctx->codec_type]= st->codec_info_nb_frames;
2509
2510         switch(avctx->codec_type) {
2511         case AVMEDIA_TYPE_AUDIO:
2512             if (!audio_disable)
2513                 st_index[AVMEDIA_TYPE_AUDIO] = i;
2514             break;
2515         case AVMEDIA_TYPE_VIDEO:
2516         case AVMEDIA_TYPE_SUBTITLE:
2517             if (!video_disable)
2518                 st_index[avctx->codec_type] = i;
2519             break;
2520         default:
2521             break;
2522         }
2523     }
2524     if (show_status) {
2525         dump_format(ic, 0, is->filename, 0);
2526     }
2527
2528     /* open the streams */
2529     if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2530         stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2531     }
2532
2533     ret=-1;
2534     if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2535         ret= stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2536     }
2537     is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2538     if(ret<0) {
2539         if (!display_disable)
2540             is->show_audio = 2;
2541     }
2542
2543     if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2544         stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2545     }
2546
2547     if (is->video_stream < 0 && is->audio_stream < 0) {
2548         fprintf(stderr, "%s: could not open codecs\n", is->filename);
2549         ret = -1;
2550         goto fail;
2551     }
2552
2553     for(;;) {
2554         if (is->abort_request)
2555             break;
2556         if (is->paused != is->last_paused) {
2557             is->last_paused = is->paused;
2558             if (is->paused)
2559                 is->read_pause_return= av_read_pause(ic);
2560             else
2561                 av_read_play(ic);
2562         }
2563 #if CONFIG_RTSP_DEMUXER
2564         if (is->paused && !strcmp(ic->iformat->name, "rtsp")) {
2565             /* wait 10 ms to avoid trying to get another packet */
2566             /* XXX: horrible */
2567             SDL_Delay(10);
2568             continue;
2569         }
2570 #endif
2571         if (is->seek_req) {
2572             int64_t seek_target= is->seek_pos;
2573             int64_t seek_min= is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2574             int64_t seek_max= is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2575 //FIXME the +-2 is due to rounding being not done in the correct direction in generation
2576 //      of the seek_pos/seek_rel variables
2577
2578             ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2579             if (ret < 0) {
2580                 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2581             }else{
2582                 if (is->audio_stream >= 0) {
2583                     packet_queue_flush(&is->audioq);
2584                     packet_queue_put(&is->audioq, &flush_pkt);
2585                 }
2586                 if (is->subtitle_stream >= 0) {
2587                     packet_queue_flush(&is->subtitleq);
2588                     packet_queue_put(&is->subtitleq, &flush_pkt);
2589                 }
2590                 if (is->video_stream >= 0) {
2591                     packet_queue_flush(&is->videoq);
2592                     packet_queue_put(&is->videoq, &flush_pkt);
2593                 }
2594             }
2595             is->seek_req = 0;
2596             eof= 0;
2597         }
2598
2599         /* if the queue are full, no need to read more */
2600         if (   is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2601             || (   (is->audioq   .size  > MIN_AUDIOQ_SIZE || is->audio_stream<0)
2602                 && (is->videoq   .nb_packets > MIN_FRAMES || is->video_stream<0)
2603                 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream<0))) {
2604             /* wait 10 ms */
2605             SDL_Delay(10);
2606             continue;
2607         }
2608         if(url_feof(ic->pb) || eof) {
2609             if(is->video_stream >= 0){
2610                 av_init_packet(pkt);
2611                 pkt->data=NULL;
2612                 pkt->size=0;
2613                 pkt->stream_index= is->video_stream;
2614                 packet_queue_put(&is->videoq, pkt);
2615             }
2616             SDL_Delay(10);
2617             if(is->audioq.size + is->videoq.size + is->subtitleq.size ==0){
2618                 if(loop!=1 && (!loop || --loop)){
2619                     stream_seek(cur_stream, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2620                 }else if(autoexit){
2621                     ret=AVERROR_EOF;
2622                     goto fail;
2623                 }
2624             }
2625             continue;
2626         }
2627         ret = av_read_frame(ic, pkt);
2628         if (ret < 0) {
2629             if (ret == AVERROR_EOF)
2630                 eof=1;
2631             if (url_ferror(ic->pb))
2632                 break;
2633             SDL_Delay(100); /* wait for user event */
2634             continue;
2635         }
2636         /* check if packet is in play range specified by user, then queue, otherwise discard */
2637         pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2638                 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2639                 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2640                 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0)/1000000
2641                 <= ((double)duration/1000000);
2642         if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2643             packet_queue_put(&is->audioq, pkt);
2644         } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2645             packet_queue_put(&is->videoq, pkt);
2646         } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2647             packet_queue_put(&is->subtitleq, pkt);
2648         } else {
2649             av_free_packet(pkt);
2650         }
2651     }
2652     /* wait until the end */
2653     while (!is->abort_request) {
2654         SDL_Delay(100);
2655     }
2656
2657     ret = 0;
2658  fail:
2659     /* disable interrupting */
2660     global_video_state = NULL;
2661
2662     /* close each stream */
2663     if (is->audio_stream >= 0)
2664         stream_component_close(is, is->audio_stream);
2665     if (is->video_stream >= 0)
2666         stream_component_close(is, is->video_stream);
2667     if (is->subtitle_stream >= 0)
2668         stream_component_close(is, is->subtitle_stream);
2669     if (is->ic) {
2670         av_close_input_file(is->ic);
2671         is->ic = NULL; /* safety */
2672     }
2673     url_set_interrupt_cb(NULL);
2674
2675     if (ret != 0) {
2676         SDL_Event event;
2677
2678         event.type = FF_QUIT_EVENT;
2679         event.user.data1 = is;
2680         SDL_PushEvent(&event);
2681     }
2682     return 0;
2683 }
2684
2685 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2686 {
2687     VideoState *is;
2688
2689     is = av_mallocz(sizeof(VideoState));
2690     if (!is)
2691         return NULL;
2692     av_strlcpy(is->filename, filename, sizeof(is->filename));
2693     is->iformat = iformat;
2694     is->ytop = 0;
2695     is->xleft = 0;
2696
2697     /* start video display */
2698     is->pictq_mutex = SDL_CreateMutex();
2699     is->pictq_cond = SDL_CreateCond();
2700
2701     is->subpq_mutex = SDL_CreateMutex();
2702     is->subpq_cond = SDL_CreateCond();
2703
2704     is->av_sync_type = av_sync_type;
2705     is->parse_tid = SDL_CreateThread(decode_thread, is);
2706     if (!is->parse_tid) {
2707         av_free(is);
2708         return NULL;
2709     }
2710     return is;
2711 }
2712
2713 static void stream_cycle_channel(VideoState *is, int codec_type)
2714 {
2715     AVFormatContext *ic = is->ic;
2716     int start_index, stream_index;
2717     AVStream *st;
2718
2719     if (codec_type == AVMEDIA_TYPE_VIDEO)
2720         start_index = is->video_stream;
2721     else if (codec_type == AVMEDIA_TYPE_AUDIO)
2722         start_index = is->audio_stream;
2723     else
2724         start_index = is->subtitle_stream;
2725     if (start_index < (codec_type == AVMEDIA_TYPE_SUBTITLE ? -1 : 0))
2726         return;
2727     stream_index = start_index;
2728     for(;;) {
2729         if (++stream_index >= is->ic->nb_streams)
2730         {
2731             if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2732             {
2733                 stream_index = -1;
2734                 goto the_end;
2735             } else
2736                 stream_index = 0;
2737         }
2738         if (stream_index == start_index)
2739             return;
2740         st = ic->streams[stream_index];
2741         if (st->codec->codec_type == codec_type) {
2742             /* check that parameters are OK */
2743             switch(codec_type) {
2744             case AVMEDIA_TYPE_AUDIO:
2745                 if (st->codec->sample_rate != 0 &&
2746                     st->codec->channels != 0)
2747                     goto the_end;
2748                 break;
2749             case AVMEDIA_TYPE_VIDEO:
2750             case AVMEDIA_TYPE_SUBTITLE:
2751                 goto the_end;
2752             default:
2753                 break;
2754             }
2755         }
2756     }
2757  the_end:
2758     stream_component_close(is, start_index);
2759     stream_component_open(is, stream_index);
2760 }
2761
2762
2763 static void toggle_full_screen(void)
2764 {
2765     is_full_screen = !is_full_screen;
2766     if (!fs_screen_width) {
2767         /* use default SDL method */
2768 //        SDL_WM_ToggleFullScreen(screen);
2769     }
2770     video_open(cur_stream);
2771 }
2772
2773 static void toggle_pause(void)
2774 {
2775     if (cur_stream)
2776         stream_pause(cur_stream);
2777     step = 0;
2778 }
2779
2780 static void step_to_next_frame(void)
2781 {
2782     if (cur_stream) {
2783         /* if the stream is paused unpause it, then step */
2784         if (cur_stream->paused)
2785             stream_pause(cur_stream);
2786     }
2787     step = 1;
2788 }
2789
2790 static void toggle_audio_display(void)
2791 {
2792     if (cur_stream) {
2793         int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2794         cur_stream->show_audio = (cur_stream->show_audio + 1) % 3;
2795         fill_rectangle(screen,
2796                     cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height,
2797                     bgcolor);
2798         SDL_UpdateRect(screen, cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height);
2799     }
2800 }
2801
2802 /* handle an event sent by the GUI */
2803 static void event_loop(void)
2804 {
2805     SDL_Event event;
2806     double incr, pos, frac;
2807
2808     for(;;) {
2809         double x;
2810         SDL_WaitEvent(&event);
2811         switch(event.type) {
2812         case SDL_KEYDOWN:
2813             if (exit_on_keydown) {
2814                 do_exit();
2815                 break;
2816             }
2817             switch(event.key.keysym.sym) {
2818             case SDLK_ESCAPE:
2819             case SDLK_q:
2820                 do_exit();
2821                 break;
2822             case SDLK_f:
2823                 toggle_full_screen();
2824                 break;
2825             case SDLK_p:
2826             case SDLK_SPACE:
2827                 toggle_pause();
2828                 break;
2829             case SDLK_s: //S: Step to next frame
2830                 step_to_next_frame();
2831                 break;
2832             case SDLK_a:
2833                 if (cur_stream)
2834                     stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2835                 break;
2836             case SDLK_v:
2837                 if (cur_stream)
2838                     stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2839                 break;
2840             case SDLK_t:
2841                 if (cur_stream)
2842                     stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2843                 break;
2844             case SDLK_w:
2845                 toggle_audio_display();
2846                 break;
2847             case SDLK_LEFT:
2848                 incr = -10.0;
2849                 goto do_seek;
2850             case SDLK_RIGHT:
2851                 incr = 10.0;
2852                 goto do_seek;
2853             case SDLK_UP:
2854                 incr = 60.0;
2855                 goto do_seek;
2856             case SDLK_DOWN:
2857                 incr = -60.0;
2858             do_seek:
2859                 if (cur_stream) {
2860                     if (seek_by_bytes) {
2861                         if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos>=0){
2862                             pos= cur_stream->video_current_pos;
2863                         }else if(cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos>=0){
2864                             pos= cur_stream->audio_pkt.pos;
2865                         }else
2866                             pos = url_ftell(cur_stream->ic->pb);
2867                         if (cur_stream->ic->bit_rate)
2868                             incr *= cur_stream->ic->bit_rate / 8.0;
2869                         else
2870                             incr *= 180000.0;
2871                         pos += incr;
2872                         stream_seek(cur_stream, pos, incr, 1);
2873                     } else {
2874                         pos = get_master_clock(cur_stream);
2875                         pos += incr;
2876                         stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2877                     }
2878                 }
2879                 break;
2880             default:
2881                 break;
2882             }
2883             break;
2884         case SDL_MOUSEBUTTONDOWN:
2885             if (exit_on_mousedown) {
2886                 do_exit();
2887                 break;
2888             }
2889         case SDL_MOUSEMOTION:
2890             if(event.type ==SDL_MOUSEBUTTONDOWN){
2891                 x= event.button.x;
2892             }else{
2893                 if(event.motion.state != SDL_PRESSED)
2894                     break;
2895                 x= event.motion.x;
2896             }
2897             if (cur_stream) {
2898                 if(seek_by_bytes || cur_stream->ic->duration<=0){
2899                     uint64_t size=  url_fsize(cur_stream->ic->pb);
2900                     stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2901                 }else{
2902                     int64_t ts;
2903                     int ns, hh, mm, ss;
2904                     int tns, thh, tmm, tss;
2905                     tns = cur_stream->ic->duration/1000000LL;
2906                     thh = tns/3600;
2907                     tmm = (tns%3600)/60;
2908                     tss = (tns%60);
2909                     frac = x/cur_stream->width;
2910                     ns = frac*tns;
2911                     hh = ns/3600;
2912                     mm = (ns%3600)/60;
2913                     ss = (ns%60);
2914                     fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d)       \n", frac*100,
2915                             hh, mm, ss, thh, tmm, tss);
2916                     ts = frac*cur_stream->ic->duration;
2917                     if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2918                         ts += cur_stream->ic->start_time;
2919                     stream_seek(cur_stream, ts, 0, 0);
2920                 }
2921             }
2922             break;
2923         case SDL_VIDEORESIZE:
2924             if (cur_stream) {
2925                 screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2926                                           SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2927                 screen_width = cur_stream->width = event.resize.w;
2928                 screen_height= cur_stream->height= event.resize.h;
2929             }
2930             break;
2931         case SDL_QUIT:
2932         case FF_QUIT_EVENT:
2933             do_exit();
2934             break;
2935         case FF_ALLOC_EVENT:
2936             video_open(event.user.data1);
2937             alloc_picture(event.user.data1);
2938             break;
2939         case FF_REFRESH_EVENT:
2940             video_refresh_timer(event.user.data1);
2941             cur_stream->refresh=0;
2942             break;
2943         default:
2944             break;
2945         }
2946     }
2947 }
2948
2949 static void opt_frame_size(const char *arg)
2950 {
2951     if (av_parse_video_size(&frame_width, &frame_height, arg) < 0) {
2952         fprintf(stderr, "Incorrect frame size\n");
2953         exit(1);
2954     }
2955     if ((frame_width % 2) != 0 || (frame_height % 2) != 0) {
2956         fprintf(stderr, "Frame size must be a multiple of 2\n");
2957         exit(1);
2958     }
2959 }
2960
2961 static int opt_width(const char *opt, const char *arg)
2962 {
2963     screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2964     return 0;
2965 }
2966
2967 static int opt_height(const char *opt, const char *arg)
2968 {
2969     screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2970     return 0;
2971 }
2972
2973 static void opt_format(const char *arg)
2974 {
2975     file_iformat = av_find_input_format(arg);
2976     if (!file_iformat) {
2977         fprintf(stderr, "Unknown input format: %s\n", arg);
2978         exit(1);
2979     }
2980 }
2981
2982 static void opt_frame_pix_fmt(const char *arg)
2983 {
2984     frame_pix_fmt = av_get_pix_fmt(arg);
2985 }
2986
2987 static int opt_sync(const char *opt, const char *arg)
2988 {
2989     if (!strcmp(arg, "audio"))
2990         av_sync_type = AV_SYNC_AUDIO_MASTER;
2991     else if (!strcmp(arg, "video"))
2992         av_sync_type = AV_SYNC_VIDEO_MASTER;
2993     else if (!strcmp(arg, "ext"))
2994         av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
2995     else {
2996         fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
2997         exit(1);
2998     }
2999     return 0;
3000 }
3001
3002 static int opt_seek(const char *opt, const char *arg)
3003 {
3004     start_time = parse_time_or_die(opt, arg, 1);
3005     return 0;
3006 }
3007
3008 static int opt_duration(const char *opt, const char *arg)
3009 {
3010     duration = parse_time_or_die(opt, arg, 1);
3011     return 0;
3012 }
3013
3014 static int opt_debug(const char *opt, const char *arg)
3015 {
3016     av_log_set_level(99);
3017     debug = parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
3018     return 0;
3019 }
3020
3021 static int opt_vismv(const char *opt, const char *arg)
3022 {
3023     debug_mv = parse_number_or_die(opt, arg, OPT_INT64, INT_MIN, INT_MAX);
3024     return 0;
3025 }
3026
3027 static int opt_thread_count(const char *opt, const char *arg)
3028 {
3029     thread_count= parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
3030 #if !HAVE_THREADS
3031     fprintf(stderr, "Warning: not compiled with thread support, using thread emulation\n");
3032 #endif
3033     return 0;
3034 }
3035
3036 static const OptionDef options[] = {
3037 #include "cmdutils_common_opts.h"
3038     { "x", HAS_ARG | OPT_FUNC2, {(void*)opt_width}, "force displayed width", "width" },
3039     { "y", HAS_ARG | OPT_FUNC2, {(void*)opt_height}, "force displayed height", "height" },
3040     { "s", HAS_ARG | OPT_VIDEO, {(void*)opt_frame_size}, "set frame size (WxH or abbreviation)", "size" },
3041     { "fs", OPT_BOOL, {(void*)&is_full_screen}, "force full screen" },
3042     { "an", OPT_BOOL, {(void*)&audio_disable}, "disable audio" },
3043     { "vn", OPT_BOOL, {(void*)&video_disable}, "disable video" },
3044     { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_AUDIO]}, "select desired audio stream", "stream_number" },
3045     { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_VIDEO]}, "select desired video stream", "stream_number" },
3046     { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE]}, "select desired subtitle stream", "stream_number" },
3047     { "ss", HAS_ARG | OPT_FUNC2, {(void*)&opt_seek}, "seek to a given position in seconds", "pos" },
3048     { "t", HAS_ARG | OPT_FUNC2, {(void*)&opt_duration}, "play  \"duration\" seconds of audio/video", "duration" },
3049     { "bytes", OPT_INT | HAS_ARG, {(void*)&seek_by_bytes}, "seek by bytes 0=off 1=on -1=auto", "val" },
3050     { "nodisp", OPT_BOOL, {(void*)&display_disable}, "disable graphical display" },
3051     { "f", HAS_ARG, {(void*)opt_format}, "force format", "fmt" },
3052     { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, {(void*)opt_frame_pix_fmt}, "set pixel format", "format" },
3053     { "stats", OPT_BOOL | OPT_EXPERT, {(void*)&show_status}, "show status", "" },
3054     { "debug", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_debug}, "print specific debug info", "" },
3055     { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&workaround_bugs}, "workaround bugs", "" },
3056     { "vismv", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_vismv}, "visualize motion vectors", "" },
3057     { "fast", OPT_BOOL | OPT_EXPERT, {(void*)&fast}, "non spec compliant optimizations", "" },
3058     { "genpts", OPT_BOOL | OPT_EXPERT, {(void*)&genpts}, "generate pts", "" },
3059     { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&decoder_reorder_pts}, "let decoder reorder pts 0=off 1=on -1=auto", ""},
3060     { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&lowres}, "", "" },
3061     { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_loop_filter}, "", "" },
3062     { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_frame}, "", "" },
3063     { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_idct}, "", "" },
3064     { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&idct}, "set idct algo",  "algo" },
3065     { "er", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_recognition}, "set error detection threshold (0-4)",  "threshold" },
3066     { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_concealment}, "set error concealment options",  "bit_mask" },
3067     { "sync", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_sync}, "set audio-video sync. type (type=audio/video/ext)", "type" },
3068     { "threads", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_thread_count}, "thread count", "count" },
3069     { "autoexit", OPT_BOOL | OPT_EXPERT, {(void*)&autoexit}, "exit at the end", "" },
3070     { "exitonkeydown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_keydown}, "exit on key down", "" },
3071     { "exitonmousedown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_mousedown}, "exit on mouse down", "" },
3072     { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&loop}, "set number of times the playback shall be looped", "loop count" },
3073     { "framedrop", OPT_BOOL | OPT_EXPERT, {(void*)&framedrop}, "drop frames when cpu is too slow", "" },
3074     { "window_title", OPT_STRING | HAS_ARG, {(void*)&window_title}, "set window title", "window title" },
3075 #if CONFIG_AVFILTER
3076     { "vf", OPT_STRING | HAS_ARG, {(void*)&vfilters}, "video filters", "filter list" },
3077 #endif
3078     { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, {(void*)&rdftspeed}, "rdft speed", "msecs" },
3079     { "default", OPT_FUNC2 | HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, {(void*)opt_default}, "generic catch all option", "" },
3080     { NULL, },
3081 };
3082
3083 static void show_usage(void)
3084 {
3085     printf("Simple media player\n");
3086     printf("usage: ffplay [options] input_file\n");
3087     printf("\n");
3088 }
3089
3090 static void show_help(void)
3091 {
3092     av_log_set_callback(log_callback_help);
3093     show_usage();
3094     show_help_options(options, "Main options:\n",
3095                       OPT_EXPERT, 0);
3096     show_help_options(options, "\nAdvanced options:\n",
3097                       OPT_EXPERT, OPT_EXPERT);
3098     printf("\n");
3099     av_opt_show2(avcodec_opts[0], NULL,
3100                  AV_OPT_FLAG_DECODING_PARAM, 0);
3101     printf("\n");
3102     av_opt_show2(avformat_opts, NULL,
3103                  AV_OPT_FLAG_DECODING_PARAM, 0);
3104 #if !CONFIG_AVFILTER
3105     printf("\n");
3106     av_opt_show2(sws_opts, NULL,
3107                  AV_OPT_FLAG_ENCODING_PARAM, 0);
3108 #endif
3109     printf("\nWhile playing:\n"
3110            "q, ESC              quit\n"
3111            "f                   toggle full screen\n"
3112            "p, SPC              pause\n"
3113            "a                   cycle audio channel\n"
3114            "v                   cycle video channel\n"
3115            "t                   cycle subtitle channel\n"
3116            "w                   show audio waves\n"
3117            "s                   activate frame-step mode\n"
3118            "left/right          seek backward/forward 10 seconds\n"
3119            "down/up             seek backward/forward 1 minute\n"
3120            "mouse click         seek to percentage in file corresponding to fraction of width\n"
3121            );
3122 }
3123
3124 static void opt_input_file(const char *filename)
3125 {
3126     if (input_filename) {
3127         fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
3128                 filename, input_filename);
3129         exit(1);
3130     }
3131     if (!strcmp(filename, "-"))
3132         filename = "pipe:";
3133     input_filename = filename;
3134 }
3135
3136 /* Called from the main */
3137 int main(int argc, char **argv)
3138 {
3139     int flags;
3140
3141     av_log_set_flags(AV_LOG_SKIP_REPEATED);
3142
3143     /* register all codecs, demux and protocols */
3144     avcodec_register_all();
3145 #if CONFIG_AVDEVICE
3146     avdevice_register_all();
3147 #endif
3148 #if CONFIG_AVFILTER
3149     avfilter_register_all();
3150 #endif
3151     av_register_all();
3152
3153     init_opts();
3154
3155     show_banner();
3156
3157     parse_options(argc, argv, options, opt_input_file);
3158
3159     if (!input_filename) {
3160         show_usage();
3161         fprintf(stderr, "An input file must be specified\n");
3162         fprintf(stderr, "Use -h to get full help or, even better, run 'man ffplay'\n");
3163         exit(1);
3164     }
3165
3166     if (display_disable) {
3167         video_disable = 1;
3168     }
3169     flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3170 #if !defined(__MINGW32__) && !defined(__APPLE__)
3171     flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3172 #endif
3173     if (SDL_Init (flags)) {
3174         fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3175         exit(1);
3176     }
3177
3178     if (!display_disable) {
3179 #if HAVE_SDL_VIDEO_SIZE
3180         const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3181         fs_screen_width = vi->current_w;
3182         fs_screen_height = vi->current_h;
3183 #endif
3184     }
3185
3186     SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3187     SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3188     SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3189
3190     av_init_packet(&flush_pkt);
3191     flush_pkt.data= "FLUSH";
3192
3193     cur_stream = stream_open(input_filename, file_iformat);
3194
3195     event_loop();
3196
3197     /* never returns */
3198
3199     return 0;
3200 }