]> git.sesse.net Git - ffmpeg/blob - ffplay.c
libvorbis: proper error logging and return.
[ffmpeg] / ffplay.c
1 /*
2  * Copyright (c) 2003 Fabrice Bellard
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20
21 /**
22  * @file
23  * simple media player based on the FFmpeg libraries
24  */
25
26 #include "config.h"
27 #include <inttypes.h>
28 #include <math.h>
29 #include <limits.h>
30 #include "libavutil/avstring.h"
31 #include "libavutil/colorspace.h"
32 #include "libavutil/mathematics.h"
33 #include "libavutil/pixdesc.h"
34 #include "libavutil/imgutils.h"
35 #include "libavutil/dict.h"
36 #include "libavutil/parseutils.h"
37 #include "libavutil/samplefmt.h"
38 #include "libavutil/avassert.h"
39 #include "libavformat/avformat.h"
40 #include "libavdevice/avdevice.h"
41 #include "libswscale/swscale.h"
42 #include "libavcodec/audioconvert.h"
43 #include "libavutil/opt.h"
44 #include "libavcodec/avfft.h"
45 #include "libswresample/swresample.h"
46
47 #if CONFIG_AVFILTER
48 # include "libavfilter/avcodec.h"
49 # include "libavfilter/avfilter.h"
50 # include "libavfilter/avfiltergraph.h"
51 # include "libavfilter/buffersink.h"
52 #endif
53
54 #include <SDL.h>
55 #include <SDL_thread.h>
56
57 #include "cmdutils.h"
58
59 #include <unistd.h>
60 #include <assert.h>
61
62 const char program_name[] = "ffplay";
63 const int program_birth_year = 2003;
64
65 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
66 #define MIN_AUDIOQ_SIZE (20 * 16 * 1024)
67 #define MIN_FRAMES 5
68
69 /* SDL audio buffer size, in samples. Should be small to have precise
70    A/V sync as SDL does not have hardware buffer fullness info. */
71 #define SDL_AUDIO_BUFFER_SIZE 1024
72
73 /* no AV sync correction is done if below the AV sync threshold */
74 #define AV_SYNC_THRESHOLD 0.01
75 /* no AV correction is done if too big error */
76 #define AV_NOSYNC_THRESHOLD 10.0
77
78 /* maximum audio speed change to get correct sync */
79 #define SAMPLE_CORRECTION_PERCENT_MAX 10
80
81 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
82 #define AUDIO_DIFF_AVG_NB   20
83
84 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
85 #define SAMPLE_ARRAY_SIZE (2*65536)
86
87 static int sws_flags = SWS_BICUBIC;
88
89 typedef struct PacketQueue {
90     AVPacketList *first_pkt, *last_pkt;
91     int nb_packets;
92     int size;
93     int abort_request;
94     SDL_mutex *mutex;
95     SDL_cond *cond;
96 } PacketQueue;
97
98 #define VIDEO_PICTURE_QUEUE_SIZE 2
99 #define SUBPICTURE_QUEUE_SIZE 4
100
101 typedef struct VideoPicture {
102     double pts;                                  ///<presentation time stamp for this picture
103     double duration;                             ///<expected duration of the frame
104     int64_t pos;                                 ///<byte position in file
105     int skip;
106     SDL_Overlay *bmp;
107     int width, height; /* source height & width */
108     int allocated;
109     int reallocate;
110     enum PixelFormat pix_fmt;
111
112 #if CONFIG_AVFILTER
113     AVFilterBufferRef *picref;
114 #endif
115 } VideoPicture;
116
117 typedef struct SubPicture {
118     double pts; /* presentation time stamp for this picture */
119     AVSubtitle sub;
120 } SubPicture;
121
122 enum {
123     AV_SYNC_AUDIO_MASTER, /* default choice */
124     AV_SYNC_VIDEO_MASTER,
125     AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
126 };
127
128 typedef struct VideoState {
129     SDL_Thread *read_tid;
130     SDL_Thread *video_tid;
131     SDL_Thread *refresh_tid;
132     AVInputFormat *iformat;
133     int no_background;
134     int abort_request;
135     int paused;
136     int last_paused;
137     int seek_req;
138     int seek_flags;
139     int64_t seek_pos;
140     int64_t seek_rel;
141     int read_pause_return;
142     AVFormatContext *ic;
143
144     int audio_stream;
145
146     int av_sync_type;
147     double external_clock; /* external clock base */
148     int64_t external_clock_time;
149
150     double audio_clock;
151     double audio_diff_cum; /* used for AV difference average computation */
152     double audio_diff_avg_coef;
153     double audio_diff_threshold;
154     int audio_diff_avg_count;
155     AVStream *audio_st;
156     PacketQueue audioq;
157     int audio_hw_buf_size;
158     DECLARE_ALIGNED(16,uint8_t,audio_buf2)[AVCODEC_MAX_AUDIO_FRAME_SIZE * 4];
159     uint8_t silence_buf[SDL_AUDIO_BUFFER_SIZE];
160     uint8_t *audio_buf;
161     uint8_t *audio_buf1;
162     unsigned int audio_buf_size; /* in bytes */
163     int audio_buf_index; /* in bytes */
164     int audio_write_buf_size;
165     AVPacket audio_pkt_temp;
166     AVPacket audio_pkt;
167     enum AVSampleFormat audio_src_fmt;
168     enum AVSampleFormat audio_tgt_fmt;
169     int audio_src_channels;
170     int audio_tgt_channels;
171     int64_t audio_src_channel_layout;
172     int64_t audio_tgt_channel_layout;
173     int audio_src_freq;
174     int audio_tgt_freq;
175     struct SwrContext *swr_ctx;
176     double audio_current_pts;
177     double audio_current_pts_drift;
178     int frame_drops_early;
179     int frame_drops_late;
180     AVFrame *frame;
181
182     enum ShowMode {
183         SHOW_MODE_NONE = -1, SHOW_MODE_VIDEO = 0, SHOW_MODE_WAVES, SHOW_MODE_RDFT, SHOW_MODE_NB
184     } show_mode;
185     int16_t sample_array[SAMPLE_ARRAY_SIZE];
186     int sample_array_index;
187     int last_i_start;
188     RDFTContext *rdft;
189     int rdft_bits;
190     FFTSample *rdft_data;
191     int xpos;
192
193     SDL_Thread *subtitle_tid;
194     int subtitle_stream;
195     int subtitle_stream_changed;
196     AVStream *subtitle_st;
197     PacketQueue subtitleq;
198     SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
199     int subpq_size, subpq_rindex, subpq_windex;
200     SDL_mutex *subpq_mutex;
201     SDL_cond *subpq_cond;
202
203     double frame_timer;
204     double frame_last_pts;
205     double frame_last_duration;
206     double frame_last_dropped_pts;
207     double frame_last_returned_time;
208     double frame_last_filter_delay;
209     int64_t frame_last_dropped_pos;
210     double video_clock;                          ///<pts of last decoded frame / predicted pts of next decoded frame
211     int video_stream;
212     AVStream *video_st;
213     PacketQueue videoq;
214     double video_current_pts;                    ///<current displayed pts (different from video_clock if frame fifos are used)
215     double video_current_pts_drift;              ///<video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
216     int64_t video_current_pos;                   ///<current displayed file pos
217     VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
218     int pictq_size, pictq_rindex, pictq_windex;
219     SDL_mutex *pictq_mutex;
220     SDL_cond *pictq_cond;
221 #if !CONFIG_AVFILTER
222     struct SwsContext *img_convert_ctx;
223 #endif
224
225     char filename[1024];
226     int width, height, xleft, ytop;
227     int step;
228
229 #if CONFIG_AVFILTER
230     AVFilterContext *out_video_filter;          ///<the last filter in the video chain
231 #endif
232
233     int refresh;
234 } VideoState;
235
236 static int opt_help(const char *opt, const char *arg);
237
238 /* options specified by the user */
239 static AVInputFormat *file_iformat;
240 static const char *input_filename;
241 static const char *window_title;
242 static int fs_screen_width;
243 static int fs_screen_height;
244 static int screen_width = 0;
245 static int screen_height = 0;
246 static int audio_disable;
247 static int video_disable;
248 static int wanted_stream[AVMEDIA_TYPE_NB]={
249     [AVMEDIA_TYPE_AUDIO]=-1,
250     [AVMEDIA_TYPE_VIDEO]=-1,
251     [AVMEDIA_TYPE_SUBTITLE]=-1,
252 };
253 static int seek_by_bytes=-1;
254 static int display_disable;
255 static int show_status = 1;
256 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
257 static int64_t start_time = AV_NOPTS_VALUE;
258 static int64_t duration = AV_NOPTS_VALUE;
259 static int workaround_bugs = 1;
260 static int fast = 0;
261 static int genpts = 0;
262 static int lowres = 0;
263 static int idct = FF_IDCT_AUTO;
264 static enum AVDiscard skip_frame= AVDISCARD_DEFAULT;
265 static enum AVDiscard skip_idct= AVDISCARD_DEFAULT;
266 static enum AVDiscard skip_loop_filter= AVDISCARD_DEFAULT;
267 static int error_recognition = FF_ER_CAREFUL;
268 static int error_concealment = 3;
269 static int decoder_reorder_pts= -1;
270 static int autoexit;
271 static int exit_on_keydown;
272 static int exit_on_mousedown;
273 static int loop=1;
274 static int framedrop=-1;
275 static enum ShowMode show_mode = SHOW_MODE_NONE;
276 static const char *audio_codec_name;
277 static const char *subtitle_codec_name;
278 static const char *video_codec_name;
279
280 static int rdftspeed=20;
281 #if CONFIG_AVFILTER
282 static char *vfilters = NULL;
283 #endif
284
285 /* current context */
286 static int is_full_screen;
287 static int64_t audio_callback_time;
288
289 static AVPacket flush_pkt;
290
291 #define FF_ALLOC_EVENT   (SDL_USEREVENT)
292 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
293 #define FF_QUIT_EVENT    (SDL_USEREVENT + 2)
294
295 static SDL_Surface *screen;
296
297 void av_noreturn exit_program(int ret)
298 {
299     exit(ret);
300 }
301
302 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
303 {
304     AVPacketList *pkt1;
305
306     /* duplicate the packet */
307     if (pkt!=&flush_pkt && av_dup_packet(pkt) < 0)
308         return -1;
309
310     pkt1 = av_malloc(sizeof(AVPacketList));
311     if (!pkt1)
312         return -1;
313     pkt1->pkt = *pkt;
314     pkt1->next = NULL;
315
316
317     SDL_LockMutex(q->mutex);
318
319     if (!q->last_pkt)
320
321         q->first_pkt = pkt1;
322     else
323         q->last_pkt->next = pkt1;
324     q->last_pkt = pkt1;
325     q->nb_packets++;
326     q->size += pkt1->pkt.size + sizeof(*pkt1);
327     /* XXX: should duplicate packet data in DV case */
328     SDL_CondSignal(q->cond);
329
330     SDL_UnlockMutex(q->mutex);
331     return 0;
332 }
333
334 /* packet queue handling */
335 static void packet_queue_init(PacketQueue *q)
336 {
337     memset(q, 0, sizeof(PacketQueue));
338     q->mutex = SDL_CreateMutex();
339     q->cond = SDL_CreateCond();
340     packet_queue_put(q, &flush_pkt);
341 }
342
343 static void packet_queue_flush(PacketQueue *q)
344 {
345     AVPacketList *pkt, *pkt1;
346
347     SDL_LockMutex(q->mutex);
348     for(pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
349         pkt1 = pkt->next;
350         av_free_packet(&pkt->pkt);
351         av_freep(&pkt);
352     }
353     q->last_pkt = NULL;
354     q->first_pkt = NULL;
355     q->nb_packets = 0;
356     q->size = 0;
357     SDL_UnlockMutex(q->mutex);
358 }
359
360 static void packet_queue_end(PacketQueue *q)
361 {
362     packet_queue_flush(q);
363     SDL_DestroyMutex(q->mutex);
364     SDL_DestroyCond(q->cond);
365 }
366
367 static void packet_queue_abort(PacketQueue *q)
368 {
369     SDL_LockMutex(q->mutex);
370
371     q->abort_request = 1;
372
373     SDL_CondSignal(q->cond);
374
375     SDL_UnlockMutex(q->mutex);
376 }
377
378 /* return < 0 if aborted, 0 if no packet and > 0 if packet.  */
379 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
380 {
381     AVPacketList *pkt1;
382     int ret;
383
384     SDL_LockMutex(q->mutex);
385
386     for(;;) {
387         if (q->abort_request) {
388             ret = -1;
389             break;
390         }
391
392         pkt1 = q->first_pkt;
393         if (pkt1) {
394             q->first_pkt = pkt1->next;
395             if (!q->first_pkt)
396                 q->last_pkt = NULL;
397             q->nb_packets--;
398             q->size -= pkt1->pkt.size + sizeof(*pkt1);
399             *pkt = pkt1->pkt;
400             av_free(pkt1);
401             ret = 1;
402             break;
403         } else if (!block) {
404             ret = 0;
405             break;
406         } else {
407             SDL_CondWait(q->cond, q->mutex);
408         }
409     }
410     SDL_UnlockMutex(q->mutex);
411     return ret;
412 }
413
414 static inline void fill_rectangle(SDL_Surface *screen,
415                                   int x, int y, int w, int h, int color)
416 {
417     SDL_Rect rect;
418     rect.x = x;
419     rect.y = y;
420     rect.w = w;
421     rect.h = h;
422     SDL_FillRect(screen, &rect, color);
423 }
424
425 #define ALPHA_BLEND(a, oldp, newp, s)\
426 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
427
428 #define RGBA_IN(r, g, b, a, s)\
429 {\
430     unsigned int v = ((const uint32_t *)(s))[0];\
431     a = (v >> 24) & 0xff;\
432     r = (v >> 16) & 0xff;\
433     g = (v >> 8) & 0xff;\
434     b = v & 0xff;\
435 }
436
437 #define YUVA_IN(y, u, v, a, s, pal)\
438 {\
439     unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
440     a = (val >> 24) & 0xff;\
441     y = (val >> 16) & 0xff;\
442     u = (val >> 8) & 0xff;\
443     v = val & 0xff;\
444 }
445
446 #define YUVA_OUT(d, y, u, v, a)\
447 {\
448     ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
449 }
450
451
452 #define BPP 1
453
454 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
455 {
456     int wrap, wrap3, width2, skip2;
457     int y, u, v, a, u1, v1, a1, w, h;
458     uint8_t *lum, *cb, *cr;
459     const uint8_t *p;
460     const uint32_t *pal;
461     int dstx, dsty, dstw, dsth;
462
463     dstw = av_clip(rect->w, 0, imgw);
464     dsth = av_clip(rect->h, 0, imgh);
465     dstx = av_clip(rect->x, 0, imgw - dstw);
466     dsty = av_clip(rect->y, 0, imgh - dsth);
467     lum = dst->data[0] + dsty * dst->linesize[0];
468     cb = dst->data[1] + (dsty >> 1) * dst->linesize[1];
469     cr = dst->data[2] + (dsty >> 1) * dst->linesize[2];
470
471     width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
472     skip2 = dstx >> 1;
473     wrap = dst->linesize[0];
474     wrap3 = rect->pict.linesize[0];
475     p = rect->pict.data[0];
476     pal = (const uint32_t *)rect->pict.data[1];  /* Now in YCrCb! */
477
478     if (dsty & 1) {
479         lum += dstx;
480         cb += skip2;
481         cr += skip2;
482
483         if (dstx & 1) {
484             YUVA_IN(y, u, v, a, p, pal);
485             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
486             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
487             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
488             cb++;
489             cr++;
490             lum++;
491             p += BPP;
492         }
493         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
494             YUVA_IN(y, u, v, a, p, pal);
495             u1 = u;
496             v1 = v;
497             a1 = a;
498             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
499
500             YUVA_IN(y, u, v, a, p + BPP, pal);
501             u1 += u;
502             v1 += v;
503             a1 += a;
504             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
505             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
506             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
507             cb++;
508             cr++;
509             p += 2 * BPP;
510             lum += 2;
511         }
512         if (w) {
513             YUVA_IN(y, u, v, a, p, pal);
514             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
515             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
516             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
517             p++;
518             lum++;
519         }
520         p += wrap3 - dstw * BPP;
521         lum += wrap - dstw - dstx;
522         cb += dst->linesize[1] - width2 - skip2;
523         cr += dst->linesize[2] - width2 - skip2;
524     }
525     for(h = dsth - (dsty & 1); h >= 2; h -= 2) {
526         lum += dstx;
527         cb += skip2;
528         cr += skip2;
529
530         if (dstx & 1) {
531             YUVA_IN(y, u, v, a, p, pal);
532             u1 = u;
533             v1 = v;
534             a1 = a;
535             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
536             p += wrap3;
537             lum += wrap;
538             YUVA_IN(y, u, v, a, p, pal);
539             u1 += u;
540             v1 += v;
541             a1 += a;
542             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
543             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
544             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
545             cb++;
546             cr++;
547             p += -wrap3 + BPP;
548             lum += -wrap + 1;
549         }
550         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
551             YUVA_IN(y, u, v, a, p, pal);
552             u1 = u;
553             v1 = v;
554             a1 = a;
555             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
556
557             YUVA_IN(y, u, v, a, p + BPP, pal);
558             u1 += u;
559             v1 += v;
560             a1 += a;
561             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
562             p += wrap3;
563             lum += wrap;
564
565             YUVA_IN(y, u, v, a, p, pal);
566             u1 += u;
567             v1 += v;
568             a1 += a;
569             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
570
571             YUVA_IN(y, u, v, a, p + BPP, pal);
572             u1 += u;
573             v1 += v;
574             a1 += a;
575             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
576
577             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
578             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
579
580             cb++;
581             cr++;
582             p += -wrap3 + 2 * BPP;
583             lum += -wrap + 2;
584         }
585         if (w) {
586             YUVA_IN(y, u, v, a, p, pal);
587             u1 = u;
588             v1 = v;
589             a1 = a;
590             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
591             p += wrap3;
592             lum += wrap;
593             YUVA_IN(y, u, v, a, p, pal);
594             u1 += u;
595             v1 += v;
596             a1 += a;
597             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
598             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
599             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
600             cb++;
601             cr++;
602             p += -wrap3 + BPP;
603             lum += -wrap + 1;
604         }
605         p += wrap3 + (wrap3 - dstw * BPP);
606         lum += wrap + (wrap - dstw - dstx);
607         cb += dst->linesize[1] - width2 - skip2;
608         cr += dst->linesize[2] - width2 - skip2;
609     }
610     /* handle odd height */
611     if (h) {
612         lum += dstx;
613         cb += skip2;
614         cr += skip2;
615
616         if (dstx & 1) {
617             YUVA_IN(y, u, v, a, p, pal);
618             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
619             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
620             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
621             cb++;
622             cr++;
623             lum++;
624             p += BPP;
625         }
626         for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
627             YUVA_IN(y, u, v, a, p, pal);
628             u1 = u;
629             v1 = v;
630             a1 = a;
631             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
632
633             YUVA_IN(y, u, v, a, p + BPP, pal);
634             u1 += u;
635             v1 += v;
636             a1 += a;
637             lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
638             cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
639             cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
640             cb++;
641             cr++;
642             p += 2 * BPP;
643             lum += 2;
644         }
645         if (w) {
646             YUVA_IN(y, u, v, a, p, pal);
647             lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
648             cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
649             cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
650         }
651     }
652 }
653
654 static void free_subpicture(SubPicture *sp)
655 {
656     avsubtitle_free(&sp->sub);
657 }
658
659 static void video_image_display(VideoState *is)
660 {
661     VideoPicture *vp;
662     SubPicture *sp;
663     AVPicture pict;
664     float aspect_ratio;
665     int width, height, x, y;
666     SDL_Rect rect;
667     int i;
668
669     vp = &is->pictq[is->pictq_rindex];
670     if (vp->bmp) {
671 #if CONFIG_AVFILTER
672          if (vp->picref->video->sample_aspect_ratio.num == 0)
673              aspect_ratio = 0;
674          else
675              aspect_ratio = av_q2d(vp->picref->video->sample_aspect_ratio);
676 #else
677
678         /* XXX: use variable in the frame */
679         if (is->video_st->sample_aspect_ratio.num)
680             aspect_ratio = av_q2d(is->video_st->sample_aspect_ratio);
681         else if (is->video_st->codec->sample_aspect_ratio.num)
682             aspect_ratio = av_q2d(is->video_st->codec->sample_aspect_ratio);
683         else
684             aspect_ratio = 0;
685 #endif
686         if (aspect_ratio <= 0.0)
687             aspect_ratio = 1.0;
688         aspect_ratio *= (float)vp->width / (float)vp->height;
689
690         if (is->subtitle_st) {
691             if (is->subpq_size > 0) {
692                 sp = &is->subpq[is->subpq_rindex];
693
694                 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000)) {
695                     SDL_LockYUVOverlay (vp->bmp);
696
697                     pict.data[0] = vp->bmp->pixels[0];
698                     pict.data[1] = vp->bmp->pixels[2];
699                     pict.data[2] = vp->bmp->pixels[1];
700
701                     pict.linesize[0] = vp->bmp->pitches[0];
702                     pict.linesize[1] = vp->bmp->pitches[2];
703                     pict.linesize[2] = vp->bmp->pitches[1];
704
705                     for (i = 0; i < sp->sub.num_rects; i++)
706                         blend_subrect(&pict, sp->sub.rects[i],
707                                       vp->bmp->w, vp->bmp->h);
708
709                     SDL_UnlockYUVOverlay (vp->bmp);
710                 }
711             }
712         }
713
714
715         /* XXX: we suppose the screen has a 1.0 pixel ratio */
716         height = is->height;
717         width = ((int)rint(height * aspect_ratio)) & ~1;
718         if (width > is->width) {
719             width = is->width;
720             height = ((int)rint(width / aspect_ratio)) & ~1;
721         }
722         x = (is->width - width) / 2;
723         y = (is->height - height) / 2;
724         is->no_background = 0;
725         rect.x = is->xleft + x;
726         rect.y = is->ytop  + y;
727         rect.w = FFMAX(width,  1);
728         rect.h = FFMAX(height, 1);
729         SDL_DisplayYUVOverlay(vp->bmp, &rect);
730     }
731 }
732
733 static inline int compute_mod(int a, int b)
734 {
735     return a < 0 ? a%b + b : a%b;
736 }
737
738 static void video_audio_display(VideoState *s)
739 {
740     int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
741     int ch, channels, h, h2, bgcolor, fgcolor;
742     int16_t time_diff;
743     int rdft_bits, nb_freq;
744
745     for(rdft_bits=1; (1<<rdft_bits)<2*s->height; rdft_bits++)
746         ;
747     nb_freq= 1<<(rdft_bits-1);
748
749     /* compute display index : center on currently output samples */
750     channels = s->audio_tgt_channels;
751     nb_display_channels = channels;
752     if (!s->paused) {
753         int data_used= s->show_mode == SHOW_MODE_WAVES ? s->width : (2*nb_freq);
754         n = 2 * channels;
755         delay = s->audio_write_buf_size;
756         delay /= n;
757
758         /* to be more precise, we take into account the time spent since
759            the last buffer computation */
760         if (audio_callback_time) {
761             time_diff = av_gettime() - audio_callback_time;
762             delay -= (time_diff * s->audio_tgt_freq) / 1000000;
763         }
764
765         delay += 2*data_used;
766         if (delay < data_used)
767             delay = data_used;
768
769         i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
770         if (s->show_mode == SHOW_MODE_WAVES) {
771             h= INT_MIN;
772             for(i=0; i<1000; i+=channels){
773                 int idx= (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
774                 int a= s->sample_array[idx];
775                 int b= s->sample_array[(idx + 4*channels)%SAMPLE_ARRAY_SIZE];
776                 int c= s->sample_array[(idx + 5*channels)%SAMPLE_ARRAY_SIZE];
777                 int d= s->sample_array[(idx + 9*channels)%SAMPLE_ARRAY_SIZE];
778                 int score= a-d;
779                 if(h<score && (b^c)<0){
780                     h= score;
781                     i_start= idx;
782                 }
783             }
784         }
785
786         s->last_i_start = i_start;
787     } else {
788         i_start = s->last_i_start;
789     }
790
791     bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
792     if (s->show_mode == SHOW_MODE_WAVES) {
793         fill_rectangle(screen,
794                        s->xleft, s->ytop, s->width, s->height,
795                        bgcolor);
796
797         fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
798
799         /* total height for one channel */
800         h = s->height / nb_display_channels;
801         /* graph height / 2 */
802         h2 = (h * 9) / 20;
803         for(ch = 0;ch < nb_display_channels; ch++) {
804             i = i_start + ch;
805             y1 = s->ytop + ch * h + (h / 2); /* position of center line */
806             for(x = 0; x < s->width; x++) {
807                 y = (s->sample_array[i] * h2) >> 15;
808                 if (y < 0) {
809                     y = -y;
810                     ys = y1 - y;
811                 } else {
812                     ys = y1;
813                 }
814                 fill_rectangle(screen,
815                                s->xleft + x, ys, 1, y,
816                                fgcolor);
817                 i += channels;
818                 if (i >= SAMPLE_ARRAY_SIZE)
819                     i -= SAMPLE_ARRAY_SIZE;
820             }
821         }
822
823         fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
824
825         for(ch = 1;ch < nb_display_channels; ch++) {
826             y = s->ytop + ch * h;
827             fill_rectangle(screen,
828                            s->xleft, y, s->width, 1,
829                            fgcolor);
830         }
831         SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
832     }else{
833         nb_display_channels= FFMIN(nb_display_channels, 2);
834         if(rdft_bits != s->rdft_bits){
835             av_rdft_end(s->rdft);
836             av_free(s->rdft_data);
837             s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
838             s->rdft_bits= rdft_bits;
839             s->rdft_data= av_malloc(4*nb_freq*sizeof(*s->rdft_data));
840         }
841         {
842             FFTSample *data[2];
843             for(ch = 0;ch < nb_display_channels; ch++) {
844                 data[ch] = s->rdft_data + 2*nb_freq*ch;
845                 i = i_start + ch;
846                 for(x = 0; x < 2*nb_freq; x++) {
847                     double w= (x-nb_freq)*(1.0/nb_freq);
848                     data[ch][x]= s->sample_array[i]*(1.0-w*w);
849                     i += channels;
850                     if (i >= SAMPLE_ARRAY_SIZE)
851                         i -= SAMPLE_ARRAY_SIZE;
852                 }
853                 av_rdft_calc(s->rdft, data[ch]);
854             }
855             //least efficient way to do this, we should of course directly access it but its more than fast enough
856             for(y=0; y<s->height; y++){
857                 double w= 1/sqrt(nb_freq);
858                 int a= sqrt(w*sqrt(data[0][2*y+0]*data[0][2*y+0] + data[0][2*y+1]*data[0][2*y+1]));
859                 int b= (nb_display_channels == 2 ) ? sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0]
860                        + data[1][2*y+1]*data[1][2*y+1])) : a;
861                 a= FFMIN(a,255);
862                 b= FFMIN(b,255);
863                 fgcolor = SDL_MapRGB(screen->format, a, b, (a+b)/2);
864
865                 fill_rectangle(screen,
866                             s->xpos, s->height-y, 1, 1,
867                             fgcolor);
868             }
869         }
870         SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
871         s->xpos++;
872         if(s->xpos >= s->width)
873             s->xpos= s->xleft;
874     }
875 }
876
877 static void stream_close(VideoState *is)
878 {
879     VideoPicture *vp;
880     int i;
881     /* XXX: use a special url_shutdown call to abort parse cleanly */
882     is->abort_request = 1;
883     SDL_WaitThread(is->read_tid, NULL);
884     SDL_WaitThread(is->refresh_tid, NULL);
885
886     /* free all pictures */
887     for(i=0;i<VIDEO_PICTURE_QUEUE_SIZE; i++) {
888         vp = &is->pictq[i];
889 #if CONFIG_AVFILTER
890         if (vp->picref) {
891             avfilter_unref_buffer(vp->picref);
892             vp->picref = NULL;
893         }
894 #endif
895         if (vp->bmp) {
896             SDL_FreeYUVOverlay(vp->bmp);
897             vp->bmp = NULL;
898         }
899     }
900     SDL_DestroyMutex(is->pictq_mutex);
901     SDL_DestroyCond(is->pictq_cond);
902     SDL_DestroyMutex(is->subpq_mutex);
903     SDL_DestroyCond(is->subpq_cond);
904 #if !CONFIG_AVFILTER
905     if (is->img_convert_ctx)
906         sws_freeContext(is->img_convert_ctx);
907 #endif
908     av_free(is);
909 }
910
911 static void do_exit(VideoState *is)
912 {
913     if (is) {
914         stream_close(is);
915     }
916     av_lockmgr_register(NULL);
917     uninit_opts();
918 #if CONFIG_AVFILTER
919     avfilter_uninit();
920 #endif
921     avformat_network_deinit();
922     if (show_status)
923         printf("\n");
924     SDL_Quit();
925     av_log(NULL, AV_LOG_QUIET, "%s", "");
926     exit(0);
927 }
928
929 static int video_open(VideoState *is, int force_set_video_mode){
930     int flags = SDL_HWSURFACE|SDL_ASYNCBLIT|SDL_HWACCEL;
931     int w,h;
932
933     if(is_full_screen) flags |= SDL_FULLSCREEN;
934     else               flags |= SDL_RESIZABLE;
935
936     if (is_full_screen && fs_screen_width) {
937         w = fs_screen_width;
938         h = fs_screen_height;
939     } else if(!is_full_screen && screen_width){
940         w = screen_width;
941         h = screen_height;
942 #if CONFIG_AVFILTER
943     }else if (is->out_video_filter && is->out_video_filter->inputs[0]){
944         w = is->out_video_filter->inputs[0]->w;
945         h = is->out_video_filter->inputs[0]->h;
946 #else
947     }else if (is->video_st && is->video_st->codec->width){
948         w = is->video_st->codec->width;
949         h = is->video_st->codec->height;
950 #endif
951     } else {
952         w = 640;
953         h = 480;
954     }
955     if(screen && is->width == screen->w && screen->w == w
956        && is->height== screen->h && screen->h == h && !force_set_video_mode)
957         return 0;
958     screen = SDL_SetVideoMode(w, h, 0, flags);
959     if (!screen) {
960         fprintf(stderr, "SDL: could not set video mode - exiting\n");
961         do_exit(is);
962     }
963     if (!window_title)
964         window_title = input_filename;
965     SDL_WM_SetCaption(window_title, window_title);
966
967     is->width = screen->w;
968     is->height = screen->h;
969
970     return 0;
971 }
972
973 /* display the current picture, if any */
974 static void video_display(VideoState *is)
975 {
976     if(!screen)
977         video_open(is, 0);
978     if (is->audio_st && is->show_mode != SHOW_MODE_VIDEO)
979         video_audio_display(is);
980     else if (is->video_st)
981         video_image_display(is);
982 }
983
984 static int refresh_thread(void *opaque)
985 {
986     VideoState *is= opaque;
987     while(!is->abort_request){
988         SDL_Event event;
989         event.type = FF_REFRESH_EVENT;
990         event.user.data1 = opaque;
991         if(!is->refresh){
992             is->refresh=1;
993             SDL_PushEvent(&event);
994         }
995         //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
996         usleep(is->audio_st && is->show_mode != SHOW_MODE_VIDEO ? rdftspeed*1000 : 5000);
997     }
998     return 0;
999 }
1000
1001 /* get the current audio clock value */
1002 static double get_audio_clock(VideoState *is)
1003 {
1004     if (is->paused) {
1005         return is->audio_current_pts;
1006     } else {
1007         return is->audio_current_pts_drift + av_gettime() / 1000000.0;
1008     }
1009 }
1010
1011 /* get the current video clock value */
1012 static double get_video_clock(VideoState *is)
1013 {
1014     if (is->paused) {
1015         return is->video_current_pts;
1016     } else {
1017         return is->video_current_pts_drift + av_gettime() / 1000000.0;
1018     }
1019 }
1020
1021 /* get the current external clock value */
1022 static double get_external_clock(VideoState *is)
1023 {
1024     int64_t ti;
1025     ti = av_gettime();
1026     return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1027 }
1028
1029 /* get the current master clock value */
1030 static double get_master_clock(VideoState *is)
1031 {
1032     double val;
1033
1034     if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1035         if (is->video_st)
1036             val = get_video_clock(is);
1037         else
1038             val = get_audio_clock(is);
1039     } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1040         if (is->audio_st)
1041             val = get_audio_clock(is);
1042         else
1043             val = get_video_clock(is);
1044     } else {
1045         val = get_external_clock(is);
1046     }
1047     return val;
1048 }
1049
1050 /* seek in the stream */
1051 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1052 {
1053     if (!is->seek_req) {
1054         is->seek_pos = pos;
1055         is->seek_rel = rel;
1056         is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1057         if (seek_by_bytes)
1058             is->seek_flags |= AVSEEK_FLAG_BYTE;
1059         is->seek_req = 1;
1060     }
1061 }
1062
1063 /* pause or resume the video */
1064 static void stream_toggle_pause(VideoState *is)
1065 {
1066     if (is->paused) {
1067         is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1068         if(is->read_pause_return != AVERROR(ENOSYS)){
1069             is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1070         }
1071         is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1072     }
1073     is->paused = !is->paused;
1074 }
1075
1076 static double compute_target_delay(double delay, VideoState *is)
1077 {
1078     double sync_threshold, diff;
1079
1080     /* update delay to follow master synchronisation source */
1081     if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1082          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1083         /* if video is slave, we try to correct big delays by
1084            duplicating or deleting a frame */
1085         diff = get_video_clock(is) - get_master_clock(is);
1086
1087         /* skip or repeat frame. We take into account the
1088            delay to compute the threshold. I still don't know
1089            if it is the best guess */
1090         sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1091         if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1092             if (diff <= -sync_threshold)
1093                 delay = 0;
1094             else if (diff >= sync_threshold)
1095                 delay = 2 * delay;
1096         }
1097     }
1098
1099     av_dlog(NULL, "video: delay=%0.3f A-V=%f\n",
1100             delay, -diff);
1101
1102     return delay;
1103 }
1104
1105 static void pictq_next_picture(VideoState *is) {
1106     /* update queue size and signal for next picture */
1107     if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1108         is->pictq_rindex = 0;
1109
1110     SDL_LockMutex(is->pictq_mutex);
1111     is->pictq_size--;
1112     SDL_CondSignal(is->pictq_cond);
1113     SDL_UnlockMutex(is->pictq_mutex);
1114 }
1115
1116 static void update_video_pts(VideoState *is, double pts, int64_t pos) {
1117     double time = av_gettime() / 1000000.0;
1118     /* update current video pts */
1119     is->video_current_pts = pts;
1120     is->video_current_pts_drift = is->video_current_pts - time;
1121     is->video_current_pos = pos;
1122     is->frame_last_pts = pts;
1123 }
1124
1125 /* called to display each frame */
1126 static void video_refresh(void *opaque)
1127 {
1128     VideoState *is = opaque;
1129     VideoPicture *vp;
1130     double time;
1131
1132     SubPicture *sp, *sp2;
1133
1134     if (is->video_st) {
1135 retry:
1136         if (is->pictq_size == 0) {
1137             SDL_LockMutex(is->pictq_mutex);
1138             if (is->frame_last_dropped_pts != AV_NOPTS_VALUE && is->frame_last_dropped_pts > is->frame_last_pts) {
1139                 update_video_pts(is, is->frame_last_dropped_pts, is->frame_last_dropped_pos);
1140                 is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1141             }
1142             SDL_UnlockMutex(is->pictq_mutex);
1143             //nothing to do, no picture to display in the que
1144         } else {
1145             double last_duration, duration, delay;
1146             /* dequeue the picture */
1147             vp = &is->pictq[is->pictq_rindex];
1148
1149             if (vp->skip) {
1150                 pictq_next_picture(is);
1151                 goto retry;
1152             }
1153
1154             /* compute nominal last_duration */
1155             last_duration = vp->pts - is->frame_last_pts;
1156             if (last_duration > 0 && last_duration < 10.0) {
1157                 /* if duration of the last frame was sane, update last_duration in video state */
1158                 is->frame_last_duration = last_duration;
1159             }
1160             delay = compute_target_delay(is->frame_last_duration, is);
1161
1162             time= av_gettime()/1000000.0;
1163             if(time < is->frame_timer + delay)
1164                 return;
1165
1166             if (delay > 0)
1167                 is->frame_timer += delay * FFMAX(1, floor((time-is->frame_timer) / delay));
1168
1169             SDL_LockMutex(is->pictq_mutex);
1170             update_video_pts(is, vp->pts, vp->pos);
1171             SDL_UnlockMutex(is->pictq_mutex);
1172
1173             if(is->pictq_size > 1) {
1174                  VideoPicture *nextvp= &is->pictq[(is->pictq_rindex+1)%VIDEO_PICTURE_QUEUE_SIZE];
1175                  duration = nextvp->pts - vp->pts; // More accurate this way, 1/time_base is often not reflecting FPS
1176             } else {
1177                  duration = vp->duration;
1178             }
1179
1180             if((framedrop>0 || (framedrop && is->audio_st)) && time > is->frame_timer + duration){
1181                 if(is->pictq_size > 1){
1182                     is->frame_drops_late++;
1183                     pictq_next_picture(is);
1184                     goto retry;
1185                 }
1186             }
1187
1188             if(is->subtitle_st) {
1189                 if (is->subtitle_stream_changed) {
1190                     SDL_LockMutex(is->subpq_mutex);
1191
1192                     while (is->subpq_size) {
1193                         free_subpicture(&is->subpq[is->subpq_rindex]);
1194
1195                         /* update queue size and signal for next picture */
1196                         if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1197                             is->subpq_rindex = 0;
1198
1199                         is->subpq_size--;
1200                     }
1201                     is->subtitle_stream_changed = 0;
1202
1203                     SDL_CondSignal(is->subpq_cond);
1204                     SDL_UnlockMutex(is->subpq_mutex);
1205                 } else {
1206                     if (is->subpq_size > 0) {
1207                         sp = &is->subpq[is->subpq_rindex];
1208
1209                         if (is->subpq_size > 1)
1210                             sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1211                         else
1212                             sp2 = NULL;
1213
1214                         if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1215                                 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1216                         {
1217                             free_subpicture(sp);
1218
1219                             /* update queue size and signal for next picture */
1220                             if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1221                                 is->subpq_rindex = 0;
1222
1223                             SDL_LockMutex(is->subpq_mutex);
1224                             is->subpq_size--;
1225                             SDL_CondSignal(is->subpq_cond);
1226                             SDL_UnlockMutex(is->subpq_mutex);
1227                         }
1228                     }
1229                 }
1230             }
1231
1232             /* display picture */
1233             if (!display_disable)
1234                 video_display(is);
1235
1236             pictq_next_picture(is);
1237         }
1238     } else if (is->audio_st) {
1239         /* draw the next audio frame */
1240
1241         /* if only audio stream, then display the audio bars (better
1242            than nothing, just to test the implementation */
1243
1244         /* display picture */
1245         if (!display_disable)
1246             video_display(is);
1247     }
1248     if (show_status) {
1249         static int64_t last_time;
1250         int64_t cur_time;
1251         int aqsize, vqsize, sqsize;
1252         double av_diff;
1253
1254         cur_time = av_gettime();
1255         if (!last_time || (cur_time - last_time) >= 30000) {
1256             aqsize = 0;
1257             vqsize = 0;
1258             sqsize = 0;
1259             if (is->audio_st)
1260                 aqsize = is->audioq.size;
1261             if (is->video_st)
1262                 vqsize = is->videoq.size;
1263             if (is->subtitle_st)
1264                 sqsize = is->subtitleq.size;
1265             av_diff = 0;
1266             if (is->audio_st && is->video_st)
1267                 av_diff = get_audio_clock(is) - get_video_clock(is);
1268             printf("%7.2f A-V:%7.3f fd=%4d aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64"   \r",
1269                    get_master_clock(is),
1270                    av_diff,
1271                    is->frame_drops_early + is->frame_drops_late,
1272                    aqsize / 1024,
1273                    vqsize / 1024,
1274                    sqsize,
1275                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_dts : 0,
1276                    is->video_st ? is->video_st->codec->pts_correction_num_faulty_pts : 0);
1277             fflush(stdout);
1278             last_time = cur_time;
1279         }
1280     }
1281 }
1282
1283 /* allocate a picture (needs to do that in main thread to avoid
1284    potential locking problems */
1285 static void alloc_picture(void *opaque)
1286 {
1287     VideoState *is = opaque;
1288     VideoPicture *vp;
1289
1290     vp = &is->pictq[is->pictq_windex];
1291
1292     if (vp->bmp)
1293         SDL_FreeYUVOverlay(vp->bmp);
1294
1295 #if CONFIG_AVFILTER
1296     if (vp->picref)
1297         avfilter_unref_buffer(vp->picref);
1298     vp->picref = NULL;
1299
1300     vp->width   = is->out_video_filter->inputs[0]->w;
1301     vp->height  = is->out_video_filter->inputs[0]->h;
1302     vp->pix_fmt = is->out_video_filter->inputs[0]->format;
1303 #else
1304     vp->width   = is->video_st->codec->width;
1305     vp->height  = is->video_st->codec->height;
1306     vp->pix_fmt = is->video_st->codec->pix_fmt;
1307 #endif
1308
1309     vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1310                                    SDL_YV12_OVERLAY,
1311                                    screen);
1312     if (!vp->bmp || vp->bmp->pitches[0] < vp->width) {
1313         /* SDL allocates a buffer smaller than requested if the video
1314          * overlay hardware is unable to support the requested size. */
1315         fprintf(stderr, "Error: the video system does not support an image\n"
1316                         "size of %dx%d pixels. Try using -lowres or -vf \"scale=w:h\"\n"
1317                         "to reduce the image size.\n", vp->width, vp->height );
1318         do_exit(is);
1319     }
1320
1321     SDL_LockMutex(is->pictq_mutex);
1322     vp->allocated = 1;
1323     SDL_CondSignal(is->pictq_cond);
1324     SDL_UnlockMutex(is->pictq_mutex);
1325 }
1326
1327 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1328 {
1329     VideoPicture *vp;
1330     double frame_delay, pts = pts1;
1331
1332     /* compute the exact PTS for the picture if it is omitted in the stream
1333      * pts1 is the dts of the pkt / pts of the frame */
1334     if (pts != 0) {
1335         /* update video clock with pts, if present */
1336         is->video_clock = pts;
1337     } else {
1338         pts = is->video_clock;
1339     }
1340     /* update video clock for next frame */
1341     frame_delay = av_q2d(is->video_st->codec->time_base);
1342     /* for MPEG2, the frame can be repeated, so we update the
1343        clock accordingly */
1344     frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1345     is->video_clock += frame_delay;
1346
1347 #if defined(DEBUG_SYNC) && 0
1348     printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1349            av_get_picture_type_char(src_frame->pict_type), pts, pts1);
1350 #endif
1351
1352     /* wait until we have space to put a new picture */
1353     SDL_LockMutex(is->pictq_mutex);
1354
1355     while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1356            !is->videoq.abort_request) {
1357         SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1358     }
1359     SDL_UnlockMutex(is->pictq_mutex);
1360
1361     if (is->videoq.abort_request)
1362         return -1;
1363
1364     vp = &is->pictq[is->pictq_windex];
1365
1366     vp->duration = frame_delay;
1367
1368     /* alloc or resize hardware picture buffer */
1369     if (!vp->bmp || vp->reallocate ||
1370 #if CONFIG_AVFILTER
1371         vp->width  != is->out_video_filter->inputs[0]->w ||
1372         vp->height != is->out_video_filter->inputs[0]->h) {
1373 #else
1374         vp->width != is->video_st->codec->width ||
1375         vp->height != is->video_st->codec->height) {
1376 #endif
1377         SDL_Event event;
1378
1379         vp->allocated  = 0;
1380         vp->reallocate = 0;
1381
1382         /* the allocation must be done in the main thread to avoid
1383            locking problems */
1384         event.type = FF_ALLOC_EVENT;
1385         event.user.data1 = is;
1386         SDL_PushEvent(&event);
1387
1388         /* wait until the picture is allocated */
1389         SDL_LockMutex(is->pictq_mutex);
1390         while (!vp->allocated && !is->videoq.abort_request) {
1391             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1392         }
1393         /* if the queue is aborted, we have to pop the pending ALLOC event or wait for the allocation to complete */
1394         if (is->videoq.abort_request && SDL_PeepEvents(&event, 1, SDL_GETEVENT, SDL_EVENTMASK(FF_ALLOC_EVENT)) != 1) {
1395             while (!vp->allocated) {
1396                 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1397             }
1398         }
1399         SDL_UnlockMutex(is->pictq_mutex);
1400
1401         if (is->videoq.abort_request)
1402             return -1;
1403     }
1404
1405     /* if the frame is not skipped, then display it */
1406     if (vp->bmp) {
1407         AVPicture pict;
1408 #if CONFIG_AVFILTER
1409         if(vp->picref)
1410             avfilter_unref_buffer(vp->picref);
1411         vp->picref = src_frame->opaque;
1412 #endif
1413
1414         /* get a pointer on the bitmap */
1415         SDL_LockYUVOverlay (vp->bmp);
1416
1417         memset(&pict,0,sizeof(AVPicture));
1418         pict.data[0] = vp->bmp->pixels[0];
1419         pict.data[1] = vp->bmp->pixels[2];
1420         pict.data[2] = vp->bmp->pixels[1];
1421
1422         pict.linesize[0] = vp->bmp->pitches[0];
1423         pict.linesize[1] = vp->bmp->pitches[2];
1424         pict.linesize[2] = vp->bmp->pitches[1];
1425
1426 #if CONFIG_AVFILTER
1427         //FIXME use direct rendering
1428         av_picture_copy(&pict, (AVPicture *)src_frame,
1429                         vp->pix_fmt, vp->width, vp->height);
1430 #else
1431         sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1432         is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1433             vp->width, vp->height, vp->pix_fmt, vp->width, vp->height,
1434             PIX_FMT_YUV420P, sws_flags, NULL, NULL, NULL);
1435         if (is->img_convert_ctx == NULL) {
1436             fprintf(stderr, "Cannot initialize the conversion context\n");
1437             exit(1);
1438         }
1439         sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1440                   0, vp->height, pict.data, pict.linesize);
1441 #endif
1442         /* update the bitmap content */
1443         SDL_UnlockYUVOverlay(vp->bmp);
1444
1445         vp->pts = pts;
1446         vp->pos = pos;
1447         vp->skip = 0;
1448
1449         /* now we can update the picture count */
1450         if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1451             is->pictq_windex = 0;
1452         SDL_LockMutex(is->pictq_mutex);
1453         is->pictq_size++;
1454         SDL_UnlockMutex(is->pictq_mutex);
1455     }
1456     return 0;
1457 }
1458
1459 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1460 {
1461     int got_picture, i;
1462
1463     if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1464         return -1;
1465
1466     if (pkt->data == flush_pkt.data) {
1467         avcodec_flush_buffers(is->video_st->codec);
1468
1469         SDL_LockMutex(is->pictq_mutex);
1470         //Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1471         for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
1472             is->pictq[i].skip = 1;
1473         }
1474         while (is->pictq_size && !is->videoq.abort_request) {
1475             SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1476         }
1477         is->video_current_pos = -1;
1478         is->frame_last_pts = AV_NOPTS_VALUE;
1479         is->frame_last_duration = 0;
1480         is->frame_timer = (double)av_gettime() / 1000000.0;
1481         is->frame_last_dropped_pts = AV_NOPTS_VALUE;
1482         SDL_UnlockMutex(is->pictq_mutex);
1483
1484         return 0;
1485     }
1486
1487     avcodec_decode_video2(is->video_st->codec, frame, &got_picture, pkt);
1488
1489     if (got_picture) {
1490         int ret = 1;
1491
1492         if (decoder_reorder_pts == -1) {
1493             *pts = *(int64_t*)av_opt_ptr(avcodec_get_frame_class(), frame, "best_effort_timestamp");
1494         } else if (decoder_reorder_pts) {
1495             *pts = frame->pkt_pts;
1496         } else {
1497             *pts = frame->pkt_dts;
1498         }
1499
1500         if (*pts == AV_NOPTS_VALUE) {
1501             *pts = 0;
1502         }
1503
1504         if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) || is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK) &&
1505              (framedrop>0 || (framedrop && is->audio_st))) {
1506             SDL_LockMutex(is->pictq_mutex);
1507             if (is->frame_last_pts != AV_NOPTS_VALUE && *pts) {
1508                 double clockdiff = get_video_clock(is) - get_master_clock(is);
1509                 double dpts = av_q2d(is->video_st->time_base) * *pts;
1510                 double ptsdiff = dpts - is->frame_last_pts;
1511                 if (fabs(clockdiff) < AV_NOSYNC_THRESHOLD &&
1512                      ptsdiff > 0 && ptsdiff < AV_NOSYNC_THRESHOLD &&
1513                      clockdiff + ptsdiff - is->frame_last_filter_delay < 0) {
1514                     is->frame_last_dropped_pos = pkt->pos;
1515                     is->frame_last_dropped_pts = dpts;
1516                     is->frame_drops_early++;
1517                     ret = 0;
1518                 }
1519             }
1520             SDL_UnlockMutex(is->pictq_mutex);
1521         }
1522
1523         if (ret)
1524             is->frame_last_returned_time = av_gettime() / 1000000.0;
1525
1526         return ret;
1527     }
1528     return 0;
1529 }
1530
1531 #if CONFIG_AVFILTER
1532 typedef struct {
1533     VideoState *is;
1534     AVFrame *frame;
1535     int use_dr1;
1536 } FilterPriv;
1537
1538 static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
1539 {
1540     AVFilterContext *ctx = codec->opaque;
1541     AVFilterBufferRef  *ref;
1542     int perms = AV_PERM_WRITE;
1543     int i, w, h, stride[4];
1544     unsigned edge;
1545     int pixel_size;
1546
1547     av_assert0(codec->flags & CODEC_FLAG_EMU_EDGE);
1548
1549     if (codec->codec->capabilities & CODEC_CAP_NEG_LINESIZES)
1550         perms |= AV_PERM_NEG_LINESIZES;
1551
1552     if(pic->buffer_hints & FF_BUFFER_HINTS_VALID) {
1553         if(pic->buffer_hints & FF_BUFFER_HINTS_READABLE) perms |= AV_PERM_READ;
1554         if(pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) perms |= AV_PERM_PRESERVE;
1555         if(pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) perms |= AV_PERM_REUSE2;
1556     }
1557     if(pic->reference) perms |= AV_PERM_READ | AV_PERM_PRESERVE;
1558
1559     w = codec->width;
1560     h = codec->height;
1561
1562     if(av_image_check_size(w, h, 0, codec))
1563         return -1;
1564
1565     avcodec_align_dimensions2(codec, &w, &h, stride);
1566     edge = codec->flags & CODEC_FLAG_EMU_EDGE ? 0 : avcodec_get_edge_width();
1567     w += edge << 1;
1568     h += edge << 1;
1569     if (codec->pix_fmt != ctx->outputs[0]->format) {
1570         av_log(codec, AV_LOG_ERROR, "Pixel format mismatches %d %d\n", codec->pix_fmt, ctx->outputs[0]->format);
1571         return -1;
1572     }
1573     if(!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
1574         return -1;
1575
1576     pixel_size = av_pix_fmt_descriptors[ref->format].comp[0].step_minus1+1;
1577     ref->video->w = codec->width;
1578     ref->video->h = codec->height;
1579     for(i = 0; i < 4; i ++) {
1580         unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_w : 0;
1581         unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_h : 0;
1582
1583         if (ref->data[i]) {
1584             ref->data[i]    += ((edge * pixel_size) >> hshift) + ((edge * ref->linesize[i]) >> vshift);
1585         }
1586         pic->data[i]     = ref->data[i];
1587         pic->linesize[i] = ref->linesize[i];
1588     }
1589     pic->opaque = ref;
1590     pic->type   = FF_BUFFER_TYPE_USER;
1591     pic->reordered_opaque = codec->reordered_opaque;
1592     if(codec->pkt) pic->pkt_pts = codec->pkt->pts;
1593     else           pic->pkt_pts = AV_NOPTS_VALUE;
1594     return 0;
1595 }
1596
1597 static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
1598 {
1599     memset(pic->data, 0, sizeof(pic->data));
1600     avfilter_unref_buffer(pic->opaque);
1601 }
1602
1603 static int input_reget_buffer(AVCodecContext *codec, AVFrame *pic)
1604 {
1605     AVFilterBufferRef *ref = pic->opaque;
1606
1607     if (pic->data[0] == NULL) {
1608         pic->buffer_hints |= FF_BUFFER_HINTS_READABLE;
1609         return codec->get_buffer(codec, pic);
1610     }
1611
1612     if ((codec->width != ref->video->w) || (codec->height != ref->video->h) ||
1613         (codec->pix_fmt != ref->format)) {
1614         av_log(codec, AV_LOG_ERROR, "Picture properties changed.\n");
1615         return -1;
1616     }
1617
1618     pic->reordered_opaque = codec->reordered_opaque;
1619     if(codec->pkt) pic->pkt_pts = codec->pkt->pts;
1620     else           pic->pkt_pts = AV_NOPTS_VALUE;
1621     return 0;
1622 }
1623
1624 static int input_init(AVFilterContext *ctx, const char *args, void *opaque)
1625 {
1626     FilterPriv *priv = ctx->priv;
1627     AVCodecContext *codec;
1628     if(!opaque) return -1;
1629
1630     priv->is = opaque;
1631     codec    = priv->is->video_st->codec;
1632     codec->opaque = ctx;
1633     if((codec->codec->capabilities & CODEC_CAP_DR1)
1634     ) {
1635         av_assert0(codec->flags & CODEC_FLAG_EMU_EDGE);
1636         priv->use_dr1 = 1;
1637         codec->get_buffer     = input_get_buffer;
1638         codec->release_buffer = input_release_buffer;
1639         codec->reget_buffer   = input_reget_buffer;
1640         codec->thread_safe_callbacks = 1;
1641     }
1642
1643     priv->frame = avcodec_alloc_frame();
1644
1645     return 0;
1646 }
1647
1648 static void input_uninit(AVFilterContext *ctx)
1649 {
1650     FilterPriv *priv = ctx->priv;
1651     av_free(priv->frame);
1652 }
1653
1654 static int input_request_frame(AVFilterLink *link)
1655 {
1656     FilterPriv *priv = link->src->priv;
1657     AVFilterBufferRef *picref;
1658     int64_t pts = 0;
1659     AVPacket pkt;
1660     int ret;
1661
1662     while (!(ret = get_video_frame(priv->is, priv->frame, &pts, &pkt)))
1663         av_free_packet(&pkt);
1664     if (ret < 0)
1665         return -1;
1666
1667     if(priv->use_dr1 && priv->frame->opaque) {
1668         picref = avfilter_ref_buffer(priv->frame->opaque, ~0);
1669     } else {
1670         picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
1671         av_image_copy(picref->data, picref->linesize,
1672                       priv->frame->data, priv->frame->linesize,
1673                       picref->format, link->w, link->h);
1674     }
1675     av_free_packet(&pkt);
1676
1677     avfilter_copy_frame_props(picref, priv->frame);
1678     picref->pts = pts;
1679
1680     avfilter_start_frame(link, picref);
1681     avfilter_draw_slice(link, 0, link->h, 1);
1682     avfilter_end_frame(link);
1683
1684     return 0;
1685 }
1686
1687 static int input_query_formats(AVFilterContext *ctx)
1688 {
1689     FilterPriv *priv = ctx->priv;
1690     enum PixelFormat pix_fmts[] = {
1691         priv->is->video_st->codec->pix_fmt, PIX_FMT_NONE
1692     };
1693
1694     avfilter_set_common_pixel_formats(ctx, avfilter_make_format_list(pix_fmts));
1695     return 0;
1696 }
1697
1698 static int input_config_props(AVFilterLink *link)
1699 {
1700     FilterPriv *priv  = link->src->priv;
1701     AVStream *s = priv->is->video_st;
1702
1703     link->w = s->codec->width;
1704     link->h = s->codec->height;
1705     link->sample_aspect_ratio = s->sample_aspect_ratio.num ?
1706         s->sample_aspect_ratio : s->codec->sample_aspect_ratio;
1707     link->time_base = s->time_base;
1708
1709     return 0;
1710 }
1711
1712 static AVFilter input_filter =
1713 {
1714     .name      = "ffplay_input",
1715
1716     .priv_size = sizeof(FilterPriv),
1717
1718     .init      = input_init,
1719     .uninit    = input_uninit,
1720
1721     .query_formats = input_query_formats,
1722
1723     .inputs    = (AVFilterPad[]) {{ .name = NULL }},
1724     .outputs   = (AVFilterPad[]) {{ .name = "default",
1725                                     .type = AVMEDIA_TYPE_VIDEO,
1726                                     .request_frame = input_request_frame,
1727                                     .config_props  = input_config_props, },
1728                                   { .name = NULL }},
1729 };
1730
1731 static int configure_video_filters(AVFilterGraph *graph, VideoState *is, const char *vfilters)
1732 {
1733     char sws_flags_str[128];
1734     int ret;
1735     enum PixelFormat pix_fmts[] = { PIX_FMT_YUV420P, PIX_FMT_NONE };
1736     AVBufferSinkParams *buffersink_params = av_buffersink_params_alloc();
1737     AVFilterContext *filt_src = NULL, *filt_out = NULL;
1738     snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1739     graph->scale_sws_opts = av_strdup(sws_flags_str);
1740
1741     if ((ret = avfilter_graph_create_filter(&filt_src, &input_filter, "src",
1742                                             NULL, is, graph)) < 0)
1743         return ret;
1744 #if FF_API_OLD_VSINK_API
1745     ret = avfilter_graph_create_filter(&filt_out, avfilter_get_by_name("buffersink"), "out",
1746                                        NULL, pix_fmts, graph);
1747 #else
1748     buffersink_params->pixel_fmts = pix_fmts;
1749     ret = avfilter_graph_create_filter(&filt_out, avfilter_get_by_name("buffersink"), "out",
1750                                        NULL, buffersink_params, graph);
1751 #endif
1752     av_freep(&buffersink_params);
1753     if (ret < 0)
1754         return ret;
1755
1756     if(vfilters) {
1757         AVFilterInOut *outputs = avfilter_inout_alloc();
1758         AVFilterInOut *inputs  = avfilter_inout_alloc();
1759
1760         outputs->name    = av_strdup("in");
1761         outputs->filter_ctx = filt_src;
1762         outputs->pad_idx = 0;
1763         outputs->next    = NULL;
1764
1765         inputs->name    = av_strdup("out");
1766         inputs->filter_ctx = filt_out;
1767         inputs->pad_idx = 0;
1768         inputs->next    = NULL;
1769
1770         if ((ret = avfilter_graph_parse(graph, vfilters, &inputs, &outputs, NULL)) < 0)
1771             return ret;
1772     } else {
1773         if ((ret = avfilter_link(filt_src, 0, filt_out, 0)) < 0)
1774             return ret;
1775     }
1776
1777     if ((ret = avfilter_graph_config(graph, NULL)) < 0)
1778         return ret;
1779
1780     is->out_video_filter = filt_out;
1781
1782     return ret;
1783 }
1784
1785 #endif  /* CONFIG_AVFILTER */
1786
1787 static int video_thread(void *arg)
1788 {
1789     VideoState *is = arg;
1790     AVFrame *frame= avcodec_alloc_frame();
1791     int64_t pts_int = AV_NOPTS_VALUE, pos = -1;
1792     double pts;
1793     int ret;
1794
1795 #if CONFIG_AVFILTER
1796     AVFilterGraph *graph = avfilter_graph_alloc();
1797     AVFilterContext *filt_out = NULL;
1798     int last_w = is->video_st->codec->width;
1799     int last_h = is->video_st->codec->height;
1800
1801     if ((ret = configure_video_filters(graph, is, vfilters)) < 0)
1802         goto the_end;
1803     filt_out = is->out_video_filter;
1804 #endif
1805
1806     for(;;) {
1807 #if !CONFIG_AVFILTER
1808         AVPacket pkt;
1809 #else
1810         AVFilterBufferRef *picref;
1811         AVRational tb = filt_out->inputs[0]->time_base;
1812 #endif
1813         while (is->paused && !is->videoq.abort_request)
1814             SDL_Delay(10);
1815 #if CONFIG_AVFILTER
1816         if (   last_w != is->video_st->codec->width
1817             || last_h != is->video_st->codec->height) {
1818             av_log(NULL, AV_LOG_INFO, "Frame changed from size:%dx%d to size:%dx%d\n",
1819                    last_w, last_h, is->video_st->codec->width, is->video_st->codec->height);
1820             avfilter_graph_free(&graph);
1821             graph = avfilter_graph_alloc();
1822             if ((ret = configure_video_filters(graph, is, vfilters)) < 0)
1823                 goto the_end;
1824             filt_out = is->out_video_filter;
1825             last_w = is->video_st->codec->width;
1826             last_h = is->video_st->codec->height;
1827         }
1828         ret = av_buffersink_get_buffer_ref(filt_out, &picref, 0);
1829         if (picref) {
1830             avfilter_fill_frame_from_video_buffer_ref(frame, picref);
1831             pts_int = picref->pts;
1832             pos     = picref->pos;
1833             frame->opaque = picref;
1834         }
1835
1836         if (av_cmp_q(tb, is->video_st->time_base)) {
1837             av_unused int64_t pts1 = pts_int;
1838             pts_int = av_rescale_q(pts_int, tb, is->video_st->time_base);
1839             av_dlog(NULL, "video_thread(): "
1840                     "tb:%d/%d pts:%"PRId64" -> tb:%d/%d pts:%"PRId64"\n",
1841                     tb.num, tb.den, pts1,
1842                     is->video_st->time_base.num, is->video_st->time_base.den, pts_int);
1843         }
1844 #else
1845         ret = get_video_frame(is, frame, &pts_int, &pkt);
1846         pos = pkt.pos;
1847         av_free_packet(&pkt);
1848 #endif
1849
1850         if (ret < 0) goto the_end;
1851
1852         is->frame_last_filter_delay = av_gettime() / 1000000.0 - is->frame_last_returned_time;
1853         if (fabs(is->frame_last_filter_delay) > AV_NOSYNC_THRESHOLD / 10.0)
1854             is->frame_last_filter_delay = 0;
1855
1856 #if CONFIG_AVFILTER
1857         if (!picref)
1858             continue;
1859 #endif
1860
1861         pts = pts_int*av_q2d(is->video_st->time_base);
1862
1863         ret = queue_picture(is, frame, pts, pos);
1864
1865         if (ret < 0)
1866             goto the_end;
1867
1868         if (is->step)
1869             stream_toggle_pause(is);
1870     }
1871  the_end:
1872 #if CONFIG_AVFILTER
1873     avfilter_graph_free(&graph);
1874 #endif
1875     av_free(frame);
1876     return 0;
1877 }
1878
1879 static int subtitle_thread(void *arg)
1880 {
1881     VideoState *is = arg;
1882     SubPicture *sp;
1883     AVPacket pkt1, *pkt = &pkt1;
1884     int got_subtitle;
1885     double pts;
1886     int i, j;
1887     int r, g, b, y, u, v, a;
1888
1889     for(;;) {
1890         while (is->paused && !is->subtitleq.abort_request) {
1891             SDL_Delay(10);
1892         }
1893         if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1894             break;
1895
1896         if(pkt->data == flush_pkt.data){
1897             avcodec_flush_buffers(is->subtitle_st->codec);
1898             continue;
1899         }
1900         SDL_LockMutex(is->subpq_mutex);
1901         while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1902                !is->subtitleq.abort_request) {
1903             SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1904         }
1905         SDL_UnlockMutex(is->subpq_mutex);
1906
1907         if (is->subtitleq.abort_request)
1908             return 0;
1909
1910         sp = &is->subpq[is->subpq_windex];
1911
1912        /* NOTE: ipts is the PTS of the _first_ picture beginning in
1913            this packet, if any */
1914         pts = 0;
1915         if (pkt->pts != AV_NOPTS_VALUE)
1916             pts = av_q2d(is->subtitle_st->time_base)*pkt->pts;
1917
1918         avcodec_decode_subtitle2(is->subtitle_st->codec, &sp->sub,
1919                                  &got_subtitle, pkt);
1920
1921         if (got_subtitle && sp->sub.format == 0) {
1922             sp->pts = pts;
1923
1924             for (i = 0; i < sp->sub.num_rects; i++)
1925             {
1926                 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1927                 {
1928                     RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1929                     y = RGB_TO_Y_CCIR(r, g, b);
1930                     u = RGB_TO_U_CCIR(r, g, b, 0);
1931                     v = RGB_TO_V_CCIR(r, g, b, 0);
1932                     YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1933                 }
1934             }
1935
1936             /* now we can update the picture count */
1937             if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1938                 is->subpq_windex = 0;
1939             SDL_LockMutex(is->subpq_mutex);
1940             is->subpq_size++;
1941             SDL_UnlockMutex(is->subpq_mutex);
1942         }
1943         av_free_packet(pkt);
1944     }
1945     return 0;
1946 }
1947
1948 /* copy samples for viewing in editor window */
1949 static void update_sample_display(VideoState *is, short *samples, int samples_size)
1950 {
1951     int size, len;
1952
1953     size = samples_size / sizeof(short);
1954     while (size > 0) {
1955         len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
1956         if (len > size)
1957             len = size;
1958         memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
1959         samples += len;
1960         is->sample_array_index += len;
1961         if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
1962             is->sample_array_index = 0;
1963         size -= len;
1964     }
1965 }
1966
1967 /* return the new audio buffer size (samples can be added or deleted
1968    to get better sync if video or external master clock) */
1969 static int synchronize_audio(VideoState *is, short *samples,
1970                              int samples_size1, double pts)
1971 {
1972     int n, samples_size;
1973     double ref_clock;
1974
1975     n = av_get_bytes_per_sample(is->audio_tgt_fmt) * is->audio_tgt_channels;
1976     samples_size = samples_size1;
1977
1978     /* if not master, then we try to remove or add samples to correct the clock */
1979     if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
1980          is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1981         double diff, avg_diff;
1982         int wanted_size, min_size, max_size, nb_samples;
1983
1984         ref_clock = get_master_clock(is);
1985         diff = get_audio_clock(is) - ref_clock;
1986
1987         if (diff < AV_NOSYNC_THRESHOLD) {
1988             is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
1989             if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
1990                 /* not enough measures to have a correct estimate */
1991                 is->audio_diff_avg_count++;
1992             } else {
1993                 /* estimate the A-V difference */
1994                 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
1995
1996                 if (fabs(avg_diff) >= is->audio_diff_threshold) {
1997                     wanted_size = samples_size + ((int)(diff * is->audio_tgt_freq) * n);
1998                     nb_samples = samples_size / n;
1999
2000                     min_size = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2001                     max_size = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2002                     if (wanted_size < min_size)
2003                         wanted_size = min_size;
2004                     else if (wanted_size > FFMIN3(max_size, samples_size, sizeof(is->audio_buf2)))
2005                         wanted_size = FFMIN3(max_size, samples_size, sizeof(is->audio_buf2));
2006
2007                     /* add or remove samples to correction the synchro */
2008                     if (wanted_size < samples_size) {
2009                         /* remove samples */
2010                         samples_size = wanted_size;
2011                     } else if (wanted_size > samples_size) {
2012                         uint8_t *samples_end, *q;
2013                         int nb;
2014
2015                         /* add samples */
2016                         nb = (samples_size - wanted_size);
2017                         samples_end = (uint8_t *)samples + samples_size - n;
2018                         q = samples_end + n;
2019                         while (nb > 0) {
2020                             memcpy(q, samples_end, n);
2021                             q += n;
2022                             nb -= n;
2023                         }
2024                         samples_size = wanted_size;
2025                     }
2026                 }
2027                 av_dlog(NULL, "diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
2028                         diff, avg_diff, samples_size - samples_size1,
2029                         is->audio_clock, is->video_clock, is->audio_diff_threshold);
2030             }
2031         } else {
2032             /* too big difference : may be initial PTS errors, so
2033                reset A-V filter */
2034             is->audio_diff_avg_count = 0;
2035             is->audio_diff_cum = 0;
2036         }
2037     }
2038
2039     return samples_size;
2040 }
2041
2042 /* decode one audio frame and returns its uncompressed size */
2043 static int audio_decode_frame(VideoState *is, double *pts_ptr)
2044 {
2045     AVPacket *pkt_temp = &is->audio_pkt_temp;
2046     AVPacket *pkt = &is->audio_pkt;
2047     AVCodecContext *dec= is->audio_st->codec;
2048     int len1, len2, data_size, resampled_data_size;
2049     int64_t dec_channel_layout;
2050     int got_frame;
2051     double pts;
2052     int new_packet = 0;
2053     int flush_complete = 0;
2054
2055     for(;;) {
2056         /* NOTE: the audio packet can contain several frames */
2057         while (pkt_temp->size > 0 || (!pkt_temp->data && new_packet)) {
2058             if (!is->frame) {
2059                 if (!(is->frame = avcodec_alloc_frame()))
2060                     return AVERROR(ENOMEM);
2061             } else
2062                 avcodec_get_frame_defaults(is->frame);
2063
2064             if (flush_complete)
2065                 break;
2066             new_packet = 0;
2067             len1 = avcodec_decode_audio4(dec, is->frame, &got_frame, pkt_temp);
2068             if (len1 < 0) {
2069                 /* if error, we skip the frame */
2070                 pkt_temp->size = 0;
2071                 break;
2072             }
2073
2074             pkt_temp->data += len1;
2075             pkt_temp->size -= len1;
2076
2077             if (!got_frame) {
2078                 /* stop sending empty packets if the decoder is finished */
2079                 if (!pkt_temp->data && dec->codec->capabilities & CODEC_CAP_DELAY)
2080                     flush_complete = 1;
2081                 continue;
2082             }
2083             data_size = av_samples_get_buffer_size(NULL, dec->channels,
2084                                                    is->frame->nb_samples,
2085                                                    dec->sample_fmt, 1);
2086
2087             dec_channel_layout = (dec->channel_layout && dec->channels == av_get_channel_layout_nb_channels(dec->channel_layout)) ? dec->channel_layout : av_get_default_channel_layout(dec->channels);
2088
2089             if (dec->sample_fmt != is->audio_src_fmt || dec_channel_layout != is->audio_src_channel_layout || dec->sample_rate != is->audio_src_freq) {
2090                 if (is->swr_ctx)
2091                     swr_free(&is->swr_ctx);
2092                 is->swr_ctx = swr_alloc_set_opts(NULL,
2093                                                  is->audio_tgt_channel_layout, is->audio_tgt_fmt, is->audio_tgt_freq,
2094                                                  dec_channel_layout,           dec->sample_fmt,   dec->sample_rate,
2095                                                  0, NULL);
2096                 if (!is->swr_ctx || swr_init(is->swr_ctx) < 0) {
2097                     fprintf(stderr, "Cannot create sample rate converter for conversion of %d Hz %s %d channels to %d Hz %s %d channels!\n",
2098                         dec->sample_rate,
2099                         av_get_sample_fmt_name(dec->sample_fmt),
2100                         dec->channels,
2101                         is->audio_tgt_freq,
2102                         av_get_sample_fmt_name(is->audio_tgt_fmt),
2103                         is->audio_tgt_channels);
2104                     break;
2105                 }
2106                 is->audio_src_channel_layout = dec_channel_layout;
2107                 is->audio_src_channels = dec->channels;
2108                 is->audio_src_freq = dec->sample_rate;
2109                 is->audio_src_fmt = dec->sample_fmt;
2110             }
2111
2112             resampled_data_size = data_size;
2113             if (is->swr_ctx) {
2114                 const uint8_t *in[] = { is->frame->data[0] };
2115                 uint8_t *out[] = {is->audio_buf2};
2116                 len2 = swr_convert(is->swr_ctx, out, sizeof(is->audio_buf2) / is->audio_tgt_channels / av_get_bytes_per_sample(is->audio_tgt_fmt),
2117                                                 in, data_size / dec->channels / av_get_bytes_per_sample(dec->sample_fmt));
2118                 if (len2 < 0) {
2119                     fprintf(stderr, "audio_resample() failed\n");
2120                     break;
2121                 }
2122                 if (len2 == sizeof(is->audio_buf2) / is->audio_tgt_channels / av_get_bytes_per_sample(is->audio_tgt_fmt)) {
2123                     fprintf(stderr, "warning: audio buffer is probably too small\n");
2124                     swr_init(is->swr_ctx);
2125                 }
2126                 is->audio_buf = is->audio_buf2;
2127                 resampled_data_size = len2 * is->audio_tgt_channels * av_get_bytes_per_sample(is->audio_tgt_fmt);
2128             } else {
2129                 is->audio_buf = is->frame->data[0];
2130             }
2131
2132             /* if no pts, then compute it */
2133             pts = is->audio_clock;
2134             *pts_ptr = pts;
2135             is->audio_clock += (double)data_size / (dec->channels * dec->sample_rate * av_get_bytes_per_sample(dec->sample_fmt));
2136 #ifdef DEBUG
2137             {
2138                 static double last_clock;
2139                 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2140                        is->audio_clock - last_clock,
2141                        is->audio_clock, pts);
2142                 last_clock = is->audio_clock;
2143             }
2144 #endif
2145             return resampled_data_size;
2146         }
2147
2148         /* free the current packet */
2149         if (pkt->data)
2150             av_free_packet(pkt);
2151         memset(pkt_temp, 0, sizeof(*pkt_temp));
2152
2153         if (is->paused || is->audioq.abort_request) {
2154             return -1;
2155         }
2156
2157         /* read next packet */
2158         if ((new_packet = packet_queue_get(&is->audioq, pkt, 1)) < 0)
2159             return -1;
2160
2161         if (pkt->data == flush_pkt.data)
2162             avcodec_flush_buffers(dec);
2163
2164         *pkt_temp = *pkt;
2165
2166         /* if update the audio clock with the pts */
2167         if (pkt->pts != AV_NOPTS_VALUE) {
2168             is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2169         }
2170     }
2171 }
2172
2173 /* prepare a new audio buffer */
2174 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2175 {
2176     VideoState *is = opaque;
2177     int audio_size, len1;
2178     int bytes_per_sec;
2179     double pts;
2180
2181     audio_callback_time = av_gettime();
2182
2183     while (len > 0) {
2184         if (is->audio_buf_index >= is->audio_buf_size) {
2185            audio_size = audio_decode_frame(is, &pts);
2186            if (audio_size < 0) {
2187                 /* if error, just output silence */
2188                is->audio_buf      = is->silence_buf;
2189                is->audio_buf_size = sizeof(is->silence_buf);
2190            } else {
2191                if (is->show_mode != SHOW_MODE_VIDEO)
2192                    update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2193                audio_size = synchronize_audio(is, (int16_t *)is->audio_buf, audio_size,
2194                                               pts);
2195                is->audio_buf_size = audio_size;
2196            }
2197            is->audio_buf_index = 0;
2198         }
2199         len1 = is->audio_buf_size - is->audio_buf_index;
2200         if (len1 > len)
2201             len1 = len;
2202         memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2203         len -= len1;
2204         stream += len1;
2205         is->audio_buf_index += len1;
2206     }
2207     bytes_per_sec = is->audio_tgt_freq * is->audio_tgt_channels * av_get_bytes_per_sample(is->audio_tgt_fmt);
2208     is->audio_write_buf_size = is->audio_buf_size - is->audio_buf_index;
2209     /* Let's assume the audio driver that is used by SDL has two periods. */
2210     is->audio_current_pts = is->audio_clock - (double)(2 * is->audio_hw_buf_size + is->audio_write_buf_size) / bytes_per_sec;
2211     is->audio_current_pts_drift = is->audio_current_pts - audio_callback_time / 1000000.0;
2212 }
2213
2214 /* open a given stream. Return 0 if OK */
2215 static int stream_component_open(VideoState *is, int stream_index)
2216 {
2217     AVFormatContext *ic = is->ic;
2218     AVCodecContext *avctx;
2219     AVCodec *codec;
2220     SDL_AudioSpec wanted_spec, spec;
2221     AVDictionary *opts;
2222     AVDictionaryEntry *t = NULL;
2223     int64_t wanted_channel_layout = 0;
2224
2225     if (stream_index < 0 || stream_index >= ic->nb_streams)
2226         return -1;
2227     avctx = ic->streams[stream_index]->codec;
2228
2229     codec = avcodec_find_decoder(avctx->codec_id);
2230     opts = filter_codec_opts(codec_opts, codec, ic, ic->streams[stream_index]);
2231
2232     switch(avctx->codec_type){
2233         case AVMEDIA_TYPE_AUDIO   : if(audio_codec_name   ) codec= avcodec_find_decoder_by_name(   audio_codec_name); break;
2234         case AVMEDIA_TYPE_SUBTITLE: if(subtitle_codec_name) codec= avcodec_find_decoder_by_name(subtitle_codec_name); break;
2235         case AVMEDIA_TYPE_VIDEO   : if(video_codec_name   ) codec= avcodec_find_decoder_by_name(   video_codec_name); break;
2236     }
2237     if (!codec)
2238         return -1;
2239
2240     avctx->workaround_bugs = workaround_bugs;
2241     avctx->lowres = lowres;
2242     if(avctx->lowres > codec->max_lowres){
2243         av_log(avctx, AV_LOG_WARNING, "The maximum value for lowres supported by the decoder is %d\n",
2244                 codec->max_lowres);
2245         avctx->lowres= codec->max_lowres;
2246     }
2247     if(avctx->lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2248     avctx->idct_algo= idct;
2249     if(fast) avctx->flags2 |= CODEC_FLAG2_FAST;
2250     avctx->skip_frame= skip_frame;
2251     avctx->skip_idct= skip_idct;
2252     avctx->skip_loop_filter= skip_loop_filter;
2253     avctx->error_recognition= error_recognition;
2254     avctx->error_concealment= error_concealment;
2255
2256     if(codec->capabilities & CODEC_CAP_DR1)
2257         avctx->flags |= CODEC_FLAG_EMU_EDGE;
2258
2259     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2260         wanted_channel_layout = (avctx->channel_layout && avctx->channels == av_get_channel_layout_nb_channels(avctx->channels)) ? avctx->channel_layout : av_get_default_channel_layout(avctx->channels);
2261         wanted_channel_layout &= ~AV_CH_LAYOUT_STEREO_DOWNMIX;
2262         wanted_spec.channels = av_get_channel_layout_nb_channels(wanted_channel_layout);
2263         wanted_spec.freq = avctx->sample_rate;
2264         if (wanted_spec.freq <= 0 || wanted_spec.channels <= 0) {
2265             fprintf(stderr, "Invalid sample rate or channel count!\n");
2266             return -1;
2267         }
2268     }
2269
2270     if (!codec ||
2271         avcodec_open2(avctx, codec, &opts) < 0)
2272         return -1;
2273     if ((t = av_dict_get(opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2274         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2275         return AVERROR_OPTION_NOT_FOUND;
2276     }
2277
2278     /* prepare audio output */
2279     if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2280         wanted_spec.format = AUDIO_S16SYS;
2281         wanted_spec.silence = 0;
2282         wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2283         wanted_spec.callback = sdl_audio_callback;
2284         wanted_spec.userdata = is;
2285         if (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2286             fprintf(stderr, "SDL_OpenAudio: %s\n", SDL_GetError());
2287             return -1;
2288         }
2289         is->audio_hw_buf_size = spec.size;
2290         if (spec.format != AUDIO_S16SYS) {
2291             fprintf(stderr, "SDL advised audio format %d is not supported!\n", spec.format);
2292             return -1;
2293         }
2294         if (spec.channels != wanted_spec.channels) {
2295             wanted_channel_layout = av_get_default_channel_layout(spec.channels);
2296             if (!wanted_channel_layout) {
2297                 fprintf(stderr, "SDL advised channel count %d is not supported!\n", spec.channels);
2298                 return -1;
2299             }
2300         }
2301         is->audio_src_fmt = is->audio_tgt_fmt = AV_SAMPLE_FMT_S16;
2302         is->audio_src_freq = is->audio_tgt_freq = spec.freq;
2303         is->audio_src_channel_layout = is->audio_tgt_channel_layout = wanted_channel_layout;
2304         is->audio_src_channels = is->audio_tgt_channels = spec.channels;
2305     }
2306
2307     ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2308     switch(avctx->codec_type) {
2309     case AVMEDIA_TYPE_AUDIO:
2310         is->audio_stream = stream_index;
2311         is->audio_st = ic->streams[stream_index];
2312         is->audio_buf_size = 0;
2313         is->audio_buf_index = 0;
2314
2315         /* init averaging filter */
2316         is->audio_diff_avg_coef = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2317         is->audio_diff_avg_count = 0;
2318         /* since we do not have a precise anough audio fifo fullness,
2319            we correct audio sync only if larger than this threshold */
2320         is->audio_diff_threshold = 2.0 * SDL_AUDIO_BUFFER_SIZE / wanted_spec.freq;
2321
2322         memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2323         packet_queue_init(&is->audioq);
2324         SDL_PauseAudio(0);
2325         break;
2326     case AVMEDIA_TYPE_VIDEO:
2327         is->video_stream = stream_index;
2328         is->video_st = ic->streams[stream_index];
2329
2330         packet_queue_init(&is->videoq);
2331         is->video_tid = SDL_CreateThread(video_thread, is);
2332         break;
2333     case AVMEDIA_TYPE_SUBTITLE:
2334         is->subtitle_stream = stream_index;
2335         is->subtitle_st = ic->streams[stream_index];
2336         packet_queue_init(&is->subtitleq);
2337
2338         is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2339         break;
2340     default:
2341         break;
2342     }
2343     return 0;
2344 }
2345
2346 static void stream_component_close(VideoState *is, int stream_index)
2347 {
2348     AVFormatContext *ic = is->ic;
2349     AVCodecContext *avctx;
2350
2351     if (stream_index < 0 || stream_index >= ic->nb_streams)
2352         return;
2353     avctx = ic->streams[stream_index]->codec;
2354
2355     switch(avctx->codec_type) {
2356     case AVMEDIA_TYPE_AUDIO:
2357         packet_queue_abort(&is->audioq);
2358
2359         SDL_CloseAudio();
2360
2361         packet_queue_end(&is->audioq);
2362         if (is->swr_ctx)
2363             swr_free(&is->swr_ctx);
2364         av_free_packet(&is->audio_pkt);
2365         av_freep(&is->audio_buf1);
2366         is->audio_buf = NULL;
2367         av_freep(&is->frame);
2368
2369         if (is->rdft) {
2370             av_rdft_end(is->rdft);
2371             av_freep(&is->rdft_data);
2372             is->rdft = NULL;
2373             is->rdft_bits = 0;
2374         }
2375         break;
2376     case AVMEDIA_TYPE_VIDEO:
2377         packet_queue_abort(&is->videoq);
2378
2379         /* note: we also signal this mutex to make sure we deblock the
2380            video thread in all cases */
2381         SDL_LockMutex(is->pictq_mutex);
2382         SDL_CondSignal(is->pictq_cond);
2383         SDL_UnlockMutex(is->pictq_mutex);
2384
2385         SDL_WaitThread(is->video_tid, NULL);
2386
2387         packet_queue_end(&is->videoq);
2388         break;
2389     case AVMEDIA_TYPE_SUBTITLE:
2390         packet_queue_abort(&is->subtitleq);
2391
2392         /* note: we also signal this mutex to make sure we deblock the
2393            video thread in all cases */
2394         SDL_LockMutex(is->subpq_mutex);
2395         is->subtitle_stream_changed = 1;
2396
2397         SDL_CondSignal(is->subpq_cond);
2398         SDL_UnlockMutex(is->subpq_mutex);
2399
2400         SDL_WaitThread(is->subtitle_tid, NULL);
2401
2402         packet_queue_end(&is->subtitleq);
2403         break;
2404     default:
2405         break;
2406     }
2407
2408     ic->streams[stream_index]->discard = AVDISCARD_ALL;
2409     avcodec_close(avctx);
2410     switch(avctx->codec_type) {
2411     case AVMEDIA_TYPE_AUDIO:
2412         is->audio_st = NULL;
2413         is->audio_stream = -1;
2414         break;
2415     case AVMEDIA_TYPE_VIDEO:
2416         is->video_st = NULL;
2417         is->video_stream = -1;
2418         break;
2419     case AVMEDIA_TYPE_SUBTITLE:
2420         is->subtitle_st = NULL;
2421         is->subtitle_stream = -1;
2422         break;
2423     default:
2424         break;
2425     }
2426 }
2427
2428 static int decode_interrupt_cb(void *ctx)
2429 {
2430     VideoState *is = ctx;
2431     return is->abort_request;
2432 }
2433
2434 /* this thread gets the stream from the disk or the network */
2435 static int read_thread(void *arg)
2436 {
2437     VideoState *is = arg;
2438     AVFormatContext *ic = NULL;
2439     int err, i, ret;
2440     int st_index[AVMEDIA_TYPE_NB];
2441     AVPacket pkt1, *pkt = &pkt1;
2442     int eof=0;
2443     int pkt_in_play_range = 0;
2444     AVDictionaryEntry *t;
2445     AVDictionary **opts;
2446     int orig_nb_streams;
2447
2448     memset(st_index, -1, sizeof(st_index));
2449     is->video_stream = -1;
2450     is->audio_stream = -1;
2451     is->subtitle_stream = -1;
2452
2453     ic = avformat_alloc_context();
2454     ic->interrupt_callback.callback = decode_interrupt_cb;
2455     ic->interrupt_callback.opaque = is;
2456     err = avformat_open_input(&ic, is->filename, is->iformat, &format_opts);
2457     if (err < 0) {
2458         print_error(is->filename, err);
2459         ret = -1;
2460         goto fail;
2461     }
2462     if ((t = av_dict_get(format_opts, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
2463         av_log(NULL, AV_LOG_ERROR, "Option %s not found.\n", t->key);
2464         ret = AVERROR_OPTION_NOT_FOUND;
2465         goto fail;
2466     }
2467     is->ic = ic;
2468
2469     if(genpts)
2470         ic->flags |= AVFMT_FLAG_GENPTS;
2471
2472     opts = setup_find_stream_info_opts(ic, codec_opts);
2473     orig_nb_streams = ic->nb_streams;
2474
2475     err = avformat_find_stream_info(ic, opts);
2476     if (err < 0) {
2477         fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2478         ret = -1;
2479         goto fail;
2480     }
2481     for (i = 0; i < orig_nb_streams; i++)
2482         av_dict_free(&opts[i]);
2483     av_freep(&opts);
2484
2485     if(ic->pb)
2486         ic->pb->eof_reached= 0; //FIXME hack, ffplay maybe should not use url_feof() to test for the end
2487
2488     if(seek_by_bytes<0)
2489         seek_by_bytes= !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2490
2491     /* if seeking requested, we execute it */
2492     if (start_time != AV_NOPTS_VALUE) {
2493         int64_t timestamp;
2494
2495         timestamp = start_time;
2496         /* add the stream start time */
2497         if (ic->start_time != AV_NOPTS_VALUE)
2498             timestamp += ic->start_time;
2499         ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2500         if (ret < 0) {
2501             fprintf(stderr, "%s: could not seek to position %0.3f\n",
2502                     is->filename, (double)timestamp / AV_TIME_BASE);
2503         }
2504     }
2505
2506     for (i = 0; i < ic->nb_streams; i++)
2507         ic->streams[i]->discard = AVDISCARD_ALL;
2508     if (!video_disable)
2509         st_index[AVMEDIA_TYPE_VIDEO] =
2510             av_find_best_stream(ic, AVMEDIA_TYPE_VIDEO,
2511                                 wanted_stream[AVMEDIA_TYPE_VIDEO], -1, NULL, 0);
2512     if (!audio_disable)
2513         st_index[AVMEDIA_TYPE_AUDIO] =
2514             av_find_best_stream(ic, AVMEDIA_TYPE_AUDIO,
2515                                 wanted_stream[AVMEDIA_TYPE_AUDIO],
2516                                 st_index[AVMEDIA_TYPE_VIDEO],
2517                                 NULL, 0);
2518     if (!video_disable)
2519         st_index[AVMEDIA_TYPE_SUBTITLE] =
2520             av_find_best_stream(ic, AVMEDIA_TYPE_SUBTITLE,
2521                                 wanted_stream[AVMEDIA_TYPE_SUBTITLE],
2522                                 (st_index[AVMEDIA_TYPE_AUDIO] >= 0 ?
2523                                  st_index[AVMEDIA_TYPE_AUDIO] :
2524                                  st_index[AVMEDIA_TYPE_VIDEO]),
2525                                 NULL, 0);
2526     if (show_status) {
2527         av_dump_format(ic, 0, is->filename, 0);
2528     }
2529
2530     is->show_mode = show_mode;
2531
2532     /* open the streams */
2533     if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2534         stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2535     }
2536
2537     ret=-1;
2538     if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2539         ret= stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2540     }
2541     is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2542     if (is->show_mode == SHOW_MODE_NONE)
2543         is->show_mode = ret >= 0 ? SHOW_MODE_VIDEO : SHOW_MODE_RDFT;
2544
2545     if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2546         stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2547     }
2548
2549     if (is->video_stream < 0 && is->audio_stream < 0) {
2550         fprintf(stderr, "%s: could not open codecs\n", is->filename);
2551         ret = -1;
2552         goto fail;
2553     }
2554
2555     for(;;) {
2556         if (is->abort_request)
2557             break;
2558         if (is->paused != is->last_paused) {
2559             is->last_paused = is->paused;
2560             if (is->paused)
2561                 is->read_pause_return= av_read_pause(ic);
2562             else
2563                 av_read_play(ic);
2564         }
2565 #if CONFIG_RTSP_DEMUXER || CONFIG_MMSH_PROTOCOL
2566         if (is->paused &&
2567                 (!strcmp(ic->iformat->name, "rtsp") ||
2568                  (ic->pb && !strncmp(input_filename, "mmsh:", 5)))) {
2569             /* wait 10 ms to avoid trying to get another packet */
2570             /* XXX: horrible */
2571             SDL_Delay(10);
2572             continue;
2573         }
2574 #endif
2575         if (is->seek_req) {
2576             int64_t seek_target= is->seek_pos;
2577             int64_t seek_min= is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2578             int64_t seek_max= is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2579 //FIXME the +-2 is due to rounding being not done in the correct direction in generation
2580 //      of the seek_pos/seek_rel variables
2581
2582             ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2583             if (ret < 0) {
2584                 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2585             }else{
2586                 if (is->audio_stream >= 0) {
2587                     packet_queue_flush(&is->audioq);
2588                     packet_queue_put(&is->audioq, &flush_pkt);
2589                 }
2590                 if (is->subtitle_stream >= 0) {
2591                     packet_queue_flush(&is->subtitleq);
2592                     packet_queue_put(&is->subtitleq, &flush_pkt);
2593                 }
2594                 if (is->video_stream >= 0) {
2595                     packet_queue_flush(&is->videoq);
2596                     packet_queue_put(&is->videoq, &flush_pkt);
2597                 }
2598             }
2599             is->seek_req = 0;
2600             eof= 0;
2601         }
2602
2603         /* if the queue are full, no need to read more */
2604         if (   is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2605             || (   (is->audioq   .size  > MIN_AUDIOQ_SIZE || is->audio_stream<0)
2606                 && (is->videoq   .nb_packets > MIN_FRAMES || is->video_stream<0)
2607                 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream<0))) {
2608             /* wait 10 ms */
2609             SDL_Delay(10);
2610             continue;
2611         }
2612         if(eof) {
2613             if(is->video_stream >= 0){
2614                 av_init_packet(pkt);
2615                 pkt->data=NULL;
2616                 pkt->size=0;
2617                 pkt->stream_index= is->video_stream;
2618                 packet_queue_put(&is->videoq, pkt);
2619             }
2620             if (is->audio_stream >= 0 &&
2621                 is->audio_st->codec->codec->capabilities & CODEC_CAP_DELAY) {
2622                 av_init_packet(pkt);
2623                 pkt->data = NULL;
2624                 pkt->size = 0;
2625                 pkt->stream_index = is->audio_stream;
2626                 packet_queue_put(&is->audioq, pkt);
2627             }
2628             SDL_Delay(10);
2629             if(is->audioq.size + is->videoq.size + is->subtitleq.size ==0){
2630                 if(loop!=1 && (!loop || --loop)){
2631                     stream_seek(is, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2632                 }else if(autoexit){
2633                     ret=AVERROR_EOF;
2634                     goto fail;
2635                 }
2636             }
2637             eof=0;
2638             continue;
2639         }
2640         ret = av_read_frame(ic, pkt);
2641         if (ret < 0) {
2642             if (ret == AVERROR_EOF || url_feof(ic->pb))
2643                 eof=1;
2644             if (ic->pb && ic->pb->error)
2645                 break;
2646             SDL_Delay(100); /* wait for user event */
2647             continue;
2648         }
2649         /* check if packet is in play range specified by user, then queue, otherwise discard */
2650         pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2651                 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2652                 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2653                 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0)/1000000
2654                 <= ((double)duration/1000000);
2655         if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2656             packet_queue_put(&is->audioq, pkt);
2657         } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2658             packet_queue_put(&is->videoq, pkt);
2659         } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2660             packet_queue_put(&is->subtitleq, pkt);
2661         } else {
2662             av_free_packet(pkt);
2663         }
2664     }
2665     /* wait until the end */
2666     while (!is->abort_request) {
2667         SDL_Delay(100);
2668     }
2669
2670     ret = 0;
2671  fail:
2672     /* close each stream */
2673     if (is->audio_stream >= 0)
2674         stream_component_close(is, is->audio_stream);
2675     if (is->video_stream >= 0)
2676         stream_component_close(is, is->video_stream);
2677     if (is->subtitle_stream >= 0)
2678         stream_component_close(is, is->subtitle_stream);
2679     if (is->ic) {
2680         avformat_close_input(&is->ic);
2681     }
2682
2683     if (ret != 0) {
2684         SDL_Event event;
2685
2686         event.type = FF_QUIT_EVENT;
2687         event.user.data1 = is;
2688         SDL_PushEvent(&event);
2689     }
2690     return 0;
2691 }
2692
2693 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2694 {
2695     VideoState *is;
2696
2697     is = av_mallocz(sizeof(VideoState));
2698     if (!is)
2699         return NULL;
2700     av_strlcpy(is->filename, filename, sizeof(is->filename));
2701     is->iformat = iformat;
2702     is->ytop = 0;
2703     is->xleft = 0;
2704
2705     /* start video display */
2706     is->pictq_mutex = SDL_CreateMutex();
2707     is->pictq_cond = SDL_CreateCond();
2708
2709     is->subpq_mutex = SDL_CreateMutex();
2710     is->subpq_cond = SDL_CreateCond();
2711
2712     is->av_sync_type = av_sync_type;
2713     is->read_tid = SDL_CreateThread(read_thread, is);
2714     if (!is->read_tid) {
2715         av_free(is);
2716         return NULL;
2717     }
2718     return is;
2719 }
2720
2721 static void stream_cycle_channel(VideoState *is, int codec_type)
2722 {
2723     AVFormatContext *ic = is->ic;
2724     int start_index, stream_index;
2725     AVStream *st;
2726
2727     if (codec_type == AVMEDIA_TYPE_VIDEO)
2728         start_index = is->video_stream;
2729     else if (codec_type == AVMEDIA_TYPE_AUDIO)
2730         start_index = is->audio_stream;
2731     else
2732         start_index = is->subtitle_stream;
2733     if (start_index < (codec_type == AVMEDIA_TYPE_SUBTITLE ? -1 : 0))
2734         return;
2735     stream_index = start_index;
2736     for(;;) {
2737         if (++stream_index >= is->ic->nb_streams)
2738         {
2739             if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2740             {
2741                 stream_index = -1;
2742                 goto the_end;
2743             } else
2744                 stream_index = 0;
2745         }
2746         if (stream_index == start_index)
2747             return;
2748         st = ic->streams[stream_index];
2749         if (st->codec->codec_type == codec_type) {
2750             /* check that parameters are OK */
2751             switch(codec_type) {
2752             case AVMEDIA_TYPE_AUDIO:
2753                 if (st->codec->sample_rate != 0 &&
2754                     st->codec->channels != 0)
2755                     goto the_end;
2756                 break;
2757             case AVMEDIA_TYPE_VIDEO:
2758             case AVMEDIA_TYPE_SUBTITLE:
2759                 goto the_end;
2760             default:
2761                 break;
2762             }
2763         }
2764     }
2765  the_end:
2766     stream_component_close(is, start_index);
2767     stream_component_open(is, stream_index);
2768 }
2769
2770
2771 static void toggle_full_screen(VideoState *is)
2772 {
2773     int i;
2774     is_full_screen = !is_full_screen;
2775 #if defined(__APPLE__) && SDL_VERSION_ATLEAST(1, 2, 14)
2776     /* OS X needs to reallocate the SDL overlays */
2777     for (i = 0; i < VIDEO_PICTURE_QUEUE_SIZE; i++) {
2778         is->pictq[i].reallocate = 1;
2779     }
2780 #endif
2781     video_open(is, 1);
2782 }
2783
2784 static void toggle_pause(VideoState *is)
2785 {
2786     stream_toggle_pause(is);
2787     is->step = 0;
2788 }
2789
2790 static void step_to_next_frame(VideoState *is)
2791 {
2792     /* if the stream is paused unpause it, then step */
2793     if (is->paused)
2794         stream_toggle_pause(is);
2795     is->step = 1;
2796 }
2797
2798 static void toggle_audio_display(VideoState *is)
2799 {
2800     int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2801     is->show_mode = (is->show_mode + 1) % SHOW_MODE_NB;
2802     fill_rectangle(screen,
2803                 is->xleft, is->ytop, is->width, is->height,
2804                 bgcolor);
2805     SDL_UpdateRect(screen, is->xleft, is->ytop, is->width, is->height);
2806 }
2807
2808 /* handle an event sent by the GUI */
2809 static void event_loop(VideoState *cur_stream)
2810 {
2811     SDL_Event event;
2812     double incr, pos, frac;
2813
2814     for(;;) {
2815         double x;
2816         SDL_WaitEvent(&event);
2817         switch(event.type) {
2818         case SDL_KEYDOWN:
2819             if (exit_on_keydown) {
2820                 do_exit(cur_stream);
2821                 break;
2822             }
2823             switch(event.key.keysym.sym) {
2824             case SDLK_ESCAPE:
2825             case SDLK_q:
2826                 do_exit(cur_stream);
2827                 break;
2828             case SDLK_f:
2829                 toggle_full_screen(cur_stream);
2830                 break;
2831             case SDLK_p:
2832             case SDLK_SPACE:
2833                 toggle_pause(cur_stream);
2834                 break;
2835             case SDLK_s: //S: Step to next frame
2836                 step_to_next_frame(cur_stream);
2837                 break;
2838             case SDLK_a:
2839                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2840                 break;
2841             case SDLK_v:
2842                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2843                 break;
2844             case SDLK_t:
2845                 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2846                 break;
2847             case SDLK_w:
2848                 toggle_audio_display(cur_stream);
2849                 break;
2850             case SDLK_PAGEUP:
2851                 incr = 600.0;
2852                 goto do_seek;
2853             case SDLK_PAGEDOWN:
2854                 incr = -600.0;
2855                 goto do_seek;
2856             case SDLK_LEFT:
2857                 incr = -10.0;
2858                 goto do_seek;
2859             case SDLK_RIGHT:
2860                 incr = 10.0;
2861                 goto do_seek;
2862             case SDLK_UP:
2863                 incr = 60.0;
2864                 goto do_seek;
2865             case SDLK_DOWN:
2866                 incr = -60.0;
2867             do_seek:
2868                 if (seek_by_bytes) {
2869                     if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos>=0){
2870                         pos= cur_stream->video_current_pos;
2871                     }else if(cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos>=0){
2872                         pos= cur_stream->audio_pkt.pos;
2873                     }else
2874                         pos = avio_tell(cur_stream->ic->pb);
2875                     if (cur_stream->ic->bit_rate)
2876                         incr *= cur_stream->ic->bit_rate / 8.0;
2877                     else
2878                         incr *= 180000.0;
2879                     pos += incr;
2880                     stream_seek(cur_stream, pos, incr, 1);
2881                 } else {
2882                     pos = get_master_clock(cur_stream);
2883                     pos += incr;
2884                     stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2885                 }
2886                 break;
2887             default:
2888                 break;
2889             }
2890             break;
2891         case SDL_MOUSEBUTTONDOWN:
2892             if (exit_on_mousedown) {
2893                 do_exit(cur_stream);
2894                 break;
2895             }
2896         case SDL_MOUSEMOTION:
2897             if(event.type ==SDL_MOUSEBUTTONDOWN){
2898                 x= event.button.x;
2899             }else{
2900                 if(event.motion.state != SDL_PRESSED)
2901                     break;
2902                 x= event.motion.x;
2903             }
2904             if(seek_by_bytes || cur_stream->ic->duration<=0){
2905                 uint64_t size=  avio_size(cur_stream->ic->pb);
2906                 stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2907             }else{
2908                 int64_t ts;
2909                 int ns, hh, mm, ss;
2910                 int tns, thh, tmm, tss;
2911                 tns = cur_stream->ic->duration/1000000LL;
2912                 thh = tns/3600;
2913                 tmm = (tns%3600)/60;
2914                 tss = (tns%60);
2915                 frac = x/cur_stream->width;
2916                 ns = frac*tns;
2917                 hh = ns/3600;
2918                 mm = (ns%3600)/60;
2919                 ss = (ns%60);
2920                 fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d)       \n", frac*100,
2921                         hh, mm, ss, thh, tmm, tss);
2922                 ts = frac*cur_stream->ic->duration;
2923                 if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2924                     ts += cur_stream->ic->start_time;
2925                 stream_seek(cur_stream, ts, 0, 0);
2926             }
2927             break;
2928         case SDL_VIDEORESIZE:
2929             screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2930                                       SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2931             screen_width = cur_stream->width = event.resize.w;
2932             screen_height= cur_stream->height= event.resize.h;
2933             break;
2934         case SDL_QUIT:
2935         case FF_QUIT_EVENT:
2936             do_exit(cur_stream);
2937             break;
2938         case FF_ALLOC_EVENT:
2939             video_open(event.user.data1, 0);
2940             alloc_picture(event.user.data1);
2941             break;
2942         case FF_REFRESH_EVENT:
2943             video_refresh(event.user.data1);
2944             cur_stream->refresh=0;
2945             break;
2946         default:
2947             break;
2948         }
2949     }
2950 }
2951
2952 static int opt_frame_size(const char *opt, const char *arg)
2953 {
2954     av_log(NULL, AV_LOG_WARNING, "Option -s is deprecated, use -video_size.\n");
2955     return opt_default("video_size", arg);
2956 }
2957
2958 static int opt_width(const char *opt, const char *arg)
2959 {
2960     screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2961     return 0;
2962 }
2963
2964 static int opt_height(const char *opt, const char *arg)
2965 {
2966     screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2967     return 0;
2968 }
2969
2970 static int opt_format(const char *opt, const char *arg)
2971 {
2972     file_iformat = av_find_input_format(arg);
2973     if (!file_iformat) {
2974         fprintf(stderr, "Unknown input format: %s\n", arg);
2975         return AVERROR(EINVAL);
2976     }
2977     return 0;
2978 }
2979
2980 static int opt_frame_pix_fmt(const char *opt, const char *arg)
2981 {
2982     av_log(NULL, AV_LOG_WARNING, "Option -pix_fmt is deprecated, use -pixel_format.\n");
2983     return opt_default("pixel_format", arg);
2984 }
2985
2986 static int opt_sync(const char *opt, const char *arg)
2987 {
2988     if (!strcmp(arg, "audio"))
2989         av_sync_type = AV_SYNC_AUDIO_MASTER;
2990     else if (!strcmp(arg, "video"))
2991         av_sync_type = AV_SYNC_VIDEO_MASTER;
2992     else if (!strcmp(arg, "ext"))
2993         av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
2994     else {
2995         fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
2996         exit(1);
2997     }
2998     return 0;
2999 }
3000
3001 static int opt_seek(const char *opt, const char *arg)
3002 {
3003     start_time = parse_time_or_die(opt, arg, 1);
3004     return 0;
3005 }
3006
3007 static int opt_duration(const char *opt, const char *arg)
3008 {
3009     duration = parse_time_or_die(opt, arg, 1);
3010     return 0;
3011 }
3012
3013 static int opt_show_mode(const char *opt, const char *arg)
3014 {
3015     show_mode = !strcmp(arg, "video") ? SHOW_MODE_VIDEO :
3016                 !strcmp(arg, "waves") ? SHOW_MODE_WAVES :
3017                 !strcmp(arg, "rdft" ) ? SHOW_MODE_RDFT  :
3018                 parse_number_or_die(opt, arg, OPT_INT, 0, SHOW_MODE_NB-1);
3019     return 0;
3020 }
3021
3022 static void opt_input_file(void *optctx, const char *filename)
3023 {
3024     if (input_filename) {
3025         fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
3026                 filename, input_filename);
3027         exit_program(1);
3028     }
3029     if (!strcmp(filename, "-"))
3030         filename = "pipe:";
3031     input_filename = filename;
3032 }
3033
3034 static int opt_codec(void *o, const char *opt, const char *arg)
3035 {
3036     switch(opt[strlen(opt)-1]){
3037     case 'a' :    audio_codec_name = arg; break;
3038     case 's' : subtitle_codec_name = arg; break;
3039     case 'v' :    video_codec_name = arg; break;
3040     }
3041     return 0;
3042 }
3043
3044 static int dummy;
3045
3046 static const OptionDef options[] = {
3047 #include "cmdutils_common_opts.h"
3048     { "x", HAS_ARG, {(void*)opt_width}, "force displayed width", "width" },
3049     { "y", HAS_ARG, {(void*)opt_height}, "force displayed height", "height" },
3050     { "s", HAS_ARG | OPT_VIDEO, {(void*)opt_frame_size}, "set frame size (WxH or abbreviation)", "size" },
3051     { "fs", OPT_BOOL, {(void*)&is_full_screen}, "force full screen" },
3052     { "an", OPT_BOOL, {(void*)&audio_disable}, "disable audio" },
3053     { "vn", OPT_BOOL, {(void*)&video_disable}, "disable video" },
3054     { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_AUDIO]}, "select desired audio stream", "stream_number" },
3055     { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_VIDEO]}, "select desired video stream", "stream_number" },
3056     { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE]}, "select desired subtitle stream", "stream_number" },
3057     { "ss", HAS_ARG, {(void*)&opt_seek}, "seek to a given position in seconds", "pos" },
3058     { "t", HAS_ARG, {(void*)&opt_duration}, "play  \"duration\" seconds of audio/video", "duration" },
3059     { "bytes", OPT_INT | HAS_ARG, {(void*)&seek_by_bytes}, "seek by bytes 0=off 1=on -1=auto", "val" },
3060     { "nodisp", OPT_BOOL, {(void*)&display_disable}, "disable graphical display" },
3061     { "f", HAS_ARG, {(void*)opt_format}, "force format", "fmt" },
3062     { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, {(void*)opt_frame_pix_fmt}, "set pixel format", "format" },
3063     { "stats", OPT_BOOL | OPT_EXPERT, {(void*)&show_status}, "show status", "" },
3064     { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&workaround_bugs}, "workaround bugs", "" },
3065     { "fast", OPT_BOOL | OPT_EXPERT, {(void*)&fast}, "non spec compliant optimizations", "" },
3066     { "genpts", OPT_BOOL | OPT_EXPERT, {(void*)&genpts}, "generate pts", "" },
3067     { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&decoder_reorder_pts}, "let decoder reorder pts 0=off 1=on -1=auto", ""},
3068     { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&lowres}, "", "" },
3069     { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_loop_filter}, "", "" },
3070     { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_frame}, "", "" },
3071     { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_idct}, "", "" },
3072     { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&idct}, "set idct algo",  "algo" },
3073     { "er", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_recognition}, "set error detection threshold (0-4)",  "threshold" },
3074     { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_concealment}, "set error concealment options",  "bit_mask" },
3075     { "sync", HAS_ARG | OPT_EXPERT, {(void*)opt_sync}, "set audio-video sync. type (type=audio/video/ext)", "type" },
3076     { "autoexit", OPT_BOOL | OPT_EXPERT, {(void*)&autoexit}, "exit at the end", "" },
3077     { "exitonkeydown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_keydown}, "exit on key down", "" },
3078     { "exitonmousedown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_mousedown}, "exit on mouse down", "" },
3079     { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&loop}, "set number of times the playback shall be looped", "loop count" },
3080     { "framedrop", OPT_BOOL | OPT_EXPERT, {(void*)&framedrop}, "drop frames when cpu is too slow", "" },
3081     { "window_title", OPT_STRING | HAS_ARG, {(void*)&window_title}, "set window title", "window title" },
3082 #if CONFIG_AVFILTER
3083     { "vf", OPT_STRING | HAS_ARG, {(void*)&vfilters}, "video filters", "filter list" },
3084 #endif
3085     { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, {(void*)&rdftspeed}, "rdft speed", "msecs" },
3086     { "showmode", HAS_ARG, {(void*)opt_show_mode}, "select show mode (0 = video, 1 = waves, 2 = RDFT)", "mode" },
3087     { "default", HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, {(void*)opt_default}, "generic catch all option", "" },
3088     { "i", OPT_BOOL, {(void *)&dummy}, "read specified file", "input_file"},
3089     { "codec", HAS_ARG | OPT_FUNC2, {(void*)opt_codec}, "force decoder", "decoder" },
3090     { NULL, },
3091 };
3092
3093 static void show_usage(void)
3094 {
3095     printf("Simple media player\n");
3096     printf("usage: %s [options] input_file\n", program_name);
3097     printf("\n");
3098 }
3099
3100 static int opt_help(const char *opt, const char *arg)
3101 {
3102     av_log_set_callback(log_callback_help);
3103     show_usage();
3104     show_help_options(options, "Main options:\n",
3105                       OPT_EXPERT, 0);
3106     show_help_options(options, "\nAdvanced options:\n",
3107                       OPT_EXPERT, OPT_EXPERT);
3108     printf("\n");
3109     show_help_children(avcodec_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3110     show_help_children(avformat_get_class(), AV_OPT_FLAG_DECODING_PARAM);
3111 #if !CONFIG_AVFILTER
3112     show_help_children(sws_get_class(), AV_OPT_FLAG_ENCODING_PARAM);
3113 #endif
3114     printf("\nWhile playing:\n"
3115            "q, ESC              quit\n"
3116            "f                   toggle full screen\n"
3117            "p, SPC              pause\n"
3118            "a                   cycle audio channel\n"
3119            "v                   cycle video channel\n"
3120            "t                   cycle subtitle channel\n"
3121            "w                   show audio waves\n"
3122            "s                   activate frame-step mode\n"
3123            "left/right          seek backward/forward 10 seconds\n"
3124            "down/up             seek backward/forward 1 minute\n"
3125            "page down/page up   seek backward/forward 10 minutes\n"
3126            "mouse click         seek to percentage in file corresponding to fraction of width\n"
3127            );
3128     return 0;
3129 }
3130
3131 static int lockmgr(void **mtx, enum AVLockOp op)
3132 {
3133    switch(op) {
3134       case AV_LOCK_CREATE:
3135           *mtx = SDL_CreateMutex();
3136           if(!*mtx)
3137               return 1;
3138           return 0;
3139       case AV_LOCK_OBTAIN:
3140           return !!SDL_LockMutex(*mtx);
3141       case AV_LOCK_RELEASE:
3142           return !!SDL_UnlockMutex(*mtx);
3143       case AV_LOCK_DESTROY:
3144           SDL_DestroyMutex(*mtx);
3145           return 0;
3146    }
3147    return 1;
3148 }
3149
3150 /* Called from the main */
3151 int main(int argc, char **argv)
3152 {
3153     int flags;
3154     VideoState *is;
3155
3156     av_log_set_flags(AV_LOG_SKIP_REPEATED);
3157     parse_loglevel(argc, argv, options);
3158
3159     /* register all codecs, demux and protocols */
3160     avcodec_register_all();
3161 #if CONFIG_AVDEVICE
3162     avdevice_register_all();
3163 #endif
3164 #if CONFIG_AVFILTER
3165     avfilter_register_all();
3166 #endif
3167     av_register_all();
3168     avformat_network_init();
3169
3170     init_opts();
3171
3172     show_banner(argc, argv, options);
3173
3174     parse_options(NULL, argc, argv, options, opt_input_file);
3175
3176     if (!input_filename) {
3177         show_usage();
3178         fprintf(stderr, "An input file must be specified\n");
3179         fprintf(stderr, "Use -h to get full help or, even better, run 'man %s'\n", program_name);
3180         exit(1);
3181     }
3182
3183     if (display_disable) {
3184         video_disable = 1;
3185     }
3186     flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3187     if (audio_disable)
3188         flags &= ~SDL_INIT_AUDIO;
3189 #if !defined(__MINGW32__) && !defined(__APPLE__)
3190     flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3191 #endif
3192     if (SDL_Init (flags)) {
3193         fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3194         fprintf(stderr, "(Did you set the DISPLAY variable?)\n");
3195         exit(1);
3196     }
3197
3198     if (!display_disable) {
3199 #if HAVE_SDL_VIDEO_SIZE
3200         const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3201         fs_screen_width = vi->current_w;
3202         fs_screen_height = vi->current_h;
3203 #endif
3204     }
3205
3206     SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3207     SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3208     SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3209
3210     if (av_lockmgr_register(lockmgr)) {
3211         fprintf(stderr, "Could not initialize lock manager!\n");
3212         do_exit(NULL);
3213     }
3214
3215     av_init_packet(&flush_pkt);
3216     flush_pkt.data= "FLUSH";
3217
3218     is = stream_open(input_filename, file_iformat);
3219     if (!is) {
3220         fprintf(stderr, "Failed to initialize VideoState!\n");
3221         do_exit(NULL);
3222     }
3223
3224     event_loop(is);
3225
3226     /* never returns */
3227
3228     return 0;
3229 }