#include "libavutil/avstring.h"
#include "libavutil/colorspace.h"
#include "libavutil/pixdesc.h"
+#include "libavcore/parseutils.h"
#include "libavformat/avformat.h"
#include "libavdevice/avdevice.h"
#include "libswscale/swscale.h"
enum PixelFormat pix_fmt;
#if CONFIG_AVFILTER
- AVFilterPicRef *picref;
+ AVFilterBufferRef *picref;
#endif
} VideoPicture;
static void free_subpicture(SubPicture *sp)
{
- int i;
-
- for (i = 0; i < sp->sub.num_rects; i++)
- {
- av_freep(&sp->sub.rects[i]->pict.data[0]);
- av_freep(&sp->sub.rects[i]->pict.data[1]);
- av_freep(&sp->sub.rects[i]);
- }
-
- av_free(sp->sub.rects);
-
- memset(&sp->sub, 0, sizeof(AVSubtitle));
+ avsubtitle_free(&sp->sub);
}
static void video_image_display(VideoState *is)
vp = &is->pictq[is->pictq_rindex];
if (vp->bmp) {
#if CONFIG_AVFILTER
- if (vp->picref->pixel_aspect.num == 0)
+ if (vp->picref->video->pixel_aspect.num == 0)
aspect_ratio = 0;
else
- aspect_ratio = av_q2d(vp->picref->pixel_aspect);
+ aspect_ratio = av_q2d(vp->picref->video->pixel_aspect);
#else
/* XXX: use variable in the frame */
for(y=0; y<s->height; y++){
double w= 1/sqrt(nb_freq);
int a= sqrt(w*sqrt(data[0][2*y+0]*data[0][2*y+0] + data[0][2*y+1]*data[0][2*y+1]));
- int b= sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0] + data[1][2*y+1]*data[1][2*y+1]));
+ int b= (nb_display_channels == 2 ) ? sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0]
+ + data[1][2*y+1]*data[1][2*y+1])) : a;
a= FFMIN(a,255);
b= FFMIN(b,255);
fgcolor = SDL_MapRGB(screen->format, a, b, (a+b)/2);
#if CONFIG_AVFILTER
if (vp->picref)
- avfilter_unref_pic(vp->picref);
+ avfilter_unref_buffer(vp->picref);
vp->picref = NULL;
vp->width = is->out_video_filter->inputs[0]->w;
AVPicture pict;
#if CONFIG_AVFILTER
if(vp->picref)
- avfilter_unref_pic(vp->picref);
+ avfilter_unref_buffer(vp->picref);
vp->picref = src_frame->opaque;
#endif
static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
{
AVFilterContext *ctx = codec->opaque;
- AVFilterPicRef *ref;
+ AVFilterBufferRef *ref;
int perms = AV_PERM_WRITE;
int i, w, h, stride[4];
unsigned edge;
if(!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
return -1;
- ref->w = codec->width;
- ref->h = codec->height;
+ ref->video->w = codec->width;
+ ref->video->h = codec->height;
for(i = 0; i < 4; i ++) {
- unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->pic->format].log2_chroma_w : 0;
- unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->pic->format].log2_chroma_h : 0;
+ unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_w : 0;
+ unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_h : 0;
if (ref->data[i]) {
ref->data[i] += (edge >> hshift) + ((edge * ref->linesize[i]) >> vshift);
static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
{
memset(pic->data, 0, sizeof(pic->data));
- avfilter_unref_pic(pic->opaque);
+ avfilter_unref_buffer(pic->opaque);
}
static int input_reget_buffer(AVCodecContext *codec, AVFrame *pic)
{
- AVFilterPicRef *ref = pic->opaque;
+ AVFilterBufferRef *ref = pic->opaque;
if (pic->data[0] == NULL) {
pic->buffer_hints |= FF_BUFFER_HINTS_READABLE;
return codec->get_buffer(codec, pic);
}
- if ((codec->width != ref->w) || (codec->height != ref->h) ||
- (codec->pix_fmt != ref->pic->format)) {
+ if ((codec->width != ref->video->w) || (codec->height != ref->video->h) ||
+ (codec->pix_fmt != ref->format)) {
av_log(codec, AV_LOG_ERROR, "Picture properties changed.\n");
return -1;
}
static int input_request_frame(AVFilterLink *link)
{
FilterPriv *priv = link->src->priv;
- AVFilterPicRef *picref;
+ AVFilterBufferRef *picref;
int64_t pts = 0;
AVPacket pkt;
int ret;
return -1;
if(priv->use_dr1) {
- picref = avfilter_ref_pic(priv->frame->opaque, ~0);
+ picref = avfilter_ref_buffer(priv->frame->opaque, ~0);
} else {
picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
- av_picture_copy((AVPicture *)&picref->data, (AVPicture *)priv->frame,
- picref->pic->format, link->w, link->h);
+ av_picture_data_copy(picref->data, picref->linesize,
+ priv->frame->data, priv->frame->linesize,
+ picref->format, link->w, link->h);
}
av_free_packet(&pkt);
picref->pts = pts;
picref->pos = pkt.pos;
- picref->pixel_aspect = priv->is->video_st->codec->sample_aspect_ratio;
+ picref->video->pixel_aspect = priv->is->video_st->codec->sample_aspect_ratio;
avfilter_start_frame(link, picref);
avfilter_draw_slice(link, 0, link->h, 1);
avfilter_end_frame(link);
static int get_filtered_video_frame(AVFilterContext *ctx, AVFrame *frame,
int64_t *pts, int64_t *pos)
{
- AVFilterPicRef *pic;
+ AVFilterBufferRef *pic;
if(avfilter_request_frame(ctx->inputs[0]))
return -1;
- if(!(pic = ctx->inputs[0]->cur_pic))
+ if(!(pic = ctx->inputs[0]->cur_buf))
return -1;
- ctx->inputs[0]->cur_pic = NULL;
+ ctx->inputs[0]->cur_buf = NULL;
frame->opaque = pic;
*pts = pic->pts;
snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
graph->scale_sws_opts = av_strdup(sws_flags_str);
- if(!(filt_src = avfilter_open(&input_filter, "src"))) goto the_end;
- if(!(filt_out = avfilter_open(&output_filter, "out"))) goto the_end;
+ if (avfilter_open(&filt_src, &input_filter, "src") < 0) goto the_end;
+ if (avfilter_open(&filt_out, &output_filter, "out") < 0) goto the_end;
if(avfilter_init_filter(filt_src, NULL, is)) goto the_end;
if(avfilter_init_filter(filt_out, NULL, frame)) goto the_end;
vp = &is->pictq[i];
#if CONFIG_AVFILTER
if (vp->picref) {
- avfilter_unref_pic(vp->picref);
+ avfilter_unref_buffer(vp->picref);
vp->picref = NULL;
}
#endif
static void opt_frame_size(const char *arg)
{
- if (av_parse_video_frame_size(&frame_width, &frame_height, arg) < 0) {
+ if (av_parse_video_size(&frame_width, &frame_height, arg) < 0) {
fprintf(stderr, "Incorrect frame size\n");
exit(1);
}