/*
- * copyright (c) 2007 Bobby Bingham
+ * Copyright (c) 2007 Bobby Bingham
*
- * This file is part of FFmpeg.
+ * This file is part of Libav.
*
- * FFmpeg is free software; you can redistribute it and/or
+ * Libav is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
- * FFmpeg is distributed in the hope that it will be useful,
+ * Libav is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
+ * License along with Libav; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
* video crop filter
*/
+#include <stdio.h>
+
#include "avfilter.h"
-#include "libavcore/imgutils.h"
+#include "formats.h"
+#include "internal.h"
+#include "video.h"
+#include "libavutil/eval.h"
+#include "libavutil/avstring.h"
+#include "libavutil/internal.h"
+#include "libavutil/libm.h"
+#include "libavutil/imgutils.h"
+#include "libavutil/mathematics.h"
+#include "libavutil/opt.h"
+
+static const char *const var_names[] = {
+ "E",
+ "PHI",
+ "PI",
+ "in_w", "iw", ///< width of the input video
+ "in_h", "ih", ///< height of the input video
+ "out_w", "ow", ///< width of the cropped video
+ "out_h", "oh", ///< height of the cropped video
+ "x",
+ "y",
+ "n", ///< number of frame
+ "pos", ///< position in the file
+ "t", ///< timestamp expressed in seconds
+ NULL
+};
+
+enum var_name {
+ VAR_E,
+ VAR_PHI,
+ VAR_PI,
+ VAR_IN_W, VAR_IW,
+ VAR_IN_H, VAR_IH,
+ VAR_OUT_W, VAR_OW,
+ VAR_OUT_H, VAR_OH,
+ VAR_X,
+ VAR_Y,
+ VAR_N,
+ VAR_T,
+ VAR_VARS_NB
+};
typedef struct {
+ const AVClass *class;
int x; ///< x offset of the non-cropped area with respect to the input area
int y; ///< y offset of the non-cropped area with respect to the input area
int w; ///< width of the cropped area
int max_step[4]; ///< max pixel step for each plane, expressed as a number of bytes
int hsub, vsub; ///< chroma subsampling
+ char *x_expr, *y_expr, *ow_expr, *oh_expr;
+ AVExpr *x_pexpr, *y_pexpr; /* parsed expressions for x and y */
+ double var_values[VAR_VARS_NB];
} CropContext;
static int query_formats(AVFilterContext *ctx)
{
- static const enum PixelFormat pix_fmts[] = {
- PIX_FMT_RGB48BE, PIX_FMT_RGB48LE,
- PIX_FMT_ARGB, PIX_FMT_RGBA,
- PIX_FMT_ABGR, PIX_FMT_BGRA,
- PIX_FMT_RGB24, PIX_FMT_BGR24,
- PIX_FMT_RGB565BE, PIX_FMT_RGB565LE,
- PIX_FMT_RGB555BE, PIX_FMT_RGB555LE,
- PIX_FMT_BGR565BE, PIX_FMT_BGR565LE,
- PIX_FMT_BGR555BE, PIX_FMT_BGR555LE,
- PIX_FMT_GRAY16BE, PIX_FMT_GRAY16LE,
- PIX_FMT_YUV420P16LE, PIX_FMT_YUV420P16BE,
- PIX_FMT_YUV422P16LE, PIX_FMT_YUV422P16BE,
- PIX_FMT_YUV444P16LE, PIX_FMT_YUV444P16BE,
- PIX_FMT_YUV444P, PIX_FMT_YUV422P,
- PIX_FMT_YUV420P, PIX_FMT_YUV411P,
- PIX_FMT_YUV410P, PIX_FMT_YUV440P,
- PIX_FMT_YUVJ444P, PIX_FMT_YUVJ422P,
- PIX_FMT_YUVJ420P, PIX_FMT_YUVJ440P,
- PIX_FMT_YUVA420P,
- PIX_FMT_RGB8, PIX_FMT_BGR8,
- PIX_FMT_RGB4_BYTE, PIX_FMT_BGR4_BYTE,
- PIX_FMT_PAL8, PIX_FMT_GRAY8,
- PIX_FMT_NONE
+ static const enum AVPixelFormat pix_fmts[] = {
+ AV_PIX_FMT_RGB48BE, AV_PIX_FMT_RGB48LE,
+ AV_PIX_FMT_BGR48BE, AV_PIX_FMT_BGR48LE,
+ AV_PIX_FMT_ARGB, AV_PIX_FMT_RGBA,
+ AV_PIX_FMT_ABGR, AV_PIX_FMT_BGRA,
+ AV_PIX_FMT_RGB24, AV_PIX_FMT_BGR24,
+ AV_PIX_FMT_RGB565BE, AV_PIX_FMT_RGB565LE,
+ AV_PIX_FMT_RGB555BE, AV_PIX_FMT_RGB555LE,
+ AV_PIX_FMT_BGR565BE, AV_PIX_FMT_BGR565LE,
+ AV_PIX_FMT_BGR555BE, AV_PIX_FMT_BGR555LE,
+ AV_PIX_FMT_GRAY16BE, AV_PIX_FMT_GRAY16LE,
+ AV_PIX_FMT_YUV420P16LE, AV_PIX_FMT_YUV420P16BE,
+ AV_PIX_FMT_YUV422P16LE, AV_PIX_FMT_YUV422P16BE,
+ AV_PIX_FMT_YUV444P16LE, AV_PIX_FMT_YUV444P16BE,
+ AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P,
+ AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P,
+ AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV440P,
+ AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ422P,
+ AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ440P,
+ AV_PIX_FMT_YUVA420P,
+ AV_PIX_FMT_RGB8, AV_PIX_FMT_BGR8,
+ AV_PIX_FMT_RGB4_BYTE, AV_PIX_FMT_BGR4_BYTE,
+ AV_PIX_FMT_PAL8, AV_PIX_FMT_GRAY8,
+ AV_PIX_FMT_NONE
};
- avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
+ ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
return 0;
}
-static av_cold int init(AVFilterContext *ctx, const char *args, void *opaque)
+static av_cold void uninit(AVFilterContext *ctx)
{
- CropContext *crop = ctx->priv;
+ CropContext *s = ctx->priv;
- if (args)
- sscanf(args, "%d:%d:%d:%d", &crop->x, &crop->y, &crop->w, &crop->h);
-
- return 0;
+ av_expr_free(s->x_pexpr);
+ s->x_pexpr = NULL;
+ av_expr_free(s->y_pexpr);
+ s->y_pexpr = NULL;
}
-static int config_input(AVFilterLink *link)
+static inline int normalize_double(int *n, double d)
{
- AVFilterContext *ctx = link->dst;
- CropContext *crop = ctx->priv;
- const AVPixFmtDescriptor *pix_desc = &av_pix_fmt_descriptors[link->format];
+ int ret = 0;
- av_image_fill_max_pixsteps(crop->max_step, NULL, pix_desc);
- crop->hsub = av_pix_fmt_descriptors[link->format].log2_chroma_w;
- crop->vsub = av_pix_fmt_descriptors[link->format].log2_chroma_h;
+ if (isnan(d)) {
+ ret = AVERROR(EINVAL);
+ } else if (d > INT_MAX || d < INT_MIN) {
+ *n = d > INT_MAX ? INT_MAX : INT_MIN;
+ ret = AVERROR(EINVAL);
+ } else
+ *n = round(d);
- if (crop->w == 0)
- crop->w = link->w - crop->x;
- if (crop->h == 0)
- crop->h = link->h - crop->y;
+ return ret;
+}
- crop->x &= ~((1 << crop->hsub) - 1);
- crop->y &= ~((1 << crop->vsub) - 1);
+static int config_input(AVFilterLink *link)
+{
+ AVFilterContext *ctx = link->dst;
+ CropContext *s = ctx->priv;
+ const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(link->format);
+ int ret;
+ const char *expr;
+ double res;
+
+ s->var_values[VAR_E] = M_E;
+ s->var_values[VAR_PHI] = M_PHI;
+ s->var_values[VAR_PI] = M_PI;
+ s->var_values[VAR_IN_W] = s->var_values[VAR_IW] = ctx->inputs[0]->w;
+ s->var_values[VAR_IN_H] = s->var_values[VAR_IH] = ctx->inputs[0]->h;
+ s->var_values[VAR_X] = NAN;
+ s->var_values[VAR_Y] = NAN;
+ s->var_values[VAR_OUT_W] = s->var_values[VAR_OW] = NAN;
+ s->var_values[VAR_OUT_H] = s->var_values[VAR_OH] = NAN;
+ s->var_values[VAR_N] = 0;
+ s->var_values[VAR_T] = NAN;
+
+ av_image_fill_max_pixsteps(s->max_step, NULL, pix_desc);
+ s->hsub = pix_desc->log2_chroma_w;
+ s->vsub = pix_desc->log2_chroma_h;
+
+ if ((ret = av_expr_parse_and_eval(&res, (expr = s->ow_expr),
+ var_names, s->var_values,
+ NULL, NULL, NULL, NULL, NULL, 0, ctx)) < 0)
+ goto fail_expr;
+ s->var_values[VAR_OUT_W] = s->var_values[VAR_OW] = res;
+ if ((ret = av_expr_parse_and_eval(&res, (expr = s->oh_expr),
+ var_names, s->var_values,
+ NULL, NULL, NULL, NULL, NULL, 0, ctx)) < 0)
+ goto fail_expr;
+ s->var_values[VAR_OUT_H] = s->var_values[VAR_OH] = res;
+ /* evaluate again ow as it may depend on oh */
+ if ((ret = av_expr_parse_and_eval(&res, (expr = s->ow_expr),
+ var_names, s->var_values,
+ NULL, NULL, NULL, NULL, NULL, 0, ctx)) < 0)
+ goto fail_expr;
+
+ s->var_values[VAR_OUT_W] = s->var_values[VAR_OW] = res;
+ if (normalize_double(&s->w, s->var_values[VAR_OUT_W]) < 0 ||
+ normalize_double(&s->h, s->var_values[VAR_OUT_H]) < 0) {
+ av_log(ctx, AV_LOG_ERROR,
+ "Too big value or invalid expression for out_w/ow or out_h/oh. "
+ "Maybe the expression for out_w:'%s' or for out_h:'%s' is self-referencing.\n",
+ s->ow_expr, s->oh_expr);
+ return AVERROR(EINVAL);
+ }
+ s->w &= ~((1 << s->hsub) - 1);
+ s->h &= ~((1 << s->vsub) - 1);
+
+ av_expr_free(s->x_pexpr);
+ av_expr_free(s->y_pexpr);
+ s->x_pexpr = s->y_pexpr = NULL;
+ if ((ret = av_expr_parse(&s->x_pexpr, s->x_expr, var_names,
+ NULL, NULL, NULL, NULL, 0, ctx)) < 0 ||
+ (ret = av_expr_parse(&s->y_pexpr, s->y_expr, var_names,
+ NULL, NULL, NULL, NULL, 0, ctx)) < 0)
+ return AVERROR(EINVAL);
- av_log(link->dst, AV_LOG_INFO, "x:%d y:%d w:%d h:%d\n",
- crop->x, crop->y, crop->w, crop->h);
+ av_log(ctx, AV_LOG_VERBOSE, "w:%d h:%d -> w:%d h:%d\n",
+ link->w, link->h, s->w, s->h);
- if (crop->x < 0 || crop->y < 0 ||
- crop->w <= 0 || crop->h <= 0 ||
- (unsigned)crop->x + (unsigned)crop->w > link->w ||
- (unsigned)crop->y + (unsigned)crop->h > link->h) {
+ if (s->w <= 0 || s->h <= 0 ||
+ s->w > link->w || s->h > link->h) {
av_log(ctx, AV_LOG_ERROR,
- "Output area %d:%d:%d:%d not within the input area 0:0:%d:%d or zero-sized\n",
- crop->x, crop->y, crop->w, crop->h, link->w, link->h);
+ "Invalid too big or non positive size for width '%d' or height '%d'\n",
+ s->w, s->h);
return AVERROR(EINVAL);
}
+ /* set default, required in the case the first computed value for x/y is NAN */
+ s->x = (link->w - s->w) / 2;
+ s->y = (link->h - s->h) / 2;
+ s->x &= ~((1 << s->hsub) - 1);
+ s->y &= ~((1 << s->vsub) - 1);
return 0;
+
+fail_expr:
+ av_log(NULL, AV_LOG_ERROR, "Error when evaluating the expression '%s'\n", expr);
+ return ret;
}
static int config_output(AVFilterLink *link)
{
- CropContext *crop = link->src->priv;
+ CropContext *s = link->src->priv;
- link->w = crop->w;
- link->h = crop->h;
+ link->w = s->w;
+ link->h = s->h;
return 0;
}
-static void start_frame(AVFilterLink *link, AVFilterBufferRef *picref)
+static int filter_frame(AVFilterLink *link, AVFrame *frame)
{
- CropContext *crop = link->dst->priv;
- AVFilterBufferRef *ref2 = avfilter_ref_buffer(picref, ~0);
+ AVFilterContext *ctx = link->dst;
+ CropContext *s = ctx->priv;
+ const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(link->format);
int i;
- picref->video->w = crop->w;
- picref->video->h = crop->h;
-
- ref2->data[0] += crop->y * ref2->linesize[0];
- ref2->data[0] += crop->x * crop->max_step[0];
-
- if (!(av_pix_fmt_descriptors[link->format].flags & PIX_FMT_PAL)) {
+ frame->width = s->w;
+ frame->height = s->h;
+
+ s->var_values[VAR_T] = frame->pts == AV_NOPTS_VALUE ?
+ NAN : frame->pts * av_q2d(link->time_base);
+ s->var_values[VAR_X] = av_expr_eval(s->x_pexpr, s->var_values, NULL);
+ s->var_values[VAR_Y] = av_expr_eval(s->y_pexpr, s->var_values, NULL);
+ s->var_values[VAR_X] = av_expr_eval(s->x_pexpr, s->var_values, NULL);
+
+ normalize_double(&s->x, s->var_values[VAR_X]);
+ normalize_double(&s->y, s->var_values[VAR_Y]);
+
+ if (s->x < 0)
+ s->x = 0;
+ if (s->y < 0)
+ s->y = 0;
+ if ((unsigned)s->x + (unsigned)s->w > link->w)
+ s->x = link->w - s->w;
+ if ((unsigned)s->y + (unsigned)s->h > link->h)
+ s->y = link->h - s->h;
+ s->x &= ~((1 << s->hsub) - 1);
+ s->y &= ~((1 << s->vsub) - 1);
+
+ av_dlog(ctx, "n:%d t:%f x:%d y:%d x+w:%d y+h:%d\n",
+ (int)s->var_values[VAR_N], s->var_values[VAR_T], s->x,
+ s->y, s->x+s->w, s->y+s->h);
+
+ frame->data[0] += s->y * frame->linesize[0];
+ frame->data[0] += s->x * s->max_step[0];
+
+ if (!(desc->flags & AV_PIX_FMT_FLAG_PAL || desc->flags & AV_PIX_FMT_FLAG_PSEUDOPAL)) {
for (i = 1; i < 3; i ++) {
- if (ref2->data[i]) {
- ref2->data[i] += (crop->y >> crop->vsub) * ref2->linesize[i];
- ref2->data[i] += (crop->x * crop->max_step[i]) >> crop->hsub;
+ if (frame->data[i]) {
+ frame->data[i] += (s->y >> s->vsub) * frame->linesize[i];
+ frame->data[i] += (s->x * s->max_step[i]) >> s->hsub;
}
}
}
/* alpha plane */
- if (ref2->data[3]) {
- ref2->data[3] += crop->y * ref2->linesize[3];
- ref2->data[3] += crop->x * crop->max_step[3];
+ if (frame->data[3]) {
+ frame->data[3] += s->y * frame->linesize[3];
+ frame->data[3] += s->x * s->max_step[3];
}
- avfilter_start_frame(link->dst->outputs[0], ref2);
+ s->var_values[VAR_N] += 1.0;
+
+ return ff_filter_frame(link->dst->outputs[0], frame);
}
-static void draw_slice(AVFilterLink *link, int y, int h, int slice_dir)
-{
- AVFilterContext *ctx = link->dst;
- CropContext *crop = ctx->priv;
+#define OFFSET(x) offsetof(CropContext, x)
+#define FLAGS AV_OPT_FLAG_VIDEO_PARAM
+static const AVOption options[] = {
+ { "out_w", "Output video width", OFFSET(ow_expr), AV_OPT_TYPE_STRING, { .str = "iw" }, .flags = FLAGS },
+ { "out_h", "Output video height", OFFSET(oh_expr), AV_OPT_TYPE_STRING, { .str = "ih" }, .flags = FLAGS },
+ { "x", "Horizontal position in the input video of the left edge of the cropped output video",
+ OFFSET(x_expr), AV_OPT_TYPE_STRING, { .str = "(in_w - out_w) / 2" }, .flags = FLAGS },
+ { "y", "Vertical position in the input video of the top edge of the cropped output video",
+ OFFSET(y_expr), AV_OPT_TYPE_STRING, { .str = "(in_h - out_h) / 2" }, .flags = FLAGS },
+ { NULL },
+};
- if (y >= crop->y + crop->h || y + h <= crop->y)
- return;
+static const AVClass crop_class = {
+ .class_name = "crop",
+ .item_name = av_default_item_name,
+ .option = options,
+ .version = LIBAVUTIL_VERSION_INT,
+};
- if (y < crop->y) {
- h -= crop->y - y;
- y = crop->y;
- }
- if (y + h > crop->y + crop->h)
- h = crop->y + crop->h - y;
+static const AVFilterPad avfilter_vf_crop_inputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .filter_frame = filter_frame,
+ .get_video_buffer = ff_null_get_video_buffer,
+ .config_props = config_input,
+ },
+ { NULL }
+};
- avfilter_draw_slice(ctx->outputs[0], y - crop->y, h, slice_dir);
-}
+static const AVFilterPad avfilter_vf_crop_outputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .config_props = config_output,
+ },
+ { NULL }
+};
AVFilter avfilter_vf_crop = {
.name = "crop",
- .description = NULL_IF_CONFIG_SMALL("Crop the input video to x:y:width:height."),
+ .description = NULL_IF_CONFIG_SMALL("Crop the input video to width:height:x:y."),
.priv_size = sizeof(CropContext),
+ .priv_class = &crop_class,
.query_formats = query_formats,
- .init = init,
-
- .inputs = (AVFilterPad[]) {{ .name = "default",
- .type = AVMEDIA_TYPE_VIDEO,
- .start_frame = start_frame,
- .draw_slice = draw_slice,
- .get_video_buffer = avfilter_null_get_video_buffer,
- .config_props = config_input, },
- { .name = NULL}},
- .outputs = (AVFilterPad[]) {{ .name = "default",
- .type = AVMEDIA_TYPE_VIDEO,
- .config_props = config_output, },
- { .name = NULL}},
+ .uninit = uninit,
+
+ .inputs = avfilter_vf_crop_inputs,
+ .outputs = avfilter_vf_crop_outputs,
};