X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=libavfilter%2Favf_concat.c;h=df6414704d05d8760773bb54a0ca85d1a4679f65;hb=0a071f7124beaf0929f772a8618ac1b6c17b0222;hp=46bd42359b301343d7b4820eb5a496d24ccfdad4;hpb=52b44e9d15c0ee3c118ed68a0c2c737a9eb50ae9;p=ffmpeg diff --git a/libavfilter/avf_concat.c b/libavfilter/avf_concat.c index 46bd42359b3..df6414704d0 100644 --- a/libavfilter/avf_concat.c +++ b/libavfilter/avf_concat.c @@ -28,8 +28,7 @@ #include "libavutil/channel_layout.h" #include "libavutil/opt.h" #include "avfilter.h" -#define FF_BUFQUEUE_SIZE 256 -#include "bufferqueue.h" +#include "filters.h" #include "internal.h" #include "video.h" #include "audio.h" @@ -48,7 +47,6 @@ typedef struct ConcatContext { int64_t pts; int64_t nb_frames; unsigned eof; - struct FFBufQueue queue; } *in; } ConcatContext; @@ -89,25 +87,25 @@ static int query_formats(AVFilterContext *ctx) /* Set the output formats */ formats = ff_all_formats(type); - if ((ret = ff_formats_ref(formats, &ctx->outputs[idx]->in_formats)) < 0) + if ((ret = ff_formats_ref(formats, &ctx->outputs[idx]->incfg.formats)) < 0) return ret; if (type == AVMEDIA_TYPE_AUDIO) { rates = ff_all_samplerates(); - if ((ret = ff_formats_ref(rates, &ctx->outputs[idx]->in_samplerates)) < 0) + if ((ret = ff_formats_ref(rates, &ctx->outputs[idx]->incfg.samplerates)) < 0) return ret; layouts = ff_all_channel_layouts(); - if ((ret = ff_channel_layouts_ref(layouts, &ctx->outputs[idx]->in_channel_layouts)) < 0) + if ((ret = ff_channel_layouts_ref(layouts, &ctx->outputs[idx]->incfg.channel_layouts)) < 0) return ret; } /* Set the same formats for each corresponding input */ for (seg = 0; seg < cat->nb_segments; seg++) { - if ((ret = ff_formats_ref(formats, &ctx->inputs[idx]->out_formats)) < 0) + if ((ret = ff_formats_ref(formats, &ctx->inputs[idx]->outcfg.formats)) < 0) return ret; if (type == AVMEDIA_TYPE_AUDIO) { - if ((ret = ff_formats_ref(rates, &ctx->inputs[idx]->out_samplerates)) < 0 || - (ret = ff_channel_layouts_ref(layouts, &ctx->inputs[idx]->out_channel_layouts)) < 0) + if ((ret = ff_formats_ref(rates, &ctx->inputs[idx]->outcfg.samplerates)) < 0 || + (ret = ff_channel_layouts_ref(layouts, &ctx->inputs[idx]->outcfg.channel_layouts)) < 0) return ret; } idx += ctx->nb_outputs; @@ -133,8 +131,21 @@ static int config_output(AVFilterLink *outlink) outlink->h = inlink->h; outlink->sample_aspect_ratio = inlink->sample_aspect_ratio; outlink->format = inlink->format; + outlink->frame_rate = inlink->frame_rate; + + for (seg = 1; seg < cat->nb_segments; seg++) { + inlink = ctx->inputs[in_no + seg * ctx->nb_outputs]; + if (outlink->frame_rate.num != inlink->frame_rate.num || + outlink->frame_rate.den != inlink->frame_rate.den) { + av_log(ctx, AV_LOG_VERBOSE, + "Video inputs have different frame rates, output will be VFR\n"); + outlink->frame_rate = av_make_q(1, 0); + break; + } + } + for (seg = 1; seg < cat->nb_segments; seg++) { - inlink = ctx->inputs[in_no += ctx->nb_outputs]; + inlink = ctx->inputs[in_no + seg * ctx->nb_outputs]; if (!outlink->sample_aspect_ratio.num) outlink->sample_aspect_ratio = inlink->sample_aspect_ratio; /* possible enhancement: unsafe mode, do not check */ @@ -185,24 +196,6 @@ static int push_frame(AVFilterContext *ctx, unsigned in_no, AVFrame *buf) return ff_filter_frame(outlink, buf); } -static int process_frame(AVFilterLink *inlink, AVFrame *buf) -{ - AVFilterContext *ctx = inlink->dst; - ConcatContext *cat = ctx->priv; - unsigned in_no = FF_INLINK_IDX(inlink); - - if (in_no < cat->cur_idx) { - av_log(ctx, AV_LOG_ERROR, "Frame after EOF on input %s\n", - ctx->input_pads[in_no].name); - av_frame_free(&buf); - } else if (in_no >= cat->cur_idx + ctx->nb_outputs) { - ff_bufqueue_add(ctx, &cat->in[in_no].queue, buf); - } else { - return push_frame(ctx, in_no, buf); - } - return 0; -} - static AVFrame *get_video_buffer(AVFilterLink *inlink, int w, int h) { AVFilterContext *ctx = inlink->dst; @@ -221,11 +214,6 @@ static AVFrame *get_audio_buffer(AVFilterLink *inlink, int nb_samples) return ff_get_audio_buffer(outlink, nb_samples); } -static int filter_frame(AVFilterLink *inlink, AVFrame *buf) -{ - return process_frame(inlink, buf); -} - static void close_input(AVFilterContext *ctx, unsigned in_no) { ConcatContext *cat = ctx->priv; @@ -263,6 +251,10 @@ static int send_silence(AVFilterContext *ctx, unsigned in_no, unsigned out_no, if (!rate_tb.den) return AVERROR_BUG; + if (cat->in[in_no].pts < INT64_MIN + seg_delta) + return AVERROR_INVALIDDATA; + if (seg_delta < cat->in[in_no].pts) + return AVERROR_INVALIDDATA; nb_samples = av_rescale_q(seg_delta - cat->in[in_no].pts, outlink->time_base, rate_tb); frame_nb_samples = FFMAX(9600, rate_tb.den / 5); /* arbitrary */ @@ -306,57 +298,10 @@ static int flush_segment(AVFilterContext *ctx) if (ret < 0) return ret; } - /* flush queued buffers */ - /* possible enhancement: flush in PTS order */ - str_max = cat->cur_idx + ctx->nb_outputs; - for (str = cat->cur_idx; str < str_max; str++) { - while (cat->in[str].queue.available) { - ret = push_frame(ctx, str, ff_bufqueue_get(&cat->in[str].queue)); - if (ret < 0) - return ret; - } - } } return 0; } -static int request_frame(AVFilterLink *outlink) -{ - AVFilterContext *ctx = outlink->src; - ConcatContext *cat = ctx->priv; - unsigned out_no = FF_OUTLINK_IDX(outlink); - unsigned in_no = out_no + cat->cur_idx; - unsigned str, str_max; - int ret; - - while (1) { - if (in_no >= ctx->nb_inputs) - return AVERROR_EOF; - if (!cat->in[in_no].eof) { - ret = ff_request_frame(ctx->inputs[in_no]); - if (ret != AVERROR_EOF) - return ret; - close_input(ctx, in_no); - } - /* cycle on all inputs to finish the segment */ - /* possible enhancement: request in PTS order */ - str_max = cat->cur_idx + ctx->nb_outputs - 1; - for (str = cat->cur_idx; cat->nb_in_active; - str = str == str_max ? cat->cur_idx : str + 1) { - if (cat->in[str].eof) - continue; - ret = ff_request_frame(ctx->inputs[str]); - if (ret != AVERROR_EOF) - return ret; - close_input(ctx, str); - } - ret = flush_segment(ctx); - if (ret < 0) - return ret; - in_no += ctx->nb_outputs; - } -} - static av_cold int init(AVFilterContext *ctx) { ConcatContext *cat = ctx->priv; @@ -371,7 +316,6 @@ static av_cold int init(AVFilterContext *ctx) .type = type, .get_video_buffer = get_video_buffer, .get_audio_buffer = get_audio_buffer, - .filter_frame = filter_frame, }; pad.name = av_asprintf("in%d:%c%d", seg, "va"[type], str); if ((ret = ff_insert_inpad(ctx, ctx->nb_inputs, &pad)) < 0) { @@ -387,7 +331,6 @@ static av_cold int init(AVFilterContext *ctx) AVFilterPad pad = { .type = type, .config_props = config_output, - .request_frame = request_frame, }; pad.name = av_asprintf("out:%c%d", "va"[type], str); if ((ret = ff_insert_outpad(ctx, ctx->nb_outputs, &pad)) < 0) { @@ -409,15 +352,88 @@ static av_cold void uninit(AVFilterContext *ctx) ConcatContext *cat = ctx->priv; unsigned i; - for (i = 0; i < ctx->nb_inputs; i++) { + for (i = 0; i < ctx->nb_inputs; i++) av_freep(&ctx->input_pads[i].name); - ff_bufqueue_discard_all(&cat->in[i].queue); - } for (i = 0; i < ctx->nb_outputs; i++) av_freep(&ctx->output_pads[i].name); av_freep(&cat->in); } +static int activate(AVFilterContext *ctx) +{ + ConcatContext *cat = ctx->priv; + AVFrame *frame; + unsigned i, j; + int ret, status; + int64_t pts; + + /* Forward status back */ + for (i = 0; i < ctx->nb_outputs; i++) { + status = ff_outlink_get_status(ctx->outputs[i]); + if (!status) + continue; + for (j = i; j < ctx->nb_inputs; j += ctx->nb_outputs) { + if (!cat->in[j].eof) { + cat->in[j].eof = 1; + ff_inlink_set_status(ctx->inputs[j], status); + return 0; + } + } + + } + + /* Forward available frames */ + if (cat->cur_idx < ctx->nb_inputs) { + for (i = 0; i < ctx->nb_outputs; i++) { + ret = ff_inlink_consume_frame(ctx->inputs[cat->cur_idx + i], &frame); + if (ret < 0) + return ret; + if (ret) { + ff_filter_set_ready(ctx, 10); + return push_frame(ctx, cat->cur_idx + i, frame); + } + } + } + + /* Forward status change */ + if (cat->cur_idx < ctx->nb_inputs) { + for (i = 0; i < ctx->nb_outputs; i++) { + ret = ff_inlink_acknowledge_status(ctx->inputs[cat->cur_idx + i], &status, &pts); + /* TODO use pts */ + if (ret > 0) { + close_input(ctx, cat->cur_idx + i); + if (cat->cur_idx + ctx->nb_outputs >= ctx->nb_inputs) { + ff_outlink_set_status(ctx->outputs[i], status, pts); + } + if (!cat->nb_in_active) { + ret = flush_segment(ctx); + if (ret < 0) + return ret; + } + ff_filter_set_ready(ctx, 10); + return 0; + } + } + } + + ret = FFERROR_NOT_READY; + for (i = 0; i < ctx->nb_outputs; i++) { + if (ff_outlink_frame_wanted(ctx->outputs[i])) { + if (cat->in[cat->cur_idx + i].eof) { + for (j = 0; j < ctx->nb_outputs; j++) + if (!cat->in[cat->cur_idx + j].eof) + ff_inlink_request_frame(ctx->inputs[cat->cur_idx + j]); + return 0; + } else { + ff_inlink_request_frame(ctx->inputs[cat->cur_idx + i]); + ret = 0; + } + } + } + + return ret; +} + static int process_command(AVFilterContext *ctx, const char *cmd, const char *args, char *res, int res_len, int flags) { @@ -437,6 +453,7 @@ AVFilter ff_avf_concat = { .init = init, .uninit = uninit, .query_formats = query_formats, + .activate = activate, .priv_size = sizeof(ConcatContext), .inputs = NULL, .outputs = NULL,