* fade audio filter
*/
-#include "libavutil/audio_fifo.h"
#include "libavutil/opt.h"
#include "audio.h"
#include "avfilter.h"
+#include "filters.h"
#include "internal.h"
typedef struct AudioFadeContext {
int overlap;
int cf0_eof;
int crossfade_is_over;
- AVAudioFifo *fifo[2];
int64_t pts;
void (*fade_samples)(uint8_t **dst, uint8_t * const *src,
int curve0, int curve1);
} AudioFadeContext;
-enum CurveType { TRI, QSIN, ESIN, HSIN, LOG, IPAR, QUA, CUB, SQU, CBR, PAR, EXP, IQSIN, IHSIN, DESE, DESI, NB_CURVES };
+enum CurveType { TRI, QSIN, ESIN, HSIN, LOG, IPAR, QUA, CUB, SQU, CBR, PAR, EXP, IQSIN, IHSIN, DESE, DESI, LOSI, NONE, NB_CURVES };
#define OFFSET(x) offsetof(AudioFadeContext, x)
#define FLAGS AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
case DESI:
gain = gain <= 0.5 ? CUBE(2 * gain) / 2: 1 - CUBE(2 * (1 - gain)) / 2;
break;
+ case LOSI: {
+ const double a = 1. / (1. - 0.787) - 1;
+ double A = 1. / (1.0 + exp(0 -((gain-0.5) * a * 2.0)));
+ double B = 1. / (1.0 + exp(a));
+ double C = 1. / (1.0 + exp(0-a));
+ gain = (A - B) / (C - B);
+ }
+ break;
+ case NONE:
+ gain = 1.0;
+ break;
}
return gain;
{ "ns", "set number of samples for fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT64, {.i64 = 44100}, 1, INT64_MAX, FLAGS },
{ "start_time", "set time to start fading", OFFSET(start_time), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
{ "st", "set time to start fading", OFFSET(start_time), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
- { "duration", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
- { "d", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
+ { "duration", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
+ { "d", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
{ "curve", "set fade curve type", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
{ "c", "set fade curve type", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
{ "tri", "linear slope", 0, AV_OPT_TYPE_CONST, {.i64 = TRI }, 0, 0, FLAGS, "curve" },
{ "ihsin", "inverted half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IHSIN}, 0, 0, FLAGS, "curve" },
{ "dese", "double-exponential seat", 0, AV_OPT_TYPE_CONST, {.i64 = DESE }, 0, 0, FLAGS, "curve" },
{ "desi", "double-exponential sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = DESI }, 0, 0, FLAGS, "curve" },
+ { "losi", "logistic sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = LOSI }, 0, 0, FLAGS, "curve" },
+ { "nofade", "no fade; keep audio as-is", 0, AV_OPT_TYPE_CONST, {.i64 = NONE }, 0, 0, FLAGS, "curve" },
{ NULL }
};
static const AVOption acrossfade_options[] = {
{ "nb_samples", "set number of samples for cross fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 44100}, 1, INT32_MAX/10, FLAGS },
{ "ns", "set number of samples for cross fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 44100}, 1, INT32_MAX/10, FLAGS },
- { "duration", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60, FLAGS },
- { "d", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60, FLAGS },
+ { "duration", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60000000, FLAGS },
+ { "d", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60000000, FLAGS },
{ "overlap", "overlap 1st stream end with 2nd stream start", OFFSET(overlap), AV_OPT_TYPE_BOOL, {.i64 = 1 }, 0, 1, FLAGS },
{ "o", "overlap 1st stream end with 2nd stream start", OFFSET(overlap), AV_OPT_TYPE_BOOL, {.i64 = 1 }, 0, 1, FLAGS },
{ "curve1", "set fade curve type for 1st stream", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
{ "ihsin", "inverted half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IHSIN}, 0, 0, FLAGS, "curve" },
{ "dese", "double-exponential seat", 0, AV_OPT_TYPE_CONST, {.i64 = DESE }, 0, 0, FLAGS, "curve" },
{ "desi", "double-exponential sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = DESI }, 0, 0, FLAGS, "curve" },
+ { "losi", "logistic sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = LOSI }, 0, 0, FLAGS, "curve" },
+ { "nofade", "no fade; keep audio as-is", 0, AV_OPT_TYPE_CONST, {.i64 = NONE }, 0, 0, FLAGS, "curve" },
{ "curve2", "set fade curve type for 2nd stream", OFFSET(curve2), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
{ "c2", "set fade curve type for 2nd stream", OFFSET(curve2), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
{ NULL }
CROSSFADE(s16, int16_t)
CROSSFADE(s32, int32_t)
-static int acrossfade_filter_frame(AVFilterLink *inlink, AVFrame *in)
+static int activate(AVFilterContext *ctx)
{
- AVFilterContext *ctx = inlink->dst;
AudioFadeContext *s = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
- AVFrame *out, *cf[2] = { NULL };
- int ret = 0, nb_samples;
+ AVFrame *in = NULL, *out, *cf[2] = { NULL };
+ int ret = 0, nb_samples, status;
+ int64_t pts;
- if (s->crossfade_is_over) {
- in->pts = s->pts;
- s->pts += av_rescale_q(in->nb_samples,
- (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
- return ff_filter_frame(outlink, in);
- } else if (inlink == ctx->inputs[0]) {
- av_audio_fifo_write(s->fifo[0], (void **)in->extended_data, in->nb_samples);
+ FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx);
- nb_samples = av_audio_fifo_size(s->fifo[0]) - s->nb_samples;
- if (nb_samples > 0) {
- out = ff_get_audio_buffer(outlink, nb_samples);
- if (!out) {
- ret = AVERROR(ENOMEM);
- goto fail;
+ if (s->crossfade_is_over) {
+ ret = ff_inlink_consume_frame(ctx->inputs[1], &in);
+ if (ret > 0) {
+ in->pts = s->pts;
+ s->pts += av_rescale_q(in->nb_samples,
+ (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+ return ff_filter_frame(outlink, in);
+ } else if (ret < 0) {
+ return ret;
+ } else if (ff_inlink_acknowledge_status(ctx->inputs[1], &status, &pts)) {
+ ff_outlink_set_status(ctx->outputs[0], status, pts);
+ return 0;
+ } else if (!ret) {
+ if (ff_outlink_frame_wanted(ctx->outputs[0])) {
+ ff_inlink_request_frame(ctx->inputs[1]);
+ return 0;
}
- av_audio_fifo_read(s->fifo[0], (void **)out->extended_data, nb_samples);
- out->pts = s->pts;
- s->pts += av_rescale_q(nb_samples,
- (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
- ret = ff_filter_frame(outlink, out);
}
- } else if (av_audio_fifo_size(s->fifo[1]) < s->nb_samples) {
- if (!s->overlap && av_audio_fifo_size(s->fifo[0]) > 0) {
- nb_samples = av_audio_fifo_size(s->fifo[0]);
-
- cf[0] = ff_get_audio_buffer(outlink, nb_samples);
- out = ff_get_audio_buffer(outlink, nb_samples);
- if (!out || !cf[0]) {
- ret = AVERROR(ENOMEM);
- goto fail;
- }
- av_audio_fifo_read(s->fifo[0], (void **)cf[0]->extended_data, nb_samples);
+ }
- s->fade_samples(out->extended_data, cf[0]->extended_data, nb_samples,
- outlink->channels, -1, nb_samples - 1, nb_samples, s->curve);
- out->pts = s->pts;
- s->pts += av_rescale_q(nb_samples,
- (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
- ret = ff_filter_frame(outlink, out);
- if (ret < 0)
- goto fail;
+ if (ff_inlink_queued_samples(ctx->inputs[0]) > s->nb_samples) {
+ nb_samples = ff_inlink_queued_samples(ctx->inputs[0]) - s->nb_samples;
+ if (nb_samples > 0) {
+ ret = ff_inlink_consume_samples(ctx->inputs[0], nb_samples, nb_samples, &in);
+ if (ret < 0) {
+ return ret;
+ }
}
-
- av_audio_fifo_write(s->fifo[1], (void **)in->extended_data, in->nb_samples);
- } else if (av_audio_fifo_size(s->fifo[1]) >= s->nb_samples) {
- av_audio_fifo_write(s->fifo[1], (void **)in->extended_data, in->nb_samples);
-
+ in->pts = s->pts;
+ s->pts += av_rescale_q(in->nb_samples,
+ (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+ return ff_filter_frame(outlink, in);
+ } else if (ff_inlink_queued_samples(ctx->inputs[0]) >= s->nb_samples &&
+ ff_inlink_queued_samples(ctx->inputs[1]) >= s->nb_samples && s->cf0_eof) {
if (s->overlap) {
- cf[0] = ff_get_audio_buffer(outlink, s->nb_samples);
- cf[1] = ff_get_audio_buffer(outlink, s->nb_samples);
out = ff_get_audio_buffer(outlink, s->nb_samples);
- if (!out || !cf[0] || !cf[1]) {
+ if (!out)
+ return AVERROR(ENOMEM);
+
+ ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
+ if (ret < 0) {
av_frame_free(&out);
- ret = AVERROR(ENOMEM);
- goto fail;
+ return ret;
}
- av_audio_fifo_read(s->fifo[0], (void **)cf[0]->extended_data, s->nb_samples);
- av_audio_fifo_read(s->fifo[1], (void **)cf[1]->extended_data, s->nb_samples);
+ ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
+ if (ret < 0) {
+ av_frame_free(&out);
+ return ret;
+ }
s->crossfade_samples(out->extended_data, cf[0]->extended_data,
cf[1]->extended_data,
- s->nb_samples, in->channels,
+ s->nb_samples, out->channels,
s->curve, s->curve2);
out->pts = s->pts;
s->pts += av_rescale_q(s->nb_samples,
(AVRational){ 1, outlink->sample_rate }, outlink->time_base);
- ret = ff_filter_frame(outlink, out);
- if (ret < 0)
- goto fail;
+ s->crossfade_is_over = 1;
+ av_frame_free(&cf[0]);
+ av_frame_free(&cf[1]);
+ return ff_filter_frame(outlink, out);
} else {
out = ff_get_audio_buffer(outlink, s->nb_samples);
- cf[1] = ff_get_audio_buffer(outlink, s->nb_samples);
- if (!out || !cf[1]) {
- ret = AVERROR(ENOMEM);
+ if (!out)
+ return AVERROR(ENOMEM);
+
+ ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
+ if (ret < 0) {
av_frame_free(&out);
- goto fail;
+ return ret;
}
- av_audio_fifo_read(s->fifo[1], (void **)cf[1]->extended_data, s->nb_samples);
-
- s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples,
- outlink->channels, 1, 0, s->nb_samples, s->curve2);
+ s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples,
+ outlink->channels, -1, s->nb_samples - 1, s->nb_samples, s->curve);
out->pts = s->pts;
s->pts += av_rescale_q(s->nb_samples,
(AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+ av_frame_free(&cf[0]);
ret = ff_filter_frame(outlink, out);
if (ret < 0)
- goto fail;
- }
+ return ret;
- nb_samples = av_audio_fifo_size(s->fifo[1]);
- if (nb_samples > 0) {
- out = ff_get_audio_buffer(outlink, nb_samples);
- if (!out) {
- ret = AVERROR(ENOMEM);
- goto fail;
+ out = ff_get_audio_buffer(outlink, s->nb_samples);
+ if (!out)
+ return AVERROR(ENOMEM);
+
+ ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
+ if (ret < 0) {
+ av_frame_free(&out);
+ return ret;
}
- av_audio_fifo_read(s->fifo[1], (void **)out->extended_data, nb_samples);
+ s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples,
+ outlink->channels, 1, 0, s->nb_samples, s->curve2);
out->pts = s->pts;
- s->pts += av_rescale_q(nb_samples,
+ s->pts += av_rescale_q(s->nb_samples,
(AVRational){ 1, outlink->sample_rate }, outlink->time_base);
- ret = ff_filter_frame(outlink, out);
+ s->crossfade_is_over = 1;
+ av_frame_free(&cf[1]);
+ return ff_filter_frame(outlink, out);
}
- s->crossfade_is_over = 1;
- }
-
-fail:
- av_frame_free(&in);
- av_frame_free(&cf[0]);
- av_frame_free(&cf[1]);
- return ret;
-}
-
-static int acrossfade_request_frame(AVFilterLink *outlink)
-{
- AVFilterContext *ctx = outlink->src;
- AudioFadeContext *s = ctx->priv;
- int ret = 0;
-
- if (!s->cf0_eof) {
- AVFilterLink *cf0 = ctx->inputs[0];
- ret = ff_request_frame(cf0);
- if (ret < 0 && ret != AVERROR_EOF)
- return ret;
- if (ret == AVERROR_EOF) {
+ } else if (ff_outlink_frame_wanted(ctx->outputs[0])) {
+ if (!s->cf0_eof && ff_outlink_get_status(ctx->inputs[0])) {
s->cf0_eof = 1;
- ret = 0;
}
- } else {
- AVFilterLink *cf1 = ctx->inputs[1];
- int nb_samples = av_audio_fifo_size(s->fifo[1]);
-
- ret = ff_request_frame(cf1);
- if (ret == AVERROR_EOF && nb_samples > 0) {
- AVFrame *out = ff_get_audio_buffer(outlink, nb_samples);
- if (!out)
- return AVERROR(ENOMEM);
-
- av_audio_fifo_read(s->fifo[1], (void **)out->extended_data, nb_samples);
- ret = ff_filter_frame(outlink, out);
+ if (ff_outlink_get_status(ctx->inputs[1])) {
+ ff_outlink_set_status(ctx->outputs[0], AVERROR_EOF, AV_NOPTS_VALUE);
+ return 0;
}
+ if (!s->cf0_eof)
+ ff_inlink_request_frame(ctx->inputs[0]);
+ else
+ ff_inlink_request_frame(ctx->inputs[1]);
+ return 0;
}
return ret;
config_output(outlink);
- s->fifo[0] = av_audio_fifo_alloc(outlink->format, outlink->channels, s->nb_samples);
- s->fifo[1] = av_audio_fifo_alloc(outlink->format, outlink->channels, s->nb_samples);
- if (!s->fifo[0] || !s->fifo[1])
- return AVERROR(ENOMEM);
-
return 0;
}
-static av_cold void uninit(AVFilterContext *ctx)
-{
- AudioFadeContext *s = ctx->priv;
-
- av_audio_fifo_free(s->fifo[0]);
- av_audio_fifo_free(s->fifo[1]);
-}
-
static const AVFilterPad avfilter_af_acrossfade_inputs[] = {
{
.name = "crossfade0",
.type = AVMEDIA_TYPE_AUDIO,
- .filter_frame = acrossfade_filter_frame,
},
{
.name = "crossfade1",
.type = AVMEDIA_TYPE_AUDIO,
- .filter_frame = acrossfade_filter_frame,
},
{ NULL }
};
{
.name = "default",
.type = AVMEDIA_TYPE_AUDIO,
- .request_frame = acrossfade_request_frame,
.config_props = acrossfade_config_output,
},
{ NULL }
.description = NULL_IF_CONFIG_SMALL("Cross fade two input audio streams."),
.query_formats = query_formats,
.priv_size = sizeof(AudioFadeContext),
- .uninit = uninit,
+ .activate = activate,
.priv_class = &acrossfade_class,
.inputs = avfilter_af_acrossfade_inputs,
.outputs = avfilter_af_acrossfade_outputs,