2 * Copyright (c) 2013-2015 Paul B Mahol
4 * This file is part of FFmpeg.
6 * FFmpeg is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2.1 of the License, or (at your option) any later version.
11 * FFmpeg is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with FFmpeg; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
26 #include "libavutil/opt.h"
32 typedef struct AudioFadeContext {
42 int crossfade_is_over;
45 void (*fade_samples)(uint8_t **dst, uint8_t * const *src,
46 int nb_samples, int channels, int direction,
47 int64_t start, int64_t range, int curve);
48 void (*crossfade_samples)(uint8_t **dst, uint8_t * const *cf0,
50 int nb_samples, int channels,
51 int curve0, int curve1);
54 enum CurveType { TRI, QSIN, ESIN, HSIN, LOG, IPAR, QUA, CUB, SQU, CBR, PAR, EXP, IQSIN, IHSIN, DESE, DESI, LOSI, NONE, NB_CURVES };
56 #define OFFSET(x) offsetof(AudioFadeContext, x)
57 #define FLAGS AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
59 static int query_formats(AVFilterContext *ctx)
61 AVFilterFormats *formats;
62 AVFilterChannelLayouts *layouts;
63 static const enum AVSampleFormat sample_fmts[] = {
64 AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
65 AV_SAMPLE_FMT_S32, AV_SAMPLE_FMT_S32P,
66 AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_FLTP,
67 AV_SAMPLE_FMT_DBL, AV_SAMPLE_FMT_DBLP,
72 layouts = ff_all_channel_counts();
74 return AVERROR(ENOMEM);
75 ret = ff_set_common_channel_layouts(ctx, layouts);
79 formats = ff_make_format_list(sample_fmts);
81 return AVERROR(ENOMEM);
82 ret = ff_set_common_formats(ctx, formats);
86 formats = ff_all_samplerates();
88 return AVERROR(ENOMEM);
89 return ff_set_common_samplerates(ctx, formats);
92 static double fade_gain(int curve, int64_t index, int64_t range)
94 #define CUBE(a) ((a)*(a)*(a))
97 gain = av_clipd(1.0 * index / range, 0, 1.0);
101 gain = sin(gain * M_PI / 2.0);
104 /* 0.6... = 2 / M_PI */
105 gain = 0.6366197723675814 * asin(gain);
108 gain = 1.0 - cos(M_PI / 4.0 * (CUBE(2.0*gain - 1) + 1));
111 gain = (1.0 - cos(gain * M_PI)) / 2.0;
114 /* 0.3... = 1 / M_PI */
115 gain = 0.3183098861837907 * acos(1 - 2 * gain);
118 /* -11.5... = 5*ln(0.1) */
119 gain = exp(-11.512925464970227 * (1 - gain));
122 gain = av_clipd(1 + 0.2 * log10(gain), 0, 1.0);
125 gain = 1 - sqrt(1 - gain);
128 gain = (1 - (1 - gain) * (1 - gain));
143 gain = gain <= 0.5 ? cbrt(2 * gain) / 2: 1 - cbrt(2 * (1 - gain)) / 2;
146 gain = gain <= 0.5 ? CUBE(2 * gain) / 2: 1 - CUBE(2 * (1 - gain)) / 2;
149 const double a = 1. / (1. - 0.787) - 1;
150 double A = 1. / (1.0 + exp(0 -((gain-0.5) * a * 2.0)));
151 double B = 1. / (1.0 + exp(a));
152 double C = 1. / (1.0 + exp(0-a));
153 gain = (A - B) / (C - B);
164 #define FADE_PLANAR(name, type) \
165 static void fade_samples_## name ##p(uint8_t **dst, uint8_t * const *src, \
166 int nb_samples, int channels, int dir, \
167 int64_t start, int64_t range, int curve) \
171 for (i = 0; i < nb_samples; i++) { \
172 double gain = fade_gain(curve, start + i * dir, range); \
173 for (c = 0; c < channels; c++) { \
174 type *d = (type *)dst[c]; \
175 const type *s = (type *)src[c]; \
177 d[i] = s[i] * gain; \
182 #define FADE(name, type) \
183 static void fade_samples_## name (uint8_t **dst, uint8_t * const *src, \
184 int nb_samples, int channels, int dir, \
185 int64_t start, int64_t range, int curve) \
187 type *d = (type *)dst[0]; \
188 const type *s = (type *)src[0]; \
191 for (i = 0; i < nb_samples; i++) { \
192 double gain = fade_gain(curve, start + i * dir, range); \
193 for (c = 0; c < channels; c++, k++) \
194 d[k] = s[k] * gain; \
198 FADE_PLANAR(dbl, double)
199 FADE_PLANAR(flt, float)
200 FADE_PLANAR(s16, int16_t)
201 FADE_PLANAR(s32, int32_t)
208 static int config_output(AVFilterLink *outlink)
210 AVFilterContext *ctx = outlink->src;
211 AudioFadeContext *s = ctx->priv;
213 switch (outlink->format) {
214 case AV_SAMPLE_FMT_DBL: s->fade_samples = fade_samples_dbl; break;
215 case AV_SAMPLE_FMT_DBLP: s->fade_samples = fade_samples_dblp; break;
216 case AV_SAMPLE_FMT_FLT: s->fade_samples = fade_samples_flt; break;
217 case AV_SAMPLE_FMT_FLTP: s->fade_samples = fade_samples_fltp; break;
218 case AV_SAMPLE_FMT_S16: s->fade_samples = fade_samples_s16; break;
219 case AV_SAMPLE_FMT_S16P: s->fade_samples = fade_samples_s16p; break;
220 case AV_SAMPLE_FMT_S32: s->fade_samples = fade_samples_s32; break;
221 case AV_SAMPLE_FMT_S32P: s->fade_samples = fade_samples_s32p; break;
225 s->nb_samples = av_rescale(s->duration, outlink->sample_rate, AV_TIME_BASE);
227 s->start_sample = av_rescale(s->start_time, outlink->sample_rate, AV_TIME_BASE);
232 #if CONFIG_AFADE_FILTER
234 static const AVOption afade_options[] = {
235 { "type", "set the fade direction", OFFSET(type), AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 1, FLAGS, "type" },
236 { "t", "set the fade direction", OFFSET(type), AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 1, FLAGS, "type" },
237 { "in", "fade-in", 0, AV_OPT_TYPE_CONST, {.i64 = 0 }, 0, 0, FLAGS, "type" },
238 { "out", "fade-out", 0, AV_OPT_TYPE_CONST, {.i64 = 1 }, 0, 0, FLAGS, "type" },
239 { "start_sample", "set number of first sample to start fading", OFFSET(start_sample), AV_OPT_TYPE_INT64, {.i64 = 0 }, 0, INT64_MAX, FLAGS },
240 { "ss", "set number of first sample to start fading", OFFSET(start_sample), AV_OPT_TYPE_INT64, {.i64 = 0 }, 0, INT64_MAX, FLAGS },
241 { "nb_samples", "set number of samples for fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT64, {.i64 = 44100}, 1, INT64_MAX, FLAGS },
242 { "ns", "set number of samples for fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT64, {.i64 = 44100}, 1, INT64_MAX, FLAGS },
243 { "start_time", "set time to start fading", OFFSET(start_time), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
244 { "st", "set time to start fading", OFFSET(start_time), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
245 { "duration", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
246 { "d", "set fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT64_MAX, FLAGS },
247 { "curve", "set fade curve type", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
248 { "c", "set fade curve type", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
249 { "tri", "linear slope", 0, AV_OPT_TYPE_CONST, {.i64 = TRI }, 0, 0, FLAGS, "curve" },
250 { "qsin", "quarter of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = QSIN }, 0, 0, FLAGS, "curve" },
251 { "esin", "exponential sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = ESIN }, 0, 0, FLAGS, "curve" },
252 { "hsin", "half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = HSIN }, 0, 0, FLAGS, "curve" },
253 { "log", "logarithmic", 0, AV_OPT_TYPE_CONST, {.i64 = LOG }, 0, 0, FLAGS, "curve" },
254 { "ipar", "inverted parabola", 0, AV_OPT_TYPE_CONST, {.i64 = IPAR }, 0, 0, FLAGS, "curve" },
255 { "qua", "quadratic", 0, AV_OPT_TYPE_CONST, {.i64 = QUA }, 0, 0, FLAGS, "curve" },
256 { "cub", "cubic", 0, AV_OPT_TYPE_CONST, {.i64 = CUB }, 0, 0, FLAGS, "curve" },
257 { "squ", "square root", 0, AV_OPT_TYPE_CONST, {.i64 = SQU }, 0, 0, FLAGS, "curve" },
258 { "cbr", "cubic root", 0, AV_OPT_TYPE_CONST, {.i64 = CBR }, 0, 0, FLAGS, "curve" },
259 { "par", "parabola", 0, AV_OPT_TYPE_CONST, {.i64 = PAR }, 0, 0, FLAGS, "curve" },
260 { "exp", "exponential", 0, AV_OPT_TYPE_CONST, {.i64 = EXP }, 0, 0, FLAGS, "curve" },
261 { "iqsin", "inverted quarter of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IQSIN}, 0, 0, FLAGS, "curve" },
262 { "ihsin", "inverted half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IHSIN}, 0, 0, FLAGS, "curve" },
263 { "dese", "double-exponential seat", 0, AV_OPT_TYPE_CONST, {.i64 = DESE }, 0, 0, FLAGS, "curve" },
264 { "desi", "double-exponential sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = DESI }, 0, 0, FLAGS, "curve" },
265 { "losi", "logistic sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = LOSI }, 0, 0, FLAGS, "curve" },
266 { "nofade", "no fade; keep audio as-is", 0, AV_OPT_TYPE_CONST, {.i64 = NONE }, 0, 0, FLAGS, "curve" },
270 AVFILTER_DEFINE_CLASS(afade);
272 static av_cold int init(AVFilterContext *ctx)
274 AudioFadeContext *s = ctx->priv;
276 if (INT64_MAX - s->nb_samples < s->start_sample)
277 return AVERROR(EINVAL);
282 static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
284 AudioFadeContext *s = inlink->dst->priv;
285 AVFilterLink *outlink = inlink->dst->outputs[0];
286 int nb_samples = buf->nb_samples;
288 int64_t cur_sample = av_rescale_q(buf->pts, inlink->time_base, (AVRational){1, inlink->sample_rate});
290 if ((!s->type && (s->start_sample + s->nb_samples < cur_sample)) ||
291 ( s->type && (cur_sample + nb_samples < s->start_sample)))
292 return ff_filter_frame(outlink, buf);
294 if (av_frame_is_writable(buf)) {
297 out_buf = ff_get_audio_buffer(outlink, nb_samples);
299 return AVERROR(ENOMEM);
300 av_frame_copy_props(out_buf, buf);
303 if ((!s->type && (cur_sample + nb_samples < s->start_sample)) ||
304 ( s->type && (s->start_sample + s->nb_samples < cur_sample))) {
305 av_samples_set_silence(out_buf->extended_data, 0, nb_samples,
306 out_buf->channels, out_buf->format);
311 start = cur_sample - s->start_sample;
313 start = s->start_sample + s->nb_samples - cur_sample;
315 s->fade_samples(out_buf->extended_data, buf->extended_data,
316 nb_samples, buf->channels,
317 s->type ? -1 : 1, start,
318 s->nb_samples, s->curve);
324 return ff_filter_frame(outlink, out_buf);
327 static const AVFilterPad avfilter_af_afade_inputs[] = {
330 .type = AVMEDIA_TYPE_AUDIO,
331 .filter_frame = filter_frame,
336 static const AVFilterPad avfilter_af_afade_outputs[] = {
339 .type = AVMEDIA_TYPE_AUDIO,
340 .config_props = config_output,
345 AVFilter ff_af_afade = {
347 .description = NULL_IF_CONFIG_SMALL("Fade in/out input audio."),
348 .query_formats = query_formats,
349 .priv_size = sizeof(AudioFadeContext),
351 .inputs = avfilter_af_afade_inputs,
352 .outputs = avfilter_af_afade_outputs,
353 .priv_class = &afade_class,
354 .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC,
357 #endif /* CONFIG_AFADE_FILTER */
359 #if CONFIG_ACROSSFADE_FILTER
361 static const AVOption acrossfade_options[] = {
362 { "nb_samples", "set number of samples for cross fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 44100}, 1, INT32_MAX/10, FLAGS },
363 { "ns", "set number of samples for cross fade duration", OFFSET(nb_samples), AV_OPT_TYPE_INT, {.i64 = 44100}, 1, INT32_MAX/10, FLAGS },
364 { "duration", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60000000, FLAGS },
365 { "d", "set cross fade duration", OFFSET(duration), AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, 60000000, FLAGS },
366 { "overlap", "overlap 1st stream end with 2nd stream start", OFFSET(overlap), AV_OPT_TYPE_BOOL, {.i64 = 1 }, 0, 1, FLAGS },
367 { "o", "overlap 1st stream end with 2nd stream start", OFFSET(overlap), AV_OPT_TYPE_BOOL, {.i64 = 1 }, 0, 1, FLAGS },
368 { "curve1", "set fade curve type for 1st stream", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
369 { "c1", "set fade curve type for 1st stream", OFFSET(curve), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
370 { "tri", "linear slope", 0, AV_OPT_TYPE_CONST, {.i64 = TRI }, 0, 0, FLAGS, "curve" },
371 { "qsin", "quarter of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = QSIN }, 0, 0, FLAGS, "curve" },
372 { "esin", "exponential sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = ESIN }, 0, 0, FLAGS, "curve" },
373 { "hsin", "half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = HSIN }, 0, 0, FLAGS, "curve" },
374 { "log", "logarithmic", 0, AV_OPT_TYPE_CONST, {.i64 = LOG }, 0, 0, FLAGS, "curve" },
375 { "ipar", "inverted parabola", 0, AV_OPT_TYPE_CONST, {.i64 = IPAR }, 0, 0, FLAGS, "curve" },
376 { "qua", "quadratic", 0, AV_OPT_TYPE_CONST, {.i64 = QUA }, 0, 0, FLAGS, "curve" },
377 { "cub", "cubic", 0, AV_OPT_TYPE_CONST, {.i64 = CUB }, 0, 0, FLAGS, "curve" },
378 { "squ", "square root", 0, AV_OPT_TYPE_CONST, {.i64 = SQU }, 0, 0, FLAGS, "curve" },
379 { "cbr", "cubic root", 0, AV_OPT_TYPE_CONST, {.i64 = CBR }, 0, 0, FLAGS, "curve" },
380 { "par", "parabola", 0, AV_OPT_TYPE_CONST, {.i64 = PAR }, 0, 0, FLAGS, "curve" },
381 { "exp", "exponential", 0, AV_OPT_TYPE_CONST, {.i64 = EXP }, 0, 0, FLAGS, "curve" },
382 { "iqsin", "inverted quarter of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IQSIN}, 0, 0, FLAGS, "curve" },
383 { "ihsin", "inverted half of sine wave", 0, AV_OPT_TYPE_CONST, {.i64 = IHSIN}, 0, 0, FLAGS, "curve" },
384 { "dese", "double-exponential seat", 0, AV_OPT_TYPE_CONST, {.i64 = DESE }, 0, 0, FLAGS, "curve" },
385 { "desi", "double-exponential sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = DESI }, 0, 0, FLAGS, "curve" },
386 { "losi", "logistic sigmoid", 0, AV_OPT_TYPE_CONST, {.i64 = LOSI }, 0, 0, FLAGS, "curve" },
387 { "nofade", "no fade; keep audio as-is", 0, AV_OPT_TYPE_CONST, {.i64 = NONE }, 0, 0, FLAGS, "curve" },
388 { "curve2", "set fade curve type for 2nd stream", OFFSET(curve2), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
389 { "c2", "set fade curve type for 2nd stream", OFFSET(curve2), AV_OPT_TYPE_INT, {.i64 = TRI }, 0, NB_CURVES - 1, FLAGS, "curve" },
393 AVFILTER_DEFINE_CLASS(acrossfade);
395 #define CROSSFADE_PLANAR(name, type) \
396 static void crossfade_samples_## name ##p(uint8_t **dst, uint8_t * const *cf0, \
397 uint8_t * const *cf1, \
398 int nb_samples, int channels, \
399 int curve0, int curve1) \
403 for (i = 0; i < nb_samples; i++) { \
404 double gain0 = fade_gain(curve0, nb_samples - 1 - i, nb_samples); \
405 double gain1 = fade_gain(curve1, i, nb_samples); \
406 for (c = 0; c < channels; c++) { \
407 type *d = (type *)dst[c]; \
408 const type *s0 = (type *)cf0[c]; \
409 const type *s1 = (type *)cf1[c]; \
411 d[i] = s0[i] * gain0 + s1[i] * gain1; \
416 #define CROSSFADE(name, type) \
417 static void crossfade_samples_## name (uint8_t **dst, uint8_t * const *cf0, \
418 uint8_t * const *cf1, \
419 int nb_samples, int channels, \
420 int curve0, int curve1) \
422 type *d = (type *)dst[0]; \
423 const type *s0 = (type *)cf0[0]; \
424 const type *s1 = (type *)cf1[0]; \
427 for (i = 0; i < nb_samples; i++) { \
428 double gain0 = fade_gain(curve0, nb_samples - 1 - i, nb_samples); \
429 double gain1 = fade_gain(curve1, i, nb_samples); \
430 for (c = 0; c < channels; c++, k++) \
431 d[k] = s0[k] * gain0 + s1[k] * gain1; \
435 CROSSFADE_PLANAR(dbl, double)
436 CROSSFADE_PLANAR(flt, float)
437 CROSSFADE_PLANAR(s16, int16_t)
438 CROSSFADE_PLANAR(s32, int32_t)
440 CROSSFADE(dbl, double)
441 CROSSFADE(flt, float)
442 CROSSFADE(s16, int16_t)
443 CROSSFADE(s32, int32_t)
445 static int activate(AVFilterContext *ctx)
447 AudioFadeContext *s = ctx->priv;
448 AVFilterLink *outlink = ctx->outputs[0];
449 AVFrame *in = NULL, *out, *cf[2] = { NULL };
450 int ret = 0, nb_samples, status;
453 FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx);
455 if (s->crossfade_is_over) {
456 ret = ff_inlink_consume_frame(ctx->inputs[1], &in);
459 } else if (ff_inlink_acknowledge_status(ctx->inputs[1], &status, &pts)) {
460 ff_outlink_set_status(ctx->outputs[0], status, pts);
463 if (ff_outlink_frame_wanted(ctx->outputs[0]) && !in) {
464 ff_inlink_request_frame(ctx->inputs[1]);
469 s->pts += av_rescale_q(in->nb_samples,
470 (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
471 return ff_filter_frame(outlink, in);
474 if (ff_inlink_queued_samples(ctx->inputs[0]) > s->nb_samples) {
475 nb_samples = ff_inlink_queued_samples(ctx->inputs[0]) - s->nb_samples;
476 if (nb_samples > 0) {
477 ret = ff_inlink_consume_samples(ctx->inputs[0], nb_samples, nb_samples, &in);
483 s->pts += av_rescale_q(in->nb_samples,
484 (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
485 return ff_filter_frame(outlink, in);
486 } else if (ff_inlink_queued_samples(ctx->inputs[1]) >= s->nb_samples) {
488 out = ff_get_audio_buffer(outlink, s->nb_samples);
490 return AVERROR(ENOMEM);
492 ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
498 ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
504 s->crossfade_samples(out->extended_data, cf[0]->extended_data,
505 cf[1]->extended_data,
506 s->nb_samples, out->channels,
507 s->curve, s->curve2);
509 s->pts += av_rescale_q(s->nb_samples,
510 (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
511 s->crossfade_is_over = 1;
512 av_frame_free(&cf[0]);
513 av_frame_free(&cf[1]);
514 return ff_filter_frame(outlink, out);
516 out = ff_get_audio_buffer(outlink, s->nb_samples);
518 return AVERROR(ENOMEM);
520 ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
526 s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples,
527 outlink->channels, -1, s->nb_samples - 1, s->nb_samples, s->curve);
529 s->pts += av_rescale_q(s->nb_samples,
530 (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
531 av_frame_free(&cf[0]);
532 ret = ff_filter_frame(outlink, out);
536 out = ff_get_audio_buffer(outlink, s->nb_samples);
538 return AVERROR(ENOMEM);
540 ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
546 s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples,
547 outlink->channels, 1, 0, s->nb_samples, s->curve2);
549 s->pts += av_rescale_q(s->nb_samples,
550 (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
551 s->crossfade_is_over = 1;
552 av_frame_free(&cf[1]);
553 return ff_filter_frame(outlink, out);
555 } else if (ff_outlink_frame_wanted(ctx->outputs[0])) {
556 if (!s->cf0_eof && ff_outlink_get_status(ctx->inputs[0])) {
559 if (ff_outlink_get_status(ctx->inputs[1])) {
560 ff_outlink_set_status(ctx->outputs[0], AVERROR_EOF, AV_NOPTS_VALUE);
564 ff_inlink_request_frame(ctx->inputs[0]);
566 ff_inlink_request_frame(ctx->inputs[1]);
573 static int acrossfade_config_output(AVFilterLink *outlink)
575 AVFilterContext *ctx = outlink->src;
576 AudioFadeContext *s = ctx->priv;
578 if (ctx->inputs[0]->sample_rate != ctx->inputs[1]->sample_rate) {
579 av_log(ctx, AV_LOG_ERROR,
580 "Inputs must have the same sample rate "
581 "%d for in0 vs %d for in1\n",
582 ctx->inputs[0]->sample_rate, ctx->inputs[1]->sample_rate);
583 return AVERROR(EINVAL);
586 outlink->sample_rate = ctx->inputs[0]->sample_rate;
587 outlink->time_base = ctx->inputs[0]->time_base;
588 outlink->channel_layout = ctx->inputs[0]->channel_layout;
589 outlink->channels = ctx->inputs[0]->channels;
591 switch (outlink->format) {
592 case AV_SAMPLE_FMT_DBL: s->crossfade_samples = crossfade_samples_dbl; break;
593 case AV_SAMPLE_FMT_DBLP: s->crossfade_samples = crossfade_samples_dblp; break;
594 case AV_SAMPLE_FMT_FLT: s->crossfade_samples = crossfade_samples_flt; break;
595 case AV_SAMPLE_FMT_FLTP: s->crossfade_samples = crossfade_samples_fltp; break;
596 case AV_SAMPLE_FMT_S16: s->crossfade_samples = crossfade_samples_s16; break;
597 case AV_SAMPLE_FMT_S16P: s->crossfade_samples = crossfade_samples_s16p; break;
598 case AV_SAMPLE_FMT_S32: s->crossfade_samples = crossfade_samples_s32; break;
599 case AV_SAMPLE_FMT_S32P: s->crossfade_samples = crossfade_samples_s32p; break;
602 config_output(outlink);
607 static const AVFilterPad avfilter_af_acrossfade_inputs[] = {
609 .name = "crossfade0",
610 .type = AVMEDIA_TYPE_AUDIO,
613 .name = "crossfade1",
614 .type = AVMEDIA_TYPE_AUDIO,
619 static const AVFilterPad avfilter_af_acrossfade_outputs[] = {
622 .type = AVMEDIA_TYPE_AUDIO,
623 .config_props = acrossfade_config_output,
628 AVFilter ff_af_acrossfade = {
629 .name = "acrossfade",
630 .description = NULL_IF_CONFIG_SMALL("Cross fade two input audio streams."),
631 .query_formats = query_formats,
632 .priv_size = sizeof(AudioFadeContext),
633 .activate = activate,
634 .priv_class = &acrossfade_class,
635 .inputs = avfilter_af_acrossfade_inputs,
636 .outputs = avfilter_af_acrossfade_outputs,
639 #endif /* CONFIG_ACROSSFADE_FILTER */