/**
* @file
- * FIFO buffering video filter
+ * FIFO buffering filter
*/
+#include "libavutil/avassert.h"
+#include "libavutil/channel_layout.h"
+#include "libavutil/common.h"
+#include "libavutil/mathematics.h"
+#include "libavutil/samplefmt.h"
+
+#include "audio.h"
#include "avfilter.h"
#include "internal.h"
#include "video.h"
-typedef struct BufPic {
- AVFilterBufferRef *picref;
- struct BufPic *next;
-} BufPic;
+typedef struct Buf {
+ AVFrame *frame;
+ struct Buf *next;
+} Buf;
typedef struct {
- BufPic root;
- BufPic *last; ///< last buffered picture
+ Buf root;
+ Buf *last; ///< last buffered frame
+
+ /**
+ * When a specific number of output samples is requested, the partial
+ * buffer is stored here
+ */
+ AVFrame *out;
+ int allocated_samples; ///< number of samples out was allocated for
} FifoContext;
-static av_cold int init(AVFilterContext *ctx, const char *args, void *opaque)
+static av_cold int init(AVFilterContext *ctx, const char *args)
{
FifoContext *fifo = ctx->priv;
fifo->last = &fifo->root;
- av_log(ctx, AV_LOG_INFO, "\n");
return 0;
}
static av_cold void uninit(AVFilterContext *ctx)
{
FifoContext *fifo = ctx->priv;
- BufPic *pic, *tmp;
+ Buf *buf, *tmp;
- for (pic = fifo->root.next; pic; pic = tmp) {
- tmp = pic->next;
- avfilter_unref_buffer(pic->picref);
- av_free(pic);
+ for (buf = fifo->root.next; buf; buf = tmp) {
+ tmp = buf->next;
+ av_frame_free(&buf->frame);
+ av_free(buf);
}
+
+ av_frame_free(&fifo->out);
}
-static void start_frame(AVFilterLink *inlink, AVFilterBufferRef *picref)
+static int add_to_queue(AVFilterLink *inlink, AVFrame *frame)
{
FifoContext *fifo = inlink->dst->priv;
- fifo->last->next = av_mallocz(sizeof(BufPic));
+ fifo->last->next = av_mallocz(sizeof(Buf));
+ if (!fifo->last->next) {
+ av_frame_free(&frame);
+ return AVERROR(ENOMEM);
+ }
+
fifo->last = fifo->last->next;
- fifo->last->picref = picref;
+ fifo->last->frame = frame;
+
+ return 0;
+}
+
+static void queue_pop(FifoContext *s)
+{
+ Buf *tmp = s->root.next->next;
+ if (s->last == s->root.next)
+ s->last = &s->root;
+ av_freep(&s->root.next);
+ s->root.next = tmp;
+}
+
+/**
+ * Move data pointers and pts offset samples forward.
+ */
+static void buffer_offset(AVFilterLink *link, AVFrame *frame,
+ int offset)
+{
+ int nb_channels = av_get_channel_layout_nb_channels(link->channel_layout);
+ int planar = av_sample_fmt_is_planar(link->format);
+ int planes = planar ? nb_channels : 1;
+ int block_align = av_get_bytes_per_sample(link->format) * (planar ? 1 : nb_channels);
+ int i;
+
+ av_assert0(frame->nb_samples > offset);
+
+ for (i = 0; i < planes; i++)
+ frame->extended_data[i] += block_align * offset;
+ if (frame->data != frame->extended_data)
+ memcpy(frame->data, frame->extended_data,
+ FFMIN(planes, FF_ARRAY_ELEMS(frame->data)) * sizeof(*frame->data));
+ frame->linesize[0] -= block_align*offset;
+ frame->nb_samples -= offset;
+
+ if (frame->pts != AV_NOPTS_VALUE) {
+ frame->pts += av_rescale_q(offset, (AVRational){1, link->sample_rate},
+ link->time_base);
+ }
+}
+
+static int calc_ptr_alignment(AVFrame *frame)
+{
+ int planes = av_sample_fmt_is_planar(frame->format) ?
+ av_get_channel_layout_nb_channels(frame->channel_layout) : 1;
+ int min_align = 128;
+ int p;
+
+ for (p = 0; p < planes; p++) {
+ int cur_align = 128;
+ while ((intptr_t)frame->extended_data[p] % cur_align)
+ cur_align >>= 1;
+ if (cur_align < min_align)
+ min_align = cur_align;
+ }
+ return min_align;
}
-static void end_frame(AVFilterLink *inlink) { }
+static int return_audio_frame(AVFilterContext *ctx)
+{
+ AVFilterLink *link = ctx->outputs[0];
+ FifoContext *s = ctx->priv;
+ AVFrame *head = s->root.next->frame;
+ AVFrame *out;
+ int ret;
+
+ if (!s->out &&
+ head->nb_samples >= link->request_samples &&
+ calc_ptr_alignment(head) >= 32) {
+ if (head->nb_samples == link->request_samples) {
+ out = head;
+ queue_pop(s);
+ } else {
+ out = av_frame_clone(head);
+ if (!out)
+ return AVERROR(ENOMEM);
+
+ out->nb_samples = link->request_samples;
+ buffer_offset(link, head, link->request_samples);
+ }
+ } else {
+ int nb_channels = av_get_channel_layout_nb_channels(link->channel_layout);
-static void draw_slice(AVFilterLink *inlink, int y, int h, int slice_dir) { }
+ if (!s->out) {
+ s->out = ff_get_audio_buffer(link, link->request_samples);
+ if (!s->out)
+ return AVERROR(ENOMEM);
+
+ s->out->nb_samples = 0;
+ s->out->pts = head->pts;
+ s->allocated_samples = link->request_samples;
+ } else if (link->request_samples != s->allocated_samples) {
+ av_log(ctx, AV_LOG_ERROR, "request_samples changed before the "
+ "buffer was returned.\n");
+ return AVERROR(EINVAL);
+ }
+
+ while (s->out->nb_samples < s->allocated_samples) {
+ int len = FFMIN(s->allocated_samples - s->out->nb_samples,
+ head->nb_samples);
+
+ av_samples_copy(s->out->extended_data, head->extended_data,
+ s->out->nb_samples, 0, len, nb_channels,
+ link->format);
+ s->out->nb_samples += len;
+
+ if (len == head->nb_samples) {
+ av_frame_free(&head);
+ queue_pop(s);
+
+ if (!s->root.next &&
+ (ret = ff_request_frame(ctx->inputs[0])) < 0) {
+ if (ret == AVERROR_EOF) {
+ av_samples_set_silence(s->out->extended_data,
+ s->out->nb_samples,
+ s->allocated_samples -
+ s->out->nb_samples,
+ nb_channels, link->format);
+ s->out->nb_samples = s->allocated_samples;
+ break;
+ }
+ return ret;
+ }
+ head = s->root.next->frame;
+ } else {
+ buffer_offset(link, head, len);
+ }
+ }
+ out = s->out;
+ s->out = NULL;
+ }
+ return ff_filter_frame(link, out);
+}
static int request_frame(AVFilterLink *outlink)
{
FifoContext *fifo = outlink->src->priv;
- BufPic *tmp;
- int ret;
+ int ret = 0;
if (!fifo->root.next) {
if ((ret = ff_request_frame(outlink->src->inputs[0])) < 0)
return ret;
}
- /* by doing this, we give ownership of the reference to the next filter,
- * so we don't have to worry about dereferencing it ourselves. */
- ff_start_frame(outlink, fifo->root.next->picref);
- ff_draw_slice (outlink, 0, outlink->h, 1);
- ff_end_frame (outlink);
-
- if (fifo->last == fifo->root.next)
- fifo->last = &fifo->root;
- tmp = fifo->root.next->next;
- av_free(fifo->root.next);
- fifo->root.next = tmp;
+ if (outlink->request_samples) {
+ return return_audio_frame(outlink->src);
+ } else {
+ ret = ff_filter_frame(outlink, fifo->root.next->frame);
+ queue_pop(fifo);
+ }
- return 0;
+ return ret;
}
+static const AVFilterPad avfilter_vf_fifo_inputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .get_video_buffer = ff_null_get_video_buffer,
+ .filter_frame = add_to_queue,
+ },
+ { NULL }
+};
+
+static const AVFilterPad avfilter_vf_fifo_outputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .request_frame = request_frame,
+ },
+ { NULL }
+};
+
AVFilter avfilter_vf_fifo = {
.name = "fifo",
.description = NULL_IF_CONFIG_SMALL("Buffer input images and send them when they are requested."),
.priv_size = sizeof(FifoContext),
- .inputs = (AVFilterPad[]) {{ .name = "default",
- .type = AVMEDIA_TYPE_VIDEO,
- .get_video_buffer= ff_null_get_video_buffer,
- .start_frame = start_frame,
- .draw_slice = draw_slice,
- .end_frame = end_frame,
- .rej_perms = AV_PERM_REUSE2, },
- { .name = NULL}},
- .outputs = (AVFilterPad[]) {{ .name = "default",
- .type = AVMEDIA_TYPE_VIDEO,
- .request_frame = request_frame, },
- { .name = NULL}},
+ .inputs = avfilter_vf_fifo_inputs,
+ .outputs = avfilter_vf_fifo_outputs,
+};
+
+static const AVFilterPad avfilter_af_afifo_inputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_AUDIO,
+ .get_audio_buffer = ff_null_get_audio_buffer,
+ .filter_frame = add_to_queue,
+ },
+ { NULL }
+};
+
+static const AVFilterPad avfilter_af_afifo_outputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_AUDIO,
+ .request_frame = request_frame,
+ },
+ { NULL }
+};
+
+AVFilter avfilter_af_afifo = {
+ .name = "afifo",
+ .description = NULL_IF_CONFIG_SMALL("Buffer input frames and send them when they are requested."),
+
+ .init = init,
+ .uninit = uninit,
+
+ .priv_size = sizeof(FifoContext),
+
+ .inputs = avfilter_af_afifo_inputs,
+ .outputs = avfilter_af_afifo_outputs,
};