int ir_len;
int air_len;
- int nb_inputs;
+ int nb_hrir_inputs;
int nb_irs;
FFTContext *fft[2], *ifft[2];
FFTComplex *data_hrtf[2];
- AVFloatDSPContext *fdsp;
- struct headphone_inputs {
+ float (*scalarproduct_float)(const float *v1, const float *v2, int len);
+ struct hrir_inputs {
int ir_len;
int eof;
- } *in;
+ } hrir_in[64];
uint64_t mapping[64];
} HeadphoneContext;
char *arg, *tokenizer, *p;
uint64_t used_channels = 0;
- s->nb_inputs = 1;
-
p = s->map;
while ((arg = av_strtok(p, "|", &tokenizer))) {
uint64_t out_channel;
}
if (s->hrir_fmt == HRIR_MULTI)
- s->nb_inputs = 2;
+ s->nb_hrir_inputs = 1;
else
- s->nb_inputs = s->nb_irs + 1;
+ s->nb_hrir_inputs = s->nb_irs;
}
typedef struct ThreadData {
}
for (i = 0; i < in->nb_samples; i++) {
- const float *temp_ir = ir;
+ const float *cur_ir = ir;
*dst = 0;
for (l = 0; l < in_channels; l++) {
*(buffer[l] + wr) = src[l];
}
- for (l = 0; l < in_channels; l++) {
+ for (l = 0; l < in_channels; cur_ir += air_len, l++) {
const float *const bptr = buffer[l];
if (l == s->lfe_channel) {
*dst += *(buffer[s->lfe_channel] + wr) * s->gain_lfe;
- temp_ir += air_len;
continue;
}
- read = (wr - (ir_len - 1) + buffer_length) & modulo;
+ read = (wr - (ir_len - 1)) & modulo;
if (read + ir_len < buffer_length) {
memcpy(temp_src, bptr + read, ir_len * sizeof(*temp_src));
memcpy(temp_src + len, bptr, (air_len - len) * sizeof(*temp_src));
}
- dst[0] += s->fdsp->scalarproduct_float(temp_ir, temp_src, FFALIGN(ir_len, 32));
- temp_ir += air_len;
+ dst[0] += s->scalarproduct_float(cur_ir, temp_src, FFALIGN(ir_len, 32));
}
if (fabsf(dst[0]) > 1)
for (j = 0; j < in->nb_samples; j++) {
dst[2 * j] += fft_acc[j].re * fft_scale;
+ if (fabsf(dst[2 * j]) > 1)
+ n_clippings[0]++;
}
for (j = 0; j < ir_len - 1; j++) {
*(ringbuffer + write_pos) += fft_acc[in->nb_samples + j].re * fft_scale;
}
- for (i = 0; i < out->nb_samples; i++) {
- if (fabsf(dst[0]) > 1) {
- n_clippings[0]++;
- }
-
- dst += 2;
- }
-
*write = wr;
return 0;
av_log(ctx, AV_LOG_ERROR, "Too big length of IRs: %d > %d.\n", ir_len, max_ir_len);
return AVERROR(EINVAL);
}
- s->in[input_number].ir_len = ir_len;
+ s->hrir_in[input_number].ir_len = ir_len;
s->ir_len = FFMAX(ir_len, s->ir_len);
return 0;
}
}
- for (i = 0; i < s->nb_inputs - 1; av_frame_free(&frame), i++) {
- int len = s->in[i + 1].ir_len;
+ for (i = 0; i < s->nb_hrir_inputs; av_frame_free(&frame), i++) {
+ int len = s->hrir_in[i].ir_len;
float *ptr;
ret = ff_inlink_consume_samples(ctx->inputs[i + 1], len, len, &frame);
FF_FILTER_FORWARD_STATUS_BACK_ALL(ctx->outputs[0], ctx);
if (!s->eof_hrirs) {
int eof = 1;
- for (i = 1; i < s->nb_inputs; i++) {
- if (s->in[i].eof)
+ for (i = 0; i < s->nb_hrir_inputs; i++) {
+ AVFilterLink *input = ctx->inputs[i + 1];
+
+ if (s->hrir_in[i].eof)
continue;
- if ((ret = check_ir(ctx->inputs[i], i)) < 0)
+ if ((ret = check_ir(input, i)) < 0)
return ret;
- if (ff_outlink_get_status(ctx->inputs[i]) == AVERROR_EOF) {
- if (!ff_inlink_queued_samples(ctx->inputs[i])) {
+ if (ff_outlink_get_status(input) == AVERROR_EOF) {
+ if (!ff_inlink_queued_samples(input)) {
av_log(ctx, AV_LOG_ERROR, "No samples provided for "
- "HRIR stream %d.\n", i - 1);
+ "HRIR stream %d.\n", i);
return AVERROR_INVALIDDATA;
}
- s->in[i].eof = 1;
+ s->hrir_in[i].eof = 1;
} else {
if (ff_outlink_frame_wanted(ctx->outputs[0]))
- ff_inlink_request_frame(ctx->inputs[i]);
+ ff_inlink_request_frame(input);
eof = 0;
}
}
if (ret)
return ret;
} else {
- for (i = 1; i < s->nb_inputs; i++) {
+ for (i = 1; i <= s->nb_hrir_inputs; i++) {
ret = ff_channel_layouts_ref(stereo_layout, &ctx->inputs[i]->outcfg.channel_layouts);
if (ret)
return ret;
parse_map(ctx);
- s->in = av_calloc(s->nb_inputs, sizeof(*s->in));
- if (!s->in)
- return AVERROR(ENOMEM);
-
- for (i = 1; i < s->nb_inputs; i++) {
- char *name = av_asprintf("hrir%d", i - 1);
+ for (i = 0; i < s->nb_hrir_inputs; i++) {
+ char *name = av_asprintf("hrir%d", i);
AVFilterPad pad = {
.name = name,
.type = AVMEDIA_TYPE_AUDIO,
};
if (!name)
return AVERROR(ENOMEM);
- if ((ret = ff_insert_inpad(ctx, i, &pad)) < 0) {
+ if ((ret = ff_insert_inpad(ctx, i + 1, &pad)) < 0) {
av_freep(&pad.name);
return ret;
}
}
- s->fdsp = avpriv_float_dsp_alloc(0);
- if (!s->fdsp)
- return AVERROR(ENOMEM);
+ if (s->type == TIME_DOMAIN) {
+ AVFloatDSPContext *fdsp = avpriv_float_dsp_alloc(0);
+ if (!fdsp)
+ return AVERROR(ENOMEM);
+ s->scalarproduct_float = fdsp->scalarproduct_float;
+ av_free(fdsp);
+ }
return 0;
}
av_freep(&s->temp_afft[1]);
av_freep(&s->data_hrtf[0]);
av_freep(&s->data_hrtf[1]);
- av_freep(&s->fdsp);
- av_freep(&s->in);
for (unsigned i = 1; i < ctx->nb_inputs; i++)
av_freep(&ctx->input_pads[i].name);
}
{ NULL }
};
-AVFilter ff_af_headphone = {
+const AVFilter ff_af_headphone = {
.name = "headphone",
.description = NULL_IF_CONFIG_SMALL("Apply headphone binaural spatialization with HRTFs in additional streams."),
.priv_size = sizeof(HeadphoneContext),