typedef struct WmallDecodeCtx {
/* generic decoder variables */
AVCodecContext *avctx;
- AVFrame frame;
+ AVFrame *frame;
uint8_t frame_data[MAX_FRAMESIZE + FF_INPUT_BUFFER_PADDING_SIZE]; ///< compressed frame data
PutBitContext pb; ///< context for filling the frame_data buffer
avctx->sample_fmt = AV_SAMPLE_FMT_S16P;
else if (s->bits_per_sample == 24) {
avctx->sample_fmt = AV_SAMPLE_FMT_S32P;
- av_log_missing_feature(avctx, "Bit-depth higher than 16", 0);
+ avpriv_report_missing_feature(avctx, "Bit-depth higher than 16");
return AVERROR_PATCHWELCOME;
} else {
av_log(avctx, AV_LOG_ERROR, "Unknown bit-depth: %d\n",
av_dlog(avctx, "\n");
} else {
- av_log_ask_for_sample(avctx, "Unsupported extradata size\n");
- return AVERROR_INVALIDDATA;
+ avpriv_request_sample(avctx, "Unsupported extradata size");
+ return AVERROR_PATCHWELCOME;
}
/* generic init */
s->num_channels);
return AVERROR_INVALIDDATA;
} else if (s->num_channels > WMALL_MAX_CHANNELS) {
- av_log_ask_for_sample(avctx, "unsupported number of channels\n");
+ avpriv_request_sample(avctx,
+ "More than %d channels", WMALL_MAX_CHANNELS);
return AVERROR_PATCHWELCOME;
}
- avcodec_get_frame_defaults(&s->frame);
- avctx->coded_frame = &s->frame;
+ s->frame = av_frame_alloc();
+ if (!s->frame)
+ return AVERROR(ENOMEM);
+
avctx->channel_layout = channel_mask;
return 0;
}
s->do_arith_coding = get_bits1(&s->gb);
if (s->do_arith_coding) {
- av_log_missing_feature(s->avctx, "Arithmetic coding", 1);
+ avpriv_request_sample(s->avctx, "Arithmetic coding");
return AVERROR_PATCHWELCOME;
}
s->do_ac_filter = get_bits1(&s->gb);
} else if (!s->cdlms[0][0].order) {
av_log(s->avctx, AV_LOG_DEBUG,
"Waiting for seekable tile\n");
- s->frame.nb_samples = 0;
+ av_frame_unref(s->frame);
return -1;
}
s->do_lpc = get_bits1(&s->gb);
if (s->do_lpc) {
decode_lpc(s);
- av_log_ask_for_sample(s->avctx, "Inverse LPC filter not "
- "implemented. Expect wrong output.\n");
+ avpriv_request_sample(s->avctx, "Expect wrong output since "
+ "inverse LPC filter");
}
} else
s->do_lpc = 0;
GetBitContext* gb = &s->gb;
int more_frames = 0, len = 0, i, ret;
- s->frame.nb_samples = s->samples_per_frame;
- if ((ret = s->avctx->get_buffer(s->avctx, &s->frame)) < 0) {
+ s->frame->nb_samples = s->samples_per_frame;
+ if ((ret = ff_get_buffer(s->avctx, s->frame, 0)) < 0) {
/* return an error if no frame could be decoded at all */
av_log(s->avctx, AV_LOG_ERROR,
"not enough space for the output samples\n");
return ret;
}
for (i = 0; i < s->num_channels; i++) {
- s->samples_16[i] = (int16_t *)s->frame.extended_data[i];
- s->samples_32[i] = (int32_t *)s->frame.extended_data[i];
+ s->samples_16[i] = (int16_t *)s->frame->extended_data[i];
+ s->samples_32[i] = (int32_t *)s->frame->extended_data[i];
}
/* get frame length */
buflen = (s->num_saved_bits + len + 8) >> 3;
if (len <= 0 || buflen > MAX_FRAMESIZE) {
- av_log_ask_for_sample(s->avctx, "input buffer too small\n");
+ avpriv_request_sample(s->avctx, "Too small input buffer");
s->packet_loss = 1;
return;
}
int buf_size = avpkt->size;
int num_bits_prev_frame, packet_sequence_number, spliced_packet;
- s->frame.nb_samples = 0;
+ s->frame->nb_samples = 0;
if (s->packet_done || s->packet_loss) {
s->packet_done = 0;
skip_bits(gb, 1); // Skip seekable_frame_in_packet, currently ununused
spliced_packet = get_bits1(gb);
if (spliced_packet)
- av_log_missing_feature(avctx, "Bitstream splicing", 1);
+ avpriv_request_sample(avctx, "Bitstream splicing");
/* get number of bits that need to be added to the previous frame */
num_bits_prev_frame = get_bits(gb, s->log2_frame_size);
save_bits(s, gb, remaining_bits(s, gb), 0);
}
- *(AVFrame *)data = s->frame;
- *got_frame_ptr = s->frame.nb_samples > 0;
+ *got_frame_ptr = s->frame->nb_samples > 0;
+ av_frame_move_ref(data, s->frame);
+
s->packet_offset = get_bits_count(gb) & 7;
return (s->packet_loss) ? AVERROR_INVALIDDATA : get_bits_count(gb) >> 3;
s->frame_offset = 0;
s->next_packet_start = 0;
s->cdlms[0][0].order = 0;
- s->frame.nb_samples = 0;
+ s->frame->nb_samples = 0;
init_put_bits(&s->pb, s->frame_data, MAX_FRAMESIZE);
}
+static av_cold int decode_close(AVCodecContext *avctx)
+{
+ WmallDecodeCtx *s = avctx->priv_data;
+
+ av_frame_free(&s->frame);
+
+ return 0;
+}
+
AVCodec ff_wmalossless_decoder = {
.name = "wmalossless",
.type = AVMEDIA_TYPE_AUDIO,
.id = AV_CODEC_ID_WMALOSSLESS,
.priv_data_size = sizeof(WmallDecodeCtx),
.init = decode_init,
+ .close = decode_close,
.decode = decode_packet,
.flush = flush,
.capabilities = CODEC_CAP_SUBFRAMES | CODEC_CAP_DR1 | CODEC_CAP_DELAY,