typedef struct ADPCMDecodeContext {
ADPCMChannelStatus status[14];
int vqa_version; /**< VQA version. Used for ADPCM_IMA_WS */
- int has_status;
+ int has_status; /**< Status flag. Reset to 0 after a flush. */
} ADPCMDecodeContext;
+static void adpcm_flush(AVCodecContext *avctx);
+
static av_cold int adpcm_decode_init(AVCodecContext * avctx)
{
ADPCMDecodeContext *c = avctx->priv_data;
unsigned int min_channels = 1;
unsigned int max_channels = 2;
+ adpcm_flush(avctx);
+
switch(avctx->codec->id) {
- case AV_CODEC_ID_ADPCM_IMA_CUNNING:
+ case AV_CODEC_ID_ADPCM_IMA_AMV:
max_channels = 1;
break;
case AV_CODEC_ID_ADPCM_DTK:
min_channels = 2;
max_channels = 8;
if (avctx->channels & 1) {
- avpriv_request_sample(avctx, "channel count %d\n", avctx->channels);
+ avpriv_request_sample(avctx, "channel count %d", avctx->channels);
return AVERROR_PATCHWELCOME;
}
break;
case AV_CODEC_ID_ADPCM_PSX:
max_channels = 8;
+ if (avctx->channels <= 0 || avctx->block_align % (16 * avctx->channels))
+ return AVERROR_INVALIDDATA;
break;
case AV_CODEC_ID_ADPCM_IMA_DAT4:
case AV_CODEC_ID_ADPCM_THP:
}
switch(avctx->codec->id) {
- case AV_CODEC_ID_ADPCM_CT:
- c->status[0].step = c->status[1].step = 511;
- break;
case AV_CODEC_ID_ADPCM_IMA_WAV:
if (avctx->bits_per_coded_sample < 2 || avctx->bits_per_coded_sample > 5)
return AVERROR_INVALIDDATA;
break;
- case AV_CODEC_ID_ADPCM_IMA_APC:
- if (avctx->extradata && avctx->extradata_size >= 8) {
- c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata ), 18);
- c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 4), 18);
- }
- break;
- case AV_CODEC_ID_ADPCM_IMA_APM:
- if (avctx->extradata) {
- if (avctx->extradata_size >= 28) {
- c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 16), 18);
- c->status[0].step_index = av_clip(AV_RL32(avctx->extradata + 20), 0, 88);
- c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 4), 18);
- c->status[1].step_index = av_clip(AV_RL32(avctx->extradata + 8), 0, 88);
- } else if (avctx->extradata_size >= 16) {
- c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 0), 18);
- c->status[0].step_index = av_clip(AV_RL32(avctx->extradata + 4), 0, 88);
- c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 8), 18);
- c->status[1].step_index = av_clip(AV_RL32(avctx->extradata + 12), 0, 88);
- }
- }
- break;
- case AV_CODEC_ID_ADPCM_IMA_WS:
- if (avctx->extradata && avctx->extradata_size >= 2)
- c->vqa_version = AV_RL16(avctx->extradata);
- break;
case AV_CODEC_ID_ADPCM_ARGO:
if (avctx->bits_per_coded_sample != 4 || avctx->block_align != 17 * avctx->channels)
return AVERROR_INVALIDDATA;
switch (avctx->codec->id) {
case AV_CODEC_ID_ADPCM_AICA:
+ case AV_CODEC_ID_ADPCM_IMA_CUNNING:
case AV_CODEC_ID_ADPCM_IMA_DAT4:
case AV_CODEC_ID_ADPCM_IMA_QT:
case AV_CODEC_ID_ADPCM_IMA_WAV:
default:
avctx->sample_fmt = AV_SAMPLE_FMT_S16;
}
-
return 0;
}
case AV_CODEC_ID_ADPCM_IMA_DAT4:
case AV_CODEC_ID_ADPCM_IMA_MOFLEX:
case AV_CODEC_ID_ADPCM_IMA_ISS: header_size = 4 * ch; break;
- case AV_CODEC_ID_ADPCM_IMA_AMV: header_size = 8; break;
case AV_CODEC_ID_ADPCM_IMA_SMJPEG: header_size = 4 * ch; break;
}
if (header_size > 0)
/* more complex formats */
switch (avctx->codec->id) {
+ case AV_CODEC_ID_ADPCM_IMA_AMV:
+ bytestream2_skip(gb, 4);
+ has_coded_samples = 1;
+ *coded_samples = bytestream2_get_le32u(gb);
+ nb_samples = FFMIN((buf_size - 8) * 2, *coded_samples);
+ bytestream2_seek(gb, -8, SEEK_CUR);
+ break;
case AV_CODEC_ID_ADPCM_EA:
has_coded_samples = 1;
*coded_samples = bytestream2_get_le32(gb);
}
break;
case AV_CODEC_ID_ADPCM_IMA_CUNNING:
- for (n = 0; n < nb_samples / 2; n++) {
- int v = bytestream2_get_byteu(&gb);
- *samples++ = adpcm_ima_cunning_expand_nibble(&c->status[0], v & 0x0F);
- *samples++ = adpcm_ima_cunning_expand_nibble(&c->status[0], v >> 4);
+ for (channel = 0; channel < avctx->channels; channel++) {
+ int16_t *smp = samples_p[channel];
+ for (n = 0; n < nb_samples / 2; n++) {
+ int v = bytestream2_get_byteu(&gb);
+ *smp++ = adpcm_ima_cunning_expand_nibble(&c->status[channel], v & 0x0F);
+ *smp++ = adpcm_ima_cunning_expand_nibble(&c->status[channel], v >> 4);
+ }
}
break;
case AV_CODEC_ID_ADPCM_IMA_OKI:
for (n = nb_samples / 2; n > 0; n--) {
int v = bytestream2_get_byteu(&gb);
- *smp++ = adpcm_ima_expand_nibble(&c->status[channel], v >> 4 , 3);
*smp++ = adpcm_ima_expand_nibble(&c->status[channel], v & 0x0F, 3);
+ *smp++ = adpcm_ima_expand_nibble(&c->status[channel], v >> 4 , 3);
}
}
} else {
for (n = nb_samples / 2; n > 0; n--) {
for (channel = 0; channel < avctx->channels; channel++) {
int v = bytestream2_get_byteu(&gb);
- *samples++ = adpcm_ima_expand_nibble(&c->status[channel], v >> 4 , 3);
- samples[st] = adpcm_ima_expand_nibble(&c->status[channel], v & 0x0F, 3);
+ *samples++ = adpcm_ima_expand_nibble(&c->status[channel], v & 0x0F, 3);
+ samples[st] = adpcm_ima_expand_nibble(&c->status[channel], v >> 4 , 3);
}
samples += avctx->channels;
}
}
break;
case AV_CODEC_ID_ADPCM_IMA_AMV:
+ av_assert0(avctx->channels == 1);
+
+ /*
+ * Header format:
+ * int16_t predictor;
+ * uint8_t step_index;
+ * uint8_t reserved;
+ * uint32_t frame_size;
+ *
+ * Some implementations have step_index as 16-bits, but others
+ * only use the lower 8 and store garbage in the upper 8.
+ */
c->status[0].predictor = sign_extend(bytestream2_get_le16u(&gb), 16);
c->status[0].step_index = bytestream2_get_byteu(&gb);
bytestream2_skipu(&gb, 5);
return AVERROR_INVALIDDATA;
}
- for (n = nb_samples >> (1 - st); n > 0; n--) {
+ for (n = nb_samples >> 1; n > 0; n--) {
int v = bytestream2_get_byteu(&gb);
*samples++ = adpcm_ima_expand_nibble(&c->status[0], v >> 4, 3);
*samples++ = adpcm_ima_expand_nibble(&c->status[0], v & 0xf, 3);
}
+
+ if (nb_samples & 1) {
+ int v = bytestream2_get_byteu(&gb);
+ *samples++ = adpcm_ima_expand_nibble(&c->status[0], v >> 4, 3);
+
+ if (v & 0x0F) {
+ /* Holds true on all the http://samples.mplayerhq.hu/amv samples. */
+ av_log(avctx, AV_LOG_WARNING, "Last nibble set on packet with odd sample count.\n");
+ av_log(avctx, AV_LOG_WARNING, "Sample will be skipped.\n");
+ }
+ }
break;
case AV_CODEC_ID_ADPCM_IMA_SMJPEG:
for (i = 0; i < avctx->channels; i++) {
}
break;
case AV_CODEC_ID_ADPCM_AICA:
- if (!c->has_status) {
- for (channel = 0; channel < avctx->channels; channel++)
- c->status[channel].step = 0;
- c->has_status = 1;
- }
for (channel = 0; channel < avctx->channels; channel++) {
samples = samples_p[channel];
for (n = nb_samples >> 1; n > 0; n--) {
int nb_samples_per_block = 28 * FFMAX(avctx->block_align, 16 * avctx->channels) / (16 * avctx->channels);
for (channel = 0; channel < avctx->channels; channel++) {
samples = samples_p[channel] + block * nb_samples_per_block;
+ av_assert0((block + 1) * nb_samples_per_block <= nb_samples);
/* Read in every sample for this channel. */
for (i = 0; i < nb_samples_per_block / 28; i++) {
}
break;
case AV_CODEC_ID_ADPCM_ZORK:
- if (!c->has_status) {
- for (channel = 0; channel < avctx->channels; channel++) {
- c->status[channel].predictor = 0;
- c->status[channel].step_index = 0;
- }
- c->has_status = 1;
- }
for (n = 0; n < nb_samples * avctx->channels; n++) {
int v = bytestream2_get_byteu(&gb);
*samples++ = adpcm_zork_expand_nibble(&c->status[n % avctx->channels], v);
static void adpcm_flush(AVCodecContext *avctx)
{
ADPCMDecodeContext *c = avctx->priv_data;
- c->has_status = 0;
+
+ /* Just nuke the entire state and re-init. */
+ memset(c, 0, sizeof(ADPCMDecodeContext));
+
+ switch(avctx->codec_id) {
+ case AV_CODEC_ID_ADPCM_CT:
+ c->status[0].step = c->status[1].step = 511;
+ break;
+
+ case AV_CODEC_ID_ADPCM_IMA_APC:
+ if (avctx->extradata && avctx->extradata_size >= 8) {
+ c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata ), 18);
+ c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 4), 18);
+ }
+ break;
+
+ case AV_CODEC_ID_ADPCM_IMA_APM:
+ if (avctx->extradata) {
+ if (avctx->extradata_size >= 28) {
+ c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 16), 18);
+ c->status[0].step_index = av_clip(AV_RL32(avctx->extradata + 20), 0, 88);
+ c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 4), 18);
+ c->status[1].step_index = av_clip(AV_RL32(avctx->extradata + 8), 0, 88);
+ } else if (avctx->extradata_size >= 16) {
+ c->status[0].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 0), 18);
+ c->status[0].step_index = av_clip(AV_RL32(avctx->extradata + 4), 0, 88);
+ c->status[1].predictor = av_clip_intp2(AV_RL32(avctx->extradata + 8), 18);
+ c->status[1].step_index = av_clip(AV_RL32(avctx->extradata + 12), 0, 88);
+ }
+ }
+ break;
+
+ case AV_CODEC_ID_ADPCM_IMA_WS:
+ if (avctx->extradata && avctx->extradata_size >= 2)
+ c->vqa_version = AV_RL16(avctx->extradata);
+ break;
+ default:
+ /* Other codecs may want to handle this during decoding. */
+ c->has_status = 0;
+ return;
+ }
+
+ c->has_status = 1;
}
AV_SAMPLE_FMT_NONE };
#define ADPCM_DECODER(id_, sample_fmts_, name_, long_name_) \
-AVCodec ff_ ## name_ ## _decoder = { \
+const AVCodec ff_ ## name_ ## _decoder = { \
.name = #name_, \
.long_name = NULL_IF_CONFIG_SMALL(long_name_), \
.type = AVMEDIA_TYPE_AUDIO, \
.flush = adpcm_flush, \
.capabilities = AV_CODEC_CAP_DR1, \
.sample_fmts = sample_fmts_, \
+ .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE, \
}
/* Note: Do not forget to add new entries to the Makefile as well. */
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_AMV, sample_fmts_s16, adpcm_ima_amv, "ADPCM IMA AMV");
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_APC, sample_fmts_s16, adpcm_ima_apc, "ADPCM IMA CRYO APC");
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_APM, sample_fmts_s16, adpcm_ima_apm, "ADPCM IMA Ubisoft APM");
-ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_CUNNING, sample_fmts_s16, adpcm_ima_cunning, "ADPCM IMA Cunning Developments");
+ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_CUNNING, sample_fmts_s16p, adpcm_ima_cunning, "ADPCM IMA Cunning Developments");
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_DAT4, sample_fmts_s16, adpcm_ima_dat4, "ADPCM IMA Eurocom DAT4");
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_DK3, sample_fmts_s16, adpcm_ima_dk3, "ADPCM IMA Duck DK3");
ADPCM_DECODER(AV_CODEC_ID_ADPCM_IMA_DK4, sample_fmts_s16, adpcm_ima_dk4, "ADPCM IMA Duck DK4");