X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=libavcodec%2Faura.c;h=b51aedeccf12ae4a38fad2a79213dc6af85c7f82;hb=6fee1b90ce3bf4fbdfde7016e0890057c9000487;hp=9f910212fba1b6add28b3c06e26b1a7829d82fca;hpb=716d413c13981da15323c7a3821860536eefdbbb;p=ffmpeg diff --git a/libavcodec/aura.c b/libavcodec/aura.c index 9f910212fba..b51aedeccf1 100644 --- a/libavcodec/aura.c +++ b/libavcodec/aura.c @@ -24,35 +24,27 @@ */ #include "avcodec.h" +#include "internal.h" #include "libavutil/internal.h" -typedef struct AuraDecodeContext { - AVCodecContext *avctx; - AVFrame frame; -} AuraDecodeContext; - static av_cold int aura_decode_init(AVCodecContext *avctx) { - AuraDecodeContext *s = avctx->priv_data; - - s->avctx = avctx; /* width needs to be divisible by 4 for this codec to work */ if (avctx->width & 0x3) - return -1; + return AVERROR(EINVAL); avctx->pix_fmt = AV_PIX_FMT_YUV422P; return 0; } static int aura_decode_frame(AVCodecContext *avctx, - void *data, int *data_size, + void *data, int *got_frame, AVPacket *pkt) { - AuraDecodeContext *s=avctx->priv_data; - + AVFrame *frame = data; uint8_t *Y, *U, *V; uint8_t val; - int x, y; + int x, y, ret; const uint8_t *buf = pkt->data; /* prediction error tables (make it clear that they are signed values) */ @@ -61,75 +53,57 @@ static int aura_decode_frame(AVCodecContext *avctx, if (pkt->size != 48 + avctx->height * avctx->width) { av_log(avctx, AV_LOG_ERROR, "got a buffer with %d bytes when %d were expected\n", pkt->size, 48 + avctx->height * avctx->width); - return -1; + return AVERROR_INVALIDDATA; } /* pixel data starts 48 bytes in, after 3x16-byte tables */ buf += 48; - if(s->frame.data[0]) - avctx->release_buffer(avctx, &s->frame); - - s->frame.buffer_hints = FF_BUFFER_HINTS_VALID; - s->frame.reference = 0; - if(avctx->get_buffer(avctx, &s->frame) < 0) { + if ((ret = ff_get_buffer(avctx, frame, 0)) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); - return -1; + return ret; } - Y = s->frame.data[0]; - U = s->frame.data[1]; - V = s->frame.data[2]; + Y = frame->data[0]; + U = frame->data[1]; + V = frame->data[2]; /* iterate through each line in the height */ for (y = 0; y < avctx->height; y++) { /* reset predictors */ - val = *buf++; + val = *buf++; U[0] = val & 0xF0; Y[0] = val << 4; - val = *buf++; + val = *buf++; V[0] = val & 0xF0; Y[1] = Y[0] + delta_table[val & 0xF]; - Y += 2; U++; V++; + Y += 2; U++; V++; /* iterate through the remaining pixel groups (4 pixels/group) */ for (x = 1; x < (avctx->width >> 1); x++) { - val = *buf++; + val = *buf++; U[0] = U[-1] + delta_table[val >> 4]; Y[0] = Y[-1] + delta_table[val & 0xF]; - val = *buf++; + val = *buf++; V[0] = V[-1] + delta_table[val >> 4]; Y[1] = Y[ 0] + delta_table[val & 0xF]; - Y += 2; U++; V++; + Y += 2; U++; V++; } - Y += s->frame.linesize[0] - avctx->width; - U += s->frame.linesize[1] - (avctx->width >> 1); - V += s->frame.linesize[2] - (avctx->width >> 1); + Y += frame->linesize[0] - avctx->width; + U += frame->linesize[1] - (avctx->width >> 1); + V += frame->linesize[2] - (avctx->width >> 1); } - *data_size=sizeof(AVFrame); - *(AVFrame*)data= s->frame; + *got_frame = 1; return pkt->size; } -static av_cold int aura_decode_end(AVCodecContext *avctx) -{ - AuraDecodeContext *s = avctx->priv_data; - - if (s->frame.data[0]) - avctx->release_buffer(avctx, &s->frame); - - return 0; -} - AVCodec ff_aura2_decoder = { .name = "aura2", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_AURA2, - .priv_data_size = sizeof(AuraDecodeContext), .init = aura_decode_init, - .close = aura_decode_end, .decode = aura_decode_frame, .capabilities = CODEC_CAP_DR1, .long_name = NULL_IF_CONFIG_SMALL("Auravision Aura 2"),