#include "avcodec.h"
#include "bytestream.h"
+#include "internal.h"
static av_cold int encode_init(AVCodecContext *avctx)
{
return AVERROR(EINVAL);
}
- if (avctx->pix_fmt != PIX_FMT_YUV422P10) {
- av_log(avctx, AV_LOG_ERROR, "v210 needs YUV422P10\n");
- return -1;
- }
-
if (avctx->bits_per_raw_sample != 10)
av_log(avctx, AV_LOG_WARNING, "bits per raw sample: %d != 10-bit\n",
avctx->bits_per_raw_sample);
return 0;
}
-static int encode_frame(AVCodecContext *avctx, unsigned char *buf,
- int buf_size, void *data)
+static int encode_frame(AVCodecContext *avctx, AVPacket *pkt,
+ const AVFrame *pic, int *got_packet)
{
- const AVFrame *pic = data;
- int aligned_width = FFALIGN(avctx->width, 48);
+ int aligned_width = ((avctx->width + 47) / 48) * 48;
int stride = aligned_width * 8 / 3;
- int h, w;
+ int line_padding = stride - ((avctx->width * 8 + 11) / 12) * 4;
+ int h, w, ret;
const uint16_t *y = (const uint16_t*)pic->data[0];
const uint16_t *u = (const uint16_t*)pic->data[1];
const uint16_t *v = (const uint16_t*)pic->data[2];
- uint8_t *p = buf;
- uint8_t *pdst = buf;
+ PutByteContext p;
- if (buf_size < avctx->height * stride) {
- av_log(avctx, AV_LOG_ERROR, "output buffer too small\n");
- return AVERROR(ENOMEM);
+ if ((ret = ff_alloc_packet(pkt, avctx->height * stride)) < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Error getting output packet.\n");
+ return ret;
}
+ bytestream2_init_writer(&p, pkt->data, pkt->size);
+
#define CLIP(v) av_clip(v, 4, 1019)
#define WRITE_PIXELS(a, b, c) \
val = CLIP(*a++); \
val |= (CLIP(*b++) << 10) | \
(CLIP(*c++) << 20); \
- bytestream_put_le32(&p, val); \
+ bytestream2_put_le32u(&p, val); \
} while (0)
for (h = 0; h < avctx->height; h++) {
val = CLIP(*y++);
if (w == avctx->width - 2)
- bytestream_put_le32(&p, val);
+ bytestream2_put_le32u(&p, val);
}
if (w < avctx->width - 3) {
val |= (CLIP(*u++) << 10) | (CLIP(*y++) << 20);
- bytestream_put_le32(&p, val);
+ bytestream2_put_le32u(&p, val);
val = CLIP(*v++) | (CLIP(*y++) << 10);
- bytestream_put_le32(&p, val);
+ bytestream2_put_le32u(&p, val);
}
- pdst += stride;
- memset(p, 0, pdst - p);
- p = pdst;
+ bytestream2_set_buffer(&p, 0, line_padding);
+
y += pic->linesize[0] / 2 - avctx->width;
u += pic->linesize[1] / 2 - avctx->width / 2;
v += pic->linesize[2] / 2 - avctx->width / 2;
}
- return p - buf;
+ pkt->flags |= AV_PKT_FLAG_KEY;
+ *got_packet = 1;
+ return 0;
}
static av_cold int encode_close(AVCodecContext *avctx)
.type = AVMEDIA_TYPE_VIDEO,
.id = CODEC_ID_V210,
.init = encode_init,
- .encode = encode_frame,
+ .encode2 = encode_frame,
.close = encode_close,
- .pix_fmts = (const enum PixelFormat[]){PIX_FMT_YUV422P10, PIX_FMT_NONE},
- .long_name = NULL_IF_CONFIG_SMALL("Uncompressed 4:2:2 10-bit"),
+ .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV422P10, PIX_FMT_NONE },
+ .long_name = NULL_IF_CONFIG_SMALL("Uncompressed 4:2:2 10-bit"),
};