#include "libavutil/pixdesc.h"
#include "avcodec.h"
-#include "dsputil.h"
+#include "idctdsp.h"
#include "internal.h"
+#include "jpegtables.h"
+#include "mjpegenc_common.h"
#include "mpegvideo.h"
#include "mjpeg.h"
#include "mjpegenc.h"
typedef struct LJpegEncContext {
- DSPContext dsp;
+ IDCTDSPContext idsp;
ScanTable scantable;
uint16_t matrix[64];
uint16_t (*scratch)[4];
} LJpegEncContext;
+static int ljpeg_encode_bgr(AVCodecContext *avctx, PutBitContext *pb,
+ const AVFrame *frame)
+{
+ LJpegEncContext *s = avctx->priv_data;
+ const int width = frame->width;
+ const int height = frame->height;
+ const int linesize = frame->linesize[0];
+ uint16_t (*buffer)[4] = s->scratch;
+ const int predictor = avctx->prediction_method+1;
+ int left[3], top[3], topleft[3];
+ int x, y, i;
+
+ for (i = 0; i < 3; i++)
+ buffer[0][i] = 1 << (9 - 1);
+
+ for (y = 0; y < height; y++) {
+ const int modified_predictor = y ? predictor : 1;
+ uint8_t *ptr = frame->data[0] + (linesize * y);
+
+ if (pb->buf_end - pb->buf - (put_bits_count(pb) >> 3) < width * 3 * 3) {
+ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
+ return -1;
+ }
+
+ for (i = 0; i < 3; i++)
+ top[i]= left[i]= topleft[i]= buffer[0][i];
+
+ for (x = 0; x < width; x++) {
+ buffer[x][1] = ptr[3 * x + 0] - ptr[3 * x + 1] + 0x100;
+ buffer[x][2] = ptr[3 * x + 2] - ptr[3 * x + 1] + 0x100;
+ buffer[x][0] = (ptr[3 * x + 0] + 2 * ptr[3 * x + 1] + ptr[3 * x + 2]) >> 2;
+
+ for (i = 0; i < 3; i++) {
+ int pred, diff;
+
+ PREDICT(pred, topleft[i], top[i], left[i], modified_predictor);
+
+ topleft[i] = top[i];
+ top[i] = buffer[x+1][i];
+
+ left[i] = buffer[x][i];
+
+ diff = ((left[i] - pred + 0x100) & 0x1FF) - 0x100;
+
+ if (i == 0)
+ ff_mjpeg_encode_dc(pb, diff, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
+ else
+ ff_mjpeg_encode_dc(pb, diff, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
+ }
+ }
+ }
+
+ return 0;
+}
+
+static inline void ljpeg_encode_yuv_mb(LJpegEncContext *s, PutBitContext *pb,
+ const AVFrame *frame, int predictor,
+ int mb_x, int mb_y)
+{
+ int i;
+
+ if (mb_x == 0 || mb_y == 0) {
+ for (i = 0; i < 3; i++) {
+ uint8_t *ptr;
+ int x, y, h, v, linesize;
+ h = s->hsample[i];
+ v = s->vsample[i];
+ linesize = frame->linesize[i];
+
+ for (y = 0; y < v; y++) {
+ for (x = 0; x < h; x++) {
+ int pred;
+
+ ptr = frame->data[i] + (linesize * (v * mb_y + y)) + (h * mb_x + x); //FIXME optimize this crap
+ if (y == 0 && mb_y == 0) {
+ if (x == 0 && mb_x == 0)
+ pred = 128;
+ else
+ pred = ptr[-1];
+ } else {
+ if (x == 0 && mb_x == 0) {
+ pred = ptr[-linesize];
+ } else {
+ PREDICT(pred, ptr[-linesize - 1], ptr[-linesize],
+ ptr[-1], predictor);
+ }
+ }
+
+ if (i == 0)
+ ff_mjpeg_encode_dc(pb, *ptr - pred, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
+ else
+ ff_mjpeg_encode_dc(pb, *ptr - pred, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
+ }
+ }
+ }
+ } else {
+ for (i = 0; i < 3; i++) {
+ uint8_t *ptr;
+ int x, y, h, v, linesize;
+ h = s->hsample[i];
+ v = s->vsample[i];
+ linesize = frame->linesize[i];
+
+ for (y = 0; y < v; y++) {
+ for (x = 0; x < h; x++) {
+ int pred;
+
+ ptr = frame->data[i] + (linesize * (v * mb_y + y)) + (h * mb_x + x); //FIXME optimize this crap
+ PREDICT(pred, ptr[-linesize - 1], ptr[-linesize], ptr[-1], predictor);
+
+ if (i == 0)
+ ff_mjpeg_encode_dc(pb, *ptr - pred, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
+ else
+ ff_mjpeg_encode_dc(pb, *ptr - pred, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
+ }
+ }
+ }
+ }
+}
+
+static int ljpeg_encode_yuv(AVCodecContext *avctx, PutBitContext *pb,
+ const AVFrame *frame)
+{
+ const int predictor = avctx->prediction_method + 1;
+ LJpegEncContext *s = avctx->priv_data;
+ const int mb_width = (avctx->width + s->hsample[0] - 1) / s->hsample[0];
+ const int mb_height = (avctx->height + s->vsample[0] - 1) / s->vsample[0];
+ int mb_x, mb_y;
+
+ for (mb_y = 0; mb_y < mb_height; mb_y++) {
+ if (pb->buf_end - pb->buf - (put_bits_count(pb) >> 3) <
+ mb_width * 4 * 3 * s->hsample[0] * s->vsample[0]) {
+ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
+ return -1;
+ }
+
+ for (mb_x = 0; mb_x < mb_width; mb_x++)
+ ljpeg_encode_yuv_mb(s, pb, frame, predictor, mb_x, mb_y);
+ }
+
+ return 0;
+}
+
static int ljpeg_encode_frame(AVCodecContext *avctx, AVPacket *pkt,
const AVFrame *pict, int *got_packet)
{
PutBitContext pb;
const int width = avctx->width;
const int height = avctx->height;
- const int predictor= avctx->prediction_method+1;
const int mb_width = (width + s->hsample[0] - 1) / s->hsample[0];
const int mb_height = (height + s->vsample[0] - 1) / s->vsample[0];
int max_pkt_size = FF_MIN_BUFFER_SIZE;
header_bits = put_bits_count(&pb);
- if (avctx->pix_fmt == AV_PIX_FMT_BGR24) {
- int x, y, i;
- const int linesize = pict->linesize[0];
- uint16_t (*buffer)[4] = s->scratch;
- int left[3], top[3], topleft[3];
-
- for(i=0; i<3; i++){
- buffer[0][i]= 1 << (9 - 1);
- }
-
- for(y = 0; y < height; y++) {
- const int modified_predictor= y ? predictor : 1;
- uint8_t *ptr = pict->data[0] + (linesize * y);
-
- if(pb.buf_end - pb.buf - (put_bits_count(&pb) >> 3) < width * 3 * 3) {
- av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
- return -1;
- }
-
- for(i=0; i<3; i++){
- top[i]= left[i]= topleft[i]= buffer[0][i];
- }
- for(x = 0; x < width; x++) {
- buffer[x][1] = ptr[3 * x + 0] - ptr[3 * x + 1] + 0x100;
- buffer[x][2] = ptr[3 * x + 2] - ptr[3 * x + 1] + 0x100;
- buffer[x][0] = (ptr[3 * x + 0] + 2 * ptr[3 * x + 1] + ptr[3 * x + 2]) >> 2;
-
- for(i=0;i<3;i++) {
- int pred, diff;
-
- PREDICT(pred, topleft[i], top[i], left[i], modified_predictor);
-
- topleft[i]= top[i];
- top[i]= buffer[x+1][i];
-
- left[i]= buffer[x][i];
-
- diff= ((left[i] - pred + 0x100)&0x1FF) - 0x100;
-
- if(i==0)
- ff_mjpeg_encode_dc(&pb, diff, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
- else
- ff_mjpeg_encode_dc(&pb, diff, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
- }
- }
- }
- }else{
- int mb_x, mb_y, i;
-
- for(mb_y = 0; mb_y < mb_height; mb_y++) {
- if (pb.buf_end - pb.buf - (put_bits_count(&pb) >> 3) <
- mb_width * 4 * 3 * s->hsample[0] * s->vsample[0]) {
- av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
- return -1;
- }
- for(mb_x = 0; mb_x < mb_width; mb_x++) {
- if(mb_x==0 || mb_y==0){
- for(i=0;i<3;i++) {
- uint8_t *ptr;
- int x, y, h, v, linesize;
- h = s->hsample[i];
- v = s->vsample[i];
- linesize = pict->linesize[i];
-
- for(y=0; y<v; y++){
- for(x=0; x<h; x++){
- int pred;
-
- ptr = pict->data[i] + (linesize * (v * mb_y + y)) + (h * mb_x + x); //FIXME optimize this crap
- if(y==0 && mb_y==0){
- if(x==0 && mb_x==0){
- pred= 128;
- }else{
- pred= ptr[-1];
- }
- }else{
- if(x==0 && mb_x==0){
- pred= ptr[-linesize];
- }else{
- PREDICT(pred, ptr[-linesize-1], ptr[-linesize], ptr[-1], predictor);
- }
- }
-
- if(i==0)
- ff_mjpeg_encode_dc(&pb, *ptr - pred, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
- else
- ff_mjpeg_encode_dc(&pb, *ptr - pred, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
- }
- }
- }
- }else{
- for(i=0;i<3;i++) {
- uint8_t *ptr;
- int x, y, h, v, linesize;
- h = s->hsample[i];
- v = s->vsample[i];
- linesize = pict->linesize[i];
-
- for(y=0; y<v; y++){
- for(x=0; x<h; x++){
- int pred;
-
- ptr = pict->data[i] + (linesize * (v * mb_y + y)) + (h * mb_x + x); //FIXME optimize this crap
- PREDICT(pred, ptr[-linesize-1], ptr[-linesize], ptr[-1], predictor);
-
- if(i==0)
- ff_mjpeg_encode_dc(&pb, *ptr - pred, s->huff_size_dc_luminance, s->huff_code_dc_luminance); //FIXME ugly
- else
- ff_mjpeg_encode_dc(&pb, *ptr - pred, s->huff_size_dc_chrominance, s->huff_code_dc_chrominance);
- }
- }
- }
- }
- }
- }
- }
+ if (avctx->pix_fmt == AV_PIX_FMT_BGR24)
+ ret = ljpeg_encode_bgr(avctx, &pb, pict);
+ else
+ ret = ljpeg_encode_yuv(avctx, &pb, pict);
+ if (ret < 0)
+ return ret;
emms_c();
{
LJpegEncContext *s = avctx->priv_data;
- av_frame_free(&avctx->coded_frame);
av_freep(&s->scratch);
return 0;
if ((avctx->pix_fmt == AV_PIX_FMT_YUV420P ||
avctx->pix_fmt == AV_PIX_FMT_YUV422P ||
- avctx->pix_fmt == AV_PIX_FMT_YUV444P) &&
+ avctx->pix_fmt == AV_PIX_FMT_YUV444P ||
+ avctx->color_range == AVCOL_RANGE_MPEG) &&
avctx->strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL) {
av_log(avctx, AV_LOG_ERROR,
"Limited range YUV is non-standard, set strict_std_compliance to "
return AVERROR(EINVAL);
}
- avctx->coded_frame = av_frame_alloc();
- if (!avctx->coded_frame)
- return AVERROR(ENOMEM);
-
+#if FF_API_CODED_FRAME
+FF_DISABLE_DEPRECATION_WARNINGS
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I;
avctx->coded_frame->key_frame = 1;
+FF_ENABLE_DEPRECATION_WARNINGS
+#endif
s->scratch = av_malloc_array(avctx->width + 1, sizeof(*s->scratch));
- ff_dsputil_init(&s->dsp, avctx);
- ff_init_scantable(s->dsp.idct_permutation, &s->scantable, ff_zigzag_direct);
+ ff_idctdsp_init(&s->idsp, avctx);
+ ff_init_scantable(s->idsp.idct_permutation, &s->scantable,
+ ff_zigzag_direct);
av_pix_fmt_get_chroma_sub_sample(avctx->pix_fmt, &chroma_h_shift,
&chroma_v_shift);
AV_PIX_FMT_BGR24,
AV_PIX_FMT_YUV420P,
AV_PIX_FMT_YUV422P,
- AV_PIX_FMT_YUVJ444P,
+ AV_PIX_FMT_YUV444P,
AV_PIX_FMT_NONE },
};