/*
* PNM image format
- * Copyright (c) 2002, 2003 Fabrice Bellard.
+ * Copyright (c) 2002, 2003 Fabrice Bellard
*
- * This file is part of FFmpeg.
+ * This file is part of Libav.
*
- * FFmpeg is free software; you can redistribute it and/or
+ * Libav is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
- * FFmpeg is distributed in the hope that it will be useful,
+ * Libav is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
+ * License along with Libav; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
+
+#include "libavutil/pixdesc.h"
#include "avcodec.h"
#include "bytestream.h"
-#include "pnm.h"
-
-
-static av_cold int common_init(AVCodecContext *avctx){
- PNMContext *s = avctx->priv_data;
-
- avcodec_get_frame_defaults((AVFrame*)&s->picture);
- avctx->coded_frame= (AVFrame*)&s->picture;
-
- return 0;
-}
+#include "internal.h"
-static int pnm_decode_frame(AVCodecContext *avctx,
- void *data, int *data_size,
- const uint8_t *buf, int buf_size)
+static int pnm_encode_frame(AVCodecContext *avctx, AVPacket *pkt,
+ const AVFrame *pict, int *got_packet)
{
- PNMContext * const s = avctx->priv_data;
- AVFrame *picture = data;
- AVFrame * const p= (AVFrame*)&s->picture;
- int i, n, linesize, h, upgrade = 0;
- unsigned char *ptr;
-
- s->bytestream_start=
- s->bytestream= buf;
- s->bytestream_end= buf + buf_size;
-
- if(ff_pnm_decode_header(avctx, s) < 0)
- return -1;
-
- if(p->data[0])
- avctx->release_buffer(avctx, p);
-
- p->reference= 0;
- if(avctx->get_buffer(avctx, p) < 0){
- av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
- return -1;
- }
- p->pict_type= FF_I_TYPE;
- p->key_frame= 1;
-
- switch(avctx->pix_fmt) {
- default:
- return -1;
- case PIX_FMT_RGB24:
- n = avctx->width * 3;
- goto do_read;
- case PIX_FMT_GRAY8:
- n = avctx->width;
- if (s->maxval < 255)
- upgrade = 1;
- goto do_read;
- case PIX_FMT_GRAY16BE:
- case PIX_FMT_GRAY16LE:
- n = avctx->width * 2;
- if (s->maxval < 65535)
- upgrade = 2;
- goto do_read;
- case PIX_FMT_MONOWHITE:
- case PIX_FMT_MONOBLACK:
- n = (avctx->width + 7) >> 3;
- do_read:
- ptr = p->data[0];
- linesize = p->linesize[0];
- if(s->bytestream + n*avctx->height > s->bytestream_end)
- return -1;
- for(i = 0; i < avctx->height; i++) {
- if (!upgrade)
- memcpy(ptr, s->bytestream, n);
- else if (upgrade == 1) {
- unsigned int j, f = (255*128 + s->maxval/2) / s->maxval;
- for (j=0; j<n; j++)
- ptr[j] = (s->bytestream[j] * f + 64) >> 7;
- } else if (upgrade == 2) {
- unsigned int j, v, f = (65535*32768 + s->maxval/2) / s->maxval;
- for (j=0; j<n/2; j++) {
- v = be2me_16(((uint16_t *)s->bytestream)[j]);
- ((uint16_t *)ptr)[j] = (v * f + 16384) >> 15;
- }
- }
- s->bytestream += n;
- ptr += linesize;
- }
- break;
- case PIX_FMT_YUV420P:
- {
- unsigned char *ptr1, *ptr2;
-
- n = avctx->width;
- ptr = p->data[0];
- linesize = p->linesize[0];
- if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end)
- return -1;
- for(i = 0; i < avctx->height; i++) {
- memcpy(ptr, s->bytestream, n);
- s->bytestream += n;
- ptr += linesize;
- }
- ptr1 = p->data[1];
- ptr2 = p->data[2];
- n >>= 1;
- h = avctx->height >> 1;
- for(i = 0; i < h; i++) {
- memcpy(ptr1, s->bytestream, n);
- s->bytestream += n;
- memcpy(ptr2, s->bytestream, n);
- s->bytestream += n;
- ptr1 += p->linesize[1];
- ptr2 += p->linesize[2];
- }
- }
- break;
- case PIX_FMT_RGB32:
- ptr = p->data[0];
- linesize = p->linesize[0];
- if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end)
- return -1;
- for(i = 0; i < avctx->height; i++) {
- int j, r, g, b, a;
-
- for(j = 0;j < avctx->width; j++) {
- r = *s->bytestream++;
- g = *s->bytestream++;
- b = *s->bytestream++;
- a = *s->bytestream++;
- ((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b;
- }
- ptr += linesize;
- }
- break;
- }
- *picture= *(AVFrame*)&s->picture;
- *data_size = sizeof(AVPicture);
-
- return s->bytestream - s->bytestream_start;
-}
-
-static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
- PNMContext *s = avctx->priv_data;
- AVFrame *pict = data;
- AVFrame * const p= (AVFrame*)&s->picture;
- int i, h, h1, c, n, linesize;
+ uint8_t *bytestream, *bytestream_start, *bytestream_end;
+ const AVFrame * const p = pict;
+ int i, h, h1, c, n, linesize, ret;
uint8_t *ptr, *ptr1, *ptr2;
- if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
+ if ((ret = ff_alloc_packet(pkt, avpicture_get_size(avctx->pix_fmt,
+ avctx->width,
+ avctx->height) + 200)) < 0) {
av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
- return -1;
+ return ret;
}
- *p = *pict;
- p->pict_type= FF_I_TYPE;
- p->key_frame= 1;
-
- s->bytestream_start=
- s->bytestream= outbuf;
- s->bytestream_end= outbuf+buf_size;
+ bytestream_start =
+ bytestream = pkt->data;
+ bytestream_end = pkt->data + pkt->size;
- h = avctx->height;
+ h = avctx->height;
h1 = h;
- switch(avctx->pix_fmt) {
- case PIX_FMT_MONOWHITE:
- c = '4';
- n = (avctx->width + 7) >> 3;
+ switch (avctx->pix_fmt) {
+ case AV_PIX_FMT_MONOWHITE:
+ c = '4';
+ n = (avctx->width + 7) >> 3;
break;
- case PIX_FMT_GRAY8:
- c = '5';
- n = avctx->width;
+ case AV_PIX_FMT_GRAY8:
+ c = '5';
+ n = avctx->width;
break;
- case PIX_FMT_GRAY16BE:
- c = '5';
- n = avctx->width * 2;
+ case AV_PIX_FMT_GRAY16BE:
+ c = '5';
+ n = avctx->width * 2;
break;
- case PIX_FMT_RGB24:
- c = '6';
- n = avctx->width * 3;
+ case AV_PIX_FMT_RGB24:
+ c = '6';
+ n = avctx->width * 3;
break;
- case PIX_FMT_YUV420P:
- c = '5';
- n = avctx->width;
+ case AV_PIX_FMT_RGB48BE:
+ c = '6';
+ n = avctx->width * 6;
+ break;
+ case AV_PIX_FMT_YUV420P:
+ c = '5';
+ n = avctx->width;
+ h1 = (h * 3) / 2;
+ break;
+ case AV_PIX_FMT_YUV420P16BE:
+ c = '5';
+ n = avctx->width * 2;
h1 = (h * 3) / 2;
break;
default:
return -1;
}
- snprintf(s->bytestream, s->bytestream_end - s->bytestream,
- "P%c\n%d %d\n",
- c, avctx->width, h1);
- s->bytestream += strlen(s->bytestream);
- if (avctx->pix_fmt != PIX_FMT_MONOWHITE) {
- snprintf(s->bytestream, s->bytestream_end - s->bytestream,
- "%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE) ? 255 : 65535);
- s->bytestream += strlen(s->bytestream);
+ snprintf(bytestream, bytestream_end - bytestream,
+ "P%c\n%d %d\n", c, avctx->width, h1);
+ bytestream += strlen(bytestream);
+ if (avctx->pix_fmt != AV_PIX_FMT_MONOWHITE) {
+ int maxdepth = (1 << av_pix_fmt_desc_get(avctx->pix_fmt)->comp[0].depth) - 1;
+ snprintf(bytestream, bytestream_end - bytestream,
+ "%d\n", maxdepth);
+ bytestream += strlen(bytestream);
}
- ptr = p->data[0];
+ ptr = p->data[0];
linesize = p->linesize[0];
- for(i=0;i<h;i++) {
- memcpy(s->bytestream, ptr, n);
- s->bytestream += n;
- ptr += linesize;
+ for (i = 0; i < h; i++) {
+ memcpy(bytestream, ptr, n);
+ bytestream += n;
+ ptr += linesize;
}
- if (avctx->pix_fmt == PIX_FMT_YUV420P) {
+ if (avctx->pix_fmt == AV_PIX_FMT_YUV420P || avctx->pix_fmt == AV_PIX_FMT_YUV420P16BE) {
h >>= 1;
n >>= 1;
ptr1 = p->data[1];
ptr2 = p->data[2];
- for(i=0;i<h;i++) {
- memcpy(s->bytestream, ptr1, n);
- s->bytestream += n;
- memcpy(s->bytestream, ptr2, n);
- s->bytestream += n;
+ for (i = 0; i < h; i++) {
+ memcpy(bytestream, ptr1, n);
+ bytestream += n;
+ memcpy(bytestream, ptr2, n);
+ bytestream += n;
ptr1 += p->linesize[1];
ptr2 += p->linesize[2];
}
}
- return s->bytestream - s->bytestream_start;
-}
-
-static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
- PNMContext *s = avctx->priv_data;
- AVFrame *pict = data;
- AVFrame * const p= (AVFrame*)&s->picture;
- int i, h, w, n, linesize, depth, maxval;
- const char *tuple_type;
- uint8_t *ptr;
-
- if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
- av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
- return -1;
- }
-
- *p = *pict;
- p->pict_type= FF_I_TYPE;
- p->key_frame= 1;
-
- s->bytestream_start=
- s->bytestream= outbuf;
- s->bytestream_end= outbuf+buf_size;
-
- h = avctx->height;
- w = avctx->width;
- switch(avctx->pix_fmt) {
- case PIX_FMT_MONOWHITE:
- n = (w + 7) >> 3;
- depth = 1;
- maxval = 1;
- tuple_type = "BLACKANDWHITE";
- break;
- case PIX_FMT_GRAY8:
- n = w;
- depth = 1;
- maxval = 255;
- tuple_type = "GRAYSCALE";
- break;
- case PIX_FMT_RGB24:
- n = w * 3;
- depth = 3;
- maxval = 255;
- tuple_type = "RGB";
- break;
- case PIX_FMT_RGB32:
- n = w * 4;
- depth = 4;
- maxval = 255;
- tuple_type = "RGB_ALPHA";
- break;
- default:
- return -1;
- }
- snprintf(s->bytestream, s->bytestream_end - s->bytestream,
- "P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n",
- w, h, depth, maxval, tuple_type);
- s->bytestream += strlen(s->bytestream);
-
- ptr = p->data[0];
- linesize = p->linesize[0];
-
- if (avctx->pix_fmt == PIX_FMT_RGB32) {
- int j;
- unsigned int v;
-
- for(i=0;i<h;i++) {
- for(j=0;j<w;j++) {
- v = ((uint32_t *)ptr)[j];
- bytestream_put_be24(&s->bytestream, v);
- *s->bytestream++ = v >> 24;
- }
- ptr += linesize;
- }
- } else {
- for(i=0;i<h;i++) {
- memcpy(s->bytestream, ptr, n);
- s->bytestream += n;
- ptr += linesize;
- }
- }
- return s->bytestream - s->bytestream_start;
-}
+ pkt->size = bytestream - bytestream_start;
+ pkt->flags |= AV_PKT_FLAG_KEY;
+ *got_packet = 1;
-#if 0
-static int pnm_probe(AVProbeData *pd)
-{
- const char *p = pd->buf;
- if (pd->buf_size >= 8 &&
- p[0] == 'P' &&
- p[1] >= '4' && p[1] <= '6' &&
- pnm_space(p[2]) )
- return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */
- else
- return 0;
-}
-
-static int pgmyuv_probe(AVProbeData *pd)
-{
- if (match_ext(pd->filename, "pgmyuv"))
- return AVPROBE_SCORE_MAX;
- else
- return 0;
+ return 0;
}
-static int pam_probe(AVProbeData *pd)
+static av_cold int pnm_encode_init(AVCodecContext *avctx)
{
- const char *p = pd->buf;
- if (pd->buf_size >= 8 &&
- p[0] == 'P' &&
- p[1] == '7' &&
- p[2] == '\n')
- return AVPROBE_SCORE_MAX;
- else
- return 0;
-}
+#if FF_API_CODED_FRAME
+FF_DISABLE_DEPRECATION_WARNINGS
+ avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I;
+ avctx->coded_frame->key_frame = 1;
+FF_ENABLE_DEPRECATION_WARNINGS
#endif
+ return 0;
+}
-#ifdef CONFIG_PGM_ENCODER
-AVCodec pgm_encoder = {
- "pgm",
- CODEC_TYPE_VIDEO,
- CODEC_ID_PGM,
- sizeof(PNMContext),
- common_init,
- pnm_encode_frame,
- NULL, //encode_end,
- pnm_decode_frame,
- .pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE},
- .long_name= "PGM (Portable GrayMap) image",
-};
-#endif // CONFIG_PGM_ENCODER
-
-#ifdef CONFIG_PGMYUV_ENCODER
-AVCodec pgmyuv_encoder = {
- "pgmyuv",
- CODEC_TYPE_VIDEO,
- CODEC_ID_PGMYUV,
- sizeof(PNMContext),
- common_init,
- pnm_encode_frame,
- NULL, //encode_end,
- pnm_decode_frame,
- .pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE},
- .long_name= "PGMYUV (Portable GrayMap YUV) image",
+#if CONFIG_PGM_ENCODER
+AVCodec ff_pgm_encoder = {
+ .name = "pgm",
+ .long_name = NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"),
+ .type = AVMEDIA_TYPE_VIDEO,
+ .id = AV_CODEC_ID_PGM,
+ .init = pnm_encode_init,
+ .encode2 = pnm_encode_frame,
+ .pix_fmts = (const enum AVPixelFormat[]){
+ AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16BE, AV_PIX_FMT_NONE
+ },
};
-#endif // CONFIG_PGMYUV_ENCODER
+#endif
-#ifdef CONFIG_PPM_ENCODER
-AVCodec ppm_encoder = {
- "ppm",
- CODEC_TYPE_VIDEO,
- CODEC_ID_PPM,
- sizeof(PNMContext),
- common_init,
- pnm_encode_frame,
- NULL, //encode_end,
- pnm_decode_frame,
- .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_NONE},
- .long_name= "PPM (Portable PixelMap) image",
+#if CONFIG_PGMYUV_ENCODER
+AVCodec ff_pgmyuv_encoder = {
+ .name = "pgmyuv",
+ .long_name = NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"),
+ .type = AVMEDIA_TYPE_VIDEO,
+ .id = AV_CODEC_ID_PGMYUV,
+ .init = pnm_encode_init,
+ .encode2 = pnm_encode_frame,
+ .pix_fmts = (const enum AVPixelFormat[]){
+ AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV420P16BE, AV_PIX_FMT_NONE
+ },
};
-#endif // CONFIG_PPM_ENCODER
+#endif
-#ifdef CONFIG_PBM_ENCODER
-AVCodec pbm_encoder = {
- "pbm",
- CODEC_TYPE_VIDEO,
- CODEC_ID_PBM,
- sizeof(PNMContext),
- common_init,
- pnm_encode_frame,
- NULL, //encode_end,
- pnm_decode_frame,
- .pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE},
- .long_name= "PBM (Portable BitMap) image",
+#if CONFIG_PPM_ENCODER
+AVCodec ff_ppm_encoder = {
+ .name = "ppm",
+ .long_name = NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"),
+ .type = AVMEDIA_TYPE_VIDEO,
+ .id = AV_CODEC_ID_PPM,
+ .init = pnm_encode_init,
+ .encode2 = pnm_encode_frame,
+ .pix_fmts = (const enum AVPixelFormat[]){
+ AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB48BE, AV_PIX_FMT_NONE
+ },
};
-#endif // CONFIG_PBM_ENCODER
+#endif
-#ifdef CONFIG_PAM_ENCODER
-AVCodec pam_encoder = {
- "pam",
- CODEC_TYPE_VIDEO,
- CODEC_ID_PAM,
- sizeof(PNMContext),
- common_init,
- pam_encode_frame,
- NULL, //encode_end,
- pnm_decode_frame,
- .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE},
- .long_name= "PAM (Portable AnyMap) image",
+#if CONFIG_PBM_ENCODER
+AVCodec ff_pbm_encoder = {
+ .name = "pbm",
+ .long_name = NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"),
+ .type = AVMEDIA_TYPE_VIDEO,
+ .id = AV_CODEC_ID_PBM,
+ .init = pnm_encode_init,
+ .encode2 = pnm_encode_frame,
+ .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_MONOWHITE,
+ AV_PIX_FMT_NONE },
};
-#endif // CONFIG_PAM_ENCODER
+#endif