X-Git-Url: https://git.sesse.net/?a=blobdiff_plain;f=libavcodec%2Ftiffenc.c;h=883d6fb8a2dad6a674af54ea860c06fa97ff84b5;hb=fa3fde168b8c3d2d7c008a98457832575dbf9fe0;hp=7361614505f6a947568fdb1dcc3800d957b1adc5;hpb=3b0ad040b3bffaea5c5fd36bfddd9a436bc6c9c4;p=ffmpeg diff --git a/libavcodec/tiffenc.c b/libavcodec/tiffenc.c index 7361614505f..883d6fb8a2d 100644 --- a/libavcodec/tiffenc.c +++ b/libavcodec/tiffenc.c @@ -25,16 +25,17 @@ * @author Bartlomiej Wolowiec */ +#include "libavutil/imgutils.h" #include "libavutil/log.h" #include "libavutil/opt.h" #include "avcodec.h" -#include "internal.h" +#include "config.h" #if CONFIG_ZLIB #include #endif -#include "libavutil/opt.h" #include "bytestream.h" +#include "internal.h" #include "tiff.h" #include "rle.h" #include "lzw.h" @@ -43,8 +44,8 @@ #define TIFF_MAX_ENTRY 32 /** sizes of various TIFF field types (string size = 1)*/ -static const uint8_t type_sizes2[6] = { - 0, 1, 1, 2, 4, 8 +static const uint8_t type_sizes2[14] = { + 0, 1, 1, 2, 4, 8, 1, 1, 2, 4, 8, 4, 8, 4 }; typedef struct TiffEncoderContext { @@ -59,6 +60,12 @@ typedef struct TiffEncoderContext { int bpp_tab_size; ///< bpp_tab size int photometric_interpretation; ///< photometric interpretation int strips; ///< number of strips + uint32_t *strip_sizes; + unsigned int strip_sizes_size; + uint32_t *strip_offsets; + unsigned int strip_offsets_size; + uint8_t *yuv_line; + unsigned int yuv_line_size; int rps; ///< row per strip uint8_t entries[TIFF_MAX_ENTRY*12]; ///< entires in header int num_entries; ///< number of entires @@ -72,7 +79,8 @@ typedef struct TiffEncoderContext { /** - * Check free space in buffer + * Check free space in buffer. + * * @param s Tiff context * @param need Needed bytes * @return 0 - ok, 1 - no free space @@ -88,13 +96,13 @@ static inline int check_size(TiffEncoderContext * s, uint64_t need) } /** - * Put n values to buffer + * Put n values to buffer. * - * @param p Pointer to pointer to output buffer - * @param n Number of values - * @param val Pointer to values - * @param type Type of values - * @param flip =0 - normal copy, >0 - flip + * @param p pointer to pointer to output buffer + * @param n number of values + * @param val pointer to values + * @param type type of values + * @param flip = 0 - normal copy, >0 - flip */ static void tnput(uint8_t ** p, int n, const uint8_t * val, enum TiffTypes type, int flip) @@ -109,11 +117,12 @@ static void tnput(uint8_t ** p, int n, const uint8_t * val, enum TiffTypes type, /** * Add entry to directory in tiff header. + * * @param s Tiff context - * @param tag Tag that identifies the entry - * @param type Entry type - * @param count The number of values - * @param ptr_val Pointer to values + * @param tag tag that identifies the entry + * @param type entry type + * @param count the number of values + * @param ptr_val pointer to values */ static void add_entry(TiffEncoderContext * s, enum TiffTags tag, enum TiffTypes type, int count, @@ -146,14 +155,14 @@ static void add_entry1(TiffEncoderContext * s, } /** - * Encode one strip in tiff file + * Encode one strip in tiff file. * * @param s Tiff context - * @param src Input buffer - * @param dst Output buffer - * @param n Size of input buffer - * @param compr Compression method - * @return Number of output bytes. If an output error is encountered, -1 returned + * @param src input buffer + * @param dst output buffer + * @param n size of input buffer + * @param compr compression method + * @return number of output bytes. If an output error is encountered, -1 is returned */ static int encode_strip(TiffEncoderContext * s, const int8_t * src, uint8_t * dst, int n, int compr) @@ -214,6 +223,18 @@ static void pack_yuv(TiffEncoderContext * s, uint8_t * dst, int lnum) } } +static av_cold int encode_init(AVCodecContext *avctx) +{ + TiffEncoderContext *s = avctx->priv_data; + + avctx->coded_frame= &s->picture; + avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I; + avctx->coded_frame->key_frame = 1; + s->avctx = avctx; + + return 0; +} + static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, const AVFrame *pict, int *got_packet) { @@ -223,84 +244,54 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, uint8_t *ptr; uint8_t *offset; uint32_t strips; - uint32_t *strip_sizes = NULL; - uint32_t *strip_offsets = NULL; int bytes_per_row; uint32_t res[2] = { s->dpi, 1 }; // image resolution (72/1) - uint16_t bpp_tab[] = { 8, 8, 8, 8 }; + uint16_t bpp_tab[4]; int ret = -1; int is_yuv = 0; - uint8_t *yuv_line = NULL; int shift_h, shift_v; - s->avctx = avctx; - *p = *pict; - p->pict_type = AV_PICTURE_TYPE_I; - p->key_frame = 1; - avctx->coded_frame= &s->picture; s->width = avctx->width; s->height = avctx->height; s->subsampling[0] = 1; s->subsampling[1] = 1; + avctx->bits_per_coded_sample = + s->bpp = av_get_bits_per_pixel(&av_pix_fmt_descriptors[avctx->pix_fmt]); + switch (avctx->pix_fmt) { case PIX_FMT_RGBA64LE: - s->bpp = 64; - s->photometric_interpretation = 2; - bpp_tab[0] = 16; - bpp_tab[1] = 16; - bpp_tab[2] = 16; - bpp_tab[3] = 16; - break; case PIX_FMT_RGB48LE: - s->bpp = 48; - s->photometric_interpretation = 2; - bpp_tab[0] = 16; - bpp_tab[1] = 16; - bpp_tab[2] = 16; - bpp_tab[3] = 16; - break; case PIX_FMT_RGBA: - avctx->bits_per_coded_sample = - s->bpp = 32; - s->photometric_interpretation = 2; - break; case PIX_FMT_RGB24: - avctx->bits_per_coded_sample = - s->bpp = 24; s->photometric_interpretation = 2; break; case PIX_FMT_GRAY8: avctx->bits_per_coded_sample = 0x28; - s->bpp = 8; + case PIX_FMT_GRAY8A: + case PIX_FMT_GRAY16LE: s->photometric_interpretation = 1; break; case PIX_FMT_PAL8: - avctx->bits_per_coded_sample = - s->bpp = 8; s->photometric_interpretation = 3; break; case PIX_FMT_MONOBLACK: case PIX_FMT_MONOWHITE: - avctx->bits_per_coded_sample = - s->bpp = 1; s->photometric_interpretation = avctx->pix_fmt == PIX_FMT_MONOBLACK; - bpp_tab[0] = 1; break; case PIX_FMT_YUV420P: case PIX_FMT_YUV422P: + case PIX_FMT_YUV440P: case PIX_FMT_YUV444P: case PIX_FMT_YUV410P: case PIX_FMT_YUV411P: s->photometric_interpretation = 6; avcodec_get_chroma_sub_sample(avctx->pix_fmt, &shift_h, &shift_v); - s->bpp = 8 + (16 >> (shift_h + shift_v)); s->subsampling[0] = 1 << shift_h; s->subsampling[1] = 1 << shift_v; - s->bpp_tab_size = 3; is_yuv = 1; break; default: @@ -308,8 +299,10 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, "This colors format is not supported\n"); return -1; } - if (!is_yuv) - s->bpp_tab_size = (s->bpp >= 48) ? ((s->bpp + 7) >> 4):((s->bpp + 7) >> 3); + + s->bpp_tab_size = av_pix_fmt_descriptors[avctx->pix_fmt].nb_components; + for (i = 0; i < s->bpp_tab_size; i++) + bpp_tab[i] = av_pix_fmt_descriptors[avctx->pix_fmt].comp[i].depth_minus1 + 1; if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE || s->compr == TIFF_LZW) //best choose for DEFLATE @@ -338,15 +331,20 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, offset = ptr; bytestream_put_le32(&ptr, 0); - strip_sizes = av_mallocz(sizeof(*strip_sizes) * strips); - strip_offsets = av_mallocz(sizeof(*strip_offsets) * strips); + av_fast_padded_mallocz(&s->strip_sizes, &s->strip_sizes_size, sizeof(s->strip_sizes[0]) * strips); + av_fast_padded_mallocz(&s->strip_offsets, &s->strip_offsets_size, sizeof(s->strip_offsets[0]) * strips); + + if (!s->strip_sizes || !s->strip_offsets) { + ret = AVERROR(ENOMEM); + goto fail; + } bytes_per_row = (((s->width - 1)/s->subsampling[0] + 1) * s->bpp * s->subsampling[0] * s->subsampling[1] + 7) >> 3; if (is_yuv){ - yuv_line = av_malloc(bytes_per_row); - if (yuv_line == NULL){ - av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n"); + av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row); + if (s->yuv_line == NULL){ + ret = AVERROR(ENOMEM); goto fail; } } @@ -359,12 +357,12 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, zlen = bytes_per_row * s->rps; zbuf = av_malloc(zlen); - strip_offsets[0] = ptr - pkt->data; + s->strip_offsets[0] = ptr - pkt->data; zn = 0; for (j = 0; j < s->rps; j++) { if (is_yuv){ - pack_yuv(s, yuv_line, j); - memcpy(zbuf + zn, yuv_line, bytes_per_row); + pack_yuv(s, s->yuv_line, j); + memcpy(zbuf + zn, s->yuv_line, bytes_per_row); j += s->subsampling[1] - 1; } else @@ -379,23 +377,23 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, goto fail; } ptr += ret; - strip_sizes[0] = ptr - pkt->data - strip_offsets[0]; + s->strip_sizes[0] = ptr - pkt->data - s->strip_offsets[0]; } else #endif { if(s->compr == TIFF_LZW) s->lzws = av_malloc(ff_lzw_encode_state_size); for (i = 0; i < s->height; i++) { - if (strip_sizes[i / s->rps] == 0) { + if (s->strip_sizes[i / s->rps] == 0) { if(s->compr == TIFF_LZW){ ff_lzw_encode_init(s->lzws, ptr, s->buf_size - (*s->buf - s->buf_start), 12, FF_LZW_TIFF, put_bits); } - strip_offsets[i / s->rps] = ptr - pkt->data; + s->strip_offsets[i / s->rps] = ptr - pkt->data; } if (is_yuv){ - pack_yuv(s, yuv_line, i); - ret = encode_strip(s, yuv_line, ptr, bytes_per_row, s->compr); + pack_yuv(s, s->yuv_line, i); + ret = encode_strip(s, s->yuv_line, ptr, bytes_per_row, s->compr); i += s->subsampling[1] - 1; } else @@ -405,11 +403,11 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, av_log(s->avctx, AV_LOG_ERROR, "Encode strip failed\n"); goto fail; } - strip_sizes[i / s->rps] += ret; + s->strip_sizes[i / s->rps] += ret; ptr += ret; if(s->compr == TIFF_LZW && (i==s->height-1 || i%s->rps == s->rps-1)){ ret = ff_lzw_encode_flush(s->lzws, flush_put_bits); - strip_sizes[(i / s->rps )] += ret ; + s->strip_sizes[(i / s->rps )] += ret ; ptr += ret; } } @@ -428,13 +426,13 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, add_entry1(s,TIFF_COMPR, TIFF_SHORT, s->compr); add_entry1(s,TIFF_INVERT, TIFF_SHORT, s->photometric_interpretation); - add_entry(s, TIFF_STRIP_OFFS, TIFF_LONG, strips, strip_offsets); + add_entry(s, TIFF_STRIP_OFFS, TIFF_LONG, strips, s->strip_offsets); if (s->bpp_tab_size) add_entry1(s,TIFF_SAMPLES_PER_PIXEL, TIFF_SHORT, s->bpp_tab_size); add_entry1(s,TIFF_ROWSPERSTRIP, TIFF_LONG, s->rps); - add_entry(s, TIFF_STRIP_SIZE, TIFF_LONG, strips, strip_sizes); + add_entry(s, TIFF_STRIP_SIZE, TIFF_LONG, strips, s->strip_sizes); add_entry(s, TIFF_XRES, TIFF_RATIONAL, 1, res); add_entry(s, TIFF_YRES, TIFF_RATIONAL, 1, res); add_entry1(s,TIFF_RES_UNIT, TIFF_SHORT, 2); @@ -474,12 +472,20 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, *got_packet = 1; fail: - av_free(strip_sizes); - av_free(strip_offsets); - av_free(yuv_line); return ret < 0 ? ret : 0; } +static av_cold int encode_close(AVCodecContext *avctx) +{ + TiffEncoderContext *s = avctx->priv_data; + + av_freep(&s->strip_sizes); + av_freep(&s->strip_offsets); + av_freep(&s->yuv_line); + + return 0; +} + #define OFFSET(x) offsetof(TiffEncoderContext, x) #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM static const AVOption options[] = { @@ -504,13 +510,15 @@ static const AVClass tiffenc_class = { AVCodec ff_tiff_encoder = { .name = "tiff", .type = AVMEDIA_TYPE_VIDEO, - .id = CODEC_ID_TIFF, + .id = AV_CODEC_ID_TIFF, .priv_data_size = sizeof(TiffEncoderContext), + .init = encode_init, .encode2 = encode_frame, + .close = encode_close, .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_RGB24, PIX_FMT_PAL8, PIX_FMT_GRAY8, + PIX_FMT_RGB24, PIX_FMT_PAL8, PIX_FMT_GRAY8, PIX_FMT_GRAY8A, PIX_FMT_GRAY16LE, PIX_FMT_MONOBLACK, PIX_FMT_MONOWHITE, - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUV444P, + PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUV440P, PIX_FMT_YUV444P, PIX_FMT_YUV410P, PIX_FMT_YUV411P, PIX_FMT_RGB48LE, PIX_FMT_RGBA, PIX_FMT_RGBA64LE, PIX_FMT_NONE