int fill_order;
uint32_t res[4];
int is_thumbnail;
+ unsigned last_tag;
int is_bayer;
uint8_t pattern[4];
unsigned black_level;
unsigned white_level;
- const uint16_t *dng_lut; // Pointer to DNG linearization table
+ uint16_t dng_lut[65536];
uint32_t sub_ifd;
uint16_t cur_page;
dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
}
break;
- case 12: {
- uint16_t *dst16 = (uint16_t *)dst;
- GetBitContext gb;
- init_get_bits8(&gb, src, width);
- for (int i = 0; i < s->width; i++) {
- dst16[i] = get_bits(&gb, 12) << 4;
- }
- }
+ case 10:
+ case 12:
+ case 14: {
+ uint16_t *dst16 = (uint16_t *)dst;
+ int is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
+ uint8_t shift = is_dng ? 0 : 16 - bpp;
+ GetBitContext gb;
+
+ init_get_bits8(&gb, src, width);
+ for (int i = 0; i < s->width; i++) {
+ dst16[i] = get_bits(&gb, bpp) << shift;
+ }
+ }
break;
default:
if (usePtr) {
z_stream zstream = { 0 };
int zret;
- zstream.next_in = (uint8_t *)src;
+ zstream.next_in = src;
zstream.avail_in = size;
zstream.next_out = dst;
zstream.avail_out = *len;
return ret;
}
+static int dng_decode_strip(AVCodecContext *avctx, AVFrame *frame);
+
static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
const uint8_t *src, int size, int strip_start, int lines)
{
av_assert0(s->bpp == 24);
}
if (s->is_bayer) {
- width = (s->bpp * s->width + 7) >> 3;
+ av_assert0(width == (s->bpp * s->width + 7) >> 3);
}
if (p->format == AV_PIX_FMT_GRAY12) {
av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, width);
is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
+ /* Decode JPEG-encoded DNGs with strips */
+ if (s->compr == TIFF_NEWJPEG && is_dng) {
+ if (s->strips > 1) {
+ av_log(s->avctx, AV_LOG_ERROR, "More than one DNG JPEG strips unsupported\n");
+ return AVERROR_PATCHWELCOME;
+ }
+ if ((ret = dng_decode_strip(s->avctx, p)) < 0)
+ return ret;
+ return 0;
+ }
+
+ if (is_dng && stride == 0)
+ return AVERROR_INVALIDDATA;
+
for (line = 0; line < lines; line++) {
if (src - ssrc > size) {
av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
/* Color processing for DNG images with uncompressed strips (non-tiled) */
if (is_dng) {
- int is_u16, pixel_size_bytes, pixel_size_bits;
+ int is_u16, pixel_size_bytes, pixel_size_bits, elements;
- is_u16 = (s->bpp > 8);
+ is_u16 = (s->bpp / s->bppcount > 8);
pixel_size_bits = (is_u16 ? 16 : 8);
pixel_size_bytes = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
+ elements = width / pixel_size_bytes * pixel_size_bits / s->bpp * s->bppcount; // need to account for [1, 16] bpp
+ av_assert0 (elements * pixel_size_bytes <= FFABS(stride));
dng_blit(s,
dst,
0, // no stride, only 1 line
dst,
0, // no stride, only 1 line
- width / pixel_size_bytes * pixel_size_bits / s->bpp * s->bppcount, // need to account for [1, 16] bpp
+ elements,
1,
0, // single-component variation is only preset in JPEG-encoded DNGs
is_u16);
return 0;
}
-static float av_always_inline linear_to_srgb(float value) {
- if (value <= 0.0031308f)
- return value * 12.92f;
- else
- return powf(value * 1.055f, 1.0f / 2.4f) - 0.055f;
-}
-
/**
* Map stored raw sensor values into linear reference values (see: DNG Specification - Chapter 5)
- * Then convert to sRGB color space.
*/
static uint16_t av_always_inline dng_process_color16(uint16_t value,
const uint16_t *lut,
// Color scaling
value_norm = (float)value * scale_factor;
- // Color space conversion (sRGB)
- value = av_clip_uint16_c((uint16_t)(linear_to_srgb(value_norm) * 0xFFFF));
+ value = av_clip_uint16_c(value_norm * 65535);
return value;
}
}
} else {
for (line = 0; line < height; line++) {
+ uint8_t *dst_u8 = dst;
+ const uint8_t *src_u8 = src;
+
for (col = 0; col < width; col++)
- *dst++ = dng_process_color8(*src++, s->dng_lut, s->black_level, scale_factor);
+ *dst_u8++ = dng_process_color8(*src_u8++, s->dng_lut, s->black_level, scale_factor);
dst += dst_stride;
src += src_stride;
}
}
-static int dng_decode_jpeg_tile(AVCodecContext *avctx, AVFrame *frame,
- int tile_byte_count, int x, int y, int w, int h)
+static int dng_decode_jpeg(AVCodecContext *avctx, AVFrame *frame,
+ int tile_byte_count, int dst_x, int dst_y, int w, int h)
{
TiffContext *s = avctx->priv_data;
AVPacket jpkt;
int is_single_comp, is_u16, pixel_size;
int ret;
+ if (tile_byte_count < 0 || tile_byte_count > bytestream2_get_bytes_left(&s->gb))
+ return AVERROR_INVALIDDATA;
+
/* Prepare a packet and send to the MJPEG decoder */
av_init_packet(&jpkt);
jpkt.data = (uint8_t*)s->gb.buffer;
return 0;
}
+ is_u16 = (s->bpp > 8);
+
/* Copy the outputted tile's pixels from 'jpgframe' to 'frame' (final buffer) */
+ if (s->jpgframe->width != s->avctx_mjpeg->width ||
+ s->jpgframe->height != s->avctx_mjpeg->height ||
+ s->jpgframe->format != s->avctx_mjpeg->pix_fmt)
+ return AVERROR_INVALIDDATA;
+
/* See dng_blit for explanation */
- is_single_comp = (s->avctx_mjpeg->width == w * 2 && s->avctx_mjpeg->height == h / 2);
+ if (s->avctx_mjpeg->width == w * 2 &&
+ s->avctx_mjpeg->height == h / 2 &&
+ s->avctx_mjpeg->pix_fmt == AV_PIX_FMT_GRAY16LE) {
+ is_single_comp = 1;
+ } else if (s->avctx_mjpeg->width >= w &&
+ s->avctx_mjpeg->height >= h &&
+ s->avctx_mjpeg->pix_fmt == (is_u16 ? AV_PIX_FMT_GRAY16 : AV_PIX_FMT_GRAY8)
+ ) {
+ is_single_comp = 0;
+ } else
+ return AVERROR_INVALIDDATA;
- is_u16 = (s->bpp > 8);
pixel_size = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
if (is_single_comp && !is_u16) {
return AVERROR_PATCHWELCOME;
}
- dst_offset = x + frame->linesize[0] * y / pixel_size;
+ dst_offset = dst_x + frame->linesize[0] * dst_y / pixel_size;
dst_data = frame->data[0] + dst_offset * pixel_size;
src_data = s->jpgframe->data[0];
return 0;
}
-static int dng_decode_tiles(AVCodecContext *avctx, AVFrame *frame)
+static int dng_decode_tiles(AVCodecContext *avctx, AVFrame *frame, AVPacket *avpkt)
{
TiffContext *s = avctx->priv_data;
int tile_idx;
int pos_x = 0, pos_y = 0;
int ret;
+ s->jpgframe->width = s->tile_width;
+ s->jpgframe->height = s->tile_length;
+
+ s->avctx_mjpeg->width = s->tile_width;
+ s->avctx_mjpeg->height = s->tile_length;
+
has_width_leftover = (s->width % s->tile_width != 0);
has_height_leftover = (s->height % s->tile_length != 0);
bytestream2_seek(&s->gb, tile_offset, SEEK_SET);
/* Decode JPEG tile and copy it in the reference frame */
- ret = dng_decode_jpeg_tile(avctx, frame, tile_byte_count, pos_x, pos_y, tile_width, tile_length);
+ ret = dng_decode_jpeg(avctx, frame, tile_byte_count, pos_x, pos_y, tile_width, tile_length);
if (ret < 0)
return ret;
}
}
- return 0;
-}
+ /* Frame is ready to be output */
+ frame->pict_type = AV_PICTURE_TYPE_I;
+ frame->key_frame = 1;
-static int dng_decode(AVCodecContext *avctx, AVFrame *frame, AVPacket *avpkt) {
- int ret;
+ return avpkt->size;
+}
+static int dng_decode_strip(AVCodecContext *avctx, AVFrame *frame)
+{
TiffContext *s = avctx->priv_data;
- s->jpgframe->width = s->tile_width;
- s->jpgframe->height = s->tile_length;
-
- s->avctx_mjpeg->width = s->tile_width;
- s->avctx_mjpeg->height = s->tile_length;
+ s->jpgframe->width = s->width;
+ s->jpgframe->height = s->height;
- /* Decode all tiles in a frame */
- ret = dng_decode_tiles(avctx, frame);
- if (ret < 0)
- return ret;
+ s->avctx_mjpeg->width = s->width;
+ s->avctx_mjpeg->height = s->height;
- /* Frame is ready to be output */
- frame->pict_type = AV_PICTURE_TYPE_I;
- frame->key_frame = 1;
-
- return avpkt->size;
+ return dng_decode_jpeg(avctx, frame, s->stripsize, 0, 0, s->width, s->height);
}
static int init_image(TiffContext *s, ThreadFrame *frame)
return AVERROR_PATCHWELCOME;
}
break;
+ case 10101:
case 10121:
- switch (AV_RL32(s->pattern)) {
- case 0x02010100:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_RGGB16LE : AV_PIX_FMT_BAYER_RGGB16BE;
- break;
- case 0x00010102:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_BGGR16LE : AV_PIX_FMT_BAYER_BGGR16BE;
- break;
- case 0x01000201:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GBRG16LE : AV_PIX_FMT_BAYER_GBRG16BE;
- break;
- case 0x01020001:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GRBG16LE : AV_PIX_FMT_BAYER_GRBG16BE;
- break;
- default:
- av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
- AV_RL32(s->pattern));
- return AVERROR_PATCHWELCOME;
- }
- /* Force endianness as mentioned in 'DNG Specification: Chapter 3: BitsPerSample'
- NOTE: The spec actually specifies big-endian, not sure why we need little-endian, but
- such images don't work otherwise. Examples are images produced by Zenmuse X7. */
- if ((s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG)
- && (s->bpp != 8 && s->bpp != 16 && s->bpp != 32)) {
- switch (s->avctx->pix_fmt) {
- case AV_PIX_FMT_BAYER_RGGB16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB16LE; break;
- case AV_PIX_FMT_BAYER_BGGR16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR16LE; break;
- case AV_PIX_FMT_BAYER_GBRG16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG16LE; break;
- case AV_PIX_FMT_BAYER_GRBG16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG16LE; break;
- }
- }
- break;
+ case 10141:
case 10161:
switch (AV_RL32(s->pattern)) {
case 0x02010100:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_RGGB16LE : AV_PIX_FMT_BAYER_RGGB16BE;
+ s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB16;
break;
case 0x00010102:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_BGGR16LE : AV_PIX_FMT_BAYER_BGGR16BE;
+ s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR16;
break;
case 0x01000201:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GBRG16LE : AV_PIX_FMT_BAYER_GBRG16BE;
+ s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG16;
break;
case 0x01020001:
- s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GRBG16LE : AV_PIX_FMT_BAYER_GRBG16BE;
+ s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG16;
break;
default:
av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
{
+ AVFrameSideData *sd;
+ GetByteContext gb_temp;
unsigned tag, type, count, off, value = 0, value2 = 1; // value2 is a denominator so init. to 1
int i, start;
int pos;
if (ret < 0) {
goto end;
}
+ if (tag <= s->last_tag)
+ return AVERROR_INVALIDDATA;
+
+ // We ignore TIFF_STRIP_SIZE as it is sometimes in the logic but wrong order around TIFF_STRIP_OFFS
+ if (tag != TIFF_STRIP_SIZE)
+ s->last_tag = tag;
off = bytestream2_tell(&s->gb);
if (count == 1) {
case TIFF_RATIONAL:
value = ff_tget(&s->gb, TIFF_LONG, s->le);
value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
+ if (!value2) {
+ av_log(s->avctx, AV_LOG_ERROR, "Invalid denominator in rational\n");
+ return AVERROR_INVALIDDATA;
+ }
+
break;
case TIFF_STRING:
if (count <= 4) {
s->height = value;
break;
case TIFF_BPP:
- if (count > 5U) {
+ if (count > 5 || count <= 0) {
av_log(s->avctx, AV_LOG_ERROR,
"This format is not supported (bpp=%d, %d components)\n",
value, count);
"Samples per pixel requires a single value, many provided\n");
return AVERROR_INVALIDDATA;
}
- if (value > 5U) {
+ if (value > 5 || value <= 0) {
av_log(s->avctx, AV_LOG_ERROR,
- "Samples per pixel %d is too large\n", value);
+ "Invalid samples per pixel %d\n", value);
return AVERROR_INVALIDDATA;
}
if (s->bppcount == 1)
else if (count > 1)
s->sub_ifd = ff_tget(&s->gb, TIFF_LONG, s->le); /** Only get the first SubIFD */
break;
- case DNG_LINEARIZATION_TABLE: {
- uint32_t lut_offset = value;
- uint32_t lut_size = count;
- uint32_t lut_wanted_size = 1 << s->bpp;
- if (lut_wanted_size != lut_size)
- av_log(s->avctx, AV_LOG_WARNING, "DNG contains LUT with invalid size (%"PRIu32"), disabling LUT\n", lut_size);
- else if (lut_offset >= bytestream2_size(&s->gb))
- av_log(s->avctx, AV_LOG_WARNING, "DNG contains LUT with invalid offset (%"PRIu32"), disabling LUT\n", lut_offset);
- else
- s->dng_lut = (uint16_t*)(s->gb.buffer + lut_offset);
+ case DNG_LINEARIZATION_TABLE:
+ if (count > FF_ARRAY_ELEMS(s->dng_lut))
+ return AVERROR_INVALIDDATA;
+ for (int i = 0; i < count; i++)
+ s->dng_lut[i] = ff_tget(&s->gb, type, s->le);
break;
- }
case DNG_BLACK_LEVEL:
if (count > 1) { /* Use the first value in the pattern (assume they're all the same) */
if (type == TIFF_RATIONAL) {
value = ff_tget(&s->gb, TIFF_LONG, s->le);
value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
+ if (!value2) {
+ av_log(s->avctx, AV_LOG_ERROR, "Invalid black level denominator\n");
+ return AVERROR_INVALIDDATA;
+ }
s->black_level = value / value2;
} else
case TIFF_PHOTOMETRIC_SEPARATED:
case TIFF_PHOTOMETRIC_YCBCR:
case TIFF_PHOTOMETRIC_CFA:
+ case TIFF_PHOTOMETRIC_LINEAR_RAW: // Used by DNG images
s->photometric = value;
break;
case TIFF_PHOTOMETRIC_ALPHA_MASK:
case TIFF_PHOTOMETRIC_ITU_LAB:
case TIFF_PHOTOMETRIC_LOG_L:
case TIFF_PHOTOMETRIC_LOG_LUV:
- case TIFF_PHOTOMETRIC_LINEAR_RAW:
avpriv_report_missing_feature(s->avctx,
"PhotometricInterpretation 0x%04X",
value);
break;
case TIFF_GEO_KEY_DIRECTORY:
if (s->geotag_count) {
- avpriv_request_sample(s->avctx, "Multiple geo key directories\n");
+ avpriv_request_sample(s->avctx, "Multiple geo key directories");
return AVERROR_INVALIDDATA;
}
ADD_METADATA(1, "GeoTIFF_Version", NULL);
}
}
break;
+ case TIFF_ICC_PROFILE:
+ gb_temp = s->gb;
+ bytestream2_seek(&gb_temp, SEEK_SET, off);
+
+ if (bytestream2_get_bytes_left(&gb_temp) < count)
+ return AVERROR_INVALIDDATA;
+
+ sd = av_frame_new_side_data(frame, AV_FRAME_DATA_ICC_PROFILE, count);
+ if (!sd)
+ return AVERROR(ENOMEM);
+
+ bytestream2_get_bufferu(&gb_temp, sd->data, count);
+ break;
case TIFF_ARTIST:
ADD_METADATA(count, "artist", NULL);
break;
GetByteContext stripsizes;
GetByteContext stripdata;
int retry_for_subifd, retry_for_page;
+ int is_dng;
+ int has_tile_bits, has_strip_bits;
bytestream2_init(&s->gb, avpkt->data, avpkt->size);
s->is_tiled = 0;
s->is_jpeg = 0;
s->cur_page = 0;
- s->dng_lut = NULL;
+ s->last_tag = 0;
+
+ for (i = 0; i < 65536; i++)
+ s->dng_lut[i] = i;
+
free_geotags(s);
// Reset these offsets so we can tell if they were set this frame
return AVERROR_INVALIDDATA;
}
if (off <= last_off) {
- avpriv_request_sample(s->avctx, "non increasing IFD offset\n");
+ avpriv_request_sample(s->avctx, "non increasing IFD offset");
return AVERROR_INVALIDDATA;
}
if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
goto again;
}
+ /* At this point we've decided on which (Sub)IFD to process */
+
+ is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
+
for (i = 0; i<s->geotag_count; i++) {
const char *keyname = get_geokey_name(s->geotags[i].key);
if (!keyname) {
}
}
+ if (is_dng) {
+ int bps;
+
+ if (s->bpp % s->bppcount)
+ return AVERROR_INVALIDDATA;
+ bps = s->bpp / s->bppcount;
+ if (bps < 8 || bps > 32)
+ return AVERROR_INVALIDDATA;
+
+ if (s->white_level == 0)
+ s->white_level = (1LL << bps) - 1; /* Default value as per the spec */
+
+ if (s->white_level <= s->black_level) {
+ av_log(avctx, AV_LOG_ERROR, "BlackLevel (%"PRId32") must be less than WhiteLevel (%"PRId32")\n",
+ s->black_level, s->white_level);
+ return AVERROR_INVALIDDATA;
+ }
+
+ if (s->planar)
+ return AVERROR_PATCHWELCOME;
+ }
+
if (!s->is_tiled && !s->strippos && !s->stripoff) {
av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
return AVERROR_INVALIDDATA;
}
+
+ has_tile_bits = s->is_tiled || s->tile_byte_counts_offset || s->tile_offsets_offset || s->tile_width || s->tile_length || s->tile_count;
+ has_strip_bits = s->strippos || s->strips || s->stripoff || s->rps || s->sot || s->sstype || s->stripsize || s->stripsizesoff;
+
+ if (has_tile_bits && has_strip_bits) {
+ av_log(avctx, AV_LOG_WARNING, "Tiled TIFF is not allowed to strip\n");
+ }
+
/* now we have the data and may start decoding */
if ((ret = init_image(s, &frame)) < 0)
return ret;
}
}
+ if (s->photometric == TIFF_PHOTOMETRIC_LINEAR_RAW ||
+ s->photometric == TIFF_PHOTOMETRIC_CFA) {
+ p->color_trc = AVCOL_TRC_LINEAR;
+ } else if (s->photometric == TIFF_PHOTOMETRIC_BLACK_IS_ZERO) {
+ p->color_trc = AVCOL_TRC_GAMMA22;
+ }
+
/* Handle DNG images with JPEG-compressed tiles */
- if ((s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG) && s->is_tiled) {
+ if (is_dng && s->is_tiled) {
if (!s->is_jpeg) {
avpriv_report_missing_feature(avctx, "DNG uncompressed tiled images");
return AVERROR_PATCHWELCOME;
avpriv_report_missing_feature(avctx, "DNG JPG-compressed tiled non-bayer-encoded images");
return AVERROR_PATCHWELCOME;
} else {
- if ((ret = dng_decode(avctx, (AVFrame*)data, avpkt)) > 0)
+ if ((ret = dng_decode_tiles(avctx, (AVFrame*)data, avpkt)) > 0)
*got_frame = 1;
return ret;
}
FFSWAP(int, p->linesize[0], p->linesize[1]);
}
- if (s->is_bayer && s->white_level && s->bpp == 16 &&
- !(s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG)) {
+ if (s->is_bayer && s->white_level && s->bpp == 16 && !is_dng) {
uint16_t *dst = (uint16_t *)p->data[0];
for (i = 0; i < s->height; i++) {
for (j = 0; j < s->width; j++)
s->avctx_mjpeg->flags2 = avctx->flags2;
s->avctx_mjpeg->dct_algo = avctx->dct_algo;
s->avctx_mjpeg->idct_algo = avctx->idct_algo;
- ret = ff_codec_open2_recursive(s->avctx_mjpeg, codec, NULL);
+ ret = avcodec_open2(s->avctx_mjpeg, codec, NULL);
if (ret < 0) {
- av_frame_free(&s->jpgframe);
- avcodec_free_context(&s->avctx_mjpeg);
return ret;
}
.init = tiff_init,
.close = tiff_end,
.decode = decode_frame,
- .init_thread_copy = ONLY_IF_THREADS_ENABLED(tiff_init),
.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
+ .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP,
.priv_class = &tiff_decoder_class,
};