You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

2231 lines
76KB

  1. /*
  2. * Copyright (c) 2006 Konstantin Shishkov
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * TIFF image decoder
  23. * @author Konstantin Shishkov
  24. */
  25. #include "config.h"
  26. #if CONFIG_ZLIB
  27. #include <zlib.h>
  28. #endif
  29. #if CONFIG_LZMA
  30. #define LZMA_API_STATIC
  31. #include <lzma.h>
  32. #endif
  33. #include "libavutil/attributes.h"
  34. #include "libavutil/avstring.h"
  35. #include "libavutil/error.h"
  36. #include "libavutil/intreadwrite.h"
  37. #include "libavutil/imgutils.h"
  38. #include "libavutil/opt.h"
  39. #include "avcodec.h"
  40. #include "bytestream.h"
  41. #include "faxcompr.h"
  42. #include "internal.h"
  43. #include "lzw.h"
  44. #include "mathops.h"
  45. #include "tiff.h"
  46. #include "tiff_data.h"
  47. #include "mjpegdec.h"
  48. #include "thread.h"
  49. #include "get_bits.h"
  50. typedef struct TiffContext {
  51. AVClass *class;
  52. AVCodecContext *avctx;
  53. GetByteContext gb;
  54. /* JPEG decoding for DNG */
  55. AVCodecContext *avctx_mjpeg; // wrapper context for MJPEG
  56. AVPacket *jpkt; // encoded JPEG tile
  57. AVFrame *jpgframe; // decoded JPEG tile
  58. int get_subimage;
  59. uint16_t get_page;
  60. int get_thumbnail;
  61. enum TiffType tiff_type;
  62. int width, height;
  63. unsigned int bpp, bppcount;
  64. uint32_t palette[256];
  65. int palette_is_set;
  66. int le;
  67. enum TiffCompr compr;
  68. enum TiffPhotometric photometric;
  69. int planar;
  70. int subsampling[2];
  71. int fax_opts;
  72. int predictor;
  73. int fill_order;
  74. uint32_t res[4];
  75. int is_thumbnail;
  76. unsigned last_tag;
  77. int is_bayer;
  78. uint8_t pattern[4];
  79. unsigned black_level;
  80. unsigned white_level;
  81. uint16_t dng_lut[65536];
  82. uint32_t sub_ifd;
  83. uint16_t cur_page;
  84. int strips, rps, sstype;
  85. int sot;
  86. int stripsizesoff, stripsize, stripoff, strippos;
  87. LZWState *lzw;
  88. /* Tile support */
  89. int is_tiled;
  90. int tile_byte_counts_offset, tile_offsets_offset;
  91. int tile_width, tile_length;
  92. int tile_count;
  93. int is_jpeg;
  94. uint8_t *deinvert_buf;
  95. int deinvert_buf_size;
  96. uint8_t *yuv_line;
  97. unsigned int yuv_line_size;
  98. uint8_t *fax_buffer;
  99. unsigned int fax_buffer_size;
  100. int geotag_count;
  101. TiffGeoTag *geotags;
  102. } TiffContext;
  103. static void tiff_set_type(TiffContext *s, enum TiffType tiff_type) {
  104. if (s->tiff_type < tiff_type) // Prioritize higher-valued entries
  105. s->tiff_type = tiff_type;
  106. }
  107. static void free_geotags(TiffContext *const s)
  108. {
  109. int i;
  110. for (i = 0; i < s->geotag_count; i++) {
  111. if (s->geotags[i].val)
  112. av_freep(&s->geotags[i].val);
  113. }
  114. av_freep(&s->geotags);
  115. s->geotag_count = 0;
  116. }
  117. #define RET_GEOKEY(TYPE, array, element)\
  118. if (key >= TIFF_##TYPE##_KEY_ID_OFFSET &&\
  119. key - TIFF_##TYPE##_KEY_ID_OFFSET < FF_ARRAY_ELEMS(tiff_##array##_name_type_map))\
  120. return tiff_##array##_name_type_map[key - TIFF_##TYPE##_KEY_ID_OFFSET].element;
  121. static const char *get_geokey_name(int key)
  122. {
  123. RET_GEOKEY(VERT, vert, name);
  124. RET_GEOKEY(PROJ, proj, name);
  125. RET_GEOKEY(GEOG, geog, name);
  126. RET_GEOKEY(CONF, conf, name);
  127. return NULL;
  128. }
  129. static int get_geokey_type(int key)
  130. {
  131. RET_GEOKEY(VERT, vert, type);
  132. RET_GEOKEY(PROJ, proj, type);
  133. RET_GEOKEY(GEOG, geog, type);
  134. RET_GEOKEY(CONF, conf, type);
  135. return AVERROR_INVALIDDATA;
  136. }
  137. static int cmp_id_key(const void *id, const void *k)
  138. {
  139. return *(const int*)id - ((const TiffGeoTagKeyName*)k)->key;
  140. }
  141. static const char *search_keyval(const TiffGeoTagKeyName *keys, int n, int id)
  142. {
  143. TiffGeoTagKeyName *r = bsearch(&id, keys, n, sizeof(keys[0]), cmp_id_key);
  144. if(r)
  145. return r->name;
  146. return NULL;
  147. }
  148. static char *get_geokey_val(int key, int val)
  149. {
  150. char *ap;
  151. if (val == TIFF_GEO_KEY_UNDEFINED)
  152. return av_strdup("undefined");
  153. if (val == TIFF_GEO_KEY_USER_DEFINED)
  154. return av_strdup("User-Defined");
  155. #define RET_GEOKEY_VAL(TYPE, array)\
  156. if (val >= TIFF_##TYPE##_OFFSET &&\
  157. val - TIFF_##TYPE##_OFFSET < FF_ARRAY_ELEMS(tiff_##array##_codes))\
  158. return av_strdup(tiff_##array##_codes[val - TIFF_##TYPE##_OFFSET]);
  159. switch (key) {
  160. case TIFF_GT_MODEL_TYPE_GEOKEY:
  161. RET_GEOKEY_VAL(GT_MODEL_TYPE, gt_model_type);
  162. break;
  163. case TIFF_GT_RASTER_TYPE_GEOKEY:
  164. RET_GEOKEY_VAL(GT_RASTER_TYPE, gt_raster_type);
  165. break;
  166. case TIFF_GEOG_LINEAR_UNITS_GEOKEY:
  167. case TIFF_PROJ_LINEAR_UNITS_GEOKEY:
  168. case TIFF_VERTICAL_UNITS_GEOKEY:
  169. RET_GEOKEY_VAL(LINEAR_UNIT, linear_unit);
  170. break;
  171. case TIFF_GEOG_ANGULAR_UNITS_GEOKEY:
  172. case TIFF_GEOG_AZIMUTH_UNITS_GEOKEY:
  173. RET_GEOKEY_VAL(ANGULAR_UNIT, angular_unit);
  174. break;
  175. case TIFF_GEOGRAPHIC_TYPE_GEOKEY:
  176. RET_GEOKEY_VAL(GCS_TYPE, gcs_type);
  177. RET_GEOKEY_VAL(GCSE_TYPE, gcse_type);
  178. break;
  179. case TIFF_GEOG_GEODETIC_DATUM_GEOKEY:
  180. RET_GEOKEY_VAL(GEODETIC_DATUM, geodetic_datum);
  181. RET_GEOKEY_VAL(GEODETIC_DATUM_E, geodetic_datum_e);
  182. break;
  183. case TIFF_GEOG_ELLIPSOID_GEOKEY:
  184. RET_GEOKEY_VAL(ELLIPSOID, ellipsoid);
  185. break;
  186. case TIFF_GEOG_PRIME_MERIDIAN_GEOKEY:
  187. RET_GEOKEY_VAL(PRIME_MERIDIAN, prime_meridian);
  188. break;
  189. case TIFF_PROJECTED_CS_TYPE_GEOKEY:
  190. ap = av_strdup(search_keyval(tiff_proj_cs_type_codes, FF_ARRAY_ELEMS(tiff_proj_cs_type_codes), val));
  191. if(ap) return ap;
  192. break;
  193. case TIFF_PROJECTION_GEOKEY:
  194. ap = av_strdup(search_keyval(tiff_projection_codes, FF_ARRAY_ELEMS(tiff_projection_codes), val));
  195. if(ap) return ap;
  196. break;
  197. case TIFF_PROJ_COORD_TRANS_GEOKEY:
  198. RET_GEOKEY_VAL(COORD_TRANS, coord_trans);
  199. break;
  200. case TIFF_VERTICAL_CS_TYPE_GEOKEY:
  201. RET_GEOKEY_VAL(VERT_CS, vert_cs);
  202. RET_GEOKEY_VAL(ORTHO_VERT_CS, ortho_vert_cs);
  203. break;
  204. }
  205. ap = av_malloc(14);
  206. if (ap)
  207. snprintf(ap, 14, "Unknown-%d", val);
  208. return ap;
  209. }
  210. static char *doubles2str(double *dp, int count, const char *sep)
  211. {
  212. int i;
  213. char *ap, *ap0;
  214. uint64_t component_len;
  215. if (!sep) sep = ", ";
  216. component_len = 24LL + strlen(sep);
  217. if (count >= (INT_MAX - 1)/component_len)
  218. return NULL;
  219. ap = av_malloc(component_len * count + 1);
  220. if (!ap)
  221. return NULL;
  222. ap0 = ap;
  223. ap[0] = '\0';
  224. for (i = 0; i < count; i++) {
  225. unsigned l = snprintf(ap, component_len, "%.15g%s", dp[i], sep);
  226. if(l >= component_len) {
  227. av_free(ap0);
  228. return NULL;
  229. }
  230. ap += l;
  231. }
  232. ap0[strlen(ap0) - strlen(sep)] = '\0';
  233. return ap0;
  234. }
  235. static int add_metadata(int count, int type,
  236. const char *name, const char *sep, TiffContext *s, AVFrame *frame)
  237. {
  238. switch(type) {
  239. case TIFF_DOUBLE: return ff_tadd_doubles_metadata(count, name, sep, &s->gb, s->le, &frame->metadata);
  240. case TIFF_SHORT : return ff_tadd_shorts_metadata(count, name, sep, &s->gb, s->le, 0, &frame->metadata);
  241. case TIFF_STRING: return ff_tadd_string_metadata(count, name, &s->gb, s->le, &frame->metadata);
  242. default : return AVERROR_INVALIDDATA;
  243. };
  244. }
  245. static void av_always_inline dng_blit(TiffContext *s, uint8_t *dst, int dst_stride,
  246. const uint8_t *src, int src_stride, int width, int height,
  247. int is_single_comp, int is_u16);
  248. static void av_always_inline horizontal_fill(TiffContext *s,
  249. unsigned int bpp, uint8_t* dst,
  250. int usePtr, const uint8_t *src,
  251. uint8_t c, int width, int offset)
  252. {
  253. switch (bpp) {
  254. case 1:
  255. while (--width >= 0) {
  256. dst[(width+offset)*8+7] = (usePtr ? src[width] : c) & 0x1;
  257. dst[(width+offset)*8+6] = (usePtr ? src[width] : c) >> 1 & 0x1;
  258. dst[(width+offset)*8+5] = (usePtr ? src[width] : c) >> 2 & 0x1;
  259. dst[(width+offset)*8+4] = (usePtr ? src[width] : c) >> 3 & 0x1;
  260. dst[(width+offset)*8+3] = (usePtr ? src[width] : c) >> 4 & 0x1;
  261. dst[(width+offset)*8+2] = (usePtr ? src[width] : c) >> 5 & 0x1;
  262. dst[(width+offset)*8+1] = (usePtr ? src[width] : c) >> 6 & 0x1;
  263. dst[(width+offset)*8+0] = (usePtr ? src[width] : c) >> 7;
  264. }
  265. break;
  266. case 2:
  267. while (--width >= 0) {
  268. dst[(width+offset)*4+3] = (usePtr ? src[width] : c) & 0x3;
  269. dst[(width+offset)*4+2] = (usePtr ? src[width] : c) >> 2 & 0x3;
  270. dst[(width+offset)*4+1] = (usePtr ? src[width] : c) >> 4 & 0x3;
  271. dst[(width+offset)*4+0] = (usePtr ? src[width] : c) >> 6;
  272. }
  273. break;
  274. case 4:
  275. while (--width >= 0) {
  276. dst[(width+offset)*2+1] = (usePtr ? src[width] : c) & 0xF;
  277. dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
  278. }
  279. break;
  280. case 10:
  281. case 12:
  282. case 14: {
  283. uint16_t *dst16 = (uint16_t *)dst;
  284. int is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
  285. uint8_t shift = is_dng ? 0 : 16 - bpp;
  286. GetBitContext gb;
  287. init_get_bits8(&gb, src, width);
  288. for (int i = 0; i < s->width; i++) {
  289. dst16[i] = get_bits(&gb, bpp) << shift;
  290. }
  291. }
  292. break;
  293. default:
  294. if (usePtr) {
  295. memcpy(dst + offset, src, width);
  296. } else {
  297. memset(dst + offset, c, width);
  298. }
  299. }
  300. }
  301. static int deinvert_buffer(TiffContext *s, const uint8_t *src, int size)
  302. {
  303. int i;
  304. av_fast_padded_malloc(&s->deinvert_buf, &s->deinvert_buf_size, size);
  305. if (!s->deinvert_buf)
  306. return AVERROR(ENOMEM);
  307. for (i = 0; i < size; i++)
  308. s->deinvert_buf[i] = ff_reverse[src[i]];
  309. return 0;
  310. }
  311. static void unpack_gray(TiffContext *s, AVFrame *p,
  312. const uint8_t *src, int lnum, int width, int bpp)
  313. {
  314. GetBitContext gb;
  315. uint16_t *dst = (uint16_t *)(p->data[0] + lnum * p->linesize[0]);
  316. init_get_bits8(&gb, src, width);
  317. for (int i = 0; i < s->width; i++) {
  318. dst[i] = get_bits(&gb, bpp);
  319. }
  320. }
  321. static void unpack_yuv(TiffContext *s, AVFrame *p,
  322. const uint8_t *src, int lnum)
  323. {
  324. int i, j, k;
  325. int w = (s->width - 1) / s->subsampling[0] + 1;
  326. uint8_t *pu = &p->data[1][lnum / s->subsampling[1] * p->linesize[1]];
  327. uint8_t *pv = &p->data[2][lnum / s->subsampling[1] * p->linesize[2]];
  328. if (s->width % s->subsampling[0] || s->height % s->subsampling[1]) {
  329. for (i = 0; i < w; i++) {
  330. for (j = 0; j < s->subsampling[1]; j++)
  331. for (k = 0; k < s->subsampling[0]; k++)
  332. p->data[0][FFMIN(lnum + j, s->height-1) * p->linesize[0] +
  333. FFMIN(i * s->subsampling[0] + k, s->width-1)] = *src++;
  334. *pu++ = *src++;
  335. *pv++ = *src++;
  336. }
  337. }else{
  338. for (i = 0; i < w; i++) {
  339. for (j = 0; j < s->subsampling[1]; j++)
  340. for (k = 0; k < s->subsampling[0]; k++)
  341. p->data[0][(lnum + j) * p->linesize[0] +
  342. i * s->subsampling[0] + k] = *src++;
  343. *pu++ = *src++;
  344. *pv++ = *src++;
  345. }
  346. }
  347. }
  348. #if CONFIG_ZLIB
  349. static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
  350. int size)
  351. {
  352. z_stream zstream = { 0 };
  353. int zret;
  354. zstream.next_in = src;
  355. zstream.avail_in = size;
  356. zstream.next_out = dst;
  357. zstream.avail_out = *len;
  358. zret = inflateInit(&zstream);
  359. if (zret != Z_OK) {
  360. av_log(NULL, AV_LOG_ERROR, "Inflate init error: %d\n", zret);
  361. return zret;
  362. }
  363. zret = inflate(&zstream, Z_SYNC_FLUSH);
  364. inflateEnd(&zstream);
  365. *len = zstream.total_out;
  366. return zret == Z_STREAM_END ? Z_OK : zret;
  367. }
  368. static int tiff_unpack_zlib(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  369. const uint8_t *src, int size, int width, int lines,
  370. int strip_start, int is_yuv)
  371. {
  372. uint8_t *zbuf;
  373. unsigned long outlen;
  374. int ret, line;
  375. outlen = width * lines;
  376. zbuf = av_malloc(outlen);
  377. if (!zbuf)
  378. return AVERROR(ENOMEM);
  379. if (s->fill_order) {
  380. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  381. av_free(zbuf);
  382. return ret;
  383. }
  384. src = s->deinvert_buf;
  385. }
  386. ret = tiff_uncompress(zbuf, &outlen, src, size);
  387. if (ret != Z_OK) {
  388. av_log(s->avctx, AV_LOG_ERROR,
  389. "Uncompressing failed (%lu of %lu) with error %d\n", outlen,
  390. (unsigned long)width * lines, ret);
  391. av_free(zbuf);
  392. return AVERROR_UNKNOWN;
  393. }
  394. src = zbuf;
  395. for (line = 0; line < lines; line++) {
  396. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  397. horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
  398. } else {
  399. memcpy(dst, src, width);
  400. }
  401. if (is_yuv) {
  402. unpack_yuv(s, p, dst, strip_start + line);
  403. line += s->subsampling[1] - 1;
  404. }
  405. dst += stride;
  406. src += width;
  407. }
  408. av_free(zbuf);
  409. return 0;
  410. }
  411. #endif
  412. #if CONFIG_LZMA
  413. static int tiff_uncompress_lzma(uint8_t *dst, uint64_t *len, const uint8_t *src,
  414. int size)
  415. {
  416. lzma_stream stream = LZMA_STREAM_INIT;
  417. lzma_ret ret;
  418. stream.next_in = (uint8_t *)src;
  419. stream.avail_in = size;
  420. stream.next_out = dst;
  421. stream.avail_out = *len;
  422. ret = lzma_stream_decoder(&stream, UINT64_MAX, 0);
  423. if (ret != LZMA_OK) {
  424. av_log(NULL, AV_LOG_ERROR, "LZMA init error: %d\n", ret);
  425. return ret;
  426. }
  427. ret = lzma_code(&stream, LZMA_RUN);
  428. lzma_end(&stream);
  429. *len = stream.total_out;
  430. return ret == LZMA_STREAM_END ? LZMA_OK : ret;
  431. }
  432. static int tiff_unpack_lzma(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  433. const uint8_t *src, int size, int width, int lines,
  434. int strip_start, int is_yuv)
  435. {
  436. uint64_t outlen = width * (uint64_t)lines;
  437. int ret, line;
  438. uint8_t *buf = av_malloc(outlen);
  439. if (!buf)
  440. return AVERROR(ENOMEM);
  441. if (s->fill_order) {
  442. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  443. av_free(buf);
  444. return ret;
  445. }
  446. src = s->deinvert_buf;
  447. }
  448. ret = tiff_uncompress_lzma(buf, &outlen, src, size);
  449. if (ret != LZMA_OK) {
  450. av_log(s->avctx, AV_LOG_ERROR,
  451. "Uncompressing failed (%"PRIu64" of %"PRIu64") with error %d\n", outlen,
  452. (uint64_t)width * lines, ret);
  453. av_free(buf);
  454. return AVERROR_UNKNOWN;
  455. }
  456. src = buf;
  457. for (line = 0; line < lines; line++) {
  458. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  459. horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
  460. } else {
  461. memcpy(dst, src, width);
  462. }
  463. if (is_yuv) {
  464. unpack_yuv(s, p, dst, strip_start + line);
  465. line += s->subsampling[1] - 1;
  466. }
  467. dst += stride;
  468. src += width;
  469. }
  470. av_free(buf);
  471. return 0;
  472. }
  473. #endif
  474. static int tiff_unpack_fax(TiffContext *s, uint8_t *dst, int stride,
  475. const uint8_t *src, int size, int width, int lines)
  476. {
  477. int i, ret = 0;
  478. int line;
  479. uint8_t *src2;
  480. av_fast_padded_malloc(&s->fax_buffer, &s->fax_buffer_size, size);
  481. src2 = s->fax_buffer;
  482. if (!src2) {
  483. av_log(s->avctx, AV_LOG_ERROR,
  484. "Error allocating temporary buffer\n");
  485. return AVERROR(ENOMEM);
  486. }
  487. if (!s->fill_order) {
  488. memcpy(src2, src, size);
  489. } else {
  490. for (i = 0; i < size; i++)
  491. src2[i] = ff_reverse[src[i]];
  492. }
  493. memset(src2 + size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
  494. ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride,
  495. s->compr, s->fax_opts);
  496. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  497. for (line = 0; line < lines; line++) {
  498. horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
  499. dst += stride;
  500. }
  501. return ret;
  502. }
  503. static int dng_decode_strip(AVCodecContext *avctx, AVFrame *frame);
  504. static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  505. const uint8_t *src, int size, int strip_start, int lines)
  506. {
  507. PutByteContext pb;
  508. int c, line, pixels, code, ret;
  509. const uint8_t *ssrc = src;
  510. int width = ((s->width * s->bpp) + 7) >> 3;
  511. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(p->format);
  512. int is_yuv = !(desc->flags & AV_PIX_FMT_FLAG_RGB) &&
  513. (desc->flags & AV_PIX_FMT_FLAG_PLANAR) &&
  514. desc->nb_components >= 3;
  515. int is_dng;
  516. if (s->planar)
  517. width /= s->bppcount;
  518. if (size <= 0)
  519. return AVERROR_INVALIDDATA;
  520. if (is_yuv) {
  521. int bytes_per_row = (((s->width - 1) / s->subsampling[0] + 1) * s->bpp *
  522. s->subsampling[0] * s->subsampling[1] + 7) >> 3;
  523. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row);
  524. if (s->yuv_line == NULL) {
  525. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  526. return AVERROR(ENOMEM);
  527. }
  528. dst = s->yuv_line;
  529. stride = 0;
  530. width = (s->width - 1) / s->subsampling[0] + 1;
  531. width = width * s->subsampling[0] * s->subsampling[1] + 2*width;
  532. av_assert0(width <= bytes_per_row);
  533. av_assert0(s->bpp == 24);
  534. }
  535. if (s->is_bayer) {
  536. av_assert0(width == (s->bpp * s->width + 7) >> 3);
  537. }
  538. if (p->format == AV_PIX_FMT_GRAY12) {
  539. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, width);
  540. if (s->yuv_line == NULL) {
  541. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  542. return AVERROR(ENOMEM);
  543. }
  544. dst = s->yuv_line;
  545. stride = 0;
  546. }
  547. if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
  548. #if CONFIG_ZLIB
  549. return tiff_unpack_zlib(s, p, dst, stride, src, size, width, lines,
  550. strip_start, is_yuv);
  551. #else
  552. av_log(s->avctx, AV_LOG_ERROR,
  553. "zlib support not enabled, "
  554. "deflate compression not supported\n");
  555. return AVERROR(ENOSYS);
  556. #endif
  557. }
  558. if (s->compr == TIFF_LZMA) {
  559. #if CONFIG_LZMA
  560. return tiff_unpack_lzma(s, p, dst, stride, src, size, width, lines,
  561. strip_start, is_yuv);
  562. #else
  563. av_log(s->avctx, AV_LOG_ERROR,
  564. "LZMA support not enabled\n");
  565. return AVERROR(ENOSYS);
  566. #endif
  567. }
  568. if (s->compr == TIFF_LZW) {
  569. if (s->fill_order) {
  570. if ((ret = deinvert_buffer(s, src, size)) < 0)
  571. return ret;
  572. ssrc = src = s->deinvert_buf;
  573. }
  574. if (size > 1 && !src[0] && (src[1]&1)) {
  575. av_log(s->avctx, AV_LOG_ERROR, "Old style LZW is unsupported\n");
  576. }
  577. if ((ret = ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF)) < 0) {
  578. av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
  579. return ret;
  580. }
  581. for (line = 0; line < lines; line++) {
  582. pixels = ff_lzw_decode(s->lzw, dst, width);
  583. if (pixels < width) {
  584. av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
  585. pixels, width);
  586. return AVERROR_INVALIDDATA;
  587. }
  588. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  589. horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
  590. if (is_yuv) {
  591. unpack_yuv(s, p, dst, strip_start + line);
  592. line += s->subsampling[1] - 1;
  593. } else if (p->format == AV_PIX_FMT_GRAY12) {
  594. unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
  595. }
  596. dst += stride;
  597. }
  598. return 0;
  599. }
  600. if (s->compr == TIFF_CCITT_RLE ||
  601. s->compr == TIFF_G3 ||
  602. s->compr == TIFF_G4) {
  603. if (is_yuv || p->format == AV_PIX_FMT_GRAY12)
  604. return AVERROR_INVALIDDATA;
  605. return tiff_unpack_fax(s, dst, stride, src, size, width, lines);
  606. }
  607. bytestream2_init(&s->gb, src, size);
  608. bytestream2_init_writer(&pb, dst, is_yuv ? s->yuv_line_size : (stride * lines));
  609. is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
  610. /* Decode JPEG-encoded DNGs with strips */
  611. if (s->compr == TIFF_NEWJPEG && is_dng) {
  612. if (s->strips > 1) {
  613. av_log(s->avctx, AV_LOG_ERROR, "More than one DNG JPEG strips unsupported\n");
  614. return AVERROR_PATCHWELCOME;
  615. }
  616. if ((ret = dng_decode_strip(s->avctx, p)) < 0)
  617. return ret;
  618. return 0;
  619. }
  620. if (is_dng && stride == 0)
  621. return AVERROR_INVALIDDATA;
  622. for (line = 0; line < lines; line++) {
  623. if (src - ssrc > size) {
  624. av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
  625. return AVERROR_INVALIDDATA;
  626. }
  627. if (bytestream2_get_bytes_left(&s->gb) == 0 || bytestream2_get_eof(&pb))
  628. break;
  629. bytestream2_seek_p(&pb, stride * line, SEEK_SET);
  630. switch (s->compr) {
  631. case TIFF_RAW:
  632. if (ssrc + size - src < width)
  633. return AVERROR_INVALIDDATA;
  634. if (!s->fill_order) {
  635. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 || s->is_bayer),
  636. dst, 1, src, 0, width, 0);
  637. } else {
  638. int i;
  639. for (i = 0; i < width; i++)
  640. dst[i] = ff_reverse[src[i]];
  641. }
  642. /* Color processing for DNG images with uncompressed strips (non-tiled) */
  643. if (is_dng) {
  644. int is_u16, pixel_size_bytes, pixel_size_bits, elements;
  645. is_u16 = (s->bpp / s->bppcount > 8);
  646. pixel_size_bits = (is_u16 ? 16 : 8);
  647. pixel_size_bytes = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
  648. elements = width / pixel_size_bytes * pixel_size_bits / s->bpp * s->bppcount; // need to account for [1, 16] bpp
  649. av_assert0 (elements * pixel_size_bytes <= FFABS(stride));
  650. dng_blit(s,
  651. dst,
  652. 0, // no stride, only 1 line
  653. dst,
  654. 0, // no stride, only 1 line
  655. elements,
  656. 1,
  657. 0, // single-component variation is only preset in JPEG-encoded DNGs
  658. is_u16);
  659. }
  660. src += width;
  661. break;
  662. case TIFF_PACKBITS:
  663. for (pixels = 0; pixels < width;) {
  664. if (ssrc + size - src < 2) {
  665. av_log(s->avctx, AV_LOG_ERROR, "Read went out of bounds\n");
  666. return AVERROR_INVALIDDATA;
  667. }
  668. code = s->fill_order ? (int8_t) ff_reverse[*src++]: (int8_t) *src++;
  669. if (code >= 0) {
  670. code++;
  671. if (pixels + code > width ||
  672. ssrc + size - src < code) {
  673. av_log(s->avctx, AV_LOG_ERROR,
  674. "Copy went out of bounds\n");
  675. return AVERROR_INVALIDDATA;
  676. }
  677. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  678. dst, 1, src, 0, code, pixels);
  679. src += code;
  680. pixels += code;
  681. } else if (code != -128) { // -127..-1
  682. code = (-code) + 1;
  683. if (pixels + code > width) {
  684. av_log(s->avctx, AV_LOG_ERROR,
  685. "Run went out of bounds\n");
  686. return AVERROR_INVALIDDATA;
  687. }
  688. c = *src++;
  689. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  690. dst, 0, NULL, c, code, pixels);
  691. pixels += code;
  692. }
  693. }
  694. if (s->fill_order) {
  695. int i;
  696. for (i = 0; i < width; i++)
  697. dst[i] = ff_reverse[dst[i]];
  698. }
  699. break;
  700. }
  701. if (is_yuv) {
  702. unpack_yuv(s, p, dst, strip_start + line);
  703. line += s->subsampling[1] - 1;
  704. } else if (p->format == AV_PIX_FMT_GRAY12) {
  705. unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
  706. }
  707. dst += stride;
  708. }
  709. return 0;
  710. }
  711. /**
  712. * Map stored raw sensor values into linear reference values (see: DNG Specification - Chapter 5)
  713. */
  714. static uint16_t av_always_inline dng_process_color16(uint16_t value,
  715. const uint16_t *lut,
  716. uint16_t black_level,
  717. float scale_factor) {
  718. float value_norm;
  719. // Lookup table lookup
  720. if (lut)
  721. value = lut[value];
  722. // Black level subtraction
  723. value = av_clip_uint16_c((unsigned)value - black_level);
  724. // Color scaling
  725. value_norm = (float)value * scale_factor;
  726. value = av_clip_uint16_c(value_norm * 65535);
  727. return value;
  728. }
  729. static uint16_t av_always_inline dng_process_color8(uint16_t value,
  730. const uint16_t *lut,
  731. uint16_t black_level,
  732. float scale_factor) {
  733. return dng_process_color16(value, lut, black_level, scale_factor) >> 8;
  734. }
  735. static void dng_blit(TiffContext *s, uint8_t *dst, int dst_stride,
  736. const uint8_t *src, int src_stride,
  737. int width, int height, int is_single_comp, int is_u16)
  738. {
  739. int line, col;
  740. float scale_factor;
  741. scale_factor = 1.0f / (s->white_level - s->black_level);
  742. if (is_single_comp) {
  743. if (!is_u16)
  744. return; /* <= 8bpp unsupported */
  745. /* Image is double the width and half the height we need, each row comprises 2 rows of the output
  746. (split vertically in the middle). */
  747. for (line = 0; line < height / 2; line++) {
  748. uint16_t *dst_u16 = (uint16_t *)dst;
  749. uint16_t *src_u16 = (uint16_t *)src;
  750. /* Blit first half of input row row to initial row of output */
  751. for (col = 0; col < width; col++)
  752. *dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
  753. /* Advance the destination pointer by a row (source pointer remains in the same place) */
  754. dst += dst_stride * sizeof(uint16_t);
  755. dst_u16 = (uint16_t *)dst;
  756. /* Blit second half of input row row to next row of output */
  757. for (col = 0; col < width; col++)
  758. *dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
  759. dst += dst_stride * sizeof(uint16_t);
  760. src += src_stride * sizeof(uint16_t);
  761. }
  762. } else {
  763. /* Input and output image are the same size and the MJpeg decoder has done per-component
  764. deinterleaving, so blitting here is straightforward. */
  765. if (is_u16) {
  766. for (line = 0; line < height; line++) {
  767. uint16_t *dst_u16 = (uint16_t *)dst;
  768. uint16_t *src_u16 = (uint16_t *)src;
  769. for (col = 0; col < width; col++)
  770. *dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
  771. dst += dst_stride * sizeof(uint16_t);
  772. src += src_stride * sizeof(uint16_t);
  773. }
  774. } else {
  775. for (line = 0; line < height; line++) {
  776. uint8_t *dst_u8 = dst;
  777. const uint8_t *src_u8 = src;
  778. for (col = 0; col < width; col++)
  779. *dst_u8++ = dng_process_color8(*src_u8++, s->dng_lut, s->black_level, scale_factor);
  780. dst += dst_stride;
  781. src += src_stride;
  782. }
  783. }
  784. }
  785. }
  786. static int dng_decode_jpeg(AVCodecContext *avctx, AVFrame *frame,
  787. int tile_byte_count, int dst_x, int dst_y, int w, int h)
  788. {
  789. TiffContext *s = avctx->priv_data;
  790. uint8_t *dst_data, *src_data;
  791. uint32_t dst_offset; /* offset from dst buffer in pixels */
  792. int is_single_comp, is_u16, pixel_size;
  793. int ret;
  794. if (tile_byte_count < 0 || tile_byte_count > bytestream2_get_bytes_left(&s->gb))
  795. return AVERROR_INVALIDDATA;
  796. /* Prepare a packet and send to the MJPEG decoder */
  797. av_packet_unref(s->jpkt);
  798. s->jpkt->data = (uint8_t*)s->gb.buffer;
  799. s->jpkt->size = tile_byte_count;
  800. if (s->is_bayer) {
  801. MJpegDecodeContext *mjpegdecctx = s->avctx_mjpeg->priv_data;
  802. /* We have to set this information here, there is no way to know if a given JPEG is a DNG-embedded
  803. image or not from its own data (and we need that information when decoding it). */
  804. mjpegdecctx->bayer = 1;
  805. }
  806. ret = avcodec_send_packet(s->avctx_mjpeg, s->jpkt);
  807. if (ret < 0) {
  808. av_log(avctx, AV_LOG_ERROR, "Error submitting a packet for decoding\n");
  809. return ret;
  810. }
  811. ret = avcodec_receive_frame(s->avctx_mjpeg, s->jpgframe);
  812. if (ret < 0) {
  813. av_log(avctx, AV_LOG_ERROR, "JPEG decoding error: %s.\n", av_err2str(ret));
  814. /* Normally skip, error if explode */
  815. if (avctx->err_recognition & AV_EF_EXPLODE)
  816. return AVERROR_INVALIDDATA;
  817. else
  818. return 0;
  819. }
  820. is_u16 = (s->bpp > 8);
  821. /* Copy the outputted tile's pixels from 'jpgframe' to 'frame' (final buffer) */
  822. if (s->jpgframe->width != s->avctx_mjpeg->width ||
  823. s->jpgframe->height != s->avctx_mjpeg->height ||
  824. s->jpgframe->format != s->avctx_mjpeg->pix_fmt)
  825. return AVERROR_INVALIDDATA;
  826. /* See dng_blit for explanation */
  827. if (s->avctx_mjpeg->width == w * 2 &&
  828. s->avctx_mjpeg->height == h / 2 &&
  829. s->avctx_mjpeg->pix_fmt == AV_PIX_FMT_GRAY16LE) {
  830. is_single_comp = 1;
  831. } else if (s->avctx_mjpeg->width >= w &&
  832. s->avctx_mjpeg->height >= h &&
  833. s->avctx_mjpeg->pix_fmt == (is_u16 ? AV_PIX_FMT_GRAY16 : AV_PIX_FMT_GRAY8)
  834. ) {
  835. is_single_comp = 0;
  836. } else
  837. return AVERROR_INVALIDDATA;
  838. pixel_size = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
  839. if (is_single_comp && !is_u16) {
  840. av_log(s->avctx, AV_LOG_ERROR, "DNGs with bpp <= 8 and 1 component are unsupported\n");
  841. av_frame_unref(s->jpgframe);
  842. return AVERROR_PATCHWELCOME;
  843. }
  844. dst_offset = dst_x + frame->linesize[0] * dst_y / pixel_size;
  845. dst_data = frame->data[0] + dst_offset * pixel_size;
  846. src_data = s->jpgframe->data[0];
  847. dng_blit(s,
  848. dst_data,
  849. frame->linesize[0] / pixel_size,
  850. src_data,
  851. s->jpgframe->linesize[0] / pixel_size,
  852. w,
  853. h,
  854. is_single_comp,
  855. is_u16);
  856. av_frame_unref(s->jpgframe);
  857. return 0;
  858. }
  859. static int dng_decode_tiles(AVCodecContext *avctx, AVFrame *frame,
  860. const AVPacket *avpkt)
  861. {
  862. TiffContext *s = avctx->priv_data;
  863. int tile_idx;
  864. int tile_offset_offset, tile_offset;
  865. int tile_byte_count_offset, tile_byte_count;
  866. int tile_count_x, tile_count_y;
  867. int tile_width, tile_length;
  868. int has_width_leftover, has_height_leftover;
  869. int tile_x = 0, tile_y = 0;
  870. int pos_x = 0, pos_y = 0;
  871. int ret;
  872. s->jpgframe->width = s->tile_width;
  873. s->jpgframe->height = s->tile_length;
  874. s->avctx_mjpeg->width = s->tile_width;
  875. s->avctx_mjpeg->height = s->tile_length;
  876. has_width_leftover = (s->width % s->tile_width != 0);
  877. has_height_leftover = (s->height % s->tile_length != 0);
  878. /* Calculate tile counts (round up) */
  879. tile_count_x = (s->width + s->tile_width - 1) / s->tile_width;
  880. tile_count_y = (s->height + s->tile_length - 1) / s->tile_length;
  881. /* Iterate over the number of tiles */
  882. for (tile_idx = 0; tile_idx < s->tile_count; tile_idx++) {
  883. tile_x = tile_idx % tile_count_x;
  884. tile_y = tile_idx / tile_count_x;
  885. if (has_width_leftover && tile_x == tile_count_x - 1) // If on the right-most tile
  886. tile_width = s->width % s->tile_width;
  887. else
  888. tile_width = s->tile_width;
  889. if (has_height_leftover && tile_y == tile_count_y - 1) // If on the bottom-most tile
  890. tile_length = s->height % s->tile_length;
  891. else
  892. tile_length = s->tile_length;
  893. /* Read tile offset */
  894. tile_offset_offset = s->tile_offsets_offset + tile_idx * sizeof(int);
  895. bytestream2_seek(&s->gb, tile_offset_offset, SEEK_SET);
  896. tile_offset = ff_tget_long(&s->gb, s->le);
  897. /* Read tile byte size */
  898. tile_byte_count_offset = s->tile_byte_counts_offset + tile_idx * sizeof(int);
  899. bytestream2_seek(&s->gb, tile_byte_count_offset, SEEK_SET);
  900. tile_byte_count = ff_tget_long(&s->gb, s->le);
  901. /* Seek to tile data */
  902. bytestream2_seek(&s->gb, tile_offset, SEEK_SET);
  903. /* Decode JPEG tile and copy it in the reference frame */
  904. ret = dng_decode_jpeg(avctx, frame, tile_byte_count, pos_x, pos_y, tile_width, tile_length);
  905. if (ret < 0)
  906. return ret;
  907. /* Advance current positions */
  908. pos_x += tile_width;
  909. if (tile_x == tile_count_x - 1) { // If on the right edge
  910. pos_x = 0;
  911. pos_y += tile_length;
  912. }
  913. }
  914. /* Frame is ready to be output */
  915. frame->pict_type = AV_PICTURE_TYPE_I;
  916. frame->key_frame = 1;
  917. return avpkt->size;
  918. }
  919. static int dng_decode_strip(AVCodecContext *avctx, AVFrame *frame)
  920. {
  921. TiffContext *s = avctx->priv_data;
  922. s->jpgframe->width = s->width;
  923. s->jpgframe->height = s->height;
  924. s->avctx_mjpeg->width = s->width;
  925. s->avctx_mjpeg->height = s->height;
  926. return dng_decode_jpeg(avctx, frame, s->stripsize, 0, 0, s->width, s->height);
  927. }
  928. static int init_image(TiffContext *s, ThreadFrame *frame)
  929. {
  930. int ret;
  931. int create_gray_palette = 0;
  932. // make sure there is no aliasing in the following switch
  933. if (s->bpp >= 100 || s->bppcount >= 10) {
  934. av_log(s->avctx, AV_LOG_ERROR,
  935. "Unsupported image parameters: bpp=%d, bppcount=%d\n",
  936. s->bpp, s->bppcount);
  937. return AVERROR_INVALIDDATA;
  938. }
  939. switch (s->planar * 1000 + s->bpp * 10 + s->bppcount + s->is_bayer * 10000) {
  940. case 11:
  941. if (!s->palette_is_set) {
  942. s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK;
  943. break;
  944. }
  945. case 21:
  946. case 41:
  947. s->avctx->pix_fmt = AV_PIX_FMT_PAL8;
  948. if (!s->palette_is_set) {
  949. create_gray_palette = 1;
  950. }
  951. break;
  952. case 81:
  953. s->avctx->pix_fmt = s->palette_is_set ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8;
  954. break;
  955. case 121:
  956. s->avctx->pix_fmt = AV_PIX_FMT_GRAY12;
  957. break;
  958. case 10081:
  959. switch (AV_RL32(s->pattern)) {
  960. case 0x02010100:
  961. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB8;
  962. break;
  963. case 0x00010102:
  964. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR8;
  965. break;
  966. case 0x01000201:
  967. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG8;
  968. break;
  969. case 0x01020001:
  970. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG8;
  971. break;
  972. default:
  973. av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
  974. AV_RL32(s->pattern));
  975. return AVERROR_PATCHWELCOME;
  976. }
  977. break;
  978. case 10101:
  979. case 10121:
  980. case 10141:
  981. case 10161:
  982. switch (AV_RL32(s->pattern)) {
  983. case 0x02010100:
  984. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB16;
  985. break;
  986. case 0x00010102:
  987. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR16;
  988. break;
  989. case 0x01000201:
  990. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG16;
  991. break;
  992. case 0x01020001:
  993. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG16;
  994. break;
  995. default:
  996. av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
  997. AV_RL32(s->pattern));
  998. return AVERROR_PATCHWELCOME;
  999. }
  1000. break;
  1001. case 243:
  1002. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1003. if (s->subsampling[0] == 1 && s->subsampling[1] == 1) {
  1004. s->avctx->pix_fmt = AV_PIX_FMT_YUV444P;
  1005. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 1) {
  1006. s->avctx->pix_fmt = AV_PIX_FMT_YUV422P;
  1007. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 1) {
  1008. s->avctx->pix_fmt = AV_PIX_FMT_YUV411P;
  1009. } else if (s->subsampling[0] == 1 && s->subsampling[1] == 2) {
  1010. s->avctx->pix_fmt = AV_PIX_FMT_YUV440P;
  1011. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 2) {
  1012. s->avctx->pix_fmt = AV_PIX_FMT_YUV420P;
  1013. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 4) {
  1014. s->avctx->pix_fmt = AV_PIX_FMT_YUV410P;
  1015. } else {
  1016. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr subsampling\n");
  1017. return AVERROR_PATCHWELCOME;
  1018. }
  1019. } else
  1020. s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
  1021. break;
  1022. case 161:
  1023. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE;
  1024. break;
  1025. case 162:
  1026. s->avctx->pix_fmt = AV_PIX_FMT_YA8;
  1027. break;
  1028. case 322:
  1029. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_YA16LE : AV_PIX_FMT_YA16BE;
  1030. break;
  1031. case 324:
  1032. s->avctx->pix_fmt = s->photometric == TIFF_PHOTOMETRIC_SEPARATED ? AV_PIX_FMT_RGB0 : AV_PIX_FMT_RGBA;
  1033. break;
  1034. case 405:
  1035. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED)
  1036. s->avctx->pix_fmt = AV_PIX_FMT_RGBA;
  1037. else {
  1038. av_log(s->avctx, AV_LOG_ERROR,
  1039. "bpp=40 without PHOTOMETRIC_SEPARATED is unsupported\n");
  1040. return AVERROR_PATCHWELCOME;
  1041. }
  1042. break;
  1043. case 483:
  1044. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE;
  1045. break;
  1046. case 644:
  1047. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGBA64LE : AV_PIX_FMT_RGBA64BE;
  1048. break;
  1049. case 1243:
  1050. s->avctx->pix_fmt = AV_PIX_FMT_GBRP;
  1051. break;
  1052. case 1324:
  1053. s->avctx->pix_fmt = AV_PIX_FMT_GBRAP;
  1054. break;
  1055. case 1483:
  1056. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRP16LE : AV_PIX_FMT_GBRP16BE;
  1057. break;
  1058. case 1644:
  1059. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRAP16LE : AV_PIX_FMT_GBRAP16BE;
  1060. break;
  1061. default:
  1062. av_log(s->avctx, AV_LOG_ERROR,
  1063. "This format is not supported (bpp=%d, bppcount=%d)\n",
  1064. s->bpp, s->bppcount);
  1065. return AVERROR_INVALIDDATA;
  1066. }
  1067. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1068. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->avctx->pix_fmt);
  1069. if((desc->flags & AV_PIX_FMT_FLAG_RGB) ||
  1070. !(desc->flags & AV_PIX_FMT_FLAG_PLANAR) ||
  1071. desc->nb_components < 3) {
  1072. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr variant\n");
  1073. return AVERROR_INVALIDDATA;
  1074. }
  1075. }
  1076. if (s->width != s->avctx->width || s->height != s->avctx->height) {
  1077. ret = ff_set_dimensions(s->avctx, s->width, s->height);
  1078. if (ret < 0)
  1079. return ret;
  1080. }
  1081. if ((ret = ff_thread_get_buffer(s->avctx, frame, 0)) < 0)
  1082. return ret;
  1083. if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  1084. if (!create_gray_palette)
  1085. memcpy(frame->f->data[1], s->palette, sizeof(s->palette));
  1086. else {
  1087. /* make default grayscale pal */
  1088. int i;
  1089. uint32_t *pal = (uint32_t *)frame->f->data[1];
  1090. for (i = 0; i < 1<<s->bpp; i++)
  1091. pal[i] = 0xFFU << 24 | i * 255 / ((1<<s->bpp) - 1) * 0x010101;
  1092. }
  1093. }
  1094. return 0;
  1095. }
  1096. static void set_sar(TiffContext *s, unsigned tag, unsigned num, unsigned den)
  1097. {
  1098. int offset = tag == TIFF_YRES ? 2 : 0;
  1099. s->res[offset++] = num;
  1100. s->res[offset] = den;
  1101. if (s->res[0] && s->res[1] && s->res[2] && s->res[3]) {
  1102. uint64_t num = s->res[2] * (uint64_t)s->res[1];
  1103. uint64_t den = s->res[0] * (uint64_t)s->res[3];
  1104. if (num > INT64_MAX || den > INT64_MAX) {
  1105. num = num >> 1;
  1106. den = den >> 1;
  1107. }
  1108. av_reduce(&s->avctx->sample_aspect_ratio.num, &s->avctx->sample_aspect_ratio.den,
  1109. num, den, INT32_MAX);
  1110. if (!s->avctx->sample_aspect_ratio.den)
  1111. s->avctx->sample_aspect_ratio = (AVRational) {0, 1};
  1112. }
  1113. }
  1114. static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
  1115. {
  1116. AVFrameSideData *sd;
  1117. GetByteContext gb_temp;
  1118. unsigned tag, type, count, off, value = 0, value2 = 1; // value2 is a denominator so init. to 1
  1119. int i, start;
  1120. int pos;
  1121. int ret;
  1122. double *dp;
  1123. ret = ff_tread_tag(&s->gb, s->le, &tag, &type, &count, &start);
  1124. if (ret < 0) {
  1125. goto end;
  1126. }
  1127. if (tag <= s->last_tag)
  1128. return AVERROR_INVALIDDATA;
  1129. // We ignore TIFF_STRIP_SIZE as it is sometimes in the logic but wrong order around TIFF_STRIP_OFFS
  1130. if (tag != TIFF_STRIP_SIZE)
  1131. s->last_tag = tag;
  1132. off = bytestream2_tell(&s->gb);
  1133. if (count == 1) {
  1134. switch (type) {
  1135. case TIFF_BYTE:
  1136. case TIFF_SHORT:
  1137. case TIFF_LONG:
  1138. value = ff_tget(&s->gb, type, s->le);
  1139. break;
  1140. case TIFF_RATIONAL:
  1141. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  1142. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  1143. if (!value2) {
  1144. av_log(s->avctx, AV_LOG_ERROR, "Invalid denominator in rational\n");
  1145. return AVERROR_INVALIDDATA;
  1146. }
  1147. break;
  1148. case TIFF_STRING:
  1149. if (count <= 4) {
  1150. break;
  1151. }
  1152. default:
  1153. value = UINT_MAX;
  1154. }
  1155. }
  1156. switch (tag) {
  1157. case TIFF_SUBFILE:
  1158. s->is_thumbnail = (value != 0);
  1159. break;
  1160. case TIFF_WIDTH:
  1161. s->width = value;
  1162. break;
  1163. case TIFF_HEIGHT:
  1164. s->height = value;
  1165. break;
  1166. case TIFF_BPP:
  1167. if (count > 5 || count <= 0) {
  1168. av_log(s->avctx, AV_LOG_ERROR,
  1169. "This format is not supported (bpp=%d, %d components)\n",
  1170. value, count);
  1171. return AVERROR_INVALIDDATA;
  1172. }
  1173. s->bppcount = count;
  1174. if (count == 1)
  1175. s->bpp = value;
  1176. else {
  1177. switch (type) {
  1178. case TIFF_BYTE:
  1179. case TIFF_SHORT:
  1180. case TIFF_LONG:
  1181. s->bpp = 0;
  1182. if (bytestream2_get_bytes_left(&s->gb) < type_sizes[type] * count)
  1183. return AVERROR_INVALIDDATA;
  1184. for (i = 0; i < count; i++)
  1185. s->bpp += ff_tget(&s->gb, type, s->le);
  1186. break;
  1187. default:
  1188. s->bpp = -1;
  1189. }
  1190. }
  1191. break;
  1192. case TIFF_SAMPLES_PER_PIXEL:
  1193. if (count != 1) {
  1194. av_log(s->avctx, AV_LOG_ERROR,
  1195. "Samples per pixel requires a single value, many provided\n");
  1196. return AVERROR_INVALIDDATA;
  1197. }
  1198. if (value > 5 || value <= 0) {
  1199. av_log(s->avctx, AV_LOG_ERROR,
  1200. "Invalid samples per pixel %d\n", value);
  1201. return AVERROR_INVALIDDATA;
  1202. }
  1203. if (s->bppcount == 1)
  1204. s->bpp *= value;
  1205. s->bppcount = value;
  1206. break;
  1207. case TIFF_COMPR:
  1208. s->compr = value;
  1209. av_log(s->avctx, AV_LOG_DEBUG, "compression: %d\n", s->compr);
  1210. s->predictor = 0;
  1211. switch (s->compr) {
  1212. case TIFF_RAW:
  1213. case TIFF_PACKBITS:
  1214. case TIFF_LZW:
  1215. case TIFF_CCITT_RLE:
  1216. break;
  1217. case TIFF_G3:
  1218. case TIFF_G4:
  1219. s->fax_opts = 0;
  1220. break;
  1221. case TIFF_DEFLATE:
  1222. case TIFF_ADOBE_DEFLATE:
  1223. #if CONFIG_ZLIB
  1224. break;
  1225. #else
  1226. av_log(s->avctx, AV_LOG_ERROR, "Deflate: ZLib not compiled in\n");
  1227. return AVERROR(ENOSYS);
  1228. #endif
  1229. case TIFF_JPEG:
  1230. case TIFF_NEWJPEG:
  1231. s->is_jpeg = 1;
  1232. break;
  1233. case TIFF_LZMA:
  1234. #if CONFIG_LZMA
  1235. break;
  1236. #else
  1237. av_log(s->avctx, AV_LOG_ERROR, "LZMA not compiled in\n");
  1238. return AVERROR(ENOSYS);
  1239. #endif
  1240. default:
  1241. av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
  1242. s->compr);
  1243. return AVERROR_INVALIDDATA;
  1244. }
  1245. break;
  1246. case TIFF_ROWSPERSTRIP:
  1247. if (!value || (type == TIFF_LONG && value == UINT_MAX))
  1248. value = s->height;
  1249. s->rps = FFMIN(value, s->height);
  1250. break;
  1251. case TIFF_STRIP_OFFS:
  1252. if (count == 1) {
  1253. if (value > INT_MAX) {
  1254. av_log(s->avctx, AV_LOG_ERROR,
  1255. "strippos %u too large\n", value);
  1256. return AVERROR_INVALIDDATA;
  1257. }
  1258. s->strippos = 0;
  1259. s->stripoff = value;
  1260. } else
  1261. s->strippos = off;
  1262. s->strips = count;
  1263. if (s->strips == 1)
  1264. s->rps = s->height;
  1265. s->sot = type;
  1266. break;
  1267. case TIFF_STRIP_SIZE:
  1268. if (count == 1) {
  1269. if (value > INT_MAX) {
  1270. av_log(s->avctx, AV_LOG_ERROR,
  1271. "stripsize %u too large\n", value);
  1272. return AVERROR_INVALIDDATA;
  1273. }
  1274. s->stripsizesoff = 0;
  1275. s->stripsize = value;
  1276. s->strips = 1;
  1277. } else {
  1278. s->stripsizesoff = off;
  1279. }
  1280. s->strips = count;
  1281. s->sstype = type;
  1282. break;
  1283. case TIFF_XRES:
  1284. case TIFF_YRES:
  1285. set_sar(s, tag, value, value2);
  1286. break;
  1287. case TIFF_TILE_OFFSETS:
  1288. s->tile_offsets_offset = off;
  1289. s->tile_count = count;
  1290. s->is_tiled = 1;
  1291. break;
  1292. case TIFF_TILE_BYTE_COUNTS:
  1293. s->tile_byte_counts_offset = off;
  1294. break;
  1295. case TIFF_TILE_LENGTH:
  1296. s->tile_length = value;
  1297. break;
  1298. case TIFF_TILE_WIDTH:
  1299. s->tile_width = value;
  1300. break;
  1301. case TIFF_PREDICTOR:
  1302. s->predictor = value;
  1303. break;
  1304. case TIFF_SUB_IFDS:
  1305. if (count == 1)
  1306. s->sub_ifd = value;
  1307. else if (count > 1)
  1308. s->sub_ifd = ff_tget(&s->gb, TIFF_LONG, s->le); /** Only get the first SubIFD */
  1309. break;
  1310. case DNG_LINEARIZATION_TABLE:
  1311. if (count > FF_ARRAY_ELEMS(s->dng_lut))
  1312. return AVERROR_INVALIDDATA;
  1313. for (int i = 0; i < count; i++)
  1314. s->dng_lut[i] = ff_tget(&s->gb, type, s->le);
  1315. break;
  1316. case DNG_BLACK_LEVEL:
  1317. if (count > 1) { /* Use the first value in the pattern (assume they're all the same) */
  1318. if (type == TIFF_RATIONAL) {
  1319. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  1320. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  1321. if (!value2) {
  1322. av_log(s->avctx, AV_LOG_ERROR, "Invalid black level denominator\n");
  1323. return AVERROR_INVALIDDATA;
  1324. }
  1325. s->black_level = value / value2;
  1326. } else
  1327. s->black_level = ff_tget(&s->gb, type, s->le);
  1328. av_log(s->avctx, AV_LOG_WARNING, "Assuming black level pattern values are identical\n");
  1329. } else {
  1330. s->black_level = value / value2;
  1331. }
  1332. break;
  1333. case DNG_WHITE_LEVEL:
  1334. s->white_level = value;
  1335. break;
  1336. case TIFF_CFA_PATTERN_DIM:
  1337. if (count != 2 || (ff_tget(&s->gb, type, s->le) != 2 &&
  1338. ff_tget(&s->gb, type, s->le) != 2)) {
  1339. av_log(s->avctx, AV_LOG_ERROR, "CFA Pattern dimensions are not 2x2\n");
  1340. return AVERROR_INVALIDDATA;
  1341. }
  1342. break;
  1343. case TIFF_CFA_PATTERN:
  1344. s->is_bayer = 1;
  1345. s->pattern[0] = ff_tget(&s->gb, type, s->le);
  1346. s->pattern[1] = ff_tget(&s->gb, type, s->le);
  1347. s->pattern[2] = ff_tget(&s->gb, type, s->le);
  1348. s->pattern[3] = ff_tget(&s->gb, type, s->le);
  1349. break;
  1350. case TIFF_PHOTOMETRIC:
  1351. switch (value) {
  1352. case TIFF_PHOTOMETRIC_WHITE_IS_ZERO:
  1353. case TIFF_PHOTOMETRIC_BLACK_IS_ZERO:
  1354. case TIFF_PHOTOMETRIC_RGB:
  1355. case TIFF_PHOTOMETRIC_PALETTE:
  1356. case TIFF_PHOTOMETRIC_SEPARATED:
  1357. case TIFF_PHOTOMETRIC_YCBCR:
  1358. case TIFF_PHOTOMETRIC_CFA:
  1359. case TIFF_PHOTOMETRIC_LINEAR_RAW: // Used by DNG images
  1360. s->photometric = value;
  1361. break;
  1362. case TIFF_PHOTOMETRIC_ALPHA_MASK:
  1363. case TIFF_PHOTOMETRIC_CIE_LAB:
  1364. case TIFF_PHOTOMETRIC_ICC_LAB:
  1365. case TIFF_PHOTOMETRIC_ITU_LAB:
  1366. case TIFF_PHOTOMETRIC_LOG_L:
  1367. case TIFF_PHOTOMETRIC_LOG_LUV:
  1368. avpriv_report_missing_feature(s->avctx,
  1369. "PhotometricInterpretation 0x%04X",
  1370. value);
  1371. return AVERROR_PATCHWELCOME;
  1372. default:
  1373. av_log(s->avctx, AV_LOG_ERROR, "PhotometricInterpretation %u is "
  1374. "unknown\n", value);
  1375. return AVERROR_INVALIDDATA;
  1376. }
  1377. break;
  1378. case TIFF_FILL_ORDER:
  1379. if (value < 1 || value > 2) {
  1380. av_log(s->avctx, AV_LOG_ERROR,
  1381. "Unknown FillOrder value %d, trying default one\n", value);
  1382. value = 1;
  1383. }
  1384. s->fill_order = value - 1;
  1385. break;
  1386. case TIFF_PAL: {
  1387. GetByteContext pal_gb[3];
  1388. off = type_sizes[type];
  1389. if (count / 3 > 256 ||
  1390. bytestream2_get_bytes_left(&s->gb) < count / 3 * off * 3)
  1391. return AVERROR_INVALIDDATA;
  1392. pal_gb[0] = pal_gb[1] = pal_gb[2] = s->gb;
  1393. bytestream2_skip(&pal_gb[1], count / 3 * off);
  1394. bytestream2_skip(&pal_gb[2], count / 3 * off * 2);
  1395. off = (type_sizes[type] - 1) << 3;
  1396. if (off > 31U) {
  1397. av_log(s->avctx, AV_LOG_ERROR, "palette shift %d is out of range\n", off);
  1398. return AVERROR_INVALIDDATA;
  1399. }
  1400. for (i = 0; i < count / 3; i++) {
  1401. uint32_t p = 0xFF000000;
  1402. p |= (ff_tget(&pal_gb[0], type, s->le) >> off) << 16;
  1403. p |= (ff_tget(&pal_gb[1], type, s->le) >> off) << 8;
  1404. p |= ff_tget(&pal_gb[2], type, s->le) >> off;
  1405. s->palette[i] = p;
  1406. }
  1407. s->palette_is_set = 1;
  1408. break;
  1409. }
  1410. case TIFF_PLANAR:
  1411. s->planar = value == 2;
  1412. break;
  1413. case TIFF_YCBCR_SUBSAMPLING:
  1414. if (count != 2) {
  1415. av_log(s->avctx, AV_LOG_ERROR, "subsample count invalid\n");
  1416. return AVERROR_INVALIDDATA;
  1417. }
  1418. for (i = 0; i < count; i++) {
  1419. s->subsampling[i] = ff_tget(&s->gb, type, s->le);
  1420. if (s->subsampling[i] <= 0) {
  1421. av_log(s->avctx, AV_LOG_ERROR, "subsampling %d is invalid\n", s->subsampling[i]);
  1422. s->subsampling[i] = 1;
  1423. return AVERROR_INVALIDDATA;
  1424. }
  1425. }
  1426. break;
  1427. case TIFF_T4OPTIONS:
  1428. if (s->compr == TIFF_G3)
  1429. s->fax_opts = value;
  1430. break;
  1431. case TIFF_T6OPTIONS:
  1432. if (s->compr == TIFF_G4)
  1433. s->fax_opts = value;
  1434. break;
  1435. #define ADD_METADATA(count, name, sep)\
  1436. if ((ret = add_metadata(count, type, name, sep, s, frame)) < 0) {\
  1437. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");\
  1438. goto end;\
  1439. }
  1440. case TIFF_MODEL_PIXEL_SCALE:
  1441. ADD_METADATA(count, "ModelPixelScaleTag", NULL);
  1442. break;
  1443. case TIFF_MODEL_TRANSFORMATION:
  1444. ADD_METADATA(count, "ModelTransformationTag", NULL);
  1445. break;
  1446. case TIFF_MODEL_TIEPOINT:
  1447. ADD_METADATA(count, "ModelTiepointTag", NULL);
  1448. break;
  1449. case TIFF_GEO_KEY_DIRECTORY:
  1450. if (s->geotag_count) {
  1451. avpriv_request_sample(s->avctx, "Multiple geo key directories");
  1452. return AVERROR_INVALIDDATA;
  1453. }
  1454. ADD_METADATA(1, "GeoTIFF_Version", NULL);
  1455. ADD_METADATA(2, "GeoTIFF_Key_Revision", ".");
  1456. s->geotag_count = ff_tget_short(&s->gb, s->le);
  1457. if (s->geotag_count > count / 4 - 1) {
  1458. s->geotag_count = count / 4 - 1;
  1459. av_log(s->avctx, AV_LOG_WARNING, "GeoTIFF key directory buffer shorter than specified\n");
  1460. }
  1461. if ( bytestream2_get_bytes_left(&s->gb) < s->geotag_count * sizeof(int16_t) * 4
  1462. || s->geotag_count == 0) {
  1463. s->geotag_count = 0;
  1464. return -1;
  1465. }
  1466. s->geotags = av_mallocz_array(s->geotag_count, sizeof(TiffGeoTag));
  1467. if (!s->geotags) {
  1468. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1469. s->geotag_count = 0;
  1470. goto end;
  1471. }
  1472. for (i = 0; i < s->geotag_count; i++) {
  1473. s->geotags[i].key = ff_tget_short(&s->gb, s->le);
  1474. s->geotags[i].type = ff_tget_short(&s->gb, s->le);
  1475. s->geotags[i].count = ff_tget_short(&s->gb, s->le);
  1476. if (!s->geotags[i].type)
  1477. s->geotags[i].val = get_geokey_val(s->geotags[i].key, ff_tget_short(&s->gb, s->le));
  1478. else
  1479. s->geotags[i].offset = ff_tget_short(&s->gb, s->le);
  1480. }
  1481. break;
  1482. case TIFF_GEO_DOUBLE_PARAMS:
  1483. if (count >= INT_MAX / sizeof(int64_t))
  1484. return AVERROR_INVALIDDATA;
  1485. if (bytestream2_get_bytes_left(&s->gb) < count * sizeof(int64_t))
  1486. return AVERROR_INVALIDDATA;
  1487. dp = av_malloc_array(count, sizeof(double));
  1488. if (!dp) {
  1489. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1490. goto end;
  1491. }
  1492. for (i = 0; i < count; i++)
  1493. dp[i] = ff_tget_double(&s->gb, s->le);
  1494. for (i = 0; i < s->geotag_count; i++) {
  1495. if (s->geotags[i].type == TIFF_GEO_DOUBLE_PARAMS) {
  1496. if (s->geotags[i].count == 0
  1497. || s->geotags[i].offset + s->geotags[i].count > count) {
  1498. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1499. } else if (s->geotags[i].val) {
  1500. av_log(s->avctx, AV_LOG_WARNING, "Duplicate GeoTIFF key %d\n", s->geotags[i].key);
  1501. } else {
  1502. char *ap = doubles2str(&dp[s->geotags[i].offset], s->geotags[i].count, ", ");
  1503. if (!ap) {
  1504. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1505. av_freep(&dp);
  1506. return AVERROR(ENOMEM);
  1507. }
  1508. s->geotags[i].val = ap;
  1509. }
  1510. }
  1511. }
  1512. av_freep(&dp);
  1513. break;
  1514. case TIFF_GEO_ASCII_PARAMS:
  1515. pos = bytestream2_tell(&s->gb);
  1516. for (i = 0; i < s->geotag_count; i++) {
  1517. if (s->geotags[i].type == TIFF_GEO_ASCII_PARAMS) {
  1518. if (s->geotags[i].count == 0
  1519. || s->geotags[i].offset + s->geotags[i].count > count) {
  1520. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1521. } else {
  1522. char *ap;
  1523. bytestream2_seek(&s->gb, pos + s->geotags[i].offset, SEEK_SET);
  1524. if (bytestream2_get_bytes_left(&s->gb) < s->geotags[i].count)
  1525. return AVERROR_INVALIDDATA;
  1526. if (s->geotags[i].val)
  1527. return AVERROR_INVALIDDATA;
  1528. ap = av_malloc(s->geotags[i].count);
  1529. if (!ap) {
  1530. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1531. return AVERROR(ENOMEM);
  1532. }
  1533. bytestream2_get_bufferu(&s->gb, ap, s->geotags[i].count);
  1534. ap[s->geotags[i].count - 1] = '\0'; //replace the "|" delimiter with a 0 byte
  1535. s->geotags[i].val = ap;
  1536. }
  1537. }
  1538. }
  1539. break;
  1540. case TIFF_ICC_PROFILE:
  1541. gb_temp = s->gb;
  1542. bytestream2_seek(&gb_temp, SEEK_SET, off);
  1543. if (bytestream2_get_bytes_left(&gb_temp) < count)
  1544. return AVERROR_INVALIDDATA;
  1545. sd = av_frame_new_side_data(frame, AV_FRAME_DATA_ICC_PROFILE, count);
  1546. if (!sd)
  1547. return AVERROR(ENOMEM);
  1548. bytestream2_get_bufferu(&gb_temp, sd->data, count);
  1549. break;
  1550. case TIFF_ARTIST:
  1551. ADD_METADATA(count, "artist", NULL);
  1552. break;
  1553. case TIFF_COPYRIGHT:
  1554. ADD_METADATA(count, "copyright", NULL);
  1555. break;
  1556. case TIFF_DATE:
  1557. ADD_METADATA(count, "date", NULL);
  1558. break;
  1559. case TIFF_DOCUMENT_NAME:
  1560. ADD_METADATA(count, "document_name", NULL);
  1561. break;
  1562. case TIFF_HOST_COMPUTER:
  1563. ADD_METADATA(count, "computer", NULL);
  1564. break;
  1565. case TIFF_IMAGE_DESCRIPTION:
  1566. ADD_METADATA(count, "description", NULL);
  1567. break;
  1568. case TIFF_MAKE:
  1569. ADD_METADATA(count, "make", NULL);
  1570. break;
  1571. case TIFF_MODEL:
  1572. ADD_METADATA(count, "model", NULL);
  1573. break;
  1574. case TIFF_PAGE_NAME:
  1575. ADD_METADATA(count, "page_name", NULL);
  1576. break;
  1577. case TIFF_PAGE_NUMBER:
  1578. ADD_METADATA(count, "page_number", " / ");
  1579. // need to seek back to re-read the page number
  1580. bytestream2_seek(&s->gb, -count * sizeof(uint16_t), SEEK_CUR);
  1581. // read the page number
  1582. s->cur_page = ff_tget(&s->gb, TIFF_SHORT, s->le);
  1583. // get back to where we were before the previous seek
  1584. bytestream2_seek(&s->gb, count * sizeof(uint16_t) - sizeof(uint16_t), SEEK_CUR);
  1585. break;
  1586. case TIFF_SOFTWARE_NAME:
  1587. ADD_METADATA(count, "software", NULL);
  1588. break;
  1589. case DNG_VERSION:
  1590. if (count == 4) {
  1591. unsigned int ver[4];
  1592. ver[0] = ff_tget(&s->gb, type, s->le);
  1593. ver[1] = ff_tget(&s->gb, type, s->le);
  1594. ver[2] = ff_tget(&s->gb, type, s->le);
  1595. ver[3] = ff_tget(&s->gb, type, s->le);
  1596. av_log(s->avctx, AV_LOG_DEBUG, "DNG file, version %u.%u.%u.%u\n",
  1597. ver[0], ver[1], ver[2], ver[3]);
  1598. tiff_set_type(s, TIFF_TYPE_DNG);
  1599. }
  1600. break;
  1601. case CINEMADNG_TIME_CODES:
  1602. case CINEMADNG_FRAME_RATE:
  1603. case CINEMADNG_T_STOP:
  1604. case CINEMADNG_REEL_NAME:
  1605. case CINEMADNG_CAMERA_LABEL:
  1606. tiff_set_type(s, TIFF_TYPE_CINEMADNG);
  1607. break;
  1608. default:
  1609. if (s->avctx->err_recognition & AV_EF_EXPLODE) {
  1610. av_log(s->avctx, AV_LOG_ERROR,
  1611. "Unknown or unsupported tag %d/0x%0X\n",
  1612. tag, tag);
  1613. return AVERROR_INVALIDDATA;
  1614. }
  1615. }
  1616. end:
  1617. if (s->bpp > 64U) {
  1618. av_log(s->avctx, AV_LOG_ERROR,
  1619. "This format is not supported (bpp=%d, %d components)\n",
  1620. s->bpp, count);
  1621. s->bpp = 0;
  1622. return AVERROR_INVALIDDATA;
  1623. }
  1624. bytestream2_seek(&s->gb, start, SEEK_SET);
  1625. return 0;
  1626. }
  1627. static int decode_frame(AVCodecContext *avctx,
  1628. void *data, int *got_frame, AVPacket *avpkt)
  1629. {
  1630. TiffContext *const s = avctx->priv_data;
  1631. AVFrame *const p = data;
  1632. ThreadFrame frame = { .f = data };
  1633. unsigned off, last_off;
  1634. int le, ret, plane, planes;
  1635. int i, j, entries, stride;
  1636. unsigned soff, ssize;
  1637. uint8_t *dst;
  1638. GetByteContext stripsizes;
  1639. GetByteContext stripdata;
  1640. int retry_for_subifd, retry_for_page;
  1641. int is_dng;
  1642. int has_tile_bits, has_strip_bits;
  1643. bytestream2_init(&s->gb, avpkt->data, avpkt->size);
  1644. // parse image header
  1645. if ((ret = ff_tdecode_header(&s->gb, &le, &off))) {
  1646. av_log(avctx, AV_LOG_ERROR, "Invalid TIFF header\n");
  1647. return ret;
  1648. } else if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1649. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1650. return AVERROR_INVALIDDATA;
  1651. }
  1652. s->le = le;
  1653. // TIFF_BPP is not a required tag and defaults to 1
  1654. s->tiff_type = TIFF_TYPE_TIFF;
  1655. again:
  1656. s->is_thumbnail = 0;
  1657. s->bppcount = s->bpp = 1;
  1658. s->photometric = TIFF_PHOTOMETRIC_NONE;
  1659. s->compr = TIFF_RAW;
  1660. s->fill_order = 0;
  1661. s->white_level = 0;
  1662. s->is_bayer = 0;
  1663. s->is_tiled = 0;
  1664. s->is_jpeg = 0;
  1665. s->cur_page = 0;
  1666. s->last_tag = 0;
  1667. for (i = 0; i < 65536; i++)
  1668. s->dng_lut[i] = i;
  1669. free_geotags(s);
  1670. // Reset these offsets so we can tell if they were set this frame
  1671. s->stripsizesoff = s->strippos = 0;
  1672. /* parse image file directory */
  1673. bytestream2_seek(&s->gb, off, SEEK_SET);
  1674. entries = ff_tget_short(&s->gb, le);
  1675. if (bytestream2_get_bytes_left(&s->gb) < entries * 12)
  1676. return AVERROR_INVALIDDATA;
  1677. for (i = 0; i < entries; i++) {
  1678. if ((ret = tiff_decode_tag(s, p)) < 0)
  1679. return ret;
  1680. }
  1681. if (s->get_thumbnail && !s->is_thumbnail) {
  1682. av_log(avctx, AV_LOG_INFO, "No embedded thumbnail present\n");
  1683. return AVERROR_EOF;
  1684. }
  1685. /** whether we should process this IFD's SubIFD */
  1686. retry_for_subifd = s->sub_ifd && (s->get_subimage || (!s->get_thumbnail && s->is_thumbnail));
  1687. /** whether we should process this multi-page IFD's next page */
  1688. retry_for_page = s->get_page && s->cur_page + 1 < s->get_page; // get_page is 1-indexed
  1689. last_off = off;
  1690. if (retry_for_page) {
  1691. // set offset to the next IFD
  1692. off = ff_tget_long(&s->gb, le);
  1693. } else if (retry_for_subifd) {
  1694. // set offset to the SubIFD
  1695. off = s->sub_ifd;
  1696. }
  1697. if (retry_for_subifd || retry_for_page) {
  1698. if (!off) {
  1699. av_log(avctx, AV_LOG_ERROR, "Requested entry not found\n");
  1700. return AVERROR_INVALIDDATA;
  1701. }
  1702. if (off <= last_off) {
  1703. avpriv_request_sample(s->avctx, "non increasing IFD offset");
  1704. return AVERROR_INVALIDDATA;
  1705. }
  1706. if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1707. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1708. return AVERROR_INVALIDDATA;
  1709. }
  1710. s->sub_ifd = 0;
  1711. goto again;
  1712. }
  1713. /* At this point we've decided on which (Sub)IFD to process */
  1714. is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
  1715. for (i = 0; i<s->geotag_count; i++) {
  1716. const char *keyname = get_geokey_name(s->geotags[i].key);
  1717. if (!keyname) {
  1718. av_log(avctx, AV_LOG_WARNING, "Unknown or unsupported GeoTIFF key %d\n", s->geotags[i].key);
  1719. continue;
  1720. }
  1721. if (get_geokey_type(s->geotags[i].key) != s->geotags[i].type) {
  1722. av_log(avctx, AV_LOG_WARNING, "Type of GeoTIFF key %d is wrong\n", s->geotags[i].key);
  1723. continue;
  1724. }
  1725. ret = av_dict_set(&p->metadata, keyname, s->geotags[i].val, 0);
  1726. if (ret<0) {
  1727. av_log(avctx, AV_LOG_ERROR, "Writing metadata with key '%s' failed\n", keyname);
  1728. return ret;
  1729. }
  1730. }
  1731. if (is_dng) {
  1732. int bps;
  1733. if (s->bpp % s->bppcount)
  1734. return AVERROR_INVALIDDATA;
  1735. bps = s->bpp / s->bppcount;
  1736. if (bps < 8 || bps > 32)
  1737. return AVERROR_INVALIDDATA;
  1738. if (s->white_level == 0)
  1739. s->white_level = (1LL << bps) - 1; /* Default value as per the spec */
  1740. if (s->white_level <= s->black_level) {
  1741. av_log(avctx, AV_LOG_ERROR, "BlackLevel (%"PRId32") must be less than WhiteLevel (%"PRId32")\n",
  1742. s->black_level, s->white_level);
  1743. return AVERROR_INVALIDDATA;
  1744. }
  1745. if (s->planar)
  1746. return AVERROR_PATCHWELCOME;
  1747. }
  1748. if (!s->is_tiled && !s->strippos && !s->stripoff) {
  1749. av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
  1750. return AVERROR_INVALIDDATA;
  1751. }
  1752. has_tile_bits = s->is_tiled || s->tile_byte_counts_offset || s->tile_offsets_offset || s->tile_width || s->tile_length || s->tile_count;
  1753. has_strip_bits = s->strippos || s->strips || s->stripoff || s->rps || s->sot || s->sstype || s->stripsize || s->stripsizesoff;
  1754. if (has_tile_bits && has_strip_bits) {
  1755. int tiled_dng = s->is_tiled && is_dng;
  1756. av_log(avctx, tiled_dng ? AV_LOG_WARNING : AV_LOG_ERROR, "Tiled TIFF is not allowed to strip\n");
  1757. if (!tiled_dng)
  1758. return AVERROR_INVALIDDATA;
  1759. }
  1760. /* now we have the data and may start decoding */
  1761. if ((ret = init_image(s, &frame)) < 0)
  1762. return ret;
  1763. if (!s->is_tiled || has_strip_bits) {
  1764. if (s->strips == 1 && !s->stripsize) {
  1765. av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
  1766. s->stripsize = avpkt->size - s->stripoff;
  1767. }
  1768. if (s->stripsizesoff) {
  1769. if (s->stripsizesoff >= (unsigned)avpkt->size)
  1770. return AVERROR_INVALIDDATA;
  1771. bytestream2_init(&stripsizes, avpkt->data + s->stripsizesoff,
  1772. avpkt->size - s->stripsizesoff);
  1773. }
  1774. if (s->strippos) {
  1775. if (s->strippos >= (unsigned)avpkt->size)
  1776. return AVERROR_INVALIDDATA;
  1777. bytestream2_init(&stripdata, avpkt->data + s->strippos,
  1778. avpkt->size - s->strippos);
  1779. }
  1780. if (s->rps <= 0 || s->rps % s->subsampling[1]) {
  1781. av_log(avctx, AV_LOG_ERROR, "rps %d invalid\n", s->rps);
  1782. return AVERROR_INVALIDDATA;
  1783. }
  1784. }
  1785. if (s->photometric == TIFF_PHOTOMETRIC_LINEAR_RAW ||
  1786. s->photometric == TIFF_PHOTOMETRIC_CFA) {
  1787. p->color_trc = AVCOL_TRC_LINEAR;
  1788. } else if (s->photometric == TIFF_PHOTOMETRIC_BLACK_IS_ZERO) {
  1789. p->color_trc = AVCOL_TRC_GAMMA22;
  1790. }
  1791. /* Handle DNG images with JPEG-compressed tiles */
  1792. if (is_dng && s->is_tiled) {
  1793. if (!s->is_jpeg) {
  1794. avpriv_report_missing_feature(avctx, "DNG uncompressed tiled images");
  1795. return AVERROR_PATCHWELCOME;
  1796. } else if (!s->is_bayer) {
  1797. avpriv_report_missing_feature(avctx, "DNG JPG-compressed tiled non-bayer-encoded images");
  1798. return AVERROR_PATCHWELCOME;
  1799. } else {
  1800. if ((ret = dng_decode_tiles(avctx, (AVFrame*)data, avpkt)) > 0)
  1801. *got_frame = 1;
  1802. return ret;
  1803. }
  1804. }
  1805. /* Handle TIFF images and DNG images with uncompressed strips (non-tiled) */
  1806. planes = s->planar ? s->bppcount : 1;
  1807. for (plane = 0; plane < planes; plane++) {
  1808. uint8_t *five_planes = NULL;
  1809. int remaining = avpkt->size;
  1810. int decoded_height;
  1811. stride = p->linesize[plane];
  1812. dst = p->data[plane];
  1813. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1814. s->avctx->pix_fmt == AV_PIX_FMT_RGBA) {
  1815. stride = stride * 5 / 4;
  1816. five_planes =
  1817. dst = av_malloc(stride * s->height);
  1818. if (!dst)
  1819. return AVERROR(ENOMEM);
  1820. }
  1821. for (i = 0; i < s->height; i += s->rps) {
  1822. if (i)
  1823. dst += s->rps * stride;
  1824. if (s->stripsizesoff)
  1825. ssize = ff_tget(&stripsizes, s->sstype, le);
  1826. else
  1827. ssize = s->stripsize;
  1828. if (s->strippos)
  1829. soff = ff_tget(&stripdata, s->sot, le);
  1830. else
  1831. soff = s->stripoff;
  1832. if (soff > avpkt->size || ssize > avpkt->size - soff || ssize > remaining) {
  1833. av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
  1834. av_freep(&five_planes);
  1835. return AVERROR_INVALIDDATA;
  1836. }
  1837. remaining -= ssize;
  1838. if ((ret = tiff_unpack_strip(s, p, dst, stride, avpkt->data + soff, ssize, i,
  1839. FFMIN(s->rps, s->height - i))) < 0) {
  1840. if (avctx->err_recognition & AV_EF_EXPLODE) {
  1841. av_freep(&five_planes);
  1842. return ret;
  1843. }
  1844. break;
  1845. }
  1846. }
  1847. decoded_height = FFMIN(i, s->height);
  1848. if (s->predictor == 2) {
  1849. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1850. av_log(s->avctx, AV_LOG_ERROR, "predictor == 2 with YUV is unsupported");
  1851. return AVERROR_PATCHWELCOME;
  1852. }
  1853. dst = five_planes ? five_planes : p->data[plane];
  1854. soff = s->bpp >> 3;
  1855. if (s->planar)
  1856. soff = FFMAX(soff / s->bppcount, 1);
  1857. ssize = s->width * soff;
  1858. if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48LE ||
  1859. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64LE ||
  1860. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16LE ||
  1861. s->avctx->pix_fmt == AV_PIX_FMT_YA16LE ||
  1862. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16LE ||
  1863. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16LE) {
  1864. for (i = 0; i < decoded_height; i++) {
  1865. for (j = soff; j < ssize; j += 2)
  1866. AV_WL16(dst + j, AV_RL16(dst + j) + AV_RL16(dst + j - soff));
  1867. dst += stride;
  1868. }
  1869. } else if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48BE ||
  1870. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE ||
  1871. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16BE ||
  1872. s->avctx->pix_fmt == AV_PIX_FMT_YA16BE ||
  1873. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16BE ||
  1874. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16BE) {
  1875. for (i = 0; i < decoded_height; i++) {
  1876. for (j = soff; j < ssize; j += 2)
  1877. AV_WB16(dst + j, AV_RB16(dst + j) + AV_RB16(dst + j - soff));
  1878. dst += stride;
  1879. }
  1880. } else {
  1881. for (i = 0; i < decoded_height; i++) {
  1882. for (j = soff; j < ssize; j++)
  1883. dst[j] += dst[j - soff];
  1884. dst += stride;
  1885. }
  1886. }
  1887. }
  1888. if (s->photometric == TIFF_PHOTOMETRIC_WHITE_IS_ZERO) {
  1889. int c = (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255);
  1890. dst = p->data[plane];
  1891. for (i = 0; i < s->height; i++) {
  1892. for (j = 0; j < stride; j++)
  1893. dst[j] = c - dst[j];
  1894. dst += stride;
  1895. }
  1896. }
  1897. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1898. (s->avctx->pix_fmt == AV_PIX_FMT_RGB0 || s->avctx->pix_fmt == AV_PIX_FMT_RGBA)) {
  1899. int x = s->avctx->pix_fmt == AV_PIX_FMT_RGB0 ? 4 : 5;
  1900. uint8_t *src = five_planes ? five_planes : p->data[plane];
  1901. dst = p->data[plane];
  1902. for (i = 0; i < s->height; i++) {
  1903. for (j = 0; j < s->width; j++) {
  1904. int k = 255 - src[x * j + 3];
  1905. int r = (255 - src[x * j ]) * k;
  1906. int g = (255 - src[x * j + 1]) * k;
  1907. int b = (255 - src[x * j + 2]) * k;
  1908. dst[4 * j ] = r * 257 >> 16;
  1909. dst[4 * j + 1] = g * 257 >> 16;
  1910. dst[4 * j + 2] = b * 257 >> 16;
  1911. dst[4 * j + 3] = s->avctx->pix_fmt == AV_PIX_FMT_RGBA ? src[x * j + 4] : 255;
  1912. }
  1913. src += stride;
  1914. dst += p->linesize[plane];
  1915. }
  1916. av_freep(&five_planes);
  1917. } else if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1918. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE) {
  1919. dst = p->data[plane];
  1920. for (i = 0; i < s->height; i++) {
  1921. for (j = 0; j < s->width; j++) {
  1922. uint64_t k = 65535 - AV_RB16(dst + 8 * j + 6);
  1923. uint64_t r = (65535 - AV_RB16(dst + 8 * j )) * k;
  1924. uint64_t g = (65535 - AV_RB16(dst + 8 * j + 2)) * k;
  1925. uint64_t b = (65535 - AV_RB16(dst + 8 * j + 4)) * k;
  1926. AV_WB16(dst + 8 * j , r * 65537 >> 32);
  1927. AV_WB16(dst + 8 * j + 2, g * 65537 >> 32);
  1928. AV_WB16(dst + 8 * j + 4, b * 65537 >> 32);
  1929. AV_WB16(dst + 8 * j + 6, 65535);
  1930. }
  1931. dst += p->linesize[plane];
  1932. }
  1933. }
  1934. }
  1935. if (s->planar && s->bppcount > 2) {
  1936. FFSWAP(uint8_t*, p->data[0], p->data[2]);
  1937. FFSWAP(int, p->linesize[0], p->linesize[2]);
  1938. FFSWAP(uint8_t*, p->data[0], p->data[1]);
  1939. FFSWAP(int, p->linesize[0], p->linesize[1]);
  1940. }
  1941. if (s->is_bayer && s->white_level && s->bpp == 16 && !is_dng) {
  1942. uint16_t *dst = (uint16_t *)p->data[0];
  1943. for (i = 0; i < s->height; i++) {
  1944. for (j = 0; j < s->width; j++)
  1945. dst[j] = FFMIN((dst[j] / (float)s->white_level) * 65535, 65535);
  1946. dst += stride / 2;
  1947. }
  1948. }
  1949. *got_frame = 1;
  1950. return avpkt->size;
  1951. }
  1952. static av_cold int tiff_init(AVCodecContext *avctx)
  1953. {
  1954. TiffContext *s = avctx->priv_data;
  1955. const AVCodec *codec;
  1956. int ret;
  1957. s->width = 0;
  1958. s->height = 0;
  1959. s->subsampling[0] =
  1960. s->subsampling[1] = 1;
  1961. s->avctx = avctx;
  1962. ff_lzw_decode_open(&s->lzw);
  1963. if (!s->lzw)
  1964. return AVERROR(ENOMEM);
  1965. ff_ccitt_unpack_init();
  1966. /* Allocate JPEG frame */
  1967. s->jpgframe = av_frame_alloc();
  1968. s->jpkt = av_packet_alloc();
  1969. if (!s->jpgframe || !s->jpkt)
  1970. return AVERROR(ENOMEM);
  1971. /* Prepare everything needed for JPEG decoding */
  1972. codec = avcodec_find_decoder(AV_CODEC_ID_MJPEG);
  1973. if (!codec)
  1974. return AVERROR_BUG;
  1975. s->avctx_mjpeg = avcodec_alloc_context3(codec);
  1976. if (!s->avctx_mjpeg)
  1977. return AVERROR(ENOMEM);
  1978. s->avctx_mjpeg->flags = avctx->flags;
  1979. s->avctx_mjpeg->flags2 = avctx->flags2;
  1980. s->avctx_mjpeg->dct_algo = avctx->dct_algo;
  1981. s->avctx_mjpeg->idct_algo = avctx->idct_algo;
  1982. ret = avcodec_open2(s->avctx_mjpeg, codec, NULL);
  1983. if (ret < 0) {
  1984. return ret;
  1985. }
  1986. return 0;
  1987. }
  1988. static av_cold int tiff_end(AVCodecContext *avctx)
  1989. {
  1990. TiffContext *const s = avctx->priv_data;
  1991. free_geotags(s);
  1992. ff_lzw_decode_close(&s->lzw);
  1993. av_freep(&s->deinvert_buf);
  1994. s->deinvert_buf_size = 0;
  1995. av_freep(&s->yuv_line);
  1996. s->yuv_line_size = 0;
  1997. av_freep(&s->fax_buffer);
  1998. s->fax_buffer_size = 0;
  1999. av_frame_free(&s->jpgframe);
  2000. av_packet_free(&s->jpkt);
  2001. avcodec_free_context(&s->avctx_mjpeg);
  2002. return 0;
  2003. }
  2004. #define OFFSET(x) offsetof(TiffContext, x)
  2005. static const AVOption tiff_options[] = {
  2006. { "subimage", "decode subimage instead if available", OFFSET(get_subimage), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  2007. { "thumbnail", "decode embedded thumbnail subimage instead if available", OFFSET(get_thumbnail), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  2008. { "page", "page number of multi-page image to decode (starting from 1)", OFFSET(get_page), AV_OPT_TYPE_INT, {.i64=0}, 0, UINT16_MAX, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  2009. { NULL },
  2010. };
  2011. static const AVClass tiff_decoder_class = {
  2012. .class_name = "TIFF decoder",
  2013. .item_name = av_default_item_name,
  2014. .option = tiff_options,
  2015. .version = LIBAVUTIL_VERSION_INT,
  2016. };
  2017. AVCodec ff_tiff_decoder = {
  2018. .name = "tiff",
  2019. .long_name = NULL_IF_CONFIG_SMALL("TIFF image"),
  2020. .type = AVMEDIA_TYPE_VIDEO,
  2021. .id = AV_CODEC_ID_TIFF,
  2022. .priv_data_size = sizeof(TiffContext),
  2023. .init = tiff_init,
  2024. .close = tiff_end,
  2025. .decode = decode_frame,
  2026. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
  2027. .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP,
  2028. .priv_class = &tiff_decoder_class,
  2029. };