You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

2101 lines
72KB

  1. /*
  2. * Copyright (c) 2006 Konstantin Shishkov
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * TIFF image decoder
  23. * @author Konstantin Shishkov
  24. */
  25. #include "config.h"
  26. #if CONFIG_ZLIB
  27. #include <zlib.h>
  28. #endif
  29. #if CONFIG_LZMA
  30. #define LZMA_API_STATIC
  31. #include <lzma.h>
  32. #endif
  33. #include "libavutil/attributes.h"
  34. #include "libavutil/avstring.h"
  35. #include "libavutil/error.h"
  36. #include "libavutil/intreadwrite.h"
  37. #include "libavutil/imgutils.h"
  38. #include "libavutil/opt.h"
  39. #include "avcodec.h"
  40. #include "bytestream.h"
  41. #include "faxcompr.h"
  42. #include "internal.h"
  43. #include "lzw.h"
  44. #include "mathops.h"
  45. #include "tiff.h"
  46. #include "tiff_data.h"
  47. #include "mjpegdec.h"
  48. #include "thread.h"
  49. #include "get_bits.h"
  50. typedef struct TiffContext {
  51. AVClass *class;
  52. AVCodecContext *avctx;
  53. GetByteContext gb;
  54. /* JPEG decoding for DNG */
  55. AVCodecContext *avctx_mjpeg; // wrapper context for MJPEG
  56. AVFrame *jpgframe; // decoded JPEG tile
  57. int get_subimage;
  58. uint16_t get_page;
  59. int get_thumbnail;
  60. enum TiffType tiff_type;
  61. int width, height;
  62. unsigned int bpp, bppcount;
  63. uint32_t palette[256];
  64. int palette_is_set;
  65. int le;
  66. enum TiffCompr compr;
  67. enum TiffPhotometric photometric;
  68. int planar;
  69. int subsampling[2];
  70. int fax_opts;
  71. int predictor;
  72. int fill_order;
  73. uint32_t res[4];
  74. int is_thumbnail;
  75. int is_bayer;
  76. uint8_t pattern[4];
  77. unsigned black_level;
  78. unsigned white_level;
  79. const uint16_t *dng_lut; // Pointer to DNG linearization table
  80. uint32_t sub_ifd;
  81. uint16_t cur_page;
  82. int strips, rps, sstype;
  83. int sot;
  84. int stripsizesoff, stripsize, stripoff, strippos;
  85. LZWState *lzw;
  86. /* Tile support */
  87. int is_tiled;
  88. int tile_byte_counts_offset, tile_offsets_offset;
  89. int tile_width, tile_length;
  90. int tile_count;
  91. int is_jpeg;
  92. uint8_t *deinvert_buf;
  93. int deinvert_buf_size;
  94. uint8_t *yuv_line;
  95. unsigned int yuv_line_size;
  96. uint8_t *fax_buffer;
  97. unsigned int fax_buffer_size;
  98. int geotag_count;
  99. TiffGeoTag *geotags;
  100. } TiffContext;
  101. static void tiff_set_type(TiffContext *s, enum TiffType tiff_type) {
  102. if (s->tiff_type < tiff_type) // Prioritize higher-valued entries
  103. s->tiff_type = tiff_type;
  104. }
  105. static void free_geotags(TiffContext *const s)
  106. {
  107. int i;
  108. for (i = 0; i < s->geotag_count; i++) {
  109. if (s->geotags[i].val)
  110. av_freep(&s->geotags[i].val);
  111. }
  112. av_freep(&s->geotags);
  113. s->geotag_count = 0;
  114. }
  115. #define RET_GEOKEY(TYPE, array, element)\
  116. if (key >= TIFF_##TYPE##_KEY_ID_OFFSET &&\
  117. key - TIFF_##TYPE##_KEY_ID_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_name_type_map))\
  118. return ff_tiff_##array##_name_type_map[key - TIFF_##TYPE##_KEY_ID_OFFSET].element;
  119. static const char *get_geokey_name(int key)
  120. {
  121. RET_GEOKEY(VERT, vert, name);
  122. RET_GEOKEY(PROJ, proj, name);
  123. RET_GEOKEY(GEOG, geog, name);
  124. RET_GEOKEY(CONF, conf, name);
  125. return NULL;
  126. }
  127. static int get_geokey_type(int key)
  128. {
  129. RET_GEOKEY(VERT, vert, type);
  130. RET_GEOKEY(PROJ, proj, type);
  131. RET_GEOKEY(GEOG, geog, type);
  132. RET_GEOKEY(CONF, conf, type);
  133. return AVERROR_INVALIDDATA;
  134. }
  135. static int cmp_id_key(const void *id, const void *k)
  136. {
  137. return *(const int*)id - ((const TiffGeoTagKeyName*)k)->key;
  138. }
  139. static const char *search_keyval(const TiffGeoTagKeyName *keys, int n, int id)
  140. {
  141. TiffGeoTagKeyName *r = bsearch(&id, keys, n, sizeof(keys[0]), cmp_id_key);
  142. if(r)
  143. return r->name;
  144. return NULL;
  145. }
  146. static char *get_geokey_val(int key, int val)
  147. {
  148. char *ap;
  149. if (val == TIFF_GEO_KEY_UNDEFINED)
  150. return av_strdup("undefined");
  151. if (val == TIFF_GEO_KEY_USER_DEFINED)
  152. return av_strdup("User-Defined");
  153. #define RET_GEOKEY_VAL(TYPE, array)\
  154. if (val >= TIFF_##TYPE##_OFFSET &&\
  155. val - TIFF_##TYPE##_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_codes))\
  156. return av_strdup(ff_tiff_##array##_codes[val - TIFF_##TYPE##_OFFSET]);
  157. switch (key) {
  158. case TIFF_GT_MODEL_TYPE_GEOKEY:
  159. RET_GEOKEY_VAL(GT_MODEL_TYPE, gt_model_type);
  160. break;
  161. case TIFF_GT_RASTER_TYPE_GEOKEY:
  162. RET_GEOKEY_VAL(GT_RASTER_TYPE, gt_raster_type);
  163. break;
  164. case TIFF_GEOG_LINEAR_UNITS_GEOKEY:
  165. case TIFF_PROJ_LINEAR_UNITS_GEOKEY:
  166. case TIFF_VERTICAL_UNITS_GEOKEY:
  167. RET_GEOKEY_VAL(LINEAR_UNIT, linear_unit);
  168. break;
  169. case TIFF_GEOG_ANGULAR_UNITS_GEOKEY:
  170. case TIFF_GEOG_AZIMUTH_UNITS_GEOKEY:
  171. RET_GEOKEY_VAL(ANGULAR_UNIT, angular_unit);
  172. break;
  173. case TIFF_GEOGRAPHIC_TYPE_GEOKEY:
  174. RET_GEOKEY_VAL(GCS_TYPE, gcs_type);
  175. RET_GEOKEY_VAL(GCSE_TYPE, gcse_type);
  176. break;
  177. case TIFF_GEOG_GEODETIC_DATUM_GEOKEY:
  178. RET_GEOKEY_VAL(GEODETIC_DATUM, geodetic_datum);
  179. RET_GEOKEY_VAL(GEODETIC_DATUM_E, geodetic_datum_e);
  180. break;
  181. case TIFF_GEOG_ELLIPSOID_GEOKEY:
  182. RET_GEOKEY_VAL(ELLIPSOID, ellipsoid);
  183. break;
  184. case TIFF_GEOG_PRIME_MERIDIAN_GEOKEY:
  185. RET_GEOKEY_VAL(PRIME_MERIDIAN, prime_meridian);
  186. break;
  187. case TIFF_PROJECTED_CS_TYPE_GEOKEY:
  188. ap = av_strdup(search_keyval(ff_tiff_proj_cs_type_codes, FF_ARRAY_ELEMS(ff_tiff_proj_cs_type_codes), val));
  189. if(ap) return ap;
  190. break;
  191. case TIFF_PROJECTION_GEOKEY:
  192. ap = av_strdup(search_keyval(ff_tiff_projection_codes, FF_ARRAY_ELEMS(ff_tiff_projection_codes), val));
  193. if(ap) return ap;
  194. break;
  195. case TIFF_PROJ_COORD_TRANS_GEOKEY:
  196. RET_GEOKEY_VAL(COORD_TRANS, coord_trans);
  197. break;
  198. case TIFF_VERTICAL_CS_TYPE_GEOKEY:
  199. RET_GEOKEY_VAL(VERT_CS, vert_cs);
  200. RET_GEOKEY_VAL(ORTHO_VERT_CS, ortho_vert_cs);
  201. break;
  202. }
  203. ap = av_malloc(14);
  204. if (ap)
  205. snprintf(ap, 14, "Unknown-%d", val);
  206. return ap;
  207. }
  208. static char *doubles2str(double *dp, int count, const char *sep)
  209. {
  210. int i;
  211. char *ap, *ap0;
  212. uint64_t component_len;
  213. if (!sep) sep = ", ";
  214. component_len = 24LL + strlen(sep);
  215. if (count >= (INT_MAX - 1)/component_len)
  216. return NULL;
  217. ap = av_malloc(component_len * count + 1);
  218. if (!ap)
  219. return NULL;
  220. ap0 = ap;
  221. ap[0] = '\0';
  222. for (i = 0; i < count; i++) {
  223. unsigned l = snprintf(ap, component_len, "%.15g%s", dp[i], sep);
  224. if(l >= component_len) {
  225. av_free(ap0);
  226. return NULL;
  227. }
  228. ap += l;
  229. }
  230. ap0[strlen(ap0) - strlen(sep)] = '\0';
  231. return ap0;
  232. }
  233. static int add_metadata(int count, int type,
  234. const char *name, const char *sep, TiffContext *s, AVFrame *frame)
  235. {
  236. switch(type) {
  237. case TIFF_DOUBLE: return ff_tadd_doubles_metadata(count, name, sep, &s->gb, s->le, &frame->metadata);
  238. case TIFF_SHORT : return ff_tadd_shorts_metadata(count, name, sep, &s->gb, s->le, 0, &frame->metadata);
  239. case TIFF_STRING: return ff_tadd_string_metadata(count, name, &s->gb, s->le, &frame->metadata);
  240. default : return AVERROR_INVALIDDATA;
  241. };
  242. }
  243. static void av_always_inline dng_blit(TiffContext *s, uint8_t *dst, int dst_stride,
  244. const uint8_t *src, int src_stride, int width, int height, int is_u16);
  245. static void av_always_inline horizontal_fill(TiffContext *s,
  246. unsigned int bpp, uint8_t* dst,
  247. int usePtr, const uint8_t *src,
  248. uint8_t c, int width, int offset)
  249. {
  250. switch (bpp) {
  251. case 1:
  252. while (--width >= 0) {
  253. dst[(width+offset)*8+7] = (usePtr ? src[width] : c) & 0x1;
  254. dst[(width+offset)*8+6] = (usePtr ? src[width] : c) >> 1 & 0x1;
  255. dst[(width+offset)*8+5] = (usePtr ? src[width] : c) >> 2 & 0x1;
  256. dst[(width+offset)*8+4] = (usePtr ? src[width] : c) >> 3 & 0x1;
  257. dst[(width+offset)*8+3] = (usePtr ? src[width] : c) >> 4 & 0x1;
  258. dst[(width+offset)*8+2] = (usePtr ? src[width] : c) >> 5 & 0x1;
  259. dst[(width+offset)*8+1] = (usePtr ? src[width] : c) >> 6 & 0x1;
  260. dst[(width+offset)*8+0] = (usePtr ? src[width] : c) >> 7;
  261. }
  262. break;
  263. case 2:
  264. while (--width >= 0) {
  265. dst[(width+offset)*4+3] = (usePtr ? src[width] : c) & 0x3;
  266. dst[(width+offset)*4+2] = (usePtr ? src[width] : c) >> 2 & 0x3;
  267. dst[(width+offset)*4+1] = (usePtr ? src[width] : c) >> 4 & 0x3;
  268. dst[(width+offset)*4+0] = (usePtr ? src[width] : c) >> 6;
  269. }
  270. break;
  271. case 4:
  272. while (--width >= 0) {
  273. dst[(width+offset)*2+1] = (usePtr ? src[width] : c) & 0xF;
  274. dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
  275. }
  276. break;
  277. case 12: {
  278. uint16_t *dst16 = (uint16_t *)dst;
  279. GetBitContext gb;
  280. init_get_bits8(&gb, src, width);
  281. for (int i = 0; i < s->width; i++) {
  282. dst16[i] = get_bits(&gb, 12) << 4;
  283. }
  284. }
  285. break;
  286. default:
  287. if (usePtr) {
  288. memcpy(dst + offset, src, width);
  289. } else {
  290. memset(dst + offset, c, width);
  291. }
  292. }
  293. }
  294. static int deinvert_buffer(TiffContext *s, const uint8_t *src, int size)
  295. {
  296. int i;
  297. av_fast_padded_malloc(&s->deinvert_buf, &s->deinvert_buf_size, size);
  298. if (!s->deinvert_buf)
  299. return AVERROR(ENOMEM);
  300. for (i = 0; i < size; i++)
  301. s->deinvert_buf[i] = ff_reverse[src[i]];
  302. return 0;
  303. }
  304. static void unpack_gray(TiffContext *s, AVFrame *p,
  305. const uint8_t *src, int lnum, int width, int bpp)
  306. {
  307. GetBitContext gb;
  308. uint16_t *dst = (uint16_t *)(p->data[0] + lnum * p->linesize[0]);
  309. init_get_bits8(&gb, src, width);
  310. for (int i = 0; i < s->width; i++) {
  311. dst[i] = get_bits(&gb, bpp);
  312. }
  313. }
  314. static void unpack_yuv(TiffContext *s, AVFrame *p,
  315. const uint8_t *src, int lnum)
  316. {
  317. int i, j, k;
  318. int w = (s->width - 1) / s->subsampling[0] + 1;
  319. uint8_t *pu = &p->data[1][lnum / s->subsampling[1] * p->linesize[1]];
  320. uint8_t *pv = &p->data[2][lnum / s->subsampling[1] * p->linesize[2]];
  321. if (s->width % s->subsampling[0] || s->height % s->subsampling[1]) {
  322. for (i = 0; i < w; i++) {
  323. for (j = 0; j < s->subsampling[1]; j++)
  324. for (k = 0; k < s->subsampling[0]; k++)
  325. p->data[0][FFMIN(lnum + j, s->height-1) * p->linesize[0] +
  326. FFMIN(i * s->subsampling[0] + k, s->width-1)] = *src++;
  327. *pu++ = *src++;
  328. *pv++ = *src++;
  329. }
  330. }else{
  331. for (i = 0; i < w; i++) {
  332. for (j = 0; j < s->subsampling[1]; j++)
  333. for (k = 0; k < s->subsampling[0]; k++)
  334. p->data[0][(lnum + j) * p->linesize[0] +
  335. i * s->subsampling[0] + k] = *src++;
  336. *pu++ = *src++;
  337. *pv++ = *src++;
  338. }
  339. }
  340. }
  341. #if CONFIG_ZLIB
  342. static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
  343. int size)
  344. {
  345. z_stream zstream = { 0 };
  346. int zret;
  347. zstream.next_in = (uint8_t *)src;
  348. zstream.avail_in = size;
  349. zstream.next_out = dst;
  350. zstream.avail_out = *len;
  351. zret = inflateInit(&zstream);
  352. if (zret != Z_OK) {
  353. av_log(NULL, AV_LOG_ERROR, "Inflate init error: %d\n", zret);
  354. return zret;
  355. }
  356. zret = inflate(&zstream, Z_SYNC_FLUSH);
  357. inflateEnd(&zstream);
  358. *len = zstream.total_out;
  359. return zret == Z_STREAM_END ? Z_OK : zret;
  360. }
  361. static int tiff_unpack_zlib(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  362. const uint8_t *src, int size, int width, int lines,
  363. int strip_start, int is_yuv)
  364. {
  365. uint8_t *zbuf;
  366. unsigned long outlen;
  367. int ret, line;
  368. outlen = width * lines;
  369. zbuf = av_malloc(outlen);
  370. if (!zbuf)
  371. return AVERROR(ENOMEM);
  372. if (s->fill_order) {
  373. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  374. av_free(zbuf);
  375. return ret;
  376. }
  377. src = s->deinvert_buf;
  378. }
  379. ret = tiff_uncompress(zbuf, &outlen, src, size);
  380. if (ret != Z_OK) {
  381. av_log(s->avctx, AV_LOG_ERROR,
  382. "Uncompressing failed (%lu of %lu) with error %d\n", outlen,
  383. (unsigned long)width * lines, ret);
  384. av_free(zbuf);
  385. return AVERROR_UNKNOWN;
  386. }
  387. src = zbuf;
  388. for (line = 0; line < lines; line++) {
  389. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  390. horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
  391. } else {
  392. memcpy(dst, src, width);
  393. }
  394. if (is_yuv) {
  395. unpack_yuv(s, p, dst, strip_start + line);
  396. line += s->subsampling[1] - 1;
  397. }
  398. dst += stride;
  399. src += width;
  400. }
  401. av_free(zbuf);
  402. return 0;
  403. }
  404. #endif
  405. #if CONFIG_LZMA
  406. static int tiff_uncompress_lzma(uint8_t *dst, uint64_t *len, const uint8_t *src,
  407. int size)
  408. {
  409. lzma_stream stream = LZMA_STREAM_INIT;
  410. lzma_ret ret;
  411. stream.next_in = (uint8_t *)src;
  412. stream.avail_in = size;
  413. stream.next_out = dst;
  414. stream.avail_out = *len;
  415. ret = lzma_stream_decoder(&stream, UINT64_MAX, 0);
  416. if (ret != LZMA_OK) {
  417. av_log(NULL, AV_LOG_ERROR, "LZMA init error: %d\n", ret);
  418. return ret;
  419. }
  420. ret = lzma_code(&stream, LZMA_RUN);
  421. lzma_end(&stream);
  422. *len = stream.total_out;
  423. return ret == LZMA_STREAM_END ? LZMA_OK : ret;
  424. }
  425. static int tiff_unpack_lzma(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  426. const uint8_t *src, int size, int width, int lines,
  427. int strip_start, int is_yuv)
  428. {
  429. uint64_t outlen = width * (uint64_t)lines;
  430. int ret, line;
  431. uint8_t *buf = av_malloc(outlen);
  432. if (!buf)
  433. return AVERROR(ENOMEM);
  434. if (s->fill_order) {
  435. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  436. av_free(buf);
  437. return ret;
  438. }
  439. src = s->deinvert_buf;
  440. }
  441. ret = tiff_uncompress_lzma(buf, &outlen, src, size);
  442. if (ret != LZMA_OK) {
  443. av_log(s->avctx, AV_LOG_ERROR,
  444. "Uncompressing failed (%"PRIu64" of %"PRIu64") with error %d\n", outlen,
  445. (uint64_t)width * lines, ret);
  446. av_free(buf);
  447. return AVERROR_UNKNOWN;
  448. }
  449. src = buf;
  450. for (line = 0; line < lines; line++) {
  451. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  452. horizontal_fill(s, s->bpp, dst, 1, src, 0, width, 0);
  453. } else {
  454. memcpy(dst, src, width);
  455. }
  456. if (is_yuv) {
  457. unpack_yuv(s, p, dst, strip_start + line);
  458. line += s->subsampling[1] - 1;
  459. }
  460. dst += stride;
  461. src += width;
  462. }
  463. av_free(buf);
  464. return 0;
  465. }
  466. #endif
  467. static int tiff_unpack_fax(TiffContext *s, uint8_t *dst, int stride,
  468. const uint8_t *src, int size, int width, int lines)
  469. {
  470. int i, ret = 0;
  471. int line;
  472. uint8_t *src2;
  473. av_fast_padded_malloc(&s->fax_buffer, &s->fax_buffer_size, size);
  474. src2 = s->fax_buffer;
  475. if (!src2) {
  476. av_log(s->avctx, AV_LOG_ERROR,
  477. "Error allocating temporary buffer\n");
  478. return AVERROR(ENOMEM);
  479. }
  480. if (!s->fill_order) {
  481. memcpy(src2, src, size);
  482. } else {
  483. for (i = 0; i < size; i++)
  484. src2[i] = ff_reverse[src[i]];
  485. }
  486. memset(src2 + size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
  487. ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride,
  488. s->compr, s->fax_opts);
  489. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  490. for (line = 0; line < lines; line++) {
  491. horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
  492. dst += stride;
  493. }
  494. return ret;
  495. }
  496. static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  497. const uint8_t *src, int size, int strip_start, int lines)
  498. {
  499. PutByteContext pb;
  500. int c, line, pixels, code, ret;
  501. const uint8_t *ssrc = src;
  502. int width = ((s->width * s->bpp) + 7) >> 3;
  503. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(p->format);
  504. int is_yuv = !(desc->flags & AV_PIX_FMT_FLAG_RGB) &&
  505. (desc->flags & AV_PIX_FMT_FLAG_PLANAR) &&
  506. desc->nb_components >= 3;
  507. int is_dng;
  508. if (s->planar)
  509. width /= s->bppcount;
  510. if (size <= 0)
  511. return AVERROR_INVALIDDATA;
  512. if (is_yuv) {
  513. int bytes_per_row = (((s->width - 1) / s->subsampling[0] + 1) * s->bpp *
  514. s->subsampling[0] * s->subsampling[1] + 7) >> 3;
  515. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row);
  516. if (s->yuv_line == NULL) {
  517. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  518. return AVERROR(ENOMEM);
  519. }
  520. dst = s->yuv_line;
  521. stride = 0;
  522. width = (s->width - 1) / s->subsampling[0] + 1;
  523. width = width * s->subsampling[0] * s->subsampling[1] + 2*width;
  524. av_assert0(width <= bytes_per_row);
  525. av_assert0(s->bpp == 24);
  526. }
  527. if (s->is_bayer) {
  528. width = (s->bpp * s->width + 7) >> 3;
  529. }
  530. if (p->format == AV_PIX_FMT_GRAY12) {
  531. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, width);
  532. if (s->yuv_line == NULL) {
  533. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  534. return AVERROR(ENOMEM);
  535. }
  536. dst = s->yuv_line;
  537. stride = 0;
  538. }
  539. if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
  540. #if CONFIG_ZLIB
  541. return tiff_unpack_zlib(s, p, dst, stride, src, size, width, lines,
  542. strip_start, is_yuv);
  543. #else
  544. av_log(s->avctx, AV_LOG_ERROR,
  545. "zlib support not enabled, "
  546. "deflate compression not supported\n");
  547. return AVERROR(ENOSYS);
  548. #endif
  549. }
  550. if (s->compr == TIFF_LZMA) {
  551. #if CONFIG_LZMA
  552. return tiff_unpack_lzma(s, p, dst, stride, src, size, width, lines,
  553. strip_start, is_yuv);
  554. #else
  555. av_log(s->avctx, AV_LOG_ERROR,
  556. "LZMA support not enabled\n");
  557. return AVERROR(ENOSYS);
  558. #endif
  559. }
  560. if (s->compr == TIFF_LZW) {
  561. if (s->fill_order) {
  562. if ((ret = deinvert_buffer(s, src, size)) < 0)
  563. return ret;
  564. ssrc = src = s->deinvert_buf;
  565. }
  566. if (size > 1 && !src[0] && (src[1]&1)) {
  567. av_log(s->avctx, AV_LOG_ERROR, "Old style LZW is unsupported\n");
  568. }
  569. if ((ret = ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF)) < 0) {
  570. av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
  571. return ret;
  572. }
  573. for (line = 0; line < lines; line++) {
  574. pixels = ff_lzw_decode(s->lzw, dst, width);
  575. if (pixels < width) {
  576. av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
  577. pixels, width);
  578. return AVERROR_INVALIDDATA;
  579. }
  580. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  581. horizontal_fill(s, s->bpp, dst, 1, dst, 0, width, 0);
  582. if (is_yuv) {
  583. unpack_yuv(s, p, dst, strip_start + line);
  584. line += s->subsampling[1] - 1;
  585. } else if (p->format == AV_PIX_FMT_GRAY12) {
  586. unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
  587. }
  588. dst += stride;
  589. }
  590. return 0;
  591. }
  592. if (s->compr == TIFF_CCITT_RLE ||
  593. s->compr == TIFF_G3 ||
  594. s->compr == TIFF_G4) {
  595. if (is_yuv || p->format == AV_PIX_FMT_GRAY12)
  596. return AVERROR_INVALIDDATA;
  597. return tiff_unpack_fax(s, dst, stride, src, size, width, lines);
  598. }
  599. bytestream2_init(&s->gb, src, size);
  600. bytestream2_init_writer(&pb, dst, is_yuv ? s->yuv_line_size : (stride * lines));
  601. is_dng = (s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG);
  602. for (line = 0; line < lines; line++) {
  603. if (src - ssrc > size) {
  604. av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
  605. return AVERROR_INVALIDDATA;
  606. }
  607. if (bytestream2_get_bytes_left(&s->gb) == 0 || bytestream2_get_eof(&pb))
  608. break;
  609. bytestream2_seek_p(&pb, stride * line, SEEK_SET);
  610. switch (s->compr) {
  611. case TIFF_RAW:
  612. if (ssrc + size - src < width)
  613. return AVERROR_INVALIDDATA;
  614. if (!s->fill_order) {
  615. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 || s->is_bayer),
  616. dst, 1, src, 0, width, 0);
  617. } else {
  618. int i;
  619. for (i = 0; i < width; i++)
  620. dst[i] = ff_reverse[src[i]];
  621. }
  622. /* Color processing for DNG images with uncompressed strips (non-tiled) */
  623. if (is_dng) {
  624. int is_u16, pixel_size_bytes, pixel_size_bits;
  625. is_u16 = (s->bpp > 8);
  626. pixel_size_bits = (is_u16 ? 16 : 8);
  627. pixel_size_bytes = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
  628. dng_blit(s,
  629. dst,
  630. 0, // no stride, only 1 line
  631. dst,
  632. 0, // no stride, only 1 line
  633. width / pixel_size_bytes * pixel_size_bits / s->bpp * s->bppcount, // need to account for [1, 16] bpp
  634. 1,
  635. is_u16);
  636. }
  637. src += width;
  638. break;
  639. case TIFF_PACKBITS:
  640. for (pixels = 0; pixels < width;) {
  641. if (ssrc + size - src < 2) {
  642. av_log(s->avctx, AV_LOG_ERROR, "Read went out of bounds\n");
  643. return AVERROR_INVALIDDATA;
  644. }
  645. code = s->fill_order ? (int8_t) ff_reverse[*src++]: (int8_t) *src++;
  646. if (code >= 0) {
  647. code++;
  648. if (pixels + code > width ||
  649. ssrc + size - src < code) {
  650. av_log(s->avctx, AV_LOG_ERROR,
  651. "Copy went out of bounds\n");
  652. return AVERROR_INVALIDDATA;
  653. }
  654. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  655. dst, 1, src, 0, code, pixels);
  656. src += code;
  657. pixels += code;
  658. } else if (code != -128) { // -127..-1
  659. code = (-code) + 1;
  660. if (pixels + code > width) {
  661. av_log(s->avctx, AV_LOG_ERROR,
  662. "Run went out of bounds\n");
  663. return AVERROR_INVALIDDATA;
  664. }
  665. c = *src++;
  666. horizontal_fill(s, s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  667. dst, 0, NULL, c, code, pixels);
  668. pixels += code;
  669. }
  670. }
  671. if (s->fill_order) {
  672. int i;
  673. for (i = 0; i < width; i++)
  674. dst[i] = ff_reverse[dst[i]];
  675. }
  676. break;
  677. }
  678. if (is_yuv) {
  679. unpack_yuv(s, p, dst, strip_start + line);
  680. line += s->subsampling[1] - 1;
  681. } else if (p->format == AV_PIX_FMT_GRAY12) {
  682. unpack_gray(s, p, dst, strip_start + line, width, s->bpp);
  683. }
  684. dst += stride;
  685. }
  686. return 0;
  687. }
  688. static float av_always_inline linear_to_srgb(float value) {
  689. if (value <= 0.0031308f)
  690. return value * 12.92f;
  691. else
  692. return powf(value * 1.055f, 1.0f / 2.4f) - 0.055f;
  693. }
  694. /**
  695. * Map stored raw sensor values into linear reference values (see: DNG Specification - Chapter 5)
  696. * Then convert to sRGB color space.
  697. */
  698. static uint16_t av_always_inline dng_process_color16(uint16_t value,
  699. const uint16_t *lut,
  700. uint16_t black_level,
  701. float scale_factor) {
  702. float value_norm;
  703. // Lookup table lookup
  704. if (lut)
  705. value = lut[value];
  706. // Black level subtraction
  707. value = av_clip_uint16_c((unsigned)value - black_level);
  708. // Color scaling
  709. value_norm = (float)value * scale_factor;
  710. // Color space conversion (sRGB)
  711. value = av_clip_uint16_c((uint16_t)(linear_to_srgb(value_norm) * 0xFFFF));
  712. return value;
  713. }
  714. static uint16_t av_always_inline dng_process_color8(uint16_t value,
  715. const uint16_t *lut,
  716. uint16_t black_level,
  717. float scale_factor) {
  718. return dng_process_color16(value, lut, black_level, scale_factor) >> 8;
  719. }
  720. static void dng_blit(TiffContext *s, uint8_t *dst, int dst_stride,
  721. const uint8_t *src, int src_stride,
  722. int width, int height, int is_u16)
  723. {
  724. int line, col;
  725. float scale_factor;
  726. scale_factor = 1.0f / (s->white_level - s->black_level);
  727. if (is_u16) {
  728. for (line = 0; line < height; line++) {
  729. uint16_t *dst_u16 = (uint16_t *)dst;
  730. uint16_t *src_u16 = (uint16_t *)src;
  731. for (col = 0; col < width; col++)
  732. *dst_u16++ = dng_process_color16(*src_u16++, s->dng_lut, s->black_level, scale_factor);
  733. dst += dst_stride * sizeof(uint16_t);
  734. src += src_stride * sizeof(uint16_t);
  735. }
  736. } else {
  737. for (line = 0; line < height; line++) {
  738. for (col = 0; col < width; col++)
  739. *dst++ = dng_process_color8(*src++, s->dng_lut, s->black_level, scale_factor);
  740. dst += dst_stride;
  741. src += src_stride;
  742. }
  743. }
  744. }
  745. static int dng_decode_jpeg_tile(AVCodecContext *avctx, AVFrame *frame,
  746. int tile_byte_count, int x, int y, int w, int h)
  747. {
  748. TiffContext *s = avctx->priv_data;
  749. AVPacket jpkt;
  750. uint8_t *dst_data, *src_data;
  751. uint32_t dst_offset; /* offset from dst buffer in pixels */
  752. int is_u16, pixel_size;
  753. int ret;
  754. /* Prepare a packet and send to the MJPEG decoder */
  755. av_init_packet(&jpkt);
  756. jpkt.data = (uint8_t*)s->gb.buffer;
  757. jpkt.size = tile_byte_count;
  758. ret = avcodec_send_packet(s->avctx_mjpeg, &jpkt);
  759. if (ret < 0) {
  760. av_log(avctx, AV_LOG_ERROR, "Error submitting a packet for decoding\n");
  761. return ret;
  762. }
  763. ret = avcodec_receive_frame(s->avctx_mjpeg, s->jpgframe);
  764. if (ret < 0) {
  765. av_log(avctx, AV_LOG_ERROR, "JPEG decoding error: %s.\n", av_err2str(ret));
  766. /* Normally skip, error if explode */
  767. if (avctx->err_recognition & AV_EF_EXPLODE)
  768. return AVERROR_INVALIDDATA;
  769. else
  770. return 0;
  771. }
  772. /* Copy the outputted tile's pixels from 'jpgframe' to 'frame' (final buffer) */
  773. is_u16 = (s->bpp > 8);
  774. pixel_size = (is_u16 ? sizeof(uint16_t) : sizeof(uint8_t));
  775. dst_offset = x + frame->linesize[0] * y / pixel_size;
  776. dst_data = frame->data[0] + dst_offset * pixel_size;
  777. src_data = s->jpgframe->data[0];
  778. dng_blit(s,
  779. dst_data,
  780. frame->linesize[0] / pixel_size,
  781. src_data,
  782. s->jpgframe->linesize[0] / pixel_size,
  783. w,
  784. h,
  785. is_u16);
  786. av_frame_unref(s->jpgframe);
  787. return 0;
  788. }
  789. static int dng_decode_tiles(AVCodecContext *avctx, AVFrame *frame)
  790. {
  791. TiffContext *s = avctx->priv_data;
  792. int tile_idx;
  793. int tile_offset_offset, tile_offset;
  794. int tile_byte_count_offset, tile_byte_count;
  795. int tile_count_x, tile_count_y;
  796. int tile_width, tile_length;
  797. int has_width_leftover, has_height_leftover;
  798. int tile_x = 0, tile_y = 0;
  799. int pos_x = 0, pos_y = 0;
  800. int ret;
  801. has_width_leftover = (s->width % s->tile_width != 0);
  802. has_height_leftover = (s->height % s->tile_length != 0);
  803. /* Calculate tile counts (round up) */
  804. tile_count_x = (s->width + s->tile_width - 1) / s->tile_width;
  805. tile_count_y = (s->height + s->tile_length - 1) / s->tile_length;
  806. /* Iterate over the number of tiles */
  807. for (tile_idx = 0; tile_idx < s->tile_count; tile_idx++) {
  808. tile_x = tile_idx % tile_count_x;
  809. tile_y = tile_idx / tile_count_x;
  810. if (has_width_leftover && tile_x == tile_count_x - 1) // If on the right-most tile
  811. tile_width = s->width % s->tile_width;
  812. else
  813. tile_width = s->tile_width;
  814. if (has_height_leftover && tile_y == tile_count_y - 1) // If on the bottom-most tile
  815. tile_length = s->height % s->tile_length;
  816. else
  817. tile_length = s->tile_length;
  818. /* Read tile offset */
  819. tile_offset_offset = s->tile_offsets_offset + tile_idx * sizeof(int);
  820. bytestream2_seek(&s->gb, tile_offset_offset, SEEK_SET);
  821. tile_offset = ff_tget_long(&s->gb, s->le);
  822. /* Read tile byte size */
  823. tile_byte_count_offset = s->tile_byte_counts_offset + tile_idx * sizeof(int);
  824. bytestream2_seek(&s->gb, tile_byte_count_offset, SEEK_SET);
  825. tile_byte_count = ff_tget_long(&s->gb, s->le);
  826. /* Seek to tile data */
  827. bytestream2_seek(&s->gb, tile_offset, SEEK_SET);
  828. /* Decode JPEG tile and copy it in the reference frame */
  829. ret = dng_decode_jpeg_tile(avctx, frame, tile_byte_count, pos_x, pos_y, tile_width, tile_length);
  830. if (ret < 0)
  831. return ret;
  832. /* Advance current positions */
  833. pos_x += tile_width;
  834. if (tile_x == tile_count_x - 1) { // If on the right edge
  835. pos_x = 0;
  836. pos_y += tile_length;
  837. }
  838. }
  839. return 0;
  840. }
  841. static int dng_decode(AVCodecContext *avctx, AVFrame *frame, AVPacket *avpkt) {
  842. int ret;
  843. TiffContext *s = avctx->priv_data;
  844. s->jpgframe->width = s->tile_width;
  845. s->jpgframe->height = s->tile_length;
  846. s->avctx_mjpeg->width = s->tile_width;
  847. s->avctx_mjpeg->height = s->tile_length;
  848. /* Decode all tiles in a frame */
  849. ret = dng_decode_tiles(avctx, frame);
  850. if (ret < 0)
  851. return ret;
  852. /* Frame is ready to be output */
  853. frame->pict_type = AV_PICTURE_TYPE_I;
  854. frame->key_frame = 1;
  855. return avpkt->size;
  856. }
  857. static int init_image(TiffContext *s, ThreadFrame *frame)
  858. {
  859. int ret;
  860. int create_gray_palette = 0;
  861. // make sure there is no aliasing in the following switch
  862. if (s->bpp >= 100 || s->bppcount >= 10) {
  863. av_log(s->avctx, AV_LOG_ERROR,
  864. "Unsupported image parameters: bpp=%d, bppcount=%d\n",
  865. s->bpp, s->bppcount);
  866. return AVERROR_INVALIDDATA;
  867. }
  868. switch (s->planar * 1000 + s->bpp * 10 + s->bppcount + s->is_bayer * 10000) {
  869. case 11:
  870. if (!s->palette_is_set) {
  871. s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK;
  872. break;
  873. }
  874. case 21:
  875. case 41:
  876. s->avctx->pix_fmt = AV_PIX_FMT_PAL8;
  877. if (!s->palette_is_set) {
  878. create_gray_palette = 1;
  879. }
  880. break;
  881. case 81:
  882. s->avctx->pix_fmt = s->palette_is_set ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8;
  883. break;
  884. case 121:
  885. s->avctx->pix_fmt = AV_PIX_FMT_GRAY12;
  886. break;
  887. case 10081:
  888. switch (AV_RL32(s->pattern)) {
  889. case 0x02010100:
  890. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB8;
  891. break;
  892. case 0x00010102:
  893. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR8;
  894. break;
  895. case 0x01000201:
  896. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG8;
  897. break;
  898. case 0x01020001:
  899. s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG8;
  900. break;
  901. default:
  902. av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
  903. AV_RL32(s->pattern));
  904. return AVERROR_PATCHWELCOME;
  905. }
  906. break;
  907. case 10121:
  908. switch (AV_RL32(s->pattern)) {
  909. case 0x02010100:
  910. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_RGGB16LE : AV_PIX_FMT_BAYER_RGGB16BE;
  911. break;
  912. case 0x00010102:
  913. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_BGGR16LE : AV_PIX_FMT_BAYER_BGGR16BE;
  914. break;
  915. case 0x01000201:
  916. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GBRG16LE : AV_PIX_FMT_BAYER_GBRG16BE;
  917. break;
  918. case 0x01020001:
  919. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GRBG16LE : AV_PIX_FMT_BAYER_GRBG16BE;
  920. break;
  921. default:
  922. av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
  923. AV_RL32(s->pattern));
  924. return AVERROR_PATCHWELCOME;
  925. }
  926. /* Force endianness as mentioned in 'DNG Specification: Chapter 3: BitsPerSample'
  927. NOTE: The spec actually specifies big-endian, not sure why we need little-endian, but
  928. such images don't work otherwise. Examples are images produced by Zenmuse X7. */
  929. if ((s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG)
  930. && (s->bpp != 8 && s->bpp != 16 && s->bpp != 32)) {
  931. switch (s->avctx->pix_fmt) {
  932. case AV_PIX_FMT_BAYER_RGGB16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_RGGB16LE; break;
  933. case AV_PIX_FMT_BAYER_BGGR16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_BGGR16LE; break;
  934. case AV_PIX_FMT_BAYER_GBRG16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GBRG16LE; break;
  935. case AV_PIX_FMT_BAYER_GRBG16BE: s->avctx->pix_fmt = AV_PIX_FMT_BAYER_GRBG16LE; break;
  936. }
  937. }
  938. break;
  939. case 10161:
  940. switch (AV_RL32(s->pattern)) {
  941. case 0x02010100:
  942. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_RGGB16LE : AV_PIX_FMT_BAYER_RGGB16BE;
  943. break;
  944. case 0x00010102:
  945. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_BGGR16LE : AV_PIX_FMT_BAYER_BGGR16BE;
  946. break;
  947. case 0x01000201:
  948. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GBRG16LE : AV_PIX_FMT_BAYER_GBRG16BE;
  949. break;
  950. case 0x01020001:
  951. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_BAYER_GRBG16LE : AV_PIX_FMT_BAYER_GRBG16BE;
  952. break;
  953. default:
  954. av_log(s->avctx, AV_LOG_ERROR, "Unsupported Bayer pattern: 0x%X\n",
  955. AV_RL32(s->pattern));
  956. return AVERROR_PATCHWELCOME;
  957. }
  958. break;
  959. case 243:
  960. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  961. if (s->subsampling[0] == 1 && s->subsampling[1] == 1) {
  962. s->avctx->pix_fmt = AV_PIX_FMT_YUV444P;
  963. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 1) {
  964. s->avctx->pix_fmt = AV_PIX_FMT_YUV422P;
  965. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 1) {
  966. s->avctx->pix_fmt = AV_PIX_FMT_YUV411P;
  967. } else if (s->subsampling[0] == 1 && s->subsampling[1] == 2) {
  968. s->avctx->pix_fmt = AV_PIX_FMT_YUV440P;
  969. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 2) {
  970. s->avctx->pix_fmt = AV_PIX_FMT_YUV420P;
  971. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 4) {
  972. s->avctx->pix_fmt = AV_PIX_FMT_YUV410P;
  973. } else {
  974. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr subsampling\n");
  975. return AVERROR_PATCHWELCOME;
  976. }
  977. } else
  978. s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
  979. break;
  980. case 161:
  981. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE;
  982. break;
  983. case 162:
  984. s->avctx->pix_fmt = AV_PIX_FMT_YA8;
  985. break;
  986. case 322:
  987. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_YA16LE : AV_PIX_FMT_YA16BE;
  988. break;
  989. case 324:
  990. s->avctx->pix_fmt = s->photometric == TIFF_PHOTOMETRIC_SEPARATED ? AV_PIX_FMT_RGB0 : AV_PIX_FMT_RGBA;
  991. break;
  992. case 405:
  993. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED)
  994. s->avctx->pix_fmt = AV_PIX_FMT_RGBA;
  995. else {
  996. av_log(s->avctx, AV_LOG_ERROR,
  997. "bpp=40 without PHOTOMETRIC_SEPARATED is unsupported\n");
  998. return AVERROR_PATCHWELCOME;
  999. }
  1000. break;
  1001. case 483:
  1002. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE;
  1003. break;
  1004. case 644:
  1005. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGBA64LE : AV_PIX_FMT_RGBA64BE;
  1006. break;
  1007. case 1243:
  1008. s->avctx->pix_fmt = AV_PIX_FMT_GBRP;
  1009. break;
  1010. case 1324:
  1011. s->avctx->pix_fmt = AV_PIX_FMT_GBRAP;
  1012. break;
  1013. case 1483:
  1014. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRP16LE : AV_PIX_FMT_GBRP16BE;
  1015. break;
  1016. case 1644:
  1017. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRAP16LE : AV_PIX_FMT_GBRAP16BE;
  1018. break;
  1019. default:
  1020. av_log(s->avctx, AV_LOG_ERROR,
  1021. "This format is not supported (bpp=%d, bppcount=%d)\n",
  1022. s->bpp, s->bppcount);
  1023. return AVERROR_INVALIDDATA;
  1024. }
  1025. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1026. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->avctx->pix_fmt);
  1027. if((desc->flags & AV_PIX_FMT_FLAG_RGB) ||
  1028. !(desc->flags & AV_PIX_FMT_FLAG_PLANAR) ||
  1029. desc->nb_components < 3) {
  1030. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr variant\n");
  1031. return AVERROR_INVALIDDATA;
  1032. }
  1033. }
  1034. if (s->width != s->avctx->width || s->height != s->avctx->height) {
  1035. ret = ff_set_dimensions(s->avctx, s->width, s->height);
  1036. if (ret < 0)
  1037. return ret;
  1038. }
  1039. if ((ret = ff_thread_get_buffer(s->avctx, frame, 0)) < 0)
  1040. return ret;
  1041. if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  1042. if (!create_gray_palette)
  1043. memcpy(frame->f->data[1], s->palette, sizeof(s->palette));
  1044. else {
  1045. /* make default grayscale pal */
  1046. int i;
  1047. uint32_t *pal = (uint32_t *)frame->f->data[1];
  1048. for (i = 0; i < 1<<s->bpp; i++)
  1049. pal[i] = 0xFFU << 24 | i * 255 / ((1<<s->bpp) - 1) * 0x010101;
  1050. }
  1051. }
  1052. return 0;
  1053. }
  1054. static void set_sar(TiffContext *s, unsigned tag, unsigned num, unsigned den)
  1055. {
  1056. int offset = tag == TIFF_YRES ? 2 : 0;
  1057. s->res[offset++] = num;
  1058. s->res[offset] = den;
  1059. if (s->res[0] && s->res[1] && s->res[2] && s->res[3]) {
  1060. uint64_t num = s->res[2] * (uint64_t)s->res[1];
  1061. uint64_t den = s->res[0] * (uint64_t)s->res[3];
  1062. if (num > INT64_MAX || den > INT64_MAX) {
  1063. num = num >> 1;
  1064. den = den >> 1;
  1065. }
  1066. av_reduce(&s->avctx->sample_aspect_ratio.num, &s->avctx->sample_aspect_ratio.den,
  1067. num, den, INT32_MAX);
  1068. if (!s->avctx->sample_aspect_ratio.den)
  1069. s->avctx->sample_aspect_ratio = (AVRational) {0, 1};
  1070. }
  1071. }
  1072. static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
  1073. {
  1074. unsigned tag, type, count, off, value = 0, value2 = 1; // value2 is a denominator so init. to 1
  1075. int i, start;
  1076. int pos;
  1077. int ret;
  1078. double *dp;
  1079. ret = ff_tread_tag(&s->gb, s->le, &tag, &type, &count, &start);
  1080. if (ret < 0) {
  1081. goto end;
  1082. }
  1083. off = bytestream2_tell(&s->gb);
  1084. if (count == 1) {
  1085. switch (type) {
  1086. case TIFF_BYTE:
  1087. case TIFF_SHORT:
  1088. case TIFF_LONG:
  1089. value = ff_tget(&s->gb, type, s->le);
  1090. break;
  1091. case TIFF_RATIONAL:
  1092. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  1093. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  1094. break;
  1095. case TIFF_STRING:
  1096. if (count <= 4) {
  1097. break;
  1098. }
  1099. default:
  1100. value = UINT_MAX;
  1101. }
  1102. }
  1103. switch (tag) {
  1104. case TIFF_SUBFILE:
  1105. s->is_thumbnail = (value != 0);
  1106. break;
  1107. case TIFF_WIDTH:
  1108. s->width = value;
  1109. break;
  1110. case TIFF_HEIGHT:
  1111. s->height = value;
  1112. break;
  1113. case TIFF_BPP:
  1114. if (count > 5U) {
  1115. av_log(s->avctx, AV_LOG_ERROR,
  1116. "This format is not supported (bpp=%d, %d components)\n",
  1117. value, count);
  1118. return AVERROR_INVALIDDATA;
  1119. }
  1120. s->bppcount = count;
  1121. if (count == 1)
  1122. s->bpp = value;
  1123. else {
  1124. switch (type) {
  1125. case TIFF_BYTE:
  1126. case TIFF_SHORT:
  1127. case TIFF_LONG:
  1128. s->bpp = 0;
  1129. if (bytestream2_get_bytes_left(&s->gb) < type_sizes[type] * count)
  1130. return AVERROR_INVALIDDATA;
  1131. for (i = 0; i < count; i++)
  1132. s->bpp += ff_tget(&s->gb, type, s->le);
  1133. break;
  1134. default:
  1135. s->bpp = -1;
  1136. }
  1137. }
  1138. break;
  1139. case TIFF_SAMPLES_PER_PIXEL:
  1140. if (count != 1) {
  1141. av_log(s->avctx, AV_LOG_ERROR,
  1142. "Samples per pixel requires a single value, many provided\n");
  1143. return AVERROR_INVALIDDATA;
  1144. }
  1145. if (value > 5U) {
  1146. av_log(s->avctx, AV_LOG_ERROR,
  1147. "Samples per pixel %d is too large\n", value);
  1148. return AVERROR_INVALIDDATA;
  1149. }
  1150. if (s->bppcount == 1)
  1151. s->bpp *= value;
  1152. s->bppcount = value;
  1153. break;
  1154. case TIFF_COMPR:
  1155. s->compr = value;
  1156. av_log(s->avctx, AV_LOG_DEBUG, "compression: %d\n", s->compr);
  1157. s->predictor = 0;
  1158. switch (s->compr) {
  1159. case TIFF_RAW:
  1160. case TIFF_PACKBITS:
  1161. case TIFF_LZW:
  1162. case TIFF_CCITT_RLE:
  1163. break;
  1164. case TIFF_G3:
  1165. case TIFF_G4:
  1166. s->fax_opts = 0;
  1167. break;
  1168. case TIFF_DEFLATE:
  1169. case TIFF_ADOBE_DEFLATE:
  1170. #if CONFIG_ZLIB
  1171. break;
  1172. #else
  1173. av_log(s->avctx, AV_LOG_ERROR, "Deflate: ZLib not compiled in\n");
  1174. return AVERROR(ENOSYS);
  1175. #endif
  1176. case TIFF_JPEG:
  1177. case TIFF_NEWJPEG:
  1178. s->is_jpeg = 1;
  1179. break;
  1180. case TIFF_LZMA:
  1181. #if CONFIG_LZMA
  1182. break;
  1183. #else
  1184. av_log(s->avctx, AV_LOG_ERROR, "LZMA not compiled in\n");
  1185. return AVERROR(ENOSYS);
  1186. #endif
  1187. default:
  1188. av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
  1189. s->compr);
  1190. return AVERROR_INVALIDDATA;
  1191. }
  1192. break;
  1193. case TIFF_ROWSPERSTRIP:
  1194. if (!value || (type == TIFF_LONG && value == UINT_MAX))
  1195. value = s->height;
  1196. s->rps = FFMIN(value, s->height);
  1197. break;
  1198. case TIFF_STRIP_OFFS:
  1199. if (count == 1) {
  1200. if (value > INT_MAX) {
  1201. av_log(s->avctx, AV_LOG_ERROR,
  1202. "strippos %u too large\n", value);
  1203. return AVERROR_INVALIDDATA;
  1204. }
  1205. s->strippos = 0;
  1206. s->stripoff = value;
  1207. } else
  1208. s->strippos = off;
  1209. s->strips = count;
  1210. if (s->strips == 1)
  1211. s->rps = s->height;
  1212. s->sot = type;
  1213. break;
  1214. case TIFF_STRIP_SIZE:
  1215. if (count == 1) {
  1216. if (value > INT_MAX) {
  1217. av_log(s->avctx, AV_LOG_ERROR,
  1218. "stripsize %u too large\n", value);
  1219. return AVERROR_INVALIDDATA;
  1220. }
  1221. s->stripsizesoff = 0;
  1222. s->stripsize = value;
  1223. s->strips = 1;
  1224. } else {
  1225. s->stripsizesoff = off;
  1226. }
  1227. s->strips = count;
  1228. s->sstype = type;
  1229. break;
  1230. case TIFF_XRES:
  1231. case TIFF_YRES:
  1232. set_sar(s, tag, value, value2);
  1233. break;
  1234. case TIFF_TILE_OFFSETS:
  1235. s->tile_offsets_offset = off;
  1236. s->tile_count = count;
  1237. s->is_tiled = 1;
  1238. break;
  1239. case TIFF_TILE_BYTE_COUNTS:
  1240. s->tile_byte_counts_offset = off;
  1241. break;
  1242. case TIFF_TILE_LENGTH:
  1243. s->tile_length = value;
  1244. break;
  1245. case TIFF_TILE_WIDTH:
  1246. s->tile_width = value;
  1247. break;
  1248. case TIFF_PREDICTOR:
  1249. s->predictor = value;
  1250. break;
  1251. case TIFF_SUB_IFDS:
  1252. if (count == 1)
  1253. s->sub_ifd = value;
  1254. else if (count > 1)
  1255. s->sub_ifd = ff_tget(&s->gb, TIFF_LONG, s->le); /** Only get the first SubIFD */
  1256. break;
  1257. case DNG_LINEARIZATION_TABLE: {
  1258. uint32_t lut_offset = value;
  1259. uint32_t lut_size = count;
  1260. uint32_t lut_wanted_size = 1 << s->bpp;
  1261. if (lut_wanted_size != lut_size)
  1262. av_log(s->avctx, AV_LOG_WARNING, "DNG contains LUT with invalid size (%"PRIu32"), disabling LUT\n", lut_size);
  1263. else if (lut_offset >= bytestream2_size(&s->gb))
  1264. av_log(s->avctx, AV_LOG_WARNING, "DNG contains LUT with invalid offset (%"PRIu32"), disabling LUT\n", lut_offset);
  1265. else
  1266. s->dng_lut = (uint16_t*)(s->gb.buffer + lut_offset);
  1267. break;
  1268. }
  1269. case DNG_BLACK_LEVEL:
  1270. if (count > 1) { /* Use the first value in the pattern (assume they're all the same) */
  1271. if (type == TIFF_RATIONAL) {
  1272. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  1273. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  1274. s->black_level = value / value2;
  1275. } else
  1276. s->black_level = ff_tget(&s->gb, type, s->le);
  1277. av_log(s->avctx, AV_LOG_WARNING, "Assuming black level pattern values are identical\n");
  1278. } else {
  1279. s->black_level = value / value2;
  1280. }
  1281. break;
  1282. case DNG_WHITE_LEVEL:
  1283. s->white_level = value;
  1284. break;
  1285. case TIFF_CFA_PATTERN_DIM:
  1286. if (count != 2 || (ff_tget(&s->gb, type, s->le) != 2 &&
  1287. ff_tget(&s->gb, type, s->le) != 2)) {
  1288. av_log(s->avctx, AV_LOG_ERROR, "CFA Pattern dimensions are not 2x2\n");
  1289. return AVERROR_INVALIDDATA;
  1290. }
  1291. break;
  1292. case TIFF_CFA_PATTERN:
  1293. s->is_bayer = 1;
  1294. s->pattern[0] = ff_tget(&s->gb, type, s->le);
  1295. s->pattern[1] = ff_tget(&s->gb, type, s->le);
  1296. s->pattern[2] = ff_tget(&s->gb, type, s->le);
  1297. s->pattern[3] = ff_tget(&s->gb, type, s->le);
  1298. break;
  1299. case TIFF_PHOTOMETRIC:
  1300. switch (value) {
  1301. case TIFF_PHOTOMETRIC_WHITE_IS_ZERO:
  1302. case TIFF_PHOTOMETRIC_BLACK_IS_ZERO:
  1303. case TIFF_PHOTOMETRIC_RGB:
  1304. case TIFF_PHOTOMETRIC_PALETTE:
  1305. case TIFF_PHOTOMETRIC_SEPARATED:
  1306. case TIFF_PHOTOMETRIC_YCBCR:
  1307. case TIFF_PHOTOMETRIC_CFA:
  1308. s->photometric = value;
  1309. break;
  1310. case TIFF_PHOTOMETRIC_ALPHA_MASK:
  1311. case TIFF_PHOTOMETRIC_CIE_LAB:
  1312. case TIFF_PHOTOMETRIC_ICC_LAB:
  1313. case TIFF_PHOTOMETRIC_ITU_LAB:
  1314. case TIFF_PHOTOMETRIC_LOG_L:
  1315. case TIFF_PHOTOMETRIC_LOG_LUV:
  1316. case TIFF_PHOTOMETRIC_LINEAR_RAW:
  1317. avpriv_report_missing_feature(s->avctx,
  1318. "PhotometricInterpretation 0x%04X",
  1319. value);
  1320. return AVERROR_PATCHWELCOME;
  1321. default:
  1322. av_log(s->avctx, AV_LOG_ERROR, "PhotometricInterpretation %u is "
  1323. "unknown\n", value);
  1324. return AVERROR_INVALIDDATA;
  1325. }
  1326. break;
  1327. case TIFF_FILL_ORDER:
  1328. if (value < 1 || value > 2) {
  1329. av_log(s->avctx, AV_LOG_ERROR,
  1330. "Unknown FillOrder value %d, trying default one\n", value);
  1331. value = 1;
  1332. }
  1333. s->fill_order = value - 1;
  1334. break;
  1335. case TIFF_PAL: {
  1336. GetByteContext pal_gb[3];
  1337. off = type_sizes[type];
  1338. if (count / 3 > 256 ||
  1339. bytestream2_get_bytes_left(&s->gb) < count / 3 * off * 3)
  1340. return AVERROR_INVALIDDATA;
  1341. pal_gb[0] = pal_gb[1] = pal_gb[2] = s->gb;
  1342. bytestream2_skip(&pal_gb[1], count / 3 * off);
  1343. bytestream2_skip(&pal_gb[2], count / 3 * off * 2);
  1344. off = (type_sizes[type] - 1) << 3;
  1345. if (off > 31U) {
  1346. av_log(s->avctx, AV_LOG_ERROR, "palette shift %d is out of range\n", off);
  1347. return AVERROR_INVALIDDATA;
  1348. }
  1349. for (i = 0; i < count / 3; i++) {
  1350. uint32_t p = 0xFF000000;
  1351. p |= (ff_tget(&pal_gb[0], type, s->le) >> off) << 16;
  1352. p |= (ff_tget(&pal_gb[1], type, s->le) >> off) << 8;
  1353. p |= ff_tget(&pal_gb[2], type, s->le) >> off;
  1354. s->palette[i] = p;
  1355. }
  1356. s->palette_is_set = 1;
  1357. break;
  1358. }
  1359. case TIFF_PLANAR:
  1360. s->planar = value == 2;
  1361. break;
  1362. case TIFF_YCBCR_SUBSAMPLING:
  1363. if (count != 2) {
  1364. av_log(s->avctx, AV_LOG_ERROR, "subsample count invalid\n");
  1365. return AVERROR_INVALIDDATA;
  1366. }
  1367. for (i = 0; i < count; i++) {
  1368. s->subsampling[i] = ff_tget(&s->gb, type, s->le);
  1369. if (s->subsampling[i] <= 0) {
  1370. av_log(s->avctx, AV_LOG_ERROR, "subsampling %d is invalid\n", s->subsampling[i]);
  1371. s->subsampling[i] = 1;
  1372. return AVERROR_INVALIDDATA;
  1373. }
  1374. }
  1375. break;
  1376. case TIFF_T4OPTIONS:
  1377. if (s->compr == TIFF_G3)
  1378. s->fax_opts = value;
  1379. break;
  1380. case TIFF_T6OPTIONS:
  1381. if (s->compr == TIFF_G4)
  1382. s->fax_opts = value;
  1383. break;
  1384. #define ADD_METADATA(count, name, sep)\
  1385. if ((ret = add_metadata(count, type, name, sep, s, frame)) < 0) {\
  1386. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");\
  1387. goto end;\
  1388. }
  1389. case TIFF_MODEL_PIXEL_SCALE:
  1390. ADD_METADATA(count, "ModelPixelScaleTag", NULL);
  1391. break;
  1392. case TIFF_MODEL_TRANSFORMATION:
  1393. ADD_METADATA(count, "ModelTransformationTag", NULL);
  1394. break;
  1395. case TIFF_MODEL_TIEPOINT:
  1396. ADD_METADATA(count, "ModelTiepointTag", NULL);
  1397. break;
  1398. case TIFF_GEO_KEY_DIRECTORY:
  1399. if (s->geotag_count) {
  1400. avpriv_request_sample(s->avctx, "Multiple geo key directories\n");
  1401. return AVERROR_INVALIDDATA;
  1402. }
  1403. ADD_METADATA(1, "GeoTIFF_Version", NULL);
  1404. ADD_METADATA(2, "GeoTIFF_Key_Revision", ".");
  1405. s->geotag_count = ff_tget_short(&s->gb, s->le);
  1406. if (s->geotag_count > count / 4 - 1) {
  1407. s->geotag_count = count / 4 - 1;
  1408. av_log(s->avctx, AV_LOG_WARNING, "GeoTIFF key directory buffer shorter than specified\n");
  1409. }
  1410. if ( bytestream2_get_bytes_left(&s->gb) < s->geotag_count * sizeof(int16_t) * 4
  1411. || s->geotag_count == 0) {
  1412. s->geotag_count = 0;
  1413. return -1;
  1414. }
  1415. s->geotags = av_mallocz_array(s->geotag_count, sizeof(TiffGeoTag));
  1416. if (!s->geotags) {
  1417. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1418. s->geotag_count = 0;
  1419. goto end;
  1420. }
  1421. for (i = 0; i < s->geotag_count; i++) {
  1422. s->geotags[i].key = ff_tget_short(&s->gb, s->le);
  1423. s->geotags[i].type = ff_tget_short(&s->gb, s->le);
  1424. s->geotags[i].count = ff_tget_short(&s->gb, s->le);
  1425. if (!s->geotags[i].type)
  1426. s->geotags[i].val = get_geokey_val(s->geotags[i].key, ff_tget_short(&s->gb, s->le));
  1427. else
  1428. s->geotags[i].offset = ff_tget_short(&s->gb, s->le);
  1429. }
  1430. break;
  1431. case TIFF_GEO_DOUBLE_PARAMS:
  1432. if (count >= INT_MAX / sizeof(int64_t))
  1433. return AVERROR_INVALIDDATA;
  1434. if (bytestream2_get_bytes_left(&s->gb) < count * sizeof(int64_t))
  1435. return AVERROR_INVALIDDATA;
  1436. dp = av_malloc_array(count, sizeof(double));
  1437. if (!dp) {
  1438. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1439. goto end;
  1440. }
  1441. for (i = 0; i < count; i++)
  1442. dp[i] = ff_tget_double(&s->gb, s->le);
  1443. for (i = 0; i < s->geotag_count; i++) {
  1444. if (s->geotags[i].type == TIFF_GEO_DOUBLE_PARAMS) {
  1445. if (s->geotags[i].count == 0
  1446. || s->geotags[i].offset + s->geotags[i].count > count) {
  1447. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1448. } else if (s->geotags[i].val) {
  1449. av_log(s->avctx, AV_LOG_WARNING, "Duplicate GeoTIFF key %d\n", s->geotags[i].key);
  1450. } else {
  1451. char *ap = doubles2str(&dp[s->geotags[i].offset], s->geotags[i].count, ", ");
  1452. if (!ap) {
  1453. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1454. av_freep(&dp);
  1455. return AVERROR(ENOMEM);
  1456. }
  1457. s->geotags[i].val = ap;
  1458. }
  1459. }
  1460. }
  1461. av_freep(&dp);
  1462. break;
  1463. case TIFF_GEO_ASCII_PARAMS:
  1464. pos = bytestream2_tell(&s->gb);
  1465. for (i = 0; i < s->geotag_count; i++) {
  1466. if (s->geotags[i].type == TIFF_GEO_ASCII_PARAMS) {
  1467. if (s->geotags[i].count == 0
  1468. || s->geotags[i].offset + s->geotags[i].count > count) {
  1469. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1470. } else {
  1471. char *ap;
  1472. bytestream2_seek(&s->gb, pos + s->geotags[i].offset, SEEK_SET);
  1473. if (bytestream2_get_bytes_left(&s->gb) < s->geotags[i].count)
  1474. return AVERROR_INVALIDDATA;
  1475. if (s->geotags[i].val)
  1476. return AVERROR_INVALIDDATA;
  1477. ap = av_malloc(s->geotags[i].count);
  1478. if (!ap) {
  1479. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1480. return AVERROR(ENOMEM);
  1481. }
  1482. bytestream2_get_bufferu(&s->gb, ap, s->geotags[i].count);
  1483. ap[s->geotags[i].count - 1] = '\0'; //replace the "|" delimiter with a 0 byte
  1484. s->geotags[i].val = ap;
  1485. }
  1486. }
  1487. }
  1488. break;
  1489. case TIFF_ARTIST:
  1490. ADD_METADATA(count, "artist", NULL);
  1491. break;
  1492. case TIFF_COPYRIGHT:
  1493. ADD_METADATA(count, "copyright", NULL);
  1494. break;
  1495. case TIFF_DATE:
  1496. ADD_METADATA(count, "date", NULL);
  1497. break;
  1498. case TIFF_DOCUMENT_NAME:
  1499. ADD_METADATA(count, "document_name", NULL);
  1500. break;
  1501. case TIFF_HOST_COMPUTER:
  1502. ADD_METADATA(count, "computer", NULL);
  1503. break;
  1504. case TIFF_IMAGE_DESCRIPTION:
  1505. ADD_METADATA(count, "description", NULL);
  1506. break;
  1507. case TIFF_MAKE:
  1508. ADD_METADATA(count, "make", NULL);
  1509. break;
  1510. case TIFF_MODEL:
  1511. ADD_METADATA(count, "model", NULL);
  1512. break;
  1513. case TIFF_PAGE_NAME:
  1514. ADD_METADATA(count, "page_name", NULL);
  1515. break;
  1516. case TIFF_PAGE_NUMBER:
  1517. ADD_METADATA(count, "page_number", " / ");
  1518. // need to seek back to re-read the page number
  1519. bytestream2_seek(&s->gb, -count * sizeof(uint16_t), SEEK_CUR);
  1520. // read the page number
  1521. s->cur_page = ff_tget(&s->gb, TIFF_SHORT, s->le);
  1522. // get back to where we were before the previous seek
  1523. bytestream2_seek(&s->gb, count * sizeof(uint16_t) - sizeof(uint16_t), SEEK_CUR);
  1524. break;
  1525. case TIFF_SOFTWARE_NAME:
  1526. ADD_METADATA(count, "software", NULL);
  1527. break;
  1528. case DNG_VERSION:
  1529. if (count == 4) {
  1530. unsigned int ver[4];
  1531. ver[0] = ff_tget(&s->gb, type, s->le);
  1532. ver[1] = ff_tget(&s->gb, type, s->le);
  1533. ver[2] = ff_tget(&s->gb, type, s->le);
  1534. ver[3] = ff_tget(&s->gb, type, s->le);
  1535. av_log(s->avctx, AV_LOG_DEBUG, "DNG file, version %u.%u.%u.%u\n",
  1536. ver[0], ver[1], ver[2], ver[3]);
  1537. tiff_set_type(s, TIFF_TYPE_DNG);
  1538. }
  1539. break;
  1540. case CINEMADNG_TIME_CODES:
  1541. case CINEMADNG_FRAME_RATE:
  1542. case CINEMADNG_T_STOP:
  1543. case CINEMADNG_REEL_NAME:
  1544. case CINEMADNG_CAMERA_LABEL:
  1545. tiff_set_type(s, TIFF_TYPE_CINEMADNG);
  1546. break;
  1547. default:
  1548. if (s->avctx->err_recognition & AV_EF_EXPLODE) {
  1549. av_log(s->avctx, AV_LOG_ERROR,
  1550. "Unknown or unsupported tag %d/0x%0X\n",
  1551. tag, tag);
  1552. return AVERROR_INVALIDDATA;
  1553. }
  1554. }
  1555. end:
  1556. if (s->bpp > 64U) {
  1557. av_log(s->avctx, AV_LOG_ERROR,
  1558. "This format is not supported (bpp=%d, %d components)\n",
  1559. s->bpp, count);
  1560. s->bpp = 0;
  1561. return AVERROR_INVALIDDATA;
  1562. }
  1563. bytestream2_seek(&s->gb, start, SEEK_SET);
  1564. return 0;
  1565. }
  1566. static int decode_frame(AVCodecContext *avctx,
  1567. void *data, int *got_frame, AVPacket *avpkt)
  1568. {
  1569. TiffContext *const s = avctx->priv_data;
  1570. AVFrame *const p = data;
  1571. ThreadFrame frame = { .f = data };
  1572. unsigned off, last_off;
  1573. int le, ret, plane, planes;
  1574. int i, j, entries, stride;
  1575. unsigned soff, ssize;
  1576. uint8_t *dst;
  1577. GetByteContext stripsizes;
  1578. GetByteContext stripdata;
  1579. int retry_for_subifd, retry_for_page;
  1580. bytestream2_init(&s->gb, avpkt->data, avpkt->size);
  1581. // parse image header
  1582. if ((ret = ff_tdecode_header(&s->gb, &le, &off))) {
  1583. av_log(avctx, AV_LOG_ERROR, "Invalid TIFF header\n");
  1584. return ret;
  1585. } else if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1586. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1587. return AVERROR_INVALIDDATA;
  1588. }
  1589. s->le = le;
  1590. // TIFF_BPP is not a required tag and defaults to 1
  1591. s->tiff_type = TIFF_TYPE_TIFF;
  1592. again:
  1593. s->is_thumbnail = 0;
  1594. s->bppcount = s->bpp = 1;
  1595. s->photometric = TIFF_PHOTOMETRIC_NONE;
  1596. s->compr = TIFF_RAW;
  1597. s->fill_order = 0;
  1598. s->white_level = 0;
  1599. s->is_bayer = 0;
  1600. s->is_tiled = 0;
  1601. s->is_jpeg = 0;
  1602. s->cur_page = 0;
  1603. s->dng_lut = NULL;
  1604. free_geotags(s);
  1605. // Reset these offsets so we can tell if they were set this frame
  1606. s->stripsizesoff = s->strippos = 0;
  1607. /* parse image file directory */
  1608. bytestream2_seek(&s->gb, off, SEEK_SET);
  1609. entries = ff_tget_short(&s->gb, le);
  1610. if (bytestream2_get_bytes_left(&s->gb) < entries * 12)
  1611. return AVERROR_INVALIDDATA;
  1612. for (i = 0; i < entries; i++) {
  1613. if ((ret = tiff_decode_tag(s, p)) < 0)
  1614. return ret;
  1615. }
  1616. if (s->get_thumbnail && !s->is_thumbnail) {
  1617. av_log(avctx, AV_LOG_INFO, "No embedded thumbnail present\n");
  1618. return AVERROR_EOF;
  1619. }
  1620. /** whether we should process this IFD's SubIFD */
  1621. retry_for_subifd = s->sub_ifd && (s->get_subimage || (!s->get_thumbnail && s->is_thumbnail));
  1622. /** whether we should process this multi-page IFD's next page */
  1623. retry_for_page = s->get_page && s->cur_page + 1 < s->get_page; // get_page is 1-indexed
  1624. last_off = off;
  1625. if (retry_for_page) {
  1626. // set offset to the next IFD
  1627. off = ff_tget_long(&s->gb, le);
  1628. } else if (retry_for_subifd) {
  1629. // set offset to the SubIFD
  1630. off = s->sub_ifd;
  1631. }
  1632. if (retry_for_subifd || retry_for_page) {
  1633. if (!off) {
  1634. av_log(avctx, AV_LOG_ERROR, "Requested entry not found\n");
  1635. return AVERROR_INVALIDDATA;
  1636. }
  1637. if (off <= last_off) {
  1638. avpriv_request_sample(s->avctx, "non increasing IFD offset\n");
  1639. return AVERROR_INVALIDDATA;
  1640. }
  1641. if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1642. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1643. return AVERROR_INVALIDDATA;
  1644. }
  1645. s->sub_ifd = 0;
  1646. goto again;
  1647. }
  1648. for (i = 0; i<s->geotag_count; i++) {
  1649. const char *keyname = get_geokey_name(s->geotags[i].key);
  1650. if (!keyname) {
  1651. av_log(avctx, AV_LOG_WARNING, "Unknown or unsupported GeoTIFF key %d\n", s->geotags[i].key);
  1652. continue;
  1653. }
  1654. if (get_geokey_type(s->geotags[i].key) != s->geotags[i].type) {
  1655. av_log(avctx, AV_LOG_WARNING, "Type of GeoTIFF key %d is wrong\n", s->geotags[i].key);
  1656. continue;
  1657. }
  1658. ret = av_dict_set(&p->metadata, keyname, s->geotags[i].val, 0);
  1659. if (ret<0) {
  1660. av_log(avctx, AV_LOG_ERROR, "Writing metadata with key '%s' failed\n", keyname);
  1661. return ret;
  1662. }
  1663. }
  1664. if (!s->is_tiled && !s->strippos && !s->stripoff) {
  1665. av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
  1666. return AVERROR_INVALIDDATA;
  1667. }
  1668. /* now we have the data and may start decoding */
  1669. if ((ret = init_image(s, &frame)) < 0)
  1670. return ret;
  1671. if (!s->is_tiled) {
  1672. if (s->strips == 1 && !s->stripsize) {
  1673. av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
  1674. s->stripsize = avpkt->size - s->stripoff;
  1675. }
  1676. if (s->stripsizesoff) {
  1677. if (s->stripsizesoff >= (unsigned)avpkt->size)
  1678. return AVERROR_INVALIDDATA;
  1679. bytestream2_init(&stripsizes, avpkt->data + s->stripsizesoff,
  1680. avpkt->size - s->stripsizesoff);
  1681. }
  1682. if (s->strippos) {
  1683. if (s->strippos >= (unsigned)avpkt->size)
  1684. return AVERROR_INVALIDDATA;
  1685. bytestream2_init(&stripdata, avpkt->data + s->strippos,
  1686. avpkt->size - s->strippos);
  1687. }
  1688. if (s->rps <= 0 || s->rps % s->subsampling[1]) {
  1689. av_log(avctx, AV_LOG_ERROR, "rps %d invalid\n", s->rps);
  1690. return AVERROR_INVALIDDATA;
  1691. }
  1692. }
  1693. /* Handle DNG images with JPEG-compressed tiles */
  1694. if ((s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG) && s->is_tiled) {
  1695. if (!s->is_jpeg) {
  1696. avpriv_report_missing_feature(avctx, "DNG uncompressed tiled images");
  1697. return AVERROR_PATCHWELCOME;
  1698. } else if (!s->is_bayer) {
  1699. avpriv_report_missing_feature(avctx, "DNG JPG-compressed tiled non-bayer-encoded images");
  1700. return AVERROR_PATCHWELCOME;
  1701. } else {
  1702. if ((ret = dng_decode(avctx, (AVFrame*)data, avpkt)) > 0)
  1703. *got_frame = 1;
  1704. return ret;
  1705. }
  1706. }
  1707. /* Handle TIFF images and DNG images with uncompressed strips (non-tiled) */
  1708. planes = s->planar ? s->bppcount : 1;
  1709. for (plane = 0; plane < planes; plane++) {
  1710. uint8_t *five_planes = NULL;
  1711. int remaining = avpkt->size;
  1712. int decoded_height;
  1713. stride = p->linesize[plane];
  1714. dst = p->data[plane];
  1715. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1716. s->avctx->pix_fmt == AV_PIX_FMT_RGBA) {
  1717. stride = stride * 5 / 4;
  1718. five_planes =
  1719. dst = av_malloc(stride * s->height);
  1720. if (!dst)
  1721. return AVERROR(ENOMEM);
  1722. }
  1723. for (i = 0; i < s->height; i += s->rps) {
  1724. if (i)
  1725. dst += s->rps * stride;
  1726. if (s->stripsizesoff)
  1727. ssize = ff_tget(&stripsizes, s->sstype, le);
  1728. else
  1729. ssize = s->stripsize;
  1730. if (s->strippos)
  1731. soff = ff_tget(&stripdata, s->sot, le);
  1732. else
  1733. soff = s->stripoff;
  1734. if (soff > avpkt->size || ssize > avpkt->size - soff || ssize > remaining) {
  1735. av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
  1736. av_freep(&five_planes);
  1737. return AVERROR_INVALIDDATA;
  1738. }
  1739. remaining -= ssize;
  1740. if ((ret = tiff_unpack_strip(s, p, dst, stride, avpkt->data + soff, ssize, i,
  1741. FFMIN(s->rps, s->height - i))) < 0) {
  1742. if (avctx->err_recognition & AV_EF_EXPLODE) {
  1743. av_freep(&five_planes);
  1744. return ret;
  1745. }
  1746. break;
  1747. }
  1748. }
  1749. decoded_height = FFMIN(i, s->height);
  1750. if (s->predictor == 2) {
  1751. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1752. av_log(s->avctx, AV_LOG_ERROR, "predictor == 2 with YUV is unsupported");
  1753. return AVERROR_PATCHWELCOME;
  1754. }
  1755. dst = five_planes ? five_planes : p->data[plane];
  1756. soff = s->bpp >> 3;
  1757. if (s->planar)
  1758. soff = FFMAX(soff / s->bppcount, 1);
  1759. ssize = s->width * soff;
  1760. if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48LE ||
  1761. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64LE ||
  1762. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16LE ||
  1763. s->avctx->pix_fmt == AV_PIX_FMT_YA16LE ||
  1764. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16LE ||
  1765. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16LE) {
  1766. for (i = 0; i < decoded_height; i++) {
  1767. for (j = soff; j < ssize; j += 2)
  1768. AV_WL16(dst + j, AV_RL16(dst + j) + AV_RL16(dst + j - soff));
  1769. dst += stride;
  1770. }
  1771. } else if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48BE ||
  1772. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE ||
  1773. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16BE ||
  1774. s->avctx->pix_fmt == AV_PIX_FMT_YA16BE ||
  1775. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16BE ||
  1776. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16BE) {
  1777. for (i = 0; i < decoded_height; i++) {
  1778. for (j = soff; j < ssize; j += 2)
  1779. AV_WB16(dst + j, AV_RB16(dst + j) + AV_RB16(dst + j - soff));
  1780. dst += stride;
  1781. }
  1782. } else {
  1783. for (i = 0; i < decoded_height; i++) {
  1784. for (j = soff; j < ssize; j++)
  1785. dst[j] += dst[j - soff];
  1786. dst += stride;
  1787. }
  1788. }
  1789. }
  1790. if (s->photometric == TIFF_PHOTOMETRIC_WHITE_IS_ZERO) {
  1791. int c = (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255);
  1792. dst = p->data[plane];
  1793. for (i = 0; i < s->height; i++) {
  1794. for (j = 0; j < stride; j++)
  1795. dst[j] = c - dst[j];
  1796. dst += stride;
  1797. }
  1798. }
  1799. if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1800. (s->avctx->pix_fmt == AV_PIX_FMT_RGB0 || s->avctx->pix_fmt == AV_PIX_FMT_RGBA)) {
  1801. int x = s->avctx->pix_fmt == AV_PIX_FMT_RGB0 ? 4 : 5;
  1802. uint8_t *src = five_planes ? five_planes : p->data[plane];
  1803. dst = p->data[plane];
  1804. for (i = 0; i < s->height; i++) {
  1805. for (j = 0; j < s->width; j++) {
  1806. int k = 255 - src[x * j + 3];
  1807. int r = (255 - src[x * j ]) * k;
  1808. int g = (255 - src[x * j + 1]) * k;
  1809. int b = (255 - src[x * j + 2]) * k;
  1810. dst[4 * j ] = r * 257 >> 16;
  1811. dst[4 * j + 1] = g * 257 >> 16;
  1812. dst[4 * j + 2] = b * 257 >> 16;
  1813. dst[4 * j + 3] = s->avctx->pix_fmt == AV_PIX_FMT_RGBA ? src[x * j + 4] : 255;
  1814. }
  1815. src += stride;
  1816. dst += p->linesize[plane];
  1817. }
  1818. av_freep(&five_planes);
  1819. } else if (s->photometric == TIFF_PHOTOMETRIC_SEPARATED &&
  1820. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE) {
  1821. dst = p->data[plane];
  1822. for (i = 0; i < s->height; i++) {
  1823. for (j = 0; j < s->width; j++) {
  1824. uint64_t k = 65535 - AV_RB16(dst + 8 * j + 6);
  1825. uint64_t r = (65535 - AV_RB16(dst + 8 * j )) * k;
  1826. uint64_t g = (65535 - AV_RB16(dst + 8 * j + 2)) * k;
  1827. uint64_t b = (65535 - AV_RB16(dst + 8 * j + 4)) * k;
  1828. AV_WB16(dst + 8 * j , r * 65537 >> 32);
  1829. AV_WB16(dst + 8 * j + 2, g * 65537 >> 32);
  1830. AV_WB16(dst + 8 * j + 4, b * 65537 >> 32);
  1831. AV_WB16(dst + 8 * j + 6, 65535);
  1832. }
  1833. dst += p->linesize[plane];
  1834. }
  1835. }
  1836. }
  1837. if (s->planar && s->bppcount > 2) {
  1838. FFSWAP(uint8_t*, p->data[0], p->data[2]);
  1839. FFSWAP(int, p->linesize[0], p->linesize[2]);
  1840. FFSWAP(uint8_t*, p->data[0], p->data[1]);
  1841. FFSWAP(int, p->linesize[0], p->linesize[1]);
  1842. }
  1843. if (s->is_bayer && s->white_level && s->bpp == 16 &&
  1844. !(s->tiff_type == TIFF_TYPE_DNG || s->tiff_type == TIFF_TYPE_CINEMADNG)) {
  1845. uint16_t *dst = (uint16_t *)p->data[0];
  1846. for (i = 0; i < s->height; i++) {
  1847. for (j = 0; j < s->width; j++)
  1848. dst[j] = FFMIN((dst[j] / (float)s->white_level) * 65535, 65535);
  1849. dst += stride / 2;
  1850. }
  1851. }
  1852. *got_frame = 1;
  1853. return avpkt->size;
  1854. }
  1855. static av_cold int tiff_init(AVCodecContext *avctx)
  1856. {
  1857. TiffContext *s = avctx->priv_data;
  1858. const AVCodec *codec;
  1859. int ret;
  1860. s->width = 0;
  1861. s->height = 0;
  1862. s->subsampling[0] =
  1863. s->subsampling[1] = 1;
  1864. s->avctx = avctx;
  1865. ff_lzw_decode_open(&s->lzw);
  1866. if (!s->lzw)
  1867. return AVERROR(ENOMEM);
  1868. ff_ccitt_unpack_init();
  1869. /* Allocate JPEG frame */
  1870. s->jpgframe = av_frame_alloc();
  1871. if (!s->jpgframe)
  1872. return AVERROR(ENOMEM);
  1873. /* Prepare everything needed for JPEG decoding */
  1874. codec = avcodec_find_decoder(AV_CODEC_ID_MJPEG);
  1875. if (!codec)
  1876. return AVERROR_BUG;
  1877. s->avctx_mjpeg = avcodec_alloc_context3(codec);
  1878. if (!s->avctx_mjpeg)
  1879. return AVERROR(ENOMEM);
  1880. s->avctx_mjpeg->flags = avctx->flags;
  1881. s->avctx_mjpeg->flags2 = avctx->flags2;
  1882. s->avctx_mjpeg->dct_algo = avctx->dct_algo;
  1883. s->avctx_mjpeg->idct_algo = avctx->idct_algo;
  1884. ret = ff_codec_open2_recursive(s->avctx_mjpeg, codec, NULL);
  1885. if (ret < 0) {
  1886. av_frame_free(&s->jpgframe);
  1887. avcodec_free_context(&s->avctx_mjpeg);
  1888. return ret;
  1889. }
  1890. return 0;
  1891. }
  1892. static av_cold int tiff_end(AVCodecContext *avctx)
  1893. {
  1894. TiffContext *const s = avctx->priv_data;
  1895. free_geotags(s);
  1896. ff_lzw_decode_close(&s->lzw);
  1897. av_freep(&s->deinvert_buf);
  1898. s->deinvert_buf_size = 0;
  1899. av_freep(&s->yuv_line);
  1900. s->yuv_line_size = 0;
  1901. av_freep(&s->fax_buffer);
  1902. s->fax_buffer_size = 0;
  1903. av_frame_free(&s->jpgframe);
  1904. avcodec_free_context(&s->avctx_mjpeg);
  1905. return 0;
  1906. }
  1907. #define OFFSET(x) offsetof(TiffContext, x)
  1908. static const AVOption tiff_options[] = {
  1909. { "subimage", "decode subimage instead if available", OFFSET(get_subimage), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  1910. { "thumbnail", "decode embedded thumbnail subimage instead if available", OFFSET(get_thumbnail), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  1911. { "page", "page number of multi-page image to decode (starting from 1)", OFFSET(get_page), AV_OPT_TYPE_INT, {.i64=0}, 0, UINT16_MAX, AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM },
  1912. { NULL },
  1913. };
  1914. static const AVClass tiff_decoder_class = {
  1915. .class_name = "TIFF decoder",
  1916. .item_name = av_default_item_name,
  1917. .option = tiff_options,
  1918. .version = LIBAVUTIL_VERSION_INT,
  1919. };
  1920. AVCodec ff_tiff_decoder = {
  1921. .name = "tiff",
  1922. .long_name = NULL_IF_CONFIG_SMALL("TIFF image"),
  1923. .type = AVMEDIA_TYPE_VIDEO,
  1924. .id = AV_CODEC_ID_TIFF,
  1925. .priv_data_size = sizeof(TiffContext),
  1926. .init = tiff_init,
  1927. .close = tiff_end,
  1928. .decode = decode_frame,
  1929. .init_thread_copy = ONLY_IF_THREADS_ENABLED(tiff_init),
  1930. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
  1931. .priv_class = &tiff_decoder_class,
  1932. };