You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1389 lines
46KB

  1. /*
  2. * Copyright (c) 2006 Konstantin Shishkov
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * TIFF image decoder
  23. * @author Konstantin Shishkov
  24. */
  25. #include "config.h"
  26. #if CONFIG_ZLIB
  27. #include <zlib.h>
  28. #endif
  29. #if CONFIG_LZMA
  30. #define LZMA_API_STATIC
  31. #include <lzma.h>
  32. #endif
  33. #include "libavutil/attributes.h"
  34. #include "libavutil/avstring.h"
  35. #include "libavutil/intreadwrite.h"
  36. #include "libavutil/imgutils.h"
  37. #include "avcodec.h"
  38. #include "bytestream.h"
  39. #include "faxcompr.h"
  40. #include "internal.h"
  41. #include "lzw.h"
  42. #include "mathops.h"
  43. #include "tiff.h"
  44. #include "tiff_data.h"
  45. #include "thread.h"
  46. typedef struct TiffContext {
  47. AVCodecContext *avctx;
  48. GetByteContext gb;
  49. int width, height;
  50. unsigned int bpp, bppcount;
  51. uint32_t palette[256];
  52. int palette_is_set;
  53. int le;
  54. enum TiffCompr compr;
  55. enum TiffPhotometric photometric;
  56. int planar;
  57. int subsampling[2];
  58. int fax_opts;
  59. int predictor;
  60. int fill_order;
  61. uint32_t res[4];
  62. int strips, rps, sstype;
  63. int sot;
  64. int stripsizesoff, stripsize, stripoff, strippos;
  65. LZWState *lzw;
  66. uint8_t *deinvert_buf;
  67. int deinvert_buf_size;
  68. uint8_t *yuv_line;
  69. unsigned int yuv_line_size;
  70. int geotag_count;
  71. TiffGeoTag *geotags;
  72. } TiffContext;
  73. static void free_geotags(TiffContext *const s)
  74. {
  75. int i;
  76. for (i = 0; i < s->geotag_count; i++) {
  77. if (s->geotags[i].val)
  78. av_freep(&s->geotags[i].val);
  79. }
  80. av_freep(&s->geotags);
  81. s->geotag_count = 0;
  82. }
  83. #define RET_GEOKEY(TYPE, array, element)\
  84. if (key >= TIFF_##TYPE##_KEY_ID_OFFSET &&\
  85. key - TIFF_##TYPE##_KEY_ID_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_name_type_map))\
  86. return ff_tiff_##array##_name_type_map[key - TIFF_##TYPE##_KEY_ID_OFFSET].element;
  87. static const char *get_geokey_name(int key)
  88. {
  89. RET_GEOKEY(VERT, vert, name);
  90. RET_GEOKEY(PROJ, proj, name);
  91. RET_GEOKEY(GEOG, geog, name);
  92. RET_GEOKEY(CONF, conf, name);
  93. return NULL;
  94. }
  95. static int get_geokey_type(int key)
  96. {
  97. RET_GEOKEY(VERT, vert, type);
  98. RET_GEOKEY(PROJ, proj, type);
  99. RET_GEOKEY(GEOG, geog, type);
  100. RET_GEOKEY(CONF, conf, type);
  101. return AVERROR_INVALIDDATA;
  102. }
  103. static int cmp_id_key(const void *id, const void *k)
  104. {
  105. return *(const int*)id - ((const TiffGeoTagKeyName*)k)->key;
  106. }
  107. static const char *search_keyval(const TiffGeoTagKeyName *keys, int n, int id)
  108. {
  109. TiffGeoTagKeyName *r = bsearch(&id, keys, n, sizeof(keys[0]), cmp_id_key);
  110. if(r)
  111. return r->name;
  112. return NULL;
  113. }
  114. static char *get_geokey_val(int key, int val)
  115. {
  116. char *ap;
  117. if (val == TIFF_GEO_KEY_UNDEFINED)
  118. return av_strdup("undefined");
  119. if (val == TIFF_GEO_KEY_USER_DEFINED)
  120. return av_strdup("User-Defined");
  121. #define RET_GEOKEY_VAL(TYPE, array)\
  122. if (val >= TIFF_##TYPE##_OFFSET &&\
  123. val - TIFF_##TYPE##_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_codes))\
  124. return av_strdup(ff_tiff_##array##_codes[val - TIFF_##TYPE##_OFFSET]);
  125. switch (key) {
  126. case TIFF_GT_MODEL_TYPE_GEOKEY:
  127. RET_GEOKEY_VAL(GT_MODEL_TYPE, gt_model_type);
  128. break;
  129. case TIFF_GT_RASTER_TYPE_GEOKEY:
  130. RET_GEOKEY_VAL(GT_RASTER_TYPE, gt_raster_type);
  131. break;
  132. case TIFF_GEOG_LINEAR_UNITS_GEOKEY:
  133. case TIFF_PROJ_LINEAR_UNITS_GEOKEY:
  134. case TIFF_VERTICAL_UNITS_GEOKEY:
  135. RET_GEOKEY_VAL(LINEAR_UNIT, linear_unit);
  136. break;
  137. case TIFF_GEOG_ANGULAR_UNITS_GEOKEY:
  138. case TIFF_GEOG_AZIMUTH_UNITS_GEOKEY:
  139. RET_GEOKEY_VAL(ANGULAR_UNIT, angular_unit);
  140. break;
  141. case TIFF_GEOGRAPHIC_TYPE_GEOKEY:
  142. RET_GEOKEY_VAL(GCS_TYPE, gcs_type);
  143. RET_GEOKEY_VAL(GCSE_TYPE, gcse_type);
  144. break;
  145. case TIFF_GEOG_GEODETIC_DATUM_GEOKEY:
  146. RET_GEOKEY_VAL(GEODETIC_DATUM, geodetic_datum);
  147. RET_GEOKEY_VAL(GEODETIC_DATUM_E, geodetic_datum_e);
  148. break;
  149. case TIFF_GEOG_ELLIPSOID_GEOKEY:
  150. RET_GEOKEY_VAL(ELLIPSOID, ellipsoid);
  151. break;
  152. case TIFF_GEOG_PRIME_MERIDIAN_GEOKEY:
  153. RET_GEOKEY_VAL(PRIME_MERIDIAN, prime_meridian);
  154. break;
  155. case TIFF_PROJECTED_CS_TYPE_GEOKEY:
  156. ap = av_strdup(search_keyval(ff_tiff_proj_cs_type_codes, FF_ARRAY_ELEMS(ff_tiff_proj_cs_type_codes), val));
  157. if(ap) return ap;
  158. break;
  159. case TIFF_PROJECTION_GEOKEY:
  160. ap = av_strdup(search_keyval(ff_tiff_projection_codes, FF_ARRAY_ELEMS(ff_tiff_projection_codes), val));
  161. if(ap) return ap;
  162. break;
  163. case TIFF_PROJ_COORD_TRANS_GEOKEY:
  164. RET_GEOKEY_VAL(COORD_TRANS, coord_trans);
  165. break;
  166. case TIFF_VERTICAL_CS_TYPE_GEOKEY:
  167. RET_GEOKEY_VAL(VERT_CS, vert_cs);
  168. RET_GEOKEY_VAL(ORTHO_VERT_CS, ortho_vert_cs);
  169. break;
  170. }
  171. ap = av_malloc(14);
  172. if (ap)
  173. snprintf(ap, 14, "Unknown-%d", val);
  174. return ap;
  175. }
  176. static char *doubles2str(double *dp, int count, const char *sep)
  177. {
  178. int i;
  179. char *ap, *ap0;
  180. uint64_t component_len;
  181. if (!sep) sep = ", ";
  182. component_len = 24LL + strlen(sep);
  183. if (count >= (INT_MAX - 1)/component_len)
  184. return NULL;
  185. ap = av_malloc(component_len * count + 1);
  186. if (!ap)
  187. return NULL;
  188. ap0 = ap;
  189. ap[0] = '\0';
  190. for (i = 0; i < count; i++) {
  191. unsigned l = snprintf(ap, component_len, "%.15g%s", dp[i], sep);
  192. if(l >= component_len) {
  193. av_free(ap0);
  194. return NULL;
  195. }
  196. ap += l;
  197. }
  198. ap0[strlen(ap0) - strlen(sep)] = '\0';
  199. return ap0;
  200. }
  201. static int add_metadata(int count, int type,
  202. const char *name, const char *sep, TiffContext *s, AVFrame *frame)
  203. {
  204. switch(type) {
  205. case TIFF_DOUBLE: return ff_tadd_doubles_metadata(count, name, sep, &s->gb, s->le, avpriv_frame_get_metadatap(frame));
  206. case TIFF_SHORT : return ff_tadd_shorts_metadata(count, name, sep, &s->gb, s->le, 0, avpriv_frame_get_metadatap(frame));
  207. case TIFF_STRING: return ff_tadd_string_metadata(count, name, &s->gb, s->le, avpriv_frame_get_metadatap(frame));
  208. default : return AVERROR_INVALIDDATA;
  209. };
  210. }
  211. static void av_always_inline horizontal_fill(unsigned int bpp, uint8_t* dst,
  212. int usePtr, const uint8_t *src,
  213. uint8_t c, int width, int offset)
  214. {
  215. switch (bpp) {
  216. case 1:
  217. while (--width >= 0) {
  218. dst[(width+offset)*8+7] = (usePtr ? src[width] : c) & 0x1;
  219. dst[(width+offset)*8+6] = (usePtr ? src[width] : c) >> 1 & 0x1;
  220. dst[(width+offset)*8+5] = (usePtr ? src[width] : c) >> 2 & 0x1;
  221. dst[(width+offset)*8+4] = (usePtr ? src[width] : c) >> 3 & 0x1;
  222. dst[(width+offset)*8+3] = (usePtr ? src[width] : c) >> 4 & 0x1;
  223. dst[(width+offset)*8+2] = (usePtr ? src[width] : c) >> 5 & 0x1;
  224. dst[(width+offset)*8+1] = (usePtr ? src[width] : c) >> 6 & 0x1;
  225. dst[(width+offset)*8+0] = (usePtr ? src[width] : c) >> 7;
  226. }
  227. break;
  228. case 2:
  229. while (--width >= 0) {
  230. dst[(width+offset)*4+3] = (usePtr ? src[width] : c) & 0x3;
  231. dst[(width+offset)*4+2] = (usePtr ? src[width] : c) >> 2 & 0x3;
  232. dst[(width+offset)*4+1] = (usePtr ? src[width] : c) >> 4 & 0x3;
  233. dst[(width+offset)*4+0] = (usePtr ? src[width] : c) >> 6;
  234. }
  235. break;
  236. case 4:
  237. while (--width >= 0) {
  238. dst[(width+offset)*2+1] = (usePtr ? src[width] : c) & 0xF;
  239. dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
  240. }
  241. break;
  242. default:
  243. if (usePtr) {
  244. memcpy(dst + offset, src, width);
  245. } else {
  246. memset(dst + offset, c, width);
  247. }
  248. }
  249. }
  250. static int deinvert_buffer(TiffContext *s, const uint8_t *src, int size)
  251. {
  252. int i;
  253. av_fast_padded_malloc(&s->deinvert_buf, &s->deinvert_buf_size, size);
  254. if (!s->deinvert_buf)
  255. return AVERROR(ENOMEM);
  256. for (i = 0; i < size; i++)
  257. s->deinvert_buf[i] = ff_reverse[src[i]];
  258. return 0;
  259. }
  260. static void unpack_yuv(TiffContext *s, AVFrame *p,
  261. const uint8_t *src, int lnum)
  262. {
  263. int i, j, k;
  264. int w = (s->width - 1) / s->subsampling[0] + 1;
  265. uint8_t *pu = &p->data[1][lnum / s->subsampling[1] * p->linesize[1]];
  266. uint8_t *pv = &p->data[2][lnum / s->subsampling[1] * p->linesize[2]];
  267. if (s->width % s->subsampling[0] || s->height % s->subsampling[1]) {
  268. for (i = 0; i < w; i++) {
  269. for (j = 0; j < s->subsampling[1]; j++)
  270. for (k = 0; k < s->subsampling[0]; k++)
  271. p->data[0][FFMIN(lnum + j, s->height-1) * p->linesize[0] +
  272. FFMIN(i * s->subsampling[0] + k, s->width-1)] = *src++;
  273. *pu++ = *src++;
  274. *pv++ = *src++;
  275. }
  276. }else{
  277. for (i = 0; i < w; i++) {
  278. for (j = 0; j < s->subsampling[1]; j++)
  279. for (k = 0; k < s->subsampling[0]; k++)
  280. p->data[0][(lnum + j) * p->linesize[0] +
  281. i * s->subsampling[0] + k] = *src++;
  282. *pu++ = *src++;
  283. *pv++ = *src++;
  284. }
  285. }
  286. }
  287. #if CONFIG_ZLIB
  288. static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
  289. int size)
  290. {
  291. z_stream zstream = { 0 };
  292. int zret;
  293. zstream.next_in = (uint8_t *)src;
  294. zstream.avail_in = size;
  295. zstream.next_out = dst;
  296. zstream.avail_out = *len;
  297. zret = inflateInit(&zstream);
  298. if (zret != Z_OK) {
  299. av_log(NULL, AV_LOG_ERROR, "Inflate init error: %d\n", zret);
  300. return zret;
  301. }
  302. zret = inflate(&zstream, Z_SYNC_FLUSH);
  303. inflateEnd(&zstream);
  304. *len = zstream.total_out;
  305. return zret == Z_STREAM_END ? Z_OK : zret;
  306. }
  307. static int tiff_unpack_zlib(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  308. const uint8_t *src, int size, int width, int lines,
  309. int strip_start, int is_yuv)
  310. {
  311. uint8_t *zbuf;
  312. unsigned long outlen;
  313. int ret, line;
  314. outlen = width * lines;
  315. zbuf = av_malloc(outlen);
  316. if (!zbuf)
  317. return AVERROR(ENOMEM);
  318. if (s->fill_order) {
  319. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  320. av_free(zbuf);
  321. return ret;
  322. }
  323. src = s->deinvert_buf;
  324. }
  325. ret = tiff_uncompress(zbuf, &outlen, src, size);
  326. if (ret != Z_OK) {
  327. av_log(s->avctx, AV_LOG_ERROR,
  328. "Uncompressing failed (%lu of %lu) with error %d\n", outlen,
  329. (unsigned long)width * lines, ret);
  330. av_free(zbuf);
  331. return AVERROR_UNKNOWN;
  332. }
  333. src = zbuf;
  334. for (line = 0; line < lines; line++) {
  335. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  336. horizontal_fill(s->bpp, dst, 1, src, 0, width, 0);
  337. } else {
  338. memcpy(dst, src, width);
  339. }
  340. if (is_yuv) {
  341. unpack_yuv(s, p, dst, strip_start + line);
  342. line += s->subsampling[1] - 1;
  343. }
  344. dst += stride;
  345. src += width;
  346. }
  347. av_free(zbuf);
  348. return 0;
  349. }
  350. #endif
  351. #if CONFIG_LZMA
  352. static int tiff_uncompress_lzma(uint8_t *dst, uint64_t *len, const uint8_t *src,
  353. int size)
  354. {
  355. lzma_stream stream = LZMA_STREAM_INIT;
  356. lzma_ret ret;
  357. stream.next_in = (uint8_t *)src;
  358. stream.avail_in = size;
  359. stream.next_out = dst;
  360. stream.avail_out = *len;
  361. ret = lzma_stream_decoder(&stream, UINT64_MAX, 0);
  362. if (ret != LZMA_OK) {
  363. av_log(NULL, AV_LOG_ERROR, "LZMA init error: %d\n", ret);
  364. return ret;
  365. }
  366. ret = lzma_code(&stream, LZMA_RUN);
  367. lzma_end(&stream);
  368. *len = stream.total_out;
  369. return ret == LZMA_STREAM_END ? LZMA_OK : ret;
  370. }
  371. static int tiff_unpack_lzma(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  372. const uint8_t *src, int size, int width, int lines,
  373. int strip_start, int is_yuv)
  374. {
  375. uint64_t outlen = width * lines;
  376. int ret, line;
  377. uint8_t *buf = av_malloc(outlen);
  378. if (!buf)
  379. return AVERROR(ENOMEM);
  380. if (s->fill_order) {
  381. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  382. av_free(buf);
  383. return ret;
  384. }
  385. src = s->deinvert_buf;
  386. }
  387. ret = tiff_uncompress_lzma(buf, &outlen, src, size);
  388. if (ret != LZMA_OK) {
  389. av_log(s->avctx, AV_LOG_ERROR,
  390. "Uncompressing failed (%"PRIu64" of %"PRIu64") with error %d\n", outlen,
  391. (uint64_t)width * lines, ret);
  392. av_free(buf);
  393. return AVERROR_UNKNOWN;
  394. }
  395. src = buf;
  396. for (line = 0; line < lines; line++) {
  397. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  398. horizontal_fill(s->bpp, dst, 1, src, 0, width, 0);
  399. } else {
  400. memcpy(dst, src, width);
  401. }
  402. if (is_yuv) {
  403. unpack_yuv(s, p, dst, strip_start + line);
  404. line += s->subsampling[1] - 1;
  405. }
  406. dst += stride;
  407. src += width;
  408. }
  409. av_free(buf);
  410. return 0;
  411. }
  412. #endif
  413. static int tiff_unpack_fax(TiffContext *s, uint8_t *dst, int stride,
  414. const uint8_t *src, int size, int width, int lines)
  415. {
  416. int i, ret = 0;
  417. int line;
  418. uint8_t *src2 = av_malloc((unsigned)size +
  419. AV_INPUT_BUFFER_PADDING_SIZE);
  420. if (!src2) {
  421. av_log(s->avctx, AV_LOG_ERROR,
  422. "Error allocating temporary buffer\n");
  423. return AVERROR(ENOMEM);
  424. }
  425. if (!s->fill_order) {
  426. memcpy(src2, src, size);
  427. } else {
  428. for (i = 0; i < size; i++)
  429. src2[i] = ff_reverse[src[i]];
  430. }
  431. memset(src2 + size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
  432. ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride,
  433. s->compr, s->fax_opts);
  434. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  435. for (line = 0; line < lines; line++) {
  436. horizontal_fill(s->bpp, dst, 1, dst, 0, width, 0);
  437. dst += stride;
  438. }
  439. av_free(src2);
  440. return ret;
  441. }
  442. static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  443. const uint8_t *src, int size, int strip_start, int lines)
  444. {
  445. PutByteContext pb;
  446. int c, line, pixels, code, ret;
  447. const uint8_t *ssrc = src;
  448. int width = ((s->width * s->bpp) + 7) >> 3;
  449. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(p->format);
  450. int is_yuv = !(desc->flags & AV_PIX_FMT_FLAG_RGB) &&
  451. (desc->flags & AV_PIX_FMT_FLAG_PLANAR) &&
  452. desc->nb_components >= 3;
  453. if (s->planar)
  454. width /= s->bppcount;
  455. if (size <= 0)
  456. return AVERROR_INVALIDDATA;
  457. if (is_yuv) {
  458. int bytes_per_row = (((s->width - 1) / s->subsampling[0] + 1) * s->bpp *
  459. s->subsampling[0] * s->subsampling[1] + 7) >> 3;
  460. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row);
  461. if (s->yuv_line == NULL) {
  462. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  463. return AVERROR(ENOMEM);
  464. }
  465. dst = s->yuv_line;
  466. stride = 0;
  467. width = (s->width - 1) / s->subsampling[0] + 1;
  468. width = width * s->subsampling[0] * s->subsampling[1] + 2*width;
  469. av_assert0(width <= bytes_per_row);
  470. av_assert0(s->bpp == 24);
  471. }
  472. if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
  473. #if CONFIG_ZLIB
  474. return tiff_unpack_zlib(s, p, dst, stride, src, size, width, lines,
  475. strip_start, is_yuv);
  476. #else
  477. av_log(s->avctx, AV_LOG_ERROR,
  478. "zlib support not enabled, "
  479. "deflate compression not supported\n");
  480. return AVERROR(ENOSYS);
  481. #endif
  482. }
  483. if (s->compr == TIFF_LZMA) {
  484. #if CONFIG_LZMA
  485. return tiff_unpack_lzma(s, p, dst, stride, src, size, width, lines,
  486. strip_start, is_yuv);
  487. #else
  488. av_log(s->avctx, AV_LOG_ERROR,
  489. "LZMA support not enabled\n");
  490. return AVERROR(ENOSYS);
  491. #endif
  492. }
  493. if (s->compr == TIFF_LZW) {
  494. if (s->fill_order) {
  495. if ((ret = deinvert_buffer(s, src, size)) < 0)
  496. return ret;
  497. ssrc = src = s->deinvert_buf;
  498. }
  499. if (size > 1 && !src[0] && (src[1]&1)) {
  500. av_log(s->avctx, AV_LOG_ERROR, "Old style LZW is unsupported\n");
  501. }
  502. if ((ret = ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF)) < 0) {
  503. av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
  504. return ret;
  505. }
  506. for (line = 0; line < lines; line++) {
  507. pixels = ff_lzw_decode(s->lzw, dst, width);
  508. if (pixels < width) {
  509. av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
  510. pixels, width);
  511. return AVERROR_INVALIDDATA;
  512. }
  513. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  514. horizontal_fill(s->bpp, dst, 1, dst, 0, width, 0);
  515. if (is_yuv) {
  516. unpack_yuv(s, p, dst, strip_start + line);
  517. line += s->subsampling[1] - 1;
  518. }
  519. dst += stride;
  520. }
  521. return 0;
  522. }
  523. if (s->compr == TIFF_CCITT_RLE ||
  524. s->compr == TIFF_G3 ||
  525. s->compr == TIFF_G4) {
  526. if (is_yuv)
  527. return AVERROR_INVALIDDATA;
  528. return tiff_unpack_fax(s, dst, stride, src, size, width, lines);
  529. }
  530. bytestream2_init(&s->gb, src, size);
  531. bytestream2_init_writer(&pb, dst, is_yuv ? s->yuv_line_size : (stride * lines));
  532. for (line = 0; line < lines; line++) {
  533. if (src - ssrc > size) {
  534. av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
  535. return AVERROR_INVALIDDATA;
  536. }
  537. if (bytestream2_get_bytes_left(&s->gb) == 0 || bytestream2_get_eof(&pb))
  538. break;
  539. bytestream2_seek_p(&pb, stride * line, SEEK_SET);
  540. switch (s->compr) {
  541. case TIFF_RAW:
  542. if (ssrc + size - src < width)
  543. return AVERROR_INVALIDDATA;
  544. if (!s->fill_order) {
  545. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  546. dst, 1, src, 0, width, 0);
  547. } else {
  548. int i;
  549. for (i = 0; i < width; i++)
  550. dst[i] = ff_reverse[src[i]];
  551. }
  552. src += width;
  553. break;
  554. case TIFF_PACKBITS:
  555. for (pixels = 0; pixels < width;) {
  556. if (ssrc + size - src < 2) {
  557. av_log(s->avctx, AV_LOG_ERROR, "Read went out of bounds\n");
  558. return AVERROR_INVALIDDATA;
  559. }
  560. code = s->fill_order ? (int8_t) ff_reverse[*src++]: (int8_t) *src++;
  561. if (code >= 0) {
  562. code++;
  563. if (pixels + code > width ||
  564. ssrc + size - src < code) {
  565. av_log(s->avctx, AV_LOG_ERROR,
  566. "Copy went out of bounds\n");
  567. return AVERROR_INVALIDDATA;
  568. }
  569. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  570. dst, 1, src, 0, code, pixels);
  571. src += code;
  572. pixels += code;
  573. } else if (code != -128) { // -127..-1
  574. code = (-code) + 1;
  575. if (pixels + code > width) {
  576. av_log(s->avctx, AV_LOG_ERROR,
  577. "Run went out of bounds\n");
  578. return AVERROR_INVALIDDATA;
  579. }
  580. c = *src++;
  581. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  582. dst, 0, NULL, c, code, pixels);
  583. pixels += code;
  584. }
  585. }
  586. if (s->fill_order) {
  587. int i;
  588. for (i = 0; i < width; i++)
  589. dst[i] = ff_reverse[dst[i]];
  590. }
  591. break;
  592. }
  593. if (is_yuv) {
  594. unpack_yuv(s, p, dst, strip_start + line);
  595. line += s->subsampling[1] - 1;
  596. }
  597. dst += stride;
  598. }
  599. return 0;
  600. }
  601. static int init_image(TiffContext *s, ThreadFrame *frame)
  602. {
  603. int ret;
  604. int create_gray_palette = 0;
  605. // make sure there is no aliasing in the following switch
  606. if (s->bpp >= 100 || s->bppcount >= 10) {
  607. av_log(s->avctx, AV_LOG_ERROR,
  608. "Unsupported image parameters: bpp=%d, bppcount=%d\n",
  609. s->bpp, s->bppcount);
  610. return AVERROR_INVALIDDATA;
  611. }
  612. switch (s->planar * 1000 + s->bpp * 10 + s->bppcount) {
  613. case 11:
  614. if (!s->palette_is_set) {
  615. s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK;
  616. break;
  617. }
  618. case 21:
  619. case 41:
  620. s->avctx->pix_fmt = AV_PIX_FMT_PAL8;
  621. if (!s->palette_is_set) {
  622. create_gray_palette = 1;
  623. }
  624. break;
  625. case 81:
  626. s->avctx->pix_fmt = s->palette_is_set ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8;
  627. break;
  628. case 243:
  629. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  630. if (s->subsampling[0] == 1 && s->subsampling[1] == 1) {
  631. s->avctx->pix_fmt = AV_PIX_FMT_YUV444P;
  632. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 1) {
  633. s->avctx->pix_fmt = AV_PIX_FMT_YUV422P;
  634. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 1) {
  635. s->avctx->pix_fmt = AV_PIX_FMT_YUV411P;
  636. } else if (s->subsampling[0] == 1 && s->subsampling[1] == 2) {
  637. s->avctx->pix_fmt = AV_PIX_FMT_YUV440P;
  638. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 2) {
  639. s->avctx->pix_fmt = AV_PIX_FMT_YUV420P;
  640. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 4) {
  641. s->avctx->pix_fmt = AV_PIX_FMT_YUV410P;
  642. } else {
  643. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr subsampling\n");
  644. return AVERROR_PATCHWELCOME;
  645. }
  646. } else
  647. s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
  648. break;
  649. case 161:
  650. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE;
  651. break;
  652. case 162:
  653. s->avctx->pix_fmt = AV_PIX_FMT_YA8;
  654. break;
  655. case 322:
  656. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_YA16LE : AV_PIX_FMT_YA16BE;
  657. break;
  658. case 324:
  659. s->avctx->pix_fmt = AV_PIX_FMT_RGBA;
  660. break;
  661. case 483:
  662. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE;
  663. break;
  664. case 644:
  665. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGBA64LE : AV_PIX_FMT_RGBA64BE;
  666. break;
  667. case 1243:
  668. s->avctx->pix_fmt = AV_PIX_FMT_GBRP;
  669. break;
  670. case 1324:
  671. s->avctx->pix_fmt = AV_PIX_FMT_GBRAP;
  672. break;
  673. case 1483:
  674. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRP16LE : AV_PIX_FMT_GBRP16BE;
  675. break;
  676. case 1644:
  677. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRAP16LE : AV_PIX_FMT_GBRAP16BE;
  678. break;
  679. default:
  680. av_log(s->avctx, AV_LOG_ERROR,
  681. "This format is not supported (bpp=%d, bppcount=%d)\n",
  682. s->bpp, s->bppcount);
  683. return AVERROR_INVALIDDATA;
  684. }
  685. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  686. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->avctx->pix_fmt);
  687. if((desc->flags & AV_PIX_FMT_FLAG_RGB) ||
  688. !(desc->flags & AV_PIX_FMT_FLAG_PLANAR) ||
  689. desc->nb_components < 3) {
  690. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr variant\n");
  691. return AVERROR_INVALIDDATA;
  692. }
  693. }
  694. if (s->width != s->avctx->width || s->height != s->avctx->height) {
  695. ret = ff_set_dimensions(s->avctx, s->width, s->height);
  696. if (ret < 0)
  697. return ret;
  698. }
  699. if ((ret = ff_thread_get_buffer(s->avctx, frame, 0)) < 0)
  700. return ret;
  701. if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  702. if (!create_gray_palette)
  703. memcpy(frame->f->data[1], s->palette, sizeof(s->palette));
  704. else {
  705. /* make default grayscale pal */
  706. int i;
  707. uint32_t *pal = (uint32_t *)frame->f->data[1];
  708. for (i = 0; i < 1<<s->bpp; i++)
  709. pal[i] = 0xFFU << 24 | i * 255 / ((1<<s->bpp) - 1) * 0x010101;
  710. }
  711. }
  712. return 0;
  713. }
  714. static void set_sar(TiffContext *s, unsigned tag, unsigned num, unsigned den)
  715. {
  716. int offset = tag == TIFF_YRES ? 2 : 0;
  717. s->res[offset++] = num;
  718. s->res[offset] = den;
  719. if (s->res[0] && s->res[1] && s->res[2] && s->res[3])
  720. av_reduce(&s->avctx->sample_aspect_ratio.num, &s->avctx->sample_aspect_ratio.den,
  721. s->res[2] * (uint64_t)s->res[1], s->res[0] * (uint64_t)s->res[3], INT32_MAX);
  722. }
  723. static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
  724. {
  725. unsigned tag, type, count, off, value = 0, value2 = 0;
  726. int i, start;
  727. int pos;
  728. int ret;
  729. double *dp;
  730. ret = ff_tread_tag(&s->gb, s->le, &tag, &type, &count, &start);
  731. if (ret < 0) {
  732. goto end;
  733. }
  734. off = bytestream2_tell(&s->gb);
  735. if (count == 1) {
  736. switch (type) {
  737. case TIFF_BYTE:
  738. case TIFF_SHORT:
  739. case TIFF_LONG:
  740. value = ff_tget(&s->gb, type, s->le);
  741. break;
  742. case TIFF_RATIONAL:
  743. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  744. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  745. break;
  746. case TIFF_STRING:
  747. if (count <= 4) {
  748. break;
  749. }
  750. default:
  751. value = UINT_MAX;
  752. }
  753. }
  754. switch (tag) {
  755. case TIFF_WIDTH:
  756. s->width = value;
  757. break;
  758. case TIFF_HEIGHT:
  759. s->height = value;
  760. break;
  761. case TIFF_BPP:
  762. if (count > 4U) {
  763. av_log(s->avctx, AV_LOG_ERROR,
  764. "This format is not supported (bpp=%d, %d components)\n",
  765. value, count);
  766. return AVERROR_INVALIDDATA;
  767. }
  768. s->bppcount = count;
  769. if (count == 1)
  770. s->bpp = value;
  771. else {
  772. switch (type) {
  773. case TIFF_BYTE:
  774. case TIFF_SHORT:
  775. case TIFF_LONG:
  776. s->bpp = 0;
  777. if (bytestream2_get_bytes_left(&s->gb) < type_sizes[type] * count)
  778. return AVERROR_INVALIDDATA;
  779. for (i = 0; i < count; i++)
  780. s->bpp += ff_tget(&s->gb, type, s->le);
  781. break;
  782. default:
  783. s->bpp = -1;
  784. }
  785. }
  786. break;
  787. case TIFF_SAMPLES_PER_PIXEL:
  788. if (count != 1) {
  789. av_log(s->avctx, AV_LOG_ERROR,
  790. "Samples per pixel requires a single value, many provided\n");
  791. return AVERROR_INVALIDDATA;
  792. }
  793. if (value > 4U) {
  794. av_log(s->avctx, AV_LOG_ERROR,
  795. "Samples per pixel %d is too large\n", value);
  796. return AVERROR_INVALIDDATA;
  797. }
  798. if (s->bppcount == 1)
  799. s->bpp *= value;
  800. s->bppcount = value;
  801. break;
  802. case TIFF_COMPR:
  803. s->compr = value;
  804. s->predictor = 0;
  805. switch (s->compr) {
  806. case TIFF_RAW:
  807. case TIFF_PACKBITS:
  808. case TIFF_LZW:
  809. case TIFF_CCITT_RLE:
  810. break;
  811. case TIFF_G3:
  812. case TIFF_G4:
  813. s->fax_opts = 0;
  814. break;
  815. case TIFF_DEFLATE:
  816. case TIFF_ADOBE_DEFLATE:
  817. #if CONFIG_ZLIB
  818. break;
  819. #else
  820. av_log(s->avctx, AV_LOG_ERROR, "Deflate: ZLib not compiled in\n");
  821. return AVERROR(ENOSYS);
  822. #endif
  823. case TIFF_JPEG:
  824. case TIFF_NEWJPEG:
  825. avpriv_report_missing_feature(s->avctx, "JPEG compression");
  826. return AVERROR_PATCHWELCOME;
  827. case TIFF_LZMA:
  828. #if CONFIG_LZMA
  829. break;
  830. #else
  831. av_log(s->avctx, AV_LOG_ERROR, "LZMA not compiled in\n");
  832. return AVERROR(ENOSYS);
  833. #endif
  834. default:
  835. av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
  836. s->compr);
  837. return AVERROR_INVALIDDATA;
  838. }
  839. break;
  840. case TIFF_ROWSPERSTRIP:
  841. if (!value || (type == TIFF_LONG && value == UINT_MAX))
  842. value = s->height;
  843. s->rps = FFMIN(value, s->height);
  844. break;
  845. case TIFF_STRIP_OFFS:
  846. if (count == 1) {
  847. s->strippos = 0;
  848. s->stripoff = value;
  849. } else
  850. s->strippos = off;
  851. s->strips = count;
  852. if (s->strips == 1)
  853. s->rps = s->height;
  854. s->sot = type;
  855. break;
  856. case TIFF_STRIP_SIZE:
  857. if (count == 1) {
  858. s->stripsizesoff = 0;
  859. s->stripsize = value;
  860. s->strips = 1;
  861. } else {
  862. s->stripsizesoff = off;
  863. }
  864. s->strips = count;
  865. s->sstype = type;
  866. break;
  867. case TIFF_XRES:
  868. case TIFF_YRES:
  869. set_sar(s, tag, value, value2);
  870. break;
  871. case TIFF_TILE_BYTE_COUNTS:
  872. case TIFF_TILE_LENGTH:
  873. case TIFF_TILE_OFFSETS:
  874. case TIFF_TILE_WIDTH:
  875. av_log(s->avctx, AV_LOG_ERROR, "Tiled images are not supported\n");
  876. return AVERROR_PATCHWELCOME;
  877. break;
  878. case TIFF_PREDICTOR:
  879. s->predictor = value;
  880. break;
  881. case TIFF_PHOTOMETRIC:
  882. switch (value) {
  883. case TIFF_PHOTOMETRIC_WHITE_IS_ZERO:
  884. case TIFF_PHOTOMETRIC_BLACK_IS_ZERO:
  885. case TIFF_PHOTOMETRIC_RGB:
  886. case TIFF_PHOTOMETRIC_PALETTE:
  887. case TIFF_PHOTOMETRIC_YCBCR:
  888. s->photometric = value;
  889. break;
  890. case TIFF_PHOTOMETRIC_ALPHA_MASK:
  891. case TIFF_PHOTOMETRIC_SEPARATED:
  892. case TIFF_PHOTOMETRIC_CIE_LAB:
  893. case TIFF_PHOTOMETRIC_ICC_LAB:
  894. case TIFF_PHOTOMETRIC_ITU_LAB:
  895. case TIFF_PHOTOMETRIC_CFA:
  896. case TIFF_PHOTOMETRIC_LOG_L:
  897. case TIFF_PHOTOMETRIC_LOG_LUV:
  898. case TIFF_PHOTOMETRIC_LINEAR_RAW:
  899. avpriv_report_missing_feature(s->avctx,
  900. "PhotometricInterpretation 0x%04X",
  901. value);
  902. return AVERROR_PATCHWELCOME;
  903. default:
  904. av_log(s->avctx, AV_LOG_ERROR, "PhotometricInterpretation %u is "
  905. "unknown\n", value);
  906. return AVERROR_INVALIDDATA;
  907. }
  908. break;
  909. case TIFF_FILL_ORDER:
  910. if (value < 1 || value > 2) {
  911. av_log(s->avctx, AV_LOG_ERROR,
  912. "Unknown FillOrder value %d, trying default one\n", value);
  913. value = 1;
  914. }
  915. s->fill_order = value - 1;
  916. break;
  917. case TIFF_PAL: {
  918. GetByteContext pal_gb[3];
  919. off = type_sizes[type];
  920. if (count / 3 > 256 ||
  921. bytestream2_get_bytes_left(&s->gb) < count / 3 * off * 3)
  922. return AVERROR_INVALIDDATA;
  923. pal_gb[0] = pal_gb[1] = pal_gb[2] = s->gb;
  924. bytestream2_skip(&pal_gb[1], count / 3 * off);
  925. bytestream2_skip(&pal_gb[2], count / 3 * off * 2);
  926. off = (type_sizes[type] - 1) << 3;
  927. for (i = 0; i < count / 3; i++) {
  928. uint32_t p = 0xFF000000;
  929. p |= (ff_tget(&pal_gb[0], type, s->le) >> off) << 16;
  930. p |= (ff_tget(&pal_gb[1], type, s->le) >> off) << 8;
  931. p |= ff_tget(&pal_gb[2], type, s->le) >> off;
  932. s->palette[i] = p;
  933. }
  934. s->palette_is_set = 1;
  935. break;
  936. }
  937. case TIFF_PLANAR:
  938. s->planar = value == 2;
  939. break;
  940. case TIFF_YCBCR_SUBSAMPLING:
  941. if (count != 2) {
  942. av_log(s->avctx, AV_LOG_ERROR, "subsample count invalid\n");
  943. return AVERROR_INVALIDDATA;
  944. }
  945. for (i = 0; i < count; i++)
  946. s->subsampling[i] = ff_tget(&s->gb, type, s->le);
  947. break;
  948. case TIFF_T4OPTIONS:
  949. if (s->compr == TIFF_G3)
  950. s->fax_opts = value;
  951. break;
  952. case TIFF_T6OPTIONS:
  953. if (s->compr == TIFF_G4)
  954. s->fax_opts = value;
  955. break;
  956. #define ADD_METADATA(count, name, sep)\
  957. if ((ret = add_metadata(count, type, name, sep, s, frame)) < 0) {\
  958. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");\
  959. goto end;\
  960. }
  961. case TIFF_MODEL_PIXEL_SCALE:
  962. ADD_METADATA(count, "ModelPixelScaleTag", NULL);
  963. break;
  964. case TIFF_MODEL_TRANSFORMATION:
  965. ADD_METADATA(count, "ModelTransformationTag", NULL);
  966. break;
  967. case TIFF_MODEL_TIEPOINT:
  968. ADD_METADATA(count, "ModelTiepointTag", NULL);
  969. break;
  970. case TIFF_GEO_KEY_DIRECTORY:
  971. ADD_METADATA(1, "GeoTIFF_Version", NULL);
  972. ADD_METADATA(2, "GeoTIFF_Key_Revision", ".");
  973. s->geotag_count = ff_tget_short(&s->gb, s->le);
  974. if (s->geotag_count > count / 4 - 1) {
  975. s->geotag_count = count / 4 - 1;
  976. av_log(s->avctx, AV_LOG_WARNING, "GeoTIFF key directory buffer shorter than specified\n");
  977. }
  978. if (bytestream2_get_bytes_left(&s->gb) < s->geotag_count * sizeof(int16_t) * 4) {
  979. s->geotag_count = 0;
  980. return -1;
  981. }
  982. s->geotags = av_mallocz_array(s->geotag_count, sizeof(TiffGeoTag));
  983. if (!s->geotags) {
  984. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  985. s->geotag_count = 0;
  986. goto end;
  987. }
  988. for (i = 0; i < s->geotag_count; i++) {
  989. s->geotags[i].key = ff_tget_short(&s->gb, s->le);
  990. s->geotags[i].type = ff_tget_short(&s->gb, s->le);
  991. s->geotags[i].count = ff_tget_short(&s->gb, s->le);
  992. if (!s->geotags[i].type)
  993. s->geotags[i].val = get_geokey_val(s->geotags[i].key, ff_tget_short(&s->gb, s->le));
  994. else
  995. s->geotags[i].offset = ff_tget_short(&s->gb, s->le);
  996. }
  997. break;
  998. case TIFF_GEO_DOUBLE_PARAMS:
  999. if (count >= INT_MAX / sizeof(int64_t))
  1000. return AVERROR_INVALIDDATA;
  1001. if (bytestream2_get_bytes_left(&s->gb) < count * sizeof(int64_t))
  1002. return AVERROR_INVALIDDATA;
  1003. dp = av_malloc_array(count, sizeof(double));
  1004. if (!dp) {
  1005. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1006. goto end;
  1007. }
  1008. for (i = 0; i < count; i++)
  1009. dp[i] = ff_tget_double(&s->gb, s->le);
  1010. for (i = 0; i < s->geotag_count; i++) {
  1011. if (s->geotags[i].type == TIFF_GEO_DOUBLE_PARAMS) {
  1012. if (s->geotags[i].count == 0
  1013. || s->geotags[i].offset + s->geotags[i].count > count) {
  1014. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1015. } else {
  1016. char *ap = doubles2str(&dp[s->geotags[i].offset], s->geotags[i].count, ", ");
  1017. if (!ap) {
  1018. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1019. av_freep(&dp);
  1020. return AVERROR(ENOMEM);
  1021. }
  1022. s->geotags[i].val = ap;
  1023. }
  1024. }
  1025. }
  1026. av_freep(&dp);
  1027. break;
  1028. case TIFF_GEO_ASCII_PARAMS:
  1029. pos = bytestream2_tell(&s->gb);
  1030. for (i = 0; i < s->geotag_count; i++) {
  1031. if (s->geotags[i].type == TIFF_GEO_ASCII_PARAMS) {
  1032. if (s->geotags[i].count == 0
  1033. || s->geotags[i].offset + s->geotags[i].count > count) {
  1034. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1035. } else {
  1036. char *ap;
  1037. bytestream2_seek(&s->gb, pos + s->geotags[i].offset, SEEK_SET);
  1038. if (bytestream2_get_bytes_left(&s->gb) < s->geotags[i].count)
  1039. return AVERROR_INVALIDDATA;
  1040. ap = av_malloc(s->geotags[i].count);
  1041. if (!ap) {
  1042. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1043. return AVERROR(ENOMEM);
  1044. }
  1045. bytestream2_get_bufferu(&s->gb, ap, s->geotags[i].count);
  1046. ap[s->geotags[i].count - 1] = '\0'; //replace the "|" delimiter with a 0 byte
  1047. s->geotags[i].val = ap;
  1048. }
  1049. }
  1050. }
  1051. break;
  1052. case TIFF_ARTIST:
  1053. ADD_METADATA(count, "artist", NULL);
  1054. break;
  1055. case TIFF_COPYRIGHT:
  1056. ADD_METADATA(count, "copyright", NULL);
  1057. break;
  1058. case TIFF_DATE:
  1059. ADD_METADATA(count, "date", NULL);
  1060. break;
  1061. case TIFF_DOCUMENT_NAME:
  1062. ADD_METADATA(count, "document_name", NULL);
  1063. break;
  1064. case TIFF_HOST_COMPUTER:
  1065. ADD_METADATA(count, "computer", NULL);
  1066. break;
  1067. case TIFF_IMAGE_DESCRIPTION:
  1068. ADD_METADATA(count, "description", NULL);
  1069. break;
  1070. case TIFF_MAKE:
  1071. ADD_METADATA(count, "make", NULL);
  1072. break;
  1073. case TIFF_MODEL:
  1074. ADD_METADATA(count, "model", NULL);
  1075. break;
  1076. case TIFF_PAGE_NAME:
  1077. ADD_METADATA(count, "page_name", NULL);
  1078. break;
  1079. case TIFF_PAGE_NUMBER:
  1080. ADD_METADATA(count, "page_number", " / ");
  1081. break;
  1082. case TIFF_SOFTWARE_NAME:
  1083. ADD_METADATA(count, "software", NULL);
  1084. break;
  1085. default:
  1086. if (s->avctx->err_recognition & AV_EF_EXPLODE) {
  1087. av_log(s->avctx, AV_LOG_ERROR,
  1088. "Unknown or unsupported tag %d/0X%0X\n",
  1089. tag, tag);
  1090. return AVERROR_INVALIDDATA;
  1091. }
  1092. }
  1093. end:
  1094. if (s->bpp > 64U) {
  1095. av_log(s->avctx, AV_LOG_ERROR,
  1096. "This format is not supported (bpp=%d, %d components)\n",
  1097. s->bpp, count);
  1098. s->bpp = 0;
  1099. return AVERROR_INVALIDDATA;
  1100. }
  1101. bytestream2_seek(&s->gb, start, SEEK_SET);
  1102. return 0;
  1103. }
  1104. static int decode_frame(AVCodecContext *avctx,
  1105. void *data, int *got_frame, AVPacket *avpkt)
  1106. {
  1107. TiffContext *const s = avctx->priv_data;
  1108. AVFrame *const p = data;
  1109. ThreadFrame frame = { .f = data };
  1110. unsigned off;
  1111. int le, ret, plane, planes;
  1112. int i, j, entries, stride;
  1113. unsigned soff, ssize;
  1114. uint8_t *dst;
  1115. GetByteContext stripsizes;
  1116. GetByteContext stripdata;
  1117. bytestream2_init(&s->gb, avpkt->data, avpkt->size);
  1118. // parse image header
  1119. if ((ret = ff_tdecode_header(&s->gb, &le, &off))) {
  1120. av_log(avctx, AV_LOG_ERROR, "Invalid TIFF header\n");
  1121. return ret;
  1122. } else if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1123. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1124. return AVERROR_INVALIDDATA;
  1125. }
  1126. s->le = le;
  1127. // TIFF_BPP is not a required tag and defaults to 1
  1128. s->bppcount = s->bpp = 1;
  1129. s->photometric = TIFF_PHOTOMETRIC_NONE;
  1130. s->compr = TIFF_RAW;
  1131. s->fill_order = 0;
  1132. free_geotags(s);
  1133. // Reset these offsets so we can tell if they were set this frame
  1134. s->stripsizesoff = s->strippos = 0;
  1135. /* parse image file directory */
  1136. bytestream2_seek(&s->gb, off, SEEK_SET);
  1137. entries = ff_tget_short(&s->gb, le);
  1138. if (bytestream2_get_bytes_left(&s->gb) < entries * 12)
  1139. return AVERROR_INVALIDDATA;
  1140. for (i = 0; i < entries; i++) {
  1141. if ((ret = tiff_decode_tag(s, p)) < 0)
  1142. return ret;
  1143. }
  1144. for (i = 0; i<s->geotag_count; i++) {
  1145. const char *keyname = get_geokey_name(s->geotags[i].key);
  1146. if (!keyname) {
  1147. av_log(avctx, AV_LOG_WARNING, "Unknown or unsupported GeoTIFF key %d\n", s->geotags[i].key);
  1148. continue;
  1149. }
  1150. if (get_geokey_type(s->geotags[i].key) != s->geotags[i].type) {
  1151. av_log(avctx, AV_LOG_WARNING, "Type of GeoTIFF key %d is wrong\n", s->geotags[i].key);
  1152. continue;
  1153. }
  1154. ret = av_dict_set(avpriv_frame_get_metadatap(p), keyname, s->geotags[i].val, 0);
  1155. if (ret<0) {
  1156. av_log(avctx, AV_LOG_ERROR, "Writing metadata with key '%s' failed\n", keyname);
  1157. return ret;
  1158. }
  1159. }
  1160. if (!s->strippos && !s->stripoff) {
  1161. av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
  1162. return AVERROR_INVALIDDATA;
  1163. }
  1164. /* now we have the data and may start decoding */
  1165. if ((ret = init_image(s, &frame)) < 0)
  1166. return ret;
  1167. if (s->strips == 1 && !s->stripsize) {
  1168. av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
  1169. s->stripsize = avpkt->size - s->stripoff;
  1170. }
  1171. if (s->stripsizesoff) {
  1172. if (s->stripsizesoff >= (unsigned)avpkt->size)
  1173. return AVERROR_INVALIDDATA;
  1174. bytestream2_init(&stripsizes, avpkt->data + s->stripsizesoff,
  1175. avpkt->size - s->stripsizesoff);
  1176. }
  1177. if (s->strippos) {
  1178. if (s->strippos >= (unsigned)avpkt->size)
  1179. return AVERROR_INVALIDDATA;
  1180. bytestream2_init(&stripdata, avpkt->data + s->strippos,
  1181. avpkt->size - s->strippos);
  1182. }
  1183. if (s->rps <= 0) {
  1184. av_log(avctx, AV_LOG_ERROR, "rps %d invalid\n", s->rps);
  1185. return AVERROR_INVALIDDATA;
  1186. }
  1187. planes = s->planar ? s->bppcount : 1;
  1188. for (plane = 0; plane < planes; plane++) {
  1189. stride = p->linesize[plane];
  1190. dst = p->data[plane];
  1191. for (i = 0; i < s->height; i += s->rps) {
  1192. if (s->stripsizesoff)
  1193. ssize = ff_tget(&stripsizes, s->sstype, le);
  1194. else
  1195. ssize = s->stripsize;
  1196. if (s->strippos)
  1197. soff = ff_tget(&stripdata, s->sot, le);
  1198. else
  1199. soff = s->stripoff;
  1200. if (soff > avpkt->size || ssize > avpkt->size - soff) {
  1201. av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
  1202. return AVERROR_INVALIDDATA;
  1203. }
  1204. if ((ret = tiff_unpack_strip(s, p, dst, stride, avpkt->data + soff, ssize, i,
  1205. FFMIN(s->rps, s->height - i))) < 0) {
  1206. if (avctx->err_recognition & AV_EF_EXPLODE)
  1207. return ret;
  1208. break;
  1209. }
  1210. dst += s->rps * stride;
  1211. }
  1212. if (s->predictor == 2) {
  1213. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1214. av_log(s->avctx, AV_LOG_ERROR, "predictor == 2 with YUV is unsupported");
  1215. return AVERROR_PATCHWELCOME;
  1216. }
  1217. dst = p->data[plane];
  1218. soff = s->bpp >> 3;
  1219. if (s->planar)
  1220. soff = FFMAX(soff / s->bppcount, 1);
  1221. ssize = s->width * soff;
  1222. if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48LE ||
  1223. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64LE ||
  1224. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16LE ||
  1225. s->avctx->pix_fmt == AV_PIX_FMT_YA16LE ||
  1226. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16LE ||
  1227. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16LE) {
  1228. for (i = 0; i < s->height; i++) {
  1229. for (j = soff; j < ssize; j += 2)
  1230. AV_WL16(dst + j, AV_RL16(dst + j) + AV_RL16(dst + j - soff));
  1231. dst += stride;
  1232. }
  1233. } else if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48BE ||
  1234. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE ||
  1235. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16BE ||
  1236. s->avctx->pix_fmt == AV_PIX_FMT_YA16BE ||
  1237. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16BE ||
  1238. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16BE) {
  1239. for (i = 0; i < s->height; i++) {
  1240. for (j = soff; j < ssize; j += 2)
  1241. AV_WB16(dst + j, AV_RB16(dst + j) + AV_RB16(dst + j - soff));
  1242. dst += stride;
  1243. }
  1244. } else {
  1245. for (i = 0; i < s->height; i++) {
  1246. for (j = soff; j < ssize; j++)
  1247. dst[j] += dst[j - soff];
  1248. dst += stride;
  1249. }
  1250. }
  1251. }
  1252. if (s->photometric == TIFF_PHOTOMETRIC_WHITE_IS_ZERO) {
  1253. dst = p->data[plane];
  1254. for (i = 0; i < s->height; i++) {
  1255. for (j = 0; j < stride; j++)
  1256. dst[j] = (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255) - dst[j];
  1257. dst += stride;
  1258. }
  1259. }
  1260. }
  1261. if (s->planar && s->bppcount > 2) {
  1262. FFSWAP(uint8_t*, p->data[0], p->data[2]);
  1263. FFSWAP(int, p->linesize[0], p->linesize[2]);
  1264. FFSWAP(uint8_t*, p->data[0], p->data[1]);
  1265. FFSWAP(int, p->linesize[0], p->linesize[1]);
  1266. }
  1267. *got_frame = 1;
  1268. return avpkt->size;
  1269. }
  1270. static av_cold int tiff_init(AVCodecContext *avctx)
  1271. {
  1272. TiffContext *s = avctx->priv_data;
  1273. s->width = 0;
  1274. s->height = 0;
  1275. s->subsampling[0] =
  1276. s->subsampling[1] = 1;
  1277. s->avctx = avctx;
  1278. ff_lzw_decode_open(&s->lzw);
  1279. ff_ccitt_unpack_init();
  1280. return 0;
  1281. }
  1282. static av_cold int tiff_end(AVCodecContext *avctx)
  1283. {
  1284. TiffContext *const s = avctx->priv_data;
  1285. free_geotags(s);
  1286. ff_lzw_decode_close(&s->lzw);
  1287. av_freep(&s->deinvert_buf);
  1288. return 0;
  1289. }
  1290. AVCodec ff_tiff_decoder = {
  1291. .name = "tiff",
  1292. .long_name = NULL_IF_CONFIG_SMALL("TIFF image"),
  1293. .type = AVMEDIA_TYPE_VIDEO,
  1294. .id = AV_CODEC_ID_TIFF,
  1295. .priv_data_size = sizeof(TiffContext),
  1296. .init = tiff_init,
  1297. .close = tiff_end,
  1298. .decode = decode_frame,
  1299. .init_thread_copy = ONLY_IF_THREADS_ENABLED(tiff_init),
  1300. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
  1301. };