You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1444 lines
48KB

  1. /*
  2. * Copyright (c) 2006 Konstantin Shishkov
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * TIFF image decoder
  23. * @author Konstantin Shishkov
  24. */
  25. #include "config.h"
  26. #if CONFIG_ZLIB
  27. #include <zlib.h>
  28. #endif
  29. #if CONFIG_LZMA
  30. #define LZMA_API_STATIC
  31. #include <lzma.h>
  32. #endif
  33. #include "libavutil/attributes.h"
  34. #include "libavutil/avstring.h"
  35. #include "libavutil/intreadwrite.h"
  36. #include "libavutil/imgutils.h"
  37. #include "avcodec.h"
  38. #include "bytestream.h"
  39. #include "faxcompr.h"
  40. #include "internal.h"
  41. #include "lzw.h"
  42. #include "mathops.h"
  43. #include "tiff.h"
  44. #include "tiff_data.h"
  45. #include "thread.h"
  46. typedef struct TiffContext {
  47. AVCodecContext *avctx;
  48. GetByteContext gb;
  49. int width, height;
  50. unsigned int bpp, bppcount;
  51. uint32_t palette[256];
  52. int palette_is_set;
  53. int le;
  54. enum TiffCompr compr;
  55. enum TiffPhotometric photometric;
  56. int planar;
  57. int subsampling[2];
  58. int fax_opts;
  59. int predictor;
  60. int fill_order;
  61. uint32_t res[4];
  62. int strips, rps, sstype;
  63. int sot;
  64. int stripsizesoff, stripsize, stripoff, strippos;
  65. LZWState *lzw;
  66. uint8_t *deinvert_buf;
  67. int deinvert_buf_size;
  68. uint8_t *yuv_line;
  69. unsigned int yuv_line_size;
  70. uint8_t *fax_buffer;
  71. unsigned int fax_buffer_size;
  72. int geotag_count;
  73. TiffGeoTag *geotags;
  74. } TiffContext;
  75. static void free_geotags(TiffContext *const s)
  76. {
  77. int i;
  78. for (i = 0; i < s->geotag_count; i++) {
  79. if (s->geotags[i].val)
  80. av_freep(&s->geotags[i].val);
  81. }
  82. av_freep(&s->geotags);
  83. s->geotag_count = 0;
  84. }
  85. #define RET_GEOKEY(TYPE, array, element)\
  86. if (key >= TIFF_##TYPE##_KEY_ID_OFFSET &&\
  87. key - TIFF_##TYPE##_KEY_ID_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_name_type_map))\
  88. return ff_tiff_##array##_name_type_map[key - TIFF_##TYPE##_KEY_ID_OFFSET].element;
  89. static const char *get_geokey_name(int key)
  90. {
  91. RET_GEOKEY(VERT, vert, name);
  92. RET_GEOKEY(PROJ, proj, name);
  93. RET_GEOKEY(GEOG, geog, name);
  94. RET_GEOKEY(CONF, conf, name);
  95. return NULL;
  96. }
  97. static int get_geokey_type(int key)
  98. {
  99. RET_GEOKEY(VERT, vert, type);
  100. RET_GEOKEY(PROJ, proj, type);
  101. RET_GEOKEY(GEOG, geog, type);
  102. RET_GEOKEY(CONF, conf, type);
  103. return AVERROR_INVALIDDATA;
  104. }
  105. static int cmp_id_key(const void *id, const void *k)
  106. {
  107. return *(const int*)id - ((const TiffGeoTagKeyName*)k)->key;
  108. }
  109. static const char *search_keyval(const TiffGeoTagKeyName *keys, int n, int id)
  110. {
  111. TiffGeoTagKeyName *r = bsearch(&id, keys, n, sizeof(keys[0]), cmp_id_key);
  112. if(r)
  113. return r->name;
  114. return NULL;
  115. }
  116. static char *get_geokey_val(int key, int val)
  117. {
  118. char *ap;
  119. if (val == TIFF_GEO_KEY_UNDEFINED)
  120. return av_strdup("undefined");
  121. if (val == TIFF_GEO_KEY_USER_DEFINED)
  122. return av_strdup("User-Defined");
  123. #define RET_GEOKEY_VAL(TYPE, array)\
  124. if (val >= TIFF_##TYPE##_OFFSET &&\
  125. val - TIFF_##TYPE##_OFFSET < FF_ARRAY_ELEMS(ff_tiff_##array##_codes))\
  126. return av_strdup(ff_tiff_##array##_codes[val - TIFF_##TYPE##_OFFSET]);
  127. switch (key) {
  128. case TIFF_GT_MODEL_TYPE_GEOKEY:
  129. RET_GEOKEY_VAL(GT_MODEL_TYPE, gt_model_type);
  130. break;
  131. case TIFF_GT_RASTER_TYPE_GEOKEY:
  132. RET_GEOKEY_VAL(GT_RASTER_TYPE, gt_raster_type);
  133. break;
  134. case TIFF_GEOG_LINEAR_UNITS_GEOKEY:
  135. case TIFF_PROJ_LINEAR_UNITS_GEOKEY:
  136. case TIFF_VERTICAL_UNITS_GEOKEY:
  137. RET_GEOKEY_VAL(LINEAR_UNIT, linear_unit);
  138. break;
  139. case TIFF_GEOG_ANGULAR_UNITS_GEOKEY:
  140. case TIFF_GEOG_AZIMUTH_UNITS_GEOKEY:
  141. RET_GEOKEY_VAL(ANGULAR_UNIT, angular_unit);
  142. break;
  143. case TIFF_GEOGRAPHIC_TYPE_GEOKEY:
  144. RET_GEOKEY_VAL(GCS_TYPE, gcs_type);
  145. RET_GEOKEY_VAL(GCSE_TYPE, gcse_type);
  146. break;
  147. case TIFF_GEOG_GEODETIC_DATUM_GEOKEY:
  148. RET_GEOKEY_VAL(GEODETIC_DATUM, geodetic_datum);
  149. RET_GEOKEY_VAL(GEODETIC_DATUM_E, geodetic_datum_e);
  150. break;
  151. case TIFF_GEOG_ELLIPSOID_GEOKEY:
  152. RET_GEOKEY_VAL(ELLIPSOID, ellipsoid);
  153. break;
  154. case TIFF_GEOG_PRIME_MERIDIAN_GEOKEY:
  155. RET_GEOKEY_VAL(PRIME_MERIDIAN, prime_meridian);
  156. break;
  157. case TIFF_PROJECTED_CS_TYPE_GEOKEY:
  158. ap = av_strdup(search_keyval(ff_tiff_proj_cs_type_codes, FF_ARRAY_ELEMS(ff_tiff_proj_cs_type_codes), val));
  159. if(ap) return ap;
  160. break;
  161. case TIFF_PROJECTION_GEOKEY:
  162. ap = av_strdup(search_keyval(ff_tiff_projection_codes, FF_ARRAY_ELEMS(ff_tiff_projection_codes), val));
  163. if(ap) return ap;
  164. break;
  165. case TIFF_PROJ_COORD_TRANS_GEOKEY:
  166. RET_GEOKEY_VAL(COORD_TRANS, coord_trans);
  167. break;
  168. case TIFF_VERTICAL_CS_TYPE_GEOKEY:
  169. RET_GEOKEY_VAL(VERT_CS, vert_cs);
  170. RET_GEOKEY_VAL(ORTHO_VERT_CS, ortho_vert_cs);
  171. break;
  172. }
  173. ap = av_malloc(14);
  174. if (ap)
  175. snprintf(ap, 14, "Unknown-%d", val);
  176. return ap;
  177. }
  178. static char *doubles2str(double *dp, int count, const char *sep)
  179. {
  180. int i;
  181. char *ap, *ap0;
  182. uint64_t component_len;
  183. if (!sep) sep = ", ";
  184. component_len = 24LL + strlen(sep);
  185. if (count >= (INT_MAX - 1)/component_len)
  186. return NULL;
  187. ap = av_malloc(component_len * count + 1);
  188. if (!ap)
  189. return NULL;
  190. ap0 = ap;
  191. ap[0] = '\0';
  192. for (i = 0; i < count; i++) {
  193. unsigned l = snprintf(ap, component_len, "%.15g%s", dp[i], sep);
  194. if(l >= component_len) {
  195. av_free(ap0);
  196. return NULL;
  197. }
  198. ap += l;
  199. }
  200. ap0[strlen(ap0) - strlen(sep)] = '\0';
  201. return ap0;
  202. }
  203. static int add_metadata(int count, int type,
  204. const char *name, const char *sep, TiffContext *s, AVFrame *frame)
  205. {
  206. switch(type) {
  207. case TIFF_DOUBLE: return ff_tadd_doubles_metadata(count, name, sep, &s->gb, s->le, &frame->metadata);
  208. case TIFF_SHORT : return ff_tadd_shorts_metadata(count, name, sep, &s->gb, s->le, 0, &frame->metadata);
  209. case TIFF_STRING: return ff_tadd_string_metadata(count, name, &s->gb, s->le, &frame->metadata);
  210. default : return AVERROR_INVALIDDATA;
  211. };
  212. }
  213. static void av_always_inline horizontal_fill(unsigned int bpp, uint8_t* dst,
  214. int usePtr, const uint8_t *src,
  215. uint8_t c, int width, int offset)
  216. {
  217. switch (bpp) {
  218. case 1:
  219. while (--width >= 0) {
  220. dst[(width+offset)*8+7] = (usePtr ? src[width] : c) & 0x1;
  221. dst[(width+offset)*8+6] = (usePtr ? src[width] : c) >> 1 & 0x1;
  222. dst[(width+offset)*8+5] = (usePtr ? src[width] : c) >> 2 & 0x1;
  223. dst[(width+offset)*8+4] = (usePtr ? src[width] : c) >> 3 & 0x1;
  224. dst[(width+offset)*8+3] = (usePtr ? src[width] : c) >> 4 & 0x1;
  225. dst[(width+offset)*8+2] = (usePtr ? src[width] : c) >> 5 & 0x1;
  226. dst[(width+offset)*8+1] = (usePtr ? src[width] : c) >> 6 & 0x1;
  227. dst[(width+offset)*8+0] = (usePtr ? src[width] : c) >> 7;
  228. }
  229. break;
  230. case 2:
  231. while (--width >= 0) {
  232. dst[(width+offset)*4+3] = (usePtr ? src[width] : c) & 0x3;
  233. dst[(width+offset)*4+2] = (usePtr ? src[width] : c) >> 2 & 0x3;
  234. dst[(width+offset)*4+1] = (usePtr ? src[width] : c) >> 4 & 0x3;
  235. dst[(width+offset)*4+0] = (usePtr ? src[width] : c) >> 6;
  236. }
  237. break;
  238. case 4:
  239. while (--width >= 0) {
  240. dst[(width+offset)*2+1] = (usePtr ? src[width] : c) & 0xF;
  241. dst[(width+offset)*2+0] = (usePtr ? src[width] : c) >> 4;
  242. }
  243. break;
  244. default:
  245. if (usePtr) {
  246. memcpy(dst + offset, src, width);
  247. } else {
  248. memset(dst + offset, c, width);
  249. }
  250. }
  251. }
  252. static int deinvert_buffer(TiffContext *s, const uint8_t *src, int size)
  253. {
  254. int i;
  255. av_fast_padded_malloc(&s->deinvert_buf, &s->deinvert_buf_size, size);
  256. if (!s->deinvert_buf)
  257. return AVERROR(ENOMEM);
  258. for (i = 0; i < size; i++)
  259. s->deinvert_buf[i] = ff_reverse[src[i]];
  260. return 0;
  261. }
  262. static void unpack_yuv(TiffContext *s, AVFrame *p,
  263. const uint8_t *src, int lnum)
  264. {
  265. int i, j, k;
  266. int w = (s->width - 1) / s->subsampling[0] + 1;
  267. uint8_t *pu = &p->data[1][lnum / s->subsampling[1] * p->linesize[1]];
  268. uint8_t *pv = &p->data[2][lnum / s->subsampling[1] * p->linesize[2]];
  269. if (s->width % s->subsampling[0] || s->height % s->subsampling[1]) {
  270. for (i = 0; i < w; i++) {
  271. for (j = 0; j < s->subsampling[1]; j++)
  272. for (k = 0; k < s->subsampling[0]; k++)
  273. p->data[0][FFMIN(lnum + j, s->height-1) * p->linesize[0] +
  274. FFMIN(i * s->subsampling[0] + k, s->width-1)] = *src++;
  275. *pu++ = *src++;
  276. *pv++ = *src++;
  277. }
  278. }else{
  279. for (i = 0; i < w; i++) {
  280. for (j = 0; j < s->subsampling[1]; j++)
  281. for (k = 0; k < s->subsampling[0]; k++)
  282. p->data[0][(lnum + j) * p->linesize[0] +
  283. i * s->subsampling[0] + k] = *src++;
  284. *pu++ = *src++;
  285. *pv++ = *src++;
  286. }
  287. }
  288. }
  289. #if CONFIG_ZLIB
  290. static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
  291. int size)
  292. {
  293. z_stream zstream = { 0 };
  294. int zret;
  295. zstream.next_in = (uint8_t *)src;
  296. zstream.avail_in = size;
  297. zstream.next_out = dst;
  298. zstream.avail_out = *len;
  299. zret = inflateInit(&zstream);
  300. if (zret != Z_OK) {
  301. av_log(NULL, AV_LOG_ERROR, "Inflate init error: %d\n", zret);
  302. return zret;
  303. }
  304. zret = inflate(&zstream, Z_SYNC_FLUSH);
  305. inflateEnd(&zstream);
  306. *len = zstream.total_out;
  307. return zret == Z_STREAM_END ? Z_OK : zret;
  308. }
  309. static int tiff_unpack_zlib(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  310. const uint8_t *src, int size, int width, int lines,
  311. int strip_start, int is_yuv)
  312. {
  313. uint8_t *zbuf;
  314. unsigned long outlen;
  315. int ret, line;
  316. outlen = width * lines;
  317. zbuf = av_malloc(outlen);
  318. if (!zbuf)
  319. return AVERROR(ENOMEM);
  320. if (s->fill_order) {
  321. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  322. av_free(zbuf);
  323. return ret;
  324. }
  325. src = s->deinvert_buf;
  326. }
  327. ret = tiff_uncompress(zbuf, &outlen, src, size);
  328. if (ret != Z_OK) {
  329. av_log(s->avctx, AV_LOG_ERROR,
  330. "Uncompressing failed (%lu of %lu) with error %d\n", outlen,
  331. (unsigned long)width * lines, ret);
  332. av_free(zbuf);
  333. return AVERROR_UNKNOWN;
  334. }
  335. src = zbuf;
  336. for (line = 0; line < lines; line++) {
  337. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  338. horizontal_fill(s->bpp, dst, 1, src, 0, width, 0);
  339. } else {
  340. memcpy(dst, src, width);
  341. }
  342. if (is_yuv) {
  343. unpack_yuv(s, p, dst, strip_start + line);
  344. line += s->subsampling[1] - 1;
  345. }
  346. dst += stride;
  347. src += width;
  348. }
  349. av_free(zbuf);
  350. return 0;
  351. }
  352. #endif
  353. #if CONFIG_LZMA
  354. static int tiff_uncompress_lzma(uint8_t *dst, uint64_t *len, const uint8_t *src,
  355. int size)
  356. {
  357. lzma_stream stream = LZMA_STREAM_INIT;
  358. lzma_ret ret;
  359. stream.next_in = (uint8_t *)src;
  360. stream.avail_in = size;
  361. stream.next_out = dst;
  362. stream.avail_out = *len;
  363. ret = lzma_stream_decoder(&stream, UINT64_MAX, 0);
  364. if (ret != LZMA_OK) {
  365. av_log(NULL, AV_LOG_ERROR, "LZMA init error: %d\n", ret);
  366. return ret;
  367. }
  368. ret = lzma_code(&stream, LZMA_RUN);
  369. lzma_end(&stream);
  370. *len = stream.total_out;
  371. return ret == LZMA_STREAM_END ? LZMA_OK : ret;
  372. }
  373. static int tiff_unpack_lzma(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  374. const uint8_t *src, int size, int width, int lines,
  375. int strip_start, int is_yuv)
  376. {
  377. uint64_t outlen = width * (uint64_t)lines;
  378. int ret, line;
  379. uint8_t *buf = av_malloc(outlen);
  380. if (!buf)
  381. return AVERROR(ENOMEM);
  382. if (s->fill_order) {
  383. if ((ret = deinvert_buffer(s, src, size)) < 0) {
  384. av_free(buf);
  385. return ret;
  386. }
  387. src = s->deinvert_buf;
  388. }
  389. ret = tiff_uncompress_lzma(buf, &outlen, src, size);
  390. if (ret != LZMA_OK) {
  391. av_log(s->avctx, AV_LOG_ERROR,
  392. "Uncompressing failed (%"PRIu64" of %"PRIu64") with error %d\n", outlen,
  393. (uint64_t)width * lines, ret);
  394. av_free(buf);
  395. return AVERROR_UNKNOWN;
  396. }
  397. src = buf;
  398. for (line = 0; line < lines; line++) {
  399. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  400. horizontal_fill(s->bpp, dst, 1, src, 0, width, 0);
  401. } else {
  402. memcpy(dst, src, width);
  403. }
  404. if (is_yuv) {
  405. unpack_yuv(s, p, dst, strip_start + line);
  406. line += s->subsampling[1] - 1;
  407. }
  408. dst += stride;
  409. src += width;
  410. }
  411. av_free(buf);
  412. return 0;
  413. }
  414. #endif
  415. static int tiff_unpack_fax(TiffContext *s, uint8_t *dst, int stride,
  416. const uint8_t *src, int size, int width, int lines)
  417. {
  418. int i, ret = 0;
  419. int line;
  420. uint8_t *src2;
  421. av_fast_padded_malloc(&s->fax_buffer, &s->fax_buffer_size, size);
  422. src2 = s->fax_buffer;
  423. if (!src2) {
  424. av_log(s->avctx, AV_LOG_ERROR,
  425. "Error allocating temporary buffer\n");
  426. return AVERROR(ENOMEM);
  427. }
  428. if (!s->fill_order) {
  429. memcpy(src2, src, size);
  430. } else {
  431. for (i = 0; i < size; i++)
  432. src2[i] = ff_reverse[src[i]];
  433. }
  434. memset(src2 + size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
  435. ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride,
  436. s->compr, s->fax_opts);
  437. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  438. for (line = 0; line < lines; line++) {
  439. horizontal_fill(s->bpp, dst, 1, dst, 0, width, 0);
  440. dst += stride;
  441. }
  442. return ret;
  443. }
  444. static int tiff_unpack_strip(TiffContext *s, AVFrame *p, uint8_t *dst, int stride,
  445. const uint8_t *src, int size, int strip_start, int lines)
  446. {
  447. PutByteContext pb;
  448. int c, line, pixels, code, ret;
  449. const uint8_t *ssrc = src;
  450. int width = ((s->width * s->bpp) + 7) >> 3;
  451. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(p->format);
  452. int is_yuv = !(desc->flags & AV_PIX_FMT_FLAG_RGB) &&
  453. (desc->flags & AV_PIX_FMT_FLAG_PLANAR) &&
  454. desc->nb_components >= 3;
  455. if (s->planar)
  456. width /= s->bppcount;
  457. if (size <= 0)
  458. return AVERROR_INVALIDDATA;
  459. if (is_yuv) {
  460. int bytes_per_row = (((s->width - 1) / s->subsampling[0] + 1) * s->bpp *
  461. s->subsampling[0] * s->subsampling[1] + 7) >> 3;
  462. av_fast_padded_malloc(&s->yuv_line, &s->yuv_line_size, bytes_per_row);
  463. if (s->yuv_line == NULL) {
  464. av_log(s->avctx, AV_LOG_ERROR, "Not enough memory\n");
  465. return AVERROR(ENOMEM);
  466. }
  467. dst = s->yuv_line;
  468. stride = 0;
  469. width = (s->width - 1) / s->subsampling[0] + 1;
  470. width = width * s->subsampling[0] * s->subsampling[1] + 2*width;
  471. av_assert0(width <= bytes_per_row);
  472. av_assert0(s->bpp == 24);
  473. }
  474. if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
  475. #if CONFIG_ZLIB
  476. return tiff_unpack_zlib(s, p, dst, stride, src, size, width, lines,
  477. strip_start, is_yuv);
  478. #else
  479. av_log(s->avctx, AV_LOG_ERROR,
  480. "zlib support not enabled, "
  481. "deflate compression not supported\n");
  482. return AVERROR(ENOSYS);
  483. #endif
  484. }
  485. if (s->compr == TIFF_LZMA) {
  486. #if CONFIG_LZMA
  487. return tiff_unpack_lzma(s, p, dst, stride, src, size, width, lines,
  488. strip_start, is_yuv);
  489. #else
  490. av_log(s->avctx, AV_LOG_ERROR,
  491. "LZMA support not enabled\n");
  492. return AVERROR(ENOSYS);
  493. #endif
  494. }
  495. if (s->compr == TIFF_LZW) {
  496. if (s->fill_order) {
  497. if ((ret = deinvert_buffer(s, src, size)) < 0)
  498. return ret;
  499. ssrc = src = s->deinvert_buf;
  500. }
  501. if (size > 1 && !src[0] && (src[1]&1)) {
  502. av_log(s->avctx, AV_LOG_ERROR, "Old style LZW is unsupported\n");
  503. }
  504. if ((ret = ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF)) < 0) {
  505. av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
  506. return ret;
  507. }
  508. for (line = 0; line < lines; line++) {
  509. pixels = ff_lzw_decode(s->lzw, dst, width);
  510. if (pixels < width) {
  511. av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
  512. pixels, width);
  513. return AVERROR_INVALIDDATA;
  514. }
  515. if (s->bpp < 8 && s->avctx->pix_fmt == AV_PIX_FMT_PAL8)
  516. horizontal_fill(s->bpp, dst, 1, dst, 0, width, 0);
  517. if (is_yuv) {
  518. unpack_yuv(s, p, dst, strip_start + line);
  519. line += s->subsampling[1] - 1;
  520. }
  521. dst += stride;
  522. }
  523. return 0;
  524. }
  525. if (s->compr == TIFF_CCITT_RLE ||
  526. s->compr == TIFF_G3 ||
  527. s->compr == TIFF_G4) {
  528. if (is_yuv || p->format == AV_PIX_FMT_GRAY12)
  529. return AVERROR_INVALIDDATA;
  530. return tiff_unpack_fax(s, dst, stride, src, size, width, lines);
  531. }
  532. bytestream2_init(&s->gb, src, size);
  533. bytestream2_init_writer(&pb, dst, is_yuv ? s->yuv_line_size : (stride * lines));
  534. for (line = 0; line < lines; line++) {
  535. if (src - ssrc > size) {
  536. av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
  537. return AVERROR_INVALIDDATA;
  538. }
  539. if (bytestream2_get_bytes_left(&s->gb) == 0 || bytestream2_get_eof(&pb))
  540. break;
  541. bytestream2_seek_p(&pb, stride * line, SEEK_SET);
  542. switch (s->compr) {
  543. case TIFF_RAW:
  544. if (ssrc + size - src < width)
  545. return AVERROR_INVALIDDATA;
  546. if (!s->fill_order) {
  547. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  548. dst, 1, src, 0, width, 0);
  549. } else {
  550. int i;
  551. for (i = 0; i < width; i++)
  552. dst[i] = ff_reverse[src[i]];
  553. }
  554. src += width;
  555. break;
  556. case TIFF_PACKBITS:
  557. for (pixels = 0; pixels < width;) {
  558. if (ssrc + size - src < 2) {
  559. av_log(s->avctx, AV_LOG_ERROR, "Read went out of bounds\n");
  560. return AVERROR_INVALIDDATA;
  561. }
  562. code = s->fill_order ? (int8_t) ff_reverse[*src++]: (int8_t) *src++;
  563. if (code >= 0) {
  564. code++;
  565. if (pixels + code > width ||
  566. ssrc + size - src < code) {
  567. av_log(s->avctx, AV_LOG_ERROR,
  568. "Copy went out of bounds\n");
  569. return AVERROR_INVALIDDATA;
  570. }
  571. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  572. dst, 1, src, 0, code, pixels);
  573. src += code;
  574. pixels += code;
  575. } else if (code != -128) { // -127..-1
  576. code = (-code) + 1;
  577. if (pixels + code > width) {
  578. av_log(s->avctx, AV_LOG_ERROR,
  579. "Run went out of bounds\n");
  580. return AVERROR_INVALIDDATA;
  581. }
  582. c = *src++;
  583. horizontal_fill(s->bpp * (s->avctx->pix_fmt == AV_PIX_FMT_PAL8),
  584. dst, 0, NULL, c, code, pixels);
  585. pixels += code;
  586. }
  587. }
  588. if (s->fill_order) {
  589. int i;
  590. for (i = 0; i < width; i++)
  591. dst[i] = ff_reverse[dst[i]];
  592. }
  593. break;
  594. }
  595. if (is_yuv) {
  596. unpack_yuv(s, p, dst, strip_start + line);
  597. line += s->subsampling[1] - 1;
  598. }
  599. dst += stride;
  600. }
  601. return 0;
  602. }
  603. static int init_image(TiffContext *s, ThreadFrame *frame)
  604. {
  605. int ret;
  606. int create_gray_palette = 0;
  607. // make sure there is no aliasing in the following switch
  608. if (s->bpp >= 100 || s->bppcount >= 10) {
  609. av_log(s->avctx, AV_LOG_ERROR,
  610. "Unsupported image parameters: bpp=%d, bppcount=%d\n",
  611. s->bpp, s->bppcount);
  612. return AVERROR_INVALIDDATA;
  613. }
  614. switch (s->planar * 1000 + s->bpp * 10 + s->bppcount) {
  615. case 11:
  616. if (!s->palette_is_set) {
  617. s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK;
  618. break;
  619. }
  620. case 21:
  621. case 41:
  622. s->avctx->pix_fmt = AV_PIX_FMT_PAL8;
  623. if (!s->palette_is_set) {
  624. create_gray_palette = 1;
  625. }
  626. break;
  627. case 81:
  628. s->avctx->pix_fmt = s->palette_is_set ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8;
  629. break;
  630. case 243:
  631. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  632. if (s->subsampling[0] == 1 && s->subsampling[1] == 1) {
  633. s->avctx->pix_fmt = AV_PIX_FMT_YUV444P;
  634. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 1) {
  635. s->avctx->pix_fmt = AV_PIX_FMT_YUV422P;
  636. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 1) {
  637. s->avctx->pix_fmt = AV_PIX_FMT_YUV411P;
  638. } else if (s->subsampling[0] == 1 && s->subsampling[1] == 2) {
  639. s->avctx->pix_fmt = AV_PIX_FMT_YUV440P;
  640. } else if (s->subsampling[0] == 2 && s->subsampling[1] == 2) {
  641. s->avctx->pix_fmt = AV_PIX_FMT_YUV420P;
  642. } else if (s->subsampling[0] == 4 && s->subsampling[1] == 4) {
  643. s->avctx->pix_fmt = AV_PIX_FMT_YUV410P;
  644. } else {
  645. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr subsampling\n");
  646. return AVERROR_PATCHWELCOME;
  647. }
  648. } else
  649. s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
  650. break;
  651. case 161:
  652. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE;
  653. break;
  654. case 162:
  655. s->avctx->pix_fmt = AV_PIX_FMT_YA8;
  656. break;
  657. case 322:
  658. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_YA16LE : AV_PIX_FMT_YA16BE;
  659. break;
  660. case 324:
  661. s->avctx->pix_fmt = AV_PIX_FMT_RGBA;
  662. break;
  663. case 483:
  664. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE;
  665. break;
  666. case 644:
  667. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGBA64LE : AV_PIX_FMT_RGBA64BE;
  668. break;
  669. case 1243:
  670. s->avctx->pix_fmt = AV_PIX_FMT_GBRP;
  671. break;
  672. case 1324:
  673. s->avctx->pix_fmt = AV_PIX_FMT_GBRAP;
  674. break;
  675. case 1483:
  676. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRP16LE : AV_PIX_FMT_GBRP16BE;
  677. break;
  678. case 1644:
  679. s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GBRAP16LE : AV_PIX_FMT_GBRAP16BE;
  680. break;
  681. default:
  682. av_log(s->avctx, AV_LOG_ERROR,
  683. "This format is not supported (bpp=%d, bppcount=%d)\n",
  684. s->bpp, s->bppcount);
  685. return AVERROR_INVALIDDATA;
  686. }
  687. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  688. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->avctx->pix_fmt);
  689. if((desc->flags & AV_PIX_FMT_FLAG_RGB) ||
  690. !(desc->flags & AV_PIX_FMT_FLAG_PLANAR) ||
  691. desc->nb_components < 3) {
  692. av_log(s->avctx, AV_LOG_ERROR, "Unsupported YCbCr variant\n");
  693. return AVERROR_INVALIDDATA;
  694. }
  695. }
  696. if (s->width != s->avctx->width || s->height != s->avctx->height) {
  697. ret = ff_set_dimensions(s->avctx, s->width, s->height);
  698. if (ret < 0)
  699. return ret;
  700. }
  701. if ((ret = ff_thread_get_buffer(s->avctx, frame, 0)) < 0)
  702. return ret;
  703. if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) {
  704. if (!create_gray_palette)
  705. memcpy(frame->f->data[1], s->palette, sizeof(s->palette));
  706. else {
  707. /* make default grayscale pal */
  708. int i;
  709. uint32_t *pal = (uint32_t *)frame->f->data[1];
  710. for (i = 0; i < 1<<s->bpp; i++)
  711. pal[i] = 0xFFU << 24 | i * 255 / ((1<<s->bpp) - 1) * 0x010101;
  712. }
  713. }
  714. return 0;
  715. }
  716. static void set_sar(TiffContext *s, unsigned tag, unsigned num, unsigned den)
  717. {
  718. int offset = tag == TIFF_YRES ? 2 : 0;
  719. s->res[offset++] = num;
  720. s->res[offset] = den;
  721. if (s->res[0] && s->res[1] && s->res[2] && s->res[3]) {
  722. uint64_t num = s->res[2] * (uint64_t)s->res[1];
  723. uint64_t den = s->res[0] * (uint64_t)s->res[3];
  724. if (num > INT64_MAX || den > INT64_MAX) {
  725. num = num >> 1;
  726. den = den >> 1;
  727. }
  728. av_reduce(&s->avctx->sample_aspect_ratio.num, &s->avctx->sample_aspect_ratio.den,
  729. num, den, INT32_MAX);
  730. if (!s->avctx->sample_aspect_ratio.den)
  731. s->avctx->sample_aspect_ratio = (AVRational) {0, 1};
  732. }
  733. }
  734. static int tiff_decode_tag(TiffContext *s, AVFrame *frame)
  735. {
  736. unsigned tag, type, count, off, value = 0, value2 = 0;
  737. int i, start;
  738. int pos;
  739. int ret;
  740. double *dp;
  741. ret = ff_tread_tag(&s->gb, s->le, &tag, &type, &count, &start);
  742. if (ret < 0) {
  743. goto end;
  744. }
  745. off = bytestream2_tell(&s->gb);
  746. if (count == 1) {
  747. switch (type) {
  748. case TIFF_BYTE:
  749. case TIFF_SHORT:
  750. case TIFF_LONG:
  751. value = ff_tget(&s->gb, type, s->le);
  752. break;
  753. case TIFF_RATIONAL:
  754. value = ff_tget(&s->gb, TIFF_LONG, s->le);
  755. value2 = ff_tget(&s->gb, TIFF_LONG, s->le);
  756. break;
  757. case TIFF_STRING:
  758. if (count <= 4) {
  759. break;
  760. }
  761. default:
  762. value = UINT_MAX;
  763. }
  764. }
  765. switch (tag) {
  766. case TIFF_WIDTH:
  767. s->width = value;
  768. break;
  769. case TIFF_HEIGHT:
  770. s->height = value;
  771. break;
  772. case TIFF_BPP:
  773. if (count > 4U) {
  774. av_log(s->avctx, AV_LOG_ERROR,
  775. "This format is not supported (bpp=%d, %d components)\n",
  776. value, count);
  777. return AVERROR_INVALIDDATA;
  778. }
  779. s->bppcount = count;
  780. if (count == 1)
  781. s->bpp = value;
  782. else {
  783. switch (type) {
  784. case TIFF_BYTE:
  785. case TIFF_SHORT:
  786. case TIFF_LONG:
  787. s->bpp = 0;
  788. if (bytestream2_get_bytes_left(&s->gb) < type_sizes[type] * count)
  789. return AVERROR_INVALIDDATA;
  790. for (i = 0; i < count; i++)
  791. s->bpp += ff_tget(&s->gb, type, s->le);
  792. break;
  793. default:
  794. s->bpp = -1;
  795. }
  796. }
  797. break;
  798. case TIFF_SAMPLES_PER_PIXEL:
  799. if (count != 1) {
  800. av_log(s->avctx, AV_LOG_ERROR,
  801. "Samples per pixel requires a single value, many provided\n");
  802. return AVERROR_INVALIDDATA;
  803. }
  804. if (value > 4U) {
  805. av_log(s->avctx, AV_LOG_ERROR,
  806. "Samples per pixel %d is too large\n", value);
  807. return AVERROR_INVALIDDATA;
  808. }
  809. if (s->bppcount == 1)
  810. s->bpp *= value;
  811. s->bppcount = value;
  812. break;
  813. case TIFF_COMPR:
  814. s->compr = value;
  815. av_log(s->avctx, AV_LOG_DEBUG, "compression: %d\n", s->compr);
  816. s->predictor = 0;
  817. switch (s->compr) {
  818. case TIFF_RAW:
  819. case TIFF_PACKBITS:
  820. case TIFF_LZW:
  821. case TIFF_CCITT_RLE:
  822. break;
  823. case TIFF_G3:
  824. case TIFF_G4:
  825. s->fax_opts = 0;
  826. break;
  827. case TIFF_DEFLATE:
  828. case TIFF_ADOBE_DEFLATE:
  829. #if CONFIG_ZLIB
  830. break;
  831. #else
  832. av_log(s->avctx, AV_LOG_ERROR, "Deflate: ZLib not compiled in\n");
  833. return AVERROR(ENOSYS);
  834. #endif
  835. case TIFF_JPEG:
  836. case TIFF_NEWJPEG:
  837. avpriv_report_missing_feature(s->avctx, "JPEG compression");
  838. return AVERROR_PATCHWELCOME;
  839. case TIFF_LZMA:
  840. #if CONFIG_LZMA
  841. break;
  842. #else
  843. av_log(s->avctx, AV_LOG_ERROR, "LZMA not compiled in\n");
  844. return AVERROR(ENOSYS);
  845. #endif
  846. default:
  847. av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
  848. s->compr);
  849. return AVERROR_INVALIDDATA;
  850. }
  851. break;
  852. case TIFF_ROWSPERSTRIP:
  853. if (!value || (type == TIFF_LONG && value == UINT_MAX))
  854. value = s->height;
  855. s->rps = FFMIN(value, s->height);
  856. break;
  857. case TIFF_STRIP_OFFS:
  858. if (count == 1) {
  859. if (value > INT_MAX) {
  860. av_log(s->avctx, AV_LOG_ERROR,
  861. "strippos %u too large\n", value);
  862. return AVERROR_INVALIDDATA;
  863. }
  864. s->strippos = 0;
  865. s->stripoff = value;
  866. } else
  867. s->strippos = off;
  868. s->strips = count;
  869. if (s->strips == 1)
  870. s->rps = s->height;
  871. s->sot = type;
  872. break;
  873. case TIFF_STRIP_SIZE:
  874. if (count == 1) {
  875. if (value > INT_MAX) {
  876. av_log(s->avctx, AV_LOG_ERROR,
  877. "stripsize %u too large\n", value);
  878. return AVERROR_INVALIDDATA;
  879. }
  880. s->stripsizesoff = 0;
  881. s->stripsize = value;
  882. s->strips = 1;
  883. } else {
  884. s->stripsizesoff = off;
  885. }
  886. s->strips = count;
  887. s->sstype = type;
  888. break;
  889. case TIFF_XRES:
  890. case TIFF_YRES:
  891. set_sar(s, tag, value, value2);
  892. break;
  893. case TIFF_TILE_BYTE_COUNTS:
  894. case TIFF_TILE_LENGTH:
  895. case TIFF_TILE_OFFSETS:
  896. case TIFF_TILE_WIDTH:
  897. av_log(s->avctx, AV_LOG_ERROR, "Tiled images are not supported\n");
  898. return AVERROR_PATCHWELCOME;
  899. break;
  900. case TIFF_PREDICTOR:
  901. s->predictor = value;
  902. break;
  903. case TIFF_PHOTOMETRIC:
  904. switch (value) {
  905. case TIFF_PHOTOMETRIC_WHITE_IS_ZERO:
  906. case TIFF_PHOTOMETRIC_BLACK_IS_ZERO:
  907. case TIFF_PHOTOMETRIC_RGB:
  908. case TIFF_PHOTOMETRIC_PALETTE:
  909. case TIFF_PHOTOMETRIC_YCBCR:
  910. s->photometric = value;
  911. break;
  912. case TIFF_PHOTOMETRIC_ALPHA_MASK:
  913. case TIFF_PHOTOMETRIC_SEPARATED:
  914. case TIFF_PHOTOMETRIC_CIE_LAB:
  915. case TIFF_PHOTOMETRIC_ICC_LAB:
  916. case TIFF_PHOTOMETRIC_ITU_LAB:
  917. case TIFF_PHOTOMETRIC_CFA:
  918. case TIFF_PHOTOMETRIC_LOG_L:
  919. case TIFF_PHOTOMETRIC_LOG_LUV:
  920. case TIFF_PHOTOMETRIC_LINEAR_RAW:
  921. avpriv_report_missing_feature(s->avctx,
  922. "PhotometricInterpretation 0x%04X",
  923. value);
  924. return AVERROR_PATCHWELCOME;
  925. default:
  926. av_log(s->avctx, AV_LOG_ERROR, "PhotometricInterpretation %u is "
  927. "unknown\n", value);
  928. return AVERROR_INVALIDDATA;
  929. }
  930. break;
  931. case TIFF_FILL_ORDER:
  932. if (value < 1 || value > 2) {
  933. av_log(s->avctx, AV_LOG_ERROR,
  934. "Unknown FillOrder value %d, trying default one\n", value);
  935. value = 1;
  936. }
  937. s->fill_order = value - 1;
  938. break;
  939. case TIFF_PAL: {
  940. GetByteContext pal_gb[3];
  941. off = type_sizes[type];
  942. if (count / 3 > 256 ||
  943. bytestream2_get_bytes_left(&s->gb) < count / 3 * off * 3)
  944. return AVERROR_INVALIDDATA;
  945. pal_gb[0] = pal_gb[1] = pal_gb[2] = s->gb;
  946. bytestream2_skip(&pal_gb[1], count / 3 * off);
  947. bytestream2_skip(&pal_gb[2], count / 3 * off * 2);
  948. off = (type_sizes[type] - 1) << 3;
  949. if (off > 31U) {
  950. av_log(s->avctx, AV_LOG_ERROR, "palette shift %d is out of range\n", off);
  951. return AVERROR_INVALIDDATA;
  952. }
  953. for (i = 0; i < count / 3; i++) {
  954. uint32_t p = 0xFF000000;
  955. p |= (ff_tget(&pal_gb[0], type, s->le) >> off) << 16;
  956. p |= (ff_tget(&pal_gb[1], type, s->le) >> off) << 8;
  957. p |= ff_tget(&pal_gb[2], type, s->le) >> off;
  958. s->palette[i] = p;
  959. }
  960. s->palette_is_set = 1;
  961. break;
  962. }
  963. case TIFF_PLANAR:
  964. s->planar = value == 2;
  965. break;
  966. case TIFF_YCBCR_SUBSAMPLING:
  967. if (count != 2) {
  968. av_log(s->avctx, AV_LOG_ERROR, "subsample count invalid\n");
  969. return AVERROR_INVALIDDATA;
  970. }
  971. for (i = 0; i < count; i++) {
  972. s->subsampling[i] = ff_tget(&s->gb, type, s->le);
  973. if (s->subsampling[i] <= 0) {
  974. av_log(s->avctx, AV_LOG_ERROR, "subsampling %d is invalid\n", s->subsampling[i]);
  975. s->subsampling[i] = 1;
  976. return AVERROR_INVALIDDATA;
  977. }
  978. }
  979. break;
  980. case TIFF_T4OPTIONS:
  981. if (s->compr == TIFF_G3)
  982. s->fax_opts = value;
  983. break;
  984. case TIFF_T6OPTIONS:
  985. if (s->compr == TIFF_G4)
  986. s->fax_opts = value;
  987. break;
  988. #define ADD_METADATA(count, name, sep)\
  989. if ((ret = add_metadata(count, type, name, sep, s, frame)) < 0) {\
  990. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");\
  991. goto end;\
  992. }
  993. case TIFF_MODEL_PIXEL_SCALE:
  994. ADD_METADATA(count, "ModelPixelScaleTag", NULL);
  995. break;
  996. case TIFF_MODEL_TRANSFORMATION:
  997. ADD_METADATA(count, "ModelTransformationTag", NULL);
  998. break;
  999. case TIFF_MODEL_TIEPOINT:
  1000. ADD_METADATA(count, "ModelTiepointTag", NULL);
  1001. break;
  1002. case TIFF_GEO_KEY_DIRECTORY:
  1003. if (s->geotag_count) {
  1004. avpriv_request_sample(s->avctx, "Multiple geo key directories\n");
  1005. return AVERROR_INVALIDDATA;
  1006. }
  1007. ADD_METADATA(1, "GeoTIFF_Version", NULL);
  1008. ADD_METADATA(2, "GeoTIFF_Key_Revision", ".");
  1009. s->geotag_count = ff_tget_short(&s->gb, s->le);
  1010. if (s->geotag_count > count / 4 - 1) {
  1011. s->geotag_count = count / 4 - 1;
  1012. av_log(s->avctx, AV_LOG_WARNING, "GeoTIFF key directory buffer shorter than specified\n");
  1013. }
  1014. if ( bytestream2_get_bytes_left(&s->gb) < s->geotag_count * sizeof(int16_t) * 4
  1015. || s->geotag_count == 0) {
  1016. s->geotag_count = 0;
  1017. return -1;
  1018. }
  1019. s->geotags = av_mallocz_array(s->geotag_count, sizeof(TiffGeoTag));
  1020. if (!s->geotags) {
  1021. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1022. s->geotag_count = 0;
  1023. goto end;
  1024. }
  1025. for (i = 0; i < s->geotag_count; i++) {
  1026. s->geotags[i].key = ff_tget_short(&s->gb, s->le);
  1027. s->geotags[i].type = ff_tget_short(&s->gb, s->le);
  1028. s->geotags[i].count = ff_tget_short(&s->gb, s->le);
  1029. if (!s->geotags[i].type)
  1030. s->geotags[i].val = get_geokey_val(s->geotags[i].key, ff_tget_short(&s->gb, s->le));
  1031. else
  1032. s->geotags[i].offset = ff_tget_short(&s->gb, s->le);
  1033. }
  1034. break;
  1035. case TIFF_GEO_DOUBLE_PARAMS:
  1036. if (count >= INT_MAX / sizeof(int64_t))
  1037. return AVERROR_INVALIDDATA;
  1038. if (bytestream2_get_bytes_left(&s->gb) < count * sizeof(int64_t))
  1039. return AVERROR_INVALIDDATA;
  1040. dp = av_malloc_array(count, sizeof(double));
  1041. if (!dp) {
  1042. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1043. goto end;
  1044. }
  1045. for (i = 0; i < count; i++)
  1046. dp[i] = ff_tget_double(&s->gb, s->le);
  1047. for (i = 0; i < s->geotag_count; i++) {
  1048. if (s->geotags[i].type == TIFF_GEO_DOUBLE_PARAMS) {
  1049. if (s->geotags[i].count == 0
  1050. || s->geotags[i].offset + s->geotags[i].count > count) {
  1051. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1052. } else if (s->geotags[i].val) {
  1053. av_log(s->avctx, AV_LOG_WARNING, "Duplicate GeoTIFF key %d\n", s->geotags[i].key);
  1054. } else {
  1055. char *ap = doubles2str(&dp[s->geotags[i].offset], s->geotags[i].count, ", ");
  1056. if (!ap) {
  1057. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1058. av_freep(&dp);
  1059. return AVERROR(ENOMEM);
  1060. }
  1061. s->geotags[i].val = ap;
  1062. }
  1063. }
  1064. }
  1065. av_freep(&dp);
  1066. break;
  1067. case TIFF_GEO_ASCII_PARAMS:
  1068. pos = bytestream2_tell(&s->gb);
  1069. for (i = 0; i < s->geotag_count; i++) {
  1070. if (s->geotags[i].type == TIFF_GEO_ASCII_PARAMS) {
  1071. if (s->geotags[i].count == 0
  1072. || s->geotags[i].offset + s->geotags[i].count > count) {
  1073. av_log(s->avctx, AV_LOG_WARNING, "Invalid GeoTIFF key %d\n", s->geotags[i].key);
  1074. } else {
  1075. char *ap;
  1076. bytestream2_seek(&s->gb, pos + s->geotags[i].offset, SEEK_SET);
  1077. if (bytestream2_get_bytes_left(&s->gb) < s->geotags[i].count)
  1078. return AVERROR_INVALIDDATA;
  1079. if (s->geotags[i].val)
  1080. return AVERROR_INVALIDDATA;
  1081. ap = av_malloc(s->geotags[i].count);
  1082. if (!ap) {
  1083. av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
  1084. return AVERROR(ENOMEM);
  1085. }
  1086. bytestream2_get_bufferu(&s->gb, ap, s->geotags[i].count);
  1087. ap[s->geotags[i].count - 1] = '\0'; //replace the "|" delimiter with a 0 byte
  1088. s->geotags[i].val = ap;
  1089. }
  1090. }
  1091. }
  1092. break;
  1093. case TIFF_ARTIST:
  1094. ADD_METADATA(count, "artist", NULL);
  1095. break;
  1096. case TIFF_COPYRIGHT:
  1097. ADD_METADATA(count, "copyright", NULL);
  1098. break;
  1099. case TIFF_DATE:
  1100. ADD_METADATA(count, "date", NULL);
  1101. break;
  1102. case TIFF_DOCUMENT_NAME:
  1103. ADD_METADATA(count, "document_name", NULL);
  1104. break;
  1105. case TIFF_HOST_COMPUTER:
  1106. ADD_METADATA(count, "computer", NULL);
  1107. break;
  1108. case TIFF_IMAGE_DESCRIPTION:
  1109. ADD_METADATA(count, "description", NULL);
  1110. break;
  1111. case TIFF_MAKE:
  1112. ADD_METADATA(count, "make", NULL);
  1113. break;
  1114. case TIFF_MODEL:
  1115. ADD_METADATA(count, "model", NULL);
  1116. break;
  1117. case TIFF_PAGE_NAME:
  1118. ADD_METADATA(count, "page_name", NULL);
  1119. break;
  1120. case TIFF_PAGE_NUMBER:
  1121. ADD_METADATA(count, "page_number", " / ");
  1122. break;
  1123. case TIFF_SOFTWARE_NAME:
  1124. ADD_METADATA(count, "software", NULL);
  1125. break;
  1126. default:
  1127. if (s->avctx->err_recognition & AV_EF_EXPLODE) {
  1128. av_log(s->avctx, AV_LOG_ERROR,
  1129. "Unknown or unsupported tag %d/0X%0X\n",
  1130. tag, tag);
  1131. return AVERROR_INVALIDDATA;
  1132. }
  1133. }
  1134. end:
  1135. if (s->bpp > 64U) {
  1136. av_log(s->avctx, AV_LOG_ERROR,
  1137. "This format is not supported (bpp=%d, %d components)\n",
  1138. s->bpp, count);
  1139. s->bpp = 0;
  1140. return AVERROR_INVALIDDATA;
  1141. }
  1142. bytestream2_seek(&s->gb, start, SEEK_SET);
  1143. return 0;
  1144. }
  1145. static int decode_frame(AVCodecContext *avctx,
  1146. void *data, int *got_frame, AVPacket *avpkt)
  1147. {
  1148. TiffContext *const s = avctx->priv_data;
  1149. AVFrame *const p = data;
  1150. ThreadFrame frame = { .f = data };
  1151. unsigned off;
  1152. int le, ret, plane, planes;
  1153. int i, j, entries, stride;
  1154. unsigned soff, ssize;
  1155. uint8_t *dst;
  1156. GetByteContext stripsizes;
  1157. GetByteContext stripdata;
  1158. bytestream2_init(&s->gb, avpkt->data, avpkt->size);
  1159. // parse image header
  1160. if ((ret = ff_tdecode_header(&s->gb, &le, &off))) {
  1161. av_log(avctx, AV_LOG_ERROR, "Invalid TIFF header\n");
  1162. return ret;
  1163. } else if (off >= UINT_MAX - 14 || avpkt->size < off + 14) {
  1164. av_log(avctx, AV_LOG_ERROR, "IFD offset is greater than image size\n");
  1165. return AVERROR_INVALIDDATA;
  1166. }
  1167. s->le = le;
  1168. // TIFF_BPP is not a required tag and defaults to 1
  1169. s->bppcount = s->bpp = 1;
  1170. s->photometric = TIFF_PHOTOMETRIC_NONE;
  1171. s->compr = TIFF_RAW;
  1172. s->fill_order = 0;
  1173. free_geotags(s);
  1174. // Reset these offsets so we can tell if they were set this frame
  1175. s->stripsizesoff = s->strippos = 0;
  1176. /* parse image file directory */
  1177. bytestream2_seek(&s->gb, off, SEEK_SET);
  1178. entries = ff_tget_short(&s->gb, le);
  1179. if (bytestream2_get_bytes_left(&s->gb) < entries * 12)
  1180. return AVERROR_INVALIDDATA;
  1181. for (i = 0; i < entries; i++) {
  1182. if ((ret = tiff_decode_tag(s, p)) < 0)
  1183. return ret;
  1184. }
  1185. for (i = 0; i<s->geotag_count; i++) {
  1186. const char *keyname = get_geokey_name(s->geotags[i].key);
  1187. if (!keyname) {
  1188. av_log(avctx, AV_LOG_WARNING, "Unknown or unsupported GeoTIFF key %d\n", s->geotags[i].key);
  1189. continue;
  1190. }
  1191. if (get_geokey_type(s->geotags[i].key) != s->geotags[i].type) {
  1192. av_log(avctx, AV_LOG_WARNING, "Type of GeoTIFF key %d is wrong\n", s->geotags[i].key);
  1193. continue;
  1194. }
  1195. ret = av_dict_set(&p->metadata, keyname, s->geotags[i].val, 0);
  1196. if (ret<0) {
  1197. av_log(avctx, AV_LOG_ERROR, "Writing metadata with key '%s' failed\n", keyname);
  1198. return ret;
  1199. }
  1200. }
  1201. if (!s->strippos && !s->stripoff) {
  1202. av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
  1203. return AVERROR_INVALIDDATA;
  1204. }
  1205. /* now we have the data and may start decoding */
  1206. if ((ret = init_image(s, &frame)) < 0)
  1207. return ret;
  1208. if (s->strips == 1 && !s->stripsize) {
  1209. av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
  1210. s->stripsize = avpkt->size - s->stripoff;
  1211. }
  1212. if (s->stripsizesoff) {
  1213. if (s->stripsizesoff >= (unsigned)avpkt->size)
  1214. return AVERROR_INVALIDDATA;
  1215. bytestream2_init(&stripsizes, avpkt->data + s->stripsizesoff,
  1216. avpkt->size - s->stripsizesoff);
  1217. }
  1218. if (s->strippos) {
  1219. if (s->strippos >= (unsigned)avpkt->size)
  1220. return AVERROR_INVALIDDATA;
  1221. bytestream2_init(&stripdata, avpkt->data + s->strippos,
  1222. avpkt->size - s->strippos);
  1223. }
  1224. if (s->rps <= 0 || s->rps % s->subsampling[1]) {
  1225. av_log(avctx, AV_LOG_ERROR, "rps %d invalid\n", s->rps);
  1226. return AVERROR_INVALIDDATA;
  1227. }
  1228. planes = s->planar ? s->bppcount : 1;
  1229. for (plane = 0; plane < planes; plane++) {
  1230. int remaining = avpkt->size;
  1231. int decoded_height;
  1232. stride = p->linesize[plane];
  1233. dst = p->data[plane];
  1234. for (i = 0; i < s->height; i += s->rps) {
  1235. if (i)
  1236. dst += s->rps * stride;
  1237. if (s->stripsizesoff)
  1238. ssize = ff_tget(&stripsizes, s->sstype, le);
  1239. else
  1240. ssize = s->stripsize;
  1241. if (s->strippos)
  1242. soff = ff_tget(&stripdata, s->sot, le);
  1243. else
  1244. soff = s->stripoff;
  1245. if (soff > avpkt->size || ssize > avpkt->size - soff || ssize > remaining) {
  1246. av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
  1247. return AVERROR_INVALIDDATA;
  1248. }
  1249. remaining -= ssize;
  1250. if ((ret = tiff_unpack_strip(s, p, dst, stride, avpkt->data + soff, ssize, i,
  1251. FFMIN(s->rps, s->height - i))) < 0) {
  1252. if (avctx->err_recognition & AV_EF_EXPLODE)
  1253. return ret;
  1254. break;
  1255. }
  1256. }
  1257. decoded_height = FFMIN(i, s->height);
  1258. if (s->predictor == 2) {
  1259. if (s->photometric == TIFF_PHOTOMETRIC_YCBCR) {
  1260. av_log(s->avctx, AV_LOG_ERROR, "predictor == 2 with YUV is unsupported");
  1261. return AVERROR_PATCHWELCOME;
  1262. }
  1263. dst = p->data[plane];
  1264. soff = s->bpp >> 3;
  1265. if (s->planar)
  1266. soff = FFMAX(soff / s->bppcount, 1);
  1267. ssize = s->width * soff;
  1268. if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48LE ||
  1269. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64LE ||
  1270. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16LE ||
  1271. s->avctx->pix_fmt == AV_PIX_FMT_YA16LE ||
  1272. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16LE ||
  1273. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16LE) {
  1274. for (i = 0; i < decoded_height; i++) {
  1275. for (j = soff; j < ssize; j += 2)
  1276. AV_WL16(dst + j, AV_RL16(dst + j) + AV_RL16(dst + j - soff));
  1277. dst += stride;
  1278. }
  1279. } else if (s->avctx->pix_fmt == AV_PIX_FMT_RGB48BE ||
  1280. s->avctx->pix_fmt == AV_PIX_FMT_RGBA64BE ||
  1281. s->avctx->pix_fmt == AV_PIX_FMT_GRAY16BE ||
  1282. s->avctx->pix_fmt == AV_PIX_FMT_YA16BE ||
  1283. s->avctx->pix_fmt == AV_PIX_FMT_GBRP16BE ||
  1284. s->avctx->pix_fmt == AV_PIX_FMT_GBRAP16BE) {
  1285. for (i = 0; i < decoded_height; i++) {
  1286. for (j = soff; j < ssize; j += 2)
  1287. AV_WB16(dst + j, AV_RB16(dst + j) + AV_RB16(dst + j - soff));
  1288. dst += stride;
  1289. }
  1290. } else {
  1291. for (i = 0; i < decoded_height; i++) {
  1292. for (j = soff; j < ssize; j++)
  1293. dst[j] += dst[j - soff];
  1294. dst += stride;
  1295. }
  1296. }
  1297. }
  1298. if (s->photometric == TIFF_PHOTOMETRIC_WHITE_IS_ZERO) {
  1299. int c = (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255);
  1300. dst = p->data[plane];
  1301. for (i = 0; i < s->height; i++) {
  1302. for (j = 0; j < stride; j++)
  1303. dst[j] = c - dst[j];
  1304. dst += stride;
  1305. }
  1306. }
  1307. }
  1308. if (s->planar && s->bppcount > 2) {
  1309. FFSWAP(uint8_t*, p->data[0], p->data[2]);
  1310. FFSWAP(int, p->linesize[0], p->linesize[2]);
  1311. FFSWAP(uint8_t*, p->data[0], p->data[1]);
  1312. FFSWAP(int, p->linesize[0], p->linesize[1]);
  1313. }
  1314. *got_frame = 1;
  1315. return avpkt->size;
  1316. }
  1317. static av_cold int tiff_init(AVCodecContext *avctx)
  1318. {
  1319. TiffContext *s = avctx->priv_data;
  1320. s->width = 0;
  1321. s->height = 0;
  1322. s->subsampling[0] =
  1323. s->subsampling[1] = 1;
  1324. s->avctx = avctx;
  1325. ff_lzw_decode_open(&s->lzw);
  1326. if (!s->lzw)
  1327. return AVERROR(ENOMEM);
  1328. ff_ccitt_unpack_init();
  1329. return 0;
  1330. }
  1331. static av_cold int tiff_end(AVCodecContext *avctx)
  1332. {
  1333. TiffContext *const s = avctx->priv_data;
  1334. free_geotags(s);
  1335. ff_lzw_decode_close(&s->lzw);
  1336. av_freep(&s->deinvert_buf);
  1337. s->deinvert_buf_size = 0;
  1338. av_freep(&s->fax_buffer);
  1339. s->fax_buffer_size = 0;
  1340. return 0;
  1341. }
  1342. AVCodec ff_tiff_decoder = {
  1343. .name = "tiff",
  1344. .long_name = NULL_IF_CONFIG_SMALL("TIFF image"),
  1345. .type = AVMEDIA_TYPE_VIDEO,
  1346. .id = AV_CODEC_ID_TIFF,
  1347. .priv_data_size = sizeof(TiffContext),
  1348. .init = tiff_init,
  1349. .close = tiff_end,
  1350. .decode = decode_frame,
  1351. .init_thread_copy = ONLY_IF_THREADS_ENABLED(tiff_init),
  1352. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
  1353. };