You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1699 lines
58KB

  1. /*
  2. * MJPEG decoder
  3. * Copyright (c) 2000, 2001 Fabrice Bellard
  4. * Copyright (c) 2003 Alex Beregszaszi
  5. * Copyright (c) 2003-2004 Michael Niedermayer
  6. *
  7. * Support for external huffman table, various fixes (AVID workaround),
  8. * aspecting, new decode_frame mechanism and apple mjpeg-b support
  9. * by Alex Beregszaszi
  10. *
  11. * This file is part of Libav.
  12. *
  13. * Libav is free software; you can redistribute it and/or
  14. * modify it under the terms of the GNU Lesser General Public
  15. * License as published by the Free Software Foundation; either
  16. * version 2.1 of the License, or (at your option) any later version.
  17. *
  18. * Libav is distributed in the hope that it will be useful,
  19. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  20. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  21. * Lesser General Public License for more details.
  22. *
  23. * You should have received a copy of the GNU Lesser General Public
  24. * License along with Libav; if not, write to the Free Software
  25. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  26. */
  27. /**
  28. * @file
  29. * MJPEG decoder.
  30. */
  31. // #define DEBUG
  32. #include <assert.h>
  33. #include "libavutil/imgutils.h"
  34. #include "libavutil/opt.h"
  35. #include "avcodec.h"
  36. #include "internal.h"
  37. #include "mjpeg.h"
  38. #include "mjpegdec.h"
  39. #include "jpeglsdec.h"
  40. static int build_vlc(VLC *vlc, const uint8_t *bits_table,
  41. const uint8_t *val_table, int nb_codes,
  42. int use_static, int is_ac)
  43. {
  44. uint8_t huff_size[256] = { 0 };
  45. uint16_t huff_code[256];
  46. uint16_t huff_sym[256];
  47. int i;
  48. assert(nb_codes <= 256);
  49. ff_mjpeg_build_huffman_codes(huff_size, huff_code, bits_table, val_table);
  50. for (i = 0; i < 256; i++)
  51. huff_sym[i] = i + 16 * is_ac;
  52. if (is_ac)
  53. huff_sym[0] = 16 * 256;
  54. return ff_init_vlc_sparse(vlc, 9, nb_codes, huff_size, 1, 1,
  55. huff_code, 2, 2, huff_sym, 2, 2, use_static);
  56. }
  57. static void build_basic_mjpeg_vlc(MJpegDecodeContext *s)
  58. {
  59. build_vlc(&s->vlcs[0][0], avpriv_mjpeg_bits_dc_luminance,
  60. avpriv_mjpeg_val_dc, 12, 0, 0);
  61. build_vlc(&s->vlcs[0][1], avpriv_mjpeg_bits_dc_chrominance,
  62. avpriv_mjpeg_val_dc, 12, 0, 0);
  63. build_vlc(&s->vlcs[1][0], avpriv_mjpeg_bits_ac_luminance,
  64. avpriv_mjpeg_val_ac_luminance, 251, 0, 1);
  65. build_vlc(&s->vlcs[1][1], avpriv_mjpeg_bits_ac_chrominance,
  66. avpriv_mjpeg_val_ac_chrominance, 251, 0, 1);
  67. build_vlc(&s->vlcs[2][0], avpriv_mjpeg_bits_ac_luminance,
  68. avpriv_mjpeg_val_ac_luminance, 251, 0, 0);
  69. build_vlc(&s->vlcs[2][1], avpriv_mjpeg_bits_ac_chrominance,
  70. avpriv_mjpeg_val_ac_chrominance, 251, 0, 0);
  71. }
  72. av_cold int ff_mjpeg_decode_init(AVCodecContext *avctx)
  73. {
  74. MJpegDecodeContext *s = avctx->priv_data;
  75. if (!s->picture_ptr)
  76. s->picture_ptr = &s->picture;
  77. s->avctx = avctx;
  78. ff_hpeldsp_init(&s->hdsp, avctx->flags);
  79. ff_dsputil_init(&s->dsp, avctx);
  80. ff_init_scantable(s->dsp.idct_permutation, &s->scantable, ff_zigzag_direct);
  81. s->buffer_size = 0;
  82. s->buffer = NULL;
  83. s->start_code = -1;
  84. s->first_picture = 1;
  85. s->org_height = avctx->coded_height;
  86. avctx->chroma_sample_location = AVCHROMA_LOC_CENTER;
  87. build_basic_mjpeg_vlc(s);
  88. if (s->extern_huff) {
  89. int ret;
  90. av_log(avctx, AV_LOG_INFO, "mjpeg: using external huffman table\n");
  91. init_get_bits(&s->gb, avctx->extradata, avctx->extradata_size * 8);
  92. if ((ret = ff_mjpeg_decode_dht(s))) {
  93. av_log(avctx, AV_LOG_ERROR,
  94. "mjpeg: error using external huffman table\n");
  95. return ret;
  96. }
  97. }
  98. if (avctx->field_order == AV_FIELD_BB) { /* quicktime icefloe 019 */
  99. s->interlace_polarity = 1; /* bottom field first */
  100. av_log(avctx, AV_LOG_DEBUG, "mjpeg bottom field first\n");
  101. }
  102. if (avctx->codec->id == AV_CODEC_ID_AMV)
  103. s->flipped = 1;
  104. return 0;
  105. }
  106. /* quantize tables */
  107. int ff_mjpeg_decode_dqt(MJpegDecodeContext *s)
  108. {
  109. int len, index, i, j;
  110. len = get_bits(&s->gb, 16) - 2;
  111. while (len >= 65) {
  112. /* only 8 bit precision handled */
  113. if (get_bits(&s->gb, 4) != 0) {
  114. av_log(s->avctx, AV_LOG_ERROR, "dqt: 16bit precision\n");
  115. return -1;
  116. }
  117. index = get_bits(&s->gb, 4);
  118. if (index >= 4)
  119. return -1;
  120. av_log(s->avctx, AV_LOG_DEBUG, "index=%d\n", index);
  121. /* read quant table */
  122. for (i = 0; i < 64; i++) {
  123. j = s->scantable.permutated[i];
  124. s->quant_matrixes[index][j] = get_bits(&s->gb, 8);
  125. }
  126. // XXX FIXME finetune, and perhaps add dc too
  127. s->qscale[index] = FFMAX(s->quant_matrixes[index][s->scantable.permutated[1]],
  128. s->quant_matrixes[index][s->scantable.permutated[8]]) >> 1;
  129. av_log(s->avctx, AV_LOG_DEBUG, "qscale[%d]: %d\n",
  130. index, s->qscale[index]);
  131. len -= 65;
  132. }
  133. return 0;
  134. }
  135. /* decode huffman tables and build VLC decoders */
  136. int ff_mjpeg_decode_dht(MJpegDecodeContext *s)
  137. {
  138. int len, index, i, class, n, v, code_max;
  139. uint8_t bits_table[17];
  140. uint8_t val_table[256];
  141. int ret = 0;
  142. len = get_bits(&s->gb, 16) - 2;
  143. while (len > 0) {
  144. if (len < 17)
  145. return AVERROR_INVALIDDATA;
  146. class = get_bits(&s->gb, 4);
  147. if (class >= 2)
  148. return AVERROR_INVALIDDATA;
  149. index = get_bits(&s->gb, 4);
  150. if (index >= 4)
  151. return AVERROR_INVALIDDATA;
  152. n = 0;
  153. for (i = 1; i <= 16; i++) {
  154. bits_table[i] = get_bits(&s->gb, 8);
  155. n += bits_table[i];
  156. }
  157. len -= 17;
  158. if (len < n || n > 256)
  159. return AVERROR_INVALIDDATA;
  160. code_max = 0;
  161. for (i = 0; i < n; i++) {
  162. v = get_bits(&s->gb, 8);
  163. if (v > code_max)
  164. code_max = v;
  165. val_table[i] = v;
  166. }
  167. len -= n;
  168. /* build VLC and flush previous vlc if present */
  169. ff_free_vlc(&s->vlcs[class][index]);
  170. av_log(s->avctx, AV_LOG_DEBUG, "class=%d index=%d nb_codes=%d\n",
  171. class, index, code_max + 1);
  172. if ((ret = build_vlc(&s->vlcs[class][index], bits_table, val_table,
  173. code_max + 1, 0, class > 0)) < 0)
  174. return ret;
  175. if (class > 0) {
  176. ff_free_vlc(&s->vlcs[2][index]);
  177. if ((ret = build_vlc(&s->vlcs[2][index], bits_table, val_table,
  178. code_max + 1, 0, 0)) < 0)
  179. return ret;
  180. }
  181. }
  182. return 0;
  183. }
  184. int ff_mjpeg_decode_sof(MJpegDecodeContext *s)
  185. {
  186. int len, nb_components, i, width, height, pix_fmt_id;
  187. /* XXX: verify len field validity */
  188. len = get_bits(&s->gb, 16);
  189. s->bits = get_bits(&s->gb, 8);
  190. if (s->pegasus_rct)
  191. s->bits = 9;
  192. if (s->bits == 9 && !s->pegasus_rct)
  193. s->rct = 1; // FIXME ugly
  194. if (s->bits != 8 && !s->lossless) {
  195. av_log(s->avctx, AV_LOG_ERROR, "only 8 bits/component accepted\n");
  196. return -1;
  197. }
  198. height = get_bits(&s->gb, 16);
  199. width = get_bits(&s->gb, 16);
  200. // HACK for odd_height.mov
  201. if (s->interlaced && s->width == width && s->height == height + 1)
  202. height= s->height;
  203. av_log(s->avctx, AV_LOG_DEBUG, "sof0: picture: %dx%d\n", width, height);
  204. if (av_image_check_size(width, height, 0, s->avctx))
  205. return AVERROR_INVALIDDATA;
  206. nb_components = get_bits(&s->gb, 8);
  207. if (nb_components <= 0 ||
  208. nb_components > MAX_COMPONENTS)
  209. return -1;
  210. if (s->ls && !(s->bits <= 8 || nb_components == 1)) {
  211. avpriv_report_missing_feature(s->avctx,
  212. "JPEG-LS that is not <= 8 "
  213. "bits/component or 16-bit gray");
  214. return AVERROR_PATCHWELCOME;
  215. }
  216. s->nb_components = nb_components;
  217. s->h_max = 1;
  218. s->v_max = 1;
  219. for (i = 0; i < nb_components; i++) {
  220. /* component id */
  221. s->component_id[i] = get_bits(&s->gb, 8) - 1;
  222. s->h_count[i] = get_bits(&s->gb, 4);
  223. s->v_count[i] = get_bits(&s->gb, 4);
  224. /* compute hmax and vmax (only used in interleaved case) */
  225. if (s->h_count[i] > s->h_max)
  226. s->h_max = s->h_count[i];
  227. if (s->v_count[i] > s->v_max)
  228. s->v_max = s->v_count[i];
  229. s->quant_index[i] = get_bits(&s->gb, 8);
  230. if (s->quant_index[i] >= 4)
  231. return AVERROR_INVALIDDATA;
  232. if (!s->h_count[i] || !s->v_count[i]) {
  233. av_log(s->avctx, AV_LOG_ERROR,
  234. "Invalid sampling factor in component %d %d:%d\n",
  235. i, s->h_count[i], s->v_count[i]);
  236. return AVERROR_INVALIDDATA;
  237. }
  238. av_log(s->avctx, AV_LOG_DEBUG, "component %d %d:%d id: %d quant:%d\n",
  239. i, s->h_count[i], s->v_count[i],
  240. s->component_id[i], s->quant_index[i]);
  241. }
  242. if (s->ls && (s->h_max > 1 || s->v_max > 1)) {
  243. avpriv_report_missing_feature(s->avctx, "Subsampling in JPEG-LS");
  244. return AVERROR_PATCHWELCOME;
  245. }
  246. if (s->v_max == 1 && s->h_max == 1 && s->lossless == 1)
  247. s->rgb = 1;
  248. /* if different size, realloc/alloc picture */
  249. /* XXX: also check h_count and v_count */
  250. if (width != s->width || height != s->height) {
  251. s->width = width;
  252. s->height = height;
  253. s->interlaced = 0;
  254. /* test interlaced mode */
  255. if (s->first_picture &&
  256. s->org_height != 0 &&
  257. s->height < ((s->org_height * 3) / 4)) {
  258. s->interlaced = 1;
  259. s->bottom_field = s->interlace_polarity;
  260. s->picture_ptr->interlaced_frame = 1;
  261. s->picture_ptr->top_field_first = !s->interlace_polarity;
  262. height *= 2;
  263. }
  264. avcodec_set_dimensions(s->avctx, width, height);
  265. s->first_picture = 0;
  266. }
  267. if (!(s->interlaced && (s->bottom_field == !s->interlace_polarity))) {
  268. /* XXX: not complete test ! */
  269. pix_fmt_id = (s->h_count[0] << 28) | (s->v_count[0] << 24) |
  270. (s->h_count[1] << 20) | (s->v_count[1] << 16) |
  271. (s->h_count[2] << 12) | (s->v_count[2] << 8) |
  272. (s->h_count[3] << 4) | s->v_count[3];
  273. av_log(s->avctx, AV_LOG_DEBUG, "pix fmt id %x\n", pix_fmt_id);
  274. /* NOTE we do not allocate pictures large enough for the possible
  275. * padding of h/v_count being 4 */
  276. if (!(pix_fmt_id & 0xD0D0D0D0))
  277. pix_fmt_id -= (pix_fmt_id & 0xF0F0F0F0) >> 1;
  278. if (!(pix_fmt_id & 0x0D0D0D0D))
  279. pix_fmt_id -= (pix_fmt_id & 0x0F0F0F0F) >> 1;
  280. switch (pix_fmt_id) {
  281. case 0x11111100:
  282. if (s->rgb)
  283. s->avctx->pix_fmt = AV_PIX_FMT_BGRA;
  284. else
  285. s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV444P : AV_PIX_FMT_YUVJ444P;
  286. assert(s->nb_components == 3);
  287. break;
  288. case 0x11000000:
  289. s->avctx->pix_fmt = AV_PIX_FMT_GRAY8;
  290. break;
  291. case 0x12111100:
  292. s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV440P : AV_PIX_FMT_YUVJ440P;
  293. break;
  294. case 0x21111100:
  295. s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV422P : AV_PIX_FMT_YUVJ422P;
  296. break;
  297. case 0x22111100:
  298. s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_YUVJ420P;
  299. break;
  300. default:
  301. av_log(s->avctx, AV_LOG_ERROR, "Unhandled pixel format 0x%x\n", pix_fmt_id);
  302. return AVERROR_PATCHWELCOME;
  303. }
  304. if (s->ls) {
  305. if (s->nb_components > 1)
  306. s->avctx->pix_fmt = AV_PIX_FMT_RGB24;
  307. else if (s->bits <= 8)
  308. s->avctx->pix_fmt = AV_PIX_FMT_GRAY8;
  309. else
  310. s->avctx->pix_fmt = AV_PIX_FMT_GRAY16;
  311. }
  312. av_frame_unref(s->picture_ptr);
  313. if (ff_get_buffer(s->avctx, s->picture_ptr, AV_GET_BUFFER_FLAG_REF) < 0) {
  314. av_log(s->avctx, AV_LOG_ERROR, "get_buffer() failed\n");
  315. return -1;
  316. }
  317. s->picture_ptr->pict_type = AV_PICTURE_TYPE_I;
  318. s->picture_ptr->key_frame = 1;
  319. s->got_picture = 1;
  320. for (i = 0; i < 3; i++)
  321. s->linesize[i] = s->picture_ptr->linesize[i] << s->interlaced;
  322. av_dlog(s->avctx, "%d %d %d %d %d %d\n",
  323. s->width, s->height, s->linesize[0], s->linesize[1],
  324. s->interlaced, s->avctx->height);
  325. if (len != (8 + (3 * nb_components)))
  326. av_log(s->avctx, AV_LOG_DEBUG, "decode_sof0: error, len(%d) mismatch\n", len);
  327. }
  328. /* totally blank picture as progressive JPEG will only add details to it */
  329. if (s->progressive) {
  330. int bw = (width + s->h_max * 8 - 1) / (s->h_max * 8);
  331. int bh = (height + s->v_max * 8 - 1) / (s->v_max * 8);
  332. for (i = 0; i < s->nb_components; i++) {
  333. int size = bw * bh * s->h_count[i] * s->v_count[i];
  334. av_freep(&s->blocks[i]);
  335. av_freep(&s->last_nnz[i]);
  336. s->blocks[i] = av_malloc(size * sizeof(**s->blocks));
  337. s->last_nnz[i] = av_mallocz(size * sizeof(**s->last_nnz));
  338. s->block_stride[i] = bw * s->h_count[i];
  339. }
  340. memset(s->coefs_finished, 0, sizeof(s->coefs_finished));
  341. }
  342. return 0;
  343. }
  344. static inline int mjpeg_decode_dc(MJpegDecodeContext *s, int dc_index)
  345. {
  346. int code;
  347. code = get_vlc2(&s->gb, s->vlcs[0][dc_index].table, 9, 2);
  348. if (code < 0) {
  349. av_log(s->avctx, AV_LOG_WARNING,
  350. "mjpeg_decode_dc: bad vlc: %d:%d (%p)\n",
  351. 0, dc_index, &s->vlcs[0][dc_index]);
  352. return 0xffff;
  353. }
  354. if (code)
  355. return get_xbits(&s->gb, code);
  356. else
  357. return 0;
  358. }
  359. /* decode block and dequantize */
  360. static int decode_block(MJpegDecodeContext *s, int16_t *block, int component,
  361. int dc_index, int ac_index, int16_t *quant_matrix)
  362. {
  363. int code, i, j, level, val;
  364. /* DC coef */
  365. val = mjpeg_decode_dc(s, dc_index);
  366. if (val == 0xffff) {
  367. av_log(s->avctx, AV_LOG_ERROR, "error dc\n");
  368. return AVERROR_INVALIDDATA;
  369. }
  370. val = val * quant_matrix[0] + s->last_dc[component];
  371. s->last_dc[component] = val;
  372. block[0] = val;
  373. /* AC coefs */
  374. i = 0;
  375. {OPEN_READER(re, &s->gb);
  376. do {
  377. UPDATE_CACHE(re, &s->gb);
  378. GET_VLC(code, re, &s->gb, s->vlcs[1][ac_index].table, 9, 2);
  379. i += ((unsigned)code) >> 4;
  380. code &= 0xf;
  381. if (code) {
  382. if (code > MIN_CACHE_BITS - 16)
  383. UPDATE_CACHE(re, &s->gb);
  384. {
  385. int cache = GET_CACHE(re, &s->gb);
  386. int sign = (~cache) >> 31;
  387. level = (NEG_USR32(sign ^ cache,code) ^ sign) - sign;
  388. }
  389. LAST_SKIP_BITS(re, &s->gb, code);
  390. if (i > 63) {
  391. av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i);
  392. return AVERROR_INVALIDDATA;
  393. }
  394. j = s->scantable.permutated[i];
  395. block[j] = level * quant_matrix[j];
  396. }
  397. } while (i < 63);
  398. CLOSE_READER(re, &s->gb);}
  399. return 0;
  400. }
  401. static int decode_dc_progressive(MJpegDecodeContext *s, int16_t *block,
  402. int component, int dc_index,
  403. int16_t *quant_matrix, int Al)
  404. {
  405. int val;
  406. s->dsp.clear_block(block);
  407. val = mjpeg_decode_dc(s, dc_index);
  408. if (val == 0xffff) {
  409. av_log(s->avctx, AV_LOG_ERROR, "error dc\n");
  410. return AVERROR_INVALIDDATA;
  411. }
  412. val = (val * quant_matrix[0] << Al) + s->last_dc[component];
  413. s->last_dc[component] = val;
  414. block[0] = val;
  415. return 0;
  416. }
  417. /* decode block and dequantize - progressive JPEG version */
  418. static int decode_block_progressive(MJpegDecodeContext *s, int16_t *block,
  419. uint8_t *last_nnz, int ac_index,
  420. int16_t *quant_matrix,
  421. int ss, int se, int Al, int *EOBRUN)
  422. {
  423. int code, i, j, level, val, run;
  424. if (*EOBRUN) {
  425. (*EOBRUN)--;
  426. return 0;
  427. }
  428. {
  429. OPEN_READER(re, &s->gb);
  430. for (i = ss; ; i++) {
  431. UPDATE_CACHE(re, &s->gb);
  432. GET_VLC(code, re, &s->gb, s->vlcs[2][ac_index].table, 9, 2);
  433. run = ((unsigned) code) >> 4;
  434. code &= 0xF;
  435. if (code) {
  436. i += run;
  437. if (code > MIN_CACHE_BITS - 16)
  438. UPDATE_CACHE(re, &s->gb);
  439. {
  440. int cache = GET_CACHE(re, &s->gb);
  441. int sign = (~cache) >> 31;
  442. level = (NEG_USR32(sign ^ cache,code) ^ sign) - sign;
  443. }
  444. LAST_SKIP_BITS(re, &s->gb, code);
  445. if (i >= se) {
  446. if (i == se) {
  447. j = s->scantable.permutated[se];
  448. block[j] = level * quant_matrix[j] << Al;
  449. break;
  450. }
  451. av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i);
  452. return AVERROR_INVALIDDATA;
  453. }
  454. j = s->scantable.permutated[i];
  455. block[j] = level * quant_matrix[j] << Al;
  456. } else {
  457. if (run == 0xF) {// ZRL - skip 15 coefficients
  458. i += 15;
  459. if (i >= se) {
  460. av_log(s->avctx, AV_LOG_ERROR, "ZRL overflow: %d\n", i);
  461. return AVERROR_INVALIDDATA;
  462. }
  463. } else {
  464. val = (1 << run);
  465. if (run) {
  466. UPDATE_CACHE(re, &s->gb);
  467. val += NEG_USR32(GET_CACHE(re, &s->gb), run);
  468. LAST_SKIP_BITS(re, &s->gb, run);
  469. }
  470. *EOBRUN = val - 1;
  471. break;
  472. }
  473. }
  474. }
  475. CLOSE_READER(re, &s->gb);
  476. }
  477. if (i > *last_nnz)
  478. *last_nnz = i;
  479. return 0;
  480. }
  481. #define REFINE_BIT(j) { \
  482. UPDATE_CACHE(re, &s->gb); \
  483. sign = block[j] >> 15; \
  484. block[j] += SHOW_UBITS(re, &s->gb, 1) * \
  485. ((quant_matrix[j] ^ sign) - sign) << Al; \
  486. LAST_SKIP_BITS(re, &s->gb, 1); \
  487. }
  488. #define ZERO_RUN \
  489. for (; ; i++) { \
  490. if (i > last) { \
  491. i += run; \
  492. if (i > se) { \
  493. av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i); \
  494. return -1; \
  495. } \
  496. break; \
  497. } \
  498. j = s->scantable.permutated[i]; \
  499. if (block[j]) \
  500. REFINE_BIT(j) \
  501. else if (run-- == 0) \
  502. break; \
  503. }
  504. /* decode block and dequantize - progressive JPEG refinement pass */
  505. static int decode_block_refinement(MJpegDecodeContext *s, int16_t *block,
  506. uint8_t *last_nnz,
  507. int ac_index, int16_t *quant_matrix,
  508. int ss, int se, int Al, int *EOBRUN)
  509. {
  510. int code, i = ss, j, sign, val, run;
  511. int last = FFMIN(se, *last_nnz);
  512. OPEN_READER(re, &s->gb);
  513. if (*EOBRUN) {
  514. (*EOBRUN)--;
  515. } else {
  516. for (; ; i++) {
  517. UPDATE_CACHE(re, &s->gb);
  518. GET_VLC(code, re, &s->gb, s->vlcs[2][ac_index].table, 9, 2);
  519. if (code & 0xF) {
  520. run = ((unsigned) code) >> 4;
  521. UPDATE_CACHE(re, &s->gb);
  522. val = SHOW_UBITS(re, &s->gb, 1);
  523. LAST_SKIP_BITS(re, &s->gb, 1);
  524. ZERO_RUN;
  525. j = s->scantable.permutated[i];
  526. val--;
  527. block[j] = ((quant_matrix[j]^val) - val) << Al;
  528. if (i == se) {
  529. if (i > *last_nnz)
  530. *last_nnz = i;
  531. CLOSE_READER(re, &s->gb);
  532. return 0;
  533. }
  534. } else {
  535. run = ((unsigned) code) >> 4;
  536. if (run == 0xF) {
  537. ZERO_RUN;
  538. } else {
  539. val = run;
  540. run = (1 << run);
  541. if (val) {
  542. UPDATE_CACHE(re, &s->gb);
  543. run += SHOW_UBITS(re, &s->gb, val);
  544. LAST_SKIP_BITS(re, &s->gb, val);
  545. }
  546. *EOBRUN = run - 1;
  547. break;
  548. }
  549. }
  550. }
  551. if (i > *last_nnz)
  552. *last_nnz = i;
  553. }
  554. for (; i <= last; i++) {
  555. j = s->scantable.permutated[i];
  556. if (block[j])
  557. REFINE_BIT(j)
  558. }
  559. CLOSE_READER(re, &s->gb);
  560. return 0;
  561. }
  562. #undef REFINE_BIT
  563. #undef ZERO_RUN
  564. static int ljpeg_decode_rgb_scan(MJpegDecodeContext *s, int predictor,
  565. int point_transform)
  566. {
  567. int i, mb_x, mb_y;
  568. uint16_t (*buffer)[4];
  569. int left[3], top[3], topleft[3];
  570. const int linesize = s->linesize[0];
  571. const int mask = (1 << s->bits) - 1;
  572. av_fast_malloc(&s->ljpeg_buffer, &s->ljpeg_buffer_size,
  573. (unsigned)s->mb_width * 4 * sizeof(s->ljpeg_buffer[0][0]));
  574. buffer = s->ljpeg_buffer;
  575. for (i = 0; i < 3; i++)
  576. buffer[0][i] = 1 << (s->bits + point_transform - 1);
  577. for (mb_y = 0; mb_y < s->mb_height; mb_y++) {
  578. const int modified_predictor = mb_y ? predictor : 1;
  579. uint8_t *ptr = s->picture_ptr->data[0] + (linesize * mb_y);
  580. if (s->interlaced && s->bottom_field)
  581. ptr += linesize >> 1;
  582. for (i = 0; i < 3; i++)
  583. top[i] = left[i] = topleft[i] = buffer[0][i];
  584. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  585. if (s->restart_interval && !s->restart_count)
  586. s->restart_count = s->restart_interval;
  587. for (i = 0; i < 3; i++) {
  588. int pred;
  589. topleft[i] = top[i];
  590. top[i] = buffer[mb_x][i];
  591. PREDICT(pred, topleft[i], top[i], left[i], modified_predictor);
  592. left[i] = buffer[mb_x][i] =
  593. mask & (pred + (mjpeg_decode_dc(s, s->dc_index[i]) << point_transform));
  594. }
  595. if (s->restart_interval && !--s->restart_count) {
  596. align_get_bits(&s->gb);
  597. skip_bits(&s->gb, 16); /* skip RSTn */
  598. }
  599. }
  600. if (s->rct) {
  601. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  602. ptr[4 * mb_x + 1] = buffer[mb_x][0] - ((buffer[mb_x][1] + buffer[mb_x][2] - 0x200) >> 2);
  603. ptr[4 * mb_x + 0] = buffer[mb_x][1] + ptr[4 * mb_x + 1];
  604. ptr[4 * mb_x + 2] = buffer[mb_x][2] + ptr[4 * mb_x + 1];
  605. }
  606. } else if (s->pegasus_rct) {
  607. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  608. ptr[4 * mb_x + 1] = buffer[mb_x][0] - ((buffer[mb_x][1] + buffer[mb_x][2]) >> 2);
  609. ptr[4 * mb_x + 0] = buffer[mb_x][1] + ptr[4 * mb_x + 1];
  610. ptr[4 * mb_x + 2] = buffer[mb_x][2] + ptr[4 * mb_x + 1];
  611. }
  612. } else {
  613. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  614. ptr[4 * mb_x + 0] = buffer[mb_x][2];
  615. ptr[4 * mb_x + 1] = buffer[mb_x][1];
  616. ptr[4 * mb_x + 2] = buffer[mb_x][0];
  617. }
  618. }
  619. }
  620. return 0;
  621. }
  622. static int ljpeg_decode_yuv_scan(MJpegDecodeContext *s, int predictor,
  623. int point_transform, int nb_components)
  624. {
  625. int i, mb_x, mb_y;
  626. for (mb_y = 0; mb_y < s->mb_height; mb_y++) {
  627. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  628. if (s->restart_interval && !s->restart_count)
  629. s->restart_count = s->restart_interval;
  630. if (mb_x == 0 || mb_y == 0 || s->interlaced) {
  631. for (i = 0; i < nb_components; i++) {
  632. uint8_t *ptr;
  633. int n, h, v, x, y, c, j, linesize;
  634. n = s->nb_blocks[i];
  635. c = s->comp_index[i];
  636. h = s->h_scount[i];
  637. v = s->v_scount[i];
  638. x = 0;
  639. y = 0;
  640. linesize = s->linesize[c];
  641. for (j = 0; j < n; j++) {
  642. int pred;
  643. // FIXME optimize this crap
  644. ptr = s->picture_ptr->data[c] +
  645. (linesize * (v * mb_y + y)) +
  646. (h * mb_x + x);
  647. if (y == 0 && mb_y == 0) {
  648. if (x == 0 && mb_x == 0)
  649. pred = 128 << point_transform;
  650. else
  651. pred = ptr[-1];
  652. } else {
  653. if (x == 0 && mb_x == 0)
  654. pred = ptr[-linesize];
  655. else
  656. PREDICT(pred, ptr[-linesize - 1],
  657. ptr[-linesize], ptr[-1], predictor);
  658. }
  659. if (s->interlaced && s->bottom_field)
  660. ptr += linesize >> 1;
  661. *ptr = pred + (mjpeg_decode_dc(s, s->dc_index[i]) << point_transform);
  662. if (++x == h) {
  663. x = 0;
  664. y++;
  665. }
  666. }
  667. }
  668. } else {
  669. for (i = 0; i < nb_components; i++) {
  670. uint8_t *ptr;
  671. int n, h, v, x, y, c, j, linesize;
  672. n = s->nb_blocks[i];
  673. c = s->comp_index[i];
  674. h = s->h_scount[i];
  675. v = s->v_scount[i];
  676. x = 0;
  677. y = 0;
  678. linesize = s->linesize[c];
  679. for (j = 0; j < n; j++) {
  680. int pred;
  681. // FIXME optimize this crap
  682. ptr = s->picture_ptr->data[c] +
  683. (linesize * (v * mb_y + y)) +
  684. (h * mb_x + x);
  685. PREDICT(pred, ptr[-linesize - 1],
  686. ptr[-linesize], ptr[-1], predictor);
  687. *ptr = pred + (mjpeg_decode_dc(s, s->dc_index[i]) << point_transform);
  688. if (++x == h) {
  689. x = 0;
  690. y++;
  691. }
  692. }
  693. }
  694. }
  695. if (s->restart_interval && !--s->restart_count) {
  696. align_get_bits(&s->gb);
  697. skip_bits(&s->gb, 16); /* skip RSTn */
  698. }
  699. }
  700. }
  701. return 0;
  702. }
  703. static int mjpeg_decode_scan(MJpegDecodeContext *s, int nb_components, int Ah,
  704. int Al, const uint8_t *mb_bitmask,
  705. const AVFrame *reference)
  706. {
  707. int i, mb_x, mb_y;
  708. uint8_t *data[MAX_COMPONENTS];
  709. const uint8_t *reference_data[MAX_COMPONENTS];
  710. int linesize[MAX_COMPONENTS];
  711. GetBitContext mb_bitmask_gb;
  712. if (mb_bitmask)
  713. init_get_bits(&mb_bitmask_gb, mb_bitmask, s->mb_width * s->mb_height);
  714. if (s->flipped && s->avctx->flags & CODEC_FLAG_EMU_EDGE) {
  715. av_log(s->avctx, AV_LOG_ERROR,
  716. "Can not flip image with CODEC_FLAG_EMU_EDGE set!\n");
  717. s->flipped = 0;
  718. }
  719. for (i = 0; i < nb_components; i++) {
  720. int c = s->comp_index[i];
  721. data[c] = s->picture_ptr->data[c];
  722. reference_data[c] = reference ? reference->data[c] : NULL;
  723. linesize[c] = s->linesize[c];
  724. s->coefs_finished[c] |= 1;
  725. if (s->flipped) {
  726. // picture should be flipped upside-down for this codec
  727. int offset = (linesize[c] * (s->v_scount[i] *
  728. (8 * s->mb_height - ((s->height / s->v_max) & 7)) - 1));
  729. data[c] += offset;
  730. reference_data[c] += offset;
  731. linesize[c] *= -1;
  732. }
  733. }
  734. for (mb_y = 0; mb_y < s->mb_height; mb_y++) {
  735. for (mb_x = 0; mb_x < s->mb_width; mb_x++) {
  736. const int copy_mb = mb_bitmask && !get_bits1(&mb_bitmask_gb);
  737. if (s->restart_interval && !s->restart_count)
  738. s->restart_count = s->restart_interval;
  739. if (get_bits_left(&s->gb) < 0) {
  740. av_log(s->avctx, AV_LOG_ERROR, "overread %d\n",
  741. -get_bits_left(&s->gb));
  742. return AVERROR_INVALIDDATA;
  743. }
  744. for (i = 0; i < nb_components; i++) {
  745. uint8_t *ptr;
  746. int n, h, v, x, y, c, j;
  747. int block_offset;
  748. n = s->nb_blocks[i];
  749. c = s->comp_index[i];
  750. h = s->h_scount[i];
  751. v = s->v_scount[i];
  752. x = 0;
  753. y = 0;
  754. for (j = 0; j < n; j++) {
  755. block_offset = ((linesize[c] * (v * mb_y + y) * 8) +
  756. (h * mb_x + x) * 8);
  757. if (s->interlaced && s->bottom_field)
  758. block_offset += linesize[c] >> 1;
  759. ptr = data[c] + block_offset;
  760. if (!s->progressive) {
  761. if (copy_mb)
  762. s->hdsp.put_pixels_tab[1][0](ptr,
  763. reference_data[c] + block_offset,
  764. linesize[c], 8);
  765. else {
  766. s->dsp.clear_block(s->block);
  767. if (decode_block(s, s->block, i,
  768. s->dc_index[i], s->ac_index[i],
  769. s->quant_matrixes[s->quant_index[c]]) < 0) {
  770. av_log(s->avctx, AV_LOG_ERROR,
  771. "error y=%d x=%d\n", mb_y, mb_x);
  772. return AVERROR_INVALIDDATA;
  773. }
  774. s->dsp.idct_put(ptr, linesize[c], s->block);
  775. }
  776. } else {
  777. int block_idx = s->block_stride[c] * (v * mb_y + y) +
  778. (h * mb_x + x);
  779. int16_t *block = s->blocks[c][block_idx];
  780. if (Ah)
  781. block[0] += get_bits1(&s->gb) *
  782. s->quant_matrixes[s->quant_index[c]][0] << Al;
  783. else if (decode_dc_progressive(s, block, i, s->dc_index[i],
  784. s->quant_matrixes[s->quant_index[c]],
  785. Al) < 0) {
  786. av_log(s->avctx, AV_LOG_ERROR,
  787. "error y=%d x=%d\n", mb_y, mb_x);
  788. return AVERROR_INVALIDDATA;
  789. }
  790. }
  791. av_dlog(s->avctx, "mb: %d %d processed\n", mb_y, mb_x);
  792. av_dlog(s->avctx, "%d %d %d %d %d %d %d %d \n",
  793. mb_x, mb_y, x, y, c, s->bottom_field,
  794. (v * mb_y + y) * 8, (h * mb_x + x) * 8);
  795. if (++x == h) {
  796. x = 0;
  797. y++;
  798. }
  799. }
  800. }
  801. if (s->restart_interval) {
  802. s->restart_count--;
  803. i = 8 + ((-get_bits_count(&s->gb)) & 7);
  804. /* skip RSTn */
  805. if (show_bits(&s->gb, i) == (1 << i) - 1) {
  806. int pos = get_bits_count(&s->gb);
  807. align_get_bits(&s->gb);
  808. while (get_bits_left(&s->gb) >= 8 && show_bits(&s->gb, 8) == 0xFF)
  809. skip_bits(&s->gb, 8);
  810. if ((get_bits(&s->gb, 8) & 0xF8) == 0xD0) {
  811. for (i = 0; i < nb_components; i++) /* reset dc */
  812. s->last_dc[i] = 1024;
  813. } else
  814. skip_bits_long(&s->gb, pos - get_bits_count(&s->gb));
  815. }
  816. }
  817. }
  818. }
  819. return 0;
  820. }
  821. static int mjpeg_decode_scan_progressive_ac(MJpegDecodeContext *s, int ss,
  822. int se, int Ah, int Al,
  823. const uint8_t *mb_bitmask,
  824. const AVFrame *reference)
  825. {
  826. int mb_x, mb_y;
  827. int EOBRUN = 0;
  828. int c = s->comp_index[0];
  829. uint8_t *data = s->picture_ptr->data[c];
  830. const uint8_t *reference_data = reference ? reference->data[c] : NULL;
  831. int linesize = s->linesize[c];
  832. int last_scan = 0;
  833. int16_t *quant_matrix = s->quant_matrixes[s->quant_index[c]];
  834. GetBitContext mb_bitmask_gb;
  835. if (ss < 0 || ss >= 64 ||
  836. se < ss || se >= 64 ||
  837. Ah < 0 || Al < 0)
  838. return AVERROR_INVALIDDATA;
  839. if (mb_bitmask)
  840. init_get_bits(&mb_bitmask_gb, mb_bitmask, s->mb_width * s->mb_height);
  841. if (!Al) {
  842. s->coefs_finished[c] |= (1LL << (se + 1)) - (1LL << ss);
  843. last_scan = !~s->coefs_finished[c];
  844. }
  845. if (s->interlaced && s->bottom_field) {
  846. int offset = linesize >> 1;
  847. data += offset;
  848. reference_data += offset;
  849. }
  850. for (mb_y = 0; mb_y < s->mb_height; mb_y++) {
  851. int block_offset = mb_y * linesize * 8;
  852. uint8_t *ptr = data + block_offset;
  853. int block_idx = mb_y * s->block_stride[c];
  854. int16_t (*block)[64] = &s->blocks[c][block_idx];
  855. uint8_t *last_nnz = &s->last_nnz[c][block_idx];
  856. for (mb_x = 0; mb_x < s->mb_width; mb_x++, block++, last_nnz++) {
  857. const int copy_mb = mb_bitmask && !get_bits1(&mb_bitmask_gb);
  858. if (!copy_mb) {
  859. int ret;
  860. if (Ah)
  861. ret = decode_block_refinement(s, *block, last_nnz, s->ac_index[0],
  862. quant_matrix, ss, se, Al, &EOBRUN);
  863. else
  864. ret = decode_block_progressive(s, *block, last_nnz, s->ac_index[0],
  865. quant_matrix, ss, se, Al, &EOBRUN);
  866. if (ret < 0) {
  867. av_log(s->avctx, AV_LOG_ERROR,
  868. "error y=%d x=%d\n", mb_y, mb_x);
  869. return AVERROR_INVALIDDATA;
  870. }
  871. }
  872. if (last_scan) {
  873. if (copy_mb) {
  874. s->hdsp.put_pixels_tab[1][0](ptr,
  875. reference_data + block_offset,
  876. linesize, 8);
  877. } else {
  878. s->dsp.idct_put(ptr, linesize, *block);
  879. ptr += 8;
  880. }
  881. }
  882. }
  883. }
  884. return 0;
  885. }
  886. int ff_mjpeg_decode_sos(MJpegDecodeContext *s, const uint8_t *mb_bitmask,
  887. const AVFrame *reference)
  888. {
  889. int len, nb_components, i, h, v, predictor, point_transform;
  890. int index, id, ret;
  891. const int block_size = s->lossless ? 1 : 8;
  892. int ilv, prev_shift;
  893. /* XXX: verify len field validity */
  894. len = get_bits(&s->gb, 16);
  895. nb_components = get_bits(&s->gb, 8);
  896. if (nb_components == 0 || nb_components > MAX_COMPONENTS) {
  897. av_log(s->avctx, AV_LOG_ERROR,
  898. "decode_sos: nb_components (%d) unsupported\n", nb_components);
  899. return AVERROR_PATCHWELCOME;
  900. }
  901. if (len != 6 + 2 * nb_components) {
  902. av_log(s->avctx, AV_LOG_ERROR, "decode_sos: invalid len (%d)\n", len);
  903. return AVERROR_INVALIDDATA;
  904. }
  905. for (i = 0; i < nb_components; i++) {
  906. id = get_bits(&s->gb, 8) - 1;
  907. av_log(s->avctx, AV_LOG_DEBUG, "component: %d\n", id);
  908. /* find component index */
  909. for (index = 0; index < s->nb_components; index++)
  910. if (id == s->component_id[index])
  911. break;
  912. if (index == s->nb_components) {
  913. av_log(s->avctx, AV_LOG_ERROR,
  914. "decode_sos: index(%d) out of components\n", index);
  915. return AVERROR_INVALIDDATA;
  916. }
  917. /* Metasoft MJPEG codec has Cb and Cr swapped */
  918. if (s->avctx->codec_tag == MKTAG('M', 'T', 'S', 'J')
  919. && nb_components == 3 && s->nb_components == 3 && i)
  920. index = 3 - i;
  921. s->comp_index[i] = index;
  922. s->nb_blocks[i] = s->h_count[index] * s->v_count[index];
  923. s->h_scount[i] = s->h_count[index];
  924. s->v_scount[i] = s->v_count[index];
  925. s->dc_index[i] = get_bits(&s->gb, 4);
  926. s->ac_index[i] = get_bits(&s->gb, 4);
  927. if (s->dc_index[i] < 0 || s->ac_index[i] < 0 ||
  928. s->dc_index[i] >= 4 || s->ac_index[i] >= 4)
  929. goto out_of_range;
  930. if (!s->vlcs[0][s->dc_index[i]].table ||
  931. !s->vlcs[1][s->ac_index[i]].table)
  932. goto out_of_range;
  933. }
  934. predictor = get_bits(&s->gb, 8); /* JPEG Ss / lossless JPEG predictor /JPEG-LS NEAR */
  935. ilv = get_bits(&s->gb, 8); /* JPEG Se / JPEG-LS ILV */
  936. prev_shift = get_bits(&s->gb, 4); /* Ah */
  937. point_transform = get_bits(&s->gb, 4); /* Al */
  938. if (nb_components > 1) {
  939. /* interleaved stream */
  940. s->mb_width = (s->width + s->h_max * block_size - 1) / (s->h_max * block_size);
  941. s->mb_height = (s->height + s->v_max * block_size - 1) / (s->v_max * block_size);
  942. } else if (!s->ls) { /* skip this for JPEG-LS */
  943. h = s->h_max / s->h_scount[0];
  944. v = s->v_max / s->v_scount[0];
  945. s->mb_width = (s->width + h * block_size - 1) / (h * block_size);
  946. s->mb_height = (s->height + v * block_size - 1) / (v * block_size);
  947. s->nb_blocks[0] = 1;
  948. s->h_scount[0] = 1;
  949. s->v_scount[0] = 1;
  950. }
  951. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  952. av_log(s->avctx, AV_LOG_DEBUG, "%s %s p:%d >>:%d ilv:%d bits:%d %s\n",
  953. s->lossless ? "lossless" : "sequential DCT", s->rgb ? "RGB" : "",
  954. predictor, point_transform, ilv, s->bits,
  955. s->pegasus_rct ? "PRCT" : (s->rct ? "RCT" : ""));
  956. /* mjpeg-b can have padding bytes between sos and image data, skip them */
  957. for (i = s->mjpb_skiptosod; i > 0; i--)
  958. skip_bits(&s->gb, 8);
  959. next_field:
  960. for (i = 0; i < nb_components; i++)
  961. s->last_dc[i] = 1024;
  962. if (s->lossless) {
  963. if (CONFIG_JPEGLS_DECODER && s->ls) {
  964. // for () {
  965. // reset_ls_coding_parameters(s, 0);
  966. if ((ret = ff_jpegls_decode_picture(s, predictor,
  967. point_transform, ilv)) < 0)
  968. return ret;
  969. } else {
  970. if (s->rgb) {
  971. if ((ret = ljpeg_decode_rgb_scan(s, predictor,
  972. point_transform)) < 0)
  973. return ret;
  974. } else {
  975. if ((ret = ljpeg_decode_yuv_scan(s, predictor,
  976. point_transform,
  977. nb_components)) < 0)
  978. return ret;
  979. }
  980. }
  981. } else {
  982. if (s->progressive && predictor) {
  983. if ((ret = mjpeg_decode_scan_progressive_ac(s, predictor,
  984. ilv, prev_shift,
  985. point_transform,
  986. mb_bitmask,
  987. reference)) < 0)
  988. return ret;
  989. } else {
  990. if ((ret = mjpeg_decode_scan(s, nb_components,
  991. prev_shift, point_transform,
  992. mb_bitmask, reference)) < 0)
  993. return ret;
  994. }
  995. }
  996. if (s->interlaced &&
  997. get_bits_left(&s->gb) > 32 &&
  998. show_bits(&s->gb, 8) == 0xFF) {
  999. GetBitContext bak = s->gb;
  1000. align_get_bits(&bak);
  1001. if (show_bits(&bak, 16) == 0xFFD1) {
  1002. av_dlog(s->avctx, "AVRn interlaced picture marker found\n");
  1003. s->gb = bak;
  1004. skip_bits(&s->gb, 16);
  1005. s->bottom_field ^= 1;
  1006. goto next_field;
  1007. }
  1008. }
  1009. emms_c();
  1010. return 0;
  1011. out_of_range:
  1012. av_log(s->avctx, AV_LOG_ERROR, "decode_sos: ac/dc index out of range\n");
  1013. return AVERROR_INVALIDDATA;
  1014. }
  1015. static int mjpeg_decode_dri(MJpegDecodeContext *s)
  1016. {
  1017. if (get_bits(&s->gb, 16) != 4)
  1018. return AVERROR_INVALIDDATA;
  1019. s->restart_interval = get_bits(&s->gb, 16);
  1020. s->restart_count = 0;
  1021. av_log(s->avctx, AV_LOG_DEBUG, "restart interval: %d\n",
  1022. s->restart_interval);
  1023. return 0;
  1024. }
  1025. static int mjpeg_decode_app(MJpegDecodeContext *s)
  1026. {
  1027. int len, id, i;
  1028. len = get_bits(&s->gb, 16);
  1029. if (len < 5)
  1030. return AVERROR_INVALIDDATA;
  1031. if (8 * len > get_bits_left(&s->gb))
  1032. return AVERROR_INVALIDDATA;
  1033. id = get_bits_long(&s->gb, 32);
  1034. id = av_be2ne32(id);
  1035. len -= 6;
  1036. if (s->avctx->debug & FF_DEBUG_STARTCODE)
  1037. av_log(s->avctx, AV_LOG_DEBUG, "APPx %8X\n", id);
  1038. /* Buggy AVID, it puts EOI only at every 10th frame. */
  1039. /* Also, this fourcc is used by non-avid files too, it holds some
  1040. information, but it's always present in AVID-created files. */
  1041. if (id == AV_RL32("AVI1")) {
  1042. /* structure:
  1043. 4bytes AVI1
  1044. 1bytes polarity
  1045. 1bytes always zero
  1046. 4bytes field_size
  1047. 4bytes field_size_less_padding
  1048. */
  1049. s->buggy_avid = 1;
  1050. i = get_bits(&s->gb, 8);
  1051. if (i == 2)
  1052. s->bottom_field = 1;
  1053. else if (i == 1)
  1054. s->bottom_field = 0;
  1055. #if 0
  1056. skip_bits(&s->gb, 8);
  1057. skip_bits(&s->gb, 32);
  1058. skip_bits(&s->gb, 32);
  1059. len -= 10;
  1060. #endif
  1061. goto out;
  1062. }
  1063. // len -= 2;
  1064. if (id == AV_RL32("JFIF")) {
  1065. int t_w, t_h, v1, v2;
  1066. skip_bits(&s->gb, 8); /* the trailing zero-byte */
  1067. v1 = get_bits(&s->gb, 8);
  1068. v2 = get_bits(&s->gb, 8);
  1069. skip_bits(&s->gb, 8);
  1070. s->avctx->sample_aspect_ratio.num = get_bits(&s->gb, 16);
  1071. s->avctx->sample_aspect_ratio.den = get_bits(&s->gb, 16);
  1072. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  1073. av_log(s->avctx, AV_LOG_INFO,
  1074. "mjpeg: JFIF header found (version: %x.%x) SAR=%d/%d\n",
  1075. v1, v2,
  1076. s->avctx->sample_aspect_ratio.num,
  1077. s->avctx->sample_aspect_ratio.den);
  1078. t_w = get_bits(&s->gb, 8);
  1079. t_h = get_bits(&s->gb, 8);
  1080. if (t_w && t_h) {
  1081. /* skip thumbnail */
  1082. if (len -10 - (t_w * t_h * 3) > 0)
  1083. len -= t_w * t_h * 3;
  1084. }
  1085. len -= 10;
  1086. goto out;
  1087. }
  1088. if (id == AV_RL32("Adob") && (get_bits(&s->gb, 8) == 'e')) {
  1089. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  1090. av_log(s->avctx, AV_LOG_INFO, "mjpeg: Adobe header found\n");
  1091. skip_bits(&s->gb, 16); /* version */
  1092. skip_bits(&s->gb, 16); /* flags0 */
  1093. skip_bits(&s->gb, 16); /* flags1 */
  1094. skip_bits(&s->gb, 8); /* transform */
  1095. len -= 7;
  1096. goto out;
  1097. }
  1098. if (id == AV_RL32("LJIF")) {
  1099. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  1100. av_log(s->avctx, AV_LOG_INFO,
  1101. "Pegasus lossless jpeg header found\n");
  1102. skip_bits(&s->gb, 16); /* version ? */
  1103. skip_bits(&s->gb, 16); /* unknwon always 0? */
  1104. skip_bits(&s->gb, 16); /* unknwon always 0? */
  1105. skip_bits(&s->gb, 16); /* unknwon always 0? */
  1106. switch (get_bits(&s->gb, 8)) {
  1107. case 1:
  1108. s->rgb = 1;
  1109. s->pegasus_rct = 0;
  1110. break;
  1111. case 2:
  1112. s->rgb = 1;
  1113. s->pegasus_rct = 1;
  1114. break;
  1115. default:
  1116. av_log(s->avctx, AV_LOG_ERROR, "unknown colorspace\n");
  1117. }
  1118. len -= 9;
  1119. goto out;
  1120. }
  1121. /* Apple MJPEG-A */
  1122. if ((s->start_code == APP1) && (len > (0x28 - 8))) {
  1123. id = get_bits_long(&s->gb, 32);
  1124. id = av_be2ne32(id);
  1125. len -= 4;
  1126. /* Apple MJPEG-A */
  1127. if (id == AV_RL32("mjpg")) {
  1128. #if 0
  1129. skip_bits(&s->gb, 32); /* field size */
  1130. skip_bits(&s->gb, 32); /* pad field size */
  1131. skip_bits(&s->gb, 32); /* next off */
  1132. skip_bits(&s->gb, 32); /* quant off */
  1133. skip_bits(&s->gb, 32); /* huff off */
  1134. skip_bits(&s->gb, 32); /* image off */
  1135. skip_bits(&s->gb, 32); /* scan off */
  1136. skip_bits(&s->gb, 32); /* data off */
  1137. #endif
  1138. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  1139. av_log(s->avctx, AV_LOG_INFO, "mjpeg: Apple MJPEG-A header found\n");
  1140. }
  1141. }
  1142. out:
  1143. /* slow but needed for extreme adobe jpegs */
  1144. if (len < 0)
  1145. av_log(s->avctx, AV_LOG_ERROR,
  1146. "mjpeg: error, decode_app parser read over the end\n");
  1147. while (--len > 0)
  1148. skip_bits(&s->gb, 8);
  1149. return 0;
  1150. }
  1151. static int mjpeg_decode_com(MJpegDecodeContext *s)
  1152. {
  1153. int len = get_bits(&s->gb, 16);
  1154. if (len >= 2 && 8 * len - 16 <= get_bits_left(&s->gb)) {
  1155. char *cbuf = av_malloc(len - 1);
  1156. if (cbuf) {
  1157. int i;
  1158. for (i = 0; i < len - 2; i++)
  1159. cbuf[i] = get_bits(&s->gb, 8);
  1160. if (i > 0 && cbuf[i - 1] == '\n')
  1161. cbuf[i - 1] = 0;
  1162. else
  1163. cbuf[i] = 0;
  1164. if (s->avctx->debug & FF_DEBUG_PICT_INFO)
  1165. av_log(s->avctx, AV_LOG_INFO, "mjpeg comment: '%s'\n", cbuf);
  1166. /* buggy avid, it puts EOI only at every 10th frame */
  1167. if (!strcmp(cbuf, "AVID")) {
  1168. s->buggy_avid = 1;
  1169. } else if (!strcmp(cbuf, "CS=ITU601"))
  1170. s->cs_itu601 = 1;
  1171. else if ((len > 20 && !strncmp(cbuf, "Intel(R) JPEG Library", 21)) ||
  1172. (len > 19 && !strncmp(cbuf, "Metasoft MJPEG Codec", 20)))
  1173. s->flipped = 1;
  1174. av_free(cbuf);
  1175. }
  1176. }
  1177. return 0;
  1178. }
  1179. /* return the 8 bit start code value and update the search
  1180. state. Return -1 if no start code found */
  1181. static int find_marker(const uint8_t **pbuf_ptr, const uint8_t *buf_end)
  1182. {
  1183. const uint8_t *buf_ptr;
  1184. unsigned int v, v2;
  1185. int val;
  1186. #ifdef DEBUG
  1187. int skipped = 0;
  1188. #endif
  1189. buf_ptr = *pbuf_ptr;
  1190. while (buf_ptr < buf_end) {
  1191. v = *buf_ptr++;
  1192. v2 = *buf_ptr;
  1193. if ((v == 0xff) && (v2 >= 0xc0) && (v2 <= 0xfe) && buf_ptr < buf_end) {
  1194. val = *buf_ptr++;
  1195. goto found;
  1196. }
  1197. #ifdef DEBUG
  1198. skipped++;
  1199. #endif
  1200. }
  1201. val = -1;
  1202. found:
  1203. av_dlog(NULL, "find_marker skipped %d bytes\n", skipped);
  1204. *pbuf_ptr = buf_ptr;
  1205. return val;
  1206. }
  1207. int ff_mjpeg_find_marker(MJpegDecodeContext *s,
  1208. const uint8_t **buf_ptr, const uint8_t *buf_end,
  1209. const uint8_t **unescaped_buf_ptr,
  1210. int *unescaped_buf_size)
  1211. {
  1212. int start_code;
  1213. start_code = find_marker(buf_ptr, buf_end);
  1214. av_fast_padded_malloc(&s->buffer, &s->buffer_size, buf_end - *buf_ptr);
  1215. if (!s->buffer)
  1216. return AVERROR(ENOMEM);
  1217. /* unescape buffer of SOS, use special treatment for JPEG-LS */
  1218. if (start_code == SOS && !s->ls) {
  1219. const uint8_t *src = *buf_ptr;
  1220. uint8_t *dst = s->buffer;
  1221. while (src < buf_end) {
  1222. uint8_t x = *(src++);
  1223. *(dst++) = x;
  1224. if (s->avctx->codec_id != AV_CODEC_ID_THP) {
  1225. if (x == 0xff) {
  1226. while (src < buf_end && x == 0xff)
  1227. x = *(src++);
  1228. if (x >= 0xd0 && x <= 0xd7)
  1229. *(dst++) = x;
  1230. else if (x)
  1231. break;
  1232. }
  1233. }
  1234. }
  1235. *unescaped_buf_ptr = s->buffer;
  1236. *unescaped_buf_size = dst - s->buffer;
  1237. memset(s->buffer + *unescaped_buf_size, 0,
  1238. FF_INPUT_BUFFER_PADDING_SIZE);
  1239. av_log(s->avctx, AV_LOG_DEBUG, "escaping removed %td bytes\n",
  1240. (buf_end - *buf_ptr) - (dst - s->buffer));
  1241. } else if (start_code == SOS && s->ls) {
  1242. const uint8_t *src = *buf_ptr;
  1243. uint8_t *dst = s->buffer;
  1244. int bit_count = 0;
  1245. int t = 0, b = 0;
  1246. PutBitContext pb;
  1247. s->cur_scan++;
  1248. /* find marker */
  1249. while (src + t < buf_end) {
  1250. uint8_t x = src[t++];
  1251. if (x == 0xff) {
  1252. while ((src + t < buf_end) && x == 0xff)
  1253. x = src[t++];
  1254. if (x & 0x80) {
  1255. t -= 2;
  1256. break;
  1257. }
  1258. }
  1259. }
  1260. bit_count = t * 8;
  1261. init_put_bits(&pb, dst, t);
  1262. /* unescape bitstream */
  1263. while (b < t) {
  1264. uint8_t x = src[b++];
  1265. put_bits(&pb, 8, x);
  1266. if (x == 0xFF) {
  1267. x = src[b++];
  1268. put_bits(&pb, 7, x);
  1269. bit_count--;
  1270. }
  1271. }
  1272. flush_put_bits(&pb);
  1273. *unescaped_buf_ptr = dst;
  1274. *unescaped_buf_size = (bit_count + 7) >> 3;
  1275. memset(s->buffer + *unescaped_buf_size, 0,
  1276. FF_INPUT_BUFFER_PADDING_SIZE);
  1277. } else {
  1278. *unescaped_buf_ptr = *buf_ptr;
  1279. *unescaped_buf_size = buf_end - *buf_ptr;
  1280. }
  1281. return start_code;
  1282. }
  1283. int ff_mjpeg_decode_frame(AVCodecContext *avctx, void *data, int *got_frame,
  1284. AVPacket *avpkt)
  1285. {
  1286. const uint8_t *buf = avpkt->data;
  1287. int buf_size = avpkt->size;
  1288. MJpegDecodeContext *s = avctx->priv_data;
  1289. const uint8_t *buf_end, *buf_ptr;
  1290. const uint8_t *unescaped_buf_ptr;
  1291. int unescaped_buf_size;
  1292. int start_code;
  1293. int ret = 0;
  1294. s->got_picture = 0; // picture from previous image can not be reused
  1295. buf_ptr = buf;
  1296. buf_end = buf + buf_size;
  1297. while (buf_ptr < buf_end) {
  1298. /* find start next marker */
  1299. start_code = ff_mjpeg_find_marker(s, &buf_ptr, buf_end,
  1300. &unescaped_buf_ptr,
  1301. &unescaped_buf_size);
  1302. /* EOF */
  1303. if (start_code < 0) {
  1304. goto the_end;
  1305. } else if (unescaped_buf_size > (1U<<29)) {
  1306. av_log(avctx, AV_LOG_ERROR, "MJPEG packet 0x%x too big (0x%x/0x%x), corrupt data?\n",
  1307. start_code, unescaped_buf_size, buf_size);
  1308. return AVERROR_INVALIDDATA;
  1309. } else {
  1310. av_log(avctx, AV_LOG_DEBUG, "marker=%x avail_size_in_buf=%td\n",
  1311. start_code, buf_end - buf_ptr);
  1312. init_get_bits(&s->gb, unescaped_buf_ptr, unescaped_buf_size * 8);
  1313. s->start_code = start_code;
  1314. if (s->avctx->debug & FF_DEBUG_STARTCODE)
  1315. av_log(avctx, AV_LOG_DEBUG, "startcode: %X\n", start_code);
  1316. /* process markers */
  1317. if (start_code >= 0xd0 && start_code <= 0xd7)
  1318. av_log(avctx, AV_LOG_DEBUG,
  1319. "restart marker: %d\n", start_code & 0x0f);
  1320. /* APP fields */
  1321. else if (start_code >= APP0 && start_code <= APP15)
  1322. mjpeg_decode_app(s);
  1323. /* Comment */
  1324. else if (start_code == COM)
  1325. mjpeg_decode_com(s);
  1326. if (!CONFIG_JPEGLS_DECODER &&
  1327. (start_code == SOF48 || start_code == LSE)) {
  1328. av_log(avctx, AV_LOG_ERROR, "JPEG-LS support not enabled.\n");
  1329. return AVERROR(ENOSYS);
  1330. }
  1331. switch (start_code) {
  1332. case SOI:
  1333. s->restart_interval = 0;
  1334. s->restart_count = 0;
  1335. /* nothing to do on SOI */
  1336. break;
  1337. case DQT:
  1338. ff_mjpeg_decode_dqt(s);
  1339. break;
  1340. case DHT:
  1341. if ((ret = ff_mjpeg_decode_dht(s)) < 0) {
  1342. av_log(avctx, AV_LOG_ERROR, "huffman table decode error\n");
  1343. return ret;
  1344. }
  1345. break;
  1346. case SOF0:
  1347. case SOF1:
  1348. s->lossless = 0;
  1349. s->ls = 0;
  1350. s->progressive = 0;
  1351. if ((ret = ff_mjpeg_decode_sof(s)) < 0)
  1352. return ret;
  1353. break;
  1354. case SOF2:
  1355. s->lossless = 0;
  1356. s->ls = 0;
  1357. s->progressive = 1;
  1358. if ((ret = ff_mjpeg_decode_sof(s)) < 0)
  1359. return ret;
  1360. break;
  1361. case SOF3:
  1362. s->lossless = 1;
  1363. s->ls = 0;
  1364. s->progressive = 0;
  1365. if ((ret = ff_mjpeg_decode_sof(s)) < 0)
  1366. return ret;
  1367. break;
  1368. case SOF48:
  1369. s->lossless = 1;
  1370. s->ls = 1;
  1371. s->progressive = 0;
  1372. if ((ret = ff_mjpeg_decode_sof(s)) < 0)
  1373. return ret;
  1374. break;
  1375. case LSE:
  1376. if (!CONFIG_JPEGLS_DECODER ||
  1377. (ret = ff_jpegls_decode_lse(s)) < 0)
  1378. return ret;
  1379. break;
  1380. case EOI:
  1381. s->cur_scan = 0;
  1382. if ((s->buggy_avid && !s->interlaced) || s->restart_interval)
  1383. break;
  1384. eoi_parser:
  1385. if (!s->got_picture) {
  1386. av_log(avctx, AV_LOG_WARNING,
  1387. "Found EOI before any SOF, ignoring\n");
  1388. break;
  1389. }
  1390. if (s->interlaced) {
  1391. s->bottom_field ^= 1;
  1392. /* if not bottom field, do not output image yet */
  1393. if (s->bottom_field == !s->interlace_polarity)
  1394. goto not_the_end;
  1395. }
  1396. if ((ret = av_frame_ref(data, s->picture_ptr)) < 0)
  1397. return ret;
  1398. *got_frame = 1;
  1399. if (!s->lossless &&
  1400. avctx->debug & FF_DEBUG_QP) {
  1401. av_log(avctx, AV_LOG_DEBUG,
  1402. "QP: %d\n", FFMAX3(s->qscale[0],
  1403. s->qscale[1],
  1404. s->qscale[2]));
  1405. }
  1406. goto the_end;
  1407. case SOS:
  1408. if (!s->got_picture) {
  1409. av_log(avctx, AV_LOG_WARNING,
  1410. "Can not process SOS before SOF, skipping\n");
  1411. break;
  1412. }
  1413. if ((ret = ff_mjpeg_decode_sos(s, NULL, NULL)) < 0 &&
  1414. (avctx->err_recognition & AV_EF_EXPLODE))
  1415. return ret;
  1416. /* buggy avid puts EOI every 10-20th frame */
  1417. /* if restart period is over process EOI */
  1418. if ((s->buggy_avid && !s->interlaced) || s->restart_interval)
  1419. goto eoi_parser;
  1420. break;
  1421. case DRI:
  1422. mjpeg_decode_dri(s);
  1423. break;
  1424. case SOF5:
  1425. case SOF6:
  1426. case SOF7:
  1427. case SOF9:
  1428. case SOF10:
  1429. case SOF11:
  1430. case SOF13:
  1431. case SOF14:
  1432. case SOF15:
  1433. case JPG:
  1434. av_log(avctx, AV_LOG_ERROR,
  1435. "mjpeg: unsupported coding type (%x)\n", start_code);
  1436. break;
  1437. }
  1438. not_the_end:
  1439. /* eof process start code */
  1440. buf_ptr += (get_bits_count(&s->gb) + 7) / 8;
  1441. av_log(avctx, AV_LOG_DEBUG,
  1442. "marker parser used %d bytes (%d bits)\n",
  1443. (get_bits_count(&s->gb) + 7) / 8, get_bits_count(&s->gb));
  1444. }
  1445. }
  1446. if (s->got_picture) {
  1447. av_log(avctx, AV_LOG_WARNING, "EOI missing, emulating\n");
  1448. goto eoi_parser;
  1449. }
  1450. av_log(avctx, AV_LOG_FATAL, "No JPEG data found in image\n");
  1451. return AVERROR_INVALIDDATA;
  1452. the_end:
  1453. av_log(avctx, AV_LOG_DEBUG, "mjpeg decode frame unused %td bytes\n",
  1454. buf_end - buf_ptr);
  1455. // return buf_end - buf_ptr;
  1456. return buf_ptr - buf;
  1457. }
  1458. av_cold int ff_mjpeg_decode_end(AVCodecContext *avctx)
  1459. {
  1460. MJpegDecodeContext *s = avctx->priv_data;
  1461. int i, j;
  1462. if (s->picture_ptr)
  1463. av_frame_unref(s->picture_ptr);
  1464. av_free(s->buffer);
  1465. av_freep(&s->ljpeg_buffer);
  1466. s->ljpeg_buffer_size = 0;
  1467. for (i = 0; i < 3; i++) {
  1468. for (j = 0; j < 4; j++)
  1469. ff_free_vlc(&s->vlcs[i][j]);
  1470. }
  1471. for (i = 0; i < MAX_COMPONENTS; i++) {
  1472. av_freep(&s->blocks[i]);
  1473. av_freep(&s->last_nnz[i]);
  1474. }
  1475. return 0;
  1476. }
  1477. #define OFFSET(x) offsetof(MJpegDecodeContext, x)
  1478. #define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM
  1479. static const AVOption options[] = {
  1480. { "extern_huff", "Use external huffman table.",
  1481. OFFSET(extern_huff), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 1, VD },
  1482. { NULL },
  1483. };
  1484. static const AVClass mjpegdec_class = {
  1485. .class_name = "MJPEG decoder",
  1486. .item_name = av_default_item_name,
  1487. .option = options,
  1488. .version = LIBAVUTIL_VERSION_INT,
  1489. };
  1490. AVCodec ff_mjpeg_decoder = {
  1491. .name = "mjpeg",
  1492. .type = AVMEDIA_TYPE_VIDEO,
  1493. .id = AV_CODEC_ID_MJPEG,
  1494. .priv_data_size = sizeof(MJpegDecodeContext),
  1495. .init = ff_mjpeg_decode_init,
  1496. .close = ff_mjpeg_decode_end,
  1497. .decode = ff_mjpeg_decode_frame,
  1498. .capabilities = CODEC_CAP_DR1,
  1499. .long_name = NULL_IF_CONFIG_SMALL("MJPEG (Motion JPEG)"),
  1500. .priv_class = &mjpegdec_class,
  1501. };
  1502. AVCodec ff_thp_decoder = {
  1503. .name = "thp",
  1504. .type = AVMEDIA_TYPE_VIDEO,
  1505. .id = AV_CODEC_ID_THP,
  1506. .priv_data_size = sizeof(MJpegDecodeContext),
  1507. .init = ff_mjpeg_decode_init,
  1508. .close = ff_mjpeg_decode_end,
  1509. .decode = ff_mjpeg_decode_frame,
  1510. .capabilities = CODEC_CAP_DR1,
  1511. .long_name = NULL_IF_CONFIG_SMALL("Nintendo Gamecube THP video"),
  1512. };