You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1115 lines
46KB

  1. /*
  2. * Intel Indeo 3 (IV31, IV32, etc.) video decoder for ffmpeg
  3. * written, produced, and directed by Alan Smithee
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include <stdio.h>
  22. #include <stdlib.h>
  23. #include <string.h>
  24. #include <unistd.h>
  25. #include "avcodec.h"
  26. #include "dsputil.h"
  27. #include "bytestream.h"
  28. #include "indeo3data.h"
  29. typedef struct
  30. {
  31. uint8_t *Ybuf;
  32. uint8_t *Ubuf;
  33. uint8_t *Vbuf;
  34. unsigned short y_w, y_h;
  35. unsigned short uv_w, uv_h;
  36. } YUVBufs;
  37. typedef struct Indeo3DecodeContext {
  38. AVCodecContext *avctx;
  39. int width, height;
  40. AVFrame frame;
  41. uint8_t *buf;
  42. YUVBufs iv_frame[2];
  43. YUVBufs *cur_frame;
  44. YUVBufs *ref_frame;
  45. uint8_t *ModPred;
  46. uint8_t *corrector_type;
  47. } Indeo3DecodeContext;
  48. static const uint8_t corrector_type_0[24] = {
  49. 195, 159, 133, 115, 101, 93, 87, 77,
  50. 195, 159, 133, 115, 101, 93, 87, 77,
  51. 128, 79, 79, 79, 79, 79, 79, 79
  52. };
  53. static const uint8_t corrector_type_2[8] = { 9, 7, 6, 8, 5, 4, 3, 2 };
  54. static av_cold int build_modpred(Indeo3DecodeContext *s)
  55. {
  56. int i, j;
  57. if (!(s->ModPred = av_malloc(8 * 128)))
  58. return AVERROR(ENOMEM);
  59. for (i=0; i < 128; ++i) {
  60. s->ModPred[i+0*128] = i > 126 ? 254 : 2*(i + 1 - ((i + 1) % 2));
  61. s->ModPred[i+1*128] = i == 7 ? 20 :
  62. i == 119 ||
  63. i == 120 ? 236 : 2*(i + 2 - ((i + 1) % 3));
  64. s->ModPred[i+2*128] = i > 125 ? 248 : 2*(i + 2 - ((i + 2) % 4));
  65. s->ModPred[i+3*128] = 2*(i + 1 - ((i - 3) % 5));
  66. s->ModPred[i+4*128] = i == 8 ? 20 : 2*(i + 1 - ((i - 3) % 6));
  67. s->ModPred[i+5*128] = 2*(i + 4 - ((i + 3) % 7));
  68. s->ModPred[i+6*128] = i > 123 ? 240 : 2*(i + 4 - ((i + 4) % 8));
  69. s->ModPred[i+7*128] = 2*(i + 5 - ((i + 4) % 9));
  70. }
  71. if (!(s->corrector_type = av_malloc(24 * 256)))
  72. return AVERROR(ENOMEM);
  73. for (i=0; i < 24; ++i) {
  74. for (j=0; j < 256; ++j) {
  75. s->corrector_type[i*256+j] = j < corrector_type_0[i] ? 1 :
  76. j < 248 || (i == 16 && j == 248) ? 0 :
  77. corrector_type_2[j - 248];
  78. }
  79. }
  80. return 0;
  81. }
  82. static av_cold int iv_alloc_frames(Indeo3DecodeContext *s)
  83. {
  84. int luma_width = (s->width + 3) & ~3,
  85. luma_height = (s->height + 3) & ~3,
  86. chroma_width = ((luma_width >> 2) + 3) & ~3,
  87. chroma_height = ((luma_height >> 2) + 3) & ~3,
  88. luma_pixels = luma_width * luma_height,
  89. chroma_pixels = chroma_width * chroma_height,
  90. i;
  91. unsigned int bufsize = luma_pixels * 2 + luma_width * 3 +
  92. (chroma_pixels + chroma_width) * 4;
  93. if(!(s->buf = av_malloc(bufsize)))
  94. return AVERROR(ENOMEM);
  95. s->iv_frame[0].y_w = s->iv_frame[1].y_w = luma_width;
  96. s->iv_frame[0].y_h = s->iv_frame[1].y_h = luma_height;
  97. s->iv_frame[0].uv_w = s->iv_frame[1].uv_w = chroma_width;
  98. s->iv_frame[0].uv_h = s->iv_frame[1].uv_h = chroma_height;
  99. s->iv_frame[0].Ybuf = s->buf + luma_width;
  100. i = luma_pixels + luma_width * 2;
  101. s->iv_frame[1].Ybuf = s->buf + i;
  102. i += (luma_pixels + luma_width);
  103. s->iv_frame[0].Ubuf = s->buf + i;
  104. i += (chroma_pixels + chroma_width);
  105. s->iv_frame[1].Ubuf = s->buf + i;
  106. i += (chroma_pixels + chroma_width);
  107. s->iv_frame[0].Vbuf = s->buf + i;
  108. i += (chroma_pixels + chroma_width);
  109. s->iv_frame[1].Vbuf = s->buf + i;
  110. for(i = 1; i <= luma_width; i++)
  111. s->iv_frame[0].Ybuf[-i] = s->iv_frame[1].Ybuf[-i] =
  112. s->iv_frame[0].Ubuf[-i] = 0x80;
  113. for(i = 1; i <= chroma_width; i++) {
  114. s->iv_frame[1].Ubuf[-i] = 0x80;
  115. s->iv_frame[0].Vbuf[-i] = 0x80;
  116. s->iv_frame[1].Vbuf[-i] = 0x80;
  117. s->iv_frame[1].Vbuf[chroma_pixels+i-1] = 0x80;
  118. }
  119. return 0;
  120. }
  121. static av_cold void iv_free_func(Indeo3DecodeContext *s)
  122. {
  123. av_free(s->buf);
  124. av_free(s->ModPred);
  125. av_free(s->corrector_type);
  126. }
  127. typedef struct {
  128. long xpos;
  129. long ypos;
  130. long width;
  131. long height;
  132. long split_flag;
  133. long split_direction;
  134. long usl7;
  135. } ustr_t;
  136. #define LV1_CHECK(buf1,rle_v3,lv1,lp2) \
  137. if((lv1 & 0x80) != 0) { \
  138. if(rle_v3 != 0) \
  139. rle_v3 = 0; \
  140. else { \
  141. rle_v3 = 1; \
  142. buf1 -= 2; \
  143. } \
  144. } \
  145. lp2 = 4;
  146. #define RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3) \
  147. if(rle_v3 == 0) { \
  148. rle_v2 = *buf1; \
  149. rle_v1 = 1; \
  150. if(rle_v2 > 32) { \
  151. rle_v2 -= 32; \
  152. rle_v1 = 0; \
  153. } \
  154. rle_v3 = 1; \
  155. } \
  156. buf1--;
  157. #define LP2_CHECK(buf1,rle_v3,lp2) \
  158. if(lp2 == 0 && rle_v3 != 0) \
  159. rle_v3 = 0; \
  160. else { \
  161. buf1--; \
  162. rle_v3 = 1; \
  163. }
  164. #define RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2) \
  165. rle_v2--; \
  166. if(rle_v2 == 0) { \
  167. rle_v3 = 0; \
  168. buf1 += 2; \
  169. } \
  170. lp2 = 4;
  171. static void iv_Decode_Chunk(Indeo3DecodeContext *s,
  172. uint8_t *cur, uint8_t *ref, int width, int height,
  173. const uint8_t *buf1, long cb_offset, const uint8_t *hdr,
  174. const uint8_t *buf2, int min_width_160)
  175. {
  176. uint8_t bit_buf;
  177. unsigned long bit_pos, lv, lv1, lv2;
  178. long *width_tbl, width_tbl_arr[10];
  179. const signed char *ref_vectors;
  180. uint8_t *cur_frm_pos, *ref_frm_pos, *cp, *cp2;
  181. uint32_t *cur_lp, *ref_lp;
  182. const uint32_t *correction_lp[2], *correctionloworder_lp[2], *correctionhighorder_lp[2];
  183. uint8_t *correction_type_sp[2];
  184. ustr_t strip_tbl[20], *strip;
  185. int i, j, k, lp1, lp2, flag1, cmd, blks_width, blks_height, region_160_width,
  186. rle_v1, rle_v2, rle_v3;
  187. unsigned short res;
  188. bit_buf = 0;
  189. ref_vectors = NULL;
  190. width_tbl = width_tbl_arr + 1;
  191. i = (width < 0 ? width + 3 : width)/4;
  192. for(j = -1; j < 8; j++)
  193. width_tbl[j] = i * j;
  194. strip = strip_tbl;
  195. for(region_160_width = 0; region_160_width < (width - min_width_160); region_160_width += min_width_160);
  196. strip->ypos = strip->xpos = 0;
  197. for(strip->width = min_width_160; width > strip->width; strip->width *= 2);
  198. strip->height = height;
  199. strip->split_direction = 0;
  200. strip->split_flag = 0;
  201. strip->usl7 = 0;
  202. bit_pos = 0;
  203. rle_v1 = rle_v2 = rle_v3 = 0;
  204. while(strip >= strip_tbl) {
  205. if(bit_pos <= 0) {
  206. bit_pos = 8;
  207. bit_buf = *buf1++;
  208. }
  209. bit_pos -= 2;
  210. cmd = (bit_buf >> bit_pos) & 0x03;
  211. if(cmd == 0) {
  212. strip++;
  213. memcpy(strip, strip-1, sizeof(ustr_t));
  214. strip->split_flag = 1;
  215. strip->split_direction = 0;
  216. strip->height = (strip->height > 8 ? ((strip->height+8)>>4)<<3 : 4);
  217. continue;
  218. } else if(cmd == 1) {
  219. strip++;
  220. memcpy(strip, strip-1, sizeof(ustr_t));
  221. strip->split_flag = 1;
  222. strip->split_direction = 1;
  223. strip->width = (strip->width > 8 ? ((strip->width+8)>>4)<<3 : 4);
  224. continue;
  225. } else if(cmd == 2) {
  226. if(strip->usl7 == 0) {
  227. strip->usl7 = 1;
  228. ref_vectors = NULL;
  229. continue;
  230. }
  231. } else if(cmd == 3) {
  232. if(strip->usl7 == 0) {
  233. strip->usl7 = 1;
  234. ref_vectors = (const signed char*)buf2 + (*buf1 * 2);
  235. buf1++;
  236. continue;
  237. }
  238. }
  239. cur_frm_pos = cur + width * strip->ypos + strip->xpos;
  240. if((blks_width = strip->width) < 0)
  241. blks_width += 3;
  242. blks_width >>= 2;
  243. blks_height = strip->height;
  244. if(ref_vectors != NULL) {
  245. ref_frm_pos = ref + (ref_vectors[0] + strip->ypos) * width +
  246. ref_vectors[1] + strip->xpos;
  247. } else
  248. ref_frm_pos = cur_frm_pos - width_tbl[4];
  249. if(cmd == 2) {
  250. if(bit_pos <= 0) {
  251. bit_pos = 8;
  252. bit_buf = *buf1++;
  253. }
  254. bit_pos -= 2;
  255. cmd = (bit_buf >> bit_pos) & 0x03;
  256. if(cmd == 0 || ref_vectors != NULL) {
  257. for(lp1 = 0; lp1 < blks_width; lp1++) {
  258. for(i = 0, j = 0; i < blks_height; i++, j += width_tbl[1])
  259. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  260. cur_frm_pos += 4;
  261. ref_frm_pos += 4;
  262. }
  263. } else if(cmd != 1)
  264. return;
  265. } else {
  266. k = *buf1 >> 4;
  267. j = *buf1 & 0x0f;
  268. buf1++;
  269. lv = j + cb_offset;
  270. if((lv - 8) <= 7 && (k == 0 || k == 3 || k == 10)) {
  271. cp2 = s->ModPred + ((lv - 8) << 7);
  272. cp = ref_frm_pos;
  273. for(i = 0; i < blks_width << 2; i++) {
  274. int v = *cp >> 1;
  275. *(cp++) = cp2[v];
  276. }
  277. }
  278. if(k == 1 || k == 4) {
  279. lv = (hdr[j] & 0xf) + cb_offset;
  280. correction_type_sp[0] = s->corrector_type + (lv << 8);
  281. correction_lp[0] = correction + (lv << 8);
  282. lv = (hdr[j] >> 4) + cb_offset;
  283. correction_lp[1] = correction + (lv << 8);
  284. correction_type_sp[1] = s->corrector_type + (lv << 8);
  285. } else {
  286. correctionloworder_lp[0] = correctionloworder_lp[1] = correctionloworder + (lv << 8);
  287. correctionhighorder_lp[0] = correctionhighorder_lp[1] = correctionhighorder + (lv << 8);
  288. correction_type_sp[0] = correction_type_sp[1] = s->corrector_type + (lv << 8);
  289. correction_lp[0] = correction_lp[1] = correction + (lv << 8);
  290. }
  291. switch(k) {
  292. case 1:
  293. case 0: /********** CASE 0 **********/
  294. for( ; blks_height > 0; blks_height -= 4) {
  295. for(lp1 = 0; lp1 < blks_width; lp1++) {
  296. for(lp2 = 0; lp2 < 4; ) {
  297. k = *buf1++;
  298. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2];
  299. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2];
  300. switch(correction_type_sp[0][k]) {
  301. case 0:
  302. *cur_lp = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  303. lp2++;
  304. break;
  305. case 1:
  306. res = ((le2me_16(((unsigned short *)(ref_lp))[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  307. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  308. res = ((le2me_16(((unsigned short *)(ref_lp))[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  309. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  310. buf1++;
  311. lp2++;
  312. break;
  313. case 2:
  314. if(lp2 == 0) {
  315. for(i = 0, j = 0; i < 2; i++, j += width_tbl[1])
  316. cur_lp[j] = ref_lp[j];
  317. lp2 += 2;
  318. }
  319. break;
  320. case 3:
  321. if(lp2 < 2) {
  322. for(i = 0, j = 0; i < (3 - lp2); i++, j += width_tbl[1])
  323. cur_lp[j] = ref_lp[j];
  324. lp2 = 3;
  325. }
  326. break;
  327. case 8:
  328. if(lp2 == 0) {
  329. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  330. if(rle_v1 == 1 || ref_vectors != NULL) {
  331. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  332. cur_lp[j] = ref_lp[j];
  333. }
  334. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  335. break;
  336. } else {
  337. rle_v1 = 1;
  338. rle_v2 = *buf1 - 1;
  339. }
  340. case 5:
  341. LP2_CHECK(buf1,rle_v3,lp2)
  342. case 4:
  343. for(i = 0, j = 0; i < (4 - lp2); i++, j += width_tbl[1])
  344. cur_lp[j] = ref_lp[j];
  345. lp2 = 4;
  346. break;
  347. case 7:
  348. if(rle_v3 != 0)
  349. rle_v3 = 0;
  350. else {
  351. buf1--;
  352. rle_v3 = 1;
  353. }
  354. case 6:
  355. if(ref_vectors != NULL) {
  356. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  357. cur_lp[j] = ref_lp[j];
  358. }
  359. lp2 = 4;
  360. break;
  361. case 9:
  362. lv1 = *buf1++;
  363. lv = (lv1 & 0x7F) << 1;
  364. lv += (lv << 8);
  365. lv += (lv << 16);
  366. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  367. cur_lp[j] = lv;
  368. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  369. break;
  370. default:
  371. return;
  372. }
  373. }
  374. cur_frm_pos += 4;
  375. ref_frm_pos += 4;
  376. }
  377. cur_frm_pos += ((width - blks_width) * 4);
  378. ref_frm_pos += ((width - blks_width) * 4);
  379. }
  380. break;
  381. case 4:
  382. case 3: /********** CASE 3 **********/
  383. if(ref_vectors != NULL)
  384. return;
  385. flag1 = 1;
  386. for( ; blks_height > 0; blks_height -= 8) {
  387. for(lp1 = 0; lp1 < blks_width; lp1++) {
  388. for(lp2 = 0; lp2 < 4; ) {
  389. k = *buf1++;
  390. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  391. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  392. switch(correction_type_sp[lp2 & 0x01][k]) {
  393. case 0:
  394. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  395. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  396. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  397. else
  398. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  399. lp2++;
  400. break;
  401. case 1:
  402. res = ((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  403. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  404. res = ((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  405. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  406. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  407. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  408. else
  409. cur_lp[0] = cur_lp[width_tbl[1]];
  410. buf1++;
  411. lp2++;
  412. break;
  413. case 2:
  414. if(lp2 == 0) {
  415. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  416. cur_lp[j] = *ref_lp;
  417. lp2 += 2;
  418. }
  419. break;
  420. case 3:
  421. if(lp2 < 2) {
  422. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  423. cur_lp[j] = *ref_lp;
  424. lp2 = 3;
  425. }
  426. break;
  427. case 6:
  428. lp2 = 4;
  429. break;
  430. case 7:
  431. if(rle_v3 != 0)
  432. rle_v3 = 0;
  433. else {
  434. buf1--;
  435. rle_v3 = 1;
  436. }
  437. lp2 = 4;
  438. break;
  439. case 8:
  440. if(lp2 == 0) {
  441. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  442. if(rle_v1 == 1) {
  443. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  444. cur_lp[j] = ref_lp[j];
  445. }
  446. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  447. break;
  448. } else {
  449. rle_v2 = (*buf1) - 1;
  450. rle_v1 = 1;
  451. }
  452. case 5:
  453. LP2_CHECK(buf1,rle_v3,lp2)
  454. case 4:
  455. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  456. cur_lp[j] = *ref_lp;
  457. lp2 = 4;
  458. break;
  459. case 9:
  460. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  461. lv1 = *buf1++;
  462. lv = (lv1 & 0x7F) << 1;
  463. lv += (lv << 8);
  464. lv += (lv << 16);
  465. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  466. cur_lp[j] = lv;
  467. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  468. break;
  469. default:
  470. return;
  471. }
  472. }
  473. cur_frm_pos += 4;
  474. }
  475. cur_frm_pos += (((width * 2) - blks_width) * 4);
  476. flag1 = 0;
  477. }
  478. break;
  479. case 10: /********** CASE 10 **********/
  480. if(ref_vectors == NULL) {
  481. flag1 = 1;
  482. for( ; blks_height > 0; blks_height -= 8) {
  483. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  484. for(lp2 = 0; lp2 < 4; ) {
  485. k = *buf1++;
  486. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  487. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  488. lv1 = ref_lp[0];
  489. lv2 = ref_lp[1];
  490. if(lp2 == 0 && flag1 != 0) {
  491. #ifdef WORDS_BIGENDIAN
  492. lv1 = lv1 & 0xFF00FF00;
  493. lv1 = (lv1 >> 8) | lv1;
  494. lv2 = lv2 & 0xFF00FF00;
  495. lv2 = (lv2 >> 8) | lv2;
  496. #else
  497. lv1 = lv1 & 0x00FF00FF;
  498. lv1 = (lv1 << 8) | lv1;
  499. lv2 = lv2 & 0x00FF00FF;
  500. lv2 = (lv2 << 8) | lv2;
  501. #endif
  502. }
  503. switch(correction_type_sp[lp2 & 0x01][k]) {
  504. case 0:
  505. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  506. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionhighorder_lp[lp2 & 0x01][k]) << 1);
  507. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  508. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  509. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  510. } else {
  511. cur_lp[0] = cur_lp[width_tbl[1]];
  512. cur_lp[1] = cur_lp[width_tbl[1]+1];
  513. }
  514. lp2++;
  515. break;
  516. case 1:
  517. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][*buf1]) << 1);
  518. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  519. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  520. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  521. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  522. } else {
  523. cur_lp[0] = cur_lp[width_tbl[1]];
  524. cur_lp[1] = cur_lp[width_tbl[1]+1];
  525. }
  526. buf1++;
  527. lp2++;
  528. break;
  529. case 2:
  530. if(lp2 == 0) {
  531. if(flag1 != 0) {
  532. for(i = 0, j = width_tbl[1]; i < 3; i++, j += width_tbl[1]) {
  533. cur_lp[j] = lv1;
  534. cur_lp[j+1] = lv2;
  535. }
  536. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  537. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  538. } else {
  539. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  540. cur_lp[j] = lv1;
  541. cur_lp[j+1] = lv2;
  542. }
  543. }
  544. lp2 += 2;
  545. }
  546. break;
  547. case 3:
  548. if(lp2 < 2) {
  549. if(lp2 == 0 && flag1 != 0) {
  550. for(i = 0, j = width_tbl[1]; i < 5; i++, j += width_tbl[1]) {
  551. cur_lp[j] = lv1;
  552. cur_lp[j+1] = lv2;
  553. }
  554. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  555. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  556. } else {
  557. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  558. cur_lp[j] = lv1;
  559. cur_lp[j+1] = lv2;
  560. }
  561. }
  562. lp2 = 3;
  563. }
  564. break;
  565. case 8:
  566. if(lp2 == 0) {
  567. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  568. if(rle_v1 == 1) {
  569. if(flag1 != 0) {
  570. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  571. cur_lp[j] = lv1;
  572. cur_lp[j+1] = lv2;
  573. }
  574. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  575. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  576. } else {
  577. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  578. cur_lp[j] = lv1;
  579. cur_lp[j+1] = lv2;
  580. }
  581. }
  582. }
  583. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  584. break;
  585. } else {
  586. rle_v1 = 1;
  587. rle_v2 = (*buf1) - 1;
  588. }
  589. case 5:
  590. LP2_CHECK(buf1,rle_v3,lp2)
  591. case 4:
  592. if(lp2 == 0 && flag1 != 0) {
  593. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  594. cur_lp[j] = lv1;
  595. cur_lp[j+1] = lv2;
  596. }
  597. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  598. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  599. } else {
  600. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  601. cur_lp[j] = lv1;
  602. cur_lp[j+1] = lv2;
  603. }
  604. }
  605. lp2 = 4;
  606. break;
  607. case 6:
  608. lp2 = 4;
  609. break;
  610. case 7:
  611. if(lp2 == 0) {
  612. if(rle_v3 != 0)
  613. rle_v3 = 0;
  614. else {
  615. buf1--;
  616. rle_v3 = 1;
  617. }
  618. lp2 = 4;
  619. }
  620. break;
  621. case 9:
  622. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  623. lv1 = *buf1;
  624. lv = (lv1 & 0x7F) << 1;
  625. lv += (lv << 8);
  626. lv += (lv << 16);
  627. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  628. cur_lp[j] = lv;
  629. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  630. break;
  631. default:
  632. return;
  633. }
  634. }
  635. cur_frm_pos += 8;
  636. }
  637. cur_frm_pos += (((width * 2) - blks_width) * 4);
  638. flag1 = 0;
  639. }
  640. } else {
  641. for( ; blks_height > 0; blks_height -= 8) {
  642. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  643. for(lp2 = 0; lp2 < 4; ) {
  644. k = *buf1++;
  645. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  646. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  647. switch(correction_type_sp[lp2 & 0x01][k]) {
  648. case 0:
  649. lv1 = correctionloworder_lp[lp2 & 0x01][k];
  650. lv2 = correctionhighorder_lp[lp2 & 0x01][k];
  651. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  652. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  653. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  654. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  655. lp2++;
  656. break;
  657. case 1:
  658. lv1 = correctionloworder_lp[lp2 & 0x01][*buf1++];
  659. lv2 = correctionloworder_lp[lp2 & 0x01][k];
  660. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  661. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  662. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  663. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  664. lp2++;
  665. break;
  666. case 2:
  667. if(lp2 == 0) {
  668. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  669. cur_lp[j] = ref_lp[j];
  670. cur_lp[j+1] = ref_lp[j+1];
  671. }
  672. lp2 += 2;
  673. }
  674. break;
  675. case 3:
  676. if(lp2 < 2) {
  677. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  678. cur_lp[j] = ref_lp[j];
  679. cur_lp[j+1] = ref_lp[j+1];
  680. }
  681. lp2 = 3;
  682. }
  683. break;
  684. case 8:
  685. if(lp2 == 0) {
  686. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  687. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  688. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  689. ((uint32_t *)cur_frm_pos)[j+1] = ((uint32_t *)ref_frm_pos)[j+1];
  690. }
  691. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  692. break;
  693. } else {
  694. rle_v1 = 1;
  695. rle_v2 = (*buf1) - 1;
  696. }
  697. case 5:
  698. case 7:
  699. LP2_CHECK(buf1,rle_v3,lp2)
  700. case 6:
  701. case 4:
  702. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  703. cur_lp[j] = ref_lp[j];
  704. cur_lp[j+1] = ref_lp[j+1];
  705. }
  706. lp2 = 4;
  707. break;
  708. case 9:
  709. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  710. lv1 = *buf1;
  711. lv = (lv1 & 0x7F) << 1;
  712. lv += (lv << 8);
  713. lv += (lv << 16);
  714. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  715. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)cur_frm_pos)[j+1] = lv;
  716. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  717. break;
  718. default:
  719. return;
  720. }
  721. }
  722. cur_frm_pos += 8;
  723. ref_frm_pos += 8;
  724. }
  725. cur_frm_pos += (((width * 2) - blks_width) * 4);
  726. ref_frm_pos += (((width * 2) - blks_width) * 4);
  727. }
  728. }
  729. break;
  730. case 11: /********** CASE 11 **********/
  731. if(ref_vectors == NULL)
  732. return;
  733. for( ; blks_height > 0; blks_height -= 8) {
  734. for(lp1 = 0; lp1 < blks_width; lp1++) {
  735. for(lp2 = 0; lp2 < 4; ) {
  736. k = *buf1++;
  737. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  738. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  739. switch(correction_type_sp[lp2 & 0x01][k]) {
  740. case 0:
  741. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  742. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  743. lp2++;
  744. break;
  745. case 1:
  746. lv1 = (unsigned short)(correction_lp[lp2 & 0x01][*buf1++]);
  747. lv2 = (unsigned short)(correction_lp[lp2 & 0x01][k]);
  748. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + lv1) << 1);
  749. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  750. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + lv2) << 1);
  751. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  752. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]]) >> 1) + lv1) << 1);
  753. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  754. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]+1]) >> 1) + lv2) << 1);
  755. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  756. lp2++;
  757. break;
  758. case 2:
  759. if(lp2 == 0) {
  760. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  761. cur_lp[j] = ref_lp[j];
  762. lp2 += 2;
  763. }
  764. break;
  765. case 3:
  766. if(lp2 < 2) {
  767. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  768. cur_lp[j] = ref_lp[j];
  769. lp2 = 3;
  770. }
  771. break;
  772. case 8:
  773. if(lp2 == 0) {
  774. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  775. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  776. cur_lp[j] = ref_lp[j];
  777. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  778. break;
  779. } else {
  780. rle_v1 = 1;
  781. rle_v2 = (*buf1) - 1;
  782. }
  783. case 5:
  784. case 7:
  785. LP2_CHECK(buf1,rle_v3,lp2)
  786. case 4:
  787. case 6:
  788. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  789. cur_lp[j] = ref_lp[j];
  790. lp2 = 4;
  791. break;
  792. case 9:
  793. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  794. lv1 = *buf1++;
  795. lv = (lv1 & 0x7F) << 1;
  796. lv += (lv << 8);
  797. lv += (lv << 16);
  798. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  799. cur_lp[j] = lv;
  800. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  801. break;
  802. default:
  803. return;
  804. }
  805. }
  806. cur_frm_pos += 4;
  807. ref_frm_pos += 4;
  808. }
  809. cur_frm_pos += (((width * 2) - blks_width) * 4);
  810. ref_frm_pos += (((width * 2) - blks_width) * 4);
  811. }
  812. break;
  813. default:
  814. return;
  815. }
  816. }
  817. if(strip < strip_tbl)
  818. return;
  819. for( ; strip >= strip_tbl; strip--) {
  820. if(strip->split_flag != 0) {
  821. strip->split_flag = 0;
  822. strip->usl7 = (strip-1)->usl7;
  823. if(strip->split_direction) {
  824. strip->xpos += strip->width;
  825. strip->width = (strip-1)->width - strip->width;
  826. if(region_160_width <= strip->xpos && width < strip->width + strip->xpos)
  827. strip->width = width - strip->xpos;
  828. } else {
  829. strip->ypos += strip->height;
  830. strip->height = (strip-1)->height - strip->height;
  831. }
  832. break;
  833. }
  834. }
  835. }
  836. }
  837. static av_cold int indeo3_decode_init(AVCodecContext *avctx)
  838. {
  839. Indeo3DecodeContext *s = avctx->priv_data;
  840. int ret = 0;
  841. s->avctx = avctx;
  842. s->width = avctx->width;
  843. s->height = avctx->height;
  844. avctx->pix_fmt = PIX_FMT_YUV410P;
  845. if (!(ret = build_modpred(s)))
  846. ret = iv_alloc_frames(s);
  847. if (ret)
  848. iv_free_func(s);
  849. return ret;
  850. }
  851. static unsigned long iv_decode_frame(Indeo3DecodeContext *s,
  852. const uint8_t *buf, int buf_size)
  853. {
  854. unsigned int image_width, image_height,
  855. chroma_width, chroma_height;
  856. unsigned long flags, cb_offset, data_size,
  857. y_offset, v_offset, u_offset, mc_vector_count;
  858. const uint8_t *hdr_pos, *buf_pos;
  859. buf_pos = buf;
  860. buf_pos += 18; /* skip OS header (16 bytes) and version number */
  861. flags = bytestream_get_le16(&buf_pos);
  862. data_size = bytestream_get_le32(&buf_pos);
  863. cb_offset = *buf_pos++;
  864. buf_pos += 3; /* skip reserved byte and checksum */
  865. image_height = bytestream_get_le16(&buf_pos);
  866. image_width = bytestream_get_le16(&buf_pos);
  867. if(avcodec_check_dimensions(NULL, image_width, image_height))
  868. return -1;
  869. chroma_height = ((image_height >> 2) + 3) & 0x7ffc;
  870. chroma_width = ((image_width >> 2) + 3) & 0x7ffc;
  871. y_offset = bytestream_get_le32(&buf_pos);
  872. v_offset = bytestream_get_le32(&buf_pos);
  873. u_offset = bytestream_get_le32(&buf_pos);
  874. buf_pos += 4; /* reserved */
  875. hdr_pos = buf_pos;
  876. if(data_size == 0x80) return 4;
  877. if(flags & 0x200) {
  878. s->cur_frame = s->iv_frame + 1;
  879. s->ref_frame = s->iv_frame;
  880. } else {
  881. s->cur_frame = s->iv_frame;
  882. s->ref_frame = s->iv_frame + 1;
  883. }
  884. buf_pos = buf + 16 + y_offset;
  885. mc_vector_count = bytestream_get_le32(&buf_pos);
  886. iv_Decode_Chunk(s, s->cur_frame->Ybuf, s->ref_frame->Ybuf, image_width,
  887. image_height, buf_pos + mc_vector_count * 2, cb_offset, hdr_pos, buf_pos,
  888. FFMIN(image_width, 160));
  889. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  890. {
  891. buf_pos = buf + 16 + v_offset;
  892. mc_vector_count = bytestream_get_le32(&buf_pos);
  893. iv_Decode_Chunk(s, s->cur_frame->Vbuf, s->ref_frame->Vbuf, chroma_width,
  894. chroma_height, buf_pos + mc_vector_count * 2, cb_offset, hdr_pos, buf_pos,
  895. FFMIN(chroma_width, 40));
  896. buf_pos = buf + 16 + u_offset;
  897. mc_vector_count = bytestream_get_le32(&buf_pos);
  898. iv_Decode_Chunk(s, s->cur_frame->Ubuf, s->ref_frame->Ubuf, chroma_width,
  899. chroma_height, buf_pos + mc_vector_count * 2, cb_offset, hdr_pos, buf_pos,
  900. FFMIN(chroma_width, 40));
  901. }
  902. return 8;
  903. }
  904. static int indeo3_decode_frame(AVCodecContext *avctx,
  905. void *data, int *data_size,
  906. const uint8_t *buf, int buf_size)
  907. {
  908. Indeo3DecodeContext *s=avctx->priv_data;
  909. uint8_t *src, *dest;
  910. int y;
  911. iv_decode_frame(s, buf, buf_size);
  912. if(s->frame.data[0])
  913. avctx->release_buffer(avctx, &s->frame);
  914. s->frame.reference = 0;
  915. if(avctx->get_buffer(avctx, &s->frame) < 0) {
  916. av_log(s->avctx, AV_LOG_ERROR, "get_buffer() failed\n");
  917. return -1;
  918. }
  919. src = s->cur_frame->Ybuf;
  920. dest = s->frame.data[0];
  921. for (y = 0; y < s->height; y++) {
  922. memcpy(dest, src, s->cur_frame->y_w);
  923. src += s->cur_frame->y_w;
  924. dest += s->frame.linesize[0];
  925. }
  926. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  927. {
  928. src = s->cur_frame->Ubuf;
  929. dest = s->frame.data[1];
  930. for (y = 0; y < s->height / 4; y++) {
  931. memcpy(dest, src, s->cur_frame->uv_w);
  932. src += s->cur_frame->uv_w;
  933. dest += s->frame.linesize[1];
  934. }
  935. src = s->cur_frame->Vbuf;
  936. dest = s->frame.data[2];
  937. for (y = 0; y < s->height / 4; y++) {
  938. memcpy(dest, src, s->cur_frame->uv_w);
  939. src += s->cur_frame->uv_w;
  940. dest += s->frame.linesize[2];
  941. }
  942. }
  943. *data_size=sizeof(AVFrame);
  944. *(AVFrame*)data= s->frame;
  945. return buf_size;
  946. }
  947. static av_cold int indeo3_decode_end(AVCodecContext *avctx)
  948. {
  949. Indeo3DecodeContext *s = avctx->priv_data;
  950. iv_free_func(s);
  951. return 0;
  952. }
  953. AVCodec indeo3_decoder = {
  954. "indeo3",
  955. CODEC_TYPE_VIDEO,
  956. CODEC_ID_INDEO3,
  957. sizeof(Indeo3DecodeContext),
  958. indeo3_decode_init,
  959. NULL,
  960. indeo3_decode_end,
  961. indeo3_decode_frame,
  962. 0,
  963. NULL,
  964. .long_name = NULL_IF_CONFIG_SMALL("Intel Indeo 3"),
  965. };