You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1120 lines
46KB

  1. /*
  2. * Intel Indeo 3 (IV31, IV32, etc.) video decoder for ffmpeg
  3. * written, produced, and directed by Alan Smithee
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include <stdio.h>
  22. #include <stdlib.h>
  23. #include <string.h>
  24. #include <unistd.h>
  25. #include "avcodec.h"
  26. #include "dsputil.h"
  27. #include "bytestream.h"
  28. #include "indeo3data.h"
  29. typedef struct
  30. {
  31. uint8_t *Ybuf;
  32. uint8_t *Ubuf;
  33. uint8_t *Vbuf;
  34. unsigned int the_buf_size;
  35. unsigned short y_w, y_h;
  36. unsigned short uv_w, uv_h;
  37. } YUVBufs;
  38. typedef struct Indeo3DecodeContext {
  39. AVCodecContext *avctx;
  40. int width, height;
  41. AVFrame frame;
  42. uint8_t *buf;
  43. YUVBufs iv_frame[2];
  44. YUVBufs *cur_frame;
  45. YUVBufs *ref_frame;
  46. uint8_t *ModPred;
  47. uint8_t *corrector_type;
  48. } Indeo3DecodeContext;
  49. static const uint8_t corrector_type_0[24] = {
  50. 195, 159, 133, 115, 101, 93, 87, 77,
  51. 195, 159, 133, 115, 101, 93, 87, 77,
  52. 128, 79, 79, 79, 79, 79, 79, 79
  53. };
  54. static const uint8_t corrector_type_2[8] = { 9, 7, 6, 8, 5, 4, 3, 2 };
  55. static av_cold int build_modpred(Indeo3DecodeContext *s)
  56. {
  57. int i, j;
  58. if (!(s->ModPred = av_malloc(8 * 128)))
  59. return AVERROR(ENOMEM);
  60. for (i=0; i < 128; ++i) {
  61. s->ModPred[i+0*128] = i > 126 ? 254 : 2*(i + 1 - ((i + 1) % 2));
  62. s->ModPred[i+1*128] = i == 7 ? 20 :
  63. i == 119 ||
  64. i == 120 ? 236 : 2*(i + 2 - ((i + 1) % 3));
  65. s->ModPred[i+2*128] = i > 125 ? 248 : 2*(i + 2 - ((i + 2) % 4));
  66. s->ModPred[i+3*128] = 2*(i + 1 - ((i - 3) % 5));
  67. s->ModPred[i+4*128] = i == 8 ? 20 : 2*(i + 1 - ((i - 3) % 6));
  68. s->ModPred[i+5*128] = 2*(i + 4 - ((i + 3) % 7));
  69. s->ModPred[i+6*128] = i > 123 ? 240 : 2*(i + 4 - ((i + 4) % 8));
  70. s->ModPred[i+7*128] = 2*(i + 5 - ((i + 4) % 9));
  71. }
  72. if (!(s->corrector_type = av_malloc(24 * 256)))
  73. return AVERROR(ENOMEM);
  74. for (i=0; i < 24; ++i) {
  75. for (j=0; j < 256; ++j) {
  76. s->corrector_type[i*256+j] = j < corrector_type_0[i] ? 1 :
  77. j < 248 || (i == 16 && j == 248) ? 0 :
  78. corrector_type_2[j - 248];
  79. }
  80. }
  81. return 0;
  82. }
  83. static av_cold int iv_alloc_frames(Indeo3DecodeContext *s)
  84. {
  85. int luma_width, luma_height, luma_pixels, chroma_width, chroma_height,
  86. chroma_pixels, i;
  87. unsigned int bufsize;
  88. luma_width = (s->width + 3) & (~3);
  89. luma_height = (s->height + 3) & (~3);
  90. chroma_width = ((luma_width >> 2) + 3) & (~3);
  91. chroma_height = ((luma_height>> 2) + 3) & (~3);
  92. luma_pixels = luma_width * luma_height;
  93. chroma_pixels = chroma_width * chroma_height;
  94. bufsize = luma_pixels * 2 + luma_width * 3 +
  95. (chroma_pixels + chroma_width) * 4;
  96. if(!(s->buf = av_malloc(bufsize)))
  97. return AVERROR(ENOMEM);
  98. s->iv_frame[0].y_w = s->iv_frame[1].y_w = luma_width;
  99. s->iv_frame[0].y_h = s->iv_frame[1].y_h = luma_height;
  100. s->iv_frame[0].uv_w = s->iv_frame[1].uv_w = chroma_width;
  101. s->iv_frame[0].uv_h = s->iv_frame[1].uv_h = chroma_height;
  102. s->iv_frame[0].the_buf_size = bufsize;
  103. s->iv_frame[0].Ybuf = s->buf + luma_width;
  104. i = luma_pixels + luma_width * 2;
  105. s->iv_frame[1].Ybuf = s->buf + i;
  106. i += (luma_pixels + luma_width);
  107. s->iv_frame[0].Ubuf = s->buf + i;
  108. i += (chroma_pixels + chroma_width);
  109. s->iv_frame[1].Ubuf = s->buf + i;
  110. i += (chroma_pixels + chroma_width);
  111. s->iv_frame[0].Vbuf = s->buf + i;
  112. i += (chroma_pixels + chroma_width);
  113. s->iv_frame[1].Vbuf = s->buf + i;
  114. for(i = 1; i <= luma_width; i++)
  115. s->iv_frame[0].Ybuf[-i] = s->iv_frame[1].Ybuf[-i] =
  116. s->iv_frame[0].Ubuf[-i] = 0x80;
  117. for(i = 1; i <= chroma_width; i++) {
  118. s->iv_frame[1].Ubuf[-i] = 0x80;
  119. s->iv_frame[0].Vbuf[-i] = 0x80;
  120. s->iv_frame[1].Vbuf[-i] = 0x80;
  121. s->iv_frame[1].Vbuf[chroma_pixels+i-1] = 0x80;
  122. }
  123. return 0;
  124. }
  125. static av_cold void iv_free_func(Indeo3DecodeContext *s)
  126. {
  127. av_free(s->buf);
  128. av_free(s->ModPred);
  129. av_free(s->corrector_type);
  130. }
  131. typedef struct {
  132. long xpos;
  133. long ypos;
  134. long width;
  135. long height;
  136. long split_flag;
  137. long split_direction;
  138. long usl7;
  139. } ustr_t;
  140. #define LV1_CHECK(buf1,rle_v3,lv1,lp2) \
  141. if((lv1 & 0x80) != 0) { \
  142. if(rle_v3 != 0) \
  143. rle_v3 = 0; \
  144. else { \
  145. rle_v3 = 1; \
  146. buf1 -= 2; \
  147. } \
  148. } \
  149. lp2 = 4;
  150. #define RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3) \
  151. if(rle_v3 == 0) { \
  152. rle_v2 = *buf1; \
  153. rle_v1 = 1; \
  154. if(rle_v2 > 32) { \
  155. rle_v2 -= 32; \
  156. rle_v1 = 0; \
  157. } \
  158. rle_v3 = 1; \
  159. } \
  160. buf1--;
  161. #define LP2_CHECK(buf1,rle_v3,lp2) \
  162. if(lp2 == 0 && rle_v3 != 0) \
  163. rle_v3 = 0; \
  164. else { \
  165. buf1--; \
  166. rle_v3 = 1; \
  167. }
  168. #define RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2) \
  169. rle_v2--; \
  170. if(rle_v2 == 0) { \
  171. rle_v3 = 0; \
  172. buf1 += 2; \
  173. } \
  174. lp2 = 4;
  175. static void iv_Decode_Chunk(Indeo3DecodeContext *s,
  176. uint8_t *cur, uint8_t *ref, int width, int height,
  177. const uint8_t *buf1, long fflags2, const uint8_t *hdr,
  178. const uint8_t *buf2, int min_width_160)
  179. {
  180. uint8_t bit_buf;
  181. unsigned long bit_pos, lv, lv1, lv2;
  182. long *width_tbl, width_tbl_arr[10];
  183. const signed char *ref_vectors;
  184. uint8_t *cur_frm_pos, *ref_frm_pos, *cp, *cp2;
  185. uint32_t *cur_lp, *ref_lp;
  186. const uint32_t *correction_lp[2], *correctionloworder_lp[2], *correctionhighorder_lp[2];
  187. uint8_t *correction_type_sp[2];
  188. ustr_t strip_tbl[20], *strip;
  189. int i, j, k, lp1, lp2, flag1, cmd, blks_width, blks_height, region_160_width,
  190. rle_v1, rle_v2, rle_v3;
  191. unsigned short res;
  192. bit_buf = 0;
  193. ref_vectors = NULL;
  194. width_tbl = width_tbl_arr + 1;
  195. i = (width < 0 ? width + 3 : width)/4;
  196. for(j = -1; j < 8; j++)
  197. width_tbl[j] = i * j;
  198. strip = strip_tbl;
  199. for(region_160_width = 0; region_160_width < (width - min_width_160); region_160_width += min_width_160);
  200. strip->ypos = strip->xpos = 0;
  201. for(strip->width = min_width_160; width > strip->width; strip->width *= 2);
  202. strip->height = height;
  203. strip->split_direction = 0;
  204. strip->split_flag = 0;
  205. strip->usl7 = 0;
  206. bit_pos = 0;
  207. rle_v1 = rle_v2 = rle_v3 = 0;
  208. while(strip >= strip_tbl) {
  209. if(bit_pos <= 0) {
  210. bit_pos = 8;
  211. bit_buf = *buf1++;
  212. }
  213. bit_pos -= 2;
  214. cmd = (bit_buf >> bit_pos) & 0x03;
  215. if(cmd == 0) {
  216. strip++;
  217. memcpy(strip, strip-1, sizeof(ustr_t));
  218. strip->split_flag = 1;
  219. strip->split_direction = 0;
  220. strip->height = (strip->height > 8 ? ((strip->height+8)>>4)<<3 : 4);
  221. continue;
  222. } else if(cmd == 1) {
  223. strip++;
  224. memcpy(strip, strip-1, sizeof(ustr_t));
  225. strip->split_flag = 1;
  226. strip->split_direction = 1;
  227. strip->width = (strip->width > 8 ? ((strip->width+8)>>4)<<3 : 4);
  228. continue;
  229. } else if(cmd == 2) {
  230. if(strip->usl7 == 0) {
  231. strip->usl7 = 1;
  232. ref_vectors = NULL;
  233. continue;
  234. }
  235. } else if(cmd == 3) {
  236. if(strip->usl7 == 0) {
  237. strip->usl7 = 1;
  238. ref_vectors = (const signed char*)buf2 + (*buf1 * 2);
  239. buf1++;
  240. continue;
  241. }
  242. }
  243. cur_frm_pos = cur + width * strip->ypos + strip->xpos;
  244. if((blks_width = strip->width) < 0)
  245. blks_width += 3;
  246. blks_width >>= 2;
  247. blks_height = strip->height;
  248. if(ref_vectors != NULL) {
  249. ref_frm_pos = ref + (ref_vectors[0] + strip->ypos) * width +
  250. ref_vectors[1] + strip->xpos;
  251. } else
  252. ref_frm_pos = cur_frm_pos - width_tbl[4];
  253. if(cmd == 2) {
  254. if(bit_pos <= 0) {
  255. bit_pos = 8;
  256. bit_buf = *buf1++;
  257. }
  258. bit_pos -= 2;
  259. cmd = (bit_buf >> bit_pos) & 0x03;
  260. if(cmd == 0 || ref_vectors != NULL) {
  261. for(lp1 = 0; lp1 < blks_width; lp1++) {
  262. for(i = 0, j = 0; i < blks_height; i++, j += width_tbl[1])
  263. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  264. cur_frm_pos += 4;
  265. ref_frm_pos += 4;
  266. }
  267. } else if(cmd != 1)
  268. return;
  269. } else {
  270. k = *buf1 >> 4;
  271. j = *buf1 & 0x0f;
  272. buf1++;
  273. lv = j + fflags2;
  274. if((lv - 8) <= 7 && (k == 0 || k == 3 || k == 10)) {
  275. cp2 = s->ModPred + ((lv - 8) << 7);
  276. cp = ref_frm_pos;
  277. for(i = 0; i < blks_width << 2; i++) {
  278. int v = *cp >> 1;
  279. *(cp++) = cp2[v];
  280. }
  281. }
  282. if(k == 1 || k == 4) {
  283. lv = (hdr[j] & 0xf) + fflags2;
  284. correction_type_sp[0] = s->corrector_type + (lv << 8);
  285. correction_lp[0] = correction + (lv << 8);
  286. lv = (hdr[j] >> 4) + fflags2;
  287. correction_lp[1] = correction + (lv << 8);
  288. correction_type_sp[1] = s->corrector_type + (lv << 8);
  289. } else {
  290. correctionloworder_lp[0] = correctionloworder_lp[1] = correctionloworder + (lv << 8);
  291. correctionhighorder_lp[0] = correctionhighorder_lp[1] = correctionhighorder + (lv << 8);
  292. correction_type_sp[0] = correction_type_sp[1] = s->corrector_type + (lv << 8);
  293. correction_lp[0] = correction_lp[1] = correction + (lv << 8);
  294. }
  295. switch(k) {
  296. case 1:
  297. case 0: /********** CASE 0 **********/
  298. for( ; blks_height > 0; blks_height -= 4) {
  299. for(lp1 = 0; lp1 < blks_width; lp1++) {
  300. for(lp2 = 0; lp2 < 4; ) {
  301. k = *buf1++;
  302. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2];
  303. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2];
  304. switch(correction_type_sp[0][k]) {
  305. case 0:
  306. *cur_lp = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  307. lp2++;
  308. break;
  309. case 1:
  310. res = ((le2me_16(((unsigned short *)(ref_lp))[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  311. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  312. res = ((le2me_16(((unsigned short *)(ref_lp))[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  313. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  314. buf1++;
  315. lp2++;
  316. break;
  317. case 2:
  318. if(lp2 == 0) {
  319. for(i = 0, j = 0; i < 2; i++, j += width_tbl[1])
  320. cur_lp[j] = ref_lp[j];
  321. lp2 += 2;
  322. }
  323. break;
  324. case 3:
  325. if(lp2 < 2) {
  326. for(i = 0, j = 0; i < (3 - lp2); i++, j += width_tbl[1])
  327. cur_lp[j] = ref_lp[j];
  328. lp2 = 3;
  329. }
  330. break;
  331. case 8:
  332. if(lp2 == 0) {
  333. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  334. if(rle_v1 == 1 || ref_vectors != NULL) {
  335. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  336. cur_lp[j] = ref_lp[j];
  337. }
  338. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  339. break;
  340. } else {
  341. rle_v1 = 1;
  342. rle_v2 = *buf1 - 1;
  343. }
  344. case 5:
  345. LP2_CHECK(buf1,rle_v3,lp2)
  346. case 4:
  347. for(i = 0, j = 0; i < (4 - lp2); i++, j += width_tbl[1])
  348. cur_lp[j] = ref_lp[j];
  349. lp2 = 4;
  350. break;
  351. case 7:
  352. if(rle_v3 != 0)
  353. rle_v3 = 0;
  354. else {
  355. buf1--;
  356. rle_v3 = 1;
  357. }
  358. case 6:
  359. if(ref_vectors != NULL) {
  360. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  361. cur_lp[j] = ref_lp[j];
  362. }
  363. lp2 = 4;
  364. break;
  365. case 9:
  366. lv1 = *buf1++;
  367. lv = (lv1 & 0x7F) << 1;
  368. lv += (lv << 8);
  369. lv += (lv << 16);
  370. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  371. cur_lp[j] = lv;
  372. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  373. break;
  374. default:
  375. return;
  376. }
  377. }
  378. cur_frm_pos += 4;
  379. ref_frm_pos += 4;
  380. }
  381. cur_frm_pos += ((width - blks_width) * 4);
  382. ref_frm_pos += ((width - blks_width) * 4);
  383. }
  384. break;
  385. case 4:
  386. case 3: /********** CASE 3 **********/
  387. if(ref_vectors != NULL)
  388. return;
  389. flag1 = 1;
  390. for( ; blks_height > 0; blks_height -= 8) {
  391. for(lp1 = 0; lp1 < blks_width; lp1++) {
  392. for(lp2 = 0; lp2 < 4; ) {
  393. k = *buf1++;
  394. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  395. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  396. switch(correction_type_sp[lp2 & 0x01][k]) {
  397. case 0:
  398. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  399. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  400. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  401. else
  402. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  403. lp2++;
  404. break;
  405. case 1:
  406. res = ((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  407. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  408. res = ((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  409. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  410. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  411. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  412. else
  413. cur_lp[0] = cur_lp[width_tbl[1]];
  414. buf1++;
  415. lp2++;
  416. break;
  417. case 2:
  418. if(lp2 == 0) {
  419. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  420. cur_lp[j] = *ref_lp;
  421. lp2 += 2;
  422. }
  423. break;
  424. case 3:
  425. if(lp2 < 2) {
  426. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  427. cur_lp[j] = *ref_lp;
  428. lp2 = 3;
  429. }
  430. break;
  431. case 6:
  432. lp2 = 4;
  433. break;
  434. case 7:
  435. if(rle_v3 != 0)
  436. rle_v3 = 0;
  437. else {
  438. buf1--;
  439. rle_v3 = 1;
  440. }
  441. lp2 = 4;
  442. break;
  443. case 8:
  444. if(lp2 == 0) {
  445. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  446. if(rle_v1 == 1) {
  447. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  448. cur_lp[j] = ref_lp[j];
  449. }
  450. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  451. break;
  452. } else {
  453. rle_v2 = (*buf1) - 1;
  454. rle_v1 = 1;
  455. }
  456. case 5:
  457. LP2_CHECK(buf1,rle_v3,lp2)
  458. case 4:
  459. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  460. cur_lp[j] = *ref_lp;
  461. lp2 = 4;
  462. break;
  463. case 9:
  464. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  465. lv1 = *buf1++;
  466. lv = (lv1 & 0x7F) << 1;
  467. lv += (lv << 8);
  468. lv += (lv << 16);
  469. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  470. cur_lp[j] = lv;
  471. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  472. break;
  473. default:
  474. return;
  475. }
  476. }
  477. cur_frm_pos += 4;
  478. }
  479. cur_frm_pos += (((width * 2) - blks_width) * 4);
  480. flag1 = 0;
  481. }
  482. break;
  483. case 10: /********** CASE 10 **********/
  484. if(ref_vectors == NULL) {
  485. flag1 = 1;
  486. for( ; blks_height > 0; blks_height -= 8) {
  487. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  488. for(lp2 = 0; lp2 < 4; ) {
  489. k = *buf1++;
  490. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  491. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  492. lv1 = ref_lp[0];
  493. lv2 = ref_lp[1];
  494. if(lp2 == 0 && flag1 != 0) {
  495. #ifdef WORDS_BIGENDIAN
  496. lv1 = lv1 & 0xFF00FF00;
  497. lv1 = (lv1 >> 8) | lv1;
  498. lv2 = lv2 & 0xFF00FF00;
  499. lv2 = (lv2 >> 8) | lv2;
  500. #else
  501. lv1 = lv1 & 0x00FF00FF;
  502. lv1 = (lv1 << 8) | lv1;
  503. lv2 = lv2 & 0x00FF00FF;
  504. lv2 = (lv2 << 8) | lv2;
  505. #endif
  506. }
  507. switch(correction_type_sp[lp2 & 0x01][k]) {
  508. case 0:
  509. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  510. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionhighorder_lp[lp2 & 0x01][k]) << 1);
  511. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  512. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  513. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  514. } else {
  515. cur_lp[0] = cur_lp[width_tbl[1]];
  516. cur_lp[1] = cur_lp[width_tbl[1]+1];
  517. }
  518. lp2++;
  519. break;
  520. case 1:
  521. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][*buf1]) << 1);
  522. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  523. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  524. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  525. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  526. } else {
  527. cur_lp[0] = cur_lp[width_tbl[1]];
  528. cur_lp[1] = cur_lp[width_tbl[1]+1];
  529. }
  530. buf1++;
  531. lp2++;
  532. break;
  533. case 2:
  534. if(lp2 == 0) {
  535. if(flag1 != 0) {
  536. for(i = 0, j = width_tbl[1]; i < 3; i++, j += width_tbl[1]) {
  537. cur_lp[j] = lv1;
  538. cur_lp[j+1] = lv2;
  539. }
  540. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  541. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  542. } else {
  543. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  544. cur_lp[j] = lv1;
  545. cur_lp[j+1] = lv2;
  546. }
  547. }
  548. lp2 += 2;
  549. }
  550. break;
  551. case 3:
  552. if(lp2 < 2) {
  553. if(lp2 == 0 && flag1 != 0) {
  554. for(i = 0, j = width_tbl[1]; i < 5; i++, j += width_tbl[1]) {
  555. cur_lp[j] = lv1;
  556. cur_lp[j+1] = lv2;
  557. }
  558. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  559. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  560. } else {
  561. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  562. cur_lp[j] = lv1;
  563. cur_lp[j+1] = lv2;
  564. }
  565. }
  566. lp2 = 3;
  567. }
  568. break;
  569. case 8:
  570. if(lp2 == 0) {
  571. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  572. if(rle_v1 == 1) {
  573. if(flag1 != 0) {
  574. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  575. cur_lp[j] = lv1;
  576. cur_lp[j+1] = lv2;
  577. }
  578. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  579. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  580. } else {
  581. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  582. cur_lp[j] = lv1;
  583. cur_lp[j+1] = lv2;
  584. }
  585. }
  586. }
  587. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  588. break;
  589. } else {
  590. rle_v1 = 1;
  591. rle_v2 = (*buf1) - 1;
  592. }
  593. case 5:
  594. LP2_CHECK(buf1,rle_v3,lp2)
  595. case 4:
  596. if(lp2 == 0 && flag1 != 0) {
  597. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  598. cur_lp[j] = lv1;
  599. cur_lp[j+1] = lv2;
  600. }
  601. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  602. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  603. } else {
  604. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  605. cur_lp[j] = lv1;
  606. cur_lp[j+1] = lv2;
  607. }
  608. }
  609. lp2 = 4;
  610. break;
  611. case 6:
  612. lp2 = 4;
  613. break;
  614. case 7:
  615. if(lp2 == 0) {
  616. if(rle_v3 != 0)
  617. rle_v3 = 0;
  618. else {
  619. buf1--;
  620. rle_v3 = 1;
  621. }
  622. lp2 = 4;
  623. }
  624. break;
  625. case 9:
  626. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  627. lv1 = *buf1;
  628. lv = (lv1 & 0x7F) << 1;
  629. lv += (lv << 8);
  630. lv += (lv << 16);
  631. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  632. cur_lp[j] = lv;
  633. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  634. break;
  635. default:
  636. return;
  637. }
  638. }
  639. cur_frm_pos += 8;
  640. }
  641. cur_frm_pos += (((width * 2) - blks_width) * 4);
  642. flag1 = 0;
  643. }
  644. } else {
  645. for( ; blks_height > 0; blks_height -= 8) {
  646. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  647. for(lp2 = 0; lp2 < 4; ) {
  648. k = *buf1++;
  649. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  650. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  651. switch(correction_type_sp[lp2 & 0x01][k]) {
  652. case 0:
  653. lv1 = correctionloworder_lp[lp2 & 0x01][k];
  654. lv2 = correctionhighorder_lp[lp2 & 0x01][k];
  655. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  656. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  657. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  658. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  659. lp2++;
  660. break;
  661. case 1:
  662. lv1 = correctionloworder_lp[lp2 & 0x01][*buf1++];
  663. lv2 = correctionloworder_lp[lp2 & 0x01][k];
  664. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  665. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  666. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  667. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  668. lp2++;
  669. break;
  670. case 2:
  671. if(lp2 == 0) {
  672. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  673. cur_lp[j] = ref_lp[j];
  674. cur_lp[j+1] = ref_lp[j+1];
  675. }
  676. lp2 += 2;
  677. }
  678. break;
  679. case 3:
  680. if(lp2 < 2) {
  681. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  682. cur_lp[j] = ref_lp[j];
  683. cur_lp[j+1] = ref_lp[j+1];
  684. }
  685. lp2 = 3;
  686. }
  687. break;
  688. case 8:
  689. if(lp2 == 0) {
  690. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  691. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  692. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  693. ((uint32_t *)cur_frm_pos)[j+1] = ((uint32_t *)ref_frm_pos)[j+1];
  694. }
  695. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  696. break;
  697. } else {
  698. rle_v1 = 1;
  699. rle_v2 = (*buf1) - 1;
  700. }
  701. case 5:
  702. case 7:
  703. LP2_CHECK(buf1,rle_v3,lp2)
  704. case 6:
  705. case 4:
  706. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  707. cur_lp[j] = ref_lp[j];
  708. cur_lp[j+1] = ref_lp[j+1];
  709. }
  710. lp2 = 4;
  711. break;
  712. case 9:
  713. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  714. lv1 = *buf1;
  715. lv = (lv1 & 0x7F) << 1;
  716. lv += (lv << 8);
  717. lv += (lv << 16);
  718. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  719. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)cur_frm_pos)[j+1] = lv;
  720. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  721. break;
  722. default:
  723. return;
  724. }
  725. }
  726. cur_frm_pos += 8;
  727. ref_frm_pos += 8;
  728. }
  729. cur_frm_pos += (((width * 2) - blks_width) * 4);
  730. ref_frm_pos += (((width * 2) - blks_width) * 4);
  731. }
  732. }
  733. break;
  734. case 11: /********** CASE 11 **********/
  735. if(ref_vectors == NULL)
  736. return;
  737. for( ; blks_height > 0; blks_height -= 8) {
  738. for(lp1 = 0; lp1 < blks_width; lp1++) {
  739. for(lp2 = 0; lp2 < 4; ) {
  740. k = *buf1++;
  741. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  742. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  743. switch(correction_type_sp[lp2 & 0x01][k]) {
  744. case 0:
  745. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  746. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  747. lp2++;
  748. break;
  749. case 1:
  750. lv1 = (unsigned short)(correction_lp[lp2 & 0x01][*buf1++]);
  751. lv2 = (unsigned short)(correction_lp[lp2 & 0x01][k]);
  752. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + lv1) << 1);
  753. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  754. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + lv2) << 1);
  755. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  756. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]]) >> 1) + lv1) << 1);
  757. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  758. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]+1]) >> 1) + lv2) << 1);
  759. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  760. lp2++;
  761. break;
  762. case 2:
  763. if(lp2 == 0) {
  764. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  765. cur_lp[j] = ref_lp[j];
  766. lp2 += 2;
  767. }
  768. break;
  769. case 3:
  770. if(lp2 < 2) {
  771. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  772. cur_lp[j] = ref_lp[j];
  773. lp2 = 3;
  774. }
  775. break;
  776. case 8:
  777. if(lp2 == 0) {
  778. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  779. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  780. cur_lp[j] = ref_lp[j];
  781. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  782. break;
  783. } else {
  784. rle_v1 = 1;
  785. rle_v2 = (*buf1) - 1;
  786. }
  787. case 5:
  788. case 7:
  789. LP2_CHECK(buf1,rle_v3,lp2)
  790. case 4:
  791. case 6:
  792. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  793. cur_lp[j] = ref_lp[j];
  794. lp2 = 4;
  795. break;
  796. case 9:
  797. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  798. lv1 = *buf1++;
  799. lv = (lv1 & 0x7F) << 1;
  800. lv += (lv << 8);
  801. lv += (lv << 16);
  802. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  803. cur_lp[j] = lv;
  804. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  805. break;
  806. default:
  807. return;
  808. }
  809. }
  810. cur_frm_pos += 4;
  811. ref_frm_pos += 4;
  812. }
  813. cur_frm_pos += (((width * 2) - blks_width) * 4);
  814. ref_frm_pos += (((width * 2) - blks_width) * 4);
  815. }
  816. break;
  817. default:
  818. return;
  819. }
  820. }
  821. if(strip < strip_tbl)
  822. return;
  823. for( ; strip >= strip_tbl; strip--) {
  824. if(strip->split_flag != 0) {
  825. strip->split_flag = 0;
  826. strip->usl7 = (strip-1)->usl7;
  827. if(strip->split_direction) {
  828. strip->xpos += strip->width;
  829. strip->width = (strip-1)->width - strip->width;
  830. if(region_160_width <= strip->xpos && width < strip->width + strip->xpos)
  831. strip->width = width - strip->xpos;
  832. } else {
  833. strip->ypos += strip->height;
  834. strip->height = (strip-1)->height - strip->height;
  835. }
  836. break;
  837. }
  838. }
  839. }
  840. }
  841. static av_cold int indeo3_decode_init(AVCodecContext *avctx)
  842. {
  843. Indeo3DecodeContext *s = avctx->priv_data;
  844. int ret = 0;
  845. s->avctx = avctx;
  846. s->width = avctx->width;
  847. s->height = avctx->height;
  848. avctx->pix_fmt = PIX_FMT_YUV410P;
  849. if (!(ret = build_modpred(s)))
  850. ret = iv_alloc_frames(s);
  851. if (ret)
  852. iv_free_func(s);
  853. return ret;
  854. }
  855. static unsigned long iv_decode_frame(Indeo3DecodeContext *s,
  856. const uint8_t *buf, int buf_size)
  857. {
  858. unsigned int hdr_width, hdr_height,
  859. chroma_width, chroma_height;
  860. unsigned long fflags1, fflags2, fflags3, offs1, offs2, offs3, offs;
  861. const uint8_t *hdr_pos, *buf_pos;
  862. buf_pos = buf;
  863. buf_pos += 18;
  864. fflags1 = bytestream_get_le16(&buf_pos);
  865. fflags3 = bytestream_get_le32(&buf_pos);
  866. fflags2 = *buf_pos++;
  867. buf_pos += 3;
  868. hdr_height = bytestream_get_le16(&buf_pos);
  869. hdr_width = bytestream_get_le16(&buf_pos);
  870. if(avcodec_check_dimensions(NULL, hdr_width, hdr_height))
  871. return -1;
  872. chroma_height = ((hdr_height >> 2) + 3) & 0x7ffc;
  873. chroma_width = ((hdr_width >> 2) + 3) & 0x7ffc;
  874. offs1 = bytestream_get_le32(&buf_pos);
  875. offs2 = bytestream_get_le32(&buf_pos);
  876. offs3 = bytestream_get_le32(&buf_pos);
  877. buf_pos += 4;
  878. hdr_pos = buf_pos;
  879. if(fflags3 == 0x80) return 4;
  880. if(fflags1 & 0x200) {
  881. s->cur_frame = s->iv_frame + 1;
  882. s->ref_frame = s->iv_frame;
  883. } else {
  884. s->cur_frame = s->iv_frame;
  885. s->ref_frame = s->iv_frame + 1;
  886. }
  887. buf_pos = buf + 16 + offs1;
  888. offs = bytestream_get_le32(&buf_pos);
  889. iv_Decode_Chunk(s, s->cur_frame->Ybuf, s->ref_frame->Ybuf, hdr_width,
  890. hdr_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  891. FFMIN(hdr_width, 160));
  892. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  893. {
  894. buf_pos = buf + 16 + offs2;
  895. offs = bytestream_get_le32(&buf_pos);
  896. iv_Decode_Chunk(s, s->cur_frame->Vbuf, s->ref_frame->Vbuf, chroma_width,
  897. chroma_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  898. FFMIN(chroma_width, 40));
  899. buf_pos = buf + 16 + offs3;
  900. offs = bytestream_get_le32(&buf_pos);
  901. iv_Decode_Chunk(s, s->cur_frame->Ubuf, s->ref_frame->Ubuf, chroma_width,
  902. chroma_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  903. FFMIN(chroma_width, 40));
  904. }
  905. return 8;
  906. }
  907. static int indeo3_decode_frame(AVCodecContext *avctx,
  908. void *data, int *data_size,
  909. const uint8_t *buf, int buf_size)
  910. {
  911. Indeo3DecodeContext *s=avctx->priv_data;
  912. uint8_t *src, *dest;
  913. int y;
  914. iv_decode_frame(s, buf, buf_size);
  915. if(s->frame.data[0])
  916. avctx->release_buffer(avctx, &s->frame);
  917. s->frame.reference = 0;
  918. if(avctx->get_buffer(avctx, &s->frame) < 0) {
  919. av_log(s->avctx, AV_LOG_ERROR, "get_buffer() failed\n");
  920. return -1;
  921. }
  922. src = s->cur_frame->Ybuf;
  923. dest = s->frame.data[0];
  924. for (y = 0; y < s->height; y++) {
  925. memcpy(dest, src, s->cur_frame->y_w);
  926. src += s->cur_frame->y_w;
  927. dest += s->frame.linesize[0];
  928. }
  929. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  930. {
  931. src = s->cur_frame->Ubuf;
  932. dest = s->frame.data[1];
  933. for (y = 0; y < s->height / 4; y++) {
  934. memcpy(dest, src, s->cur_frame->uv_w);
  935. src += s->cur_frame->uv_w;
  936. dest += s->frame.linesize[1];
  937. }
  938. src = s->cur_frame->Vbuf;
  939. dest = s->frame.data[2];
  940. for (y = 0; y < s->height / 4; y++) {
  941. memcpy(dest, src, s->cur_frame->uv_w);
  942. src += s->cur_frame->uv_w;
  943. dest += s->frame.linesize[2];
  944. }
  945. }
  946. *data_size=sizeof(AVFrame);
  947. *(AVFrame*)data= s->frame;
  948. return buf_size;
  949. }
  950. static av_cold int indeo3_decode_end(AVCodecContext *avctx)
  951. {
  952. Indeo3DecodeContext *s = avctx->priv_data;
  953. iv_free_func(s);
  954. return 0;
  955. }
  956. AVCodec indeo3_decoder = {
  957. "indeo3",
  958. CODEC_TYPE_VIDEO,
  959. CODEC_ID_INDEO3,
  960. sizeof(Indeo3DecodeContext),
  961. indeo3_decode_init,
  962. NULL,
  963. indeo3_decode_end,
  964. indeo3_decode_frame,
  965. 0,
  966. NULL,
  967. .long_name = NULL_IF_CONFIG_SMALL("Intel Indeo 3"),
  968. };