You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1129 lines
38KB

  1. /*
  2. * Intel Indeo 3 (IV31, IV32, etc.) video decoder for ffmpeg
  3. * written, produced, and directed by Alan Smithee
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include <stdio.h>
  22. #include <stdlib.h>
  23. #include <string.h>
  24. #include <unistd.h>
  25. #include "avcodec.h"
  26. #include "dsputil.h"
  27. #include "bytestream.h"
  28. #include "indeo3data.h"
  29. typedef struct
  30. {
  31. uint8_t *Ybuf;
  32. uint8_t *Ubuf;
  33. uint8_t *Vbuf;
  34. unsigned int the_buf_size;
  35. unsigned short y_w, y_h;
  36. unsigned short uv_w, uv_h;
  37. } YUVBufs;
  38. typedef struct Indeo3DecodeContext {
  39. AVCodecContext *avctx;
  40. int width, height;
  41. AVFrame frame;
  42. uint8_t *buf;
  43. YUVBufs iv_frame[2];
  44. YUVBufs *cur_frame;
  45. YUVBufs *ref_frame;
  46. uint8_t *ModPred;
  47. uint8_t *corrector_type;
  48. } Indeo3DecodeContext;
  49. static const uint8_t corrector_type_0[24] = {
  50. 195, 159, 133, 115, 101, 93, 87, 77,
  51. 195, 159, 133, 115, 101, 93, 87, 77,
  52. 128, 79, 79, 79, 79, 79, 79, 79
  53. };
  54. static const uint8_t corrector_type_2[8] = { 9, 7, 6, 8, 5, 4, 3, 2 };
  55. static av_cold int build_modpred(Indeo3DecodeContext *s)
  56. {
  57. int i, j;
  58. if (!(s->ModPred = av_malloc(8 * 128)))
  59. return AVERROR(ENOMEM);
  60. for (i=0; i < 128; ++i) {
  61. s->ModPred[i+0*128] = i > 126 ? 254 : 2*(i + 1 - ((i + 1) % 2));
  62. s->ModPred[i+1*128] = i == 7 ? 20 :
  63. i == 119 ||
  64. i == 120 ? 236 : 2*(i + 2 - ((i + 1) % 3));
  65. s->ModPred[i+2*128] = i > 125 ? 248 : 2*(i + 2 - ((i + 2) % 4));
  66. s->ModPred[i+3*128] = 2*(i + 1 - ((i - 3) % 5));
  67. s->ModPred[i+4*128] = i == 8 ? 20 : 2*(i + 1 - ((i - 3) % 6));
  68. s->ModPred[i+5*128] = 2*(i + 4 - ((i + 3) % 7));
  69. s->ModPred[i+6*128] = i > 123 ? 240 : 2*(i + 4 - ((i + 4) % 8));
  70. s->ModPred[i+7*128] = 2*(i + 5 - ((i + 4) % 9));
  71. }
  72. if (!(s->corrector_type = av_malloc(24 * 256)))
  73. return AVERROR(ENOMEM);
  74. for (i=0; i < 24; ++i) {
  75. for (j=0; j < 256; ++j) {
  76. s->corrector_type[i*256+j] = j < corrector_type_0[i] ? 1 :
  77. j < 248 || (i == 16 && j == 248) ? 0 :
  78. corrector_type_2[j - 248];
  79. }
  80. }
  81. return 0;
  82. }
  83. static void iv_Decode_Chunk(Indeo3DecodeContext *s, uint8_t *cur,
  84. uint8_t *ref, int width, int height, const uint8_t *buf1,
  85. long fflags2, const uint8_t *hdr,
  86. const uint8_t *buf2, int min_width_160);
  87. /* ---------------------------------------------------------------------- */
  88. static av_cold int iv_alloc_frames(Indeo3DecodeContext *s)
  89. {
  90. int luma_width, luma_height, luma_pixels, chroma_width, chroma_height,
  91. chroma_pixels, i;
  92. unsigned int bufsize;
  93. luma_width = (s->width + 3) & (~3);
  94. luma_height = (s->height + 3) & (~3);
  95. chroma_width = ((luma_width >> 2) + 3) & (~3);
  96. chroma_height = ((luma_height>> 2) + 3) & (~3);
  97. luma_pixels = luma_width * luma_height;
  98. chroma_pixels = chroma_width * chroma_height;
  99. bufsize = luma_pixels * 2 + luma_width * 3 +
  100. (chroma_pixels + chroma_width) * 4;
  101. if(!(s->buf = av_malloc(bufsize)))
  102. return AVERROR(ENOMEM);
  103. s->iv_frame[0].y_w = s->iv_frame[1].y_w = luma_width;
  104. s->iv_frame[0].y_h = s->iv_frame[1].y_h = luma_height;
  105. s->iv_frame[0].uv_w = s->iv_frame[1].uv_w = chroma_width;
  106. s->iv_frame[0].uv_h = s->iv_frame[1].uv_h = chroma_height;
  107. s->iv_frame[0].the_buf_size = bufsize;
  108. s->iv_frame[0].Ybuf = s->buf + luma_width;
  109. i = luma_pixels + luma_width * 2;
  110. s->iv_frame[1].Ybuf = s->buf + i;
  111. i += (luma_pixels + luma_width);
  112. s->iv_frame[0].Ubuf = s->buf + i;
  113. i += (chroma_pixels + chroma_width);
  114. s->iv_frame[1].Ubuf = s->buf + i;
  115. i += (chroma_pixels + chroma_width);
  116. s->iv_frame[0].Vbuf = s->buf + i;
  117. i += (chroma_pixels + chroma_width);
  118. s->iv_frame[1].Vbuf = s->buf + i;
  119. for(i = 1; i <= luma_width; i++)
  120. s->iv_frame[0].Ybuf[-i] = s->iv_frame[1].Ybuf[-i] =
  121. s->iv_frame[0].Ubuf[-i] = 0x80;
  122. for(i = 1; i <= chroma_width; i++) {
  123. s->iv_frame[1].Ubuf[-i] = 0x80;
  124. s->iv_frame[0].Vbuf[-i] = 0x80;
  125. s->iv_frame[1].Vbuf[-i] = 0x80;
  126. s->iv_frame[1].Vbuf[chroma_pixels+i-1] = 0x80;
  127. }
  128. return 0;
  129. }
  130. /* ---------------------------------------------------------------------- */
  131. static av_cold void iv_free_func(Indeo3DecodeContext *s)
  132. {
  133. av_free(s->buf);
  134. av_free(s->ModPred);
  135. av_free(s->corrector_type);
  136. }
  137. /* ---------------------------------------------------------------------- */
  138. static unsigned long iv_decode_frame(Indeo3DecodeContext *s,
  139. const uint8_t *buf, int buf_size)
  140. {
  141. unsigned int hdr_width, hdr_height,
  142. chroma_width, chroma_height;
  143. unsigned long fflags1, fflags2, fflags3, offs1, offs2, offs3, offs;
  144. const uint8_t *hdr_pos, *buf_pos;
  145. buf_pos = buf;
  146. buf_pos += 18;
  147. fflags1 = bytestream_get_le16(&buf_pos);
  148. fflags3 = bytestream_get_le32(&buf_pos);
  149. fflags2 = *buf_pos++;
  150. buf_pos += 3;
  151. hdr_height = bytestream_get_le16(&buf_pos);
  152. hdr_width = bytestream_get_le16(&buf_pos);
  153. if(avcodec_check_dimensions(NULL, hdr_width, hdr_height))
  154. return -1;
  155. chroma_height = ((hdr_height >> 2) + 3) & 0x7ffc;
  156. chroma_width = ((hdr_width >> 2) + 3) & 0x7ffc;
  157. offs1 = bytestream_get_le32(&buf_pos);
  158. offs2 = bytestream_get_le32(&buf_pos);
  159. offs3 = bytestream_get_le32(&buf_pos);
  160. buf_pos += 4;
  161. hdr_pos = buf_pos;
  162. if(fflags3 == 0x80) return 4;
  163. if(fflags1 & 0x200) {
  164. s->cur_frame = s->iv_frame + 1;
  165. s->ref_frame = s->iv_frame;
  166. } else {
  167. s->cur_frame = s->iv_frame;
  168. s->ref_frame = s->iv_frame + 1;
  169. }
  170. buf_pos = buf + 16 + offs1;
  171. offs = bytestream_get_le32(&buf_pos);
  172. iv_Decode_Chunk(s, s->cur_frame->Ybuf, s->ref_frame->Ybuf, hdr_width,
  173. hdr_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  174. FFMIN(hdr_width, 160));
  175. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  176. {
  177. buf_pos = buf + 16 + offs2;
  178. offs = bytestream_get_le32(&buf_pos);
  179. iv_Decode_Chunk(s, s->cur_frame->Vbuf, s->ref_frame->Vbuf, chroma_width,
  180. chroma_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  181. FFMIN(chroma_width, 40));
  182. buf_pos = buf + 16 + offs3;
  183. offs = bytestream_get_le32(&buf_pos);
  184. iv_Decode_Chunk(s, s->cur_frame->Ubuf, s->ref_frame->Ubuf, chroma_width,
  185. chroma_height, buf_pos + offs * 2, fflags2, hdr_pos, buf_pos,
  186. FFMIN(chroma_width, 40));
  187. }
  188. return 8;
  189. }
  190. typedef struct {
  191. long xpos;
  192. long ypos;
  193. long width;
  194. long height;
  195. long split_flag;
  196. long split_direction;
  197. long usl7;
  198. } ustr_t;
  199. /* ---------------------------------------------------------------------- */
  200. #define LV1_CHECK(buf1,rle_v3,lv1,lp2) \
  201. if((lv1 & 0x80) != 0) { \
  202. if(rle_v3 != 0) \
  203. rle_v3 = 0; \
  204. else { \
  205. rle_v3 = 1; \
  206. buf1 -= 2; \
  207. } \
  208. } \
  209. lp2 = 4;
  210. #define RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3) \
  211. if(rle_v3 == 0) { \
  212. rle_v2 = *buf1; \
  213. rle_v1 = 1; \
  214. if(rle_v2 > 32) { \
  215. rle_v2 -= 32; \
  216. rle_v1 = 0; \
  217. } \
  218. rle_v3 = 1; \
  219. } \
  220. buf1--;
  221. #define LP2_CHECK(buf1,rle_v3,lp2) \
  222. if(lp2 == 0 && rle_v3 != 0) \
  223. rle_v3 = 0; \
  224. else { \
  225. buf1--; \
  226. rle_v3 = 1; \
  227. }
  228. #define RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2) \
  229. rle_v2--; \
  230. if(rle_v2 == 0) { \
  231. rle_v3 = 0; \
  232. buf1 += 2; \
  233. } \
  234. lp2 = 4;
  235. static void iv_Decode_Chunk(Indeo3DecodeContext *s,
  236. uint8_t *cur, uint8_t *ref, int width, int height,
  237. const uint8_t *buf1, long fflags2, const uint8_t *hdr,
  238. const uint8_t *buf2, int min_width_160)
  239. {
  240. uint8_t bit_buf;
  241. unsigned long bit_pos, lv, lv1, lv2;
  242. long *width_tbl, width_tbl_arr[10];
  243. const signed char *ref_vectors;
  244. uint8_t *cur_frm_pos, *ref_frm_pos, *cp, *cp2;
  245. uint32_t *cur_lp, *ref_lp;
  246. const uint32_t *correction_lp[2], *correctionloworder_lp[2], *correctionhighorder_lp[2];
  247. uint8_t *correction_type_sp[2];
  248. ustr_t strip_tbl[20], *strip;
  249. int i, j, k, lp1, lp2, flag1, cmd, blks_width, blks_height, region_160_width,
  250. rle_v1, rle_v2, rle_v3;
  251. unsigned short res;
  252. bit_buf = 0;
  253. ref_vectors = NULL;
  254. width_tbl = width_tbl_arr + 1;
  255. i = (width < 0 ? width + 3 : width)/4;
  256. for(j = -1; j < 8; j++)
  257. width_tbl[j] = i * j;
  258. strip = strip_tbl;
  259. for(region_160_width = 0; region_160_width < (width - min_width_160); region_160_width += min_width_160);
  260. strip->ypos = strip->xpos = 0;
  261. for(strip->width = min_width_160; width > strip->width; strip->width *= 2);
  262. strip->height = height;
  263. strip->split_direction = 0;
  264. strip->split_flag = 0;
  265. strip->usl7 = 0;
  266. bit_pos = 0;
  267. rle_v1 = rle_v2 = rle_v3 = 0;
  268. while(strip >= strip_tbl) {
  269. if(bit_pos <= 0) {
  270. bit_pos = 8;
  271. bit_buf = *buf1++;
  272. }
  273. bit_pos -= 2;
  274. cmd = (bit_buf >> bit_pos) & 0x03;
  275. if(cmd == 0) {
  276. strip++;
  277. memcpy(strip, strip-1, sizeof(ustr_t));
  278. strip->split_flag = 1;
  279. strip->split_direction = 0;
  280. strip->height = (strip->height > 8 ? ((strip->height+8)>>4)<<3 : 4);
  281. continue;
  282. } else if(cmd == 1) {
  283. strip++;
  284. memcpy(strip, strip-1, sizeof(ustr_t));
  285. strip->split_flag = 1;
  286. strip->split_direction = 1;
  287. strip->width = (strip->width > 8 ? ((strip->width+8)>>4)<<3 : 4);
  288. continue;
  289. } else if(cmd == 2) {
  290. if(strip->usl7 == 0) {
  291. strip->usl7 = 1;
  292. ref_vectors = NULL;
  293. continue;
  294. }
  295. } else if(cmd == 3) {
  296. if(strip->usl7 == 0) {
  297. strip->usl7 = 1;
  298. ref_vectors = (const signed char*)buf2 + (*buf1 * 2);
  299. buf1++;
  300. continue;
  301. }
  302. }
  303. cur_frm_pos = cur + width * strip->ypos + strip->xpos;
  304. if((blks_width = strip->width) < 0)
  305. blks_width += 3;
  306. blks_width >>= 2;
  307. blks_height = strip->height;
  308. if(ref_vectors != NULL) {
  309. ref_frm_pos = ref + (ref_vectors[0] + strip->ypos) * width +
  310. ref_vectors[1] + strip->xpos;
  311. } else
  312. ref_frm_pos = cur_frm_pos - width_tbl[4];
  313. if(cmd == 2) {
  314. if(bit_pos <= 0) {
  315. bit_pos = 8;
  316. bit_buf = *buf1++;
  317. }
  318. bit_pos -= 2;
  319. cmd = (bit_buf >> bit_pos) & 0x03;
  320. if(cmd == 0 || ref_vectors != NULL) {
  321. for(lp1 = 0; lp1 < blks_width; lp1++) {
  322. for(i = 0, j = 0; i < blks_height; i++, j += width_tbl[1])
  323. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  324. cur_frm_pos += 4;
  325. ref_frm_pos += 4;
  326. }
  327. } else if(cmd != 1)
  328. return;
  329. } else {
  330. k = *buf1 >> 4;
  331. j = *buf1 & 0x0f;
  332. buf1++;
  333. lv = j + fflags2;
  334. if((lv - 8) <= 7 && (k == 0 || k == 3 || k == 10)) {
  335. cp2 = s->ModPred + ((lv - 8) << 7);
  336. cp = ref_frm_pos;
  337. for(i = 0; i < blks_width << 2; i++) {
  338. int v = *cp >> 1;
  339. *(cp++) = cp2[v];
  340. }
  341. }
  342. if(k == 1 || k == 4) {
  343. lv = (hdr[j] & 0xf) + fflags2;
  344. correction_type_sp[0] = s->corrector_type + (lv << 8);
  345. correction_lp[0] = correction + (lv << 8);
  346. lv = (hdr[j] >> 4) + fflags2;
  347. correction_lp[1] = correction + (lv << 8);
  348. correction_type_sp[1] = s->corrector_type + (lv << 8);
  349. } else {
  350. correctionloworder_lp[0] = correctionloworder_lp[1] = correctionloworder + (lv << 8);
  351. correctionhighorder_lp[0] = correctionhighorder_lp[1] = correctionhighorder + (lv << 8);
  352. correction_type_sp[0] = correction_type_sp[1] = s->corrector_type + (lv << 8);
  353. correction_lp[0] = correction_lp[1] = correction + (lv << 8);
  354. }
  355. switch(k) {
  356. case 1:
  357. case 0: /********** CASE 0 **********/
  358. for( ; blks_height > 0; blks_height -= 4) {
  359. for(lp1 = 0; lp1 < blks_width; lp1++) {
  360. for(lp2 = 0; lp2 < 4; ) {
  361. k = *buf1++;
  362. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2];
  363. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2];
  364. switch(correction_type_sp[0][k]) {
  365. case 0:
  366. *cur_lp = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  367. lp2++;
  368. break;
  369. case 1:
  370. res = ((le2me_16(((unsigned short *)(ref_lp))[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  371. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  372. res = ((le2me_16(((unsigned short *)(ref_lp))[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  373. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  374. buf1++;
  375. lp2++;
  376. break;
  377. case 2:
  378. if(lp2 == 0) {
  379. for(i = 0, j = 0; i < 2; i++, j += width_tbl[1])
  380. cur_lp[j] = ref_lp[j];
  381. lp2 += 2;
  382. }
  383. break;
  384. case 3:
  385. if(lp2 < 2) {
  386. for(i = 0, j = 0; i < (3 - lp2); i++, j += width_tbl[1])
  387. cur_lp[j] = ref_lp[j];
  388. lp2 = 3;
  389. }
  390. break;
  391. case 8:
  392. if(lp2 == 0) {
  393. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  394. if(rle_v1 == 1 || ref_vectors != NULL) {
  395. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  396. cur_lp[j] = ref_lp[j];
  397. }
  398. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  399. break;
  400. } else {
  401. rle_v1 = 1;
  402. rle_v2 = *buf1 - 1;
  403. }
  404. case 5:
  405. LP2_CHECK(buf1,rle_v3,lp2)
  406. case 4:
  407. for(i = 0, j = 0; i < (4 - lp2); i++, j += width_tbl[1])
  408. cur_lp[j] = ref_lp[j];
  409. lp2 = 4;
  410. break;
  411. case 7:
  412. if(rle_v3 != 0)
  413. rle_v3 = 0;
  414. else {
  415. buf1--;
  416. rle_v3 = 1;
  417. }
  418. case 6:
  419. if(ref_vectors != NULL) {
  420. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  421. cur_lp[j] = ref_lp[j];
  422. }
  423. lp2 = 4;
  424. break;
  425. case 9:
  426. lv1 = *buf1++;
  427. lv = (lv1 & 0x7F) << 1;
  428. lv += (lv << 8);
  429. lv += (lv << 16);
  430. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  431. cur_lp[j] = lv;
  432. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  433. break;
  434. default:
  435. return;
  436. }
  437. }
  438. cur_frm_pos += 4;
  439. ref_frm_pos += 4;
  440. }
  441. cur_frm_pos += ((width - blks_width) * 4);
  442. ref_frm_pos += ((width - blks_width) * 4);
  443. }
  444. break;
  445. case 4:
  446. case 3: /********** CASE 3 **********/
  447. if(ref_vectors != NULL)
  448. return;
  449. flag1 = 1;
  450. for( ; blks_height > 0; blks_height -= 8) {
  451. for(lp1 = 0; lp1 < blks_width; lp1++) {
  452. for(lp2 = 0; lp2 < 4; ) {
  453. k = *buf1++;
  454. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  455. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  456. switch(correction_type_sp[lp2 & 0x01][k]) {
  457. case 0:
  458. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  459. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  460. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  461. else
  462. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  463. lp2++;
  464. break;
  465. case 1:
  466. res = ((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + correction_lp[lp2 & 0x01][*buf1]) << 1;
  467. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  468. res = ((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1;
  469. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  470. if(lp2 > 0 || flag1 == 0 || strip->ypos != 0)
  471. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  472. else
  473. cur_lp[0] = cur_lp[width_tbl[1]];
  474. buf1++;
  475. lp2++;
  476. break;
  477. case 2:
  478. if(lp2 == 0) {
  479. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  480. cur_lp[j] = *ref_lp;
  481. lp2 += 2;
  482. }
  483. break;
  484. case 3:
  485. if(lp2 < 2) {
  486. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  487. cur_lp[j] = *ref_lp;
  488. lp2 = 3;
  489. }
  490. break;
  491. case 6:
  492. lp2 = 4;
  493. break;
  494. case 7:
  495. if(rle_v3 != 0)
  496. rle_v3 = 0;
  497. else {
  498. buf1--;
  499. rle_v3 = 1;
  500. }
  501. lp2 = 4;
  502. break;
  503. case 8:
  504. if(lp2 == 0) {
  505. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  506. if(rle_v1 == 1) {
  507. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  508. cur_lp[j] = ref_lp[j];
  509. }
  510. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  511. break;
  512. } else {
  513. rle_v2 = (*buf1) - 1;
  514. rle_v1 = 1;
  515. }
  516. case 5:
  517. LP2_CHECK(buf1,rle_v3,lp2)
  518. case 4:
  519. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  520. cur_lp[j] = *ref_lp;
  521. lp2 = 4;
  522. break;
  523. case 9:
  524. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  525. lv1 = *buf1++;
  526. lv = (lv1 & 0x7F) << 1;
  527. lv += (lv << 8);
  528. lv += (lv << 16);
  529. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  530. cur_lp[j] = lv;
  531. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  532. break;
  533. default:
  534. return;
  535. }
  536. }
  537. cur_frm_pos += 4;
  538. }
  539. cur_frm_pos += (((width * 2) - blks_width) * 4);
  540. flag1 = 0;
  541. }
  542. break;
  543. case 10: /********** CASE 10 **********/
  544. if(ref_vectors == NULL) {
  545. flag1 = 1;
  546. for( ; blks_height > 0; blks_height -= 8) {
  547. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  548. for(lp2 = 0; lp2 < 4; ) {
  549. k = *buf1++;
  550. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  551. ref_lp = ((uint32_t *)cur_frm_pos) + width_tbl[(lp2 * 2) - 1];
  552. lv1 = ref_lp[0];
  553. lv2 = ref_lp[1];
  554. if(lp2 == 0 && flag1 != 0) {
  555. #ifdef WORDS_BIGENDIAN
  556. lv1 = lv1 & 0xFF00FF00;
  557. lv1 = (lv1 >> 8) | lv1;
  558. lv2 = lv2 & 0xFF00FF00;
  559. lv2 = (lv2 >> 8) | lv2;
  560. #else
  561. lv1 = lv1 & 0x00FF00FF;
  562. lv1 = (lv1 << 8) | lv1;
  563. lv2 = lv2 & 0x00FF00FF;
  564. lv2 = (lv2 << 8) | lv2;
  565. #endif
  566. }
  567. switch(correction_type_sp[lp2 & 0x01][k]) {
  568. case 0:
  569. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  570. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionhighorder_lp[lp2 & 0x01][k]) << 1);
  571. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  572. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  573. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  574. } else {
  575. cur_lp[0] = cur_lp[width_tbl[1]];
  576. cur_lp[1] = cur_lp[width_tbl[1]+1];
  577. }
  578. lp2++;
  579. break;
  580. case 1:
  581. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(lv1) >> 1) + correctionloworder_lp[lp2 & 0x01][*buf1]) << 1);
  582. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(lv2) >> 1) + correctionloworder_lp[lp2 & 0x01][k]) << 1);
  583. if(lp2 > 0 || strip->ypos != 0 || flag1 == 0) {
  584. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  585. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  586. } else {
  587. cur_lp[0] = cur_lp[width_tbl[1]];
  588. cur_lp[1] = cur_lp[width_tbl[1]+1];
  589. }
  590. buf1++;
  591. lp2++;
  592. break;
  593. case 2:
  594. if(lp2 == 0) {
  595. if(flag1 != 0) {
  596. for(i = 0, j = width_tbl[1]; i < 3; i++, j += width_tbl[1]) {
  597. cur_lp[j] = lv1;
  598. cur_lp[j+1] = lv2;
  599. }
  600. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  601. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  602. } else {
  603. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  604. cur_lp[j] = lv1;
  605. cur_lp[j+1] = lv2;
  606. }
  607. }
  608. lp2 += 2;
  609. }
  610. break;
  611. case 3:
  612. if(lp2 < 2) {
  613. if(lp2 == 0 && flag1 != 0) {
  614. for(i = 0, j = width_tbl[1]; i < 5; i++, j += width_tbl[1]) {
  615. cur_lp[j] = lv1;
  616. cur_lp[j+1] = lv2;
  617. }
  618. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  619. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  620. } else {
  621. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  622. cur_lp[j] = lv1;
  623. cur_lp[j+1] = lv2;
  624. }
  625. }
  626. lp2 = 3;
  627. }
  628. break;
  629. case 8:
  630. if(lp2 == 0) {
  631. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  632. if(rle_v1 == 1) {
  633. if(flag1 != 0) {
  634. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  635. cur_lp[j] = lv1;
  636. cur_lp[j+1] = lv2;
  637. }
  638. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  639. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  640. } else {
  641. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  642. cur_lp[j] = lv1;
  643. cur_lp[j+1] = lv2;
  644. }
  645. }
  646. }
  647. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  648. break;
  649. } else {
  650. rle_v1 = 1;
  651. rle_v2 = (*buf1) - 1;
  652. }
  653. case 5:
  654. LP2_CHECK(buf1,rle_v3,lp2)
  655. case 4:
  656. if(lp2 == 0 && flag1 != 0) {
  657. for(i = 0, j = width_tbl[1]; i < 7; i++, j += width_tbl[1]) {
  658. cur_lp[j] = lv1;
  659. cur_lp[j+1] = lv2;
  660. }
  661. cur_lp[0] = ((cur_lp[-width_tbl[1]] >> 1) + (cur_lp[width_tbl[1]] >> 1)) & 0xFEFEFEFE;
  662. cur_lp[1] = ((cur_lp[-width_tbl[1]+1] >> 1) + (cur_lp[width_tbl[1]+1] >> 1)) & 0xFEFEFEFE;
  663. } else {
  664. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  665. cur_lp[j] = lv1;
  666. cur_lp[j+1] = lv2;
  667. }
  668. }
  669. lp2 = 4;
  670. break;
  671. case 6:
  672. lp2 = 4;
  673. break;
  674. case 7:
  675. if(lp2 == 0) {
  676. if(rle_v3 != 0)
  677. rle_v3 = 0;
  678. else {
  679. buf1--;
  680. rle_v3 = 1;
  681. }
  682. lp2 = 4;
  683. }
  684. break;
  685. case 9:
  686. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  687. lv1 = *buf1;
  688. lv = (lv1 & 0x7F) << 1;
  689. lv += (lv << 8);
  690. lv += (lv << 16);
  691. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  692. cur_lp[j] = lv;
  693. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  694. break;
  695. default:
  696. return;
  697. }
  698. }
  699. cur_frm_pos += 8;
  700. }
  701. cur_frm_pos += (((width * 2) - blks_width) * 4);
  702. flag1 = 0;
  703. }
  704. } else {
  705. for( ; blks_height > 0; blks_height -= 8) {
  706. for(lp1 = 0; lp1 < blks_width; lp1 += 2) {
  707. for(lp2 = 0; lp2 < 4; ) {
  708. k = *buf1++;
  709. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  710. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  711. switch(correction_type_sp[lp2 & 0x01][k]) {
  712. case 0:
  713. lv1 = correctionloworder_lp[lp2 & 0x01][k];
  714. lv2 = correctionhighorder_lp[lp2 & 0x01][k];
  715. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  716. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  717. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  718. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  719. lp2++;
  720. break;
  721. case 1:
  722. lv1 = correctionloworder_lp[lp2 & 0x01][*buf1++];
  723. lv2 = correctionloworder_lp[lp2 & 0x01][k];
  724. cur_lp[0] = le2me_32(((le2me_32(ref_lp[0]) >> 1) + lv1) << 1);
  725. cur_lp[1] = le2me_32(((le2me_32(ref_lp[1]) >> 1) + lv2) << 1);
  726. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + lv1) << 1);
  727. cur_lp[width_tbl[1]+1] = le2me_32(((le2me_32(ref_lp[width_tbl[1]+1]) >> 1) + lv2) << 1);
  728. lp2++;
  729. break;
  730. case 2:
  731. if(lp2 == 0) {
  732. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1]) {
  733. cur_lp[j] = ref_lp[j];
  734. cur_lp[j+1] = ref_lp[j+1];
  735. }
  736. lp2 += 2;
  737. }
  738. break;
  739. case 3:
  740. if(lp2 < 2) {
  741. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1]) {
  742. cur_lp[j] = ref_lp[j];
  743. cur_lp[j+1] = ref_lp[j+1];
  744. }
  745. lp2 = 3;
  746. }
  747. break;
  748. case 8:
  749. if(lp2 == 0) {
  750. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  751. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1]) {
  752. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)ref_frm_pos)[j];
  753. ((uint32_t *)cur_frm_pos)[j+1] = ((uint32_t *)ref_frm_pos)[j+1];
  754. }
  755. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  756. break;
  757. } else {
  758. rle_v1 = 1;
  759. rle_v2 = (*buf1) - 1;
  760. }
  761. case 5:
  762. case 7:
  763. LP2_CHECK(buf1,rle_v3,lp2)
  764. case 6:
  765. case 4:
  766. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1]) {
  767. cur_lp[j] = ref_lp[j];
  768. cur_lp[j+1] = ref_lp[j+1];
  769. }
  770. lp2 = 4;
  771. break;
  772. case 9:
  773. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  774. lv1 = *buf1;
  775. lv = (lv1 & 0x7F) << 1;
  776. lv += (lv << 8);
  777. lv += (lv << 16);
  778. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  779. ((uint32_t *)cur_frm_pos)[j] = ((uint32_t *)cur_frm_pos)[j+1] = lv;
  780. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  781. break;
  782. default:
  783. return;
  784. }
  785. }
  786. cur_frm_pos += 8;
  787. ref_frm_pos += 8;
  788. }
  789. cur_frm_pos += (((width * 2) - blks_width) * 4);
  790. ref_frm_pos += (((width * 2) - blks_width) * 4);
  791. }
  792. }
  793. break;
  794. case 11: /********** CASE 11 **********/
  795. if(ref_vectors == NULL)
  796. return;
  797. for( ; blks_height > 0; blks_height -= 8) {
  798. for(lp1 = 0; lp1 < blks_width; lp1++) {
  799. for(lp2 = 0; lp2 < 4; ) {
  800. k = *buf1++;
  801. cur_lp = ((uint32_t *)cur_frm_pos) + width_tbl[lp2 * 2];
  802. ref_lp = ((uint32_t *)ref_frm_pos) + width_tbl[lp2 * 2];
  803. switch(correction_type_sp[lp2 & 0x01][k]) {
  804. case 0:
  805. cur_lp[0] = le2me_32(((le2me_32(*ref_lp) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  806. cur_lp[width_tbl[1]] = le2me_32(((le2me_32(ref_lp[width_tbl[1]]) >> 1) + correction_lp[lp2 & 0x01][k]) << 1);
  807. lp2++;
  808. break;
  809. case 1:
  810. lv1 = (unsigned short)(correction_lp[lp2 & 0x01][*buf1++]);
  811. lv2 = (unsigned short)(correction_lp[lp2 & 0x01][k]);
  812. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[0]) >> 1) + lv1) << 1);
  813. ((unsigned short *)cur_lp)[0] = le2me_16(res);
  814. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[1]) >> 1) + lv2) << 1);
  815. ((unsigned short *)cur_lp)[1] = le2me_16(res);
  816. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]]) >> 1) + lv1) << 1);
  817. ((unsigned short *)cur_lp)[width_tbl[2]] = le2me_16(res);
  818. res = (unsigned short)(((le2me_16(((unsigned short *)ref_lp)[width_tbl[2]+1]) >> 1) + lv2) << 1);
  819. ((unsigned short *)cur_lp)[width_tbl[2]+1] = le2me_16(res);
  820. lp2++;
  821. break;
  822. case 2:
  823. if(lp2 == 0) {
  824. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  825. cur_lp[j] = ref_lp[j];
  826. lp2 += 2;
  827. }
  828. break;
  829. case 3:
  830. if(lp2 < 2) {
  831. for(i = 0, j = 0; i < 6 - (lp2 * 2); i++, j += width_tbl[1])
  832. cur_lp[j] = ref_lp[j];
  833. lp2 = 3;
  834. }
  835. break;
  836. case 8:
  837. if(lp2 == 0) {
  838. RLE_V3_CHECK(buf1,rle_v1,rle_v2,rle_v3)
  839. for(i = 0, j = 0; i < 8; i++, j += width_tbl[1])
  840. cur_lp[j] = ref_lp[j];
  841. RLE_V2_CHECK(buf1,rle_v2, rle_v3,lp2)
  842. break;
  843. } else {
  844. rle_v1 = 1;
  845. rle_v2 = (*buf1) - 1;
  846. }
  847. case 5:
  848. case 7:
  849. LP2_CHECK(buf1,rle_v3,lp2)
  850. case 4:
  851. case 6:
  852. for(i = 0, j = 0; i < 8 - (lp2 * 2); i++, j += width_tbl[1])
  853. cur_lp[j] = ref_lp[j];
  854. lp2 = 4;
  855. break;
  856. case 9:
  857. av_log(s->avctx, AV_LOG_ERROR, "UNTESTED.\n");
  858. lv1 = *buf1++;
  859. lv = (lv1 & 0x7F) << 1;
  860. lv += (lv << 8);
  861. lv += (lv << 16);
  862. for(i = 0, j = 0; i < 4; i++, j += width_tbl[1])
  863. cur_lp[j] = lv;
  864. LV1_CHECK(buf1,rle_v3,lv1,lp2)
  865. break;
  866. default:
  867. return;
  868. }
  869. }
  870. cur_frm_pos += 4;
  871. ref_frm_pos += 4;
  872. }
  873. cur_frm_pos += (((width * 2) - blks_width) * 4);
  874. ref_frm_pos += (((width * 2) - blks_width) * 4);
  875. }
  876. break;
  877. default:
  878. return;
  879. }
  880. }
  881. if(strip < strip_tbl)
  882. return;
  883. for( ; strip >= strip_tbl; strip--) {
  884. if(strip->split_flag != 0) {
  885. strip->split_flag = 0;
  886. strip->usl7 = (strip-1)->usl7;
  887. if(strip->split_direction) {
  888. strip->xpos += strip->width;
  889. strip->width = (strip-1)->width - strip->width;
  890. if(region_160_width <= strip->xpos && width < strip->width + strip->xpos)
  891. strip->width = width - strip->xpos;
  892. } else {
  893. strip->ypos += strip->height;
  894. strip->height = (strip-1)->height - strip->height;
  895. }
  896. break;
  897. }
  898. }
  899. }
  900. }
  901. static av_cold int indeo3_decode_init(AVCodecContext *avctx)
  902. {
  903. Indeo3DecodeContext *s = avctx->priv_data;
  904. int ret = 0;
  905. s->avctx = avctx;
  906. s->width = avctx->width;
  907. s->height = avctx->height;
  908. avctx->pix_fmt = PIX_FMT_YUV410P;
  909. if (!(ret = build_modpred(s)))
  910. ret = iv_alloc_frames(s);
  911. if (ret)
  912. iv_free_func(s);
  913. return ret;
  914. }
  915. static int indeo3_decode_frame(AVCodecContext *avctx,
  916. void *data, int *data_size,
  917. const uint8_t *buf, int buf_size)
  918. {
  919. Indeo3DecodeContext *s=avctx->priv_data;
  920. uint8_t *src, *dest;
  921. int y;
  922. iv_decode_frame(s, buf, buf_size);
  923. if(s->frame.data[0])
  924. avctx->release_buffer(avctx, &s->frame);
  925. s->frame.reference = 0;
  926. if(avctx->get_buffer(avctx, &s->frame) < 0) {
  927. av_log(s->avctx, AV_LOG_ERROR, "get_buffer() failed\n");
  928. return -1;
  929. }
  930. src = s->cur_frame->Ybuf;
  931. dest = s->frame.data[0];
  932. for (y = 0; y < s->height; y++) {
  933. memcpy(dest, src, s->cur_frame->y_w);
  934. src += s->cur_frame->y_w;
  935. dest += s->frame.linesize[0];
  936. }
  937. if (!(s->avctx->flags & CODEC_FLAG_GRAY))
  938. {
  939. src = s->cur_frame->Ubuf;
  940. dest = s->frame.data[1];
  941. for (y = 0; y < s->height / 4; y++) {
  942. memcpy(dest, src, s->cur_frame->uv_w);
  943. src += s->cur_frame->uv_w;
  944. dest += s->frame.linesize[1];
  945. }
  946. src = s->cur_frame->Vbuf;
  947. dest = s->frame.data[2];
  948. for (y = 0; y < s->height / 4; y++) {
  949. memcpy(dest, src, s->cur_frame->uv_w);
  950. src += s->cur_frame->uv_w;
  951. dest += s->frame.linesize[2];
  952. }
  953. }
  954. *data_size=sizeof(AVFrame);
  955. *(AVFrame*)data= s->frame;
  956. return buf_size;
  957. }
  958. static av_cold int indeo3_decode_end(AVCodecContext *avctx)
  959. {
  960. Indeo3DecodeContext *s = avctx->priv_data;
  961. iv_free_func(s);
  962. return 0;
  963. }
  964. AVCodec indeo3_decoder = {
  965. "indeo3",
  966. CODEC_TYPE_VIDEO,
  967. CODEC_ID_INDEO3,
  968. sizeof(Indeo3DecodeContext),
  969. indeo3_decode_init,
  970. NULL,
  971. indeo3_decode_end,
  972. indeo3_decode_frame,
  973. 0,
  974. NULL,
  975. .long_name = NULL_IF_CONFIG_SMALL("Intel Indeo 3"),
  976. };