You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

790 lines
23KB

  1. /*
  2. * AltiVec-optimized snow DSP utils
  3. * Copyright (c) 2006 Luca Barbato <lu_zero@gentoo.org>
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include "dsputil.h"
  22. #include "gcc_fixes.h"
  23. #include "dsputil_altivec.h"
  24. #include "snow.h"
  25. #undef NDEBUG
  26. #include <assert.h>
  27. //FIXME remove this replication
  28. #define slice_buffer_get_line(slice_buf, line_num) ((slice_buf)->line[line_num] ? (slice_buf)->line[line_num] : slice_buffer_load_line((slice_buf), (line_num)))
  29. static DWTELEM * slice_buffer_load_line(slice_buffer * buf, int line)
  30. {
  31. int offset;
  32. DWTELEM * buffer;
  33. // av_log(NULL, AV_LOG_DEBUG, "Cache hit: %d\n", line);
  34. assert(buf->data_stack_top >= 0);
  35. // assert(!buf->line[line]);
  36. if (buf->line[line])
  37. return buf->line[line];
  38. offset = buf->line_width * line;
  39. buffer = buf->data_stack[buf->data_stack_top];
  40. buf->data_stack_top--;
  41. buf->line[line] = buffer;
  42. // av_log(NULL, AV_LOG_DEBUG, "slice_buffer_load_line: line: %d remaining: %d\n", line, buf->data_stack_top + 1);
  43. return buffer;
  44. }
  45. //altivec code
  46. void ff_snow_horizontal_compose97i_altivec(IDWTELEM *b, int width)
  47. {
  48. #if 0
  49. const int w2= (width+1)>>1;
  50. DECLARE_ALIGNED_16(IDWTELEM, temp[(width>>1)]);
  51. const int w_l= (width>>1);
  52. const int w_r= w2 - 1;
  53. int i;
  54. vector signed short t1, t2, x, y, tmp1, tmp2;
  55. vector signed short *vbuf, *vtmp;
  56. vector unsigned char align;
  57. { // Lift 0
  58. IDWTELEM * const ref = b + w2 - 1;
  59. IDWTELEM b_0 = b[0];
  60. vector signed short v7 = vec_splat_s16(7);
  61. vbuf = (vector signed short *)b;
  62. tmp1 = vec_ld (0, ref);
  63. align = vec_lvsl (0, ref);
  64. tmp2 = vec_ld (15, ref);
  65. t1 = vec_perm(tmp1, tmp2, align);
  66. for (i=0; i<w_l-15; i+=16) {
  67. #if 0
  68. /* b[i+0] = b[i+0] - ((3 * (ref[i+0] + ref[i+1]) + 4) >> 3);
  69. b[i+1] = b[i+1] - ((3 * (ref[i+1] + ref[i+2]) + 4) >> 3);
  70. b[i+2] = b[i+2] - ((3 * (ref[i+2] + ref[i+3]) + 4) >> 3);
  71. b[i+3] = b[i+3] - ((3 * (ref[i+3] + ref[i+4]) + 4) >> 3);*/
  72. b[i+0] = b[i+0] + ((7 * (ref[i+0] + ref[i+1])-1) >> 8);
  73. #else
  74. tmp1 = vec_ld (0, ref+8+i);
  75. tmp2 = vec_ld (15, ref+8+i);
  76. t2 = vec_perm(tmp1, tmp2, align);
  77. y = vec_add(t1, vec_sld(t1,t2,2));
  78. // y = vec_add(vec_add(y,y),y);
  79. tmp1 = vec_ld (0, ref+12+i);
  80. y = vec_add(y, vec_splat_s32(4));
  81. y = vec_sra(y, vec_splat_u32(3));
  82. tmp2 = vec_ld (15, ref+12+i);
  83. *vbuf = vec_sub(*vbuf, y);
  84. t1 = t2;
  85. vbuf++;
  86. t2 = vec_perm(tmp1, tmp2, align);
  87. y = vec_add(t1,vec_sld(t1,t2,4));
  88. y = vec_add(vec_add(y,y),y);
  89. tmp1 = vec_ld (0, ref+12+i);
  90. y = vec_add(y, vec_splat_s32(4));
  91. y = vec_sra(y, vec_splat_u32(3));
  92. tmp2 = vec_ld (15, ref+12+i);
  93. *vbuf = vec_sub(*vbuf, y);
  94. t1=t2;
  95. vbuf++;
  96. t2 = vec_perm(tmp1, tmp2, align);
  97. y = vec_add(t1,vec_sld(t1,t2,4));
  98. y = vec_add(vec_add(y,y),y);
  99. tmp1 = vec_ld (0, ref+16+i);
  100. y = vec_add(y, vec_splat_s32(4));
  101. y = vec_sra(y, vec_splat_u32(3));
  102. tmp2 = vec_ld (15, ref+16+i);
  103. *vbuf = vec_sub(*vbuf, y);
  104. t1=t2;
  105. t2 = vec_perm(tmp1, tmp2, align);
  106. y = vec_add(t1,vec_sld(t1,t2,4));
  107. y = vec_add(vec_add(y,y),y);
  108. vbuf++;
  109. y = vec_add(y, vec_splat_s32(4));
  110. y = vec_sra(y, vec_splat_u32(3));
  111. *vbuf = vec_sub(*vbuf, y);
  112. t1=t2;
  113. vbuf++;
  114. #endif
  115. }
  116. snow_horizontal_compose_lift_lead_out(i, b, b, ref, width, w_l, 0, W_DM, W_DO, W_DS);
  117. b[0] = b_0 - ((W_DM * 2 * ref[1]+W_DO)>>W_DS);
  118. }
  119. { // Lift 1
  120. DWTELEM * const dst = b+w2;
  121. i = 0;
  122. for(; (((long)&dst[i]) & 0xF) && i<w_r; i++){
  123. dst[i] = dst[i] - (b[i] + b[i + 1]);
  124. }
  125. align = vec_lvsl(0, b+i);
  126. tmp1 = vec_ld(0, b+i);
  127. vbuf = (vector signed int*) (dst + i);
  128. tmp2 = vec_ld(15, b+i);
  129. t1 = vec_perm(tmp1, tmp2, align);
  130. for (; i<w_r-3; i+=4) {
  131. #if 0
  132. dst[i] = dst[i] - (b[i] + b[i + 1]);
  133. dst[i+1] = dst[i+1] - (b[i+1] + b[i + 2]);
  134. dst[i+2] = dst[i+2] - (b[i+2] + b[i + 3]);
  135. dst[i+3] = dst[i+3] - (b[i+3] + b[i + 4]);
  136. #else
  137. tmp1 = vec_ld(0, b+4+i);
  138. tmp2 = vec_ld(15, b+4+i);
  139. t2 = vec_perm(tmp1, tmp2, align);
  140. y = vec_add(t1, vec_sld(t1,t2,4));
  141. *vbuf = vec_sub (*vbuf, y);
  142. vbuf++;
  143. t1 = t2;
  144. #endif
  145. }
  146. snow_horizontal_compose_lift_lead_out(i, dst, dst, b, width, w_r, 1, W_CM, W_CO, W_CS);
  147. }
  148. { // Lift 2
  149. DWTELEM * const ref = b+w2 - 1;
  150. DWTELEM b_0 = b[0];
  151. vbuf= (vector signed int *) b;
  152. tmp1 = vec_ld (0, ref);
  153. align = vec_lvsl (0, ref);
  154. tmp2 = vec_ld (15, ref);
  155. t1= vec_perm(tmp1, tmp2, align);
  156. i = 0;
  157. for (; i<w_l-15; i+=16) {
  158. #if 0
  159. b[i] = b[i] - (((8 -(ref[i] + ref[i+1])) - (b[i] <<2)) >> 4);
  160. b[i+1] = b[i+1] - (((8 -(ref[i+1] + ref[i+2])) - (b[i+1]<<2)) >> 4);
  161. b[i+2] = b[i+2] - (((8 -(ref[i+2] + ref[i+3])) - (b[i+2]<<2)) >> 4);
  162. b[i+3] = b[i+3] - (((8 -(ref[i+3] + ref[i+4])) - (b[i+3]<<2)) >> 4);
  163. #else
  164. tmp1 = vec_ld (0, ref+4+i);
  165. tmp2 = vec_ld (15, ref+4+i);
  166. t2 = vec_perm(tmp1, tmp2, align);
  167. y = vec_add(t1,vec_sld(t1,t2,4));
  168. y = vec_sub(vec_splat_s32(8),y);
  169. tmp1 = vec_ld (0, ref+8+i);
  170. x = vec_sl(*vbuf,vec_splat_u32(2));
  171. y = vec_sra(vec_sub(y,x),vec_splat_u32(4));
  172. tmp2 = vec_ld (15, ref+8+i);
  173. *vbuf = vec_sub( *vbuf, y);
  174. t1 = t2;
  175. vbuf++;
  176. t2 = vec_perm(tmp1, tmp2, align);
  177. y = vec_add(t1,vec_sld(t1,t2,4));
  178. y = vec_sub(vec_splat_s32(8),y);
  179. tmp1 = vec_ld (0, ref+12+i);
  180. x = vec_sl(*vbuf,vec_splat_u32(2));
  181. y = vec_sra(vec_sub(y,x),vec_splat_u32(4));
  182. tmp2 = vec_ld (15, ref+12+i);
  183. *vbuf = vec_sub( *vbuf, y);
  184. t1 = t2;
  185. vbuf++;
  186. t2 = vec_perm(tmp1, tmp2, align);
  187. y = vec_add(t1,vec_sld(t1,t2,4));
  188. y = vec_sub(vec_splat_s32(8),y);
  189. tmp1 = vec_ld (0, ref+16+i);
  190. x = vec_sl(*vbuf,vec_splat_u32(2));
  191. y = vec_sra(vec_sub(y,x),vec_splat_u32(4));
  192. tmp2 = vec_ld (15, ref+16+i);
  193. *vbuf = vec_sub( *vbuf, y);
  194. t1 = t2;
  195. vbuf++;
  196. t2 = vec_perm(tmp1, tmp2, align);
  197. y = vec_add(t1,vec_sld(t1,t2,4));
  198. y = vec_sub(vec_splat_s32(8),y);
  199. t1 = t2;
  200. x = vec_sl(*vbuf,vec_splat_u32(2));
  201. y = vec_sra(vec_sub(y,x),vec_splat_u32(4));
  202. *vbuf = vec_sub( *vbuf, y);
  203. vbuf++;
  204. #endif
  205. }
  206. snow_horizontal_compose_liftS_lead_out(i, b, b, ref, width, w_l);
  207. b[0] = b_0 - (((-2 * ref[1] + W_BO) - 4 * b_0) >> W_BS);
  208. }
  209. { // Lift 3
  210. DWTELEM * const src = b+w2;
  211. vbuf = (vector signed int *)b;
  212. vtmp = (vector signed int *)temp;
  213. i = 0;
  214. align = vec_lvsl(0, src);
  215. for (; i<w_r-3; i+=4) {
  216. #if 0
  217. temp[i] = src[i] - ((-3*(b[i] + b[i+1]))>>1);
  218. temp[i+1] = src[i+1] - ((-3*(b[i+1] + b[i+2]))>>1);
  219. temp[i+2] = src[i+2] - ((-3*(b[i+2] + b[i+3]))>>1);
  220. temp[i+3] = src[i+3] - ((-3*(b[i+3] + b[i+4]))>>1);
  221. #else
  222. tmp1 = vec_ld(0,src+i);
  223. t1 = vec_add(vbuf[0],vec_sld(vbuf[0],vbuf[1],4));
  224. tmp2 = vec_ld(15,src+i);
  225. t1 = vec_sub(vec_splat_s32(0),t1); //bad!
  226. t1 = vec_add(t1,vec_add(t1,t1));
  227. t2 = vec_perm(tmp1 ,tmp2 ,align);
  228. t1 = vec_sra(t1,vec_splat_u32(1));
  229. vbuf++;
  230. *vtmp = vec_sub(t2,t1);
  231. vtmp++;
  232. #endif
  233. }
  234. snow_horizontal_compose_lift_lead_out(i, temp, src, b, width, w_r, 1, -3, 0, 1);
  235. }
  236. {
  237. //Interleave
  238. int a;
  239. vector signed int *t = (vector signed int *)temp,
  240. *v = (vector signed int *)b;
  241. snow_interleave_line_header(&i, width, b, temp);
  242. for (; (i & 0xE) != 0xE; i-=2){
  243. b[i+1] = temp[i>>1];
  244. b[i] = b[i>>1];
  245. }
  246. for (i-=14; i>=0; i-=16){
  247. a=i/4;
  248. v[a+3]=vec_mergel(v[(a>>1)+1],t[(a>>1)+1]);
  249. v[a+2]=vec_mergeh(v[(a>>1)+1],t[(a>>1)+1]);
  250. v[a+1]=vec_mergel(v[a>>1],t[a>>1]);
  251. v[a]=vec_mergeh(v[a>>1],t[a>>1]);
  252. }
  253. }
  254. #endif
  255. }
  256. void ff_snow_vertical_compose97i_altivec(DWTELEM *b0, DWTELEM *b1, DWTELEM *b2, DWTELEM *b3, DWTELEM *b4, DWTELEM *b5, int width)
  257. {
  258. int i, w4 = width/4;
  259. vector signed int *v0, *v1,*v2,*v3,*v4,*v5;
  260. vector signed int t1, t2;
  261. v0=(vector signed int *)b0;
  262. v1=(vector signed int *)b1;
  263. v2=(vector signed int *)b2;
  264. v3=(vector signed int *)b3;
  265. v4=(vector signed int *)b4;
  266. v5=(vector signed int *)b5;
  267. for (i=0; i< w4;i++)
  268. {
  269. #if 0
  270. b4[i] -= (3*(b3[i] + b5[i])+4)>>3;
  271. b3[i] -= ((b2[i] + b4[i]));
  272. b2[i] += ((b1[i] + b3[i])+4*b2[i]+8)>>4;
  273. b1[i] += (3*(b0[i] + b2[i]))>>1;
  274. #else
  275. t1 = vec_add(v3[i], v5[i]);
  276. t2 = vec_add(t1, vec_add(t1,t1));
  277. t1 = vec_add(t2, vec_splat_s32(4));
  278. v4[i] = vec_sub(v4[i], vec_sra(t1,vec_splat_u32(3)));
  279. v3[i] = vec_sub(v3[i], vec_add(v2[i], v4[i]));
  280. t1 = vec_add(vec_splat_s32(8), vec_add(v1[i], v3[i]));
  281. t2 = vec_sl(v2[i], vec_splat_u32(2));
  282. v2[i] = vec_add(v2[i], vec_sra(vec_add(t1,t2),vec_splat_u32(4)));
  283. t1 = vec_add(v0[i], v2[i]);
  284. t2 = vec_add(t1, vec_add(t1,t1));
  285. v1[i] = vec_add(v1[i], vec_sra(t2,vec_splat_u32(1)));
  286. #endif
  287. }
  288. for(i*=4; i < width; i++)
  289. {
  290. b4[i] -= (W_DM*(b3[i] + b5[i])+W_DO)>>W_DS;
  291. b3[i] -= (W_CM*(b2[i] + b4[i])+W_CO)>>W_CS;
  292. b2[i] += (W_BM*(b1[i] + b3[i])+4*b2[i]+W_BO)>>W_BS;
  293. b1[i] += (W_AM*(b0[i] + b2[i])+W_AO)>>W_AS;
  294. }
  295. }
  296. #define LOAD_BLOCKS \
  297. tmp1 = vec_ld(0, &block[3][y*src_stride]);\
  298. align = vec_lvsl(0, &block[3][y*src_stride]);\
  299. tmp2 = vec_ld(15, &block[3][y*src_stride]);\
  300. \
  301. b3 = vec_perm(tmp1,tmp2,align);\
  302. \
  303. tmp1 = vec_ld(0, &block[2][y*src_stride]);\
  304. align = vec_lvsl(0, &block[2][y*src_stride]);\
  305. tmp2 = vec_ld(15, &block[2][y*src_stride]);\
  306. \
  307. b2 = vec_perm(tmp1,tmp2,align);\
  308. \
  309. tmp1 = vec_ld(0, &block[1][y*src_stride]);\
  310. align = vec_lvsl(0, &block[1][y*src_stride]);\
  311. tmp2 = vec_ld(15, &block[1][y*src_stride]);\
  312. \
  313. b1 = vec_perm(tmp1,tmp2,align);\
  314. \
  315. tmp1 = vec_ld(0, &block[0][y*src_stride]);\
  316. align = vec_lvsl(0, &block[0][y*src_stride]);\
  317. tmp2 = vec_ld(15, &block[0][y*src_stride]);\
  318. \
  319. b0 = vec_perm(tmp1,tmp2,align);
  320. #define LOAD_OBMCS \
  321. tmp1 = vec_ld(0, obmc1);\
  322. align = vec_lvsl(0, obmc1);\
  323. tmp2 = vec_ld(15, obmc1);\
  324. \
  325. ob1 = vec_perm(tmp1,tmp2,align);\
  326. \
  327. tmp1 = vec_ld(0, obmc2);\
  328. align = vec_lvsl(0, obmc2);\
  329. tmp2 = vec_ld(15, obmc2);\
  330. \
  331. ob2 = vec_perm(tmp1,tmp2,align);\
  332. \
  333. tmp1 = vec_ld(0, obmc3);\
  334. align = vec_lvsl(0, obmc3);\
  335. tmp2 = vec_ld(15, obmc3);\
  336. \
  337. ob3 = vec_perm(tmp1,tmp2,align);\
  338. \
  339. tmp1 = vec_ld(0, obmc4);\
  340. align = vec_lvsl(0, obmc4);\
  341. tmp2 = vec_ld(15, obmc4);\
  342. \
  343. ob4 = vec_perm(tmp1,tmp2,align);
  344. /* interleave logic
  345. * h1 <- [ a,b,a,b, a,b,a,b, a,b,a,b, a,b,a,b ]
  346. * h2 <- [ c,d,c,d, c,d,c,d, c,d,c,d, c,d,c,d ]
  347. * h <- [ a,b,c,d, a,b,c,d, a,b,c,d, a,b,c,d ]
  348. */
  349. #define STEPS_0_1\
  350. h1 = (vector unsigned short)\
  351. vec_mergeh(ob1, ob2);\
  352. \
  353. h2 = (vector unsigned short)\
  354. vec_mergeh(ob3, ob4);\
  355. \
  356. ih = (vector unsigned char)\
  357. vec_mergeh(h1,h2);\
  358. \
  359. l1 = (vector unsigned short) vec_mergeh(b3, b2);\
  360. \
  361. ih1 = (vector unsigned char) vec_mergel(h1, h2);\
  362. \
  363. l2 = (vector unsigned short) vec_mergeh(b1, b0);\
  364. \
  365. il = (vector unsigned char) vec_mergeh(l1, l2);\
  366. \
  367. v[0] = (vector signed int) vec_msum(ih, il, vec_splat_u32(0));\
  368. \
  369. il1 = (vector unsigned char) vec_mergel(l1, l2);\
  370. \
  371. v[1] = (vector signed int) vec_msum(ih1, il1, vec_splat_u32(0));
  372. #define FINAL_STEP_SCALAR\
  373. for(x=0; x<b_w; x++)\
  374. if(add){\
  375. vbuf[x] += dst[x + src_x];\
  376. vbuf[x] = (vbuf[x] + (1<<(FRAC_BITS-1))) >> FRAC_BITS;\
  377. if(vbuf[x]&(~255)) vbuf[x]= ~(vbuf[x]>>31);\
  378. dst8[x + y*src_stride] = vbuf[x];\
  379. }else{\
  380. dst[x + src_x] -= vbuf[x];\
  381. }
  382. static void inner_add_yblock_bw_8_obmc_16_altivec(uint8_t *obmc,
  383. const int obmc_stride,
  384. uint8_t * * block, int b_w,
  385. int b_h, int src_x, int src_y,
  386. int src_stride, slice_buffer * sb,
  387. int add, uint8_t * dst8)
  388. {
  389. int y, x;
  390. DWTELEM * dst;
  391. vector unsigned short h1, h2, l1, l2;
  392. vector unsigned char ih, il, ih1, il1, tmp1, tmp2, align;
  393. vector unsigned char b0,b1,b2,b3;
  394. vector unsigned char ob1,ob2,ob3,ob4;
  395. DECLARE_ALIGNED_16(int, vbuf[16]);
  396. vector signed int *v = (vector signed int *)vbuf, *d;
  397. for(y=0; y<b_h; y++){
  398. //FIXME ugly misuse of obmc_stride
  399. uint8_t *obmc1= obmc + y*obmc_stride;
  400. uint8_t *obmc2= obmc1+ (obmc_stride>>1);
  401. uint8_t *obmc3= obmc1+ obmc_stride*(obmc_stride>>1);
  402. uint8_t *obmc4= obmc3+ (obmc_stride>>1);
  403. dst = slice_buffer_get_line(sb, src_y + y);
  404. d = (vector signed int *)(dst + src_x);
  405. //FIXME i could avoid some loads!
  406. // load blocks
  407. LOAD_BLOCKS
  408. // load obmcs
  409. LOAD_OBMCS
  410. // steps 0 1
  411. STEPS_0_1
  412. FINAL_STEP_SCALAR
  413. }
  414. }
  415. #define STEPS_2_3\
  416. h1 = (vector unsigned short) vec_mergel(ob1, ob2);\
  417. \
  418. h2 = (vector unsigned short) vec_mergel(ob3, ob4);\
  419. \
  420. ih = (vector unsigned char) vec_mergeh(h1,h2);\
  421. \
  422. l1 = (vector unsigned short) vec_mergel(b3, b2);\
  423. \
  424. l2 = (vector unsigned short) vec_mergel(b1, b0);\
  425. \
  426. ih1 = (vector unsigned char) vec_mergel(h1,h2);\
  427. \
  428. il = (vector unsigned char) vec_mergeh(l1,l2);\
  429. \
  430. v[2] = (vector signed int) vec_msum(ih, il, vec_splat_u32(0));\
  431. \
  432. il1 = (vector unsigned char) vec_mergel(l1,l2);\
  433. \
  434. v[3] = (vector signed int) vec_msum(ih1, il1, vec_splat_u32(0));
  435. static void inner_add_yblock_bw_16_obmc_32_altivec(uint8_t *obmc,
  436. const int obmc_stride,
  437. uint8_t * * block, int b_w,
  438. int b_h, int src_x, int src_y,
  439. int src_stride, slice_buffer * sb,
  440. int add, uint8_t * dst8)
  441. {
  442. int y, x;
  443. DWTELEM * dst;
  444. vector unsigned short h1, h2, l1, l2;
  445. vector unsigned char ih, il, ih1, il1, tmp1, tmp2, align;
  446. vector unsigned char b0,b1,b2,b3;
  447. vector unsigned char ob1,ob2,ob3,ob4;
  448. DECLARE_ALIGNED_16(int, vbuf[b_w]);
  449. vector signed int *v = (vector signed int *)vbuf, *d;
  450. for(y=0; y<b_h; y++){
  451. //FIXME ugly misuse of obmc_stride
  452. uint8_t *obmc1= obmc + y*obmc_stride;
  453. uint8_t *obmc2= obmc1+ (obmc_stride>>1);
  454. uint8_t *obmc3= obmc1+ obmc_stride*(obmc_stride>>1);
  455. uint8_t *obmc4= obmc3+ (obmc_stride>>1);
  456. dst = slice_buffer_get_line(sb, src_y + y);
  457. d = (vector signed int *)(dst + src_x);
  458. // load blocks
  459. LOAD_BLOCKS
  460. // load obmcs
  461. LOAD_OBMCS
  462. // steps 0 1 2 3
  463. STEPS_0_1
  464. STEPS_2_3
  465. FINAL_STEP_SCALAR
  466. }
  467. }
  468. #define FINAL_STEP_VEC \
  469. \
  470. if(add)\
  471. {\
  472. for(x=0; x<b_w/4; x++)\
  473. {\
  474. v[x] = vec_add(v[x], d[x]);\
  475. v[x] = vec_sra(vec_add(v[x],\
  476. vec_sl( vec_splat_s32(1),\
  477. vec_splat_u32(7))),\
  478. vec_splat_u32(8));\
  479. \
  480. mask = (vector bool int) vec_sl((vector signed int)\
  481. vec_cmpeq(v[x],v[x]),vec_splat_u32(8));\
  482. mask = (vector bool int) vec_and(v[x],vec_nor(mask,mask));\
  483. \
  484. mask = (vector bool int)\
  485. vec_cmpeq((vector signed int)mask,\
  486. (vector signed int)vec_splat_u32(0));\
  487. \
  488. vs = vec_sra(v[x],vec_splat_u32(8));\
  489. vs = vec_sra(v[x],vec_splat_u32(8));\
  490. vs = vec_sra(v[x],vec_splat_u32(15));\
  491. \
  492. vs = vec_nor(vs,vs);\
  493. \
  494. v[x]= vec_sel(v[x],vs,mask);\
  495. }\
  496. \
  497. for(x=0; x<b_w; x++)\
  498. dst8[x + y*src_stride] = vbuf[x];\
  499. \
  500. }\
  501. else\
  502. for(x=0; x<b_w/4; x++)\
  503. d[x] = vec_sub(d[x], v[x]);
  504. static void inner_add_yblock_a_bw_8_obmc_16_altivec(uint8_t *obmc,
  505. const int obmc_stride,
  506. uint8_t * * block, int b_w,
  507. int b_h, int src_x, int src_y,
  508. int src_stride, slice_buffer * sb,
  509. int add, uint8_t * dst8)
  510. {
  511. int y, x;
  512. DWTELEM * dst;
  513. vector bool int mask;
  514. vector signed int vs;
  515. vector unsigned short h1, h2, l1, l2;
  516. vector unsigned char ih, il, ih1, il1, tmp1, tmp2, align;
  517. vector unsigned char b0,b1,b2,b3;
  518. vector unsigned char ob1,ob2,ob3,ob4;
  519. DECLARE_ALIGNED_16(int, vbuf[16]);
  520. vector signed int *v = (vector signed int *)vbuf, *d;
  521. for(y=0; y<b_h; y++){
  522. //FIXME ugly misuse of obmc_stride
  523. uint8_t *obmc1= obmc + y*obmc_stride;
  524. uint8_t *obmc2= obmc1+ (obmc_stride>>1);
  525. uint8_t *obmc3= obmc1+ obmc_stride*(obmc_stride>>1);
  526. uint8_t *obmc4= obmc3+ (obmc_stride>>1);
  527. dst = slice_buffer_get_line(sb, src_y + y);
  528. d = (vector signed int *)(dst + src_x);
  529. //FIXME i could avoid some loads!
  530. // load blocks
  531. LOAD_BLOCKS
  532. // load obmcs
  533. LOAD_OBMCS
  534. // steps 0 1
  535. STEPS_0_1
  536. FINAL_STEP_VEC
  537. }
  538. }
  539. static void inner_add_yblock_a_bw_16_obmc_32_altivec(uint8_t *obmc,
  540. const int obmc_stride,
  541. uint8_t * * block, int b_w,
  542. int b_h, int src_x, int src_y,
  543. int src_stride, slice_buffer * sb,
  544. int add, uint8_t * dst8)
  545. {
  546. int y, x;
  547. DWTELEM * dst;
  548. vector bool int mask;
  549. vector signed int vs;
  550. vector unsigned short h1, h2, l1, l2;
  551. vector unsigned char ih, il, ih1, il1, tmp1, tmp2, align;
  552. vector unsigned char b0,b1,b2,b3;
  553. vector unsigned char ob1,ob2,ob3,ob4;
  554. DECLARE_ALIGNED_16(int, vbuf[b_w]);
  555. vector signed int *v = (vector signed int *)vbuf, *d;
  556. for(y=0; y<b_h; y++){
  557. //FIXME ugly misuse of obmc_stride
  558. uint8_t *obmc1= obmc + y*obmc_stride;
  559. uint8_t *obmc2= obmc1+ (obmc_stride>>1);
  560. uint8_t *obmc3= obmc1+ obmc_stride*(obmc_stride>>1);
  561. uint8_t *obmc4= obmc3+ (obmc_stride>>1);
  562. dst = slice_buffer_get_line(sb, src_y + y);
  563. d = (vector signed int *)(dst + src_x);
  564. // load blocks
  565. LOAD_BLOCKS
  566. // load obmcs
  567. LOAD_OBMCS
  568. // steps 0 1 2 3
  569. STEPS_0_1
  570. STEPS_2_3
  571. FINAL_STEP_VEC
  572. }
  573. }
  574. void ff_snow_inner_add_yblock_altivec(uint8_t *obmc, const int obmc_stride,
  575. uint8_t * * block, int b_w, int b_h,
  576. int src_x, int src_y, int src_stride,
  577. slice_buffer * sb, int add,
  578. uint8_t * dst8)
  579. {
  580. if (src_x&15) {
  581. if (b_w == 16)
  582. inner_add_yblock_bw_16_obmc_32_altivec(obmc, obmc_stride, block,
  583. b_w, b_h, src_x, src_y,
  584. src_stride, sb, add, dst8);
  585. else if (b_w == 8)
  586. inner_add_yblock_bw_8_obmc_16_altivec(obmc, obmc_stride, block,
  587. b_w, b_h, src_x, src_y,
  588. src_stride, sb, add, dst8);
  589. else
  590. ff_snow_inner_add_yblock(obmc, obmc_stride, block, b_w, b_h, src_x,
  591. src_y, src_stride, sb, add, dst8);
  592. } else {
  593. if (b_w == 16)
  594. inner_add_yblock_a_bw_16_obmc_32_altivec(obmc, obmc_stride, block,
  595. b_w, b_h, src_x, src_y,
  596. src_stride, sb, add, dst8);
  597. else if (b_w == 8)
  598. inner_add_yblock_a_bw_8_obmc_16_altivec(obmc, obmc_stride, block,
  599. b_w, b_h, src_x, src_y,
  600. src_stride, sb, add, dst8);
  601. else
  602. ff_snow_inner_add_yblock(obmc, obmc_stride, block, b_w, b_h, src_x,
  603. src_y, src_stride, sb, add, dst8);
  604. }
  605. }
  606. void snow_init_altivec(DSPContext* c, AVCodecContext *avctx)
  607. {
  608. #if 0
  609. c->horizontal_compose97i = ff_snow_horizontal_compose97i_altivec;
  610. c->vertical_compose97i = ff_snow_vertical_compose97i_altivec;
  611. c->inner_add_yblock = ff_snow_inner_add_yblock_altivec;
  612. #endif
  613. }