You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

320 lines
7.6KB

  1. ;******************************************************************************
  2. ;* VC1 deblocking optimizations
  3. ;* Copyright (c) 2009 David Conrad
  4. ;*
  5. ;* This file is part of Libav.
  6. ;*
  7. ;* Libav is free software; you can redistribute it and/or
  8. ;* modify it under the terms of the GNU Lesser General Public
  9. ;* License as published by the Free Software Foundation; either
  10. ;* version 2.1 of the License, or (at your option) any later version.
  11. ;*
  12. ;* Libav is distributed in the hope that it will be useful,
  13. ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. ;* Lesser General Public License for more details.
  16. ;*
  17. ;* You should have received a copy of the GNU Lesser General Public
  18. ;* License along with Libav; if not, write to the Free Software
  19. ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. ;******************************************************************************
  21. %include "libavutil/x86/x86util.asm"
  22. cextern pw_4
  23. cextern pw_5
  24. section .text
  25. ; dst_low, dst_high (src), zero
  26. ; zero-extends one vector from 8 to 16 bits
  27. %macro UNPACK_8TO16 4
  28. mova m%2, m%3
  29. punpckh%1 m%3, m%4
  30. punpckl%1 m%2, m%4
  31. %endmacro
  32. %macro STORE_4_WORDS_MMX 6
  33. movd %6d, %5
  34. %if mmsize==16
  35. psrldq %5, 4
  36. %else
  37. psrlq %5, 32
  38. %endif
  39. mov %1, %6w
  40. shr %6, 16
  41. mov %2, %6w
  42. movd %6d, %5
  43. mov %3, %6w
  44. shr %6, 16
  45. mov %4, %6w
  46. %endmacro
  47. %macro STORE_4_WORDS_SSE4 6
  48. pextrw %1, %5, %6+0
  49. pextrw %2, %5, %6+1
  50. pextrw %3, %5, %6+2
  51. pextrw %4, %5, %6+3
  52. %endmacro
  53. ; in: p1 p0 q0 q1, clobbers p0
  54. ; out: p1 = (2*(p1 - q1) - 5*(p0 - q0) + 4) >> 3
  55. %macro VC1_LOOP_FILTER_A0 4
  56. psubw %1, %4
  57. psubw %2, %3
  58. paddw %1, %1
  59. pmullw %2, [pw_5]
  60. psubw %1, %2
  61. paddw %1, [pw_4]
  62. psraw %1, 3
  63. %endmacro
  64. ; in: p0 q0 a0 a1 a2
  65. ; m0 m1 m7 m6 m5
  66. ; %1: size
  67. ; out: m0=p0' m1=q0'
  68. %macro VC1_FILTER 1
  69. PABSW m4, m7
  70. PABSW m3, m6
  71. PABSW m2, m5
  72. mova m6, m4
  73. pminsw m3, m2
  74. pcmpgtw m6, m3 ; if (a2 < a0 || a1 < a0)
  75. psubw m3, m4
  76. pmullw m3, [pw_5] ; 5*(a3 - a0)
  77. PABSW m2, m3
  78. psraw m2, 3 ; abs(d/8)
  79. pxor m7, m3 ; d_sign ^= a0_sign
  80. pxor m5, m5
  81. movd m3, r2d
  82. %if %1 > 4
  83. punpcklbw m3, m3
  84. %endif
  85. punpcklbw m3, m5
  86. pcmpgtw m3, m4 ; if (a0 < pq)
  87. pand m6, m3
  88. mova m3, m0
  89. psubw m3, m1
  90. PABSW m4, m3
  91. psraw m4, 1
  92. pxor m3, m7 ; d_sign ^ clip_sign
  93. psraw m3, 15
  94. pminsw m2, m4 ; min(d, clip)
  95. pcmpgtw m4, m5
  96. pand m6, m4 ; filt3 (C return value)
  97. ; each set of 4 pixels is not filtered if the 3rd is not
  98. %if mmsize==16
  99. pshuflw m4, m6, 0xaa
  100. %if %1 > 4
  101. pshufhw m4, m4, 0xaa
  102. %endif
  103. %else
  104. pshufw m4, m6, 0xaa
  105. %endif
  106. pandn m3, m4
  107. pand m2, m6
  108. pand m3, m2 ; d final
  109. psraw m7, 15
  110. pxor m3, m7
  111. psubw m3, m7
  112. psubw m0, m3
  113. paddw m1, m3
  114. packuswb m0, m0
  115. packuswb m1, m1
  116. %endmacro
  117. ; 1st param: size of filter
  118. ; 2nd param: mov suffix equivalent to the filter size
  119. %macro VC1_V_LOOP_FILTER 2
  120. pxor m5, m5
  121. mov%2 m6, [r4]
  122. mov%2 m4, [r4+r1]
  123. mov%2 m7, [r4+2*r1]
  124. mov%2 m0, [r4+r3]
  125. punpcklbw m6, m5
  126. punpcklbw m4, m5
  127. punpcklbw m7, m5
  128. punpcklbw m0, m5
  129. VC1_LOOP_FILTER_A0 m6, m4, m7, m0
  130. mov%2 m1, [r0]
  131. mov%2 m2, [r0+r1]
  132. punpcklbw m1, m5
  133. punpcklbw m2, m5
  134. mova m4, m0
  135. VC1_LOOP_FILTER_A0 m7, m4, m1, m2
  136. mov%2 m3, [r0+2*r1]
  137. mov%2 m4, [r0+r3]
  138. punpcklbw m3, m5
  139. punpcklbw m4, m5
  140. mova m5, m1
  141. VC1_LOOP_FILTER_A0 m5, m2, m3, m4
  142. VC1_FILTER %1
  143. mov%2 [r4+r3], m0
  144. mov%2 [r0], m1
  145. %endmacro
  146. ; 1st param: size of filter
  147. ; NOTE: UNPACK_8TO16 this number of 8 bit numbers are in half a register
  148. ; 2nd (optional) param: temp register to use for storing words
  149. %macro VC1_H_LOOP_FILTER 1-2
  150. %if %1 == 4
  151. movq m0, [r0 -4]
  152. movq m1, [r0+ r1-4]
  153. movq m2, [r0+2*r1-4]
  154. movq m3, [r0+ r3-4]
  155. TRANSPOSE4x4B 0, 1, 2, 3, 4
  156. %else
  157. movq m0, [r0 -4]
  158. movq m4, [r0+ r1-4]
  159. movq m1, [r0+2*r1-4]
  160. movq m5, [r0+ r3-4]
  161. movq m2, [r4 -4]
  162. movq m6, [r4+ r1-4]
  163. movq m3, [r4+2*r1-4]
  164. movq m7, [r4+ r3-4]
  165. punpcklbw m0, m4
  166. punpcklbw m1, m5
  167. punpcklbw m2, m6
  168. punpcklbw m3, m7
  169. TRANSPOSE4x4W 0, 1, 2, 3, 4
  170. %endif
  171. pxor m5, m5
  172. UNPACK_8TO16 bw, 6, 0, 5
  173. UNPACK_8TO16 bw, 7, 1, 5
  174. VC1_LOOP_FILTER_A0 m6, m0, m7, m1
  175. UNPACK_8TO16 bw, 4, 2, 5
  176. mova m0, m1 ; m0 = p0
  177. VC1_LOOP_FILTER_A0 m7, m1, m4, m2
  178. UNPACK_8TO16 bw, 1, 3, 5
  179. mova m5, m4
  180. VC1_LOOP_FILTER_A0 m5, m2, m1, m3
  181. SWAP 1, 4 ; m1 = q0
  182. VC1_FILTER %1
  183. punpcklbw m0, m1
  184. %if %0 > 1
  185. STORE_4_WORDS_MMX [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, %2
  186. %if %1 > 4
  187. psrldq m0, 4
  188. STORE_4_WORDS_MMX [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, %2
  189. %endif
  190. %else
  191. STORE_4_WORDS_SSE4 [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, 0
  192. STORE_4_WORDS_SSE4 [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, 4
  193. %endif
  194. %endmacro
  195. %macro START_V_FILTER 0
  196. mov r4, r0
  197. lea r3, [4*r1]
  198. sub r4, r3
  199. lea r3, [r1+2*r1]
  200. imul r2, 0x01010101
  201. %endmacro
  202. %macro START_H_FILTER 1
  203. lea r3, [r1+2*r1]
  204. %if %1 > 4
  205. lea r4, [r0+4*r1]
  206. %endif
  207. imul r2, 0x01010101
  208. %endmacro
  209. %macro VC1_LF_MMX 1
  210. INIT_MMX
  211. cglobal vc1_v_loop_filter_internal_%1
  212. VC1_V_LOOP_FILTER 4, d
  213. ret
  214. cglobal vc1_h_loop_filter_internal_%1
  215. VC1_H_LOOP_FILTER 4, r4
  216. ret
  217. ; void ff_vc1_v_loop_filter4_mmx2(uint8_t *src, int stride, int pq)
  218. cglobal vc1_v_loop_filter4_%1, 3,5,0
  219. START_V_FILTER
  220. call vc1_v_loop_filter_internal_%1
  221. RET
  222. ; void ff_vc1_h_loop_filter4_mmx2(uint8_t *src, int stride, int pq)
  223. cglobal vc1_h_loop_filter4_%1, 3,5,0
  224. START_H_FILTER 4
  225. call vc1_h_loop_filter_internal_%1
  226. RET
  227. ; void ff_vc1_v_loop_filter8_mmx2(uint8_t *src, int stride, int pq)
  228. cglobal vc1_v_loop_filter8_%1, 3,5,0
  229. START_V_FILTER
  230. call vc1_v_loop_filter_internal_%1
  231. add r4, 4
  232. add r0, 4
  233. call vc1_v_loop_filter_internal_%1
  234. RET
  235. ; void ff_vc1_h_loop_filter8_mmx2(uint8_t *src, int stride, int pq)
  236. cglobal vc1_h_loop_filter8_%1, 3,5,0
  237. START_H_FILTER 4
  238. call vc1_h_loop_filter_internal_%1
  239. lea r0, [r0+4*r1]
  240. call vc1_h_loop_filter_internal_%1
  241. RET
  242. %endmacro
  243. %define PABSW PABSW_MMXEXT
  244. VC1_LF_MMX mmx2
  245. INIT_XMM
  246. ; void ff_vc1_v_loop_filter8_sse2(uint8_t *src, int stride, int pq)
  247. cglobal vc1_v_loop_filter8_sse2, 3,5,8
  248. START_V_FILTER
  249. VC1_V_LOOP_FILTER 8, q
  250. RET
  251. ; void ff_vc1_h_loop_filter8_sse2(uint8_t *src, int stride, int pq)
  252. cglobal vc1_h_loop_filter8_sse2, 3,6,8
  253. START_H_FILTER 8
  254. VC1_H_LOOP_FILTER 8, r5
  255. RET
  256. %define PABSW PABSW_SSSE3
  257. INIT_MMX
  258. ; void ff_vc1_v_loop_filter4_ssse3(uint8_t *src, int stride, int pq)
  259. cglobal vc1_v_loop_filter4_ssse3, 3,5,0
  260. START_V_FILTER
  261. VC1_V_LOOP_FILTER 4, d
  262. RET
  263. ; void ff_vc1_h_loop_filter4_ssse3(uint8_t *src, int stride, int pq)
  264. cglobal vc1_h_loop_filter4_ssse3, 3,5,0
  265. START_H_FILTER 4
  266. VC1_H_LOOP_FILTER 4, r4
  267. RET
  268. INIT_XMM
  269. ; void ff_vc1_v_loop_filter8_ssse3(uint8_t *src, int stride, int pq)
  270. cglobal vc1_v_loop_filter8_ssse3, 3,5,8
  271. START_V_FILTER
  272. VC1_V_LOOP_FILTER 8, q
  273. RET
  274. ; void ff_vc1_h_loop_filter8_ssse3(uint8_t *src, int stride, int pq)
  275. cglobal vc1_h_loop_filter8_ssse3, 3,6,8
  276. START_H_FILTER 8
  277. VC1_H_LOOP_FILTER 8, r5
  278. RET
  279. ; void ff_vc1_h_loop_filter8_sse4(uint8_t *src, int stride, int pq)
  280. cglobal vc1_h_loop_filter8_sse4, 3,5,8
  281. START_H_FILTER 8
  282. VC1_H_LOOP_FILTER 8
  283. RET