You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

337 lines
9.0KB

  1. ;******************************************************************************
  2. ;* SSE-optimized functions for the DCA decoder
  3. ;* Copyright (C) 2012-2014 Christophe Gisquet <christophe.gisquet@gmail.com>
  4. ;*
  5. ;* This file is part of Libav.
  6. ;*
  7. ;* Libav is free software; you can redistribute it and/or
  8. ;* modify it under the terms of the GNU Lesser General Public
  9. ;* License as published by the Free Software Foundation; either
  10. ;* version 2.1 of the License, or (at your option) any later version.
  11. ;*
  12. ;* Libav is distributed in the hope that it will be useful,
  13. ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. ;* Lesser General Public License for more details.
  16. ;*
  17. ;* You should have received a copy of the GNU Lesser General Public
  18. ;* License along with Libav; if not, write to the Free Software
  19. ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. ;******************************************************************************
  21. %include "libavutil/x86/x86util.asm"
  22. SECTION_RODATA
  23. pf_inv16: times 4 dd 0x3D800000 ; 1/16
  24. SECTION .text
  25. ; %1=v0/v1 %2=in1 %3=in2
  26. %macro FIR_LOOP 2-3
  27. .loop%1:
  28. %define va m1
  29. %define vb m2
  30. %if %1
  31. %define OFFSET 0
  32. %else
  33. %define OFFSET NUM_COEF*count
  34. %endif
  35. ; for v0, incrementing and for v1, decrementing
  36. mova va, [cf0q + OFFSET]
  37. mova vb, [cf0q + OFFSET + 4*NUM_COEF]
  38. %if %0 == 3
  39. mova m4, [cf0q + OFFSET + mmsize]
  40. mova m0, [cf0q + OFFSET + 4*NUM_COEF + mmsize]
  41. %endif
  42. mulps va, %2
  43. mulps vb, %2
  44. %if %0 == 3
  45. mulps m4, %3
  46. mulps m0, %3
  47. addps va, m4
  48. addps vb, m0
  49. %endif
  50. ; va = va1 va2 va3 va4
  51. ; vb = vb1 vb2 vb3 vb4
  52. %if %1
  53. SWAP va, vb
  54. %endif
  55. mova m4, va
  56. unpcklps va, vb ; va3 vb3 va4 vb4
  57. unpckhps m4, vb ; va1 vb1 va2 vb2
  58. addps m4, va ; va1+3 vb1+3 va2+4 vb2+4
  59. movhlps vb, m4 ; va1+3 vb1+3
  60. addps vb, m4 ; va0..4 vb0..4
  61. movlps [outq + count], vb
  62. %if %1
  63. sub cf0q, 8*NUM_COEF
  64. %endif
  65. add count, 8
  66. jl .loop%1
  67. %endmacro
  68. ; void dca_lfe_fir(float *out, float *in, float *coefs)
  69. %macro DCA_LFE_FIR 1
  70. cglobal dca_lfe_fir%1, 3,3,6-%1, out, in, cf0
  71. %define IN1 m3
  72. %define IN2 m5
  73. %define count inq
  74. %define NUM_COEF 4*(2-%1)
  75. %define NUM_OUT 32*(%1+1)
  76. movu IN1, [inq + 4 - 1*mmsize]
  77. shufps IN1, IN1, q0123
  78. %if %1 == 0
  79. movu IN2, [inq + 4 - 2*mmsize]
  80. shufps IN2, IN2, q0123
  81. %endif
  82. mov count, -4*NUM_OUT
  83. add cf0q, 4*NUM_COEF*NUM_OUT
  84. add outq, 4*NUM_OUT
  85. ; compute v0 first
  86. %if %1 == 0
  87. FIR_LOOP 0, IN1, IN2
  88. %else
  89. FIR_LOOP 0, IN1
  90. %endif
  91. shufps IN1, IN1, q0123
  92. mov count, -4*NUM_OUT
  93. ; cf1 already correctly positioned
  94. add outq, 4*NUM_OUT ; outq now at out2
  95. sub cf0q, 8*NUM_COEF
  96. %if %1 == 0
  97. shufps IN2, IN2, q0123
  98. FIR_LOOP 1, IN2, IN1
  99. %else
  100. FIR_LOOP 1, IN1
  101. %endif
  102. RET
  103. %endmacro
  104. INIT_XMM sse
  105. DCA_LFE_FIR 0
  106. DCA_LFE_FIR 1
  107. %macro SETZERO 1
  108. %if cpuflag(sse2) && notcpuflag(avx)
  109. pxor %1, %1
  110. %else
  111. xorps %1, %1, %1
  112. %endif
  113. %endmacro
  114. %macro SHUF 3
  115. %if cpuflag(avx)
  116. mova %3, [%2 - 16]
  117. vperm2f128 %1, %3, %3, 1
  118. vshufps %1, %1, %1, q0123
  119. %elif cpuflag(sse2)
  120. pshufd %1, [%2], q0123
  121. %else
  122. mova %1, [%2]
  123. shufps %1, %1, q0123
  124. %endif
  125. %endmacro
  126. %macro INNER_LOOP 1
  127. ; reading backwards: ptr1 = synth_buf + j + i; ptr2 = synth_buf + j - i
  128. ;~ a += window[i + j] * (-synth_buf[15 - i + j])
  129. ;~ b += window[i + j + 16] * (synth_buf[i + j])
  130. SHUF m5, ptr2 + j + (15 - 3) * 4, m6
  131. mova m6, [ptr1 + j]
  132. %if ARCH_X86_64
  133. SHUF m11, ptr2 + j + (15 - 3) * 4 - mmsize, m12
  134. mova m12, [ptr1 + j + mmsize]
  135. %endif
  136. %if cpuflag(fma3)
  137. fmaddps m2, m6, [win + %1 + j + 16 * 4], m2
  138. fnmaddps m1, m5, [win + %1 + j], m1
  139. %if ARCH_X86_64
  140. fmaddps m8, m12, [win + %1 + j + mmsize + 16 * 4], m8
  141. fnmaddps m7, m11, [win + %1 + j + mmsize], m7
  142. %endif
  143. %else ; non-FMA
  144. mulps m6, m6, [win + %1 + j + 16 * 4]
  145. mulps m5, m5, [win + %1 + j]
  146. %if ARCH_X86_64
  147. mulps m12, m12, [win + %1 + j + mmsize + 16 * 4]
  148. mulps m11, m11, [win + %1 + j + mmsize]
  149. %endif
  150. addps m2, m2, m6
  151. subps m1, m1, m5
  152. %if ARCH_X86_64
  153. addps m8, m8, m12
  154. subps m7, m7, m11
  155. %endif
  156. %endif ; cpuflag(fma3)
  157. ;~ c += window[i + j + 32] * (synth_buf[16 + i + j])
  158. ;~ d += window[i + j + 48] * (synth_buf[31 - i + j])
  159. SHUF m6, ptr2 + j + (31 - 3) * 4, m5
  160. mova m5, [ptr1 + j + 16 * 4]
  161. %if ARCH_X86_64
  162. SHUF m12, ptr2 + j + (31 - 3) * 4 - mmsize, m11
  163. mova m11, [ptr1 + j + mmsize + 16 * 4]
  164. %endif
  165. %if cpuflag(fma3)
  166. fmaddps m3, m5, [win + %1 + j + 32 * 4], m3
  167. fmaddps m4, m6, [win + %1 + j + 48 * 4], m4
  168. %if ARCH_X86_64
  169. fmaddps m9, m11, [win + %1 + j + mmsize + 32 * 4], m9
  170. fmaddps m10, m12, [win + %1 + j + mmsize + 48 * 4], m10
  171. %endif
  172. %else ; non-FMA
  173. mulps m5, m5, [win + %1 + j + 32 * 4]
  174. mulps m6, m6, [win + %1 + j + 48 * 4]
  175. %if ARCH_X86_64
  176. mulps m11, m11, [win + %1 + j + mmsize + 32 * 4]
  177. mulps m12, m12, [win + %1 + j + mmsize + 48 * 4]
  178. %endif
  179. addps m3, m3, m5
  180. addps m4, m4, m6
  181. %if ARCH_X86_64
  182. addps m9, m9, m11
  183. addps m10, m10, m12
  184. %endif
  185. %endif ; cpuflag(fma3)
  186. sub j, 64 * 4
  187. %endmacro
  188. ; void ff_synth_filter_inner_<opt>(float *synth_buf, float synth_buf2[32],
  189. ; const float window[512], float out[32],
  190. ; intptr_t offset, float scale)
  191. %macro SYNTH_FILTER 0
  192. cglobal synth_filter_inner, 0, 6 + 4 * ARCH_X86_64, 7 + 6 * ARCH_X86_64, \
  193. synth_buf, synth_buf2, window, out, off, scale
  194. %define scale m0
  195. %if ARCH_X86_32 || WIN64
  196. %if cpuflag(sse2) && notcpuflag(avx)
  197. movd scale, scalem
  198. SPLATD m0
  199. %else
  200. VBROADCASTSS m0, scalem
  201. %endif
  202. ; Make sure offset is in a register and not on the stack
  203. %define OFFQ r4q
  204. %else
  205. SPLATD xmm0
  206. %if cpuflag(avx)
  207. vinsertf128 m0, m0, xmm0, 1
  208. %endif
  209. %define OFFQ offq
  210. %endif
  211. ; prepare inner counter limit 1
  212. mov r5q, 480
  213. sub r5q, offmp
  214. and r5q, -64
  215. shl r5q, 2
  216. %if ARCH_X86_32 || notcpuflag(avx)
  217. mov OFFQ, r5q
  218. %define i r5q
  219. mov i, 16 * 4 - (ARCH_X86_64 + 1) * mmsize ; main loop counter
  220. %else
  221. %define i 0
  222. %define OFFQ r5q
  223. %endif
  224. %define buf2 synth_buf2q
  225. %if ARCH_X86_32
  226. mov buf2, synth_buf2mp
  227. %endif
  228. .mainloop
  229. ; m1 = a m2 = b m3 = c m4 = d
  230. SETZERO m3
  231. SETZERO m4
  232. mova m1, [buf2 + i]
  233. mova m2, [buf2 + i + 16 * 4]
  234. %if ARCH_X86_32
  235. %define ptr1 r0q
  236. %define ptr2 r1q
  237. %define win r2q
  238. %define j r3q
  239. mov win, windowm
  240. mov ptr1, synth_bufm
  241. %if ARCH_X86_32 || notcpuflag(avx)
  242. add win, i
  243. add ptr1, i
  244. %endif
  245. %else ; ARCH_X86_64
  246. %define ptr1 r6q
  247. %define ptr2 r7q ; must be loaded
  248. %define win r8q
  249. %define j r9q
  250. SETZERO m9
  251. SETZERO m10
  252. mova m7, [buf2 + i + mmsize]
  253. mova m8, [buf2 + i + mmsize + 16 * 4]
  254. lea win, [windowq + i]
  255. lea ptr1, [synth_bufq + i]
  256. %endif
  257. mov ptr2, synth_bufmp
  258. ; prepare the inner loop counter
  259. mov j, OFFQ
  260. %if ARCH_X86_32 || notcpuflag(avx)
  261. sub ptr2, i
  262. %endif
  263. .loop1:
  264. INNER_LOOP 0
  265. jge .loop1
  266. mov j, 448 * 4
  267. sub j, OFFQ
  268. jz .end
  269. sub ptr1, j
  270. sub ptr2, j
  271. add win, OFFQ ; now at j-64, so define OFFSET
  272. sub j, 64 * 4
  273. .loop2:
  274. INNER_LOOP 64 * 4
  275. jge .loop2
  276. .end:
  277. %if ARCH_X86_32
  278. mov buf2, synth_buf2m ; needed for next iteration anyway
  279. mov outq, outmp ; j, which will be set again during it
  280. %endif
  281. ;~ out[i] = a * scale;
  282. ;~ out[i + 16] = b * scale;
  283. mulps m1, m1, scale
  284. mulps m2, m2, scale
  285. %if ARCH_X86_64
  286. mulps m7, m7, scale
  287. mulps m8, m8, scale
  288. %endif
  289. ;~ synth_buf2[i] = c;
  290. ;~ synth_buf2[i + 16] = d;
  291. mova [buf2 + i + 0 * 4], m3
  292. mova [buf2 + i + 16 * 4], m4
  293. %if ARCH_X86_64
  294. mova [buf2 + i + 0 * 4 + mmsize], m9
  295. mova [buf2 + i + 16 * 4 + mmsize], m10
  296. %endif
  297. ;~ out[i] = a;
  298. ;~ out[i + 16] = a;
  299. mova [outq + i + 0 * 4], m1
  300. mova [outq + i + 16 * 4], m2
  301. %if ARCH_X86_64
  302. mova [outq + i + 0 * 4 + mmsize], m7
  303. mova [outq + i + 16 * 4 + mmsize], m8
  304. %endif
  305. %if ARCH_X86_32 || notcpuflag(avx)
  306. sub i, (ARCH_X86_64 + 1) * mmsize
  307. jge .mainloop
  308. %endif
  309. RET
  310. %endmacro
  311. %if ARCH_X86_32
  312. INIT_XMM sse
  313. SYNTH_FILTER
  314. %endif
  315. INIT_XMM sse2
  316. SYNTH_FILTER
  317. INIT_YMM avx
  318. SYNTH_FILTER
  319. INIT_YMM fma3
  320. SYNTH_FILTER