| 
							- ;*****************************************************************************
 - ;* x86-optimized AC-3 DSP utils
 - ;* Copyright (c) 2011 Justin Ruggles
 - ;*
 - ;* This file is part of Libav.
 - ;*
 - ;* Libav is free software; you can redistribute it and/or
 - ;* modify it under the terms of the GNU Lesser General Public
 - ;* License as published by the Free Software Foundation; either
 - ;* version 2.1 of the License, or (at your option) any later version.
 - ;*
 - ;* Libav is distributed in the hope that it will be useful,
 - ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 - ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 - ;* Lesser General Public License for more details.
 - ;*
 - ;* You should have received a copy of the GNU Lesser General Public
 - ;* License along with Libav; if not, write to the Free Software
 - ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 - ;******************************************************************************
 - 
 - %include "x86inc.asm"
 - %include "x86util.asm"
 - 
 - SECTION_RODATA
 - 
 - ; 16777216.0f - used in ff_float_to_fixed24()
 - pf_1_24: times 4 dd 0x4B800000
 - 
 - ; used in ff_ac3_compute_mantissa_size()
 - cextern ac3_bap_bits
 - pw_bap_mul1: dw 21846, 21846, 0, 32768, 21846, 21846, 0, 32768
 - pw_bap_mul2: dw 5, 7, 0, 7, 5, 7, 0, 7
 - 
 - ; used in ff_ac3_extract_exponents()
 - pd_1:   times 4 dd 1
 - pd_151: times 4 dd 151
 - pb_shuf_4dwb: db 0, 4, 8, 12
 - 
 - SECTION .text
 - 
 - ;-----------------------------------------------------------------------------
 - ; void ff_ac3_exponent_min(uint8_t *exp, int num_reuse_blocks, int nb_coefs)
 - ;-----------------------------------------------------------------------------
 - 
 - %macro AC3_EXPONENT_MIN 1
 - cglobal ac3_exponent_min_%1, 3,4,2, exp, reuse_blks, expn, offset
 -     shl  reuse_blksq, 8
 -     jz .end
 -     LOOP_ALIGN
 - .nextexp:
 -     mov      offsetq, reuse_blksq
 -     mova          m0, [expq+offsetq]
 -     sub      offsetq, 256
 -     LOOP_ALIGN
 - .nextblk:
 -     PMINUB        m0, [expq+offsetq], m1
 -     sub      offsetq, 256
 -     jae .nextblk
 -     mova      [expq], m0
 -     add         expq, mmsize
 -     sub        expnq, mmsize
 -     jg .nextexp
 - .end:
 -     REP_RET
 - %endmacro
 - 
 - %define PMINUB PMINUB_MMX
 - %define LOOP_ALIGN
 - INIT_MMX
 - AC3_EXPONENT_MIN mmx
 - %ifdef HAVE_MMX2
 - %define PMINUB PMINUB_MMXEXT
 - %define LOOP_ALIGN ALIGN 16
 - AC3_EXPONENT_MIN mmxext
 - %endif
 - %ifdef HAVE_SSE
 - INIT_XMM
 - AC3_EXPONENT_MIN sse2
 - %endif
 - %undef PMINUB
 - %undef LOOP_ALIGN
 - 
 - ;-----------------------------------------------------------------------------
 - ; int ff_ac3_max_msb_abs_int16(const int16_t *src, int len)
 - ;
 - ; This function uses 2 different methods to calculate a valid result.
 - ; 1) logical 'or' of abs of each element
 - ;        This is used for ssse3 because of the pabsw instruction.
 - ;        It is also used for mmx because of the lack of min/max instructions.
 - ; 2) calculate min/max for the array, then or(abs(min),abs(max))
 - ;        This is used for mmxext and sse2 because they have pminsw/pmaxsw.
 - ;-----------------------------------------------------------------------------
 - 
 - %macro AC3_MAX_MSB_ABS_INT16 2
 - cglobal ac3_max_msb_abs_int16_%1, 2,2,5, src, len
 -     pxor        m2, m2
 -     pxor        m3, m3
 - .loop:
 - %ifidn %2, min_max
 -     mova        m0, [srcq]
 -     mova        m1, [srcq+mmsize]
 -     pminsw      m2, m0
 -     pminsw      m2, m1
 -     pmaxsw      m3, m0
 -     pmaxsw      m3, m1
 - %else ; or_abs
 - %ifidn %1, mmx
 -     mova        m0, [srcq]
 -     mova        m1, [srcq+mmsize]
 -     ABS2        m0, m1, m3, m4
 - %else ; ssse3
 -     ; using memory args is faster for ssse3
 -     pabsw       m0, [srcq]
 -     pabsw       m1, [srcq+mmsize]
 - %endif
 -     por         m2, m0
 -     por         m2, m1
 - %endif
 -     add       srcq, mmsize*2
 -     sub       lend, mmsize
 -     ja .loop
 - %ifidn %2, min_max
 -     ABS2        m2, m3, m0, m1
 -     por         m2, m3
 - %endif
 - %ifidn mmsize, 16
 -     movhlps     m0, m2
 -     por         m2, m0
 - %endif
 -     PSHUFLW     m0, m2, 0xe
 -     por         m2, m0
 -     PSHUFLW     m0, m2, 0x1
 -     por         m2, m0
 -     movd       eax, m2
 -     and        eax, 0xFFFF
 -     RET
 - %endmacro
 - 
 - INIT_MMX
 - %define ABS2 ABS2_MMX
 - %define PSHUFLW pshufw
 - AC3_MAX_MSB_ABS_INT16 mmx, or_abs
 - %define ABS2 ABS2_MMX2
 - AC3_MAX_MSB_ABS_INT16 mmxext, min_max
 - INIT_XMM
 - %define PSHUFLW pshuflw
 - AC3_MAX_MSB_ABS_INT16 sse2, min_max
 - %define ABS2 ABS2_SSSE3
 - AC3_MAX_MSB_ABS_INT16 ssse3, or_abs
 - 
 - ;-----------------------------------------------------------------------------
 - ; macro used for ff_ac3_lshift_int16() and ff_ac3_rshift_int32()
 - ;-----------------------------------------------------------------------------
 - 
 - %macro AC3_SHIFT 4 ; l/r, 16/32, shift instruction, instruction set
 - cglobal ac3_%1shift_int%2_%4, 3,3,5, src, len, shift
 -     movd      m0, shiftd
 - .loop:
 -     mova      m1, [srcq         ]
 -     mova      m2, [srcq+mmsize  ]
 -     mova      m3, [srcq+mmsize*2]
 -     mova      m4, [srcq+mmsize*3]
 -     %3        m1, m0
 -     %3        m2, m0
 -     %3        m3, m0
 -     %3        m4, m0
 -     mova  [srcq         ], m1
 -     mova  [srcq+mmsize  ], m2
 -     mova  [srcq+mmsize*2], m3
 -     mova  [srcq+mmsize*3], m4
 -     add     srcq, mmsize*4
 -     sub     lend, mmsize*32/%2
 -     ja .loop
 - .end:
 -     REP_RET
 - %endmacro
 - 
 - ;-----------------------------------------------------------------------------
 - ; void ff_ac3_lshift_int16(int16_t *src, unsigned int len, unsigned int shift)
 - ;-----------------------------------------------------------------------------
 - 
 - INIT_MMX
 - AC3_SHIFT l, 16, psllw, mmx
 - INIT_XMM
 - AC3_SHIFT l, 16, psllw, sse2
 - 
 - ;-----------------------------------------------------------------------------
 - ; void ff_ac3_rshift_int32(int32_t *src, unsigned int len, unsigned int shift)
 - ;-----------------------------------------------------------------------------
 - 
 - INIT_MMX
 - AC3_SHIFT r, 32, psrad, mmx
 - INIT_XMM
 - AC3_SHIFT r, 32, psrad, sse2
 - 
 - ;-----------------------------------------------------------------------------
 - ; void ff_float_to_fixed24(int32_t *dst, const float *src, unsigned int len)
 - ;-----------------------------------------------------------------------------
 - 
 - ; The 3DNow! version is not bit-identical because pf2id uses truncation rather
 - ; than round-to-nearest.
 - INIT_MMX
 - cglobal float_to_fixed24_3dnow, 3,3,0, dst, src, len
 -     movq   m0, [pf_1_24]
 - .loop:
 -     movq   m1, [srcq   ]
 -     movq   m2, [srcq+8 ]
 -     movq   m3, [srcq+16]
 -     movq   m4, [srcq+24]
 -     pfmul  m1, m0
 -     pfmul  m2, m0
 -     pfmul  m3, m0
 -     pfmul  m4, m0
 -     pf2id  m1, m1
 -     pf2id  m2, m2
 -     pf2id  m3, m3
 -     pf2id  m4, m4
 -     movq  [dstq   ], m1
 -     movq  [dstq+8 ], m2
 -     movq  [dstq+16], m3
 -     movq  [dstq+24], m4
 -     add  srcq, 32
 -     add  dstq, 32
 -     sub  lend, 8
 -     ja .loop
 -     REP_RET
 - 
 - INIT_XMM
 - cglobal float_to_fixed24_sse, 3,3,3, dst, src, len
 -     movaps     m0, [pf_1_24]
 - .loop:
 -     movaps     m1, [srcq   ]
 -     movaps     m2, [srcq+16]
 -     mulps      m1, m0
 -     mulps      m2, m0
 -     cvtps2pi  mm0, m1
 -     movhlps    m1, m1
 -     cvtps2pi  mm1, m1
 -     cvtps2pi  mm2, m2
 -     movhlps    m2, m2
 -     cvtps2pi  mm3, m2
 -     movq  [dstq   ], mm0
 -     movq  [dstq+ 8], mm1
 -     movq  [dstq+16], mm2
 -     movq  [dstq+24], mm3
 -     add      srcq, 32
 -     add      dstq, 32
 -     sub      lend, 8
 -     ja .loop
 -     REP_RET
 - 
 - INIT_XMM
 - cglobal float_to_fixed24_sse2, 3,3,9, dst, src, len
 -     movaps     m0, [pf_1_24]
 - .loop:
 -     movaps     m1, [srcq    ]
 -     movaps     m2, [srcq+16 ]
 -     movaps     m3, [srcq+32 ]
 -     movaps     m4, [srcq+48 ]
 - %ifdef m8
 -     movaps     m5, [srcq+64 ]
 -     movaps     m6, [srcq+80 ]
 -     movaps     m7, [srcq+96 ]
 -     movaps     m8, [srcq+112]
 - %endif
 -     mulps      m1, m0
 -     mulps      m2, m0
 -     mulps      m3, m0
 -     mulps      m4, m0
 - %ifdef m8
 -     mulps      m5, m0
 -     mulps      m6, m0
 -     mulps      m7, m0
 -     mulps      m8, m0
 - %endif
 -     cvtps2dq   m1, m1
 -     cvtps2dq   m2, m2
 -     cvtps2dq   m3, m3
 -     cvtps2dq   m4, m4
 - %ifdef m8
 -     cvtps2dq   m5, m5
 -     cvtps2dq   m6, m6
 -     cvtps2dq   m7, m7
 -     cvtps2dq   m8, m8
 - %endif
 -     movdqa  [dstq    ], m1
 -     movdqa  [dstq+16 ], m2
 -     movdqa  [dstq+32 ], m3
 -     movdqa  [dstq+48 ], m4
 - %ifdef m8
 -     movdqa  [dstq+64 ], m5
 -     movdqa  [dstq+80 ], m6
 -     movdqa  [dstq+96 ], m7
 -     movdqa  [dstq+112], m8
 -     add      srcq, 128
 -     add      dstq, 128
 -     sub      lenq, 32
 - %else
 -     add      srcq, 64
 -     add      dstq, 64
 -     sub      lenq, 16
 - %endif
 -     ja .loop
 -     REP_RET
 - 
 - ;------------------------------------------------------------------------------
 - ; int ff_ac3_compute_mantissa_size(uint16_t mant_cnt[6][16])
 - ;------------------------------------------------------------------------------
 - 
 - %macro PHADDD4 2 ; xmm src, xmm tmp
 -     movhlps  %2, %1
 -     paddd    %1, %2
 -     pshufd   %2, %1, 0x1
 -     paddd    %1, %2
 - %endmacro
 - 
 - INIT_XMM
 - cglobal ac3_compute_mantissa_size_sse2, 1,2,4, mant_cnt, sum
 -     movdqa      m0, [mant_cntq      ]
 -     movdqa      m1, [mant_cntq+ 1*16]
 -     paddw       m0, [mant_cntq+ 2*16]
 -     paddw       m1, [mant_cntq+ 3*16]
 -     paddw       m0, [mant_cntq+ 4*16]
 -     paddw       m1, [mant_cntq+ 5*16]
 -     paddw       m0, [mant_cntq+ 6*16]
 -     paddw       m1, [mant_cntq+ 7*16]
 -     paddw       m0, [mant_cntq+ 8*16]
 -     paddw       m1, [mant_cntq+ 9*16]
 -     paddw       m0, [mant_cntq+10*16]
 -     paddw       m1, [mant_cntq+11*16]
 -     pmaddwd     m0, [ac3_bap_bits   ]
 -     pmaddwd     m1, [ac3_bap_bits+16]
 -     paddd       m0, m1
 -     PHADDD4     m0, m1
 -     movd      sumd, m0
 -     movdqa      m3, [pw_bap_mul1]
 -     movhpd      m0, [mant_cntq     +2]
 -     movlpd      m0, [mant_cntq+1*32+2]
 -     movhpd      m1, [mant_cntq+2*32+2]
 -     movlpd      m1, [mant_cntq+3*32+2]
 -     movhpd      m2, [mant_cntq+4*32+2]
 -     movlpd      m2, [mant_cntq+5*32+2]
 -     pmulhuw     m0, m3
 -     pmulhuw     m1, m3
 -     pmulhuw     m2, m3
 -     paddusw     m0, m1
 -     paddusw     m0, m2
 -     pmaddwd     m0, [pw_bap_mul2]
 -     PHADDD4     m0, m1
 -     movd       eax, m0
 -     add        eax, sumd
 -     RET
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_ac3_extract_exponents(uint8_t *exp, int32_t *coef, int nb_coefs)
 - ;------------------------------------------------------------------------------
 - 
 - %macro PABSD_MMX 2 ; src/dst, tmp
 -     pxor     %2, %2
 -     pcmpgtd  %2, %1
 -     pxor     %1, %2
 -     psubd    %1, %2
 - %endmacro
 - 
 - %macro PABSD_SSSE3 1-2 ; src/dst, unused
 -     pabsd    %1, %1
 - %endmacro
 - 
 - %ifdef HAVE_AMD3DNOW
 - INIT_MMX
 - cglobal ac3_extract_exponents_3dnow, 3,3,0, exp, coef, len
 -     add      expq, lenq
 -     lea     coefq, [coefq+4*lenq]
 -     neg      lenq
 -     movq       m3, [pd_1]
 -     movq       m4, [pd_151]
 - .loop:
 -     movq       m0, [coefq+4*lenq  ]
 -     movq       m1, [coefq+4*lenq+8]
 -     PABSD_MMX  m0, m2
 -     PABSD_MMX  m1, m2
 -     pslld      m0, 1
 -     por        m0, m3
 -     pi2fd      m2, m0
 -     psrld      m2, 23
 -     movq       m0, m4
 -     psubd      m0, m2
 -     pslld      m1, 1
 -     por        m1, m3
 -     pi2fd      m2, m1
 -     psrld      m2, 23
 -     movq       m1, m4
 -     psubd      m1, m2
 -     packssdw   m0, m0
 -     packuswb   m0, m0
 -     packssdw   m1, m1
 -     packuswb   m1, m1
 -     punpcklwd  m0, m1
 -     movd  [expq+lenq], m0
 -     add      lenq, 4
 -     jl .loop
 -     REP_RET
 - %endif
 - 
 - %macro AC3_EXTRACT_EXPONENTS 1
 - cglobal ac3_extract_exponents_%1, 3,3,5, exp, coef, len
 -     add     expq, lenq
 -     lea    coefq, [coefq+4*lenq]
 -     neg     lenq
 -     mova      m2, [pd_1]
 -     mova      m3, [pd_151]
 - %ifidn %1, ssse3 ;
 -     movd      m4, [pb_shuf_4dwb]
 - %endif
 - .loop:
 -     ; move 4 32-bit coefs to xmm0
 -     mova      m0, [coefq+4*lenq]
 -     ; absolute value
 -     PABSD     m0, m1
 -     ; convert to float and extract exponents
 -     pslld     m0, 1
 -     por       m0, m2
 -     cvtdq2ps  m1, m0
 -     psrld     m1, 23
 -     mova      m0, m3
 -     psubd     m0, m1
 -     ; move the lowest byte in each of 4 dwords to the low dword
 - %ifidn %1, ssse3
 -     pshufb    m0, m4
 - %else
 -     packssdw  m0, m0
 -     packuswb  m0, m0
 - %endif
 -     movd  [expq+lenq], m0
 - 
 -     add     lenq, 4
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - %ifdef HAVE_SSE
 - INIT_XMM
 - %define PABSD PABSD_MMX
 - AC3_EXTRACT_EXPONENTS sse2
 - %ifdef HAVE_SSSE3
 - %define PABSD PABSD_SSSE3
 - AC3_EXTRACT_EXPONENTS ssse3
 - %endif
 - %endif
 
 
  |