| 
							- ;******************************************************************************
 - ;* x86 optimized Format Conversion Utils
 - ;* Copyright (c) 2008 Loren Merritt
 - ;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
 - ;*
 - ;* This file is part of FFmpeg.
 - ;*
 - ;* FFmpeg is free software; you can redistribute it and/or
 - ;* modify it under the terms of the GNU Lesser General Public
 - ;* License as published by the Free Software Foundation; either
 - ;* version 2.1 of the License, or (at your option) any later version.
 - ;*
 - ;* FFmpeg is distributed in the hope that it will be useful,
 - ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 - ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 - ;* Lesser General Public License for more details.
 - ;*
 - ;* You should have received a copy of the GNU Lesser General Public
 - ;* License along with FFmpeg; if not, write to the Free Software
 - ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 - ;******************************************************************************
 - 
 - %include "libavutil/x86/x86util.asm"
 - %include "util.asm"
 - 
 - SECTION_RODATA 32
 - 
 - pf_s32_inv_scale: times 8 dd 0x30000000
 - pf_s32_scale:     times 8 dd 0x4f000000
 - pf_s32_clip:      times 8 dd 0x4effffff
 - pf_s16_inv_scale: times 4 dd 0x38000000
 - pf_s16_scale:     times 4 dd 0x47000000
 - pb_shuf_unpack_even:      db -1, -1,  0,  1, -1, -1,  2,  3, -1, -1,  8,  9, -1, -1, 10, 11
 - pb_shuf_unpack_odd:       db -1, -1,  4,  5, -1, -1,  6,  7, -1, -1, 12, 13, -1, -1, 14, 15
 - pb_interleave_words: SHUFFLE_MASK_W  0,  4,  1,  5,  2,  6,  3,  7
 - pb_deinterleave_words: SHUFFLE_MASK_W  0,  2,  4,  6,  1,  3,  5,  7
 - pw_zero_even:     times 4 dw 0x0000, 0xffff
 - 
 - SECTION_TEXT
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - INIT_XMM sse2
 - cglobal conv_s16_to_s32, 3,3,3, dst, src, len
 -     lea      lenq, [2*lend]
 -     lea      dstq, [dstq+2*lenq]
 -     add      srcq, lenq
 -     neg      lenq
 - .loop:
 -     mova       m2, [srcq+lenq]
 -     pxor       m0, m0
 -     pxor       m1, m1
 -     punpcklwd  m0, m2
 -     punpckhwd  m1, m2
 -     mova  [dstq+2*lenq       ], m0
 -     mova  [dstq+2*lenq+mmsize], m1
 -     add      lenq, mmsize
 -     jl .loop
 -     REP_RET
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16_TO_FLT 0
 - cglobal conv_s16_to_flt, 3,3,3, dst, src, len
 -     lea      lenq, [2*lend]
 -     add      srcq, lenq
 -     lea      dstq, [dstq + 2*lenq]
 -     neg      lenq
 -     mova       m2, [pf_s16_inv_scale]
 -     ALIGN 16
 - .loop:
 -     mova       m0, [srcq+lenq]
 -     S16_TO_S32_SX 0, 1
 -     cvtdq2ps   m0, m0
 -     cvtdq2ps   m1, m1
 -     mulps      m0, m2
 -     mulps      m1, m2
 -     mova  [dstq+2*lenq       ], m0
 -     mova  [dstq+2*lenq+mmsize], m1
 -     add      lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16_TO_FLT
 - INIT_XMM sse4
 - CONV_S16_TO_FLT
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S32_TO_S16 0
 - cglobal conv_s32_to_s16, 3,3,4, dst, src, len
 -     lea     lenq, [2*lend]
 -     lea     srcq, [srcq+2*lenq]
 -     add     dstq, lenq
 -     neg     lenq
 - .loop:
 -     mova      m0, [srcq+2*lenq         ]
 -     mova      m1, [srcq+2*lenq+  mmsize]
 -     mova      m2, [srcq+2*lenq+2*mmsize]
 -     mova      m3, [srcq+2*lenq+3*mmsize]
 -     psrad     m0, 16
 -     psrad     m1, 16
 -     psrad     m2, 16
 -     psrad     m3, 16
 -     packssdw  m0, m1
 -     packssdw  m2, m3
 -     mova  [dstq+lenq       ], m0
 -     mova  [dstq+lenq+mmsize], m2
 -     add     lenq, mmsize*2
 -     jl .loop
 - %if mmsize == 8
 -     emms
 -     RET
 - %else
 -     REP_RET
 - %endif
 - %endmacro
 - 
 - INIT_MMX mmx
 - CONV_S32_TO_S16
 - INIT_XMM sse2
 - CONV_S32_TO_S16
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S32_TO_FLT 0
 - cglobal conv_s32_to_flt, 3,3,3, dst, src, len
 -     lea     lenq, [4*lend]
 -     add     srcq, lenq
 -     add     dstq, lenq
 -     neg     lenq
 -     mova      m0, [pf_s32_inv_scale]
 -     ALIGN 16
 - .loop:
 -     cvtdq2ps  m1, [srcq+lenq       ]
 -     cvtdq2ps  m2, [srcq+lenq+mmsize]
 -     mulps     m1, m1, m0
 -     mulps     m2, m2, m0
 -     mova  [dstq+lenq       ], m1
 -     mova  [dstq+lenq+mmsize], m2
 -     add     lenq, mmsize*2
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S32_TO_FLT
 - %if HAVE_AVX_EXTERNAL
 - INIT_YMM avx
 - CONV_S32_TO_FLT
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - INIT_XMM sse2
 - cglobal conv_flt_to_s16, 3,3,5, dst, src, len
 -     lea     lenq, [2*lend]
 -     lea     srcq, [srcq+2*lenq]
 -     add     dstq, lenq
 -     neg     lenq
 -     mova      m4, [pf_s16_scale]
 - .loop:
 -     mova      m0, [srcq+2*lenq         ]
 -     mova      m1, [srcq+2*lenq+1*mmsize]
 -     mova      m2, [srcq+2*lenq+2*mmsize]
 -     mova      m3, [srcq+2*lenq+3*mmsize]
 -     mulps     m0, m4
 -     mulps     m1, m4
 -     mulps     m2, m4
 -     mulps     m3, m4
 -     cvtps2dq  m0, m0
 -     cvtps2dq  m1, m1
 -     cvtps2dq  m2, m2
 -     cvtps2dq  m3, m3
 -     packssdw  m0, m1
 -     packssdw  m2, m3
 -     mova  [dstq+lenq       ], m0
 -     mova  [dstq+lenq+mmsize], m2
 -     add     lenq, mmsize*2
 -     jl .loop
 -     REP_RET
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLT_TO_S32 0
 - cglobal conv_flt_to_s32, 3,3,6, dst, src, len
 -     lea     lenq, [lend*4]
 -     add     srcq, lenq
 -     add     dstq, lenq
 -     neg     lenq
 -     mova      m4, [pf_s32_scale]
 -     mova      m5, [pf_s32_clip]
 - .loop:
 -     mulps     m0, m4, [srcq+lenq         ]
 -     mulps     m1, m4, [srcq+lenq+1*mmsize]
 -     mulps     m2, m4, [srcq+lenq+2*mmsize]
 -     mulps     m3, m4, [srcq+lenq+3*mmsize]
 -     minps     m0, m0, m5
 -     minps     m1, m1, m5
 -     minps     m2, m2, m5
 -     minps     m3, m3, m5
 -     cvtps2dq  m0, m0
 -     cvtps2dq  m1, m1
 -     cvtps2dq  m2, m2
 -     cvtps2dq  m3, m3
 -     mova  [dstq+lenq         ], m0
 -     mova  [dstq+lenq+1*mmsize], m1
 -     mova  [dstq+lenq+2*mmsize], m2
 -     mova  [dstq+lenq+3*mmsize], m3
 -     add     lenq, mmsize*4
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_FLT_TO_S32
 - %if HAVE_AVX_EXTERNAL
 - INIT_YMM avx
 - CONV_FLT_TO_S32
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16p_to_s16_2ch(int16_t *dst, int16_t *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16P_TO_S16_2CH 0
 - cglobal conv_s16p_to_s16_2ch, 3,4,5, dst, src0, len, src1
 -     mov       src1q, [src0q+gprsize]
 -     mov       src0q, [src0q        ]
 -     lea        lenq, [2*lend]
 -     add       src0q, lenq
 -     add       src1q, lenq
 -     lea        dstq, [dstq+2*lenq]
 -     neg        lenq
 - .loop:
 -     mova         m0, [src0q+lenq       ]
 -     mova         m1, [src1q+lenq       ]
 -     mova         m2, [src0q+lenq+mmsize]
 -     mova         m3, [src1q+lenq+mmsize]
 -     SBUTTERFLY2  wd, 0, 1, 4
 -     SBUTTERFLY2  wd, 2, 3, 4
 -     mova  [dstq+2*lenq+0*mmsize], m0
 -     mova  [dstq+2*lenq+1*mmsize], m1
 -     mova  [dstq+2*lenq+2*mmsize], m2
 -     mova  [dstq+2*lenq+3*mmsize], m3
 -     add        lenq, 2*mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16P_TO_S16_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16P_TO_S16_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16p_to_s16_6ch(int16_t *dst, int16_t *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - ;------------------------------------------------------------------------------
 - ; NOTE: In the 6-channel functions, len could be used as an index on x86-64
 - ;       instead of just a counter, which would avoid incrementing the
 - ;       pointers, but the extra complexity and amount of code is not worth
 - ;       the small gain. On x86-32 there are not enough registers to use len
 - ;       as an index without keeping two of the pointers on the stack and
 - ;       loading them in each iteration.
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16P_TO_S16_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_s16p_to_s16_6ch, 3,8,7, dst, src0, len, src1, src2, src3, src4, src5
 - %else
 - cglobal conv_s16p_to_s16_6ch, 2,7,7, dst, src0, src1, src2, src3, src4, src5
 - %define lend dword r2m
 - %endif
 -     mov      src1q, [src0q+1*gprsize]
 -     mov      src2q, [src0q+2*gprsize]
 -     mov      src3q, [src0q+3*gprsize]
 -     mov      src4q, [src0q+4*gprsize]
 -     mov      src5q, [src0q+5*gprsize]
 -     mov      src0q, [src0q]
 -     sub      src1q, src0q
 -     sub      src2q, src0q
 -     sub      src3q, src0q
 -     sub      src4q, src0q
 -     sub      src5q, src0q
 - .loop:
 - %if cpuflag(sse2slow)
 -     movq        m0, [src0q      ]   ; m0 =  0,  6, 12, 18,  x,  x,  x,  x
 -     movq        m1, [src0q+src1q]   ; m1 =  1,  7, 13, 19,  x,  x,  x,  x
 -     movq        m2, [src0q+src2q]   ; m2 =  2,  8, 14, 20,  x,  x,  x,  x
 -     movq        m3, [src0q+src3q]   ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 -     movq        m4, [src0q+src4q]   ; m4 =  4, 10, 16, 22,  x,  x,  x,  x
 -     movq        m5, [src0q+src5q]   ; m5 =  5, 11, 17, 23,  x,  x,  x,  x
 -                                     ; unpack words:
 -     punpcklwd   m0, m1              ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 -     punpcklwd   m2, m3              ; m2 =  4,  5, 10, 11, 16, 17, 22, 23
 -     punpcklwd   m4, m5              ; m4 =  2,  3,  8,  9, 14, 15, 20, 21
 -                                     ; blend dwords
 -     shufps      m1, m0, m2, q2020   ; m1 =  0,  1, 12, 13,  2,  3, 14, 15
 -     shufps      m0, m4, q2031       ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 -     shufps      m2, m4, q3131       ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 -                                     ; shuffle dwords
 -     pshufd      m0, m0, q1302       ; m0 =  4,  5,  6,  7, 16, 17, 18, 19
 -     pshufd      m1, m1, q3120       ; m1 =  0,  1,  2,  3, 12, 13, 14, 15
 -     pshufd      m2, m2, q3120       ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 -     movq   [dstq+0*mmsize/2], m1
 -     movq   [dstq+1*mmsize/2], m0
 -     movq   [dstq+2*mmsize/2], m2
 -     movhps [dstq+3*mmsize/2], m1
 -     movhps [dstq+4*mmsize/2], m0
 -     movhps [dstq+5*mmsize/2], m2
 -     add      src0q, mmsize/2
 -     add       dstq, mmsize*3
 -     sub       lend, mmsize/4
 - %else
 -     mova        m0, [src0q      ]   ; m0 =  0,  6, 12, 18, 24, 30, 36, 42
 -     mova        m1, [src0q+src1q]   ; m1 =  1,  7, 13, 19, 25, 31, 37, 43
 -     mova        m2, [src0q+src2q]   ; m2 =  2,  8, 14, 20, 26, 32, 38, 44
 -     mova        m3, [src0q+src3q]   ; m3 =  3,  9, 15, 21, 27, 33, 39, 45
 -     mova        m4, [src0q+src4q]   ; m4 =  4, 10, 16, 22, 28, 34, 40, 46
 -     mova        m5, [src0q+src5q]   ; m5 =  5, 11, 17, 23, 29, 35, 41, 47
 -                                     ; unpack words:
 -     SBUTTERFLY2 wd, 0, 1, 6         ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 -                                     ; m1 = 24, 25, 30, 31, 36, 37, 42, 43
 -     SBUTTERFLY2 wd, 2, 3, 6         ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 -                                     ; m3 = 26, 27, 32, 33, 38, 39, 44, 45
 -     SBUTTERFLY2 wd, 4, 5, 6         ; m4 =  4,  5, 10, 11, 16, 17, 22, 23
 -                                     ; m5 = 28, 29, 34, 35, 40, 41, 46, 47
 -                                     ; blend dwords
 -     shufps      m6, m0, m2, q2020   ; m6 =  0,  1, 12, 13,  2,  3, 14, 15
 -     shufps      m0, m4, q2031       ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 -     shufps      m2, m4, q3131       ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 -     SWAP 4,6                        ; m4 =  0,  1, 12, 13,  2,  3, 14, 15
 -     shufps      m6, m1, m3, q2020   ; m6 = 24, 25, 36, 37, 26, 27, 38, 39
 -     shufps      m1, m5, q2031       ; m1 = 30, 31, 42, 43, 28, 29, 40, 41
 -     shufps      m3, m5, q3131       ; m3 = 32, 33, 44, 45, 34, 35, 46, 47
 -     SWAP 5,6                        ; m5 = 24, 25, 36, 37, 26, 27, 38, 39
 -                                     ; shuffle dwords
 -     pshufd      m0, m0, q1302       ; m0 =  4,  5,  6,  7, 16, 17, 18, 19
 -     pshufd      m2, m2, q3120       ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 -     pshufd      m4, m4, q3120       ; m4 =  0,  1,  2,  3, 12, 13, 14, 15
 -     pshufd      m1, m1, q1302       ; m1 = 28, 29, 30, 31, 40, 41, 42, 43
 -     pshufd      m3, m3, q3120       ; m3 = 32, 33, 34, 35, 44, 45, 46, 47
 -     pshufd      m5, m5, q3120       ; m5 = 24, 25, 26, 27, 36, 37, 38, 39
 -                                     ; shuffle qwords
 -     punpcklqdq  m6, m4, m0          ; m6 =  0,  1,  2,  3,  4,  5,  6,  7
 -     punpckhqdq  m0, m2              ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
 -     shufps      m2, m4, q3210       ; m2 =  8,  9, 10, 11, 12, 13, 14, 15
 -     SWAP 4,6                        ; m4 =  0,  1,  2,  3,  4,  5,  6,  7
 -     punpcklqdq  m6, m5, m1          ; m6 = 24, 25, 26, 27, 28, 29, 30, 31
 -     punpckhqdq  m1, m3              ; m1 = 40, 41, 42, 43, 44, 45, 46, 47
 -     shufps      m3, m5, q3210       ; m3 = 32, 33, 34, 35, 36, 37, 38, 39
 -     SWAP 5,6                        ; m5 = 24, 25, 26, 27, 28, 29, 30, 31
 -     mova   [dstq+0*mmsize], m4
 -     mova   [dstq+1*mmsize], m2
 -     mova   [dstq+2*mmsize], m0
 -     mova   [dstq+3*mmsize], m5
 -     mova   [dstq+4*mmsize], m3
 -     mova   [dstq+5*mmsize], m1
 -     add      src0q, mmsize
 -     add       dstq, mmsize*6
 -     sub       lend, mmsize/2
 - %endif
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16P_TO_S16_6CH
 - INIT_XMM sse2slow
 - CONV_S16P_TO_S16_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16P_TO_S16_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16p_to_flt_2ch(float *dst, int16_t *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16P_TO_FLT_2CH 0
 - cglobal conv_s16p_to_flt_2ch, 3,4,6, dst, src0, len, src1
 -     lea       lenq, [2*lend]
 -     mov      src1q, [src0q+gprsize]
 -     mov      src0q, [src0q        ]
 -     lea       dstq, [dstq+4*lenq]
 -     add      src0q, lenq
 -     add      src1q, lenq
 -     neg       lenq
 -     mova        m5, [pf_s32_inv_scale]
 - .loop:
 -     mova        m2, [src0q+lenq]    ; m2 =  0,  2,  4,  6,  8, 10, 12, 14
 -     mova        m4, [src1q+lenq]    ; m4 =  1,  3,  5,  7,  9, 11, 13, 15
 -     SBUTTERFLY2 wd, 2, 4, 3         ; m2 =  0,  1,  2,  3,  4,  5,  6,  7
 -                                     ; m4 =  8,  9, 10, 11, 12, 13, 14, 15
 -     pxor        m3, m3
 -     punpcklwd   m0, m3, m2          ; m0 =      0,      1,      2,      3
 -     punpckhwd   m1, m3, m2          ; m1 =      4,      5,      6,      7
 -     punpcklwd   m2, m3, m4          ; m2 =      8,      9,     10,     11
 -     punpckhwd   m3, m4              ; m3 =     12,     13,     14,     15
 -     cvtdq2ps    m0, m0
 -     cvtdq2ps    m1, m1
 -     cvtdq2ps    m2, m2
 -     cvtdq2ps    m3, m3
 -     mulps       m0, m5
 -     mulps       m1, m5
 -     mulps       m2, m5
 -     mulps       m3, m5
 -     mova  [dstq+4*lenq         ], m0
 -     mova  [dstq+4*lenq+  mmsize], m1
 -     mova  [dstq+4*lenq+2*mmsize], m2
 -     mova  [dstq+4*lenq+3*mmsize], m3
 -     add       lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16P_TO_FLT_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16P_TO_FLT_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16p_to_flt_6ch(float *dst, int16_t *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16P_TO_FLT_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_s16p_to_flt_6ch, 3,8,8, dst, src, len, src1, src2, src3, src4, src5
 - %else
 - cglobal conv_s16p_to_flt_6ch, 2,7,8, dst, src, src1, src2, src3, src4, src5
 - %define lend dword r2m
 - %endif
 -     mov     src1q, [srcq+1*gprsize]
 -     mov     src2q, [srcq+2*gprsize]
 -     mov     src3q, [srcq+3*gprsize]
 -     mov     src4q, [srcq+4*gprsize]
 -     mov     src5q, [srcq+5*gprsize]
 -     mov      srcq, [srcq]
 -     sub     src1q, srcq
 -     sub     src2q, srcq
 -     sub     src3q, srcq
 -     sub     src4q, srcq
 -     sub     src5q, srcq
 -     mova       m7, [pf_s32_inv_scale]
 - %if cpuflag(ssse3)
 -     %define unpack_even m6
 -     mova       m6, [pb_shuf_unpack_even]
 - %if ARCH_X86_64
 -     %define unpack_odd m8
 -     mova       m8, [pb_shuf_unpack_odd]
 - %else
 -     %define unpack_odd [pb_shuf_unpack_odd]
 - %endif
 - %endif
 - .loop:
 -     movq       m0, [srcq      ]  ; m0 =  0,  6, 12, 18,  x,  x,  x,  x
 -     movq       m1, [srcq+src1q]  ; m1 =  1,  7, 13, 19,  x,  x,  x,  x
 -     movq       m2, [srcq+src2q]  ; m2 =  2,  8, 14, 20,  x,  x,  x,  x
 -     movq       m3, [srcq+src3q]  ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 -     movq       m4, [srcq+src4q]  ; m4 =  4, 10, 16, 22,  x,  x,  x,  x
 -     movq       m5, [srcq+src5q]  ; m5 =  5, 11, 17, 23,  x,  x,  x,  x
 -                                  ; unpack words:
 -     punpcklwd  m0, m1            ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 -     punpcklwd  m2, m3            ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 -     punpcklwd  m4, m5            ; m4 =  4,  5, 10, 11, 16, 17, 22, 23
 -                                  ; blend dwords
 -     shufps     m1, m4, m0, q3120 ; m1 =  4,  5, 16, 17,  6,  7, 18, 19
 -     shufps         m0, m2, q2020 ; m0 =  0,  1, 12, 13,  2,  3, 14, 15
 -     shufps         m2, m4, q3131 ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 - %if cpuflag(ssse3)
 -     pshufb     m3, m0, unpack_odd   ; m3 =  12,     13,     14,     15
 -     pshufb         m0, unpack_even  ; m0 =   0,      1,      2,      3
 -     pshufb     m4, m1, unpack_odd   ; m4 =  16,     17,     18,     19
 -     pshufb         m1, unpack_even  ; m1 =   4,      5,      6,      7
 -     pshufb     m5, m2, unpack_odd   ; m5 =  20,     21,     22,     23
 -     pshufb         m2, unpack_even  ; m2 =   8,      9,     10,     11
 - %else
 -                                  ; shuffle dwords
 -     pshufd     m0, m0, q3120     ; m0 =  0,  1,  2,  3, 12, 13, 14, 15
 -     pshufd     m1, m1, q3120     ; m1 =  4,  5,  6,  7, 16, 17, 18, 19
 -     pshufd     m2, m2, q3120     ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 -     pxor       m6, m6            ; convert s16 in m0-m2 to s32 in m0-m5
 -     punpcklwd  m3, m6, m0        ; m3 =      0,      1,      2,      3
 -     punpckhwd  m4, m6, m0        ; m4 =     12,     13,     14,     15
 -     punpcklwd  m0, m6, m1        ; m0 =      4,      5,      6,      7
 -     punpckhwd  m5, m6, m1        ; m5 =     16,     17,     18,     19
 -     punpcklwd  m1, m6, m2        ; m1 =      8,      9,     10,     11
 -     punpckhwd      m6, m2        ; m6 =     20,     21,     22,     23
 -     SWAP 6,2,1,0,3,4,5           ; swap registers 3,0,1,4,5,6 to 0,1,2,3,4,5
 - %endif
 -     cvtdq2ps   m0, m0            ; convert s32 to float
 -     cvtdq2ps   m1, m1
 -     cvtdq2ps   m2, m2
 -     cvtdq2ps   m3, m3
 -     cvtdq2ps   m4, m4
 -     cvtdq2ps   m5, m5
 -     mulps      m0, m7            ; scale float from s32 range to [-1.0,1.0]
 -     mulps      m1, m7
 -     mulps      m2, m7
 -     mulps      m3, m7
 -     mulps      m4, m7
 -     mulps      m5, m7
 -     mova  [dstq         ], m0
 -     mova  [dstq+  mmsize], m1
 -     mova  [dstq+2*mmsize], m2
 -     mova  [dstq+3*mmsize], m3
 -     mova  [dstq+4*mmsize], m4
 -     mova  [dstq+5*mmsize], m5
 -     add      srcq, mmsize/2
 -     add      dstq, mmsize*6
 -     sub      lend, mmsize/4
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16P_TO_FLT_6CH
 - INIT_XMM ssse3
 - CONV_S16P_TO_FLT_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16P_TO_FLT_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_fltp_to_s16_2ch(int16_t *dst, float *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLTP_TO_S16_2CH 0
 - cglobal conv_fltp_to_s16_2ch, 3,4,3, dst, src0, len, src1
 -     lea      lenq, [4*lend]
 -     mov     src1q, [src0q+gprsize]
 -     mov     src0q, [src0q        ]
 -     add      dstq, lenq
 -     add     src0q, lenq
 -     add     src1q, lenq
 -     neg      lenq
 -     mova       m2, [pf_s16_scale]
 - %if cpuflag(ssse3)
 -     mova       m3, [pb_interleave_words]
 - %endif
 - .loop:
 -     mulps      m0, m2, [src0q+lenq] ; m0 =    0,    2,    4,    6
 -     mulps      m1, m2, [src1q+lenq] ; m1 =    1,    3,    5,    7
 -     cvtps2dq   m0, m0
 -     cvtps2dq   m1, m1
 - %if cpuflag(ssse3)
 -     packssdw   m0, m1               ; m0 = 0, 2, 4, 6, 1, 3, 5, 7
 -     pshufb     m0, m3               ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
 - %else
 -     packssdw   m0, m0               ; m0 = 0, 2, 4, 6, x, x, x, x
 -     packssdw   m1, m1               ; m1 = 1, 3, 5, 7, x, x, x, x
 -     punpcklwd  m0, m1               ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
 - %endif
 -     mova  [dstq+lenq], m0
 -     add      lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_FLTP_TO_S16_2CH
 - INIT_XMM ssse3
 - CONV_FLTP_TO_S16_2CH
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_fltp_to_s16_6ch(int16_t *dst, float *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLTP_TO_S16_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_fltp_to_s16_6ch, 3,8,7, dst, src, len, src1, src2, src3, src4, src5
 - %else
 - cglobal conv_fltp_to_s16_6ch, 2,7,7, dst, src, src1, src2, src3, src4, src5
 - %define lend dword r2m
 - %endif
 -     mov        src1q, [srcq+1*gprsize]
 -     mov        src2q, [srcq+2*gprsize]
 -     mov        src3q, [srcq+3*gprsize]
 -     mov        src4q, [srcq+4*gprsize]
 -     mov        src5q, [srcq+5*gprsize]
 -     mov         srcq, [srcq]
 -     sub        src1q, srcq
 -     sub        src2q, srcq
 -     sub        src3q, srcq
 -     sub        src4q, srcq
 -     sub        src5q, srcq
 -     movaps      xmm6, [pf_s16_scale]
 - .loop:
 - %if cpuflag(sse2)
 -     mulps         m0, m6, [srcq      ]
 -     mulps         m1, m6, [srcq+src1q]
 -     mulps         m2, m6, [srcq+src2q]
 -     mulps         m3, m6, [srcq+src3q]
 -     mulps         m4, m6, [srcq+src4q]
 -     mulps         m5, m6, [srcq+src5q]
 -     cvtps2dq      m0, m0
 -     cvtps2dq      m1, m1
 -     cvtps2dq      m2, m2
 -     cvtps2dq      m3, m3
 -     cvtps2dq      m4, m4
 -     cvtps2dq      m5, m5
 -     packssdw      m0, m3            ; m0 =  0,  6, 12, 18,  3,  9, 15, 21
 -     packssdw      m1, m4            ; m1 =  1,  7, 13, 19,  4, 10, 16, 22
 -     packssdw      m2, m5            ; m2 =  2,  8, 14, 20,  5, 11, 17, 23
 -                                     ; unpack words:
 -     movhlps       m3, m0            ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 -     punpcklwd     m0, m1            ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 -     punpckhwd     m1, m2            ; m1 =  4,  5, 10, 11, 16, 17, 22, 23
 -     punpcklwd     m2, m3            ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 -                                     ; blend dwords:
 -     shufps        m3, m0, m2, q2020 ; m3 =  0,  1, 12, 13,  2,  3, 14, 15
 -     shufps        m0, m1, q2031     ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 -     shufps        m2, m1, q3131     ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 -                                     ; shuffle dwords:
 -     shufps        m1, m2, m3, q3120 ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 -     shufps        m3, m0,     q0220 ; m3 =  0,  1,  2,  3,  4,  5,  6,  7
 -     shufps        m0, m2,     q3113 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
 -     mova  [dstq+0*mmsize], m3
 -     mova  [dstq+1*mmsize], m1
 -     mova  [dstq+2*mmsize], m0
 - %else ; sse
 -     movlps      xmm0, [srcq      ]
 -     movlps      xmm1, [srcq+src1q]
 -     movlps      xmm2, [srcq+src2q]
 -     movlps      xmm3, [srcq+src3q]
 -     movlps      xmm4, [srcq+src4q]
 -     movlps      xmm5, [srcq+src5q]
 -     mulps       xmm0, xmm6
 -     mulps       xmm1, xmm6
 -     mulps       xmm2, xmm6
 -     mulps       xmm3, xmm6
 -     mulps       xmm4, xmm6
 -     mulps       xmm5, xmm6
 -     cvtps2pi     mm0, xmm0
 -     cvtps2pi     mm1, xmm1
 -     cvtps2pi     mm2, xmm2
 -     cvtps2pi     mm3, xmm3
 -     cvtps2pi     mm4, xmm4
 -     cvtps2pi     mm5, xmm5
 -     packssdw     mm0, mm3           ; m0 =  0,  6,  3,  9
 -     packssdw     mm1, mm4           ; m1 =  1,  7,  4, 10
 -     packssdw     mm2, mm5           ; m2 =  2,  8,  5, 11
 -                                     ; unpack words
 -     pshufw       mm3, mm0, q1032    ; m3 =  3,  9,  0,  6
 -     punpcklwd    mm0, mm1           ; m0 =  0,  1,  6,  7
 -     punpckhwd    mm1, mm2           ; m1 =  4,  5, 10, 11
 -     punpcklwd    mm2, mm3           ; m2 =  2,  3,  8,  9
 -                                     ; unpack dwords
 -     pshufw       mm3, mm0, q1032    ; m3 =  6,  7,  0,  1
 -     punpckldq    mm0, mm2           ; m0 =  0,  1,  2,  3 (final)
 -     punpckhdq    mm2, mm1           ; m2 =  8,  9, 10, 11 (final)
 -     punpckldq    mm1, mm3           ; m1 =  4,  5,  6,  7 (final)
 -     mova  [dstq+0*mmsize], mm0
 -     mova  [dstq+1*mmsize], mm1
 -     mova  [dstq+2*mmsize], mm2
 - %endif
 -     add       srcq, mmsize
 -     add       dstq, mmsize*3
 -     sub       lend, mmsize/4
 -     jg .loop
 - %if mmsize == 8
 -     emms
 -     RET
 - %else
 -     REP_RET
 - %endif
 - %endmacro
 - 
 - INIT_MMX sse
 - CONV_FLTP_TO_S16_6CH
 - INIT_XMM sse2
 - CONV_FLTP_TO_S16_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLTP_TO_S16_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_fltp_to_flt_2ch(float *dst, float *const *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLTP_TO_FLT_2CH 0
 - cglobal conv_fltp_to_flt_2ch, 3,4,5, dst, src0, len, src1
 -     mov  src1q, [src0q+gprsize]
 -     mov  src0q, [src0q]
 -     lea   lenq, [4*lend]
 -     add  src0q, lenq
 -     add  src1q, lenq
 -     lea   dstq, [dstq+2*lenq]
 -     neg   lenq
 - .loop:
 -     mova    m0, [src0q+lenq       ]
 -     mova    m1, [src1q+lenq       ]
 -     mova    m2, [src0q+lenq+mmsize]
 -     mova    m3, [src1q+lenq+mmsize]
 -     SBUTTERFLYPS 0, 1, 4
 -     SBUTTERFLYPS 2, 3, 4
 -     mova  [dstq+2*lenq+0*mmsize], m0
 -     mova  [dstq+2*lenq+1*mmsize], m1
 -     mova  [dstq+2*lenq+2*mmsize], m2
 -     mova  [dstq+2*lenq+3*mmsize], m3
 -     add   lenq, 2*mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse
 - CONV_FLTP_TO_FLT_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLTP_TO_FLT_2CH
 - %endif
 - 
 - ;-----------------------------------------------------------------------------
 - ; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len,
 - ;                              int channels);
 - ;-----------------------------------------------------------------------------
 - 
 - %macro CONV_FLTP_TO_FLT_6CH 0
 - cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len
 - %if ARCH_X86_64
 -     mov     lend, r2d
 - %else
 -     %define lend dword r2m
 - %endif
 -     mov    src1q, [srcq+1*gprsize]
 -     mov    src2q, [srcq+2*gprsize]
 -     mov    src3q, [srcq+3*gprsize]
 -     mov    src4q, [srcq+4*gprsize]
 -     mov    src5q, [srcq+5*gprsize]
 -     mov     srcq, [srcq]
 -     sub    src1q, srcq
 -     sub    src2q, srcq
 -     sub    src3q, srcq
 -     sub    src4q, srcq
 -     sub    src5q, srcq
 - .loop:
 -     mova      m0, [srcq      ]
 -     mova      m1, [srcq+src1q]
 -     mova      m2, [srcq+src2q]
 -     mova      m3, [srcq+src3q]
 -     mova      m4, [srcq+src4q]
 -     mova      m5, [srcq+src5q]
 - %if cpuflag(sse4)
 -     SBUTTERFLYPS 0, 1, 6
 -     SBUTTERFLYPS 2, 3, 6
 -     SBUTTERFLYPS 4, 5, 6
 - 
 -     blendps   m6, m4, m0, 1100b
 -     movlhps   m0, m2
 -     movhlps   m4, m2
 -     blendps   m2, m5, m1, 1100b
 -     movlhps   m1, m3
 -     movhlps   m5, m3
 - 
 -     movaps [dstq   ], m0
 -     movaps [dstq+16], m6
 -     movaps [dstq+32], m4
 -     movaps [dstq+48], m1
 -     movaps [dstq+64], m2
 -     movaps [dstq+80], m5
 - %else ; mmx
 -     SBUTTERFLY dq, 0, 1, 6
 -     SBUTTERFLY dq, 2, 3, 6
 -     SBUTTERFLY dq, 4, 5, 6
 - 
 -     movq   [dstq   ], m0
 -     movq   [dstq+ 8], m2
 -     movq   [dstq+16], m4
 -     movq   [dstq+24], m1
 -     movq   [dstq+32], m3
 -     movq   [dstq+40], m5
 - %endif
 -     add      srcq, mmsize
 -     add      dstq, mmsize*6
 -     sub      lend, mmsize/4
 -     jg .loop
 - %if mmsize == 8
 -     emms
 -     RET
 - %else
 -     REP_RET
 - %endif
 - %endmacro
 - 
 - INIT_MMX mmx
 - CONV_FLTP_TO_FLT_6CH
 - INIT_XMM sse4
 - CONV_FLTP_TO_FLT_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLTP_TO_FLT_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_s16p_2ch(int16_t *const *dst, int16_t *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16_TO_S16P_2CH 0
 - cglobal conv_s16_to_s16p_2ch, 3,4,4, dst0, src, len, dst1
 -     lea       lenq, [2*lend]
 -     mov      dst1q, [dst0q+gprsize]
 -     mov      dst0q, [dst0q        ]
 -     lea       srcq, [srcq+2*lenq]
 -     add      dst0q, lenq
 -     add      dst1q, lenq
 -     neg       lenq
 - %if cpuflag(ssse3)
 -     mova        m3, [pb_deinterleave_words]
 - %endif
 - .loop:
 -     mova        m0, [srcq+2*lenq       ]  ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 -     mova        m1, [srcq+2*lenq+mmsize]  ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 - %if cpuflag(ssse3)
 -     pshufb      m0, m3                    ; m0 =  0,  2,  4,  6,  1,  3,  5,  7
 -     pshufb      m1, m3                    ; m1 =  8, 10, 12, 14,  9, 11, 13, 15
 -     SBUTTERFLY2 qdq, 0, 1, 2              ; m0 =  0,  2,  4,  6,  8, 10, 12, 14
 -                                           ; m1 =  1,  3,  5,  7,  9, 11, 13, 15
 - %else ; sse2
 -     pshuflw     m0, m0, q3120             ; m0 =  0,  2,  1,  3,  4,  5,  6,  7
 -     pshufhw     m0, m0, q3120             ; m0 =  0,  2,  1,  3,  4,  6,  5,  7
 -     pshuflw     m1, m1, q3120             ; m1 =  8, 10,  9, 11, 12, 13, 14, 15
 -     pshufhw     m1, m1, q3120             ; m1 =  8, 10,  9, 11, 12, 14, 13, 15
 -     DEINT2_PS    0, 1, 2                  ; m0 =  0,  2,  4,  6,  8, 10, 12, 14
 -                                           ; m1 =  1,  3,  5,  7,  9, 11, 13, 15
 - %endif
 -     mova  [dst0q+lenq], m0
 -     mova  [dst1q+lenq], m1
 -     add       lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16_TO_S16P_2CH
 - INIT_XMM ssse3
 - CONV_S16_TO_S16P_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16_TO_S16P_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_s16p_6ch(int16_t *const *dst, int16_t *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16_TO_S16P_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_s16_to_s16p_6ch, 3,8,5, dst, src, len, dst1, dst2, dst3, dst4, dst5
 - %else
 - cglobal conv_s16_to_s16p_6ch, 2,7,5, dst, src, dst1, dst2, dst3, dst4, dst5
 - %define lend dword r2m
 - %endif
 -     mov     dst1q, [dstq+  gprsize]
 -     mov     dst2q, [dstq+2*gprsize]
 -     mov     dst3q, [dstq+3*gprsize]
 -     mov     dst4q, [dstq+4*gprsize]
 -     mov     dst5q, [dstq+5*gprsize]
 -     mov      dstq, [dstq          ]
 -     sub     dst1q, dstq
 -     sub     dst2q, dstq
 -     sub     dst3q, dstq
 -     sub     dst4q, dstq
 -     sub     dst5q, dstq
 - .loop:
 -     mova       m0, [srcq+0*mmsize]      ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 -     mova       m3, [srcq+1*mmsize]      ; m3 =  8,  9, 10, 11, 12, 13, 14, 15
 -     mova       m2, [srcq+2*mmsize]      ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 -     PALIGNR    m1, m3, m0, 12, m4       ; m1 =  6,  7,  8,  9, 10, 11,  x,  x
 -     shufps     m3, m2, q1032            ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
 -     psrldq     m2, 4                    ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 -     SBUTTERFLY2 wd, 0, 1, 4             ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 -                                         ; m1 =  4, 10,  5, 11,  x,  x,  x,  x
 -     SBUTTERFLY2 wd, 3, 2, 4             ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
 -                                         ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 -     SBUTTERFLY2 dq, 0, 3, 4             ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 -                                         ; m3 =  2,  8, 14, 20,  3,  9, 15, 21
 -     punpckldq  m1, m2                   ; m1 =  4, 10, 16, 22,  5, 11, 17, 23
 -     movq    [dstq      ], m0
 -     movhps  [dstq+dst1q], m0
 -     movq    [dstq+dst2q], m3
 -     movhps  [dstq+dst3q], m3
 -     movq    [dstq+dst4q], m1
 -     movhps  [dstq+dst5q], m1
 -     add      srcq, mmsize*3
 -     add      dstq, mmsize/2
 -     sub      lend, mmsize/4
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16_TO_S16P_6CH
 - INIT_XMM ssse3
 - CONV_S16_TO_S16P_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16_TO_S16P_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_fltp_2ch(float *const *dst, int16_t *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16_TO_FLTP_2CH 0
 - cglobal conv_s16_to_fltp_2ch, 3,4,5, dst0, src, len, dst1
 -     lea       lenq, [4*lend]
 -     mov      dst1q, [dst0q+gprsize]
 -     mov      dst0q, [dst0q        ]
 -     add       srcq, lenq
 -     add      dst0q, lenq
 -     add      dst1q, lenq
 -     neg       lenq
 -     mova        m3, [pf_s32_inv_scale]
 -     mova        m4, [pw_zero_even]
 - .loop:
 -     mova        m1, [srcq+lenq]
 -     pslld       m0, m1, 16
 -     pand        m1, m4
 -     cvtdq2ps    m0, m0
 -     cvtdq2ps    m1, m1
 -     mulps       m0, m0, m3
 -     mulps       m1, m1, m3
 -     mova  [dst0q+lenq], m0
 -     mova  [dst1q+lenq], m1
 -     add       lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16_TO_FLTP_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16_TO_FLTP_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_s16_to_fltp_6ch(float *const *dst, int16_t *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_S16_TO_FLTP_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_s16_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 - %else
 - cglobal conv_s16_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 - %define lend dword r2m
 - %endif
 -     mov     dst1q, [dstq+  gprsize]
 -     mov     dst2q, [dstq+2*gprsize]
 -     mov     dst3q, [dstq+3*gprsize]
 -     mov     dst4q, [dstq+4*gprsize]
 -     mov     dst5q, [dstq+5*gprsize]
 -     mov      dstq, [dstq          ]
 -     sub     dst1q, dstq
 -     sub     dst2q, dstq
 -     sub     dst3q, dstq
 -     sub     dst4q, dstq
 -     sub     dst5q, dstq
 -     mova       m6, [pf_s16_inv_scale]
 - .loop:
 -     mova       m0, [srcq+0*mmsize]  ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 -     mova       m3, [srcq+1*mmsize]  ; m3 =  8,  9, 10, 11, 12, 13, 14, 15
 -     mova       m2, [srcq+2*mmsize]  ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 -     PALIGNR    m1, m3, m0, 12, m4   ; m1 =  6,  7,  8,  9, 10, 11,  x,  x
 -     shufps     m3, m2, q1032        ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
 -     psrldq     m2, 4                ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 -     SBUTTERFLY2 wd, 0, 1, 4         ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 -                                     ; m1 =  4, 10,  5, 11,  x,  x,  x,  x
 -     SBUTTERFLY2 wd, 3, 2, 4         ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
 -                                     ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 -     SBUTTERFLY2 dq, 0, 3, 4         ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 -                                     ; m3 =  2,  8, 14, 20,  3,  9, 15, 21
 -     punpckldq  m1, m2               ; m1 =  4, 10, 16, 22,  5, 11, 17, 23
 -     S16_TO_S32_SX 0, 2              ; m0 =      0,      6,     12,     18
 -                                     ; m2 =      1,      7,     13,     19
 -     S16_TO_S32_SX 3, 4              ; m3 =      2,      8,     14,     20
 -                                     ; m4 =      3,      9,     15,     21
 -     S16_TO_S32_SX 1, 5              ; m1 =      4,     10,     16,     22
 -                                     ; m5 =      5,     11,     17,     23
 -     SWAP 1,2,3,4
 -     cvtdq2ps   m0, m0
 -     cvtdq2ps   m1, m1
 -     cvtdq2ps   m2, m2
 -     cvtdq2ps   m3, m3
 -     cvtdq2ps   m4, m4
 -     cvtdq2ps   m5, m5
 -     mulps      m0, m6
 -     mulps      m1, m6
 -     mulps      m2, m6
 -     mulps      m3, m6
 -     mulps      m4, m6
 -     mulps      m5, m6
 -     mova  [dstq      ], m0
 -     mova  [dstq+dst1q], m1
 -     mova  [dstq+dst2q], m2
 -     mova  [dstq+dst3q], m3
 -     mova  [dstq+dst4q], m4
 -     mova  [dstq+dst5q], m5
 -     add      srcq, mmsize*3
 -     add      dstq, mmsize
 -     sub      lend, mmsize/4
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_S16_TO_FLTP_6CH
 - INIT_XMM ssse3
 - CONV_S16_TO_FLTP_6CH
 - INIT_XMM sse4
 - CONV_S16_TO_FLTP_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_S16_TO_FLTP_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_s16p_2ch(int16_t *const *dst, float *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLT_TO_S16P_2CH 0
 - cglobal conv_flt_to_s16p_2ch, 3,4,6, dst0, src, len, dst1
 -     lea       lenq, [2*lend]
 -     mov      dst1q, [dst0q+gprsize]
 -     mov      dst0q, [dst0q        ]
 -     lea       srcq, [srcq+4*lenq]
 -     add      dst0q, lenq
 -     add      dst1q, lenq
 -     neg       lenq
 -     mova        m5, [pf_s16_scale]
 - .loop:
 -     mova       m0, [srcq+4*lenq         ]
 -     mova       m1, [srcq+4*lenq+  mmsize]
 -     mova       m2, [srcq+4*lenq+2*mmsize]
 -     mova       m3, [srcq+4*lenq+3*mmsize]
 -     DEINT2_PS   0, 1, 4
 -     DEINT2_PS   2, 3, 4
 -     mulps      m0, m0, m5
 -     mulps      m1, m1, m5
 -     mulps      m2, m2, m5
 -     mulps      m3, m3, m5
 -     cvtps2dq   m0, m0
 -     cvtps2dq   m1, m1
 -     cvtps2dq   m2, m2
 -     cvtps2dq   m3, m3
 -     packssdw   m0, m2
 -     packssdw   m1, m3
 -     mova  [dst0q+lenq], m0
 -     mova  [dst1q+lenq], m1
 -     add      lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_FLT_TO_S16P_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLT_TO_S16P_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_s16p_6ch(int16_t *const *dst, float *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLT_TO_S16P_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_flt_to_s16p_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 - %else
 - cglobal conv_flt_to_s16p_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 - %define lend dword r2m
 - %endif
 -     mov     dst1q, [dstq+  gprsize]
 -     mov     dst2q, [dstq+2*gprsize]
 -     mov     dst3q, [dstq+3*gprsize]
 -     mov     dst4q, [dstq+4*gprsize]
 -     mov     dst5q, [dstq+5*gprsize]
 -     mov      dstq, [dstq          ]
 -     sub     dst1q, dstq
 -     sub     dst2q, dstq
 -     sub     dst3q, dstq
 -     sub     dst4q, dstq
 -     sub     dst5q, dstq
 -     mova       m6, [pf_s16_scale]
 - .loop:
 -     mulps      m0, m6, [srcq+0*mmsize]
 -     mulps      m3, m6, [srcq+1*mmsize]
 -     mulps      m1, m6, [srcq+2*mmsize]
 -     mulps      m4, m6, [srcq+3*mmsize]
 -     mulps      m2, m6, [srcq+4*mmsize]
 -     mulps      m5, m6, [srcq+5*mmsize]
 -     cvtps2dq   m0, m0
 -     cvtps2dq   m1, m1
 -     cvtps2dq   m2, m2
 -     cvtps2dq   m3, m3
 -     cvtps2dq   m4, m4
 -     cvtps2dq   m5, m5
 -     packssdw   m0, m3               ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 -     packssdw   m1, m4               ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 -     packssdw   m2, m5               ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 -     PALIGNR    m3, m1, m0, 12, m4   ; m3 =  6,  7,  8,  9, 10, 11,  x,  x
 -     shufps     m1, m2, q1032        ; m1 = 12, 13, 14, 15, 16, 17, 18, 19
 -     psrldq     m2, 4                ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 -     SBUTTERFLY2 wd, 0, 3, 4         ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 -                                     ; m3 =  4, 10,  5, 11,  x,  x,  x,  x
 -     SBUTTERFLY2 wd, 1, 2, 4         ; m1 = 12, 18, 13, 19, 14, 20, 15, 21
 -                                     ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 -     SBUTTERFLY2 dq, 0, 1, 4         ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 -                                     ; m1 =  2,  8, 14, 20,  3,  9, 15, 21
 -     punpckldq  m3, m2               ; m3 =  4, 10, 16, 22,  5, 11, 17, 23
 -     movq    [dstq      ], m0
 -     movhps  [dstq+dst1q], m0
 -     movq    [dstq+dst2q], m1
 -     movhps  [dstq+dst3q], m1
 -     movq    [dstq+dst4q], m3
 -     movhps  [dstq+dst5q], m3
 -     add      srcq, mmsize*6
 -     add      dstq, mmsize/2
 -     sub      lend, mmsize/4
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_FLT_TO_S16P_6CH
 - INIT_XMM ssse3
 - CONV_FLT_TO_S16P_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLT_TO_S16P_6CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_fltp_2ch(float *const *dst, float *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLT_TO_FLTP_2CH 0
 - cglobal conv_flt_to_fltp_2ch, 3,4,3, dst0, src, len, dst1
 -     lea    lenq, [4*lend]
 -     mov   dst1q, [dst0q+gprsize]
 -     mov   dst0q, [dst0q        ]
 -     lea    srcq, [srcq+2*lenq]
 -     add   dst0q, lenq
 -     add   dst1q, lenq
 -     neg    lenq
 - .loop:
 -     mova     m0, [srcq+2*lenq       ]
 -     mova     m1, [srcq+2*lenq+mmsize]
 -     DEINT2_PS 0, 1, 2
 -     mova  [dst0q+lenq], m0
 -     mova  [dst1q+lenq], m1
 -     add    lenq, mmsize
 -     jl .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse
 - CONV_FLT_TO_FLTP_2CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLT_TO_FLTP_2CH
 - %endif
 - 
 - ;------------------------------------------------------------------------------
 - ; void ff_conv_flt_to_fltp_6ch(float *const *dst, float *src, int len,
 - ;                              int channels);
 - ;------------------------------------------------------------------------------
 - 
 - %macro CONV_FLT_TO_FLTP_6CH 0
 - %if ARCH_X86_64
 - cglobal conv_flt_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 - %else
 - cglobal conv_flt_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 - %define lend dword r2m
 - %endif
 -     mov     dst1q, [dstq+  gprsize]
 -     mov     dst2q, [dstq+2*gprsize]
 -     mov     dst3q, [dstq+3*gprsize]
 -     mov     dst4q, [dstq+4*gprsize]
 -     mov     dst5q, [dstq+5*gprsize]
 -     mov      dstq, [dstq          ]
 -     sub     dst1q, dstq
 -     sub     dst2q, dstq
 -     sub     dst3q, dstq
 -     sub     dst4q, dstq
 -     sub     dst5q, dstq
 - .loop:
 -     mova       m0, [srcq+0*mmsize]  ; m0 =  0,  1,  2,  3
 -     mova       m1, [srcq+1*mmsize]  ; m1 =  4,  5,  6,  7
 -     mova       m2, [srcq+2*mmsize]  ; m2 =  8,  9, 10, 11
 -     mova       m3, [srcq+3*mmsize]  ; m3 = 12, 13, 14, 15
 -     mova       m4, [srcq+4*mmsize]  ; m4 = 16, 17, 18, 19
 -     mova       m5, [srcq+5*mmsize]  ; m5 = 20, 21, 22, 23
 - 
 -     SBUTTERFLY2 dq, 0, 3, 6         ; m0 =  0, 12,  1, 13
 -                                     ; m3 =  2, 14,  3, 15
 -     SBUTTERFLY2 dq, 1, 4, 6         ; m1 =  4, 16,  5, 17
 -                                     ; m4 =  6, 18,  7, 19
 -     SBUTTERFLY2 dq, 2, 5, 6         ; m2 =  8, 20,  9, 21
 -                                     ; m5 = 10, 22, 11, 23
 -     SBUTTERFLY2 dq, 0, 4, 6         ; m0 =  0,  6, 12, 18
 -                                     ; m4 =  1,  7, 13, 19
 -     SBUTTERFLY2 dq, 3, 2, 6         ; m3 =  2,  8, 14, 20
 -                                     ; m2 =  3,  9, 15, 21
 -     SBUTTERFLY2 dq, 1, 5, 6         ; m1 =  4, 10, 16, 22
 -                                     ; m5 =  5, 11, 17, 23
 -     mova [dstq      ], m0
 -     mova [dstq+dst1q], m4
 -     mova [dstq+dst2q], m3
 -     mova [dstq+dst3q], m2
 -     mova [dstq+dst4q], m1
 -     mova [dstq+dst5q], m5
 -     add      srcq, mmsize*6
 -     add      dstq, mmsize
 -     sub      lend, mmsize/4
 -     jg .loop
 -     REP_RET
 - %endmacro
 - 
 - INIT_XMM sse2
 - CONV_FLT_TO_FLTP_6CH
 - %if HAVE_AVX_EXTERNAL
 - INIT_XMM avx
 - CONV_FLT_TO_FLTP_6CH
 - %endif
 
 
  |