|
- ;*****************************************************************************
- ;* SSE2-optimized weighted prediction code
- ;*****************************************************************************
- ;* Copyright (c) 2004-2005 Michael Niedermayer, Loren Merritt
- ;* Copyright (C) 2010 Eli Friedman <eli.friedman@gmail.com>
- ;*
- ;* This file is part of Libav.
- ;*
- ;* Libav is free software; you can redistribute it and/or
- ;* modify it under the terms of the GNU Lesser General Public
- ;* License as published by the Free Software Foundation; either
- ;* version 2.1 of the License, or (at your option) any later version.
- ;*
- ;* Libav is distributed in the hope that it will be useful,
- ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
- ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- ;* Lesser General Public License for more details.
- ;*
- ;* You should have received a copy of the GNU Lesser General Public
- ;* License along with Libav; if not, write to the Free Software
- ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- ;******************************************************************************
-
- %include "libavutil/x86/x86util.asm"
-
- SECTION .text
-
- ;-----------------------------------------------------------------------------
- ; biweight pred:
- ;
- ; void ff_h264_biweight_16_sse2(uint8_t *dst, uint8_t *src, int stride,
- ; int height, int log2_denom, int weightd,
- ; int weights, int offset);
- ; and
- ; void ff_h264_weight_16_sse2(uint8_t *dst, int stride, int height,
- ; int log2_denom, int weight, int offset);
- ;-----------------------------------------------------------------------------
-
- %macro WEIGHT_SETUP 0
- add r5, r5
- inc r5
- movd m3, r4d
- movd m5, r5d
- movd m6, r3d
- pslld m5, m6
- psrld m5, 1
- %if mmsize == 16
- pshuflw m3, m3, 0
- pshuflw m5, m5, 0
- punpcklqdq m3, m3
- punpcklqdq m5, m5
- %else
- pshufw m3, m3, 0
- pshufw m5, m5, 0
- %endif
- pxor m7, m7
- %endmacro
-
- %macro WEIGHT_OP 2
- movh m0, [r0+%1]
- movh m1, [r0+%2]
- punpcklbw m0, m7
- punpcklbw m1, m7
- pmullw m0, m3
- pmullw m1, m3
- paddsw m0, m5
- paddsw m1, m5
- psraw m0, m6
- psraw m1, m6
- packuswb m0, m1
- %endmacro
-
- INIT_MMX mmxext
- cglobal h264_weight_16, 6, 6, 0
- WEIGHT_SETUP
- .nextrow:
- WEIGHT_OP 0, 4
- mova [r0 ], m0
- WEIGHT_OP 8, 12
- mova [r0+8], m0
- add r0, r1
- dec r2d
- jnz .nextrow
- REP_RET
-
- %macro WEIGHT_FUNC_MM 2
- cglobal h264_weight_%1, 6, 6, %2
- WEIGHT_SETUP
- .nextrow:
- WEIGHT_OP 0, mmsize/2
- mova [r0], m0
- add r0, r1
- dec r2d
- jnz .nextrow
- REP_RET
- %endmacro
-
- INIT_MMX mmxext
- WEIGHT_FUNC_MM 8, 0
- INIT_XMM sse2
- WEIGHT_FUNC_MM 16, 8
-
- %macro WEIGHT_FUNC_HALF_MM 2
- cglobal h264_weight_%1, 6, 6, %2
- WEIGHT_SETUP
- sar r2d, 1
- lea r3, [r1*2]
- .nextrow:
- WEIGHT_OP 0, r1
- movh [r0], m0
- %if mmsize == 16
- movhps [r0+r1], m0
- %else
- psrlq m0, 32
- movh [r0+r1], m0
- %endif
- add r0, r3
- dec r2d
- jnz .nextrow
- REP_RET
- %endmacro
-
- INIT_MMX mmxext
- WEIGHT_FUNC_HALF_MM 4, 0
- INIT_XMM sse2
- WEIGHT_FUNC_HALF_MM 8, 8
-
- %macro BIWEIGHT_SETUP 0
- %if ARCH_X86_64
- %define off_regd r7d
- %else
- %define off_regd r3d
- %endif
- mov off_regd, r7m
- add off_regd, 1
- or off_regd, 1
- add r4d, 1
- cmp r6d, 128
- je .nonnormal
- cmp r5d, 128
- jne .normal
- .nonnormal:
- sar r5d, 1
- sar r6d, 1
- sar off_regd, 1
- sub r4d, 1
- .normal:
- %if cpuflag(ssse3)
- movd m4, r5d
- movd m0, r6d
- %else
- movd m3, r5d
- movd m4, r6d
- %endif
- movd m5, off_regd
- movd m6, r4d
- pslld m5, m6
- psrld m5, 1
- %if cpuflag(ssse3)
- punpcklbw m4, m0
- pshuflw m4, m4, 0
- pshuflw m5, m5, 0
- punpcklqdq m4, m4
- punpcklqdq m5, m5
-
- %else
- %if mmsize == 16
- pshuflw m3, m3, 0
- pshuflw m4, m4, 0
- pshuflw m5, m5, 0
- punpcklqdq m3, m3
- punpcklqdq m4, m4
- punpcklqdq m5, m5
- %else
- pshufw m3, m3, 0
- pshufw m4, m4, 0
- pshufw m5, m5, 0
- %endif
- pxor m7, m7
- %endif
- %endmacro
-
- %macro BIWEIGHT_STEPA 3
- movh m%1, [r0+%3]
- movh m%2, [r1+%3]
- punpcklbw m%1, m7
- punpcklbw m%2, m7
- pmullw m%1, m3
- pmullw m%2, m4
- paddsw m%1, m%2
- %endmacro
-
- %macro BIWEIGHT_STEPB 0
- paddsw m0, m5
- paddsw m1, m5
- psraw m0, m6
- psraw m1, m6
- packuswb m0, m1
- %endmacro
-
- INIT_MMX mmxext
- cglobal h264_biweight_16, 7, 8, 0
- BIWEIGHT_SETUP
- movifnidn r3d, r3m
- .nextrow:
- BIWEIGHT_STEPA 0, 1, 0
- BIWEIGHT_STEPA 1, 2, 4
- BIWEIGHT_STEPB
- mova [r0], m0
- BIWEIGHT_STEPA 0, 1, 8
- BIWEIGHT_STEPA 1, 2, 12
- BIWEIGHT_STEPB
- mova [r0+8], m0
- add r0, r2
- add r1, r2
- dec r3d
- jnz .nextrow
- REP_RET
-
- %macro BIWEIGHT_FUNC_MM 2
- cglobal h264_biweight_%1, 7, 8, %2
- BIWEIGHT_SETUP
- movifnidn r3d, r3m
- .nextrow:
- BIWEIGHT_STEPA 0, 1, 0
- BIWEIGHT_STEPA 1, 2, mmsize/2
- BIWEIGHT_STEPB
- mova [r0], m0
- add r0, r2
- add r1, r2
- dec r3d
- jnz .nextrow
- REP_RET
- %endmacro
-
- INIT_MMX mmxext
- BIWEIGHT_FUNC_MM 8, 0
- INIT_XMM sse2
- BIWEIGHT_FUNC_MM 16, 8
-
- %macro BIWEIGHT_FUNC_HALF_MM 2
- cglobal h264_biweight_%1, 7, 8, %2
- BIWEIGHT_SETUP
- movifnidn r3d, r3m
- sar r3, 1
- lea r4, [r2*2]
- .nextrow:
- BIWEIGHT_STEPA 0, 1, 0
- BIWEIGHT_STEPA 1, 2, r2
- BIWEIGHT_STEPB
- movh [r0], m0
- %if mmsize == 16
- movhps [r0+r2], m0
- %else
- psrlq m0, 32
- movh [r0+r2], m0
- %endif
- add r0, r4
- add r1, r4
- dec r3d
- jnz .nextrow
- REP_RET
- %endmacro
-
- INIT_MMX mmxext
- BIWEIGHT_FUNC_HALF_MM 4, 0
- INIT_XMM sse2
- BIWEIGHT_FUNC_HALF_MM 8, 8
-
- %macro BIWEIGHT_SSSE3_OP 0
- pmaddubsw m0, m4
- pmaddubsw m2, m4
- paddsw m0, m5
- paddsw m2, m5
- psraw m0, m6
- psraw m2, m6
- packuswb m0, m2
- %endmacro
-
- INIT_XMM ssse3
- cglobal h264_biweight_16, 7, 8, 8
- BIWEIGHT_SETUP
- movifnidn r3d, r3m
-
- .nextrow:
- movh m0, [r0]
- movh m2, [r0+8]
- movh m3, [r1+8]
- punpcklbw m0, [r1]
- punpcklbw m2, m3
- BIWEIGHT_SSSE3_OP
- mova [r0], m0
- add r0, r2
- add r1, r2
- dec r3d
- jnz .nextrow
- REP_RET
-
- INIT_XMM ssse3
- cglobal h264_biweight_8, 7, 8, 8
- BIWEIGHT_SETUP
- movifnidn r3d, r3m
- sar r3, 1
- lea r4, [r2*2]
-
- .nextrow:
- movh m0, [r0]
- movh m1, [r1]
- movh m2, [r0+r2]
- movh m3, [r1+r2]
- punpcklbw m0, m1
- punpcklbw m2, m3
- BIWEIGHT_SSSE3_OP
- movh [r0], m0
- movhps [r0+r2], m0
- add r0, r4
- add r1, r4
- dec r3d
- jnz .nextrow
- REP_RET
|