| @@ -80,7 +80,7 @@ SECTION .text | |||
| %endmacro | |||
| INIT_MMX mmx | |||
| ; ff_h264_idct_add_mmx(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct_add_8_mmx(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct_add_8, 3, 3, 0 | |||
| IDCT4_ADD r0, r1, r2 | |||
| RET | |||
| @@ -202,7 +202,7 @@ cglobal h264_idct_add_8, 3, 3, 0 | |||
| %endmacro | |||
| INIT_MMX mmx | |||
| ; ff_h264_idct8_add_mmx(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct8_add_8_mmx(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct8_add_8, 3, 4, 0 | |||
| %assign pad 128+4-(stack_offset&7) | |||
| SUB rsp, pad | |||
| @@ -270,7 +270,7 @@ cglobal h264_idct8_add_8, 3, 4, 0 | |||
| %endmacro | |||
| INIT_XMM sse2 | |||
| ; ff_h264_idct8_add_sse2(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct8_add_8_sse2(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct8_add_8, 3, 4, 10 | |||
| IDCT8_ADD_SSE r0, r1, r2, r3 | |||
| RET | |||
| @@ -307,7 +307,7 @@ cglobal h264_idct8_add_8, 3, 4, 10 | |||
| %endmacro | |||
| INIT_MMX mmxext | |||
| ; ff_h264_idct_dc_add_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct_dc_add_8_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| %if ARCH_X86_64 | |||
| cglobal h264_idct_dc_add_8, 3, 4, 0 | |||
| movsx r3, word [r1] | |||
| @@ -316,7 +316,7 @@ cglobal h264_idct_dc_add_8, 3, 4, 0 | |||
| DC_ADD_MMXEXT_OP movh, r0, r2, r3 | |||
| RET | |||
| ; ff_h264_idct8_dc_add_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct8_dc_add_8_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct8_dc_add_8, 3, 4, 0 | |||
| movsx r3, word [r1] | |||
| mov dword [r1], 0 | |||
| @@ -326,6 +326,7 @@ cglobal h264_idct8_dc_add_8, 3, 4, 0 | |||
| DC_ADD_MMXEXT_OP mova, r0, r2, r3 | |||
| RET | |||
| %else | |||
| ; ff_h264_idct_dc_add_8_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct_dc_add_8, 2, 3, 0 | |||
| movsx r2, word [r1] | |||
| mov dword [r1], 0 | |||
| @@ -334,7 +335,7 @@ cglobal h264_idct_dc_add_8, 2, 3, 0 | |||
| DC_ADD_MMXEXT_OP movh, r0, r1, r2 | |||
| RET | |||
| ; ff_h264_idct8_dc_add_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| ; ff_h264_idct8_dc_add_8_mmxext(uint8_t *dst, int16_t *block, int stride) | |||
| cglobal h264_idct8_dc_add_8, 2, 3, 0 | |||
| movsx r2, word [r1] | |||
| mov dword [r1], 0 | |||
| @@ -347,8 +348,9 @@ cglobal h264_idct8_dc_add_8, 2, 3, 0 | |||
| %endif | |||
| INIT_MMX mmx | |||
| ; ff_h264_idct_add16_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16_8_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16_8, 5, 7 + npicregs, 0, dst, block_offset, block, stride, nnzc, cntr, coeff, picreg | |||
| xor r5, r5 | |||
| %ifdef PIC | |||
| @@ -369,8 +371,9 @@ cglobal h264_idct_add16_8, 5, 7 + npicregs, 0, dst, block_offset, block, stride, | |||
| jl .nextblock | |||
| REP_RET | |||
| ; ff_h264_idct8_add4_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct8_add4_8_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct8_add4_8, 5, 7 + npicregs, 0, dst, block_offset, block, stride, nnzc, cntr, coeff, picreg | |||
| %assign pad 128+4-(stack_offset&7) | |||
| SUB rsp, pad | |||
| @@ -402,8 +405,9 @@ cglobal h264_idct8_add4_8, 5, 7 + npicregs, 0, dst, block_offset, block, stride, | |||
| RET | |||
| INIT_MMX mmxext | |||
| ; ff_h264_idct_add16_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16_8_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| xor r5, r5 | |||
| %ifdef PIC | |||
| @@ -448,8 +452,9 @@ cglobal h264_idct_add16_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride | |||
| REP_RET | |||
| INIT_MMX mmx | |||
| ; ff_h264_idct_add16intra_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16intra_8_mmx(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16intra_8, 5, 7 + npicregs, 0, dst, block_offset, block, stride, nnzc, cntr, coeff, picreg | |||
| xor r5, r5 | |||
| %ifdef PIC | |||
| @@ -472,9 +477,9 @@ cglobal h264_idct_add16intra_8, 5, 7 + npicregs, 0, dst, block_offset, block, st | |||
| REP_RET | |||
| INIT_MMX mmxext | |||
| ; ff_h264_idct_add16intra_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16intra_8_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16intra_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| xor r5, r5 | |||
| %ifdef PIC | |||
| @@ -516,9 +521,9 @@ cglobal h264_idct_add16intra_8, 5, 8 + npicregs, 0, dst1, block_offset, block, s | |||
| jl .nextblock | |||
| REP_RET | |||
| ; ff_h264_idct8_add4_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct8_add4_8_mmxext(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct8_add4_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| %assign pad 128+4-(stack_offset&7) | |||
| SUB rsp, pad | |||
| @@ -578,8 +583,9 @@ cglobal h264_idct8_add4_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride | |||
| RET | |||
| INIT_XMM sse2 | |||
| ; ff_h264_idct8_add4_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct8_add4_8_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct8_add4_8, 5, 8 + npicregs, 10, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| xor r5, r5 | |||
| %ifdef PIC | |||
| @@ -654,8 +660,8 @@ h264_idct_add8_mmx_plane: | |||
| jnz .nextblock | |||
| rep ret | |||
| ; ff_h264_idct_add8_mmx(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add8_8_mmx(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add8_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| mov r5, 16 | |||
| add r2, 512 | |||
| @@ -719,8 +725,9 @@ h264_idct_add8_mmxext_plane: | |||
| rep ret | |||
| INIT_MMX mmxext | |||
| ; ff_h264_idct_add8_mmxext(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add8_8_mmxext(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add8_8, 5, 8 + npicregs, 0, dst1, block_offset, block, stride, nnzc, cntr, coeff, dst2, picreg | |||
| mov r5, 16 | |||
| add r2, 512 | |||
| @@ -802,8 +809,9 @@ h264_add8x4_idct_sse2: | |||
| %endif | |||
| %endmacro | |||
| ; ff_h264_idct_add16_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16_8_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16_8, 5, 5 + ARCH_X86_64, 8 | |||
| %if ARCH_X86_64 | |||
| mov r5, r0 | |||
| @@ -849,8 +857,9 @@ cglobal h264_idct_add16_8, 5, 5 + ARCH_X86_64, 8 | |||
| %endif | |||
| %endmacro | |||
| ; ff_h264_idct_add16intra_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add16intra_8_sse2(uint8_t *dst, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add16intra_8, 5, 7 + ARCH_X86_64, 8 | |||
| %if ARCH_X86_64 | |||
| mov r7, r0 | |||
| @@ -900,8 +909,9 @@ cglobal h264_idct_add16intra_8, 5, 7 + ARCH_X86_64, 8 | |||
| %endif | |||
| %endmacro | |||
| ; ff_h264_idct_add8_sse2(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, const uint8_t nnzc[6*8]) | |||
| ; ff_h264_idct_add8_8_sse2(uint8_t **dest, const int *block_offset, | |||
| ; int16_t *block, int stride, | |||
| ; const uint8_t nnzc[6 * 8]) | |||
| cglobal h264_idct_add8_8, 5, 7 + ARCH_X86_64, 8 | |||
| add r2, 512 | |||
| %if ARCH_X86_64 | |||