| ;****************************************************************************** |
| ;* Copyright (c) 2012 Michael Niedermayer |
| ;* |
| ;* This file is part of FFmpeg. |
| ;* |
| ;* FFmpeg is free software; you can redistribute it and/or |
| ;* modify it under the terms of the GNU Lesser General Public |
| ;* License as published by the Free Software Foundation; either |
| ;* version 2.1 of the License, or (at your option) any later version. |
| ;* |
| ;* FFmpeg is distributed in the hope that it will be useful, |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| ;* Lesser General Public License for more details. |
| ;* |
| ;* You should have received a copy of the GNU Lesser General Public |
| ;* License along with FFmpeg; if not, write to the Free Software |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| ;****************************************************************************** |
| |
| %include "libavutil/x86/x86util.asm" |
| |
| |
| SECTION_RODATA 32 |
| dw1: times 8 dd 1 |
| w1 : times 16 dw 1 |
| |
| SECTION .text |
| |
| %macro MIX2_FLT 1 |
| cglobal mix_2_1_%1_float, 7, 7, 6, out, in1, in2, coeffp, index1, index2, len |
| %ifidn %1, a |
| test in1q, mmsize-1 |
| jne mix_2_1_float_u_int %+ SUFFIX |
| test in2q, mmsize-1 |
| jne mix_2_1_float_u_int %+ SUFFIX |
| test outq, mmsize-1 |
| jne mix_2_1_float_u_int %+ SUFFIX |
| %else |
| mix_2_1_float_u_int %+ SUFFIX: |
| %endif |
| VBROADCASTSS m4, [coeffpq + 4*index1q] |
| VBROADCASTSS m5, [coeffpq + 4*index2q] |
| shl lend , 2 |
| add in1q , lenq |
| add in2q , lenq |
| add outq , lenq |
| neg lenq |
| .next: |
| %ifidn %1, a |
| mulps m0, m4, [in1q + lenq ] |
| mulps m1, m5, [in2q + lenq ] |
| mulps m2, m4, [in1q + lenq + mmsize] |
| mulps m3, m5, [in2q + lenq + mmsize] |
| %else |
| movu m0, [in1q + lenq ] |
| movu m1, [in2q + lenq ] |
| movu m2, [in1q + lenq + mmsize] |
| movu m3, [in2q + lenq + mmsize] |
| mulps m0, m0, m4 |
| mulps m1, m1, m5 |
| mulps m2, m2, m4 |
| mulps m3, m3, m5 |
| %endif |
| addps m0, m0, m1 |
| addps m2, m2, m3 |
| mov%1 [outq + lenq ], m0 |
| mov%1 [outq + lenq + mmsize], m2 |
| add lenq, mmsize*2 |
| jl .next |
| REP_RET |
| %endmacro |
| |
| %macro MIX1_FLT 1 |
| cglobal mix_1_1_%1_float, 5, 5, 3, out, in, coeffp, index, len |
| %ifidn %1, a |
| test inq, mmsize-1 |
| jne mix_1_1_float_u_int %+ SUFFIX |
| test outq, mmsize-1 |
| jne mix_1_1_float_u_int %+ SUFFIX |
| %else |
| mix_1_1_float_u_int %+ SUFFIX: |
| %endif |
| VBROADCASTSS m2, [coeffpq + 4*indexq] |
| shl lenq , 2 |
| add inq , lenq |
| add outq , lenq |
| neg lenq |
| .next: |
| %ifidn %1, a |
| mulps m0, m2, [inq + lenq ] |
| mulps m1, m2, [inq + lenq + mmsize] |
| %else |
| movu m0, [inq + lenq ] |
| movu m1, [inq + lenq + mmsize] |
| mulps m0, m0, m2 |
| mulps m1, m1, m2 |
| %endif |
| mov%1 [outq + lenq ], m0 |
| mov%1 [outq + lenq + mmsize], m1 |
| add lenq, mmsize*2 |
| jl .next |
| REP_RET |
| %endmacro |
| |
| %macro MIX1_INT16 1 |
| cglobal mix_1_1_%1_int16, 5, 5, 6, out, in, coeffp, index, len |
| %ifidn %1, a |
| test inq, mmsize-1 |
| jne mix_1_1_int16_u_int %+ SUFFIX |
| test outq, mmsize-1 |
| jne mix_1_1_int16_u_int %+ SUFFIX |
| %else |
| mix_1_1_int16_u_int %+ SUFFIX: |
| %endif |
| movd m4, [coeffpq + 4*indexq] |
| SPLATW m5, m4 |
| psllq m4, 32 |
| psrlq m4, 48 |
| mova m0, [w1] |
| psllw m0, m4 |
| psrlw m0, 1 |
| punpcklwd m5, m0 |
| add lenq , lenq |
| add inq , lenq |
| add outq , lenq |
| neg lenq |
| .next: |
| mov%1 m0, [inq + lenq ] |
| mov%1 m2, [inq + lenq + mmsize] |
| mova m1, m0 |
| mova m3, m2 |
| punpcklwd m0, [w1] |
| punpckhwd m1, [w1] |
| punpcklwd m2, [w1] |
| punpckhwd m3, [w1] |
| pmaddwd m0, m5 |
| pmaddwd m1, m5 |
| pmaddwd m2, m5 |
| pmaddwd m3, m5 |
| psrad m0, m4 |
| psrad m1, m4 |
| psrad m2, m4 |
| psrad m3, m4 |
| packssdw m0, m1 |
| packssdw m2, m3 |
| mov%1 [outq + lenq ], m0 |
| mov%1 [outq + lenq + mmsize], m2 |
| add lenq, mmsize*2 |
| jl .next |
| %if mmsize == 8 |
| emms |
| RET |
| %else |
| REP_RET |
| %endif |
| %endmacro |
| |
| %macro MIX2_INT16 1 |
| cglobal mix_2_1_%1_int16, 7, 7, 8, out, in1, in2, coeffp, index1, index2, len |
| %ifidn %1, a |
| test in1q, mmsize-1 |
| jne mix_2_1_int16_u_int %+ SUFFIX |
| test in2q, mmsize-1 |
| jne mix_2_1_int16_u_int %+ SUFFIX |
| test outq, mmsize-1 |
| jne mix_2_1_int16_u_int %+ SUFFIX |
| %else |
| mix_2_1_int16_u_int %+ SUFFIX: |
| %endif |
| movd m4, [coeffpq + 4*index1q] |
| movd m6, [coeffpq + 4*index2q] |
| SPLATW m5, m4 |
| SPLATW m6, m6 |
| psllq m4, 32 |
| psrlq m4, 48 |
| mova m7, [dw1] |
| pslld m7, m4 |
| psrld m7, 1 |
| punpcklwd m5, m6 |
| add lend , lend |
| add in1q , lenq |
| add in2q , lenq |
| add outq , lenq |
| neg lenq |
| .next: |
| mov%1 m0, [in1q + lenq ] |
| mov%1 m2, [in2q + lenq ] |
| mova m1, m0 |
| punpcklwd m0, m2 |
| punpckhwd m1, m2 |
| |
| mov%1 m2, [in1q + lenq + mmsize] |
| mov%1 m6, [in2q + lenq + mmsize] |
| mova m3, m2 |
| punpcklwd m2, m6 |
| punpckhwd m3, m6 |
| |
| pmaddwd m0, m5 |
| pmaddwd m1, m5 |
| pmaddwd m2, m5 |
| pmaddwd m3, m5 |
| paddd m0, m7 |
| paddd m1, m7 |
| paddd m2, m7 |
| paddd m3, m7 |
| psrad m0, m4 |
| psrad m1, m4 |
| psrad m2, m4 |
| psrad m3, m4 |
| packssdw m0, m1 |
| packssdw m2, m3 |
| mov%1 [outq + lenq ], m0 |
| mov%1 [outq + lenq + mmsize], m2 |
| add lenq, mmsize*2 |
| jl .next |
| %if mmsize == 8 |
| emms |
| RET |
| %else |
| REP_RET |
| %endif |
| %endmacro |
| |
| |
| INIT_MMX mmx |
| MIX1_INT16 u |
| MIX1_INT16 a |
| MIX2_INT16 u |
| MIX2_INT16 a |
| |
| INIT_XMM sse |
| MIX2_FLT u |
| MIX2_FLT a |
| MIX1_FLT u |
| MIX1_FLT a |
| |
| INIT_XMM sse2 |
| MIX1_INT16 u |
| MIX1_INT16 a |
| MIX2_INT16 u |
| MIX2_INT16 a |
| |
| %if HAVE_AVX_EXTERNAL |
| INIT_YMM avx |
| MIX2_FLT u |
| MIX2_FLT a |
| MIX1_FLT u |
| MIX1_FLT a |
| %endif |