| ;****************************************************************************** |
| ;* linear least squares model |
| ;* |
| ;* Copyright (c) 2013 Loren Merritt |
| ;* |
| ;* This file is part of FFmpeg. |
| ;* |
| ;* FFmpeg is free software; you can redistribute it and/or |
| ;* modify it under the terms of the GNU Lesser General Public |
| ;* License as published by the Free Software Foundation; either |
| ;* version 2.1 of the License, or (at your option) any later version. |
| ;* |
| ;* FFmpeg is distributed in the hope that it will be useful, |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| ;* Lesser General Public License for more details. |
| ;* |
| ;* You should have received a copy of the GNU Lesser General Public |
| ;* License along with FFmpeg; if not, write to the Free Software |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| ;****************************************************************************** |
| |
| %include "x86util.asm" |
| |
| SECTION .text |
| |
| %define MAX_VARS 32 |
| %define MAX_VARS_ALIGN (MAX_VARS+4) |
| %define COVAR_STRIDE MAX_VARS_ALIGN*8 |
| %define COVAR(x,y) [covarq + (x)*8 + (y)*COVAR_STRIDE] |
| |
| struc LLSModel |
| .covariance: resq MAX_VARS_ALIGN*MAX_VARS_ALIGN |
| .coeff: resq MAX_VARS*MAX_VARS |
| .variance: resq MAX_VARS |
| .indep_count: resd 1 |
| endstruc |
| |
| %macro ADDPD_MEM 2 |
| %if cpuflag(avx) |
| vaddpd %2, %2, %1 |
| %else |
| addpd %2, %1 |
| %endif |
| mova %1, %2 |
| %endmacro |
| |
| INIT_XMM sse2 |
| %define movdqa movaps |
| cglobal update_lls, 2,5,8, ctx, var, i, j, covar2 |
| %define covarq ctxq |
| mov id, [ctxq + LLSModel.indep_count] |
| lea varq, [varq + iq*8] |
| neg iq |
| mov covar2q, covarq |
| .loopi: |
| ; Compute all 3 pairwise products of a 2x2 block that lies on the diagonal |
| mova m1, [varq + iq*8] |
| mova m3, [varq + iq*8 + 16] |
| pshufd m4, m1, q1010 |
| pshufd m5, m1, q3232 |
| pshufd m6, m3, q1010 |
| pshufd m7, m3, q3232 |
| mulpd m0, m1, m4 |
| mulpd m1, m1, m5 |
| lea covarq, [covar2q + 16] |
| ADDPD_MEM COVAR(-2,0), m0 |
| ADDPD_MEM COVAR(-2,1), m1 |
| lea jq, [iq + 2] |
| cmp jd, -2 |
| jg .skip4x4 |
| .loop4x4: |
| ; Compute all 16 pairwise products of a 4x4 block |
| mulpd m0, m4, m3 |
| mulpd m1, m5, m3 |
| mulpd m2, m6, m3 |
| mulpd m3, m3, m7 |
| ADDPD_MEM COVAR(0,0), m0 |
| ADDPD_MEM COVAR(0,1), m1 |
| ADDPD_MEM COVAR(0,2), m2 |
| ADDPD_MEM COVAR(0,3), m3 |
| mova m3, [varq + jq*8 + 16] |
| mulpd m0, m4, m3 |
| mulpd m1, m5, m3 |
| mulpd m2, m6, m3 |
| mulpd m3, m3, m7 |
| ADDPD_MEM COVAR(2,0), m0 |
| ADDPD_MEM COVAR(2,1), m1 |
| ADDPD_MEM COVAR(2,2), m2 |
| ADDPD_MEM COVAR(2,3), m3 |
| mova m3, [varq + jq*8 + 32] |
| add covarq, 32 |
| add jq, 4 |
| cmp jd, -2 |
| jle .loop4x4 |
| .skip4x4: |
| test jd, jd |
| jg .skip2x4 |
| mulpd m4, m3 |
| mulpd m5, m3 |
| mulpd m6, m3 |
| mulpd m7, m3 |
| ADDPD_MEM COVAR(0,0), m4 |
| ADDPD_MEM COVAR(0,1), m5 |
| ADDPD_MEM COVAR(0,2), m6 |
| ADDPD_MEM COVAR(0,3), m7 |
| .skip2x4: |
| add iq, 4 |
| add covar2q, 4*COVAR_STRIDE+32 |
| cmp id, -2 |
| jle .loopi |
| test id, id |
| jg .ret |
| mov jq, iq |
| %define covarq covar2q |
| .loop2x1: |
| movsd m0, [varq + iq*8] |
| movlhps m0, m0 |
| mulpd m0, [varq + jq*8] |
| ADDPD_MEM COVAR(0,0), m0 |
| inc iq |
| add covarq, COVAR_STRIDE |
| test id, id |
| jle .loop2x1 |
| .ret: |
| REP_RET |
| |
| %macro UPDATE_LLS 0 |
| cglobal update_lls, 3,6,8, ctx, var, count, i, j, count2 |
| %define covarq ctxq |
| mov countd, [ctxq + LLSModel.indep_count] |
| lea count2d, [countq-2] |
| xor id, id |
| .loopi: |
| ; Compute all 10 pairwise products of a 4x4 block that lies on the diagonal |
| mova ymm1, [varq + iq*8] |
| vbroadcastsd ymm4, [varq + iq*8] |
| vbroadcastsd ymm5, [varq + iq*8 + 8] |
| vbroadcastsd ymm6, [varq + iq*8 + 16] |
| vbroadcastsd ymm7, [varq + iq*8 + 24] |
| vextractf128 xmm3, ymm1, 1 |
| %if cpuflag(fma3) |
| mova ymm0, COVAR(iq ,0) |
| mova xmm2, COVAR(iq+2,2) |
| fmaddpd ymm0, ymm1, ymm4, ymm0 |
| fmaddpd xmm2, xmm3, xmm6, xmm2 |
| fmaddpd ymm1, ymm5, ymm1, COVAR(iq ,1) |
| fmaddpd xmm3, xmm7, xmm3, COVAR(iq+2,3) |
| mova COVAR(iq ,0), ymm0 |
| mova COVAR(iq ,1), ymm1 |
| mova COVAR(iq+2,2), xmm2 |
| mova COVAR(iq+2,3), xmm3 |
| %else |
| vmulpd ymm0, ymm1, ymm4 |
| vmulpd ymm1, ymm1, ymm5 |
| vmulpd xmm2, xmm3, xmm6 |
| vmulpd xmm3, xmm3, xmm7 |
| ADDPD_MEM COVAR(iq ,0), ymm0 |
| ADDPD_MEM COVAR(iq ,1), ymm1 |
| ADDPD_MEM COVAR(iq+2,2), xmm2 |
| ADDPD_MEM COVAR(iq+2,3), xmm3 |
| %endif ; cpuflag(fma3) |
| lea jd, [iq + 4] |
| cmp jd, count2d |
| jg .skip4x4 |
| .loop4x4: |
| ; Compute all 16 pairwise products of a 4x4 block |
| mova ymm3, [varq + jq*8] |
| %if cpuflag(fma3) |
| mova ymm0, COVAR(jq, 0) |
| mova ymm1, COVAR(jq, 1) |
| mova ymm2, COVAR(jq, 2) |
| fmaddpd ymm0, ymm3, ymm4, ymm0 |
| fmaddpd ymm1, ymm3, ymm5, ymm1 |
| fmaddpd ymm2, ymm3, ymm6, ymm2 |
| fmaddpd ymm3, ymm7, ymm3, COVAR(jq,3) |
| mova COVAR(jq, 0), ymm0 |
| mova COVAR(jq, 1), ymm1 |
| mova COVAR(jq, 2), ymm2 |
| mova COVAR(jq, 3), ymm3 |
| %else |
| vmulpd ymm0, ymm3, ymm4 |
| vmulpd ymm1, ymm3, ymm5 |
| vmulpd ymm2, ymm3, ymm6 |
| vmulpd ymm3, ymm3, ymm7 |
| ADDPD_MEM COVAR(jq,0), ymm0 |
| ADDPD_MEM COVAR(jq,1), ymm1 |
| ADDPD_MEM COVAR(jq,2), ymm2 |
| ADDPD_MEM COVAR(jq,3), ymm3 |
| %endif ; cpuflag(fma3) |
| add jd, 4 |
| cmp jd, count2d |
| jle .loop4x4 |
| .skip4x4: |
| cmp jd, countd |
| jg .skip2x4 |
| mova xmm3, [varq + jq*8] |
| %if cpuflag(fma3) |
| mova xmm0, COVAR(jq, 0) |
| mova xmm1, COVAR(jq, 1) |
| mova xmm2, COVAR(jq, 2) |
| fmaddpd xmm0, xmm3, xmm4, xmm0 |
| fmaddpd xmm1, xmm3, xmm5, xmm1 |
| fmaddpd xmm2, xmm3, xmm6, xmm2 |
| fmaddpd xmm3, xmm7, xmm3, COVAR(jq,3) |
| mova COVAR(jq, 0), xmm0 |
| mova COVAR(jq, 1), xmm1 |
| mova COVAR(jq, 2), xmm2 |
| mova COVAR(jq, 3), xmm3 |
| %else |
| vmulpd xmm0, xmm3, xmm4 |
| vmulpd xmm1, xmm3, xmm5 |
| vmulpd xmm2, xmm3, xmm6 |
| vmulpd xmm3, xmm3, xmm7 |
| ADDPD_MEM COVAR(jq,0), xmm0 |
| ADDPD_MEM COVAR(jq,1), xmm1 |
| ADDPD_MEM COVAR(jq,2), xmm2 |
| ADDPD_MEM COVAR(jq,3), xmm3 |
| %endif ; cpuflag(fma3) |
| .skip2x4: |
| add id, 4 |
| add covarq, 4*COVAR_STRIDE |
| cmp id, count2d |
| jle .loopi |
| cmp id, countd |
| jg .ret |
| mov jd, id |
| .loop2x1: |
| vmovddup xmm0, [varq + iq*8] |
| %if cpuflag(fma3) |
| mova xmm1, [varq + jq*8] |
| fmaddpd xmm0, xmm1, xmm0, COVAR(jq,0) |
| mova COVAR(jq,0), xmm0 |
| %else |
| vmulpd xmm0, [varq + jq*8] |
| ADDPD_MEM COVAR(jq,0), xmm0 |
| %endif ; cpuflag(fma3) |
| inc id |
| add covarq, COVAR_STRIDE |
| cmp id, countd |
| jle .loop2x1 |
| .ret: |
| REP_RET |
| %endmacro ; UPDATE_LLS |
| |
| %if HAVE_AVX_EXTERNAL |
| INIT_YMM avx |
| UPDATE_LLS |
| %endif |
| %if HAVE_FMA3_EXTERNAL |
| INIT_YMM fma3 |
| UPDATE_LLS |
| %endif |
| |
| INIT_XMM sse2 |
| cglobal evaluate_lls, 3,4,2, ctx, var, order, i |
| ; This function is often called on the same buffer as update_lls, but with |
| ; an offset. They can't both be aligned. |
| ; Load halves rather than movu to avoid store-forwarding stalls, since the |
| ; input was initialized immediately prior to this function using scalar math. |
| %define coefsq ctxq |
| mov id, orderd |
| imul orderd, MAX_VARS |
| lea coefsq, [ctxq + LLSModel.coeff + orderq*8] |
| movsd m0, [varq] |
| movhpd m0, [varq + 8] |
| mulpd m0, [coefsq] |
| lea coefsq, [coefsq + iq*8] |
| lea varq, [varq + iq*8] |
| neg iq |
| add iq, 2 |
| .loop: |
| movsd m1, [varq + iq*8] |
| movhpd m1, [varq + iq*8 + 8] |
| mulpd m1, [coefsq + iq*8] |
| addpd m0, m1 |
| add iq, 2 |
| jl .loop |
| jg .skip1 |
| movsd m1, [varq + iq*8] |
| mulsd m1, [coefsq + iq*8] |
| addpd m0, m1 |
| .skip1: |
| movhlps m1, m0 |
| addsd m0, m1 |
| %if ARCH_X86_32 |
| movsd r0m, m0 |
| fld qword r0m |
| %endif |
| RET |