| ;****************************************************************************** |
| ;* Copyright (c) 2010 David Conrad |
| ;* |
| ;* This file is part of FFmpeg. |
| ;* |
| ;* FFmpeg is free software; you can redistribute it and/or |
| ;* modify it under the terms of the GNU Lesser General Public |
| ;* License as published by the Free Software Foundation; either |
| ;* version 2.1 of the License, or (at your option) any later version. |
| ;* |
| ;* FFmpeg is distributed in the hope that it will be useful, |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| ;* Lesser General Public License for more details. |
| ;* |
| ;* You should have received a copy of the GNU Lesser General Public |
| ;* License along with FFmpeg; if not, write to the Free Software |
| ;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| ;****************************************************************************** |
| |
| %include "libavutil/x86/x86util.asm" |
| |
| SECTION_RODATA |
| pw_7: times 8 dw 7 |
| convert_to_unsigned_10bit: times 4 dd 0x200 |
| clip_10bit: times 8 dw 0x3ff |
| |
| cextern pw_3 |
| cextern pw_16 |
| cextern pw_32 |
| cextern pb_80 |
| |
| SECTION .text |
| |
| %macro UNPACK_ADD 6 |
| mov%5 %1, %3 |
| mov%6 m5, %4 |
| mova m4, %1 |
| mova %2, m5 |
| punpcklbw %1, m7 |
| punpcklbw m5, m7 |
| punpckhbw m4, m7 |
| punpckhbw %2, m7 |
| paddw %1, m5 |
| paddw %2, m4 |
| %endmacro |
| |
| %macro HPEL_FILTER 1 |
| ; dirac_hpel_filter_v_sse2(uint8_t *dst, uint8_t *src, int stride, int width); |
| cglobal dirac_hpel_filter_v_%1, 4,6,8, dst, src, stride, width, src0, stridex3 |
| mov src0q, srcq |
| lea stridex3q, [3*strideq] |
| sub src0q, stridex3q |
| pxor m7, m7 |
| .loop: |
| ; 7*(src[0] + src[1]) |
| UNPACK_ADD m0, m1, [srcq], [srcq + strideq], a,a |
| pmullw m0, [pw_7] |
| pmullw m1, [pw_7] |
| |
| ; 3*( ... + src[-2] + src[3]) |
| UNPACK_ADD m2, m3, [src0q + strideq], [srcq + stridex3q], a,a |
| paddw m0, m2 |
| paddw m1, m3 |
| pmullw m0, [pw_3] |
| pmullw m1, [pw_3] |
| |
| ; ... - 7*(src[-1] + src[2]) |
| UNPACK_ADD m2, m3, [src0q + strideq*2], [srcq + strideq*2], a,a |
| pmullw m2, [pw_7] |
| pmullw m3, [pw_7] |
| psubw m0, m2 |
| psubw m1, m3 |
| |
| ; ... - (src[-3] + src[4]) |
| UNPACK_ADD m2, m3, [src0q], [srcq + strideq*4], a,a |
| psubw m0, m2 |
| psubw m1, m3 |
| |
| paddw m0, [pw_16] |
| paddw m1, [pw_16] |
| psraw m0, 5 |
| psraw m1, 5 |
| packuswb m0, m1 |
| mova [dstq], m0 |
| add dstq, mmsize |
| add srcq, mmsize |
| add src0q, mmsize |
| sub widthd, mmsize |
| jg .loop |
| RET |
| |
| ; dirac_hpel_filter_h_sse2(uint8_t *dst, uint8_t *src, int width); |
| cglobal dirac_hpel_filter_h_%1, 3,3,8, dst, src, width |
| dec widthd |
| pxor m7, m7 |
| and widthd, ~(mmsize-1) |
| .loop: |
| ; 7*(src[0] + src[1]) |
| UNPACK_ADD m0, m1, [srcq + widthq], [srcq + widthq + 1], u,u |
| pmullw m0, [pw_7] |
| pmullw m1, [pw_7] |
| |
| ; 3*( ... + src[-2] + src[3]) |
| UNPACK_ADD m2, m3, [srcq + widthq - 2], [srcq + widthq + 3], u,u |
| paddw m0, m2 |
| paddw m1, m3 |
| pmullw m0, [pw_3] |
| pmullw m1, [pw_3] |
| |
| ; ... - 7*(src[-1] + src[2]) |
| UNPACK_ADD m2, m3, [srcq + widthq - 1], [srcq + widthq + 2], u,u |
| pmullw m2, [pw_7] |
| pmullw m3, [pw_7] |
| psubw m0, m2 |
| psubw m1, m3 |
| |
| ; ... - (src[-3] + src[4]) |
| UNPACK_ADD m2, m3, [srcq + widthq - 3], [srcq + widthq + 4], u,u |
| psubw m0, m2 |
| psubw m1, m3 |
| |
| paddw m0, [pw_16] |
| paddw m1, [pw_16] |
| psraw m0, 5 |
| psraw m1, 5 |
| packuswb m0, m1 |
| mova [dstq + widthq], m0 |
| sub widthd, mmsize |
| jge .loop |
| RET |
| %endmacro |
| |
| %macro PUT_RECT 1 |
| ; void put_rect_clamped(uint8_t *dst, int dst_stride, int16_t *src, int src_stride, int width, int height) |
| cglobal put_signed_rect_clamped_%1, 5,9,3, dst, dst_stride, src, src_stride, w, dst2, src2 |
| mova m0, [pb_80] |
| add wd, (mmsize-1) |
| and wd, ~(mmsize-1) |
| |
| %if ARCH_X86_64 |
| movsxd dst_strideq, dst_strided |
| movsxd src_strideq, src_strided |
| mov r7d, r5m |
| mov r8d, wd |
| %define wspill r8d |
| %define hd r7d |
| %else |
| mov r4m, wd |
| %define wspill r4m |
| %define hd r5mp |
| %endif |
| |
| .loopy: |
| lea src2q, [srcq+src_strideq] |
| lea dst2q, [dstq+dst_strideq] |
| .loopx: |
| sub wd, mmsize |
| mova m1, [srcq +2*wq] |
| mova m2, [src2q+2*wq] |
| packsswb m1, [srcq +2*wq+mmsize] |
| packsswb m2, [src2q+2*wq+mmsize] |
| paddb m1, m0 |
| paddb m2, m0 |
| mova [dstq +wq], m1 |
| mova [dst2q+wq], m2 |
| jg .loopx |
| |
| lea srcq, [srcq+src_strideq*2] |
| lea dstq, [dstq+dst_strideq*2] |
| sub hd, 2 |
| mov wd, wspill |
| jg .loopy |
| RET |
| %endm |
| |
| %macro ADD_RECT 1 |
| ; void add_rect_clamped(uint8_t *dst, uint16_t *src, int stride, int16_t *idwt, int idwt_stride, int width, int height) |
| cglobal add_rect_clamped_%1, 7,9,3, dst, src, stride, idwt, idwt_stride, w, h |
| mova m0, [pw_32] |
| add wd, (mmsize-1) |
| and wd, ~(mmsize-1) |
| |
| %if ARCH_X86_64 |
| movsxd strideq, strided |
| movsxd idwt_strideq, idwt_strided |
| mov r8d, wd |
| %define wspill r8d |
| %else |
| mov r5m, wd |
| %define wspill r5m |
| %endif |
| |
| .loop: |
| sub wd, mmsize |
| movu m1, [srcq +2*wq] ; FIXME: ensure alignment |
| paddw m1, m0 |
| psraw m1, 6 |
| movu m2, [srcq +2*wq+mmsize] ; FIXME: ensure alignment |
| paddw m2, m0 |
| psraw m2, 6 |
| paddw m1, [idwtq+2*wq] |
| paddw m2, [idwtq+2*wq+mmsize] |
| packuswb m1, m2 |
| mova [dstq +wq], m1 |
| jg .loop |
| |
| lea srcq, [srcq + 2*strideq] |
| add dstq, strideq |
| lea idwtq, [idwtq+ 2*idwt_strideq] |
| sub hd, 1 |
| mov wd, wspill |
| jg .loop |
| RET |
| %endm |
| |
| %macro ADD_OBMC 2 |
| ; void add_obmc(uint16_t *dst, uint8_t *src, int stride, uint8_t *obmc_weight, int yblen) |
| cglobal add_dirac_obmc%1_%2, 6,6,5, dst, src, stride, obmc, yblen |
| pxor m4, m4 |
| .loop: |
| %assign i 0 |
| %rep %1 / mmsize |
| mova m0, [srcq+i] |
| mova m1, m0 |
| punpcklbw m0, m4 |
| punpckhbw m1, m4 |
| mova m2, [obmcq+i] |
| mova m3, m2 |
| punpcklbw m2, m4 |
| punpckhbw m3, m4 |
| pmullw m0, m2 |
| pmullw m1, m3 |
| movu m2, [dstq+2*i] |
| movu m3, [dstq+2*i+mmsize] |
| paddw m0, m2 |
| paddw m1, m3 |
| movu [dstq+2*i], m0 |
| movu [dstq+2*i+mmsize], m1 |
| %assign i i+mmsize |
| %endrep |
| lea srcq, [srcq+strideq] |
| lea dstq, [dstq+2*strideq] |
| add obmcq, 32 |
| sub yblend, 1 |
| jg .loop |
| RET |
| %endm |
| |
| INIT_MMX |
| %if ARCH_X86_64 == 0 |
| PUT_RECT mmx |
| ADD_RECT mmx |
| |
| HPEL_FILTER mmx |
| ADD_OBMC 32, mmx |
| ADD_OBMC 16, mmx |
| %endif |
| ADD_OBMC 8, mmx |
| |
| INIT_XMM |
| PUT_RECT sse2 |
| ADD_RECT sse2 |
| |
| HPEL_FILTER sse2 |
| ADD_OBMC 32, sse2 |
| ADD_OBMC 16, sse2 |
| |
| INIT_XMM sse4 |
| |
| ; void dequant_subband_32(uint8_t *src, uint8_t *dst, ptrdiff_t stride, const int qf, const int qs, int tot_v, int tot_h) |
| cglobal dequant_subband_32, 7, 7, 4, src, dst, stride, qf, qs, tot_v, tot_h |
| movd m2, qfd |
| movd m3, qsd |
| SPLATD m2 |
| SPLATD m3 |
| mov r4d, tot_hd |
| mov r3, dstq |
| |
| .loop_v: |
| mov tot_hq, r4 |
| mov dstq, r3 |
| |
| .loop_h: |
| movu m0, [srcq] |
| |
| pabsd m1, m0 |
| pmulld m1, m2 |
| paddd m1, m3 |
| psrld m1, 2 |
| psignd m1, m0 |
| |
| movu [dstq], m1 |
| |
| add srcq, mmsize |
| add dstq, mmsize |
| sub tot_hq, 4 |
| jg .loop_h |
| lea srcq, [srcq + 4*tot_hq] |
| |
| add r3, strideq |
| dec tot_vd |
| jg .loop_v |
| |
| RET |
| |
| INIT_XMM sse4 |
| ; void put_signed_rect_clamped_10(uint8_t *dst, int dst_stride, const uint8_t *src, int src_stride, int width, int height) |
| %if ARCH_X86_64 |
| cglobal put_signed_rect_clamped_10, 6, 8, 5, dst, dst_stride, src, src_stride, w, h, t1, t2 |
| %else |
| cglobal put_signed_rect_clamped_10, 5, 7, 5, dst, dst_stride, src, src_stride, w, t1, t2 |
| %define hd r5mp |
| %endif |
| shl wd, 2 |
| add srcq, wq |
| neg wq |
| mov t2q, dstq |
| mov t1q, wq |
| pxor m2, m2 |
| mova m3, [clip_10bit] |
| mova m4, [convert_to_unsigned_10bit] |
| |
| .loop_h: |
| mov dstq, t2q |
| mov wq, t1q |
| |
| .loop_w: |
| movu m0, [srcq+wq+0*mmsize] |
| movu m1, [srcq+wq+1*mmsize] |
| |
| paddd m0, m4 |
| paddd m1, m4 |
| packusdw m0, m0, m1 |
| CLIPW m0, m2, m3 ; packusdw saturates so it's fine |
| |
| movu [dstq], m0 |
| |
| add dstq, 1*mmsize |
| add wq, 2*mmsize |
| jl .loop_w |
| |
| add srcq, src_strideq |
| add t2q, dst_strideq |
| sub hd, 1 |
| jg .loop_h |
| |
| RET |