| /* |
| * Copyright (c) 2009 Mans Rullgard <mans@mansr.com> |
| * |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include "libavutil/arm/asm.S" |
| |
| function ff_pix_abs16_armv6, export=1 |
| ldr r0, [sp] |
| push {r4-r9, lr} |
| mov r12, #0 |
| mov lr, #0 |
| ldm r1, {r4-r7} |
| ldr r8, [r2] |
| 1: |
| ldr r9, [r2, #4] |
| pld [r1, r3] |
| usada8 r12, r4, r8, r12 |
| ldr r8, [r2, #8] |
| pld [r2, r3] |
| usada8 lr, r5, r9, lr |
| ldr r9, [r2, #12] |
| usada8 r12, r6, r8, r12 |
| subs r0, r0, #1 |
| usada8 lr, r7, r9, lr |
| beq 2f |
| add r1, r1, r3 |
| ldm r1, {r4-r7} |
| add r2, r2, r3 |
| ldr r8, [r2] |
| b 1b |
| 2: |
| add r0, r12, lr |
| pop {r4-r9, pc} |
| endfunc |
| |
| function ff_pix_abs16_x2_armv6, export=1 |
| ldr r12, [sp] |
| push {r4-r11, lr} |
| mov r0, #0 |
| mov lr, #1 |
| orr lr, lr, lr, lsl #8 |
| orr lr, lr, lr, lsl #16 |
| 1: |
| ldr r8, [r2] |
| ldr r9, [r2, #4] |
| lsr r10, r8, #8 |
| ldr r4, [r1] |
| lsr r6, r9, #8 |
| orr r10, r10, r9, lsl #24 |
| ldr r5, [r2, #8] |
| eor r11, r8, r10 |
| uhadd8 r7, r8, r10 |
| orr r6, r6, r5, lsl #24 |
| and r11, r11, lr |
| uadd8 r7, r7, r11 |
| ldr r8, [r1, #4] |
| usada8 r0, r4, r7, r0 |
| eor r7, r9, r6 |
| lsr r10, r5, #8 |
| and r7, r7, lr |
| uhadd8 r4, r9, r6 |
| ldr r6, [r2, #12] |
| uadd8 r4, r4, r7 |
| pld [r1, r3] |
| orr r10, r10, r6, lsl #24 |
| usada8 r0, r8, r4, r0 |
| ldr r4, [r1, #8] |
| eor r11, r5, r10 |
| ldrb r7, [r2, #16] |
| and r11, r11, lr |
| uhadd8 r8, r5, r10 |
| ldr r5, [r1, #12] |
| uadd8 r8, r8, r11 |
| pld [r2, r3] |
| lsr r10, r6, #8 |
| usada8 r0, r4, r8, r0 |
| orr r10, r10, r7, lsl #24 |
| subs r12, r12, #1 |
| eor r11, r6, r10 |
| add r1, r1, r3 |
| uhadd8 r9, r6, r10 |
| and r11, r11, lr |
| uadd8 r9, r9, r11 |
| add r2, r2, r3 |
| usada8 r0, r5, r9, r0 |
| bgt 1b |
| |
| pop {r4-r11, pc} |
| endfunc |
| |
| .macro usad_y2 p0, p1, p2, p3, n0, n1, n2, n3 |
| ldr \n0, [r2] |
| eor \n1, \p0, \n0 |
| uhadd8 \p0, \p0, \n0 |
| and \n1, \n1, lr |
| ldr \n2, [r1] |
| uadd8 \p0, \p0, \n1 |
| ldr \n1, [r2, #4] |
| usada8 r0, \p0, \n2, r0 |
| pld [r1, r3] |
| eor \n3, \p1, \n1 |
| uhadd8 \p1, \p1, \n1 |
| and \n3, \n3, lr |
| ldr \p0, [r1, #4] |
| uadd8 \p1, \p1, \n3 |
| ldr \n2, [r2, #8] |
| usada8 r0, \p1, \p0, r0 |
| pld [r2, r3] |
| eor \p0, \p2, \n2 |
| uhadd8 \p2, \p2, \n2 |
| and \p0, \p0, lr |
| ldr \p1, [r1, #8] |
| uadd8 \p2, \p2, \p0 |
| ldr \n3, [r2, #12] |
| usada8 r0, \p2, \p1, r0 |
| eor \p1, \p3, \n3 |
| uhadd8 \p3, \p3, \n3 |
| and \p1, \p1, lr |
| ldr \p0, [r1, #12] |
| uadd8 \p3, \p3, \p1 |
| add r1, r1, r3 |
| usada8 r0, \p3, \p0, r0 |
| add r2, r2, r3 |
| .endm |
| |
| function ff_pix_abs16_y2_armv6, export=1 |
| pld [r1] |
| pld [r2] |
| ldr r12, [sp] |
| push {r4-r11, lr} |
| mov r0, #0 |
| mov lr, #1 |
| orr lr, lr, lr, lsl #8 |
| orr lr, lr, lr, lsl #16 |
| ldr r4, [r2] |
| ldr r5, [r2, #4] |
| ldr r6, [r2, #8] |
| ldr r7, [r2, #12] |
| add r2, r2, r3 |
| 1: |
| usad_y2 r4, r5, r6, r7, r8, r9, r10, r11 |
| subs r12, r12, #2 |
| usad_y2 r8, r9, r10, r11, r4, r5, r6, r7 |
| bgt 1b |
| |
| pop {r4-r11, pc} |
| endfunc |
| |
| function ff_pix_abs8_armv6, export=1 |
| pld [r2, r3] |
| ldr r12, [sp] |
| push {r4-r9, lr} |
| mov r0, #0 |
| mov lr, #0 |
| ldrd_post r4, r5, r1, r3 |
| 1: |
| subs r12, r12, #2 |
| ldr r7, [r2, #4] |
| ldr_post r6, r2, r3 |
| ldrd_post r8, r9, r1, r3 |
| usada8 r0, r4, r6, r0 |
| pld [r2, r3] |
| usada8 lr, r5, r7, lr |
| ldr r7, [r2, #4] |
| ldr_post r6, r2, r3 |
| beq 2f |
| ldrd_post r4, r5, r1, r3 |
| usada8 r0, r8, r6, r0 |
| pld [r2, r3] |
| usada8 lr, r9, r7, lr |
| b 1b |
| 2: |
| usada8 r0, r8, r6, r0 |
| usada8 lr, r9, r7, lr |
| add r0, r0, lr |
| pop {r4-r9, pc} |
| endfunc |
| |
| function ff_sse16_armv6, export=1 |
| ldr r12, [sp] |
| push {r4-r9, lr} |
| mov r0, #0 |
| 1: |
| ldrd r4, r5, [r1] |
| ldr r8, [r2] |
| uxtb16 lr, r4 |
| uxtb16 r4, r4, ror #8 |
| uxtb16 r9, r8 |
| uxtb16 r8, r8, ror #8 |
| ldr r7, [r2, #4] |
| usub16 lr, lr, r9 |
| usub16 r4, r4, r8 |
| smlad r0, lr, lr, r0 |
| uxtb16 r6, r5 |
| uxtb16 lr, r5, ror #8 |
| uxtb16 r8, r7 |
| uxtb16 r9, r7, ror #8 |
| smlad r0, r4, r4, r0 |
| ldrd r4, r5, [r1, #8] |
| usub16 r6, r6, r8 |
| usub16 r8, lr, r9 |
| ldr r7, [r2, #8] |
| smlad r0, r6, r6, r0 |
| uxtb16 lr, r4 |
| uxtb16 r4, r4, ror #8 |
| uxtb16 r9, r7 |
| uxtb16 r7, r7, ror #8 |
| smlad r0, r8, r8, r0 |
| ldr r8, [r2, #12] |
| usub16 lr, lr, r9 |
| usub16 r4, r4, r7 |
| smlad r0, lr, lr, r0 |
| uxtb16 r6, r5 |
| uxtb16 r5, r5, ror #8 |
| uxtb16 r9, r8 |
| uxtb16 r8, r8, ror #8 |
| smlad r0, r4, r4, r0 |
| usub16 r6, r6, r9 |
| usub16 r5, r5, r8 |
| smlad r0, r6, r6, r0 |
| add r1, r1, r3 |
| add r2, r2, r3 |
| subs r12, r12, #1 |
| smlad r0, r5, r5, r0 |
| bgt 1b |
| |
| pop {r4-r9, pc} |
| endfunc |