blob: 906cbf3fc832f84ca449a5a920d9056c919ac042 [file] [log] [blame]
;//
;// Copyright (C) 2007-2008 ARM Limited
;//
;// Licensed under the Apache License, Version 2.0 (the "License");
;// you may not use this file except in compliance with the License.
;// You may obtain a copy of the License at
;//
;// http://www.apache.org/licenses/LICENSE-2.0
;//
;// Unless required by applicable law or agreed to in writing, software
;// distributed under the License is distributed on an "AS IS" BASIS,
;// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
;// See the License for the specific language governing permissions and
;// limitations under the License.
;//
;//
;//
;// File Name: armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe_s.s
;// OpenMAX DL: v1.0.2
;// Revision: 9641
;// Date: Thursday, February 7, 2008
;//
;//
;//
;//
INCLUDE omxtypes_s.h
INCLUDE armCOMM_s.h
M_VARIANTS ARM1136JS
EXPORT armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe
IF ARM1136JS
;// Function header
;// Function:
;// armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe
;//
;// Implements vertical interpolation for a block of size 4x4. Input and output should
;// be aligned.
;//
;// Registers used as input for this function
;// r0,r1,r2,r3 where r0,r2 input pointer and r1,r3 corresponding step size
;//
;// Registers preserved for top level function
;// r0,r1,r2,r3,r4,r5,r6,r14
;//
;// Registers modified by the function
;// r7,r8,r9,r10,r11,r12
;//
;// Output registers
;// None. Function will preserve r0-r3
M_START armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe, r6
;// Declare input registers
pSrc RN 0
srcStep RN 1
pDst RN 2
dstStep RN 3
;// Declare inner loop registers
ValA RN 5
ValA0 RN 4
ValA1 RN 5
ValAF0 RN 4
ValAF1 RN 5
ValB RN 11
ValC RN 5
ValC0 RN 4
ValC1 RN 5
ValCD0 RN 12
ValCD1 RN 14
ValCF0 RN 4
ValCF1 RN 5
ValD RN 10
ValE RN 7
ValE0 RN 6
ValE1 RN 7
ValEB0 RN 10
ValEB1 RN 11
ValED0 RN 6
ValED1 RN 7
ValF RN 10
ValG RN 14
ValG0 RN 12
ValG1 RN 14
ValGB0 RN 12
ValGB1 RN 14
Acc0 RN 4
Acc1 RN 5
Acc2 RN 6
Acc3 RN 7
Temp RN 7
Height RN 3
Step RN 6
Counter RN 8
r0x00ff00ff RN 9 ;// [0 255 0 255] where 255 is offset
r0x0fe00fe0 RN 10 ;// [0 (16*255 - 16) 0 (16*255 - 16)]
LDR r0x00ff00ff, =0x00ff00ff ;// [0 255 0 255] 255 is offset to avoid negative results
MOV Counter, #2
TwoRowsLoop
M_LDR ValC, [pSrc], srcStep ;// Load [c3 c2 c1 c0]
M_LDR ValD, [pSrc], srcStep ;// Load [d3 d2 d1 d0]
M_LDR ValE, [pSrc], srcStep ;// Load [e3 e2 e1 e0]
SUB pSrc, pSrc, srcStep, LSL #2
LDR ValB, [pSrc] ;// Load [b3 b2 b1 b0]
UXTAB16 ValC0, r0x00ff00ff, ValC ;// [0 c2 0 c0] + [0 255 0 255]
UXTAB16 ValC1, r0x00ff00ff, ValC, ROR #8 ;// [0 c3 0 c1] + [0 255 0 255]
UXTAB16 ValE0, r0x00ff00ff, ValE ;// [0 e2 0 e0] + [0 255 0 255]
UXTAB16 ValE1, r0x00ff00ff, ValE, ROR #8 ;// [0 e3 0 e1] + [0 255 0 255]
UXTAB16 ValCD0, ValC0, ValD ;// [0 c2 0 c0] + [0 255 0 255] + [0 d2 0 d0]
UXTAB16 ValCD1, ValC1, ValD, ROR #8 ;// [0 c3 0 c1] + [0 255 0 255] + [0 d3 0 d1]
UXTAB16 ValEB0, ValE0, ValB ;// [0 e2 0 e0] + [0 255 0 255] + [0 b2 0 b0]
RSB ValCD0, ValEB0, ValCD0, LSL #2 ;// 4*(Off+C+D) - (Off+B+E)
LDR ValD, [pSrc, srcStep, LSL #1] ;// Load [d3 d2 d1 d0]
UXTAB16 ValEB1, ValE1, ValB, ROR #8 ;// [0 e3 0 e1] + [0 255 0 255] + [0 b3 0 b1]
RSB ValCD1, ValEB1, ValCD1, LSL #2
;// One cycle stall
UXTAB16 ValED0, ValE0, ValD ;// [0 e2 0 e0] + [0 255 0 255] + [0 d2 0 d0]
UXTAB16 ValED1, ValE1, ValD, ROR #8 ;// [0 e3 0 e1] + [0 255 0 255] + [0 d3 0 d1]
LDR ValF, [pSrc, srcStep, LSL #2] ;// Load [f3 f2 f1 f0]
M_LDR ValB, [pSrc], srcStep ;// Load [b3 b2 b1 b0]
ADD ValCD0, ValCD0, ValCD0, LSL #2 ;// 5 * [4*(Off+C+D) - (Off+B+E)]
ADD ValCD1, ValCD1, ValCD1, LSL #2
UXTAB16 ValCF1, ValC1, ValF, ROR #8 ;// [0 c3 0 c1] + [0 255 0 255] + [0 f3 0 f1]
UXTAB16 ValCF0, ValC0, ValF ;// [0 c2 0 c0] + [0 255 0 255] + [0 f2 0 f0]
RSB ValED1, ValCF1, ValED1, LSL #2
SUB ValA, pSrc, srcStep, LSL #1
LDR ValA, [ValA] ;// Load [a3 a2 a1 a0]
RSB ValED0, ValCF0, ValED0, LSL #2 ;// 4*(Off+E+D) - (Off+C+F)
ADD ValED1, ValED1, ValED1, LSL #2
ADD ValED0, ValED0, ValED0, LSL #2 ;// 5 * [4*(Off+E+D) - (Off+C+F)]
UXTAB16 ValA0, r0x00ff00ff, ValA ;// [0 a2 0 a0] + [0 255 0 255]
UXTAB16 ValA1, r0x00ff00ff, ValA, ROR #8 ;// [0 a3 0 a1] + [0 255 0 255]
UXTAB16 ValAF0, ValA0, ValF ;// [0 a2 0 a0] + [0 255 0 255] + [0 f2 0 f0]
UXTAB16 ValAF1, ValA1, ValF, ROR #8 ;// [0 a3 0 a1] + [0 255 0 255] + [0 f3 0 f1]
LDR r0x0fe00fe0, =0x0fe00fe0 ;// [0 255 0 255] 255 is offset to avoid negative results
ADD Acc1, ValCD1, ValAF1
LDR ValG, [pSrc, srcStep, LSL #2] ;// Load [g3 g2 g1 g0]
ADD Acc0, ValCD0, ValAF0 ;// Acc0 = 16*Off + (A+F) + 20*(C+D) - 5*(B+E)
UQSUB16 Acc1, Acc1, r0x0fe00fe0 ;// Acc1 -= (16*Off - 16)
UQSUB16 Acc0, Acc0, r0x0fe00fe0
UXTAB16 ValG0, r0x00ff00ff, ValG ;// [0 g2 0 g0] + [0 255 0 255]
UXTAB16 ValG1, r0x00ff00ff, ValG, ROR #8 ;// [0 g3 0 g1] + [0 255 0 255]
UXTAB16 ValGB0, ValG0, ValB ;// [0 g2 0 g0] + [0 255 0 255] + [0 b2 0 b0]
UXTAB16 ValGB1, ValG1, ValB, ROR #8 ;// [0 g3 0 g1] + [0 255 0 255] + [0 b3 0 b1]
ADD Acc2, ValED0, ValGB0 ;// Acc2 = 16*Off + (B+G) + 20*(D+E) - 5*(C+F)
ADD Acc3, ValED1, ValGB1
UQSUB16 Acc3, Acc3, r0x0fe00fe0 ;// Acc3 -= (16*Off - 16)
UQSUB16 Acc2, Acc2, r0x0fe00fe0
USAT16 Acc1, #13, Acc1 ;// Saturate to 8+5 = 13 bits
USAT16 Acc0, #13, Acc0
USAT16 Acc3, #13, Acc3
USAT16 Acc2, #13, Acc2
AND Acc1, r0x00ff00ff, Acc1, LSR #5 ;// [0 a3 0 a1]
AND Acc0, r0x00ff00ff, Acc0, LSR #5 ;// [0 a2 0 a0]
ORR Acc0, Acc0, Acc1, LSL #8 ;// [a3 a2 a1 a0]
AND Acc3, r0x00ff00ff, Acc3, LSR #5 ;// [0 b3 0 b1]
AND Acc2, r0x00ff00ff, Acc2, LSR #5 ;// [0 b2 0 b0]
M_STR Acc0, [pDst], dstStep ;// Store result & adjust pointer
ORR Acc2, Acc2, Acc3, LSL #8 ;// [b3 b2 b1 b0]
M_STR Acc2, [pDst], dstStep ;// Store result & adjust pointer
ADD pSrc, pSrc, srcStep, LSL #1
SUBS Counter, Counter, #1
BGT TwoRowsLoop
End
SUB pDst, pDst, dstStep, LSL #2
SUB pSrc, pSrc, srcStep, LSL #2
M_END
ENDIF
END