| /* |
| * quarterpel DSP function templates |
| * Copyright (c) 2000, 2001 Fabrice Bellard |
| * Copyright (c) 2002-2004 Michael Niedermayer <michaelni@gmx.at> |
| * |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| /** |
| * @file |
| * quarterpel DSP function templates |
| */ |
| |
| #define PIXOP2(OPNAME, OP) \ |
| static inline void OPNAME ## _no_rnd_pixels8_l2_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int h) \ |
| { \ |
| int i; \ |
| \ |
| for (i = 0; i < h; i++) { \ |
| uint32_t a, b; \ |
| a = AV_RN32(&src1[i * src_stride1]); \ |
| b = AV_RN32(&src2[i * src_stride2]); \ |
| OP(*((uint32_t *) &dst[i * dst_stride]), \ |
| no_rnd_avg32(a, b)); \ |
| a = AV_RN32(&src1[i * src_stride1 + 4]); \ |
| b = AV_RN32(&src2[i * src_stride2 + 4]); \ |
| OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ |
| no_rnd_avg32(a, b)); \ |
| } \ |
| } \ |
| \ |
| static inline void OPNAME ## _no_rnd_pixels16_l2_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int h) \ |
| { \ |
| OPNAME ## _no_rnd_pixels8_l2_8(dst, src1, src2, dst_stride, \ |
| src_stride1, src_stride2, h); \ |
| OPNAME ## _no_rnd_pixels8_l2_8(dst + 8, \ |
| src1 + 8, \ |
| src2 + 8, \ |
| dst_stride, src_stride1, \ |
| src_stride2, h); \ |
| } \ |
| \ |
| static inline void OPNAME ## _pixels8_l4_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| const uint8_t *src3, \ |
| const uint8_t *src4, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int src_stride3, \ |
| int src_stride4, \ |
| int h) \ |
| { \ |
| /* FIXME HIGH BIT DEPTH */ \ |
| int i; \ |
| \ |
| for (i = 0; i < h; i++) { \ |
| uint32_t a, b, c, d, l0, l1, h0, h1; \ |
| a = AV_RN32(&src1[i * src_stride1]); \ |
| b = AV_RN32(&src2[i * src_stride2]); \ |
| c = AV_RN32(&src3[i * src_stride3]); \ |
| d = AV_RN32(&src4[i * src_stride4]); \ |
| l0 = (a & 0x03030303UL) + \ |
| (b & 0x03030303UL) + \ |
| 0x02020202UL; \ |
| h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ |
| ((b & 0xFCFCFCFCUL) >> 2); \ |
| l1 = (c & 0x03030303UL) + \ |
| (d & 0x03030303UL); \ |
| h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ |
| ((d & 0xFCFCFCFCUL) >> 2); \ |
| OP(*((uint32_t *) &dst[i * dst_stride]), \ |
| h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ |
| a = AV_RN32(&src1[i * src_stride1 + 4]); \ |
| b = AV_RN32(&src2[i * src_stride2 + 4]); \ |
| c = AV_RN32(&src3[i * src_stride3 + 4]); \ |
| d = AV_RN32(&src4[i * src_stride4 + 4]); \ |
| l0 = (a & 0x03030303UL) + \ |
| (b & 0x03030303UL) + \ |
| 0x02020202UL; \ |
| h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ |
| ((b & 0xFCFCFCFCUL) >> 2); \ |
| l1 = (c & 0x03030303UL) + \ |
| (d & 0x03030303UL); \ |
| h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ |
| ((d & 0xFCFCFCFCUL) >> 2); \ |
| OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ |
| h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ |
| } \ |
| } \ |
| \ |
| static inline void OPNAME ## _no_rnd_pixels8_l4_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| const uint8_t *src3, \ |
| const uint8_t *src4, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int src_stride3, \ |
| int src_stride4, \ |
| int h) \ |
| { \ |
| /* FIXME HIGH BIT DEPTH */ \ |
| int i; \ |
| \ |
| for (i = 0; i < h; i++) { \ |
| uint32_t a, b, c, d, l0, l1, h0, h1; \ |
| a = AV_RN32(&src1[i * src_stride1]); \ |
| b = AV_RN32(&src2[i * src_stride2]); \ |
| c = AV_RN32(&src3[i * src_stride3]); \ |
| d = AV_RN32(&src4[i * src_stride4]); \ |
| l0 = (a & 0x03030303UL) + \ |
| (b & 0x03030303UL) + \ |
| 0x01010101UL; \ |
| h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ |
| ((b & 0xFCFCFCFCUL) >> 2); \ |
| l1 = (c & 0x03030303UL) + \ |
| (d & 0x03030303UL); \ |
| h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ |
| ((d & 0xFCFCFCFCUL) >> 2); \ |
| OP(*((uint32_t *) &dst[i * dst_stride]), \ |
| h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ |
| a = AV_RN32(&src1[i * src_stride1 + 4]); \ |
| b = AV_RN32(&src2[i * src_stride2 + 4]); \ |
| c = AV_RN32(&src3[i * src_stride3 + 4]); \ |
| d = AV_RN32(&src4[i * src_stride4 + 4]); \ |
| l0 = (a & 0x03030303UL) + \ |
| (b & 0x03030303UL) + \ |
| 0x01010101UL; \ |
| h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ |
| ((b & 0xFCFCFCFCUL) >> 2); \ |
| l1 = (c & 0x03030303UL) + \ |
| (d & 0x03030303UL); \ |
| h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ |
| ((d & 0xFCFCFCFCUL) >> 2); \ |
| OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ |
| h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ |
| } \ |
| } \ |
| \ |
| static inline void OPNAME ## _pixels16_l4_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| const uint8_t *src3, \ |
| const uint8_t *src4, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int src_stride3, \ |
| int src_stride4, \ |
| int h) \ |
| { \ |
| OPNAME ## _pixels8_l4_8(dst, src1, src2, src3, src4, dst_stride, \ |
| src_stride1, src_stride2, src_stride3, \ |
| src_stride4, h); \ |
| OPNAME ## _pixels8_l4_8(dst + 8, \ |
| src1 + 8, src2 + 8, \ |
| src3 + 8, src4 + 8, \ |
| dst_stride, src_stride1, src_stride2, \ |
| src_stride3, src_stride4, h); \ |
| } \ |
| \ |
| static inline void OPNAME ## _no_rnd_pixels16_l4_8(uint8_t *dst, \ |
| const uint8_t *src1, \ |
| const uint8_t *src2, \ |
| const uint8_t *src3, \ |
| const uint8_t *src4, \ |
| int dst_stride, \ |
| int src_stride1, \ |
| int src_stride2, \ |
| int src_stride3, \ |
| int src_stride4, \ |
| int h) \ |
| { \ |
| OPNAME ## _no_rnd_pixels8_l4_8(dst, src1, src2, src3, src4, \ |
| dst_stride, src_stride1, \ |
| src_stride2, src_stride3, \ |
| src_stride4, h); \ |
| OPNAME ## _no_rnd_pixels8_l4_8(dst + 8, \ |
| src1 + 8, src2 + 8, \ |
| src3 + 8, src4 + 8, \ |
| dst_stride, src_stride1, \ |
| src_stride2, src_stride3, \ |
| src_stride4, h); \ |
| } \ |
| |
| #define op_avg(a, b) a = rnd_avg32(a, b) |
| #define op_put(a, b) a = b |
| #define put_no_rnd_pixels8_8_c put_pixels8_8_c |
| PIXOP2(avg, op_avg) |
| PIXOP2(put, op_put) |
| #undef op_avg |
| #undef op_put |