/* * quarterpel DSP function templates * Copyright (c) 2000, 2001 Fabrice Bellard * Copyright (c) 2002-2004 Michael Niedermayer <michaelni@gmx.at> * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ /** * @file * quarterpel DSP function templates */ #define PIXOP2(OPNAME, OP) \ static inline void OPNAME ## _no_rnd_pixels8_l2_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int h) \ { \ int i; \ \ for (i = 0; i < h; i++) { \ uint32_t a, b; \ a = AV_RN32(&src1[i * src_stride1]); \ b = AV_RN32(&src2[i * src_stride2]); \ OP(*((uint32_t *) &dst[i * dst_stride]), \ no_rnd_avg32(a, b)); \ a = AV_RN32(&src1[i * src_stride1 + 4]); \ b = AV_RN32(&src2[i * src_stride2 + 4]); \ OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ no_rnd_avg32(a, b)); \ } \ } \ \ static inline void OPNAME ## _no_rnd_pixels16_l2_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int h) \ { \ OPNAME ## _no_rnd_pixels8_l2_8(dst, src1, src2, dst_stride, \ src_stride1, src_stride2, h); \ OPNAME ## _no_rnd_pixels8_l2_8(dst + 8, \ src1 + 8, \ src2 + 8, \ dst_stride, src_stride1, \ src_stride2, h); \ } \ \ static inline void OPNAME ## _pixels8_l4_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ const uint8_t *src3, \ const uint8_t *src4, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int src_stride3, \ int src_stride4, \ int h) \ { \ /* FIXME HIGH BIT DEPTH */ \ int i; \ \ for (i = 0; i < h; i++) { \ uint32_t a, b, c, d, l0, l1, h0, h1; \ a = AV_RN32(&src1[i * src_stride1]); \ b = AV_RN32(&src2[i * src_stride2]); \ c = AV_RN32(&src3[i * src_stride3]); \ d = AV_RN32(&src4[i * src_stride4]); \ l0 = (a & 0x03030303UL) + \ (b & 0x03030303UL) + \ 0x02020202UL; \ h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ ((b & 0xFCFCFCFCUL) >> 2); \ l1 = (c & 0x03030303UL) + \ (d & 0x03030303UL); \ h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ ((d & 0xFCFCFCFCUL) >> 2); \ OP(*((uint32_t *) &dst[i * dst_stride]), \ h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ a = AV_RN32(&src1[i * src_stride1 + 4]); \ b = AV_RN32(&src2[i * src_stride2 + 4]); \ c = AV_RN32(&src3[i * src_stride3 + 4]); \ d = AV_RN32(&src4[i * src_stride4 + 4]); \ l0 = (a & 0x03030303UL) + \ (b & 0x03030303UL) + \ 0x02020202UL; \ h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ ((b & 0xFCFCFCFCUL) >> 2); \ l1 = (c & 0x03030303UL) + \ (d & 0x03030303UL); \ h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ ((d & 0xFCFCFCFCUL) >> 2); \ OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ } \ } \ \ static inline void OPNAME ## _no_rnd_pixels8_l4_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ const uint8_t *src3, \ const uint8_t *src4, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int src_stride3, \ int src_stride4, \ int h) \ { \ /* FIXME HIGH BIT DEPTH */ \ int i; \ \ for (i = 0; i < h; i++) { \ uint32_t a, b, c, d, l0, l1, h0, h1; \ a = AV_RN32(&src1[i * src_stride1]); \ b = AV_RN32(&src2[i * src_stride2]); \ c = AV_RN32(&src3[i * src_stride3]); \ d = AV_RN32(&src4[i * src_stride4]); \ l0 = (a & 0x03030303UL) + \ (b & 0x03030303UL) + \ 0x01010101UL; \ h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ ((b & 0xFCFCFCFCUL) >> 2); \ l1 = (c & 0x03030303UL) + \ (d & 0x03030303UL); \ h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ ((d & 0xFCFCFCFCUL) >> 2); \ OP(*((uint32_t *) &dst[i * dst_stride]), \ h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ a = AV_RN32(&src1[i * src_stride1 + 4]); \ b = AV_RN32(&src2[i * src_stride2 + 4]); \ c = AV_RN32(&src3[i * src_stride3 + 4]); \ d = AV_RN32(&src4[i * src_stride4 + 4]); \ l0 = (a & 0x03030303UL) + \ (b & 0x03030303UL) + \ 0x01010101UL; \ h0 = ((a & 0xFCFCFCFCUL) >> 2) + \ ((b & 0xFCFCFCFCUL) >> 2); \ l1 = (c & 0x03030303UL) + \ (d & 0x03030303UL); \ h1 = ((c & 0xFCFCFCFCUL) >> 2) + \ ((d & 0xFCFCFCFCUL) >> 2); \ OP(*((uint32_t *) &dst[i * dst_stride + 4]), \ h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL)); \ } \ } \ \ static inline void OPNAME ## _pixels16_l4_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ const uint8_t *src3, \ const uint8_t *src4, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int src_stride3, \ int src_stride4, \ int h) \ { \ OPNAME ## _pixels8_l4_8(dst, src1, src2, src3, src4, dst_stride, \ src_stride1, src_stride2, src_stride3, \ src_stride4, h); \ OPNAME ## _pixels8_l4_8(dst + 8, \ src1 + 8, src2 + 8, \ src3 + 8, src4 + 8, \ dst_stride, src_stride1, src_stride2, \ src_stride3, src_stride4, h); \ } \ \ static inline void OPNAME ## _no_rnd_pixels16_l4_8(uint8_t *dst, \ const uint8_t *src1, \ const uint8_t *src2, \ const uint8_t *src3, \ const uint8_t *src4, \ int dst_stride, \ int src_stride1, \ int src_stride2, \ int src_stride3, \ int src_stride4, \ int h) \ { \ OPNAME ## _no_rnd_pixels8_l4_8(dst, src1, src2, src3, src4, \ dst_stride, src_stride1, \ src_stride2, src_stride3, \ src_stride4, h); \ OPNAME ## _no_rnd_pixels8_l4_8(dst + 8, \ src1 + 8, src2 + 8, \ src3 + 8, src4 + 8, \ dst_stride, src_stride1, \ src_stride2, src_stride3, \ src_stride4, h); \ } \ #define op_avg(a, b) a = rnd_avg32(a, b) #define op_put(a, b) a = b #define put_no_rnd_pixels8_8_c put_pixels8_8_c PIXOP2(avg, op_avg) PIXOP2(put, op_put) #undef op_avg #undef op_put