Commit 82eb4b0f authored by Zuxy Meng's avatar Zuxy Meng Committed by Corey Hickey

3DNow! & Extended 3DNow! versions of FFT

Patch by Zuxy Meng, zuxy <<dot>> meng >>at<< gmail <<dot>> com
Minor non-functional diff-related fixes by me.

Originally committed as revision 5125 to svn://svn.ffmpeg.org/ffmpeg/trunk
parent 8e321619
...@@ -1110,6 +1110,19 @@ if $cc -msse -o $TMPO $TMPC 2> /dev/null ; then ...@@ -1110,6 +1110,19 @@ if $cc -msse -o $TMPO $TMPC 2> /dev/null ; then
builtin_vector=yes builtin_vector=yes
fi fi
# test for mm3dnow.h
cat > $TMPC << EOF
#include <mm3dnow.h>
int main(void) {
return 0;
}
EOF
mm3dnow=no
if $cc -march=athlon -o $TMPO $TMPC 2> /dev/null ; then
mm3dnow=yes
fi
# Probe for -Wdeclaration-after-statement # Probe for -Wdeclaration-after-statement
if test "$cc" = "gcc"; then if test "$cc" = "gcc"; then
cat > $TMPC << EOF cat > $TMPC << EOF
...@@ -1328,6 +1341,7 @@ echo "broken inttypes.h $emu_fast_int" ...@@ -1328,6 +1341,7 @@ echo "broken inttypes.h $emu_fast_int"
if test $cpu = "x86" -o $cpu = "x86_64"; then if test $cpu = "x86" -o $cpu = "x86_64"; then
echo "MMX enabled $mmx" echo "MMX enabled $mmx"
echo "Vector Builtins $builtin_vector" echo "Vector Builtins $builtin_vector"
echo "3DNow! Builtins $mm3dnow"
fi fi
if test $cpu = "armv4l"; then if test $cpu = "armv4l"; then
echo "IWMMXT enabled $iwmmxt" echo "IWMMXT enabled $iwmmxt"
...@@ -1508,6 +1522,10 @@ if test "$builtin_vector" = "yes" ; then ...@@ -1508,6 +1522,10 @@ if test "$builtin_vector" = "yes" ; then
echo "TARGET_BUILTIN_VECTOR=yes" >> config.mak echo "TARGET_BUILTIN_VECTOR=yes" >> config.mak
echo "#define HAVE_BUILTIN_VECTOR 1" >> $TMPH echo "#define HAVE_BUILTIN_VECTOR 1" >> $TMPH
fi fi
if test "$mm3dnow" = "yes" ; then
echo "TARGET_BUILTIN_3DNOW=yes" >> config.mak
echo "#define HAVE_MM3DNOW 1" >> $TMPH
fi
if test "$iwmmxt" = "yes" ; then if test "$iwmmxt" = "yes" ; then
echo "TARGET_IWMMXT=yes" >> config.mak echo "TARGET_IWMMXT=yes" >> config.mak
echo "#define HAVE_IWMMXT 1" >> $TMPH echo "#define HAVE_IWMMXT 1" >> $TMPH
......
...@@ -330,7 +330,7 @@ OBJS += i386/fdct_mmx.o i386/cputest.o \ ...@@ -330,7 +330,7 @@ OBJS += i386/fdct_mmx.o i386/cputest.o \
i386/dsputil_mmx.o i386/mpegvideo_mmx.o \ i386/dsputil_mmx.o i386/mpegvideo_mmx.o \
i386/idct_mmx.o i386/motion_est_mmx.o \ i386/idct_mmx.o i386/motion_est_mmx.o \
i386/simple_idct_mmx.o i386/fft_sse.o i386/vp3dsp_mmx.o \ i386/simple_idct_mmx.o i386/fft_sse.o i386/vp3dsp_mmx.o \
i386/vp3dsp_sse2.o i386/vp3dsp_sse2.o i386/fft_3dn.o i386/fft_3dn2.o
ifeq ($(CONFIG_GPL),yes) ifeq ($(CONFIG_GPL),yes)
OBJS += i386/idct_mmx_xvid.o OBJS += i386/idct_mmx_xvid.o
endif endif
...@@ -338,6 +338,10 @@ ifdef TARGET_BUILTIN_VECTOR ...@@ -338,6 +338,10 @@ ifdef TARGET_BUILTIN_VECTOR
i386/fft_sse.o: CFLAGS+= -msse i386/fft_sse.o: CFLAGS+= -msse
depend: CFLAGS+= -msse depend: CFLAGS+= -msse
endif endif
ifdef TARGET_BUILTIN_3DNOW
i386/fft_3dn.o: CFLAGS+= -m3dnow
i386/fft_3dn2.o: CFLAGS+= -march=athlon
endif
endif endif
# armv4l specific stuff # armv4l specific stuff
......
...@@ -564,6 +564,8 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse); ...@@ -564,6 +564,8 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse);
void ff_fft_permute(FFTContext *s, FFTComplex *z); void ff_fft_permute(FFTContext *s, FFTComplex *z);
void ff_fft_calc_c(FFTContext *s, FFTComplex *z); void ff_fft_calc_c(FFTContext *s, FFTComplex *z);
void ff_fft_calc_sse(FFTContext *s, FFTComplex *z); void ff_fft_calc_sse(FFTContext *s, FFTComplex *z);
void ff_fft_calc_3dn(FFTContext *s, FFTComplex *z);
void ff_fft_calc_3dn2(FFTContext *s, FFTComplex *z);
void ff_fft_calc_altivec(FFTContext *s, FFTComplex *z); void ff_fft_calc_altivec(FFTContext *s, FFTComplex *z);
static inline void ff_fft_calc(FFTContext *s, FFTComplex *z) static inline void ff_fft_calc(FFTContext *s, FFTComplex *z)
......
...@@ -57,12 +57,12 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse) ...@@ -57,12 +57,12 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse)
s->exptab1 = NULL; s->exptab1 = NULL;
/* compute constant table for HAVE_SSE version */ /* compute constant table for HAVE_SSE version */
#if (defined(HAVE_MMX) && defined(HAVE_BUILTIN_VECTOR)) || defined(HAVE_ALTIVEC) #if (defined(HAVE_MMX) && (defined(HAVE_BUILTIN_VECTOR) || defined(HAVE_MM3DNOW))) || defined(HAVE_ALTIVEC)
{ {
int has_vectors = 0; int has_vectors = 0;
#if defined(HAVE_MMX) #if defined(HAVE_MMX)
has_vectors = mm_support() & MM_SSE; has_vectors = mm_support() & (MM_3DNOW | MM_3DNOWEXT | MM_SSE | MM_SSE2);
#endif #endif
#if defined(HAVE_ALTIVEC) && !defined(ALTIVEC_USE_REFERENCE_C_CODE) #if defined(HAVE_ALTIVEC) && !defined(ALTIVEC_USE_REFERENCE_C_CODE)
has_vectors = mm_support() & MM_ALTIVEC; has_vectors = mm_support() & MM_ALTIVEC;
...@@ -94,8 +94,24 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse) ...@@ -94,8 +94,24 @@ int ff_fft_init(FFTContext *s, int nbits, int inverse)
} while (nblocks != 0); } while (nblocks != 0);
av_freep(&s->exptab); av_freep(&s->exptab);
#if defined(HAVE_MMX) #if defined(HAVE_MMX)
s->fft_calc = ff_fft_calc_sse; #ifdef HAVE_MM3DNOW
#else if (has_vectors & MM_3DNOWEXT)
/* 3DNowEx for Athlon(XP) */
s->fft_calc = ff_fft_calc_3dn2;
else if (has_vectors & MM_3DNOW)
/* 3DNow! for K6-2/3 */
s->fft_calc = ff_fft_calc_3dn;
#endif
#ifdef HAVE_BUILTIN_VECTOR
if (has_vectors & MM_SSE2)
/* SSE for P4/K8 */
s->fft_calc = ff_fft_calc_sse;
else if ((has_vectors & MM_SSE) &&
s->fft_calc == ff_fft_calc_c)
/* SSE for P3 */
s->fft_calc = ff_fft_calc_sse;
#endif
#else /* HAVE_MMX */
s->fft_calc = ff_fft_calc_altivec; s->fft_calc = ff_fft_calc_altivec;
#endif #endif
} }
......
/*
* FFT/MDCT transform with 3DNow! optimizations
* Copyright (c) 2006 Zuxy MENG Jie.
* Based on fft_sse.c copyright (c) 2002 Fabrice Bellard.
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "../dsputil.h"
#include <math.h>
#ifdef HAVE_MM3DNOW
#include <mm3dnow.h>
static const int p1m1[2] __attribute__((aligned(8))) =
{ 0, 1 << 31 };
static const int m1p1[2] __attribute__((aligned(8))) =
{ 1 << 31, 0 };
void ff_fft_calc_3dn(FFTContext *s, FFTComplex *z)
{
int ln = s->nbits;
int j, np, np2;
int nblocks, nloops;
register FFTComplex *p, *q;
FFTComplex *cptr, *cptr1;
int k;
np = 1 << ln;
/* FEMMS not a must here but recommended by AMD */
_m_femms();
{
__m64 *r, a0, a1, b0, b1, tmp, c;
r = (__m64 *)&z[0];
if (s->inverse)
c = *(__m64 *)m1p1;
else
c = *(__m64 *)p1m1;
j = (np >> 2);
do {
/* do the pass 0 butterfly */
a0 = _m_pfadd(r[0], r[1]);
a1 = _m_pfsub(r[0], r[1]);
/* do the pass 0 butterfly */
b0 = _m_pfadd(r[2], r[3]);
b1 = _m_pfsub(r[2], r[3]);
/* multiply third by -i */
tmp = _m_punpckhdq(b1, b1);
b1 = _m_punpckldq(b1, b1);
b1 = _m_punpckldq(tmp, b1);
b1 = _m_pxor(b1, c);
/* do the pass 1 butterfly */
r[0] = _m_pfadd(a0, b0);
r[1] = _m_pfadd(a1, b1);
r[2] = _m_pfsub(a0, b0);
r[3] = _m_pfsub(a1, b1);
r += 4;
} while (--j != 0);
}
/* pass 2 .. ln-1 */
nblocks = np >> 3;
nloops = 1 << 2;
np2 = np >> 1;
cptr1 = s->exptab1;
do {
p = z;
q = z + nloops;
j = nblocks;
do {
cptr = cptr1;
k = nloops >> 1;
do {
__m64 a0, a1, b0, b1, c0, c1, t10, t11, t20, t21;
a0 = *(__m64 *)&p[0];
a1 = *(__m64 *)&p[1];
b0 = *(__m64 *)&q[0];
b1 = *(__m64 *)&q[1];
/* complex mul */
c0 = *(__m64 *)&cptr[0];
c1 = *(__m64 *)&cptr[1];
/* cre*re cim*re */
t10 = _m_pfmul(c0, _m_punpckldq(b0, b0));
t11 = _m_pfmul(c1, _m_punpckldq(b1, b1));
c0 = *(__m64 *)&cptr[2];
c1 = *(__m64 *)&cptr[3];
/* -cim*im cre*im */
t20 = _m_pfmul(c0, _m_punpckhdq(b0, b0));
t21 = _m_pfmul(c1, _m_punpckhdq(b1, b1));
b0 = _m_pfadd(t10, t20);
b1 = _m_pfadd(t11, t21);
/* butterfly */
*(__m64 *)&p[0] = _m_pfadd(a0, b0);
*(__m64 *)&p[1] = _m_pfadd(a1, b1);
*(__m64 *)&q[0] = _m_pfsub(a0, b0);
*(__m64 *)&q[1] = _m_pfsub(a1, b1);
p += 2;
q += 2;
cptr += 4;
} while (--k);
p += nloops;
q += nloops;
} while (--j);
cptr1 += nloops * 2;
nblocks = nblocks >> 1;
nloops = nloops << 1;
} while (nblocks != 0);
_m_femms();
}
#endif
/*
* FFT/MDCT transform with Extended 3DNow! optimizations
* Copyright (c) 2006 Zuxy MENG Jie.
* Based on fft_sse.c copyright (c) 2002 Fabrice Bellard.
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "../dsputil.h"
#include <math.h>
#ifdef HAVE_MM3DNOW
#include <mm3dnow.h>
static const int p1m1[2] __attribute__((aligned(8))) =
{ 0, 1 << 31 };
static const int m1p1[2] __attribute__((aligned(8))) =
{ 1 << 31, 0 };
void ff_fft_calc_3dn2(FFTContext *s, FFTComplex *z)
{
int ln = s->nbits;
int j, np, np2;
int nblocks, nloops;
register FFTComplex *p, *q;
FFTComplex *cptr, *cptr1;
int k;
np = 1 << ln;
/* FEMMS is not a must here but recommended by AMD */
_m_femms();
{
__m64 *r, a0, a1, b0, b1, c;
r = (__m64 *)&z[0];
if (s->inverse)
c = *(__m64 *)m1p1;
else
c = *(__m64 *)p1m1;
j = (np >> 2);
do {
/* do the pass 0 butterfly */
a0 = _m_pfadd(r[0], r[1]);
a1 = _m_pfsub(r[0], r[1]);
/* do the pass 0 butterfly */
b0 = _m_pfadd(r[2], r[3]);
b1 = _m_pfsub(r[2], r[3]);
/* multiply third by -i */
b1 = _m_pswapd(b1);
b1 = _m_pxor(b1, c);
r[0] = _m_pfadd(a0, b0);
r[1] = _m_pfadd(a1, b1);
r[2] = _m_pfsub(a0, b0);
r[3] = _m_pfsub(a1, b1);
r += 4;
} while (--j != 0);
}
/* pass 2 .. ln-1 */
nblocks = np >> 3;
nloops = 1 << 2;
np2 = np >> 1;
cptr1 = s->exptab1;
do {
p = z;
q = z + nloops;
j = nblocks;
do {
cptr = cptr1;
k = nloops >> 1;
do {
__m64 a0, a1, b0, b1, c0, c1, t10, t11, t20, t21;
a0 = *(__m64 *)&p[0];
a1 = *(__m64 *)&p[1];
b0 = *(__m64 *)&q[0];
b1 = *(__m64 *)&q[1];
/* complex mul */
c0 = *(__m64 *)&cptr[0];
c1 = *(__m64 *)&cptr[1];
/* cre*re cim*im */
t10 = _m_pfmul(c0, b0);
t11 = _m_pfmul(c1, b1);
/* no need to access cptr[2] & cptr[3] */
c0 = _m_pswapd(c0);
c1 = _m_pswapd(c1);
/* cim*re cre*im */
t20 = _m_pfmul(c0, b0);
t21 = _m_pfmul(c1, b1);
/* cre*re-cim*im cim*re+cre*im */
b0 = _m_pfpnacc(t10, t20);
b1 = _m_pfpnacc(t11, t21);
/* butterfly */
*(__m64 *)&p[0] = _m_pfadd(a0, b0);
*(__m64 *)&p[1] = _m_pfadd(a1, b1);
*(__m64 *)&q[0] = _m_pfsub(a0, b0);
*(__m64 *)&q[1] = _m_pfsub(a1, b1);
p += 2;
q += 2;
cptr += 4;
} while (--k);
p += nloops;
q += nloops;
} while (--j);
cptr1 += nloops * 2;
nblocks = nblocks >> 1;
nloops = nloops << 1;
} while (nblocks != 0);
_m_femms();
}
#endif
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment