summaryrefslogtreecommitdiff
path: root/libavcodec/x86
diff options
context:
space:
mode:
authorRonald S. Bultje <rsbultje@gmail.com>2012-07-31 20:23:39 -0700
committerRonald S. Bultje <rsbultje@gmail.com>2012-07-31 21:20:47 -0700
commitb6a3849adb0381a437952a785d39e22cb3b00282 (patch)
tree4f2ec873af703a90ba0a7018b618438559a8b33d /libavcodec/x86
parentddbe71b44fc810cc39b576bf0047f89090cfabcb (diff)
fft: port FFT/IMDCT 3dnow functions to yasm, and disable on x86-64.
64-bit CPUs always have SSE available, thus there is no need to compile in the 3dnow functions. This results in smaller binaries.
Diffstat (limited to 'libavcodec/x86')
-rw-r--r--libavcodec/x86/Makefile2
-rw-r--r--libavcodec/x86/fft.c2
-rw-r--r--libavcodec/x86/fft_3dn.c23
-rw-r--r--libavcodec/x86/fft_3dn2.c173
-rw-r--r--libavcodec/x86/fft_mmx.asm196
5 files changed, 170 insertions, 226 deletions
diff --git a/libavcodec/x86/Makefile b/libavcodec/x86/Makefile
index 1d2635749f..0ae70b2dd0 100644
--- a/libavcodec/x86/Makefile
+++ b/libavcodec/x86/Makefile
@@ -38,8 +38,6 @@ YASM-OBJS-$(CONFIG_AAC_DECODER) += x86/sbrdsp.o
YASM-OBJS-$(CONFIG_AC3DSP) += x86/ac3dsp.o
YASM-OBJS-$(CONFIG_DCT) += x86/dct32_sse.o
YASM-OBJS-$(CONFIG_ENCODERS) += x86/dsputilenc_yasm.o
-YASM-OBJS-FFT-$(HAVE_AMD3DNOW) += x86/fft_3dn.o
-YASM-OBJS-FFT-$(HAVE_AMD3DNOWEXT) += x86/fft_3dn2.o
YASM-OBJS-$(CONFIG_FFT) += x86/fft_mmx.o \
$(YASM-OBJS-FFT-yes)
YASM-OBJS-$(CONFIG_H264CHROMA) += x86/h264_chromamc.o \
diff --git a/libavcodec/x86/fft.c b/libavcodec/x86/fft.c
index 6349c239c3..f1c1c9d36b 100644
--- a/libavcodec/x86/fft.c
+++ b/libavcodec/x86/fft.c
@@ -25,6 +25,7 @@ av_cold void ff_fft_init_mmx(FFTContext *s)
{
#if HAVE_YASM
int has_vectors = av_get_cpu_flags();
+#if ARCH_X86_32
if (has_vectors & AV_CPU_FLAG_3DNOW && HAVE_AMD3DNOW) {
/* 3DNow! for K6-2/3 */
s->imdct_calc = ff_imdct_calc_3dnow;
@@ -37,6 +38,7 @@ av_cold void ff_fft_init_mmx(FFTContext *s)
s->imdct_half = ff_imdct_half_3dnow2;
s->fft_calc = ff_fft_calc_3dnow2;
}
+#endif
if (has_vectors & AV_CPU_FLAG_SSE && HAVE_SSE) {
/* SSE for P3/P4/K8 */
s->imdct_calc = ff_imdct_calc_sse;
diff --git a/libavcodec/x86/fft_3dn.c b/libavcodec/x86/fft_3dn.c
deleted file mode 100644
index 5a4d3ad2c8..0000000000
--- a/libavcodec/x86/fft_3dn.c
+++ /dev/null
@@ -1,23 +0,0 @@
-/*
- * FFT/MDCT transform with 3DNow! optimizations
- * Copyright (c) 2008 Loren Merritt
- *
- * This file is part of Libav.
- *
- * Libav is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * Libav is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with Libav; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#define EMULATE_3DNOWEXT
-#include "fft_3dn2.c"
diff --git a/libavcodec/x86/fft_3dn2.c b/libavcodec/x86/fft_3dn2.c
deleted file mode 100644
index e684cc745f..0000000000
--- a/libavcodec/x86/fft_3dn2.c
+++ /dev/null
@@ -1,173 +0,0 @@
-/*
- * FFT/MDCT transform with Extended 3DNow! optimizations
- * Copyright (c) 2006-2008 Zuxy MENG Jie, Loren Merritt
- *
- * This file is part of Libav.
- *
- * Libav is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * Libav is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with Libav; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "libavutil/x86_cpu.h"
-#include "libavcodec/dsputil.h"
-#include "fft.h"
-
-DECLARE_ALIGNED(8, static const unsigned int, m1m1)[2] = { 1U<<31, 1U<<31 };
-
-#ifdef EMULATE_3DNOWEXT
-#define PSWAPD(s,d)\
- "movq "#s","#d"\n"\
- "psrlq $32,"#d"\n"\
- "punpckldq "#s","#d"\n"
-#define ff_fft_calc_3dnow2 ff_fft_calc_3dnow
-#define ff_fft_dispatch_3dnow2 ff_fft_dispatch_3dnow
-#define ff_fft_dispatch_interleave_3dnow2 ff_fft_dispatch_interleave_3dnow
-#define ff_imdct_calc_3dnow2 ff_imdct_calc_3dnow
-#define ff_imdct_half_3dnow2 ff_imdct_half_3dnow
-#else
-#define PSWAPD(s,d) "pswapd "#s","#d"\n"
-#endif
-
-void ff_fft_dispatch_3dnow2(FFTComplex *z, int nbits);
-void ff_fft_dispatch_interleave_3dnow2(FFTComplex *z, int nbits);
-
-void ff_fft_calc_3dnow2(FFTContext *s, FFTComplex *z)
-{
- int n = 1<<s->nbits;
- int i;
- ff_fft_dispatch_interleave_3dnow2(z, s->nbits);
- __asm__ volatile("femms");
- if(n <= 8)
- for(i=0; i<n; i+=2)
- FFSWAP(FFTSample, z[i].im, z[i+1].re);
-}
-
-void ff_imdct_half_3dnow2(FFTContext *s, FFTSample *output, const FFTSample *input)
-{
- x86_reg j, k;
- long n = s->mdct_size;
- long n2 = n >> 1;
- long n4 = n >> 2;
- long n8 = n >> 3;
- const uint16_t *revtab = s->revtab;
- const FFTSample *tcos = s->tcos;
- const FFTSample *tsin = s->tsin;
- const FFTSample *in1, *in2;
- FFTComplex *z = (FFTComplex *)output;
-
- /* pre rotation */
- in1 = input;
- in2 = input + n2 - 1;
-#ifdef EMULATE_3DNOWEXT
- __asm__ volatile("movd %0, %%mm7" ::"r"(1U<<31));
-#endif
- for(k = 0; k < n4; k++) {
- // FIXME a single block is faster, but gcc 2.95 and 3.4.x on 32bit can't compile it
- __asm__ volatile(
- "movd %0, %%mm0 \n"
- "movd %2, %%mm1 \n"
- "punpckldq %1, %%mm0 \n"
- "punpckldq %3, %%mm1 \n"
- "movq %%mm0, %%mm2 \n"
- PSWAPD( %%mm1, %%mm3 )
- "pfmul %%mm1, %%mm0 \n"
- "pfmul %%mm3, %%mm2 \n"
-#ifdef EMULATE_3DNOWEXT
- "movq %%mm0, %%mm1 \n"
- "punpckhdq %%mm2, %%mm0 \n"
- "punpckldq %%mm2, %%mm1 \n"
- "pxor %%mm7, %%mm0 \n"
- "pfadd %%mm1, %%mm0 \n"
-#else
- "pfpnacc %%mm2, %%mm0 \n"
-#endif
- ::"m"(in2[-2*k]), "m"(in1[2*k]),
- "m"(tcos[k]), "m"(tsin[k])
- );
- __asm__ volatile(
- "movq %%mm0, %0 \n\t"
- :"=m"(z[revtab[k]])
- );
- }
-
- ff_fft_dispatch_3dnow2(z, s->nbits);
-
-#define CMUL(j,mm0,mm1)\
- "movq (%2,"#j",2), %%mm6 \n"\
- "movq 8(%2,"#j",2), "#mm0"\n"\
- "movq %%mm6, "#mm1"\n"\
- "movq "#mm0",%%mm7 \n"\
- "pfmul (%3,"#j"), %%mm6 \n"\
- "pfmul (%4,"#j"), "#mm0"\n"\
- "pfmul (%4,"#j"), "#mm1"\n"\
- "pfmul (%3,"#j"), %%mm7 \n"\
- "pfsub %%mm6, "#mm0"\n"\
- "pfadd %%mm7, "#mm1"\n"
-
- /* post rotation */
- j = -n2;
- k = n2-8;
- __asm__ volatile(
- "1: \n"
- CMUL(%0, %%mm0, %%mm1)
- CMUL(%1, %%mm2, %%mm3)
- "movd %%mm0, (%2,%0,2) \n"
- "movd %%mm1,12(%2,%1,2) \n"
- "movd %%mm2, (%2,%1,2) \n"
- "movd %%mm3,12(%2,%0,2) \n"
- "psrlq $32, %%mm0 \n"
- "psrlq $32, %%mm1 \n"
- "psrlq $32, %%mm2 \n"
- "psrlq $32, %%mm3 \n"
- "movd %%mm0, 8(%2,%0,2) \n"
- "movd %%mm1, 4(%2,%1,2) \n"
- "movd %%mm2, 8(%2,%1,2) \n"
- "movd %%mm3, 4(%2,%0,2) \n"
- "sub $8, %1 \n"
- "add $8, %0 \n"
- "jl 1b \n"
- :"+r"(j), "+r"(k)
- :"r"(z+n8), "r"(tcos+n8), "r"(tsin+n8)
- :"memory"
- );
- __asm__ volatile("femms");
-}
-
-void ff_imdct_calc_3dnow2(FFTContext *s, FFTSample *output, const FFTSample *input)
-{
- x86_reg j, k;
- long n = s->mdct_size;
- long n4 = n >> 2;
-
- ff_imdct_half_3dnow2(s, output+n4, input);
-
- j = -n;
- k = n-8;
- __asm__ volatile(
- "movq %4, %%mm7 \n"
- "1: \n"
- PSWAPD((%2,%1), %%mm0)
- PSWAPD((%3,%0), %%mm1)
- "pxor %%mm7, %%mm0 \n"
- "movq %%mm1, (%3,%1) \n"
- "movq %%mm0, (%2,%0) \n"
- "sub $8, %1 \n"
- "add $8, %0 \n"
- "jl 1b \n"
- :"+r"(j), "+r"(k)
- :"r"(output+n4), "r"(output+n4*3),
- "m"(*m1m1)
- );
- __asm__ volatile("femms");
-}
diff --git a/libavcodec/x86/fft_mmx.asm b/libavcodec/x86/fft_mmx.asm
index 5c6583b3b7..81e4411dcb 100644
--- a/libavcodec/x86/fft_mmx.asm
+++ b/libavcodec/x86/fft_mmx.asm
@@ -29,6 +29,7 @@
; i.e. {4x real, 4x imaginary, 4x real, ...} (or 2x respectively)
%include "x86inc.asm"
+%include "x86util.asm"
%if ARCH_X86_64
%define pointer resq
@@ -105,7 +106,7 @@ SECTION_TEXT
pfadd %5, %4 ; {t6,t5}
pxor %3, [ps_m1p1] ; {t8,t7}
mova %6, %1
- pswapd %3, %3
+ PSWAPD %3, %3
pfadd %1, %5 ; {r0,i0}
pfsub %6, %5 ; {r2,i2}
mova %4, %2
@@ -396,7 +397,6 @@ fft32_interleave_avx:
%endif
INIT_XMM sse
-%define movdqa movaps
align 16
fft4_avx:
@@ -469,8 +469,8 @@ fft8 %+ SUFFIX:
mova Z(2), m2
T2_3DN m4, m5, Z(4), Z(5)
T2_3DN m6, m7, Z2(6), Z2(7)
- pswapd m0, m5
- pswapd m2, m7
+ PSWAPD m0, m5
+ PSWAPD m2, m7
pxor m0, [ps_m1p1]
pxor m2, [ps_m1p1]
pfsub m5, m0
@@ -498,11 +498,11 @@ fft8 %+ SUFFIX:
ret
%endmacro
-INIT_MMX 3dnow2
-FFT48_3DN
-
-%macro pswapd 2
-%ifidn %1, %2
+%if ARCH_X86_32
+%macro PSWAPD 2
+%if cpuflag(3dnow2)
+ pswapd %1, %2
+%elifidn %1, %2
movd [r0+12], %1
punpckhdq %1, [r0+8]
%else
@@ -512,9 +512,12 @@ FFT48_3DN
%endif
%endmacro
-INIT_MMX 3dnow
+INIT_MMX 3dnow2
FFT48_3DN
+INIT_MMX 3dnow
+FFT48_3DN
+%endif
%define Z(x) [zcq + o1q*(x&6) + mmsize*(x&1)]
%define Z2(x) [zcq + o3q + mmsize*(x&1)]
@@ -588,6 +591,7 @@ INIT_XMM sse
DECL_PASS pass_sse, PASS_BIG 1
DECL_PASS pass_interleave_sse, PASS_BIG 0
+%macro FFT_CALC_FUNC 0
cglobal fft_calc, 2,5,8
mov r3d, [r0 + FFTContext.nbits]
PUSH r1
@@ -597,23 +601,43 @@ cglobal fft_calc, 2,5,8
FFT_DISPATCH _interleave %+ SUFFIX, r1
POP rcx
POP r4
- cmp rcx, 4
+ cmp rcx, 3+(mmsize/16)
jg .end
mov r2, -1
add rcx, 3
shl r2, cl
sub r4, r2
.loop
+%if mmsize == 8
+ PSWAPD m0, [r4 + r2 + 4]
+ mova [r4 + r2 + 4], m0
+%else
movaps xmm0, [r4 + r2]
movaps xmm1, xmm0
unpcklps xmm0, [r4 + r2 + 16]
unpckhps xmm1, [r4 + r2 + 16]
movaps [r4 + r2], xmm0
movaps [r4 + r2 + 16], xmm1
- add r2, 32
+%endif
+ add r2, mmsize*2
jl .loop
.end:
+%if cpuflag(3dnow)
+ femms
+ RET
+%else
REP_RET
+%endif
+%endmacro
+
+%if ARCH_X86_32
+INIT_MMX 3dnow
+FFT_CALC_FUNC
+INIT_MMX 3dnow2
+FFT_CALC_FUNC
+%endif
+INIT_XMM sse
+FFT_CALC_FUNC
cglobal fft_permute, 2,7,1
mov r4, [r0 + FFTContext.revtab]
@@ -648,6 +672,7 @@ cglobal fft_permute, 2,7,1
jl .loopcopy
REP_RET
+%macro IMDCT_CALC_FUNC 0
cglobal imdct_calc, 3,5,3
mov r3d, [r0 + FFTContext.mdctsize]
mov r4, [r0 + FFTContext.imdcthalf]
@@ -671,22 +696,45 @@ cglobal imdct_calc, 3,5,3
POP r3
lea r0, [r1 + 2*r3]
mov r2, r3
- sub r3, 16
+ sub r3, mmsize
neg r2
- movaps xmm2, [ps_m1m1m1m1]
+ mova m2, [ps_m1m1m1m1]
.loop:
- movaps xmm0, [r1 + r3]
- movaps xmm1, [r0 + r2]
- shufps xmm0, xmm0, 0x1b
- shufps xmm1, xmm1, 0x1b
- xorps xmm0, xmm2
- movaps [r0 + r3], xmm1
- movaps [r1 + r2], xmm0
- sub r3, 16
- add r2, 16
+%if mmsize == 8
+ PSWAPD m0, [r1 + r3]
+ PSWAPD m1, [r0 + r2]
+ pxor m0, m2
+%else
+ mova m0, [r1 + r3]
+ mova m1, [r0 + r2]
+ shufps m0, m0, 0x1b
+ shufps m1, m1, 0x1b
+ xorps m0, m2
+%endif
+ mova [r0 + r3], m1
+ mova [r1 + r2], m0
+ sub r3, mmsize
+ add r2, mmsize
jl .loop
+%if cpuflag(3dnow)
+ femms
+ RET
+%else
REP_RET
+%endif
+%endmacro
+%if ARCH_X86_32
+INIT_MMX 3dnow
+IMDCT_CALC_FUNC
+INIT_MMX 3dnow2
+IMDCT_CALC_FUNC
+%endif
+
+INIT_XMM sse
+IMDCT_CALC_FUNC
+
+%if ARCH_X86_32
INIT_MMX 3dnow
%define mulps pfmul
%define addps pfadd
@@ -697,6 +745,7 @@ DECL_PASS pass_3dnow, PASS_SMALL 1, [wq], [wq+o1q]
DECL_PASS pass_interleave_3dnow, PASS_BIG 0
%define pass_3dnow2 pass_3dnow
%define pass_interleave_3dnow2 pass_interleave_3dnow
+%endif
%ifdef PIC
%define SECTION_REL - $$
@@ -760,12 +809,14 @@ DECL_FFT 6, _interleave
INIT_XMM sse
DECL_FFT 5
DECL_FFT 5, _interleave
+%if ARCH_X86_32
INIT_MMX 3dnow
DECL_FFT 4
DECL_FFT 4, _interleave
INIT_MMX 3dnow2
DECL_FFT 4
DECL_FFT 4, _interleave
+%endif
INIT_XMM sse
%undef mulps
@@ -775,6 +826,37 @@ INIT_XMM sse
%undef unpckhps
%macro PREROTATER 5 ;-2*k, 2*k, input+n4, tcos+n8, tsin+n8
+%if mmsize == 8 ; j*2+2-n4, n4-2-j*2, input+n4, tcos+n8, tsin+n8
+ PSWAPD m0, [%3+%2*4]
+ movq m2, [%3+%1*4-8]
+ movq m3, m0
+ punpckldq m0, m2
+ punpckhdq m2, m3
+ movd m1, [%4+%1*2-4] ; tcos[j]
+ movd m3, [%4+%2*2] ; tcos[n4-j-1]
+ punpckldq m1, [%5+%1*2-4] ; tsin[j]
+ punpckldq m3, [%5+%2*2] ; tsin[n4-j-1]
+
+ mova m4, m0
+ PSWAPD m5, m1
+ pfmul m0, m1
+ pfmul m4, m5
+ mova m6, m2
+ PSWAPD m5, m3
+ pfmul m2, m3
+ pfmul m6, m5
+%if cpuflag(3dnow2)
+ pfpnacc m0, m4
+ pfpnacc m2, m6
+%else
+ SBUTTERFLY dq, 0, 4, 1
+ SBUTTERFLY dq, 2, 6, 3
+ pxor m4, m7
+ pxor m6, m7
+ pfadd m0, m4
+ pfadd m2, m6
+%endif
+%else
movaps xmm0, [%3+%2*4]
movaps xmm1, [%3+%1*4-0x10]
movaps xmm2, xmm0
@@ -795,6 +877,7 @@ INIT_XMM sse
movaps xmm0, xmm1
unpcklps xmm1, xmm2
unpckhps xmm0, xmm2
+%endif
%endmacro
%macro CMUL 6 ;j, xmm0, xmm1, 3, 4, 5
@@ -863,6 +946,40 @@ INIT_XMM sse
jl .post
%endmacro
+%macro CMUL_3DNOW 6
+ mova m6, [%1+%2*2]
+ mova %3, [%1+%2*2+8]
+ mova %4, m6
+ mova m7, %3
+ pfmul m6, [%5+%2]
+ pfmul %3, [%6+%2]
+ pfmul %4, [%6+%2]
+ pfmul m7, [%5+%2]
+ pfsub %3, m6
+ pfadd %4, m7
+%endmacro
+
+%macro POSROTATESHUF_3DNOW 5 ;j, k, z+n8, tcos+n8, tsin+n8
+.post:
+ CMUL_3DNOW %3, %1, m0, m1, %4, %5
+ CMUL_3DNOW %3, %2, m2, m3, %4, %5
+ movd [%3+%1*2+ 0], m0
+ movd [%3+%2*2+12], m1
+ movd [%3+%2*2+ 0], m2
+ movd [%3+%1*2+12], m3
+ psrlq m0, 32
+ psrlq m1, 32
+ psrlq m2, 32
+ psrlq m3, 32
+ movd [%3+%1*2+ 8], m0
+ movd [%3+%2*2+ 4], m1
+ movd [%3+%2*2+ 8], m2
+ movd [%3+%1*2+ 4], m3
+ sub %2, 8
+ add %1, 8
+ jl .post
+%endmacro
+
%macro DECL_IMDCT 1
cglobal imdct_half, 3,12,8; FFTContext *s, FFTSample *output, const FFTSample *input
%if ARCH_X86_64
@@ -892,22 +1009,34 @@ cglobal imdct_half, 3,12,8; FFTContext *s, FFTSample *output, const FFTSample *i
push rrevtab
%endif
- sub r3, 4
-%if ARCH_X86_64
+ sub r3, mmsize/4
+%if ARCH_X86_64 || mmsize == 8
xor r4, r4
sub r4, r3
%endif
+%if notcpuflag(3dnow2) && mmsize == 8
+ movd m7, [ps_m1m1m1m1]
+%endif
.pre:
%if ARCH_X86_64 == 0
;unspill
+%if mmsize != 8
xor r4, r4
sub r4, r3
- mov rtsin, [esp+4]
+%endif
mov rtcos, [esp+8]
+ mov rtsin, [esp+4]
%endif
PREROTATER r4, r3, r2, rtcos, rtsin
-%if ARCH_X86_64
+%if mmsize == 8
+ mov r6, [esp] ; rrevtab = ptr+n8
+ movzx r5, word [rrevtab+r4-2] ; rrevtab[j]
+ movzx r6, word [rrevtab+r3] ; rrevtab[n4-j-1]
+ mova [r1+r5*8], m0
+ mova [r1+r6*8], m2
+ add r4, 2
+%elif ARCH_X86_64
movzx r5, word [rrevtab+r4-4]
movzx r6, word [rrevtab+r4-2]
movzx r10, word [rrevtab+r3]
@@ -928,7 +1057,7 @@ cglobal imdct_half, 3,12,8; FFTContext *s, FFTSample *output, const FFTSample *i
movlps [r1+r5*8], xmm1
movhps [r1+r4*8], xmm1
%endif
- sub r3, 4
+ sub r3, mmsize/4
jns .pre
mov r5, r0
@@ -954,11 +1083,22 @@ cglobal imdct_half, 3,12,8; FFTContext *s, FFTSample *output, const FFTSample *i
%if ARCH_X86_64 == 0
add esp, 12
%endif
+%if mmsize == 8
+ femms
+%endif
RET
%endmacro
DECL_IMDCT POSROTATESHUF
+%if ARCH_X86_32
+INIT_MMX 3dnow
+DECL_IMDCT POSROTATESHUF_3DNOW
+
+INIT_MMX 3dnow2
+DECL_IMDCT POSROTATESHUF_3DNOW
+%endif
+
INIT_YMM avx
%if HAVE_AVX