summaryrefslogtreecommitdiff
path: root/libavcodec/x86/ac3dsp_init.c
diff options
context:
space:
mode:
authorMans Rullgard <mans@mansr.com>2012-09-11 17:55:11 +0100
committerMans Rullgard <mans@mansr.com>2012-09-12 23:39:50 +0100
commit97cb9236cf3c45eb8a8fba8b88bdde6ece3f496f (patch)
tree82d61b5b9f447f09a61277e34022336833853193 /libavcodec/x86/ac3dsp_init.c
parentb901c30da606ade7427fbfb0f9651cc5bfb953a7 (diff)
ac3: move ac3_downmix() from dsputil to ac3dsp
Signed-off-by: Mans Rullgard <mans@mansr.com>
Diffstat (limited to 'libavcodec/x86/ac3dsp_init.c')
-rw-r--r--libavcodec/x86/ac3dsp_init.c119
1 files changed, 119 insertions, 0 deletions
diff --git a/libavcodec/x86/ac3dsp_init.c b/libavcodec/x86/ac3dsp_init.c
index d1c45dd2ac..b23a9a1aeb 100644
--- a/libavcodec/x86/ac3dsp_init.c
+++ b/libavcodec/x86/ac3dsp_init.c
@@ -19,9 +19,11 @@
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
+#include "libavutil/mem.h"
#include "libavutil/x86/asm.h"
#include "libavutil/x86/cpu.h"
#include "dsputil_mmx.h"
+#include "libavcodec/ac3.h"
#include "libavcodec/ac3dsp.h"
extern void ff_ac3_exponent_min_mmx (uint8_t *exp, int num_reuse_blocks, int nb_coefs);
@@ -49,6 +51,119 @@ extern void ff_ac3_extract_exponents_3dnow(uint8_t *exp, int32_t *coef, int nb_c
extern void ff_ac3_extract_exponents_sse2 (uint8_t *exp, int32_t *coef, int nb_coefs);
extern void ff_ac3_extract_exponents_ssse3(uint8_t *exp, int32_t *coef, int nb_coefs);
+#if HAVE_SSE_INLINE
+
+#define IF1(x) x
+#define IF0(x)
+
+#define MIX5(mono, stereo) \
+ __asm__ volatile ( \
+ "movss 0(%2), %%xmm5 \n" \
+ "movss 8(%2), %%xmm6 \n" \
+ "movss 24(%2), %%xmm7 \n" \
+ "shufps $0, %%xmm5, %%xmm5 \n" \
+ "shufps $0, %%xmm6, %%xmm6 \n" \
+ "shufps $0, %%xmm7, %%xmm7 \n" \
+ "1: \n" \
+ "movaps (%0, %1), %%xmm0 \n" \
+ "movaps 0x400(%0, %1), %%xmm1 \n" \
+ "movaps 0x800(%0, %1), %%xmm2 \n" \
+ "movaps 0xc00(%0, %1), %%xmm3 \n" \
+ "movaps 0x1000(%0, %1), %%xmm4 \n" \
+ "mulps %%xmm5, %%xmm0 \n" \
+ "mulps %%xmm6, %%xmm1 \n" \
+ "mulps %%xmm5, %%xmm2 \n" \
+ "mulps %%xmm7, %%xmm3 \n" \
+ "mulps %%xmm7, %%xmm4 \n" \
+ stereo("addps %%xmm1, %%xmm0 \n") \
+ "addps %%xmm1, %%xmm2 \n" \
+ "addps %%xmm3, %%xmm0 \n" \
+ "addps %%xmm4, %%xmm2 \n" \
+ mono("addps %%xmm2, %%xmm0 \n") \
+ "movaps %%xmm0, (%0, %1) \n" \
+ stereo("movaps %%xmm2, 0x400(%0, %1) \n") \
+ "add $16, %0 \n" \
+ "jl 1b \n" \
+ : "+&r"(i) \
+ : "r"(samples[0] + len), "r"(matrix) \
+ : XMM_CLOBBERS("%xmm0", "%xmm1", "%xmm2", "%xmm3", \
+ "%xmm4", "%xmm5", "%xmm6", "%xmm7",) \
+ "memory" \
+ );
+
+#define MIX_MISC(stereo) \
+ __asm__ volatile ( \
+ "1: \n" \
+ "movaps (%3, %0), %%xmm0 \n" \
+ stereo("movaps %%xmm0, %%xmm1 \n") \
+ "mulps %%xmm4, %%xmm0 \n" \
+ stereo("mulps %%xmm5, %%xmm1 \n") \
+ "lea 1024(%3, %0), %1 \n" \
+ "mov %5, %2 \n" \
+ "2: \n" \
+ "movaps (%1), %%xmm2 \n" \
+ stereo("movaps %%xmm2, %%xmm3 \n") \
+ "mulps (%4, %2), %%xmm2 \n" \
+ stereo("mulps 16(%4, %2), %%xmm3 \n") \
+ "addps %%xmm2, %%xmm0 \n" \
+ stereo("addps %%xmm3, %%xmm1 \n") \
+ "add $1024, %1 \n" \
+ "add $32, %2 \n" \
+ "jl 2b \n" \
+ "movaps %%xmm0, (%3, %0) \n" \
+ stereo("movaps %%xmm1, 1024(%3, %0) \n") \
+ "add $16, %0 \n" \
+ "jl 1b \n" \
+ : "+&r"(i), "=&r"(j), "=&r"(k) \
+ : "r"(samples[0] + len), "r"(matrix_simd + in_ch), \
+ "g"((intptr_t) - 32 * (in_ch - 1)) \
+ : "memory" \
+ );
+
+static void ac3_downmix_sse(float (*samples)[256], float (*matrix)[2],
+ int out_ch, int in_ch, int len)
+{
+ int (*matrix_cmp)[2] = (int(*)[2])matrix;
+ intptr_t i, j, k;
+
+ i = -len * sizeof(float);
+ if (in_ch == 5 && out_ch == 2 &&
+ !(matrix_cmp[0][1] | matrix_cmp[2][0] |
+ matrix_cmp[3][1] | matrix_cmp[4][0] |
+ (matrix_cmp[1][0] ^ matrix_cmp[1][1]) |
+ (matrix_cmp[0][0] ^ matrix_cmp[2][1]))) {
+ MIX5(IF0, IF1);
+ } else if (in_ch == 5 && out_ch == 1 &&
+ matrix_cmp[0][0] == matrix_cmp[2][0] &&
+ matrix_cmp[3][0] == matrix_cmp[4][0]) {
+ MIX5(IF1, IF0);
+ } else {
+ DECLARE_ALIGNED(16, float, matrix_simd)[AC3_MAX_CHANNELS][2][4];
+ j = 2 * in_ch * sizeof(float);
+ __asm__ volatile (
+ "1: \n"
+ "sub $8, %0 \n"
+ "movss (%2, %0), %%xmm4 \n"
+ "movss 4(%2, %0), %%xmm5 \n"
+ "shufps $0, %%xmm4, %%xmm4 \n"
+ "shufps $0, %%xmm5, %%xmm5 \n"
+ "movaps %%xmm4, (%1, %0, 4) \n"
+ "movaps %%xmm5, 16(%1, %0, 4) \n"
+ "jg 1b \n"
+ : "+&r"(j)
+ : "r"(matrix_simd), "r"(matrix)
+ : "memory"
+ );
+ if (out_ch == 2) {
+ MIX_MISC(IF1);
+ } else {
+ MIX_MISC(IF0);
+ }
+ }
+}
+
+#endif /* HAVE_SSE_INLINE */
+
av_cold void ff_ac3dsp_init_x86(AC3DSPContext *c, int bit_exact)
{
int mm_flags = av_get_cpu_flags();
@@ -89,4 +204,8 @@ av_cold void ff_ac3dsp_init_x86(AC3DSPContext *c, int bit_exact)
c->extract_exponents = ff_ac3_extract_exponents_ssse3;
}
}
+
+ if (INLINE_SSE(mm_flags)) {
+ c->downmix = ac3_downmix_sse;
+ }
}