summaryrefslogtreecommitdiff
path: root/libavresample/x86
diff options
context:
space:
mode:
Diffstat (limited to 'libavresample/x86')
-rw-r--r--libavresample/x86/Makefile9
-rw-r--r--libavresample/x86/audio_convert.asm1261
-rw-r--r--libavresample/x86/audio_convert_init.c265
-rw-r--r--libavresample/x86/audio_mix.asm511
-rw-r--r--libavresample/x86/audio_mix_init.c215
-rw-r--r--libavresample/x86/dither.asm117
-rw-r--r--libavresample/x86/dither_init.c60
-rw-r--r--libavresample/x86/util.asm41
-rw-r--r--libavresample/x86/w64xmmtest.c31
9 files changed, 0 insertions, 2510 deletions
diff --git a/libavresample/x86/Makefile b/libavresample/x86/Makefile
deleted file mode 100644
index 55b709ce36..0000000000
--- a/libavresample/x86/Makefile
+++ /dev/null
@@ -1,9 +0,0 @@
-OBJS += x86/audio_convert_init.o \
- x86/audio_mix_init.o \
- x86/dither_init.o \
-
-OBJS-$(CONFIG_XMM_CLOBBER_TEST) += x86/w64xmmtest.o
-
-X86ASM-OBJS += x86/audio_convert.o \
- x86/audio_mix.o \
- x86/dither.o \
diff --git a/libavresample/x86/audio_convert.asm b/libavresample/x86/audio_convert.asm
deleted file mode 100644
index c6a5015282..0000000000
--- a/libavresample/x86/audio_convert.asm
+++ /dev/null
@@ -1,1261 +0,0 @@
-;******************************************************************************
-;* x86 optimized Format Conversion Utils
-;* Copyright (c) 2008 Loren Merritt
-;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
-;*
-;* This file is part of FFmpeg.
-;*
-;* FFmpeg is free software; you can redistribute it and/or
-;* modify it under the terms of the GNU Lesser General Public
-;* License as published by the Free Software Foundation; either
-;* version 2.1 of the License, or (at your option) any later version.
-;*
-;* FFmpeg is distributed in the hope that it will be useful,
-;* but WITHOUT ANY WARRANTY; without even the implied warranty of
-;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-;* Lesser General Public License for more details.
-;*
-;* You should have received a copy of the GNU Lesser General Public
-;* License along with FFmpeg; if not, write to the Free Software
-;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
-;******************************************************************************
-
-%include "libavutil/x86/x86util.asm"
-%include "util.asm"
-
-SECTION_RODATA 32
-
-pf_s32_inv_scale: times 8 dd 0x30000000
-pf_s32_scale: times 8 dd 0x4f000000
-pf_s32_clip: times 8 dd 0x4effffff
-pf_s16_inv_scale: times 4 dd 0x38000000
-pf_s16_scale: times 4 dd 0x47000000
-pb_shuf_unpack_even: db -1, -1, 0, 1, -1, -1, 2, 3, -1, -1, 8, 9, -1, -1, 10, 11
-pb_shuf_unpack_odd: db -1, -1, 4, 5, -1, -1, 6, 7, -1, -1, 12, 13, -1, -1, 14, 15
-pb_interleave_words: SHUFFLE_MASK_W 0, 4, 1, 5, 2, 6, 3, 7
-pb_deinterleave_words: SHUFFLE_MASK_W 0, 2, 4, 6, 1, 3, 5, 7
-pw_zero_even: times 4 dw 0x0000, 0xffff
-
-SECTION .text
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len);
-;------------------------------------------------------------------------------
-
-INIT_XMM sse2
-cglobal conv_s16_to_s32, 3,3,3, dst, src, len
- lea lenq, [2*lend]
- lea dstq, [dstq+2*lenq]
- add srcq, lenq
- neg lenq
-.loop:
- mova m2, [srcq+lenq]
- pxor m0, m0
- pxor m1, m1
- punpcklwd m0, m2
- punpckhwd m1, m2
- mova [dstq+2*lenq ], m0
- mova [dstq+2*lenq+mmsize], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16_TO_FLT 0
-cglobal conv_s16_to_flt, 3,3,3, dst, src, len
- lea lenq, [2*lend]
- add srcq, lenq
- lea dstq, [dstq + 2*lenq]
- neg lenq
- mova m2, [pf_s16_inv_scale]
- ALIGN 16
-.loop:
- mova m0, [srcq+lenq]
- S16_TO_S32_SX 0, 1
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- mulps m0, m2
- mulps m1, m2
- mova [dstq+2*lenq ], m0
- mova [dstq+2*lenq+mmsize], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16_TO_FLT
-INIT_XMM sse4
-CONV_S16_TO_FLT
-
-;------------------------------------------------------------------------------
-; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len);
-;------------------------------------------------------------------------------
-
-%macro CONV_S32_TO_S16 0
-cglobal conv_s32_to_s16, 3,3,4, dst, src, len
- lea lenq, [2*lend]
- lea srcq, [srcq+2*lenq]
- add dstq, lenq
- neg lenq
-.loop:
- mova m0, [srcq+2*lenq ]
- mova m1, [srcq+2*lenq+ mmsize]
- mova m2, [srcq+2*lenq+2*mmsize]
- mova m3, [srcq+2*lenq+3*mmsize]
- psrad m0, 16
- psrad m1, 16
- psrad m2, 16
- psrad m3, 16
- packssdw m0, m1
- packssdw m2, m3
- mova [dstq+lenq ], m0
- mova [dstq+lenq+mmsize], m2
- add lenq, mmsize*2
- jl .loop
-%if mmsize == 8
- emms
- RET
-%else
- REP_RET
-%endif
-%endmacro
-
-INIT_MMX mmx
-CONV_S32_TO_S16
-INIT_XMM sse2
-CONV_S32_TO_S16
-
-;------------------------------------------------------------------------------
-; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len);
-;------------------------------------------------------------------------------
-
-%macro CONV_S32_TO_FLT 0
-cglobal conv_s32_to_flt, 3,3,3, dst, src, len
- lea lenq, [4*lend]
- add srcq, lenq
- add dstq, lenq
- neg lenq
- mova m0, [pf_s32_inv_scale]
- ALIGN 16
-.loop:
- cvtdq2ps m1, [srcq+lenq ]
- cvtdq2ps m2, [srcq+lenq+mmsize]
- mulps m1, m1, m0
- mulps m2, m2, m0
- mova [dstq+lenq ], m1
- mova [dstq+lenq+mmsize], m2
- add lenq, mmsize*2
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S32_TO_FLT
-%if HAVE_AVX_EXTERNAL
-INIT_YMM avx
-CONV_S32_TO_FLT
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len);
-;------------------------------------------------------------------------------
-
-INIT_XMM sse2
-cglobal conv_flt_to_s16, 3,3,5, dst, src, len
- lea lenq, [2*lend]
- lea srcq, [srcq+2*lenq]
- add dstq, lenq
- neg lenq
- mova m4, [pf_s16_scale]
-.loop:
- mova m0, [srcq+2*lenq ]
- mova m1, [srcq+2*lenq+1*mmsize]
- mova m2, [srcq+2*lenq+2*mmsize]
- mova m3, [srcq+2*lenq+3*mmsize]
- mulps m0, m4
- mulps m1, m4
- mulps m2, m4
- mulps m3, m4
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- packssdw m0, m1
- packssdw m2, m3
- mova [dstq+lenq ], m0
- mova [dstq+lenq+mmsize], m2
- add lenq, mmsize*2
- jl .loop
- REP_RET
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLT_TO_S32 0
-cglobal conv_flt_to_s32, 3,3,6, dst, src, len
- lea lenq, [lend*4]
- add srcq, lenq
- add dstq, lenq
- neg lenq
- mova m4, [pf_s32_scale]
- mova m5, [pf_s32_clip]
-.loop:
- mulps m0, m4, [srcq+lenq ]
- mulps m1, m4, [srcq+lenq+1*mmsize]
- mulps m2, m4, [srcq+lenq+2*mmsize]
- mulps m3, m4, [srcq+lenq+3*mmsize]
- minps m0, m0, m5
- minps m1, m1, m5
- minps m2, m2, m5
- minps m3, m3, m5
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- mova [dstq+lenq ], m0
- mova [dstq+lenq+1*mmsize], m1
- mova [dstq+lenq+2*mmsize], m2
- mova [dstq+lenq+3*mmsize], m3
- add lenq, mmsize*4
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_FLT_TO_S32
-%if HAVE_AVX_EXTERNAL
-INIT_YMM avx
-CONV_FLT_TO_S32
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16p_to_s16_2ch(int16_t *dst, int16_t *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16P_TO_S16_2CH 0
-cglobal conv_s16p_to_s16_2ch, 3,4,5, dst, src0, len, src1
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q ]
- lea lenq, [2*lend]
- add src0q, lenq
- add src1q, lenq
- lea dstq, [dstq+2*lenq]
- neg lenq
-.loop:
- mova m0, [src0q+lenq ]
- mova m1, [src1q+lenq ]
- mova m2, [src0q+lenq+mmsize]
- mova m3, [src1q+lenq+mmsize]
- SBUTTERFLY2 wd, 0, 1, 4
- SBUTTERFLY2 wd, 2, 3, 4
- mova [dstq+2*lenq+0*mmsize], m0
- mova [dstq+2*lenq+1*mmsize], m1
- mova [dstq+2*lenq+2*mmsize], m2
- mova [dstq+2*lenq+3*mmsize], m3
- add lenq, 2*mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16P_TO_S16_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16P_TO_S16_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16p_to_s16_6ch(int16_t *dst, int16_t *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-;------------------------------------------------------------------------------
-; NOTE: In the 6-channel functions, len could be used as an index on x86-64
-; instead of just a counter, which would avoid incrementing the
-; pointers, but the extra complexity and amount of code is not worth
-; the small gain. On x86-32 there are not enough registers to use len
-; as an index without keeping two of the pointers on the stack and
-; loading them in each iteration.
-;------------------------------------------------------------------------------
-
-%macro CONV_S16P_TO_S16_6CH 0
-%if ARCH_X86_64
-cglobal conv_s16p_to_s16_6ch, 3,8,7, dst, src0, len, src1, src2, src3, src4, src5
-%else
-cglobal conv_s16p_to_s16_6ch, 2,7,7, dst, src0, src1, src2, src3, src4, src5
-%define lend dword r2m
-%endif
- mov src1q, [src0q+1*gprsize]
- mov src2q, [src0q+2*gprsize]
- mov src3q, [src0q+3*gprsize]
- mov src4q, [src0q+4*gprsize]
- mov src5q, [src0q+5*gprsize]
- mov src0q, [src0q]
- sub src1q, src0q
- sub src2q, src0q
- sub src3q, src0q
- sub src4q, src0q
- sub src5q, src0q
-.loop:
-%if cpuflag(sse2slow)
- movq m0, [src0q ] ; m0 = 0, 6, 12, 18, x, x, x, x
- movq m1, [src0q+src1q] ; m1 = 1, 7, 13, 19, x, x, x, x
- movq m2, [src0q+src2q] ; m2 = 2, 8, 14, 20, x, x, x, x
- movq m3, [src0q+src3q] ; m3 = 3, 9, 15, 21, x, x, x, x
- movq m4, [src0q+src4q] ; m4 = 4, 10, 16, 22, x, x, x, x
- movq m5, [src0q+src5q] ; m5 = 5, 11, 17, 23, x, x, x, x
- ; unpack words:
- punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19
- punpcklwd m2, m3 ; m2 = 4, 5, 10, 11, 16, 17, 22, 23
- punpcklwd m4, m5 ; m4 = 2, 3, 8, 9, 14, 15, 20, 21
- ; blend dwords
- shufps m1, m0, m2, q2020 ; m1 = 0, 1, 12, 13, 2, 3, 14, 15
- shufps m0, m4, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17
- shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23
- ; shuffle dwords
- pshufd m0, m0, q1302 ; m0 = 4, 5, 6, 7, 16, 17, 18, 19
- pshufd m1, m1, q3120 ; m1 = 0, 1, 2, 3, 12, 13, 14, 15
- pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23
- movq [dstq+0*mmsize/2], m1
- movq [dstq+1*mmsize/2], m0
- movq [dstq+2*mmsize/2], m2
- movhps [dstq+3*mmsize/2], m1
- movhps [dstq+4*mmsize/2], m0
- movhps [dstq+5*mmsize/2], m2
- add src0q, mmsize/2
- add dstq, mmsize*3
- sub lend, mmsize/4
-%else
- mova m0, [src0q ] ; m0 = 0, 6, 12, 18, 24, 30, 36, 42
- mova m1, [src0q+src1q] ; m1 = 1, 7, 13, 19, 25, 31, 37, 43
- mova m2, [src0q+src2q] ; m2 = 2, 8, 14, 20, 26, 32, 38, 44
- mova m3, [src0q+src3q] ; m3 = 3, 9, 15, 21, 27, 33, 39, 45
- mova m4, [src0q+src4q] ; m4 = 4, 10, 16, 22, 28, 34, 40, 46
- mova m5, [src0q+src5q] ; m5 = 5, 11, 17, 23, 29, 35, 41, 47
- ; unpack words:
- SBUTTERFLY2 wd, 0, 1, 6 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19
- ; m1 = 24, 25, 30, 31, 36, 37, 42, 43
- SBUTTERFLY2 wd, 2, 3, 6 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21
- ; m3 = 26, 27, 32, 33, 38, 39, 44, 45
- SBUTTERFLY2 wd, 4, 5, 6 ; m4 = 4, 5, 10, 11, 16, 17, 22, 23
- ; m5 = 28, 29, 34, 35, 40, 41, 46, 47
- ; blend dwords
- shufps m6, m0, m2, q2020 ; m6 = 0, 1, 12, 13, 2, 3, 14, 15
- shufps m0, m4, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17
- shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23
- SWAP 4,6 ; m4 = 0, 1, 12, 13, 2, 3, 14, 15
- shufps m6, m1, m3, q2020 ; m6 = 24, 25, 36, 37, 26, 27, 38, 39
- shufps m1, m5, q2031 ; m1 = 30, 31, 42, 43, 28, 29, 40, 41
- shufps m3, m5, q3131 ; m3 = 32, 33, 44, 45, 34, 35, 46, 47
- SWAP 5,6 ; m5 = 24, 25, 36, 37, 26, 27, 38, 39
- ; shuffle dwords
- pshufd m0, m0, q1302 ; m0 = 4, 5, 6, 7, 16, 17, 18, 19
- pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23
- pshufd m4, m4, q3120 ; m4 = 0, 1, 2, 3, 12, 13, 14, 15
- pshufd m1, m1, q1302 ; m1 = 28, 29, 30, 31, 40, 41, 42, 43
- pshufd m3, m3, q3120 ; m3 = 32, 33, 34, 35, 44, 45, 46, 47
- pshufd m5, m5, q3120 ; m5 = 24, 25, 26, 27, 36, 37, 38, 39
- ; shuffle qwords
- punpcklqdq m6, m4, m0 ; m6 = 0, 1, 2, 3, 4, 5, 6, 7
- punpckhqdq m0, m2 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
- shufps m2, m4, q3210 ; m2 = 8, 9, 10, 11, 12, 13, 14, 15
- SWAP 4,6 ; m4 = 0, 1, 2, 3, 4, 5, 6, 7
- punpcklqdq m6, m5, m1 ; m6 = 24, 25, 26, 27, 28, 29, 30, 31
- punpckhqdq m1, m3 ; m1 = 40, 41, 42, 43, 44, 45, 46, 47
- shufps m3, m5, q3210 ; m3 = 32, 33, 34, 35, 36, 37, 38, 39
- SWAP 5,6 ; m5 = 24, 25, 26, 27, 28, 29, 30, 31
- mova [dstq+0*mmsize], m4
- mova [dstq+1*mmsize], m2
- mova [dstq+2*mmsize], m0
- mova [dstq+3*mmsize], m5
- mova [dstq+4*mmsize], m3
- mova [dstq+5*mmsize], m1
- add src0q, mmsize
- add dstq, mmsize*6
- sub lend, mmsize/2
-%endif
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16P_TO_S16_6CH
-INIT_XMM sse2slow
-CONV_S16P_TO_S16_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16P_TO_S16_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16p_to_flt_2ch(float *dst, int16_t *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16P_TO_FLT_2CH 0
-cglobal conv_s16p_to_flt_2ch, 3,4,6, dst, src0, len, src1
- lea lenq, [2*lend]
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q ]
- lea dstq, [dstq+4*lenq]
- add src0q, lenq
- add src1q, lenq
- neg lenq
- mova m5, [pf_s32_inv_scale]
-.loop:
- mova m2, [src0q+lenq] ; m2 = 0, 2, 4, 6, 8, 10, 12, 14
- mova m4, [src1q+lenq] ; m4 = 1, 3, 5, 7, 9, 11, 13, 15
- SBUTTERFLY2 wd, 2, 4, 3 ; m2 = 0, 1, 2, 3, 4, 5, 6, 7
- ; m4 = 8, 9, 10, 11, 12, 13, 14, 15
- pxor m3, m3
- punpcklwd m0, m3, m2 ; m0 = 0, 1, 2, 3
- punpckhwd m1, m3, m2 ; m1 = 4, 5, 6, 7
- punpcklwd m2, m3, m4 ; m2 = 8, 9, 10, 11
- punpckhwd m3, m4 ; m3 = 12, 13, 14, 15
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- cvtdq2ps m2, m2
- cvtdq2ps m3, m3
- mulps m0, m5
- mulps m1, m5
- mulps m2, m5
- mulps m3, m5
- mova [dstq+4*lenq ], m0
- mova [dstq+4*lenq+ mmsize], m1
- mova [dstq+4*lenq+2*mmsize], m2
- mova [dstq+4*lenq+3*mmsize], m3
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16P_TO_FLT_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16P_TO_FLT_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16p_to_flt_6ch(float *dst, int16_t *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16P_TO_FLT_6CH 0
-%if ARCH_X86_64
-cglobal conv_s16p_to_flt_6ch, 3,8,8, dst, src, len, src1, src2, src3, src4, src5
-%else
-cglobal conv_s16p_to_flt_6ch, 2,7,8, dst, src, src1, src2, src3, src4, src5
-%define lend dword r2m
-%endif
- mov src1q, [srcq+1*gprsize]
- mov src2q, [srcq+2*gprsize]
- mov src3q, [srcq+3*gprsize]
- mov src4q, [srcq+4*gprsize]
- mov src5q, [srcq+5*gprsize]
- mov srcq, [srcq]
- sub src1q, srcq
- sub src2q, srcq
- sub src3q, srcq
- sub src4q, srcq
- sub src5q, srcq
- mova m7, [pf_s32_inv_scale]
-%if cpuflag(ssse3)
- %define unpack_even m6
- mova m6, [pb_shuf_unpack_even]
-%if ARCH_X86_64
- %define unpack_odd m8
- mova m8, [pb_shuf_unpack_odd]
-%else
- %define unpack_odd [pb_shuf_unpack_odd]
-%endif
-%endif
-.loop:
- movq m0, [srcq ] ; m0 = 0, 6, 12, 18, x, x, x, x
- movq m1, [srcq+src1q] ; m1 = 1, 7, 13, 19, x, x, x, x
- movq m2, [srcq+src2q] ; m2 = 2, 8, 14, 20, x, x, x, x
- movq m3, [srcq+src3q] ; m3 = 3, 9, 15, 21, x, x, x, x
- movq m4, [srcq+src4q] ; m4 = 4, 10, 16, 22, x, x, x, x
- movq m5, [srcq+src5q] ; m5 = 5, 11, 17, 23, x, x, x, x
- ; unpack words:
- punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19
- punpcklwd m2, m3 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21
- punpcklwd m4, m5 ; m4 = 4, 5, 10, 11, 16, 17, 22, 23
- ; blend dwords
- shufps m1, m4, m0, q3120 ; m1 = 4, 5, 16, 17, 6, 7, 18, 19
- shufps m0, m2, q2020 ; m0 = 0, 1, 12, 13, 2, 3, 14, 15
- shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23
-%if cpuflag(ssse3)
- pshufb m3, m0, unpack_odd ; m3 = 12, 13, 14, 15
- pshufb m0, unpack_even ; m0 = 0, 1, 2, 3
- pshufb m4, m1, unpack_odd ; m4 = 16, 17, 18, 19
- pshufb m1, unpack_even ; m1 = 4, 5, 6, 7
- pshufb m5, m2, unpack_odd ; m5 = 20, 21, 22, 23
- pshufb m2, unpack_even ; m2 = 8, 9, 10, 11
-%else
- ; shuffle dwords
- pshufd m0, m0, q3120 ; m0 = 0, 1, 2, 3, 12, 13, 14, 15
- pshufd m1, m1, q3120 ; m1 = 4, 5, 6, 7, 16, 17, 18, 19
- pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23
- pxor m6, m6 ; convert s16 in m0-m2 to s32 in m0-m5
- punpcklwd m3, m6, m0 ; m3 = 0, 1, 2, 3
- punpckhwd m4, m6, m0 ; m4 = 12, 13, 14, 15
- punpcklwd m0, m6, m1 ; m0 = 4, 5, 6, 7
- punpckhwd m5, m6, m1 ; m5 = 16, 17, 18, 19
- punpcklwd m1, m6, m2 ; m1 = 8, 9, 10, 11
- punpckhwd m6, m2 ; m6 = 20, 21, 22, 23
- SWAP 6,2,1,0,3,4,5 ; swap registers 3,0,1,4,5,6 to 0,1,2,3,4,5
-%endif
- cvtdq2ps m0, m0 ; convert s32 to float
- cvtdq2ps m1, m1
- cvtdq2ps m2, m2
- cvtdq2ps m3, m3
- cvtdq2ps m4, m4
- cvtdq2ps m5, m5
- mulps m0, m7 ; scale float from s32 range to [-1.0,1.0]
- mulps m1, m7
- mulps m2, m7
- mulps m3, m7
- mulps m4, m7
- mulps m5, m7
- mova [dstq ], m0
- mova [dstq+ mmsize], m1
- mova [dstq+2*mmsize], m2
- mova [dstq+3*mmsize], m3
- mova [dstq+4*mmsize], m4
- mova [dstq+5*mmsize], m5
- add srcq, mmsize/2
- add dstq, mmsize*6
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16P_TO_FLT_6CH
-INIT_XMM ssse3
-CONV_S16P_TO_FLT_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16P_TO_FLT_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_fltp_to_s16_2ch(int16_t *dst, float *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLTP_TO_S16_2CH 0
-cglobal conv_fltp_to_s16_2ch, 3,4,3, dst, src0, len, src1
- lea lenq, [4*lend]
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q ]
- add dstq, lenq
- add src0q, lenq
- add src1q, lenq
- neg lenq
- mova m2, [pf_s16_scale]
-%if cpuflag(ssse3)
- mova m3, [pb_interleave_words]
-%endif
-.loop:
- mulps m0, m2, [src0q+lenq] ; m0 = 0, 2, 4, 6
- mulps m1, m2, [src1q+lenq] ; m1 = 1, 3, 5, 7
- cvtps2dq m0, m0
- cvtps2dq m1, m1
-%if cpuflag(ssse3)
- packssdw m0, m1 ; m0 = 0, 2, 4, 6, 1, 3, 5, 7
- pshufb m0, m3 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
-%else
- packssdw m0, m0 ; m0 = 0, 2, 4, 6, x, x, x, x
- packssdw m1, m1 ; m1 = 1, 3, 5, 7, x, x, x, x
- punpcklwd m0, m1 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
-%endif
- mova [dstq+lenq], m0
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_FLTP_TO_S16_2CH
-INIT_XMM ssse3
-CONV_FLTP_TO_S16_2CH
-
-;------------------------------------------------------------------------------
-; void ff_conv_fltp_to_s16_6ch(int16_t *dst, float *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLTP_TO_S16_6CH 0
-%if ARCH_X86_64
-cglobal conv_fltp_to_s16_6ch, 3,8,7, dst, src, len, src1, src2, src3, src4, src5
-%else
-cglobal conv_fltp_to_s16_6ch, 2,7,7, dst, src, src1, src2, src3, src4, src5
-%define lend dword r2m
-%endif
- mov src1q, [srcq+1*gprsize]
- mov src2q, [srcq+2*gprsize]
- mov src3q, [srcq+3*gprsize]
- mov src4q, [srcq+4*gprsize]
- mov src5q, [srcq+5*gprsize]
- mov srcq, [srcq]
- sub src1q, srcq
- sub src2q, srcq
- sub src3q, srcq
- sub src4q, srcq
- sub src5q, srcq
- movaps xmm6, [pf_s16_scale]
-.loop:
-%if cpuflag(sse2)
- mulps m0, m6, [srcq ]
- mulps m1, m6, [srcq+src1q]
- mulps m2, m6, [srcq+src2q]
- mulps m3, m6, [srcq+src3q]
- mulps m4, m6, [srcq+src4q]
- mulps m5, m6, [srcq+src5q]
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- cvtps2dq m4, m4
- cvtps2dq m5, m5
- packssdw m0, m3 ; m0 = 0, 6, 12, 18, 3, 9, 15, 21
- packssdw m1, m4 ; m1 = 1, 7, 13, 19, 4, 10, 16, 22
- packssdw m2, m5 ; m2 = 2, 8, 14, 20, 5, 11, 17, 23
- ; unpack words:
- movhlps m3, m0 ; m3 = 3, 9, 15, 21, x, x, x, x
- punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19
- punpckhwd m1, m2 ; m1 = 4, 5, 10, 11, 16, 17, 22, 23
- punpcklwd m2, m3 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21
- ; blend dwords:
- shufps m3, m0, m2, q2020 ; m3 = 0, 1, 12, 13, 2, 3, 14, 15
- shufps m0, m1, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17
- shufps m2, m1, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23
- ; shuffle dwords:
- shufps m1, m2, m3, q3120 ; m1 = 8, 9, 10, 11, 12, 13, 14, 15
- shufps m3, m0, q0220 ; m3 = 0, 1, 2, 3, 4, 5, 6, 7
- shufps m0, m2, q3113 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
- mova [dstq+0*mmsize], m3
- mova [dstq+1*mmsize], m1
- mova [dstq+2*mmsize], m0
-%else ; sse
- movlps xmm0, [srcq ]
- movlps xmm1, [srcq+src1q]
- movlps xmm2, [srcq+src2q]
- movlps xmm3, [srcq+src3q]
- movlps xmm4, [srcq+src4q]
- movlps xmm5, [srcq+src5q]
- mulps xmm0, xmm6
- mulps xmm1, xmm6
- mulps xmm2, xmm6
- mulps xmm3, xmm6
- mulps xmm4, xmm6
- mulps xmm5, xmm6
- cvtps2pi mm0, xmm0
- cvtps2pi mm1, xmm1
- cvtps2pi mm2, xmm2
- cvtps2pi mm3, xmm3
- cvtps2pi mm4, xmm4
- cvtps2pi mm5, xmm5
- packssdw mm0, mm3 ; m0 = 0, 6, 3, 9
- packssdw mm1, mm4 ; m1 = 1, 7, 4, 10
- packssdw mm2, mm5 ; m2 = 2, 8, 5, 11
- ; unpack words
- pshufw mm3, mm0, q1032 ; m3 = 3, 9, 0, 6
- punpcklwd mm0, mm1 ; m0 = 0, 1, 6, 7
- punpckhwd mm1, mm2 ; m1 = 4, 5, 10, 11
- punpcklwd mm2, mm3 ; m2 = 2, 3, 8, 9
- ; unpack dwords
- pshufw mm3, mm0, q1032 ; m3 = 6, 7, 0, 1
- punpckldq mm0, mm2 ; m0 = 0, 1, 2, 3 (final)
- punpckhdq mm2, mm1 ; m2 = 8, 9, 10, 11 (final)
- punpckldq mm1, mm3 ; m1 = 4, 5, 6, 7 (final)
- mova [dstq+0*mmsize], mm0
- mova [dstq+1*mmsize], mm1
- mova [dstq+2*mmsize], mm2
-%endif
- add srcq, mmsize
- add dstq, mmsize*3
- sub lend, mmsize/4
- jg .loop
-%if mmsize == 8
- emms
- RET
-%else
- REP_RET
-%endif
-%endmacro
-
-INIT_MMX sse
-CONV_FLTP_TO_S16_6CH
-INIT_XMM sse2
-CONV_FLTP_TO_S16_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLTP_TO_S16_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_fltp_to_flt_2ch(float *dst, float *const *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLTP_TO_FLT_2CH 0
-cglobal conv_fltp_to_flt_2ch, 3,4,5, dst, src0, len, src1
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q]
- lea lenq, [4*lend]
- add src0q, lenq
- add src1q, lenq
- lea dstq, [dstq+2*lenq]
- neg lenq
-.loop:
- mova m0, [src0q+lenq ]
- mova m1, [src1q+lenq ]
- mova m2, [src0q+lenq+mmsize]
- mova m3, [src1q+lenq+mmsize]
- SBUTTERFLYPS 0, 1, 4
- SBUTTERFLYPS 2, 3, 4
- mova [dstq+2*lenq+0*mmsize], m0
- mova [dstq+2*lenq+1*mmsize], m1
- mova [dstq+2*lenq+2*mmsize], m2
- mova [dstq+2*lenq+3*mmsize], m3
- add lenq, 2*mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse
-CONV_FLTP_TO_FLT_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLTP_TO_FLT_2CH
-%endif
-
-;-----------------------------------------------------------------------------
-; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len,
-; int channels);
-;-----------------------------------------------------------------------------
-
-%macro CONV_FLTP_TO_FLT_6CH 0
-cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len
-%if ARCH_X86_64
- mov lend, r2d
-%else
- %define lend dword r2m
-%endif
- mov src1q, [srcq+1*gprsize]
- mov src2q, [srcq+2*gprsize]
- mov src3q, [srcq+3*gprsize]
- mov src4q, [srcq+4*gprsize]
- mov src5q, [srcq+5*gprsize]
- mov srcq, [srcq]
- sub src1q, srcq
- sub src2q, srcq
- sub src3q, srcq
- sub src4q, srcq
- sub src5q, srcq
-.loop:
- mova m0, [srcq ]
- mova m1, [srcq+src1q]
- mova m2, [srcq+src2q]
- mova m3, [srcq+src3q]
- mova m4, [srcq+src4q]
- mova m5, [srcq+src5q]
-%if cpuflag(sse4)
- SBUTTERFLYPS 0, 1, 6
- SBUTTERFLYPS 2, 3, 6
- SBUTTERFLYPS 4, 5, 6
-
- blendps m6, m4, m0, 1100b
- movlhps m0, m2
- movhlps m4, m2
- blendps m2, m5, m1, 1100b
- movlhps m1, m3
- movhlps m5, m3
-
- movaps [dstq ], m0
- movaps [dstq+16], m6
- movaps [dstq+32], m4
- movaps [dstq+48], m1
- movaps [dstq+64], m2
- movaps [dstq+80], m5
-%else ; mmx
- SBUTTERFLY dq, 0, 1, 6
- SBUTTERFLY dq, 2, 3, 6
- SBUTTERFLY dq, 4, 5, 6
-
- movq [dstq ], m0
- movq [dstq+ 8], m2
- movq [dstq+16], m4
- movq [dstq+24], m1
- movq [dstq+32], m3
- movq [dstq+40], m5
-%endif
- add srcq, mmsize
- add dstq, mmsize*6
- sub lend, mmsize/4
- jg .loop
-%if mmsize == 8
- emms
- RET
-%else
- REP_RET
-%endif
-%endmacro
-
-INIT_MMX mmx
-CONV_FLTP_TO_FLT_6CH
-INIT_XMM sse4
-CONV_FLTP_TO_FLT_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLTP_TO_FLT_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_s16p_2ch(int16_t *const *dst, int16_t *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16_TO_S16P_2CH 0
-cglobal conv_s16_to_s16p_2ch, 3,4,4, dst0, src, len, dst1
- lea lenq, [2*lend]
- mov dst1q, [dst0q+gprsize]
- mov dst0q, [dst0q ]
- lea srcq, [srcq+2*lenq]
- add dst0q, lenq
- add dst1q, lenq
- neg lenq
-%if cpuflag(ssse3)
- mova m3, [pb_deinterleave_words]
-%endif
-.loop:
- mova m0, [srcq+2*lenq ] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
- mova m1, [srcq+2*lenq+mmsize] ; m1 = 8, 9, 10, 11, 12, 13, 14, 15
-%if cpuflag(ssse3)
- pshufb m0, m3 ; m0 = 0, 2, 4, 6, 1, 3, 5, 7
- pshufb m1, m3 ; m1 = 8, 10, 12, 14, 9, 11, 13, 15
- SBUTTERFLY2 qdq, 0, 1, 2 ; m0 = 0, 2, 4, 6, 8, 10, 12, 14
- ; m1 = 1, 3, 5, 7, 9, 11, 13, 15
-%else ; sse2
- pshuflw m0, m0, q3120 ; m0 = 0, 2, 1, 3, 4, 5, 6, 7
- pshufhw m0, m0, q3120 ; m0 = 0, 2, 1, 3, 4, 6, 5, 7
- pshuflw m1, m1, q3120 ; m1 = 8, 10, 9, 11, 12, 13, 14, 15
- pshufhw m1, m1, q3120 ; m1 = 8, 10, 9, 11, 12, 14, 13, 15
- DEINT2_PS 0, 1, 2 ; m0 = 0, 2, 4, 6, 8, 10, 12, 14
- ; m1 = 1, 3, 5, 7, 9, 11, 13, 15
-%endif
- mova [dst0q+lenq], m0
- mova [dst1q+lenq], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16_TO_S16P_2CH
-INIT_XMM ssse3
-CONV_S16_TO_S16P_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16_TO_S16P_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_s16p_6ch(int16_t *const *dst, int16_t *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16_TO_S16P_6CH 0
-%if ARCH_X86_64
-cglobal conv_s16_to_s16p_6ch, 3,8,5, dst, src, len, dst1, dst2, dst3, dst4, dst5
-%else
-cglobal conv_s16_to_s16p_6ch, 2,7,5, dst, src, dst1, dst2, dst3, dst4, dst5
-%define lend dword r2m
-%endif
- mov dst1q, [dstq+ gprsize]
- mov dst2q, [dstq+2*gprsize]
- mov dst3q, [dstq+3*gprsize]
- mov dst4q, [dstq+4*gprsize]
- mov dst5q, [dstq+5*gprsize]
- mov dstq, [dstq ]
- sub dst1q, dstq
- sub dst2q, dstq
- sub dst3q, dstq
- sub dst4q, dstq
- sub dst5q, dstq
-.loop:
- mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
- mova m3, [srcq+1*mmsize] ; m3 = 8, 9, 10, 11, 12, 13, 14, 15
- mova m2, [srcq+2*mmsize] ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
- PALIGNR m1, m3, m0, 12, m4 ; m1 = 6, 7, 8, 9, 10, 11, x, x
- shufps m3, m2, q1032 ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
- psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x
- SBUTTERFLY2 wd, 0, 1, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9
- ; m1 = 4, 10, 5, 11, x, x, x, x
- SBUTTERFLY2 wd, 3, 2, 4 ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
- ; m2 = 16, 22, 17, 23, x, x, x, x
- SBUTTERFLY2 dq, 0, 3, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19
- ; m3 = 2, 8, 14, 20, 3, 9, 15, 21
- punpckldq m1, m2 ; m1 = 4, 10, 16, 22, 5, 11, 17, 23
- movq [dstq ], m0
- movhps [dstq+dst1q], m0
- movq [dstq+dst2q], m3
- movhps [dstq+dst3q], m3
- movq [dstq+dst4q], m1
- movhps [dstq+dst5q], m1
- add srcq, mmsize*3
- add dstq, mmsize/2
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16_TO_S16P_6CH
-INIT_XMM ssse3
-CONV_S16_TO_S16P_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16_TO_S16P_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_fltp_2ch(float *const *dst, int16_t *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16_TO_FLTP_2CH 0
-cglobal conv_s16_to_fltp_2ch, 3,4,5, dst0, src, len, dst1
- lea lenq, [4*lend]
- mov dst1q, [dst0q+gprsize]
- mov dst0q, [dst0q ]
- add srcq, lenq
- add dst0q, lenq
- add dst1q, lenq
- neg lenq
- mova m3, [pf_s32_inv_scale]
- mova m4, [pw_zero_even]
-.loop:
- mova m1, [srcq+lenq]
- pslld m0, m1, 16
- pand m1, m4
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- mulps m0, m0, m3
- mulps m1, m1, m3
- mova [dst0q+lenq], m0
- mova [dst1q+lenq], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16_TO_FLTP_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16_TO_FLTP_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_s16_to_fltp_6ch(float *const *dst, int16_t *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_S16_TO_FLTP_6CH 0
-%if ARCH_X86_64
-cglobal conv_s16_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
-%else
-cglobal conv_s16_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
-%define lend dword r2m
-%endif
- mov dst1q, [dstq+ gprsize]
- mov dst2q, [dstq+2*gprsize]
- mov dst3q, [dstq+3*gprsize]
- mov dst4q, [dstq+4*gprsize]
- mov dst5q, [dstq+5*gprsize]
- mov dstq, [dstq ]
- sub dst1q, dstq
- sub dst2q, dstq
- sub dst3q, dstq
- sub dst4q, dstq
- sub dst5q, dstq
- mova m6, [pf_s16_inv_scale]
-.loop:
- mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
- mova m3, [srcq+1*mmsize] ; m3 = 8, 9, 10, 11, 12, 13, 14, 15
- mova m2, [srcq+2*mmsize] ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
- PALIGNR m1, m3, m0, 12, m4 ; m1 = 6, 7, 8, 9, 10, 11, x, x
- shufps m3, m2, q1032 ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
- psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x
- SBUTTERFLY2 wd, 0, 1, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9
- ; m1 = 4, 10, 5, 11, x, x, x, x
- SBUTTERFLY2 wd, 3, 2, 4 ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
- ; m2 = 16, 22, 17, 23, x, x, x, x
- SBUTTERFLY2 dq, 0, 3, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19
- ; m3 = 2, 8, 14, 20, 3, 9, 15, 21
- punpckldq m1, m2 ; m1 = 4, 10, 16, 22, 5, 11, 17, 23
- S16_TO_S32_SX 0, 2 ; m0 = 0, 6, 12, 18
- ; m2 = 1, 7, 13, 19
- S16_TO_S32_SX 3, 4 ; m3 = 2, 8, 14, 20
- ; m4 = 3, 9, 15, 21
- S16_TO_S32_SX 1, 5 ; m1 = 4, 10, 16, 22
- ; m5 = 5, 11, 17, 23
- SWAP 1,2,3,4
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- cvtdq2ps m2, m2
- cvtdq2ps m3, m3
- cvtdq2ps m4, m4
- cvtdq2ps m5, m5
- mulps m0, m6
- mulps m1, m6
- mulps m2, m6
- mulps m3, m6
- mulps m4, m6
- mulps m5, m6
- mova [dstq ], m0
- mova [dstq+dst1q], m1
- mova [dstq+dst2q], m2
- mova [dstq+dst3q], m3
- mova [dstq+dst4q], m4
- mova [dstq+dst5q], m5
- add srcq, mmsize*3
- add dstq, mmsize
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_S16_TO_FLTP_6CH
-INIT_XMM ssse3
-CONV_S16_TO_FLTP_6CH
-INIT_XMM sse4
-CONV_S16_TO_FLTP_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_S16_TO_FLTP_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_s16p_2ch(int16_t *const *dst, float *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLT_TO_S16P_2CH 0
-cglobal conv_flt_to_s16p_2ch, 3,4,6, dst0, src, len, dst1
- lea lenq, [2*lend]
- mov dst1q, [dst0q+gprsize]
- mov dst0q, [dst0q ]
- lea srcq, [srcq+4*lenq]
- add dst0q, lenq
- add dst1q, lenq
- neg lenq
- mova m5, [pf_s16_scale]
-.loop:
- mova m0, [srcq+4*lenq ]
- mova m1, [srcq+4*lenq+ mmsize]
- mova m2, [srcq+4*lenq+2*mmsize]
- mova m3, [srcq+4*lenq+3*mmsize]
- DEINT2_PS 0, 1, 4
- DEINT2_PS 2, 3, 4
- mulps m0, m0, m5
- mulps m1, m1, m5
- mulps m2, m2, m5
- mulps m3, m3, m5
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- packssdw m0, m2
- packssdw m1, m3
- mova [dst0q+lenq], m0
- mova [dst1q+lenq], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_FLT_TO_S16P_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLT_TO_S16P_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_s16p_6ch(int16_t *const *dst, float *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLT_TO_S16P_6CH 0
-%if ARCH_X86_64
-cglobal conv_flt_to_s16p_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
-%else
-cglobal conv_flt_to_s16p_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
-%define lend dword r2m
-%endif
- mov dst1q, [dstq+ gprsize]
- mov dst2q, [dstq+2*gprsize]
- mov dst3q, [dstq+3*gprsize]
- mov dst4q, [dstq+4*gprsize]
- mov dst5q, [dstq+5*gprsize]
- mov dstq, [dstq ]
- sub dst1q, dstq
- sub dst2q, dstq
- sub dst3q, dstq
- sub dst4q, dstq
- sub dst5q, dstq
- mova m6, [pf_s16_scale]
-.loop:
- mulps m0, m6, [srcq+0*mmsize]
- mulps m3, m6, [srcq+1*mmsize]
- mulps m1, m6, [srcq+2*mmsize]
- mulps m4, m6, [srcq+3*mmsize]
- mulps m2, m6, [srcq+4*mmsize]
- mulps m5, m6, [srcq+5*mmsize]
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- cvtps2dq m4, m4
- cvtps2dq m5, m5
- packssdw m0, m3 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
- packssdw m1, m4 ; m1 = 8, 9, 10, 11, 12, 13, 14, 15
- packssdw m2, m5 ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
- PALIGNR m3, m1, m0, 12, m4 ; m3 = 6, 7, 8, 9, 10, 11, x, x
- shufps m1, m2, q1032 ; m1 = 12, 13, 14, 15, 16, 17, 18, 19
- psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x
- SBUTTERFLY2 wd, 0, 3, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9
- ; m3 = 4, 10, 5, 11, x, x, x, x
- SBUTTERFLY2 wd, 1, 2, 4 ; m1 = 12, 18, 13, 19, 14, 20, 15, 21
- ; m2 = 16, 22, 17, 23, x, x, x, x
- SBUTTERFLY2 dq, 0, 1, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19
- ; m1 = 2, 8, 14, 20, 3, 9, 15, 21
- punpckldq m3, m2 ; m3 = 4, 10, 16, 22, 5, 11, 17, 23
- movq [dstq ], m0
- movhps [dstq+dst1q], m0
- movq [dstq+dst2q], m1
- movhps [dstq+dst3q], m1
- movq [dstq+dst4q], m3
- movhps [dstq+dst5q], m3
- add srcq, mmsize*6
- add dstq, mmsize/2
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_FLT_TO_S16P_6CH
-INIT_XMM ssse3
-CONV_FLT_TO_S16P_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLT_TO_S16P_6CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_fltp_2ch(float *const *dst, float *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLT_TO_FLTP_2CH 0
-cglobal conv_flt_to_fltp_2ch, 3,4,3, dst0, src, len, dst1
- lea lenq, [4*lend]
- mov dst1q, [dst0q+gprsize]
- mov dst0q, [dst0q ]
- lea srcq, [srcq+2*lenq]
- add dst0q, lenq
- add dst1q, lenq
- neg lenq
-.loop:
- mova m0, [srcq+2*lenq ]
- mova m1, [srcq+2*lenq+mmsize]
- DEINT2_PS 0, 1, 2
- mova [dst0q+lenq], m0
- mova [dst1q+lenq], m1
- add lenq, mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse
-CONV_FLT_TO_FLTP_2CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLT_TO_FLTP_2CH
-%endif
-
-;------------------------------------------------------------------------------
-; void ff_conv_flt_to_fltp_6ch(float *const *dst, float *src, int len,
-; int channels);
-;------------------------------------------------------------------------------
-
-%macro CONV_FLT_TO_FLTP_6CH 0
-%if ARCH_X86_64
-cglobal conv_flt_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
-%else
-cglobal conv_flt_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
-%define lend dword r2m
-%endif
- mov dst1q, [dstq+ gprsize]
- mov dst2q, [dstq+2*gprsize]
- mov dst3q, [dstq+3*gprsize]
- mov dst4q, [dstq+4*gprsize]
- mov dst5q, [dstq+5*gprsize]
- mov dstq, [dstq ]
- sub dst1q, dstq
- sub dst2q, dstq
- sub dst3q, dstq
- sub dst4q, dstq
- sub dst5q, dstq
-.loop:
- mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3
- mova m1, [srcq+1*mmsize] ; m1 = 4, 5, 6, 7
- mova m2, [srcq+2*mmsize] ; m2 = 8, 9, 10, 11
- mova m3, [srcq+3*mmsize] ; m3 = 12, 13, 14, 15
- mova m4, [srcq+4*mmsize] ; m4 = 16, 17, 18, 19
- mova m5, [srcq+5*mmsize] ; m5 = 20, 21, 22, 23
-
- SBUTTERFLY2 dq, 0, 3, 6 ; m0 = 0, 12, 1, 13
- ; m3 = 2, 14, 3, 15
- SBUTTERFLY2 dq, 1, 4, 6 ; m1 = 4, 16, 5, 17
- ; m4 = 6, 18, 7, 19
- SBUTTERFLY2 dq, 2, 5, 6 ; m2 = 8, 20, 9, 21
- ; m5 = 10, 22, 11, 23
- SBUTTERFLY2 dq, 0, 4, 6 ; m0 = 0, 6, 12, 18
- ; m4 = 1, 7, 13, 19
- SBUTTERFLY2 dq, 3, 2, 6 ; m3 = 2, 8, 14, 20
- ; m2 = 3, 9, 15, 21
- SBUTTERFLY2 dq, 1, 5, 6 ; m1 = 4, 10, 16, 22
- ; m5 = 5, 11, 17, 23
- mova [dstq ], m0
- mova [dstq+dst1q], m4
- mova [dstq+dst2q], m3
- mova [dstq+dst3q], m2
- mova [dstq+dst4q], m1
- mova [dstq+dst5q], m5
- add srcq, mmsize*6
- add dstq, mmsize
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-CONV_FLT_TO_FLTP_6CH
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-CONV_FLT_TO_FLTP_6CH
-%endif
diff --git a/libavresample/x86/audio_convert_init.c b/libavresample/x86/audio_convert_init.c
deleted file mode 100644
index 0af4222bea..0000000000
--- a/libavresample/x86/audio_convert_init.c
+++ /dev/null
@@ -1,265 +0,0 @@
-/*
- * Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "config.h"
-#include "libavutil/cpu.h"
-#include "libavutil/x86/cpu.h"
-#include "libavresample/audio_convert.h"
-
-/* flat conversions */
-
-void ff_conv_s16_to_s32_sse2(int16_t *dst, const int32_t *src, int len);
-
-void ff_conv_s16_to_flt_sse2(float *dst, const int16_t *src, int len);
-void ff_conv_s16_to_flt_sse4(float *dst, const int16_t *src, int len);
-
-void ff_conv_s32_to_s16_mmx (int16_t *dst, const int32_t *src, int len);
-void ff_conv_s32_to_s16_sse2(int16_t *dst, const int32_t *src, int len);
-
-void ff_conv_s32_to_flt_sse2(float *dst, const int32_t *src, int len);
-void ff_conv_s32_to_flt_avx (float *dst, const int32_t *src, int len);
-
-void ff_conv_flt_to_s16_sse2(int16_t *dst, const float *src, int len);
-
-void ff_conv_flt_to_s32_sse2(int32_t *dst, const float *src, int len);
-void ff_conv_flt_to_s32_avx (int32_t *dst, const float *src, int len);
-
-/* interleave conversions */
-
-void ff_conv_s16p_to_s16_2ch_sse2(int16_t *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_s16_2ch_avx (int16_t *dst, int16_t *const *src,
- int len, int channels);
-
-void ff_conv_s16p_to_s16_6ch_sse2(int16_t *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_s16_6ch_sse2slow(int16_t *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_s16_6ch_avx (int16_t *dst, int16_t *const *src,
- int len, int channels);
-
-void ff_conv_s16p_to_flt_2ch_sse2(float *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_flt_2ch_avx (float *dst, int16_t *const *src,
- int len, int channels);
-
-void ff_conv_s16p_to_flt_6ch_sse2 (float *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_flt_6ch_ssse3(float *dst, int16_t *const *src,
- int len, int channels);
-void ff_conv_s16p_to_flt_6ch_avx (float *dst, int16_t *const *src,
- int len, int channels);
-
-void ff_conv_fltp_to_s16_2ch_sse2 (int16_t *dst, float *const *src,
- int len, int channels);
-void ff_conv_fltp_to_s16_2ch_ssse3(int16_t *dst, float *const *src,
- int len, int channels);
-
-void ff_conv_fltp_to_s16_6ch_sse (int16_t *dst, float *const *src,
- int len, int channels);
-void ff_conv_fltp_to_s16_6ch_sse2(int16_t *dst, float *const *src,
- int len, int channels);
-void ff_conv_fltp_to_s16_6ch_avx (int16_t *dst, float *const *src,
- int len, int channels);
-
-void ff_conv_fltp_to_flt_2ch_sse(float *dst, float *const *src, int len,
- int channels);
-void ff_conv_fltp_to_flt_2ch_avx(float *dst, float *const *src, int len,
- int channels);
-
-void ff_conv_fltp_to_flt_6ch_mmx (float *dst, float *const *src, int len,
- int channels);
-void ff_conv_fltp_to_flt_6ch_sse4(float *dst, float *const *src, int len,
- int channels);
-void ff_conv_fltp_to_flt_6ch_avx (float *dst, float *const *src, int len,
- int channels);
-
-/* deinterleave conversions */
-
-void ff_conv_s16_to_s16p_2ch_sse2(int16_t *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_s16p_2ch_ssse3(int16_t *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_s16p_2ch_avx (int16_t *const *dst, int16_t *src,
- int len, int channels);
-
-void ff_conv_s16_to_s16p_6ch_sse2 (int16_t *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_s16p_6ch_ssse3(int16_t *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_s16p_6ch_avx (int16_t *const *dst, int16_t *src,
- int len, int channels);
-
-void ff_conv_s16_to_fltp_2ch_sse2(float *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_fltp_2ch_avx (float *const *dst, int16_t *src,
- int len, int channels);
-
-void ff_conv_s16_to_fltp_6ch_sse2 (float *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_fltp_6ch_ssse3(float *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_fltp_6ch_sse4 (float *const *dst, int16_t *src,
- int len, int channels);
-void ff_conv_s16_to_fltp_6ch_avx (float *const *dst, int16_t *src,
- int len, int channels);
-
-void ff_conv_flt_to_s16p_2ch_sse2(int16_t *const *dst, float *src,
- int len, int channels);
-void ff_conv_flt_to_s16p_2ch_avx (int16_t *const *dst, float *src,
- int len, int channels);
-
-void ff_conv_flt_to_s16p_6ch_sse2 (int16_t *const *dst, float *src,
- int len, int channels);
-void ff_conv_flt_to_s16p_6ch_ssse3(int16_t *const *dst, float *src,
- int len, int channels);
-void ff_conv_flt_to_s16p_6ch_avx (int16_t *const *dst, float *src,
- int len, int channels);
-
-void ff_conv_flt_to_fltp_2ch_sse(float *const *dst, float *src, int len,
- int channels);
-void ff_conv_flt_to_fltp_2ch_avx(float *const *dst, float *src, int len,
- int channels);
-
-void ff_conv_flt_to_fltp_6ch_sse2(float *const *dst, float *src, int len,
- int channels);
-void ff_conv_flt_to_fltp_6ch_avx (float *const *dst, float *src, int len,
- int channels);
-
-av_cold void ff_audio_convert_init_x86(AudioConvert *ac)
-{
- int cpu_flags = av_get_cpu_flags();
-
- if (EXTERNAL_MMX(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S32,
- 0, 1, 8, "MMX", ff_conv_s32_to_s16_mmx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_FLTP,
- 6, 1, 4, "MMX", ff_conv_fltp_to_flt_6ch_mmx);
- }
- if (EXTERNAL_SSE(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLTP,
- 6, 1, 2, "SSE", ff_conv_fltp_to_s16_6ch_sse);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_FLTP,
- 2, 16, 8, "SSE", ff_conv_fltp_to_flt_2ch_sse);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_FLT,
- 2, 16, 4, "SSE", ff_conv_flt_to_fltp_2ch_sse);
- }
- if (EXTERNAL_SSE2(cpu_flags)) {
- if (!(cpu_flags & AV_CPU_FLAG_SSE2SLOW)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S32,
- 0, 16, 16, "SSE2", ff_conv_s32_to_s16_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
- 6, 16, 8, "SSE2", ff_conv_s16p_to_s16_6ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLTP,
- 6, 16, 4, "SSE2", ff_conv_fltp_to_s16_6ch_sse2);
- } else {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
- 6, 1, 4, "SSE2SLOW", ff_conv_s16p_to_s16_6ch_sse2slow);
- }
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S32, AV_SAMPLE_FMT_S16,
- 0, 16, 8, "SSE2", ff_conv_s16_to_s32_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16,
- 0, 16, 8, "SSE2", ff_conv_s16_to_flt_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S32,
- 0, 16, 8, "SSE2", ff_conv_s32_to_flt_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLT,
- 0, 16, 16, "SSE2", ff_conv_flt_to_s16_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S32, AV_SAMPLE_FMT_FLT,
- 0, 16, 16, "SSE2", ff_conv_flt_to_s32_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
- 2, 16, 16, "SSE2", ff_conv_s16p_to_s16_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16P,
- 2, 16, 8, "SSE2", ff_conv_s16p_to_flt_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16P,
- 6, 16, 4, "SSE2", ff_conv_s16p_to_flt_6ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLTP,
- 2, 16, 4, "SSE2", ff_conv_fltp_to_s16_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 2, 16, 8, "SSE2", ff_conv_s16_to_s16p_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "SSE2", ff_conv_s16_to_s16p_6ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_S16,
- 2, 16, 8, "SSE2", ff_conv_s16_to_fltp_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "SSE2", ff_conv_s16_to_fltp_6ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_FLT,
- 2, 16, 8, "SSE2", ff_conv_flt_to_s16p_2ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_FLT,
- 6, 16, 4, "SSE2", ff_conv_flt_to_s16p_6ch_sse2);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_FLT,
- 6, 16, 4, "SSE2", ff_conv_flt_to_fltp_6ch_sse2);
- }
- if (EXTERNAL_SSSE3(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16P,
- 6, 16, 4, "SSSE3", ff_conv_s16p_to_flt_6ch_ssse3);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLTP,
- 2, 16, 4, "SSSE3", ff_conv_fltp_to_s16_2ch_ssse3);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 2, 16, 8, "SSSE3", ff_conv_s16_to_s16p_2ch_ssse3);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "SSSE3", ff_conv_s16_to_s16p_6ch_ssse3);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "SSSE3", ff_conv_s16_to_fltp_6ch_ssse3);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_FLT,
- 6, 16, 4, "SSSE3", ff_conv_flt_to_s16p_6ch_ssse3);
- }
- if (EXTERNAL_SSE4(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16,
- 0, 16, 8, "SSE4", ff_conv_s16_to_flt_sse4);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_FLTP,
- 6, 16, 4, "SSE4", ff_conv_fltp_to_flt_6ch_sse4);
- }
- if (EXTERNAL_AVX_FAST(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S32,
- 0, 32, 16, "AVX", ff_conv_s32_to_flt_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S32, AV_SAMPLE_FMT_FLT,
- 0, 32, 32, "AVX", ff_conv_flt_to_s32_avx);
- }
- if (EXTERNAL_AVX(cpu_flags)) {
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
- 2, 16, 16, "AVX", ff_conv_s16p_to_s16_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S16P,
- 6, 16, 8, "AVX", ff_conv_s16p_to_s16_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16P,
- 2, 16, 8, "AVX", ff_conv_s16p_to_flt_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_S16P,
- 6, 16, 4, "AVX", ff_conv_s16p_to_flt_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_FLTP,
- 6, 16, 4, "AVX", ff_conv_fltp_to_s16_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_FLTP,
- 6, 16, 4, "AVX", ff_conv_fltp_to_flt_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 2, 16, 8, "AVX", ff_conv_s16_to_s16p_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "AVX", ff_conv_s16_to_s16p_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_S16,
- 2, 16, 8, "AVX", ff_conv_s16_to_fltp_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_S16,
- 6, 16, 4, "AVX", ff_conv_s16_to_fltp_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_FLT,
- 2, 16, 8, "AVX", ff_conv_flt_to_s16p_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_S16P, AV_SAMPLE_FMT_FLT,
- 6, 16, 4, "AVX", ff_conv_flt_to_s16p_6ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_FLT,
- 2, 16, 4, "AVX", ff_conv_flt_to_fltp_2ch_avx);
- ff_audio_convert_set_func(ac, AV_SAMPLE_FMT_FLTP, AV_SAMPLE_FMT_FLT,
- 6, 16, 4, "AVX", ff_conv_flt_to_fltp_6ch_avx);
- }
-}
diff --git a/libavresample/x86/audio_mix.asm b/libavresample/x86/audio_mix.asm
deleted file mode 100644
index fe27d6a6c9..0000000000
--- a/libavresample/x86/audio_mix.asm
+++ /dev/null
@@ -1,511 +0,0 @@
-;******************************************************************************
-;* x86 optimized channel mixing
-;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
-;*
-;* This file is part of FFmpeg.
-;*
-;* FFmpeg is free software; you can redistribute it and/or
-;* modify it under the terms of the GNU Lesser General Public
-;* License as published by the Free Software Foundation; either
-;* version 2.1 of the License, or (at your option) any later version.
-;*
-;* FFmpeg is distributed in the hope that it will be useful,
-;* but WITHOUT ANY WARRANTY; without even the implied warranty of
-;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-;* Lesser General Public License for more details.
-;*
-;* You should have received a copy of the GNU Lesser General Public
-;* License along with FFmpeg; if not, write to the Free Software
-;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
-;******************************************************************************
-
-%include "libavutil/x86/x86util.asm"
-%include "util.asm"
-
-SECTION .text
-
-;-----------------------------------------------------------------------------
-; void ff_mix_2_to_1_fltp_flt(float **src, float **matrix, int len,
-; int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-%macro MIX_2_TO_1_FLTP_FLT 0
-cglobal mix_2_to_1_fltp_flt, 3,4,6, src, matrix, len, src1
- mov src1q, [srcq+gprsize]
- mov srcq, [srcq ]
- sub src1q, srcq
- mov matrixq, [matrixq ]
- VBROADCASTSS m4, [matrixq ]
- VBROADCASTSS m5, [matrixq+4]
- ALIGN 16
-.loop:
- mulps m0, m4, [srcq ]
- mulps m1, m5, [srcq+src1q ]
- mulps m2, m4, [srcq+ mmsize]
- mulps m3, m5, [srcq+src1q+mmsize]
- addps m0, m0, m1
- addps m2, m2, m3
- mova [srcq ], m0
- mova [srcq+mmsize], m2
- add srcq, mmsize*2
- sub lend, mmsize*2/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse
-MIX_2_TO_1_FLTP_FLT
-%if HAVE_AVX_EXTERNAL
-INIT_YMM avx
-MIX_2_TO_1_FLTP_FLT
-%endif
-
-;-----------------------------------------------------------------------------
-; void ff_mix_2_to_1_s16p_flt(int16_t **src, float **matrix, int len,
-; int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-%macro MIX_2_TO_1_S16P_FLT 0
-cglobal mix_2_to_1_s16p_flt, 3,4,6, src, matrix, len, src1
- mov src1q, [srcq+gprsize]
- mov srcq, [srcq]
- sub src1q, srcq
- mov matrixq, [matrixq ]
- VBROADCASTSS m4, [matrixq ]
- VBROADCASTSS m5, [matrixq+4]
- ALIGN 16
-.loop:
- mova m0, [srcq ]
- mova m2, [srcq+src1q]
- S16_TO_S32_SX 0, 1
- S16_TO_S32_SX 2, 3
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- cvtdq2ps m2, m2
- cvtdq2ps m3, m3
- mulps m0, m4
- mulps m1, m4
- mulps m2, m5
- mulps m3, m5
- addps m0, m2
- addps m1, m3
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- packssdw m0, m1
- mova [srcq], m0
- add srcq, mmsize
- sub lend, mmsize/2
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-MIX_2_TO_1_S16P_FLT
-INIT_XMM sse4
-MIX_2_TO_1_S16P_FLT
-
-;-----------------------------------------------------------------------------
-; void ff_mix_2_to_1_s16p_q8(int16_t **src, int16_t **matrix, int len,
-; int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-INIT_XMM sse2
-cglobal mix_2_to_1_s16p_q8, 3,4,6, src, matrix, len, src1
- mov src1q, [srcq+gprsize]
- mov srcq, [srcq]
- sub src1q, srcq
- mov matrixq, [matrixq]
- movd m4, [matrixq]
- movd m5, [matrixq]
- SPLATW m4, m4, 0
- SPLATW m5, m5, 1
- pxor m0, m0
- punpcklwd m4, m0
- punpcklwd m5, m0
- ALIGN 16
-.loop:
- mova m0, [srcq ]
- mova m2, [srcq+src1q]
- punpckhwd m1, m0, m0
- punpcklwd m0, m0
- punpckhwd m3, m2, m2
- punpcklwd m2, m2
- pmaddwd m0, m4
- pmaddwd m1, m4
- pmaddwd m2, m5
- pmaddwd m3, m5
- paddd m0, m2
- paddd m1, m3
- psrad m0, 8
- psrad m1, 8
- packssdw m0, m1
- mova [srcq], m0
- add srcq, mmsize
- sub lend, mmsize/2
- jg .loop
- REP_RET
-
-;-----------------------------------------------------------------------------
-; void ff_mix_1_to_2_fltp_flt(float **src, float **matrix, int len,
-; int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-%macro MIX_1_TO_2_FLTP_FLT 0
-cglobal mix_1_to_2_fltp_flt, 3,5,4, src0, matrix0, len, src1, matrix1
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q]
- sub src1q, src0q
- mov matrix1q, [matrix0q+gprsize]
- mov matrix0q, [matrix0q]
- VBROADCASTSS m2, [matrix0q]
- VBROADCASTSS m3, [matrix1q]
- ALIGN 16
-.loop:
- mova m0, [src0q]
- mulps m1, m0, m3
- mulps m0, m0, m2
- mova [src0q ], m0
- mova [src0q+src1q], m1
- add src0q, mmsize
- sub lend, mmsize/4
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse
-MIX_1_TO_2_FLTP_FLT
-%if HAVE_AVX_EXTERNAL
-INIT_YMM avx
-MIX_1_TO_2_FLTP_FLT
-%endif
-
-;-----------------------------------------------------------------------------
-; void ff_mix_1_to_2_s16p_flt(int16_t **src, float **matrix, int len,
-; int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-%macro MIX_1_TO_2_S16P_FLT 0
-cglobal mix_1_to_2_s16p_flt, 3,5,6, src0, matrix0, len, src1, matrix1
- mov src1q, [src0q+gprsize]
- mov src0q, [src0q]
- sub src1q, src0q
- mov matrix1q, [matrix0q+gprsize]
- mov matrix0q, [matrix0q]
- VBROADCASTSS m4, [matrix0q]
- VBROADCASTSS m5, [matrix1q]
- ALIGN 16
-.loop:
- mova m0, [src0q]
- S16_TO_S32_SX 0, 2
- cvtdq2ps m0, m0
- cvtdq2ps m2, m2
- mulps m1, m0, m5
- mulps m0, m0, m4
- mulps m3, m2, m5
- mulps m2, m2, m4
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- packssdw m0, m2
- packssdw m1, m3
- mova [src0q ], m0
- mova [src0q+src1q], m1
- add src0q, mmsize
- sub lend, mmsize/2
- jg .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-MIX_1_TO_2_S16P_FLT
-INIT_XMM sse4
-MIX_1_TO_2_S16P_FLT
-%if HAVE_AVX_EXTERNAL
-INIT_XMM avx
-MIX_1_TO_2_S16P_FLT
-%endif
-
-;-----------------------------------------------------------------------------
-; void ff_mix_3_8_to_1_2_fltp/s16p_flt(float/int16_t **src, float **matrix,
-; int len, int out_ch, int in_ch);
-;-----------------------------------------------------------------------------
-
-%macro MIX_3_8_TO_1_2_FLT 3 ; %1 = in channels, %2 = out channels, %3 = s16p or fltp
-; define some names to make the code clearer
-%assign in_channels %1
-%assign out_channels %2
-%assign stereo out_channels - 1
-%ifidn %3, s16p
- %assign is_s16 1
-%else
- %assign is_s16 0
-%endif
-
-; determine how many matrix elements must go on the stack vs. mmregs
-%assign matrix_elements in_channels * out_channels
-%if is_s16
- %if stereo
- %assign needed_mmregs 7
- %else
- %assign needed_mmregs 5
- %endif
-%else
- %if stereo
- %assign needed_mmregs 4
- %else
- %assign needed_mmregs 3
- %endif
-%endif
-%assign matrix_elements_mm num_mmregs - needed_mmregs
-%if matrix_elements < matrix_elements_mm
- %assign matrix_elements_mm matrix_elements
-%endif
-%if matrix_elements_mm < matrix_elements
- %assign matrix_elements_stack matrix_elements - matrix_elements_mm
-%else
- %assign matrix_elements_stack 0
-%endif
-%assign matrix_stack_size matrix_elements_stack * mmsize
-
-%assign needed_stack_size -1 * matrix_stack_size
-%if ARCH_X86_32 && in_channels >= 7
-%assign needed_stack_size needed_stack_size - 16
-%endif
-
-cglobal mix_%1_to_%2_%3_flt, 3,in_channels+2,needed_mmregs+matrix_elements_mm, needed_stack_size, src0, src1, len, src2, src3, src4, src5, src6, src7
-
-; define src pointers on stack if needed
-%if matrix_elements_stack > 0 && ARCH_X86_32 && in_channels >= 7
- %define src5m [rsp+matrix_stack_size+0]
- %define src6m [rsp+matrix_stack_size+4]
- %define src7m [rsp+matrix_stack_size+8]
-%endif
-
-; load matrix pointers
-%define matrix0q r1q
-%define matrix1q r3q
-%if stereo
- mov matrix1q, [matrix0q+gprsize]
-%endif
- mov matrix0q, [matrix0q]
-
-; define matrix coeff names
-%assign %%i 0
-%assign %%j needed_mmregs
-%rep in_channels
- %if %%i >= matrix_elements_mm
- CAT_XDEFINE mx_stack_0_, %%i, 1
- CAT_XDEFINE mx_0_, %%i, [rsp+(%%i-matrix_elements_mm)*mmsize]
- %else
- CAT_XDEFINE mx_stack_0_, %%i, 0
- CAT_XDEFINE mx_0_, %%i, m %+ %%j
- %assign %%j %%j+1
- %endif
- %assign %%i %%i+1
-%endrep
-%if stereo
-%assign %%i 0
-%rep in_channels
- %if in_channels + %%i >= matrix_elements_mm
- CAT_XDEFINE mx_stack_1_, %%i, 1
- CAT_XDEFINE mx_1_, %%i, [rsp+(in_channels+%%i-matrix_elements_mm)*mmsize]
- %else
- CAT_XDEFINE mx_stack_1_, %%i, 0
- CAT_XDEFINE mx_1_, %%i, m %+ %%j
- %assign %%j %%j+1
- %endif
- %assign %%i %%i+1
-%endrep
-%endif
-
-; load/splat matrix coeffs
-%assign %%i 0
-%rep in_channels
- %if mx_stack_0_ %+ %%i
- VBROADCASTSS m0, [matrix0q+4*%%i]
- mova mx_0_ %+ %%i, m0
- %else
- VBROADCASTSS mx_0_ %+ %%i, [matrix0q+4*%%i]
- %endif
- %if stereo
- %if mx_stack_1_ %+ %%i
- VBROADCASTSS m0, [matrix1q+4*%%i]
- mova mx_1_ %+ %%i, m0
- %else
- VBROADCASTSS mx_1_ %+ %%i, [matrix1q+4*%%i]
- %endif
- %endif
- %assign %%i %%i+1
-%endrep
-
-; load channel pointers to registers as offsets from the first channel pointer
-%if ARCH_X86_64
- movsxd lenq, r2d
-%endif
- shl lenq, 2-is_s16
-%assign %%i 1
-%rep (in_channels - 1)
- %if ARCH_X86_32 && in_channels >= 7 && %%i >= 5
- mov src5q, [src0q+%%i*gprsize]
- add src5q, lenq
- mov src %+ %%i %+ m, src5q
- %else
- mov src %+ %%i %+ q, [src0q+%%i*gprsize]
- add src %+ %%i %+ q, lenq
- %endif
- %assign %%i %%i+1
-%endrep
- mov src0q, [src0q]
- add src0q, lenq
- neg lenq
-.loop:
-; for x86-32 with 7-8 channels we do not have enough gp registers for all src
-; pointers, so we have to load some of them from the stack each time
-%define copy_src_from_stack ARCH_X86_32 && in_channels >= 7 && %%i >= 5
-%if is_s16
- ; mix with s16p input
- mova m0, [src0q+lenq]
- S16_TO_S32_SX 0, 1
- cvtdq2ps m0, m0
- cvtdq2ps m1, m1
- %if stereo
- mulps m2, m0, mx_1_0
- mulps m3, m1, mx_1_0
- %endif
- mulps m0, m0, mx_0_0
- mulps m1, m1, mx_0_0
-%assign %%i 1
-%rep (in_channels - 1)
- %if copy_src_from_stack
- %define src_ptr src5q
- %else
- %define src_ptr src %+ %%i %+ q
- %endif
- %if stereo
- %if copy_src_from_stack
- mov src_ptr, src %+ %%i %+ m
- %endif
- mova m4, [src_ptr+lenq]
- S16_TO_S32_SX 4, 5
- cvtdq2ps m4, m4
- cvtdq2ps m5, m5
- FMULADD_PS m2, m4, mx_1_ %+ %%i, m2, m6
- FMULADD_PS m3, m5, mx_1_ %+ %%i, m3, m6
- FMULADD_PS m0, m4, mx_0_ %+ %%i, m0, m4
- FMULADD_PS m1, m5, mx_0_ %+ %%i, m1, m5
- %else
- %if copy_src_from_stack
- mov src_ptr, src %+ %%i %+ m
- %endif
- mova m2, [src_ptr+lenq]
- S16_TO_S32_SX 2, 3
- cvtdq2ps m2, m2
- cvtdq2ps m3, m3
- FMULADD_PS m0, m2, mx_0_ %+ %%i, m0, m4
- FMULADD_PS m1, m3, mx_0_ %+ %%i, m1, m4
- %endif
- %assign %%i %%i+1
-%endrep
- %if stereo
- cvtps2dq m2, m2
- cvtps2dq m3, m3
- packssdw m2, m3
- mova [src1q+lenq], m2
- %endif
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- packssdw m0, m1
- mova [src0q+lenq], m0
-%else
- ; mix with fltp input
- %if stereo || mx_stack_0_0
- mova m0, [src0q+lenq]
- %endif
- %if stereo
- mulps m1, m0, mx_1_0
- %endif
- %if stereo || mx_stack_0_0
- mulps m0, m0, mx_0_0
- %else
- mulps m0, mx_0_0, [src0q+lenq]
- %endif
-%assign %%i 1
-%rep (in_channels - 1)
- %if copy_src_from_stack
- %define src_ptr src5q
- mov src_ptr, src %+ %%i %+ m
- %else
- %define src_ptr src %+ %%i %+ q
- %endif
- ; avoid extra load for mono if matrix is in a mm register
- %if stereo || mx_stack_0_ %+ %%i
- mova m2, [src_ptr+lenq]
- %endif
- %if stereo
- FMULADD_PS m1, m2, mx_1_ %+ %%i, m1, m3
- %endif
- %if stereo || mx_stack_0_ %+ %%i
- FMULADD_PS m0, m2, mx_0_ %+ %%i, m0, m2
- %else
- FMULADD_PS m0, mx_0_ %+ %%i, [src_ptr+lenq], m0, m1
- %endif
- %assign %%i %%i+1
-%endrep
- mova [src0q+lenq], m0
- %if stereo
- mova [src1q+lenq], m1
- %endif
-%endif
-
- add lenq, mmsize
- jl .loop
-; zero ymm high halves
-%if mmsize == 32
- vzeroupper
-%endif
- RET
-%endmacro
-
-%macro MIX_3_8_TO_1_2_FLT_FUNCS 0
-%assign %%i 3
-%rep 6
- INIT_XMM sse
- MIX_3_8_TO_1_2_FLT %%i, 1, fltp
- MIX_3_8_TO_1_2_FLT %%i, 2, fltp
- INIT_XMM sse2
- MIX_3_8_TO_1_2_FLT %%i, 1, s16p
- MIX_3_8_TO_1_2_FLT %%i, 2, s16p
- INIT_XMM sse4
- MIX_3_8_TO_1_2_FLT %%i, 1, s16p
- MIX_3_8_TO_1_2_FLT %%i, 2, s16p
- ; do not use ymm AVX or FMA4 in x86-32 for 6 or more channels due to stack alignment issues
- %if HAVE_AVX_EXTERNAL
- %if ARCH_X86_64 || %%i < 6
- INIT_YMM avx
- %else
- INIT_XMM avx
- %endif
- MIX_3_8_TO_1_2_FLT %%i, 1, fltp
- MIX_3_8_TO_1_2_FLT %%i, 2, fltp
- INIT_XMM avx
- MIX_3_8_TO_1_2_FLT %%i, 1, s16p
- MIX_3_8_TO_1_2_FLT %%i, 2, s16p
- %endif
- %if HAVE_FMA4_EXTERNAL
- %if ARCH_X86_64 || %%i < 6
- INIT_YMM fma4
- %else
- INIT_XMM fma4
- %endif
- MIX_3_8_TO_1_2_FLT %%i, 1, fltp
- MIX_3_8_TO_1_2_FLT %%i, 2, fltp
- INIT_XMM fma4
- MIX_3_8_TO_1_2_FLT %%i, 1, s16p
- MIX_3_8_TO_1_2_FLT %%i, 2, s16p
- %endif
- %assign %%i %%i+1
-%endrep
-%endmacro
-
-MIX_3_8_TO_1_2_FLT_FUNCS
diff --git a/libavresample/x86/audio_mix_init.c b/libavresample/x86/audio_mix_init.c
deleted file mode 100644
index 9b86be2847..0000000000
--- a/libavresample/x86/audio_mix_init.c
+++ /dev/null
@@ -1,215 +0,0 @@
-/*
- * Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "config.h"
-#include "libavutil/cpu.h"
-#include "libavutil/x86/cpu.h"
-#include "libavresample/audio_mix.h"
-
-void ff_mix_2_to_1_fltp_flt_sse(float **src, float **matrix, int len,
- int out_ch, int in_ch);
-void ff_mix_2_to_1_fltp_flt_avx(float **src, float **matrix, int len,
- int out_ch, int in_ch);
-
-void ff_mix_2_to_1_s16p_flt_sse2(int16_t **src, float **matrix, int len,
- int out_ch, int in_ch);
-void ff_mix_2_to_1_s16p_flt_sse4(int16_t **src, float **matrix, int len,
- int out_ch, int in_ch);
-
-void ff_mix_2_to_1_s16p_q8_sse2(int16_t **src, int16_t **matrix,
- int len, int out_ch, int in_ch);
-
-void ff_mix_1_to_2_fltp_flt_sse(float **src, float **matrix, int len,
- int out_ch, int in_ch);
-void ff_mix_1_to_2_fltp_flt_avx(float **src, float **matrix, int len,
- int out_ch, int in_ch);
-
-void ff_mix_1_to_2_s16p_flt_sse2(int16_t **src, float **matrix, int len,
- int out_ch, int in_ch);
-void ff_mix_1_to_2_s16p_flt_sse4(int16_t **src, float **matrix, int len,
- int out_ch, int in_ch);
-void ff_mix_1_to_2_s16p_flt_avx (int16_t **src, float **matrix, int len,
- int out_ch, int in_ch);
-
-#define DEFINE_MIX_3_8_TO_1_2(chan) \
-void ff_mix_ ## chan ## _to_1_fltp_flt_sse(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_fltp_flt_sse(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_s16p_flt_sse2(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_s16p_flt_sse2(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_s16p_flt_sse4(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_s16p_flt_sse4(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_fltp_flt_avx(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_fltp_flt_avx(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_s16p_flt_avx(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_s16p_flt_avx(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_fltp_flt_fma4(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_fltp_flt_fma4(float **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
- \
-void ff_mix_ ## chan ## _to_1_s16p_flt_fma4(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch); \
-void ff_mix_ ## chan ## _to_2_s16p_flt_fma4(int16_t **src, \
- float **matrix, int len, \
- int out_ch, int in_ch);
-
-DEFINE_MIX_3_8_TO_1_2(3)
-DEFINE_MIX_3_8_TO_1_2(4)
-DEFINE_MIX_3_8_TO_1_2(5)
-DEFINE_MIX_3_8_TO_1_2(6)
-DEFINE_MIX_3_8_TO_1_2(7)
-DEFINE_MIX_3_8_TO_1_2(8)
-
-#define SET_MIX_3_8_TO_1_2(chan) \
- if (EXTERNAL_SSE(cpu_flags)) { \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, 16, 4, "SSE", \
- ff_mix_ ## chan ## _to_1_fltp_flt_sse); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, 16, 4, "SSE", \
- ff_mix_## chan ##_to_2_fltp_flt_sse); \
- } \
- if (EXTERNAL_SSE2(cpu_flags)) { \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, 16, 8, "SSE2", \
- ff_mix_ ## chan ## _to_1_s16p_flt_sse2); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, 16, 8, "SSE2", \
- ff_mix_ ## chan ## _to_2_s16p_flt_sse2); \
- } \
- if (EXTERNAL_SSE4(cpu_flags)) { \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, 16, 8, "SSE4", \
- ff_mix_ ## chan ## _to_1_s16p_flt_sse4); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, 16, 8, "SSE4", \
- ff_mix_ ## chan ## _to_2_s16p_flt_sse4); \
- } \
- if (EXTERNAL_AVX(cpu_flags)) { \
- int ptr_align = 32; \
- int smp_align = 8; \
- if (ARCH_X86_32 || chan >= 6) { \
- ptr_align = 16; \
- smp_align = 4; \
- } \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, ptr_align, smp_align, "AVX", \
- ff_mix_ ## chan ## _to_1_fltp_flt_avx); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, ptr_align, smp_align, "AVX", \
- ff_mix_ ## chan ## _to_2_fltp_flt_avx); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, 16, 8, "AVX", \
- ff_mix_ ## chan ## _to_1_s16p_flt_avx); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, 16, 8, "AVX", \
- ff_mix_ ## chan ## _to_2_s16p_flt_avx); \
- } \
- if (EXTERNAL_FMA4(cpu_flags)) { \
- int ptr_align = 32; \
- int smp_align = 8; \
- if (ARCH_X86_32 || chan >= 6) { \
- ptr_align = 16; \
- smp_align = 4; \
- } \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, ptr_align, smp_align, "FMA4", \
- ff_mix_ ## chan ## _to_1_fltp_flt_fma4); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, ptr_align, smp_align, "FMA4", \
- ff_mix_ ## chan ## _to_2_fltp_flt_fma4); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 1, 16, 8, "FMA4", \
- ff_mix_ ## chan ## _to_1_s16p_flt_fma4); \
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,\
- chan, 2, 16, 8, "FMA4", \
- ff_mix_ ## chan ## _to_2_s16p_flt_fma4); \
- }
-
-av_cold void ff_audio_mix_init_x86(AudioMix *am)
-{
- int cpu_flags = av_get_cpu_flags();
-
- if (EXTERNAL_SSE(cpu_flags)) {
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,
- 2, 1, 16, 8, "SSE", ff_mix_2_to_1_fltp_flt_sse);
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,
- 1, 2, 16, 4, "SSE", ff_mix_1_to_2_fltp_flt_sse);
- }
- if (EXTERNAL_SSE2(cpu_flags)) {
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,
- 2, 1, 16, 8, "SSE2", ff_mix_2_to_1_s16p_flt_sse2);
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_Q8,
- 2, 1, 16, 8, "SSE2", ff_mix_2_to_1_s16p_q8_sse2);
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,
- 1, 2, 16, 8, "SSE2", ff_mix_1_to_2_s16p_flt_sse2);
- }
- if (EXTERNAL_SSE4(cpu_flags)) {
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,
- 2, 1, 16, 8, "SSE4", ff_mix_2_to_1_s16p_flt_sse4);
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,
- 1, 2, 16, 8, "SSE4", ff_mix_1_to_2_s16p_flt_sse4);
- }
- if (EXTERNAL_AVX_FAST(cpu_flags)) {
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,
- 2, 1, 32, 16, "AVX", ff_mix_2_to_1_fltp_flt_avx);
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_FLTP, AV_MIX_COEFF_TYPE_FLT,
- 1, 2, 32, 8, "AVX", ff_mix_1_to_2_fltp_flt_avx);
- }
- if (EXTERNAL_AVX(cpu_flags)) {
- ff_audio_mix_set_func(am, AV_SAMPLE_FMT_S16P, AV_MIX_COEFF_TYPE_FLT,
- 1, 2, 16, 8, "AVX", ff_mix_1_to_2_s16p_flt_avx);
- }
-
- SET_MIX_3_8_TO_1_2(3)
- SET_MIX_3_8_TO_1_2(4)
- SET_MIX_3_8_TO_1_2(5)
- SET_MIX_3_8_TO_1_2(6)
- SET_MIX_3_8_TO_1_2(7)
- SET_MIX_3_8_TO_1_2(8)
-}
diff --git a/libavresample/x86/dither.asm b/libavresample/x86/dither.asm
deleted file mode 100644
index d677c7179a..0000000000
--- a/libavresample/x86/dither.asm
+++ /dev/null
@@ -1,117 +0,0 @@
-;******************************************************************************
-;* x86 optimized dithering format conversion
-;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
-;*
-;* This file is part of FFmpeg.
-;*
-;* FFmpeg is free software; you can redistribute it and/or
-;* modify it under the terms of the GNU Lesser General Public
-;* License as published by the Free Software Foundation; either
-;* version 2.1 of the License, or (at your option) any later version.
-;*
-;* FFmpeg is distributed in the hope that it will be useful,
-;* but WITHOUT ANY WARRANTY; without even the implied warranty of
-;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-;* Lesser General Public License for more details.
-;*
-;* You should have received a copy of the GNU Lesser General Public
-;* License along with FFmpeg; if not, write to the Free Software
-;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
-;******************************************************************************
-
-%include "libavutil/x86/x86util.asm"
-
-SECTION_RODATA 32
-
-; 1.0f / (2.0f * INT32_MAX)
-pf_dither_scale: times 8 dd 2.32830643762e-10
-
-pf_s16_scale: times 4 dd 32753.0
-
-SECTION .text
-
-;------------------------------------------------------------------------------
-; void ff_quantize(int16_t *dst, float *src, float *dither, int len);
-;------------------------------------------------------------------------------
-
-INIT_XMM sse2
-cglobal quantize, 4,4,3, dst, src, dither, len
- lea lenq, [2*lend]
- add dstq, lenq
- lea srcq, [srcq+2*lenq]
- lea ditherq, [ditherq+2*lenq]
- neg lenq
- mova m2, [pf_s16_scale]
-.loop:
- mulps m0, m2, [srcq+2*lenq]
- mulps m1, m2, [srcq+2*lenq+mmsize]
- addps m0, [ditherq+2*lenq]
- addps m1, [ditherq+2*lenq+mmsize]
- cvtps2dq m0, m0
- cvtps2dq m1, m1
- packssdw m0, m1
- mova [dstq+lenq], m0
- add lenq, mmsize
- jl .loop
- REP_RET
-
-;------------------------------------------------------------------------------
-; void ff_dither_int_to_float_rectangular(float *dst, int *src, int len)
-;------------------------------------------------------------------------------
-
-%macro DITHER_INT_TO_FLOAT_RECTANGULAR 0
-cglobal dither_int_to_float_rectangular, 3,3,3, dst, src, len
- lea lenq, [4*lend]
- add srcq, lenq
- add dstq, lenq
- neg lenq
- mova m0, [pf_dither_scale]
-.loop:
- cvtdq2ps m1, [srcq+lenq]
- cvtdq2ps m2, [srcq+lenq+mmsize]
- mulps m1, m1, m0
- mulps m2, m2, m0
- mova [dstq+lenq], m1
- mova [dstq+lenq+mmsize], m2
- add lenq, 2*mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-DITHER_INT_TO_FLOAT_RECTANGULAR
-INIT_YMM avx
-DITHER_INT_TO_FLOAT_RECTANGULAR
-
-;------------------------------------------------------------------------------
-; void ff_dither_int_to_float_triangular(float *dst, int *src0, int len)
-;------------------------------------------------------------------------------
-
-%macro DITHER_INT_TO_FLOAT_TRIANGULAR 0
-cglobal dither_int_to_float_triangular, 3,4,5, dst, src0, len, src1
- lea lenq, [4*lend]
- lea src1q, [src0q+2*lenq]
- add src0q, lenq
- add dstq, lenq
- neg lenq
- mova m0, [pf_dither_scale]
-.loop:
- cvtdq2ps m1, [src0q+lenq]
- cvtdq2ps m2, [src0q+lenq+mmsize]
- cvtdq2ps m3, [src1q+lenq]
- cvtdq2ps m4, [src1q+lenq+mmsize]
- addps m1, m1, m3
- addps m2, m2, m4
- mulps m1, m1, m0
- mulps m2, m2, m0
- mova [dstq+lenq], m1
- mova [dstq+lenq+mmsize], m2
- add lenq, 2*mmsize
- jl .loop
- REP_RET
-%endmacro
-
-INIT_XMM sse2
-DITHER_INT_TO_FLOAT_TRIANGULAR
-INIT_YMM avx
-DITHER_INT_TO_FLOAT_TRIANGULAR
diff --git a/libavresample/x86/dither_init.c b/libavresample/x86/dither_init.c
deleted file mode 100644
index ad157b96b1..0000000000
--- a/libavresample/x86/dither_init.c
+++ /dev/null
@@ -1,60 +0,0 @@
-/*
- * Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "config.h"
-#include "libavutil/cpu.h"
-#include "libavutil/x86/cpu.h"
-#include "libavresample/dither.h"
-
-void ff_quantize_sse2(int16_t *dst, const float *src, float *dither, int len);
-
-void ff_dither_int_to_float_rectangular_sse2(float *dst, int *src, int len);
-void ff_dither_int_to_float_rectangular_avx(float *dst, int *src, int len);
-
-void ff_dither_int_to_float_triangular_sse2(float *dst, int *src0, int len);
-void ff_dither_int_to_float_triangular_avx(float *dst, int *src0, int len);
-
-av_cold void ff_dither_init_x86(DitherDSPContext *ddsp,
- enum AVResampleDitherMethod method)
-{
- int cpu_flags = av_get_cpu_flags();
-
- if (EXTERNAL_SSE2(cpu_flags)) {
- ddsp->quantize = ff_quantize_sse2;
- ddsp->ptr_align = 16;
- ddsp->samples_align = 8;
- }
-
- if (method == AV_RESAMPLE_DITHER_RECTANGULAR) {
- if (EXTERNAL_SSE2(cpu_flags)) {
- ddsp->dither_int_to_float = ff_dither_int_to_float_rectangular_sse2;
- }
- if (EXTERNAL_AVX_FAST(cpu_flags)) {
- ddsp->dither_int_to_float = ff_dither_int_to_float_rectangular_avx;
- }
- } else {
- if (EXTERNAL_SSE2(cpu_flags)) {
- ddsp->dither_int_to_float = ff_dither_int_to_float_triangular_sse2;
- }
- if (EXTERNAL_AVX_FAST(cpu_flags)) {
- ddsp->dither_int_to_float = ff_dither_int_to_float_triangular_avx;
- }
- }
-}
diff --git a/libavresample/x86/util.asm b/libavresample/x86/util.asm
deleted file mode 100644
index 187a4a21ba..0000000000
--- a/libavresample/x86/util.asm
+++ /dev/null
@@ -1,41 +0,0 @@
-;******************************************************************************
-;* x86 utility macros for libavresample
-;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
-;*
-;* This file is part of FFmpeg.
-;*
-;* FFmpeg is free software; you can redistribute it and/or
-;* modify it under the terms of the GNU Lesser General Public
-;* License as published by the Free Software Foundation; either
-;* version 2.1 of the License, or (at your option) any later version.
-;*
-;* FFmpeg is distributed in the hope that it will be useful,
-;* but WITHOUT ANY WARRANTY; without even the implied warranty of
-;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-;* Lesser General Public License for more details.
-;*
-;* You should have received a copy of the GNU Lesser General Public
-;* License along with FFmpeg; if not, write to the Free Software
-;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
-;******************************************************************************
-
-%macro S16_TO_S32_SX 2 ; src/low dst, high dst
-%if cpuflag(sse4)
- pmovsxwd m%2, m%1
- psrldq m%1, 8
- pmovsxwd m%1, m%1
- SWAP %1, %2
-%else
- mova m%2, m%1
- punpckhwd m%2, m%2
- punpcklwd m%1, m%1
- psrad m%2, 16
- psrad m%1, 16
-%endif
-%endmacro
-
-%macro DEINT2_PS 3 ; src0/even dst, src1/odd dst, temp
- shufps m%3, m%1, m%2, q3131
- shufps m%1, m%2, q2020
- SWAP %2,%3
-%endmacro
diff --git a/libavresample/x86/w64xmmtest.c b/libavresample/x86/w64xmmtest.c
deleted file mode 100644
index 0f42bd185c..0000000000
--- a/libavresample/x86/w64xmmtest.c
+++ /dev/null
@@ -1,31 +0,0 @@
-/*
- * check XMM registers for clobbers on Win64
- * Copyright (c) 2013 Martin Storsjo
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "libavresample/avresample.h"
-#include "libavutil/x86/w64xmmtest.h"
-
-wrap(avresample_convert(AVAudioResampleContext *avr, uint8_t **output,
- int out_plane_size, int out_samples, uint8_t **input,
- int in_plane_size, int in_samples))
-{
- testxmmclobbers(avresample_convert, avr, output, out_plane_size,
- out_samples, input, in_plane_size, in_samples);
-}