;****************************************************************************** ;* x86 optimized Format Conversion Utils ;* Copyright (c) 2008 Loren Merritt ;* Copyright (c) 2012 Justin Ruggles ;* ;* This file is part of Libav. ;* ;* Libav is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* Libav is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with Libav; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "x86inc.asm" %include "x86util.asm" %include "util.asm" SECTION_RODATA 32 pf_s32_inv_scale: times 8 dd 0x30000000 pf_s32_scale: times 8 dd 0x4f000000 pf_s16_inv_scale: times 4 dd 0x38000000 pf_s16_scale: times 4 dd 0x47000000 SECTION_TEXT ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len); ;------------------------------------------------------------------------------ INIT_XMM sse2 cglobal conv_s16_to_s32, 3,3,3, dst, src, len lea lenq, [2*lend] lea dstq, [dstq+2*lenq] add srcq, lenq neg lenq .loop: mova m2, [srcq+lenq] pxor m0, m0 pxor m1, m1 punpcklwd m0, m2 punpckhwd m1, m2 mova [dstq+2*lenq ], m0 mova [dstq+2*lenq+mmsize], m1 add lenq, mmsize jl .loop REP_RET ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_FLT 0 cglobal conv_s16_to_flt, 3,3,3, dst, src, len lea lenq, [2*lend] add srcq, lenq lea dstq, [dstq + 2*lenq] neg lenq mova m2, [pf_s16_inv_scale] ALIGN 16 .loop: mova m0, [srcq+lenq] S16_TO_S32_SX 0, 1 cvtdq2ps m0, m0 cvtdq2ps m1, m1 mulps m0, m2 mulps m1, m2 mova [dstq+2*lenq ], m0 mova [dstq+2*lenq+mmsize], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_FLT INIT_XMM sse4 CONV_S16_TO_FLT ;------------------------------------------------------------------------------ ; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S32_TO_S16 0 cglobal conv_s32_to_s16, 3,3,4, dst, src, len lea lenq, [2*lend] lea srcq, [srcq+2*lenq] add dstq, lenq neg lenq .loop: mova m0, [srcq+2*lenq ] mova m1, [srcq+2*lenq+ mmsize] mova m2, [srcq+2*lenq+2*mmsize] mova m3, [srcq+2*lenq+3*mmsize] psrad m0, 16 psrad m1, 16 psrad m2, 16 psrad m3, 16 packssdw m0, m1 packssdw m2, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop %if mmsize == 8 emms RET %else REP_RET %endif %endmacro INIT_MMX mmx CONV_S32_TO_S16 INIT_XMM sse2 CONV_S32_TO_S16 ;------------------------------------------------------------------------------ ; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S32_TO_FLT 0 cglobal conv_s32_to_flt, 3,3,3, dst, src, len lea lenq, [4*lend] add srcq, lenq add dstq, lenq neg lenq mova m0, [pf_s32_inv_scale] ALIGN 16 .loop: cvtdq2ps m1, [srcq+lenq ] cvtdq2ps m2, [srcq+lenq+mmsize] mulps m1, m1, m0 mulps m2, m2, m0 mova [dstq+lenq ], m1 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S32_TO_FLT %if HAVE_AVX INIT_YMM avx CONV_S32_TO_FLT %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len); ;------------------------------------------------------------------------------ INIT_XMM sse2 cglobal conv_flt_to_s16, 3,3,5, dst, src, len lea lenq, [2*lend] lea srcq, [srcq+2*lenq] add dstq, lenq neg lenq mova m4, [pf_s16_scale] .loop: mova m0, [srcq+2*lenq ] mova m1, [srcq+2*lenq+1*mmsize] mova m2, [srcq+2*lenq+2*mmsize] mova m3, [srcq+2*lenq+3*mmsize] mulps m0, m4 mulps m1, m4 mulps m2, m4 mulps m3, m4 cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 packssdw m0, m1 packssdw m2, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop REP_RET ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_S32 0 cglobal conv_flt_to_s32, 3,3,5, dst, src, len lea lenq, [lend*4] add srcq, lenq add dstq, lenq neg lenq mova m4, [pf_s32_scale] .loop: mulps m0, m4, [srcq+lenq ] mulps m1, m4, [srcq+lenq+1*mmsize] mulps m2, m4, [srcq+lenq+2*mmsize] mulps m3, m4, [srcq+lenq+3*mmsize] cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+1*mmsize], m1 mova [dstq+lenq+2*mmsize], m2 mova [dstq+lenq+3*mmsize], m3 add lenq, mmsize*4 jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLT_TO_S32 %if HAVE_AVX INIT_YMM avx CONV_FLT_TO_S32 %endif ;----------------------------------------------------------------------------- ; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len, ; int channels); ;----------------------------------------------------------------------------- %macro CONV_FLTP_TO_FLT_6CH 0 cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len %if ARCH_X86_64 mov lend, r2d %else %define lend dword r2m %endif mov src1q, [srcq+1*gprsize] mov src2q, [srcq+2*gprsize] mov src3q, [srcq+3*gprsize] mov src4q, [srcq+4*gprsize] mov src5q, [srcq+5*gprsize] mov srcq, [srcq] sub src1q, srcq sub src2q, srcq sub src3q, srcq sub src4q, srcq sub src5q, srcq .loop: mova m0, [srcq ] mova m1, [srcq+src1q] mova m2, [srcq+src2q] mova m3, [srcq+src3q] mova m4, [srcq+src4q] mova m5, [srcq+src5q] %if cpuflag(sse4) SBUTTERFLYPS 0, 1, 6 SBUTTERFLYPS 2, 3, 6 SBUTTERFLYPS 4, 5, 6 blendps m6, m4, m0, 1100b movlhps m0, m2 movhlps m4, m2 blendps m2, m5, m1, 1100b movlhps m1, m3 movhlps m5, m3 movaps [dstq ], m0 movaps [dstq+16], m6 movaps [dstq+32], m4 movaps [dstq+48], m1 movaps [dstq+64], m2 movaps [dstq+80], m5 %else ; mmx SBUTTERFLY dq, 0, 1, 6 SBUTTERFLY dq, 2, 3, 6 SBUTTERFLY dq, 4, 5, 6 movq [dstq ], m0 movq [dstq+ 8], m2 movq [dstq+16], m4 movq [dstq+24], m1 movq [dstq+32], m3 movq [dstq+40], m5 %endif add srcq, mmsize add dstq, mmsize*6 sub lend, mmsize/4 jg .loop %if mmsize == 8 emms RET %else REP_RET %endif %endmacro INIT_MMX mmx CONV_FLTP_TO_FLT_6CH INIT_XMM sse4 CONV_FLTP_TO_FLT_6CH %if HAVE_AVX INIT_XMM avx CONV_FLTP_TO_FLT_6CH %endif