;***************************************************************************** ;* x86-optimized AC-3 DSP utils ;* Copyright (c) 2011 Justin Ruggles ;* ;* This file is part of Libav. ;* ;* Libav is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* Libav is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with Libav; if not, write to the Free Software ;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "x86inc.asm" %include "x86util.asm" SECTION_RODATA ; 16777216.0f - used in ff_float_to_fixed24() pf_1_24: times 4 dd 0x4B800000 SECTION .text ;----------------------------------------------------------------------------- ; void ff_ac3_exponent_min(uint8_t *exp, int num_reuse_blocks, int nb_coefs) ;----------------------------------------------------------------------------- %macro AC3_EXPONENT_MIN 1 cglobal ac3_exponent_min_%1, 3,4,2, exp, reuse_blks, expn, offset shl reuse_blksq, 8 jz .end LOOP_ALIGN .nextexp: mov offsetq, reuse_blksq mova m0, [expq+offsetq] sub offsetq, 256 LOOP_ALIGN .nextblk: PMINUB m0, [expq+offsetq], m1 sub offsetq, 256 jae .nextblk mova [expq], m0 add expq, mmsize sub expnq, mmsize jg .nextexp .end: REP_RET %endmacro %define PMINUB PMINUB_MMX %define LOOP_ALIGN INIT_MMX AC3_EXPONENT_MIN mmx %ifdef HAVE_MMX2 %define PMINUB PMINUB_MMXEXT %define LOOP_ALIGN ALIGN 16 AC3_EXPONENT_MIN mmxext %endif %ifdef HAVE_SSE INIT_XMM AC3_EXPONENT_MIN sse2 %endif %undef PMINUB %undef LOOP_ALIGN ;----------------------------------------------------------------------------- ; int ff_ac3_max_msb_abs_int16(const int16_t *src, int len) ; ; This function uses 2 different methods to calculate a valid result. ; 1) logical 'or' of abs of each element ; This is used for ssse3 because of the pabsw instruction. ; It is also used for mmx because of the lack of min/max instructions. ; 2) calculate min/max for the array, then or(abs(min),abs(max)) ; This is used for mmxext and sse2 because they have pminsw/pmaxsw. ;----------------------------------------------------------------------------- %macro AC3_MAX_MSB_ABS_INT16 2 cglobal ac3_max_msb_abs_int16_%1, 2,2,5, src, len pxor m2, m2 pxor m3, m3 .loop: %ifidn %2, min_max mova m0, [srcq] mova m1, [srcq+mmsize] pminsw m2, m0 pminsw m2, m1 pmaxsw m3, m0 pmaxsw m3, m1 %else ; or_abs %ifidn %1, mmx mova m0, [srcq] mova m1, [srcq+mmsize] ABS2 m0, m1, m3, m4 %else ; ssse3 ; using memory args is faster for ssse3 pabsw m0, [srcq] pabsw m1, [srcq+mmsize] %endif por m2, m0 por m2, m1 %endif add srcq, mmsize*2 sub lend, mmsize ja .loop %ifidn %2, min_max ABS2 m2, m3, m0, m1 por m2, m3 %endif %ifidn mmsize, 16 movhlps m0, m2 por m2, m0 %endif PSHUFLW m0, m2, 0xe por m2, m0 PSHUFLW m0, m2, 0x1 por m2, m0 movd eax, m2 and eax, 0xFFFF RET %endmacro INIT_MMX %define ABS2 ABS2_MMX %define PSHUFLW pshufw AC3_MAX_MSB_ABS_INT16 mmx, or_abs %define ABS2 ABS2_MMX2 AC3_MAX_MSB_ABS_INT16 mmxext, min_max INIT_XMM %define PSHUFLW pshuflw AC3_MAX_MSB_ABS_INT16 sse2, min_max %define ABS2 ABS2_SSSE3 AC3_MAX_MSB_ABS_INT16 ssse3, or_abs ;----------------------------------------------------------------------------- ; macro used for ff_ac3_lshift_int16() and ff_ac3_rshift_int32() ;----------------------------------------------------------------------------- %macro AC3_SHIFT 4 ; l/r, 16/32, shift instruction, instruction set cglobal ac3_%1shift_int%2_%4, 3,3,5, src, len, shift movd m0, shiftd .loop: mova m1, [srcq ] mova m2, [srcq+mmsize ] mova m3, [srcq+mmsize*2] mova m4, [srcq+mmsize*3] %3 m1, m0 %3 m2, m0 %3 m3, m0 %3 m4, m0 mova [srcq ], m1 mova [srcq+mmsize ], m2 mova [srcq+mmsize*2], m3 mova [srcq+mmsize*3], m4 add srcq, mmsize*4 sub lend, mmsize*32/%2 ja .loop .end: REP_RET %endmacro ;----------------------------------------------------------------------------- ; void ff_ac3_lshift_int16(int16_t *src, unsigned int len, unsigned int shift) ;----------------------------------------------------------------------------- INIT_MMX AC3_SHIFT l, 16, psllw, mmx INIT_XMM AC3_SHIFT l, 16, psllw, sse2 ;----------------------------------------------------------------------------- ; void ff_ac3_rshift_int32(int32_t *src, unsigned int len, unsigned int shift) ;----------------------------------------------------------------------------- INIT_MMX AC3_SHIFT r, 32, psrad, mmx INIT_XMM AC3_SHIFT r, 32, psrad, sse2 ;----------------------------------------------------------------------------- ; void ff_float_to_fixed24(int32_t *dst, const float *src, unsigned int len) ;----------------------------------------------------------------------------- ; The 3DNow! version is not bit-identical because pf2id uses truncation rather ; than round-to-nearest. INIT_MMX cglobal float_to_fixed24_3dnow, 3,3,0, dst, src, len movq m0, [pf_1_24] .loop: movq m1, [srcq ] movq m2, [srcq+8 ] movq m3, [srcq+16] movq m4, [srcq+24] pfmul m1, m0 pfmul m2, m0 pfmul m3, m0 pfmul m4, m0 pf2id m1, m1 pf2id m2, m2 pf2id m3, m3 pf2id m4, m4 movq [dstq ], m1 movq [dstq+8 ], m2 movq [dstq+16], m3 movq [dstq+24], m4 add srcq, 32 add dstq, 32 sub lend, 8 ja .loop REP_RET INIT_XMM cglobal float_to_fixed24_sse, 3,3,3, dst, src, len movaps m0, [pf_1_24] .loop: movaps m1, [srcq ] movaps m2, [srcq+16] mulps m1, m0 mulps m2, m0 cvtps2pi mm0, m1 movhlps m1, m1 cvtps2pi mm1, m1 cvtps2pi mm2, m2 movhlps m2, m2 cvtps2pi mm3, m2 movq [dstq ], mm0 movq [dstq+ 8], mm1 movq [dstq+16], mm2 movq [dstq+24], mm3 add srcq, 32 add dstq, 32 sub lend, 8 ja .loop REP_RET INIT_XMM cglobal float_to_fixed24_sse2, 3,3,9, dst, src, len movaps m0, [pf_1_24] .loop: movaps m1, [srcq ] movaps m2, [srcq+16 ] movaps m3, [srcq+32 ] movaps m4, [srcq+48 ] %ifdef m8 movaps m5, [srcq+64 ] movaps m6, [srcq+80 ] movaps m7, [srcq+96 ] movaps m8, [srcq+112] %endif mulps m1, m0 mulps m2, m0 mulps m3, m0 mulps m4, m0 %ifdef m8 mulps m5, m0 mulps m6, m0 mulps m7, m0 mulps m8, m0 %endif cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 cvtps2dq m4, m4 %ifdef m8 cvtps2dq m5, m5 cvtps2dq m6, m6 cvtps2dq m7, m7 cvtps2dq m8, m8 %endif movdqa [dstq ], m1 movdqa [dstq+16 ], m2 movdqa [dstq+32 ], m3 movdqa [dstq+48 ], m4 %ifdef m8 movdqa [dstq+64 ], m5 movdqa [dstq+80 ], m6 movdqa [dstq+96 ], m7 movdqa [dstq+112], m8 add srcq, 128 add dstq, 128 sub lenq, 32 %else add srcq, 64 add dstq, 64 sub lenq, 16 %endif ja .loop REP_RET