summaryrefslogtreecommitdiff
path: root/libavcodec/arm
diff options
context:
space:
mode:
authorMåns Rullgård <mans@mansr.com>2009-01-30 23:13:19 +0000
committerMåns Rullgård <mans@mansr.com>2009-01-30 23:13:19 +0000
commite1f7cb7fa0aec465ec0c172d1b922aebced04cc2 (patch)
tree0d7b397a30cff9049e5635f23a6a8b1aa148b899 /libavcodec/arm
parentdd927e2e62f499d3ed611042f91766bec20649ce (diff)
ARM: NEON optimised vector_fmul_window
Originally committed as revision 16868 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diffstat (limited to 'libavcodec/arm')
-rw-r--r--libavcodec/arm/dsputil_neon.c4
-rw-r--r--libavcodec/arm/dsputil_neon_s.S50
2 files changed, 54 insertions, 0 deletions
diff --git a/libavcodec/arm/dsputil_neon.c b/libavcodec/arm/dsputil_neon.c
index 2b1b2d32c3..e18a487fb7 100644
--- a/libavcodec/arm/dsputil_neon.c
+++ b/libavcodec/arm/dsputil_neon.c
@@ -147,6 +147,9 @@ void ff_h264_idct_add8_neon(uint8_t **dest, const int *block_offset,
const uint8_t nnzc[6*8]);
void ff_vector_fmul_neon(float *dst, const float *src, int len);
+void ff_vector_fmul_window_neon(float *dst, const float *src0,
+ const float *src1, const float *win,
+ float add_bias, int len);
void ff_float_to_int16_neon(int16_t *, const float *, long);
void ff_float_to_int16_interleave_neon(int16_t *, const float **, long, int);
@@ -245,6 +248,7 @@ void ff_dsputil_init_neon(DSPContext *c, AVCodecContext *avctx)
c->h264_idct_add8 = ff_h264_idct_add8_neon;
c->vector_fmul = ff_vector_fmul_neon;
+ c->vector_fmul_window = ff_vector_fmul_window_neon;
if (!(avctx->flags & CODEC_FLAG_BITEXACT)) {
c->float_to_int16 = ff_float_to_int16_neon;
diff --git a/libavcodec/arm/dsputil_neon_s.S b/libavcodec/arm/dsputil_neon_s.S
index 9027353c4a..3b39d2e75e 100644
--- a/libavcodec/arm/dsputil_neon_s.S
+++ b/libavcodec/arm/dsputil_neon_s.S
@@ -649,3 +649,53 @@ function ff_vector_fmul_neon, export=1
3: vst1.64 {d16-d19},[r3,:128]!
bx lr
.endfunc
+
+function ff_vector_fmul_window_neon, export=1
+ vld1.32 {d16[],d17[]}, [sp,:32]
+ push {r4,r5,lr}
+ ldr lr, [sp, #16]
+ sub r2, r2, #8
+ sub r5, lr, #2
+ add r2, r2, r5, lsl #2
+ add r4, r3, r5, lsl #3
+ add ip, r0, r5, lsl #3
+ mov r5, #-16
+ vld1.64 {d0,d1}, [r1,:128]!
+ vld1.64 {d2,d3}, [r2,:128], r5
+ vld1.64 {d4,d5}, [r3,:128]!
+ vld1.64 {d6,d7}, [r4,:128], r5
+1: subs lr, lr, #4
+ vmov q11, q8
+ vmla.f32 d22, d0, d4
+ vmov q10, q8
+ vmla.f32 d23, d1, d5
+ vrev64.32 q3, q3
+ vmla.f32 d20, d0, d7
+ vrev64.32 q1, q1
+ vmla.f32 d21, d1, d6
+ beq 2f
+ vmla.f32 d22, d3, d7
+ vld1.64 {d0,d1}, [r1,:128]!
+ vmla.f32 d23, d2, d6
+ vld1.64 {d18,d19},[r2,:128], r5
+ vmls.f32 d20, d3, d4
+ vld1.64 {d24,d25},[r3,:128]!
+ vmls.f32 d21, d2, d5
+ vld1.64 {d6,d7}, [r4,:128], r5
+ vmov q1, q9
+ vrev64.32 q11, q11
+ vmov q2, q12
+ vswp d22, d23
+ vst1.64 {d20,d21},[r0,:128]!
+ vst1.64 {d22,d23},[ip,:128], r5
+ b 1b
+2: vmla.f32 d22, d3, d7
+ vmla.f32 d23, d2, d6
+ vmls.f32 d20, d3, d4
+ vmls.f32 d21, d2, d5
+ vrev64.32 q11, q11
+ vswp d22, d23
+ vst1.64 {d20,d21},[r0,:128]!
+ vst1.64 {d22,d23},[ip,:128], r5
+ pop {r4,r5,pc}
+ .endfunc