summaryrefslogtreecommitdiff
path: root/libavcodec/arm
diff options
context:
space:
mode:
authorMichael Niedermayer <michaelni@gmx.at>2012-04-26 22:24:58 +0200
committerMichael Niedermayer <michaelni@gmx.at>2012-04-26 22:24:58 +0200
commite890b689628634d75e093b24ae257b008fc85443 (patch)
treeaf2189297e3da55438dc0e968633a8a2806fb5fd /libavcodec/arm
parentf1f5b39866c4deb7800aa9288010cdbd386da477 (diff)
parent0a07f2b346433a9a2677c69c6b29a1a827e39109 (diff)
Merge remote-tracking branch 'qatar/master'
* qatar/master: dsputil: fix invalid array indexing configure: add libavresample to rpath build: icc: silence some warnings fft-test: add option to set cpuflag mask cpu: recognise only cpu flag names pertinent to the architecture avutil: add av_parse_cpu_flags() function vp8: armv6: fix non-armv6t2 build vp8: armv6 optimisations vp8: arm: separate ARMv6 functions from NEON ARM: add some compatibility macros mov: support eac3 audio avf: fix faulty check in has_duration Conflicts: configure doc/APIchanges ffmpeg.c libavcodec/arm/Makefile libavcodec/arm/asm.S libavcodec/arm/vp8dsp_armv6.S libavcodec/arm/vp8dsp_init_arm.c libavutil/avutil.h Merged-by: Michael Niedermayer <michaelni@gmx.at>
Diffstat (limited to 'libavcodec/arm')
-rw-r--r--libavcodec/arm/Makefile4
-rw-r--r--libavcodec/arm/asm.S9
-rw-r--r--libavcodec/arm/vp8dsp.h78
-rw-r--r--libavcodec/arm/vp8dsp_armv6.S2808
-rw-r--r--libavcodec/arm/vp8dsp_init_arm.c290
-rw-r--r--libavcodec/arm/vp8dsp_init_armv6.c118
-rw-r--r--libavcodec/arm/vp8dsp_init_neon.c114
7 files changed, 1380 insertions, 2041 deletions
diff --git a/libavcodec/arm/Makefile b/libavcodec/arm/Makefile
index 8b96232468..3d002ccdcf 100644
--- a/libavcodec/arm/Makefile
+++ b/libavcodec/arm/Makefile
@@ -14,6 +14,7 @@ OBJS-$(CONFIG_VP5_DECODER) += arm/vp56dsp_init_arm.o
OBJS-$(CONFIG_VP6_DECODER) += arm/vp56dsp_init_arm.o
OBJS-$(CONFIG_VP8_DECODER) += arm/vp8dsp_init_arm.o
ARMV6-OBJS-$(CONFIG_VP8_DECODER) += arm/vp8_armv6.o \
+ arm/vp8dsp_init_armv6.o \
arm/vp8dsp_armv6.o
OBJS-$(CONFIG_H264DSP) += arm/h264dsp_init_arm.o
@@ -80,7 +81,8 @@ NEON-OBJS-$(CONFIG_VP5_DECODER) += arm/vp56dsp_neon.o \
NEON-OBJS-$(CONFIG_VP6_DECODER) += arm/vp56dsp_neon.o \
arm/vp3dsp_neon.o \
-NEON-OBJS-$(CONFIG_VP8_DECODER) += arm/vp8dsp_neon.o
+NEON-OBJS-$(CONFIG_VP8_DECODER) += arm/vp8dsp_init_neon.o \
+ arm/vp8dsp_neon.o
NEON-OBJS += arm/dsputil_init_neon.o \
arm/dsputil_neon.o \
diff --git a/libavcodec/arm/asm.S b/libavcodec/arm/asm.S
index e540eac9fa..28d52cfd53 100644
--- a/libavcodec/arm/asm.S
+++ b/libavcodec/arm/asm.S
@@ -88,6 +88,13 @@ ELF .size \name, . - \name
\name:
.endm
+#if !HAVE_ARMV6T2
+.macro movw rd, val
+ mov \rd, \val & 255
+ orr \rd, \val & ~255
+.endm
+#endif
+
.macro mov32 rd, val
#if HAVE_ARMV6T2
movw \rd, #(\val) & 0xffff
@@ -120,7 +127,7 @@ T sub \rn, \rn, \rm
T ldr \rt, [\rn]
.endm
-.macro ldr_dpren rt, rn, rm:vararg
+.macro ldr_nreg rt, rn, rm:vararg
A ldr \rt, [\rn, -\rm]
T sub \rt, \rn, \rm
T ldr \rt, [\rt]
diff --git a/libavcodec/arm/vp8dsp.h b/libavcodec/arm/vp8dsp.h
new file mode 100644
index 0000000000..ce00e4abf5
--- /dev/null
+++ b/libavcodec/arm/vp8dsp.h
@@ -0,0 +1,78 @@
+/*
+ * This file is part of Libav.
+ *
+ * Libav is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * Libav is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with Libav; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef AVCODEC_ARM_VP8DSP_H
+#define AVCODEC_ARM_VP8DSP_H
+
+#include "libavcodec/vp8dsp.h"
+
+void ff_vp8dsp_init_armv6(VP8DSPContext *dsp);
+void ff_vp8dsp_init_neon(VP8DSPContext *dsp);
+
+#define VP8_LF_Y(hv, inner, opt) \
+ void ff_vp8_##hv##_loop_filter16##inner##_##opt(uint8_t *dst, \
+ ptrdiff_t stride, \
+ int flim_E, int flim_I, \
+ int hev_thresh)
+
+#define VP8_LF_UV(hv, inner, opt) \
+ void ff_vp8_##hv##_loop_filter8uv##inner##_##opt(uint8_t *dstU, \
+ uint8_t *dstV, \
+ ptrdiff_t stride, \
+ int flim_E, int flim_I, \
+ int hev_thresh)
+
+#define VP8_LF_SIMPLE(hv, opt) \
+ void ff_vp8_##hv##_loop_filter16_simple_##opt(uint8_t *dst, \
+ ptrdiff_t stride, \
+ int flim)
+
+#define VP8_LF_HV(inner, opt) \
+ VP8_LF_Y(h, inner, opt); \
+ VP8_LF_Y(v, inner, opt); \
+ VP8_LF_UV(h, inner, opt); \
+ VP8_LF_UV(v, inner, opt)
+
+#define VP8_LF(opt) \
+ VP8_LF_HV(, opt); \
+ VP8_LF_HV(_inner, opt); \
+ VP8_LF_SIMPLE(h, opt); \
+ VP8_LF_SIMPLE(v, opt)
+
+#define VP8_MC(n, opt) \
+ void ff_put_vp8_##n##_##opt(uint8_t *dst, ptrdiff_t dststride, \
+ uint8_t *src, ptrdiff_t srcstride, \
+ int h, int x, int y)
+
+#define VP8_EPEL(w, opt) \
+ VP8_MC(pixels ## w, opt); \
+ VP8_MC(epel ## w ## _h4, opt); \
+ VP8_MC(epel ## w ## _h6, opt); \
+ VP8_MC(epel ## w ## _v4, opt); \
+ VP8_MC(epel ## w ## _h4v4, opt); \
+ VP8_MC(epel ## w ## _h6v4, opt); \
+ VP8_MC(epel ## w ## _v6, opt); \
+ VP8_MC(epel ## w ## _h4v6, opt); \
+ VP8_MC(epel ## w ## _h6v6, opt)
+
+#define VP8_BILIN(w, opt) \
+ VP8_MC(bilin ## w ## _h, opt); \
+ VP8_MC(bilin ## w ## _v, opt); \
+ VP8_MC(bilin ## w ## _hv, opt)
+
+#endif /* AVCODEC_ARM_VP8DSP_H */
diff --git a/libavcodec/arm/vp8dsp_armv6.S b/libavcodec/arm/vp8dsp_armv6.S
index 4e7b78361e..08054ff262 100644
--- a/libavcodec/arm/vp8dsp_armv6.S
+++ b/libavcodec/arm/vp8dsp_armv6.S
@@ -1,7 +1,7 @@
-/**
+/*
* VP8 ARMv6 optimisations
*
- * Copyright (c) 2011 The WebM project authors. All Rights Reserved.
+ * Copyright (c) 2010 Google Inc.
* Copyright (c) 2010 Rob Clark <rob@ti.com>
* Copyright (c) 2011 Mans Rullgard <mans@mansr.com>
*
@@ -23,14 +23,33 @@
*
* This code was partially ported from libvpx, which uses this license:
*
- * Use of this source code is governed by a BSD-style license
- * that can be found in the LICENSE file in the root of the source
- * tree. An additional intellectual property rights grant can be found
- * in the file PATENTS. All contributing project authors may
- * be found in the AUTHORS file in the root of the source tree.
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are
+ * met:
*
- * (Note that the "LICENSE", "AUTHORS" and "PATENTS" files can be
- * found in the libvpx source tree.)
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ *
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ *
+ * * Neither the name of Google nor the names of its contributors may
+ * be used to endorse or promote products derived from this software
+ * without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "asm.S"
@@ -39,50 +58,37 @@
@ void vp8_luma_dc_wht(DCTELEM block[4][4][16], DCTELEM dc[16])
function ff_vp8_luma_dc_wht_armv6, export=1
- push {r4 - r10, lr}
+ push {r4-r10, lr}
- @ load dc[] and zero memory
- mov r12, #0
- ldr r2, [r1] @ dc0[0,1]
- ldr r3, [r1, #4] @ dc0[2,3]
- ldr r4, [r1, #8] @ dc1[0,1]
- ldr r5, [r1, #12] @ dc1[2,3]
- ldr r6, [r1, #16] @ dc2[0,1]
- ldr r7, [r1, #20] @ dc2[2,3]
- ldr r8, [r1, #24] @ dc3[0,1]
- ldr r9, [r1, #28] @ dc3[2,3]
- str r12,[r1]
- str r12,[r1, #4]
- str r12,[r1, #8]
- str r12,[r1, #12]
- str r12,[r1, #16]
- str r12,[r1, #20]
- str r12,[r1, #24]
- str r12,[r1, #28]
-
- @ loop1
+ ldm r1, {r2-r9}
+ mov r10, #0
+ mov lr, #0
uadd16 r12, r2, r8 @ t0[0,1]
- uadd16 r14, r3, r9 @ t0[2,3]
usub16 r2, r2, r8 @ t3[0,1]
- usub16 r3, r3, r9 @ t3[2,3]
+ stm r1!, {r10, lr}
uadd16 r8, r4, r6 @ t1[0,1]
- uadd16 r9, r5, r7 @ t1[2,3]
usub16 r4, r4, r6 @ t2[0,1]
- usub16 r5, r5, r7 @ t2[2,3]
-
+ stm r1!, {r10, lr}
uadd16 r6, r12, r8 @ dc0[0,1]
- uadd16 r7, r14, r9 @ dc0[2,3]
usub16 r12, r12, r8 @ dc2[0,1]
- usub16 r14, r14, r9 @ dc2[2,3]
+ stm r1!, {r10, lr}
uadd16 r8, r2, r4 @ dc1[0,1]
- uadd16 r9, r3, r5 @ dc1[2,3]
usub16 r2, r2, r4 @ dc3[0,1]
+ stm r1!, {r10, lr}
+
+ uadd16 lr, r3, r9 @ t0[2,3]
+ usub16 r3, r3, r9 @ t3[2,3]
+ uadd16 r9, r5, r7 @ t1[2,3]
+ usub16 r5, r5, r7 @ t2[2,3]
+
+ uadd16 r7, lr, r9 @ dc0[2,3]
+ usub16 lr, lr, r9 @ dc2[2,3]
+ uadd16 r9, r3, r5 @ dc1[2,3]
usub16 r3, r3, r5 @ dc3[2,3]
mov r1, #3
orr r1, r1, #0x30000 @ 3 | 3 (round)
- @ "transpose"
pkhbt r4, r6, r8, lsl #16 @ dc{0,1}[0]
pkhtb r6, r8, r6, asr #16 @ dc{0,1}[1]
pkhbt r5, r12, r2, lsl #16 @ dc{2,3}[0]
@@ -91,33 +97,33 @@ function ff_vp8_luma_dc_wht_armv6, export=1
uadd16 r4, r4, r1
uadd16 r5, r5, r1
pkhtb r7, r9, r7, asr #16 @ dc{0,1}[3]
- pkhbt r2, r14, r3, lsl #16 @ dc{2,3}[2]
- pkhtb r14, r3, r14, asr #16 @ dc{2,3}[3]
+ pkhbt r2, lr, r3, lsl #16 @ dc{2,3}[2]
+ pkhtb lr, r3, lr, asr #16 @ dc{2,3}[3]
- @ loop2
uadd16 r9, r4, r7 @ t0[0,1]
- uadd16 r3, r5, r14 @ t0[2,3]
+ uadd16 r3, r5, lr @ t0[2,3]
usub16 r4, r4, r7 @ t3[0,1]
- usub16 r5, r5, r14 @ t3[2,3]
+ usub16 r5, r5, lr @ t3[2,3]
uadd16 r7, r6, r8 @ t1[0,1]
- uadd16 r14, r12, r2 @ t1[2,3]
+ uadd16 lr, r12, r2 @ t1[2,3]
usub16 r6, r6, r8 @ t2[0,1]
usub16 r12, r12, r2 @ t2[2,3]
uadd16 r8, r9, r7 @ block[0,1][0]
- uadd16 r2, r3, r14 @ block[2,3][0]
+ uadd16 r2, r3, lr @ block[2,3][0]
usub16 r9, r9, r7 @ block[0,1][2]
- usub16 r3, r3, r14 @ block[2,3][2]
+ usub16 r3, r3, lr @ block[2,3][2]
uadd16 r7, r4, r6 @ block[0,1][1]
- uadd16 r14, r5, r12 @ block[2,3][1]
+ uadd16 lr, r5, r12 @ block[2,3][1]
usub16 r4, r4, r6 @ block[0,1][3]
usub16 r5, r5, r12 @ block[2,3][3]
- @ store
- mov r6, r8, asr #19 @ block[1][0]
- mov r12, r7, asr #19 @ block[1][1]
- mov r1, r9, asr #19 @ block[1][2]
- mov r10, r4, asr #19 @ block[1][3]
+#if HAVE_ARMV6T2
+ sbfx r6, r8, #3, #13
+ sbfx r12, r7, #3, #13
+ sbfx r1, r9, #3, #13
+ sbfx r10, r4, #3, #13
+#else
sxth r8, r8
sxth r7, r7
sxth r9, r9
@@ -126,162 +132,172 @@ function ff_vp8_luma_dc_wht_armv6, export=1
asr r7, #3 @ block[0][1]
asr r9, #3 @ block[0][2]
asr r4, #3 @ block[0][3]
-
- strh r8, [r0], #32
- strh r7, [r0], #32
- strh r9, [r0], #32
- strh r4, [r0], #32
- strh r6, [r0], #32
- strh r12,[r0], #32
- strh r1, [r0], #32
- strh r10,[r0], #32
-
- mov r6, r2, asr #19 @ block[3][0]
- mov r12, r14, asr #19 @ block[3][1]
- mov r1, r3, asr #19 @ block[3][2]
- mov r10, r5, asr #19 @ block[3][3]
+#endif
+
+ strh r6, [r0], #32
+ asr r8, r8, #19 @ block[1][0]
+ strh r12, [r0], #32
+ asr r7, r7, #19 @ block[1][1]
+ strh r1, [r0], #32
+ asr r9, r9, #19 @ block[1][2]
+ strh r10, [r0], #32
+ asr r4, r4, #19 @ block[1][3]
+ strh r8, [r0], #32
+ asr r6, r2, #19 @ block[3][0]
+ strh r7, [r0], #32
+ asr r12, lr, #19 @ block[3][1]
+ strh r9, [r0], #32
+ asr r1, r3, #19 @ block[3][2]
+ strh r4, [r0], #32
+ asr r10, r5, #19 @ block[3][3]
+
+#if HAVE_ARMV6T2
+ sbfx r2, r2, #3, #13
+ sbfx lr, lr, #3, #13
+ sbfx r3, r3, #3, #13
+ sbfx r5, r5, #3, #13
+#else
sxth r2, r2
- sxth r14, r14
+ sxth lr, lr
sxth r3, r3
sxth r5, r5
asr r2, #3 @ block[2][0]
- asr r14, #3 @ block[2][1]
+ asr lr, #3 @ block[2][1]
asr r3, #3 @ block[2][2]
asr r5, #3 @ block[2][3]
-
- strh r2, [r0], #32
- strh r14,[r0], #32
- strh r3, [r0], #32
- strh r5, [r0], #32
- strh r6, [r0], #32
- strh r12,[r0], #32
- strh r1, [r0], #32
- strh r10,[r0], #32
-
- pop {r4 - r10, pc}
+#endif
+
+ strh r2, [r0], #32
+ strh lr, [r0], #32
+ strh r3, [r0], #32
+ strh r5, [r0], #32
+ strh r6, [r0], #32
+ strh r12, [r0], #32
+ strh r1, [r0], #32
+ strh r10, [r0], #32
+
+ pop {r4-r10, pc}
endfunc
@ void vp8_luma_dc_wht_dc(DCTELEM block[4][4][16], DCTELEM dc[16])
function ff_vp8_luma_dc_wht_dc_armv6, export=1
- ldrsh r2, [r1]
+ ldrsh r2, [r1]
mov r3, #0
add r2, r2, #3
- strh r3, [r1]
+ strh r3, [r1]
asr r2, r2, #3
.rept 16
- strh r2, [r0], #32
+ strh r2, [r0], #32
.endr
bx lr
endfunc
@ void vp8_idct_add(uint8_t *dst, DCTELEM block[16], int stride)
function ff_vp8_idct_add_armv6, export=1
- push {r4 - r11, lr}
+ push {r4-r12, lr}
sub sp, sp, #32
- mov r3, #0x00004E00 @ cos
- orr r3, r3, #0x0000007B @ cospi8sqrt2minus1 = 20091
- mov r4, #0x00008A00 @ sin
- orr r4, r4, #0x0000008C @ sinpi8sqrt2 = 35468
- mov r5, #0x2 @ i=2
+ movw r3, #20091 @ cospi8sqrt2minus1
+ movw r4, #35468 @ sinpi8sqrt2
+ mov r5, sp
1:
- ldr r6, [r1, #8] @ i5 | i4 = block1[1] | block1[0]
- ldr r12,[r1, #24] @ i13 | i12 = block3[1] | block3[0]
- ldr r14,[r1, #16] @ i9 | i8 = block2[1] | block2[0]
-
- smulwt r9, r3, r6 @ (ip[5] * cospi8sqrt2minus1) >> 16
- smulwb r7, r3, r6 @ (ip[4] * cospi8sqrt2minus1) >> 16
- smulwt r10, r4, r6 @ (ip[5] * sinpi8sqrt2) >> 16
- smulwb r8, r4, r6 @ (ip[4] * sinpi8sqrt2) >> 16
+ ldr r6, [r1, #8] @ i5 | i4 = block1[1] | block1[0]
+ ldr lr, [r1, #16] @ i9 | i8 = block2[1] | block2[0]
+ ldr r12, [r1, #24] @ i13 | i12 = block3[1] | block3[0]
+
+ smulwt r9, r3, r6 @ ip[5] * cospi8sqrt2minus1
+ smulwb r7, r3, r6 @ ip[4] * cospi8sqrt2minus1
+ smulwt r10, r4, r6 @ ip[5] * sinpi8sqrt2
+ smulwb r8, r4, r6 @ ip[4] * sinpi8sqrt2
pkhbt r7, r7, r9, lsl #16 @ 5c | 4c
- smulwt r11, r3, r12 @ (ip[13] * cospi8sqrt2minus1) >> 16
- pkhbt r8, r8, r10, lsl #16 @ 5s | 4s = t2 first half
- uadd16 r6, r6, r7 @ 5c+5 | 4c+4 = t3 first half
- smulwt r7, r4, r12 @ (ip[13] * sinpi8sqrt2) >> 16
- smulwb r9, r3, r12 @ (ip[12] * cospi8sqrt2minus1) >> 16
- smulwb r10, r4, r12 @ (ip[12] * sinpi8sqrt2) >> 16
-
- subs r5, r5, #1 @ i--
+ smulwt r11, r3, r12 @ ip[13] * cospi8sqrt2minus1
+ pkhbt r8, r8, r10, lsl #16 @ 5s | 4s = t2 first half
+ uadd16 r6, r6, r7 @ 5c+5 | 4c+4 = t3 first half
+ smulwb r9, r3, r12 @ ip[12] * cospi8sqrt2minus1
+ smulwt r7, r4, r12 @ ip[13] * sinpi8sqrt2
+ smulwb r10, r4, r12 @ ip[12] * sinpi8sqrt2
+
pkhbt r9, r9, r11, lsl #16 @ 13c | 12c
- ldr r11,[r1] @ i1 | i0
- pkhbt r10, r10, r7, lsl #16 @ 13s | 12s = t3 second half
- uadd16 r7, r12, r9 @ 13c+13 | 12c+12 = t2 second half
- usub16 r7, r8, r7 @ c = t2
+ ldr r11, [r1] @ i1 | i0
+ pkhbt r10, r10, r7, lsl #16 @ 13s | 12s = t3 second half
+ uadd16 r7, r12, r9 @ 13c+13 | 12c+12 = t2 2nd half
uadd16 r6, r6, r10 @ d = t3
- uadd16 r10, r11, r14 @ a = t0
- usub16 r8, r11, r14 @ b = t1
+ uadd16 r10, r11, lr @ a = t0
+ usub16 r7, r8, r7 @ c = t2
+ usub16 r8, r11, lr @ b = t1
uadd16 r9, r10, r6 @ a+d = tmp{0,1}[0]
usub16 r10, r10, r6 @ a-d = tmp{0,1}[3]
uadd16 r6, r8, r7 @ b+c = tmp{0,1}[1]
usub16 r7, r8, r7 @ b-c = tmp{0,1}[2]
mov r8, #0
- str r6, [sp, #8] @ o5 | o4
- str r7, [sp, #16] @ o9 | o8
- str r10,[sp, #24] @ o13 | o12
- str r9, [sp], #4 @ o1 | o0
- str r8, [r1, #24]
- str r8, [r1, #16]
- str r8, [r1, #8]
- str r8, [r1], #4
- bne 1b
-
- mov r5, #0x2 @ i=2
- sub sp, sp, #8
+ cmp sp, r5
+ str r6, [r5, #8] @ o5 | o4
+ str r7, [r5, #16] @ o9 | o8
+ str r10, [r5, #24] @ o13 | o12
+ str r9, [r5], #4 @ o1 | o0
+ str r8, [r1, #8]
+ str r8, [r1, #16]
+ str r8, [r1, #24]
+ str r8, [r1], #4
+ beq 1b
+
+ mov r5, #2
2:
- ldr r6, [sp, #8] @ i5 | i4 = tmp{0,1}[1]
- ldr r14,[sp, #4] @ i3 | i2 = tmp{2,3}[0]
- ldr r12,[sp, #12] @ i7 | i6 = tmp{2,3}[1]
- ldr r1, [sp], #16 @ i1 | i0 = tmp{0,1}[0]
- smulwt r9, r3, r6 @ (ip[5] * cospi8sqrt2minus1) >> 16
- smulwt r7, r3, r1 @ (ip[1] * cospi8sqrt2minus1) >> 16
- smulwt r10, r4, r6 @ (ip[5] * sinpi8sqrt2) >> 16
- smulwt r8, r4, r1 @ (ip[1] * sinpi8sqrt2) >> 16
- pkhbt r11, r1, r6, lsl #16 @ i4 | i0 = t0/t1 first half
+ pop {r1, r6, r12, lr}
+ smulwt r9, r3, r12 @ ip[5] * cospi8sqrt2minus1
+ smulwt r7, r3, r1 @ ip[1] * cospi8sqrt2minus1
+ smulwt r10, r4, r12 @ ip[5] * sinpi8sqrt2
+ smulwt r8, r4, r1 @ ip[1] * sinpi8sqrt2
+ pkhbt r11, r1, r12, lsl #16 @ i4 | i0 = t0/t1 first half
+ pkhtb r1, r12, r1, asr #16 @ i5 | i1
pkhbt r7, r7, r9, lsl #16 @ 5c | 1c
- pkhbt r8, r8, r10, lsl #16 @ 5s | 1s = temp1 = t2 first half
- pkhtb r1, r6, r1, asr #16 @ i5 | i1
- uadd16 r1, r7, r1 @ 5c+5 | 1c+1 = temp2 (d) = t3 first half
- pkhbt r9, r14, r12, lsl #16 @ i6 | i2 = t0/t1 second half
+ pkhbt r8, r8, r10, lsl #16 @ 5s | 1s = t2 first half
+ pkhbt r9, r6, lr, lsl #16 @ i6 | i2 = t0/t1 second half
+ pkhtb r12, lr, r6, asr #16 @ i7 | i3
+ uadd16 r1, r7, r1 @ 5c+5 | 1c+1 = t3 first half
uadd16 r10, r11, r9 @ a = t0
usub16 r9, r11, r9 @ b = t1
- pkhtb r6, r12, r14, asr #16 @ i7 | i3
- subs r5, r5, #0x1 @ i--
- smulwt r7, r3, r6 @ (ip[7] * cospi8sqrt2minus1) >> 16
- smulwt r11, r4, r6 @ (ip[7] * sinpi8sqrt2) >> 16
- smulwb r12, r3, r6 @ (ip[3] * cospi8sqrt2minus1) >> 16
- smulwb r14, r4, r6 @ (ip[3] * sinpi8sqrt2) >> 16
-
- pkhbt r7, r12, r7, lsl #16 @ 7c | 3c
- pkhbt r11, r14, r11, lsl #16 @ 7s | 3s = temp1 (d) = t3 second half
- mov r14, #0x4 @ set up 4's
- orr r14, r14, #0x40000 @ 4|4
- uadd16 r6, r7, r6 @ 7c+7 | 3c+3 = temp2 (c) = t2 second half
- usub16 r12, r8, r6 @ c (o5 | o1) = t2
- uadd16 r6, r11, r1 @ d (o7 | o3) = t3
- uadd16 r10, r10, r14 @ t0 + 4
- uadd16 r9, r9, r14 @ t1 + 4
- uadd16 r7, r10, r6 @ a+d = dst{0,1}[0]
- usub16 r6, r10, r6 @ a-d = dst{0,1}[3]
- uadd16 r10, r9, r12 @ b+c = dst{0,1}[1]
- usub16 r1, r9, r12 @ b-c = dst{0,1}[2]
-
- mov r9, r6, asr #3 @ o[1][3]
- mov r12, r1, asr #3 @ o[1][2]
- pkhtb r8, r12, r7, asr #19 @ o[1][0,2]
+ smulwt r7, r3, r12 @ ip[7] * cospi8sqrt2minus1
+ smulwb lr, r3, r12 @ ip[3] * cospi8sqrt2minus1
+ smulwt r11, r4, r12 @ ip[7] * sinpi8sqrt2
+ smulwb r6, r4, r12 @ ip[3] * sinpi8sqrt2
+ subs r5, r5, #1
+ pkhbt r7, lr, r7, lsl #16 @ 7c | 3c
+ pkhbt r11, r6, r11, lsl #16 @ 7s | 3s = t3 second half
+ mov r6, #0x4
+ orr r6, r6, #0x40000
+ uadd16 r12, r7, r12 @ 7c+7 | 3c+3 = t2 second half
+ uadd16 r10, r10, r6 @ t0 + 4
+ uadd16 r9, r9, r6 @ t1 + 4
+ usub16 lr, r8, r12 @ c (o5 | o1) = t2
+ uadd16 r12, r11, r1 @ d (o7 | o3) = t3
+ usub16 r1, r9, lr @ b-c = dst{0,1}[2]
+ uadd16 r7, r10, r12 @ a+d = dst{0,1}[0]
+ usub16 r12, r10, r12 @ a-d = dst{0,1}[3]
+ uadd16 r10, r9, lr @ b+c = dst{0,1}[1]
+
+ asr lr, r1, #3 @ o[1][2]
+ asr r9, r12, #3 @ o[1][3]
+ pkhtb r8, lr, r7, asr #19 @ o[1][0,2]
pkhtb r11, r9, r10, asr #19 @ o[1][1,3]
- ldr r12,[r0]
- ldr r9, [r0, r2]
+ ldr lr, [r0]
+ sxth r12, r12
+ ldr r9, [r0, r2]
+ sxth r1, r1
+#if HAVE_ARMV6T2
+ sbfx r7, r7, #3, #13
+ sbfx r10, r10, #3, #13
+#else
sxth r7, r7
- sxth r6, r6
sxth r10, r10
- sxth r1, r1
asr r7, #3 @ o[0][0]
asr r10, #3 @ o[0][1]
+#endif
pkhbt r7, r7, r1, lsl #13 @ o[0][0,2]
- pkhbt r10, r10, r6, lsl #13 @ o[0][1,3]
+ pkhbt r10, r10, r12, lsl #13 @ o[0][1,3]
- uxtab16 r7, r7, r12
- uxtab16 r10, r10, r12, ror #8
+ uxtab16 r7, r7, lr
+ uxtab16 r10, r10, lr, ror #8
uxtab16 r8, r8, r9
uxtab16 r11, r11, r9, ror #8
usat16 r7, #8, r7
@@ -290,26 +306,26 @@ function ff_vp8_idct_add_armv6, export=1
usat16 r11, #8, r11
orr r7, r7, r10, lsl #8
orr r8, r8, r11, lsl #8
- str r8, [r0, r2]
+ str r8, [r0, r2]
str_post r7, r0, r2, lsl #1
bne 2b
- pop {r4 - r11, pc}
+ pop {r4-r12, pc}
endfunc
@ void vp8_idct_dc_add(uint8_t *dst, DCTELEM block[16], int stride)
function ff_vp8_idct_dc_add_armv6, export=1
- push {r4 - r5, lr}
- ldrsh r3, [r1]
+ push {r4-r6, lr}
+ add r6, r0, r2, lsl #1
+ ldrsh r3, [r1]
mov r4, #0
add r3, r3, #4
+ strh r4, [r1], #32
asr r3, #3
- strh r4, [r1], #32
- ldr r4, [r0, r2]
- ldr_post r5, r0, r2, lsl #1
+ ldr r5, [r0]
+ ldr r4, [r0, r2]
pkhbt r3, r3, r3, lsl #16
-
uxtab16 lr, r3, r5 @ a1+2 | a1+0
uxtab16 r5, r3, r5, ror #8 @ a1+3 | a1+1
uxtab16 r12, r3, r4
@@ -319,14 +335,12 @@ function ff_vp8_idct_dc_add_armv6, export=1
usat16 r12, #8, r12
usat16 r4, #8, r4
orr lr, lr, r5, lsl #8
+ ldr r5, [r6]
orr r12, r12, r4, lsl #8
- ldr r5, [r0]
- ldr r4, [r0, r2]
- sub r0, r0, r2, lsl #1
- str r12,[r0, r2]
- str_post lr, r0, r2, lsl #1
-
+ ldr r4, [r6, r2]
+ str lr, [r0]
uxtab16 lr, r3, r5
+ str r12, [r0, r2]
uxtab16 r5, r3, r5, ror #8
uxtab16 r12, r3, r4
uxtab16 r4, r3, r4, ror #8
@@ -336,506 +350,340 @@ function ff_vp8_idct_dc_add_armv6, export=1
usat16 r4, #8, r4
orr lr, lr, r5, lsl #8
orr r12, r12, r4, lsl #8
-
- str r12,[r0, r2]
- str_post lr, r0, r2, lsl #1
-
- pop {r4 - r5, pc}
+ str lr, [r6]
+ str r12, [r6, r2]
+ pop {r4-r6, pc}
endfunc
@ void vp8_idct_dc_add4uv(uint8_t *dst, DCTELEM block[4][16], int stride)
function ff_vp8_idct_dc_add4uv_armv6, export=1
- push {lr}
+ push {r4, lr}
bl ff_vp8_idct_dc_add_armv6
- sub r0, r0, r2, lsl #2
add r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
+ add r0, r0, r2, lsl #2
sub r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
- sub r0, r0, r2, lsl #2
add r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
- pop {pc}
+ pop {r4, pc}
endfunc
@ void vp8_idct_dc_add4y(uint8_t *dst, DCTELEM block[4][16], int stride)
function ff_vp8_idct_dc_add4y_armv6, export=1
- push {lr}
+ push {r4, lr}
bl ff_vp8_idct_dc_add_armv6
- sub r0, r0, r2, lsl #2
add r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
- sub r0, r0, r2, lsl #2
add r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
- sub r0, r0, r2, lsl #2
add r0, r0, #4
bl ff_vp8_idct_dc_add_armv6
- pop {pc}
+ pop {r4, pc}
endfunc
@ loopfilter
-@ void vp8_v_loop_filter16_simple(uint8_t *dst, int stride, int flim)
-function ff_vp8_v_loop_filter16_simple_armv6, export=1
- push {r4 - r11, lr}
+.macro transpose o3, o2, o1, o0, i0, i1, i2, i3
+ uxtb16 \o1, \i1 @ xx 12 xx 10
+ uxtb16 \o0, \i0 @ xx 02 xx 00
+ uxtb16 \o3, \i3 @ xx 32 xx 30
+ uxtb16 \o2, \i2 @ xx 22 xx 20
+ orr \o1, \o0, \o1, lsl #8 @ 12 02 10 00
+ orr \o3, \o2, \o3, lsl #8 @ 32 22 30 20
- ldr_dpren r3, r0, r1, lsl #1 @ p1
- ldr_dpren r4, r0, r1 @ p0
- ldr r5, [r0] @ q0
- ldr r6, [r0, r1] @ q1
- orr r2, r2, r2, lsl #16
- mov r9, #4 @ count
- mov lr, #0 @ need 0 in a couple places
- orr r12, r2, r2, lsl #8 @ splat int -> byte
- ldr r2, c0x80808080
+ uxtb16 \i1, \i1, ror #8 @ xx 13 xx 11
+ uxtb16 \i3, \i3, ror #8 @ xx 33 xx 31
+ uxtb16 \i0, \i0, ror #8 @ xx 03 xx 01
+ uxtb16 \i2, \i2, ror #8 @ xx 23 xx 21
+ orr \i0, \i0, \i1, lsl #8 @ 13 03 11 01
+ orr \i2, \i2, \i3, lsl #8 @ 33 23 31 21
-1:
- @ vp8_simple_filter_mask()
+ pkhtb \o2, \o3, \o1, asr #16 @ 32 22 12 02
+ pkhbt \o0, \o1, \o3, lsl #16 @ 30 20 10 00
+ pkhtb \o3, \i2, \i0, asr #16 @ 33 23 13 03
+ pkhbt \o1, \i0, \i2, lsl #16 @ 31 21 11 01
+.endm
+
+.macro simple_filter
uqsub8 r7, r3, r6 @ p1 - q1
uqsub8 r8, r6, r3 @ q1 - p1
uqsub8 r10, r4, r5 @ p0 - q0
- uqsub8 r11, r5, r4 @ q0 - p0
- orr r8, r8, r7 @ abs(p1 - q1)
- orr r10, r10, r11 @ abs(p0 - q0)
- uqadd8 r10, r10, r10 @ abs(p0 - q0) * 2
- uhadd8 r8, r8, lr @ abs(p1 - q2) >> 1
- uqadd8 r10, r10, r8 @ abs(p0 - q0)*2 + abs(p1 - q1)/2
+ uqsub8 r9, r5, r4 @ q0 - p0
+ orr r7, r7, r8 @ abs(p1 - q1)
+ orr r9, r9, r10 @ abs(p0 - q0)
+ uhadd8 r7, r7, lr @ abs(p1 - q2) >> 1
+ uqadd8 r9, r9, r9 @ abs(p0 - q0) * 2
+ uqadd8 r7, r7, r9 @ abs(p0 - q0)*2 + abs(p1-q1)/2
mvn r8, #0
- usub8 r10, r12, r10 @ compare to flimit. usub8 sets GE flags
+ usub8 r10, r12, r7 @ compare to flimit
sel r10, r8, lr @ filter mask: F or 0
cmp r10, #0
- beq 2f @ skip filtering if all masks are 0x00
+ beq 2f
- @ vp8_simple_filter()
- eor r3, r3, r2 @ p1 offset to convert to a signed value
- eor r6, r6, r2 @ q1 offset to convert to a signed value
- eor r4, r4, r2 @ p0 offset to convert to a signed value
- eor r5, r5, r2 @ q0 offset to convert to a signed value
+ eor r3, r3, r2 @ ps1
+ eor r6, r6, r2 @ qs1
+ eor r4, r4, r2 @ ps0
+ eor r5, r5, r2 @ qs0
qsub8 r3, r3, r6 @ vp8_filter = p1 - q1
qsub8 r6, r5, r4 @ q0 - p0
qadd8 r3, r3, r6 @ += q0 - p0
- ldr r7, c0x04040404
+ lsr r7, r2, #5 @ 0x04040404
qadd8 r3, r3, r6 @ += q0 - p0
- ldr r8, c0x03030303
- qadd8 r3, r3, r6 @ vp8_filter = p1-q1 + 3*(q0-p0))
- @STALL
+ sub r9, r7, r2, lsr #7 @ 0x03030303
+ qadd8 r3, r3, r6 @ vp8_filter = p1-q1 + 3*(q0-p0)
and r3, r3, r10 @ vp8_filter &= mask
- qadd8 r7, r3, r7 @ Filter1 = vp8_filter + 4
- qadd8 r8, r3, r8 @ Filter2 = vp8_filter + 3
+ qadd8 r9, r3, r9 @ Filter2 = vp8_filter + 3
+ qadd8 r3, r3, r7 @ Filter1 = vp8_filter + 4
- shadd8 r7, r7, lr
- shadd8 r8, r8, lr
- shadd8 r7, r7, lr
- shadd8 r8, r8, lr
- shadd8 r7, r7, lr @ Filter1 >>= 3
- shadd8 r8, r8, lr @ Filter2 >>= 3
+ shadd8 r9, r9, lr
+ shadd8 r3, r3, lr
+ shadd8 r9, r9, lr
+ shadd8 r3, r3, lr
+ shadd8 r9, r9, lr @ Filter2 >>= 3
+ shadd8 r3, r3, lr @ Filter1 >>= 3
- qsub8 r5, r5, r7 @ u = q0 - Filter1
- qadd8 r4, r4, r8 @ u = p0 + Filter2
- eor r5, r5, r2 @ *oq0 = u^0x80
- eor r4, r4, r2 @ *op0 = u^0x80
-T sub r7, r0, r1
- str r5, [r0] @ store oq0 result
-A str r4, [r0, -r1] @ store op0 result
-T str r4, [r7]
+ qadd8 r4, r4, r9 @ u = p0 + Filter2
+ qsub8 r5, r5, r3 @ u = q0 - Filter1
+ eor r4, r4, r2 @ *op0 = u ^ 0x80
+ eor r5, r5, r2 @ *oq0 = u ^ 0x80
+.endm
-2:
- subs r9, r9, #1 @ counter--
- add r0, r0, #4 @ next row
-T itttt ne
-A ldrne r3, [r0, -r1, lsl #1] @ p1
-T subne r3, r0, r1, lsl #1
-T ldrne r3, [r3] @ p1
-A ldrne r4, [r0, -r1] @ p0
-T subne r4, r0, r1
-T ldrne r4, [r4] @ p0
-T itt ne
- ldrne r5, [r0] @ q0
- ldrne r6, [r0, r1] @ q1
+@ void vp8_v_loop_filter16_simple(uint8_t *dst, int stride, int flim)
+function ff_vp8_v_loop_filter16_simple_armv6, export=1
+ push {r4-r11, lr}
+ orr r2, r2, r2, lsl #16
+ mov r11, #4
+ mov lr, #0
+ orr r12, r2, r2, lsl #8
+ mov32 r2, 0x80808080
+1:
+ ldr_nreg r3, r0, r1, lsl #1 @ p1
+ ldr_nreg r4, r0, r1 @ p0
+ ldr r5, [r0] @ q0
+ ldr r6, [r0, r1] @ q1
+ simple_filter
+T sub r7, r0, r1
+ str r5, [r0] @ oq0
+A str r4, [r0, -r1] @ op0
+T str r4, [r7]
+2:
+ subs r11, r11, #1
+ add r0, r0, #4
bne 1b
- pop {r4 - r11, pc}
+ pop {r4-r11, pc}
endfunc
-c0x01010101: .long 0x01010101
-c0x03030303: .long 0x03030303
-c0x04040404: .long 0x04040404
-c0x7F7F7F7F: .long 0x7F7F7F7F
-c0x80808080: .long 0x80808080
-
-@ void vp8_v_loop_filter16_inner(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh)
-@ and
-@ void vp8_v_loop_filter8uv_inner(uint8_t *dstU, uint8_t *dstV, int stride,
-@ int fE, int fI, int hev_thresh)
-@ call:
-@ void vp8_v_loop_filter_inner(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh, int count)
-function ff_vp8_v_loop_filter_inner_armv6, export=1
- push {r4 - r11, lr}
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
- ldr r5, [sp, #40] @ counter
- ldr r6, [sp, #36] @ load thresh address
- sub sp, sp, #16 @ create temp buffer
-
- ldr r10,[r0, r1] @ p2
- ldr_post r9, r0, r1, lsl #1 @ p3
- ldr r12,[r0, r1] @ p0
- ldr_post r11, r0, r1, lsl #1 @ p1
-
- orr r2, r2, r2, lsl #16
- orr r3, r3, r3, lsl #16
- orr r6, r6, r6, lsl #16
- orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
- orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
- orr r3, r6, r6, lsl #8 @ thresh splat int -> byte
-
-1:
- @ vp8_filter_mask() function
- @ calculate breakout conditions
+.macro filter_mask_p
uqsub8 r6, r9, r10 @ p3 - p2
uqsub8 r7, r10, r9 @ p2 - p3
uqsub8 r8, r10, r11 @ p2 - p1
uqsub8 r10, r11, r10 @ p1 - p2
-
- orr r6, r6, r7 @ abs (p3-p2)
- orr r8, r8, r10 @ abs (p2-p1)
- uqsub8 lr, r6, r2 @ compare to limit. lr: vp8_filter_mask
+ orr r6, r6, r7 @ abs(p3-p2)
+ orr r8, r8, r10 @ abs(p2-p1)
+ uqsub8 lr, r6, r2 @ compare to limit
uqsub8 r8, r8, r2 @ compare to limit
uqsub8 r6, r11, r12 @ p1 - p0
orr lr, lr, r8
uqsub8 r7, r12, r11 @ p0 - p1
- ldr r10,[r0, r1] @ q1
- ldr_post r9, r0, r1, lsl #1 @ q0
- orr r6, r6, r7 @ abs (p1-p0)
+ orr r6, r6, r7 @ abs(p1-p0)
uqsub8 r7, r6, r2 @ compare to limit
- uqsub8 r8, r6, r3 @ compare to thresh -- save r8 for later
+ uqsub8 r8, r6, r3 @ compare to thresh
orr lr, lr, r7
+.endm
+.macro filter_mask_pq
uqsub8 r6, r11, r10 @ p1 - q1
uqsub8 r7, r10, r11 @ q1 - p1
uqsub8 r11, r12, r9 @ p0 - q0
uqsub8 r12, r9, r12 @ q0 - p0
- orr r6, r6, r7 @ abs (p1-q1)
- ldr r7, c0x7F7F7F7F
- orr r12, r11, r12 @ abs (p0-q0)
- ldr_post r11, r0, r1 @ q2
- uqadd8 r12, r12, r12 @ abs (p0-q0) * 2
- and r6, r7, r6, lsr #1 @ abs (p1-q1) / 2
+ orr r6, r6, r7 @ abs(p1-q1)
+ orr r12, r11, r12 @ abs(p0-q0)
+ mov32 r7, 0x7f7f7f7f
+ uqadd8 r12, r12, r12 @ abs(p0-q0) * 2
+ and r6, r7, r6, lsr #1 @ abs(p1-q1) / 2
+ uqadd8 r12, r12, r6 @ abs(p0-q0) * 2 + abs(p1-q1)/2
+.endm
+
+.macro filter_mask_v
+ filter_mask_p
+
+ ldr r10, [r0, r1] @ q1
+ ldr_post r9, r0, r1, lsl #1 @ q0
+
+ filter_mask_pq
+
+ ldr r11, [r0] @ q2
+
uqsub8 r7, r9, r10 @ q0 - q1
- uqadd8 r12, r12, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
uqsub8 r6, r10, r9 @ q1 - q0
uqsub8 r12, r12, r4 @ compare to flimit
uqsub8 r9, r11, r10 @ q2 - q1
-
- orr lr, lr, r12
-
- ldr_post r12, r0, r1 @ q3
uqsub8 r10, r10, r11 @ q1 - q2
- orr r6, r7, r6 @ abs (q1-q0)
- orr r10, r9, r10 @ abs (q2-q1)
+ orr lr, lr, r12
+ ldr r12, [r0, r1] @ q3
+ orr r6, r7, r6 @ abs(q1-q0)
+ orr r10, r9, r10 @ abs(q2-q1)
+ uqsub8 r9, r12, r11 @ q3 - q2
+ uqsub8 r11, r11, r12 @ q2 - q3
uqsub8 r7, r6, r2 @ compare to limit
uqsub8 r10, r10, r2 @ compare to limit
- uqsub8 r6, r6, r3 @ compare to thresh -- save r6 for later
+ uqsub8 r6, r6, r3 @ compare to thresh
+ orr r9, r9, r11 @ abs(q3-q2)
orr lr, lr, r7
orr lr, lr, r10
+ uqsub8 r9, r9, r2 @ compare to limit
+ orr lr, lr, r9
- uqsub8 r10, r12, r11 @ q3 - q2
- uqsub8 r9, r11, r12 @ q2 - q3
+ mov r12, #0
+ usub8 lr, r12, lr
+ mvn r11, #0
+ sel lr, r11, r12 @ filter mask
+ sub r0, r0, r1, lsl #1
+.endm
- mvn r11, #0 @ r11 == -1
+.macro filter_mask_h
+ transpose r12, r11, r10, r9, r6, r7, r8, lr
- orr r10, r10, r9 @ abs (q3-q2)
- uqsub8 r10, r10, r2 @ compare to limit
+ filter_mask_p
- mov r12, #0
- orr lr, lr, r10
+ stm sp, {r8, r11, r12, lr}
sub r0, r0, r1, lsl #2
+ add r0, r0, #4
- usub8 lr, r12, lr @ use usub8 instead of ssub8
- sel lr, r11, r12 @ filter mask: lr
+ ldr r7, [r0, r1]
+ ldr_post r6, r0, r1, lsl #1
+ ldr lr, [r0, r1]
+ ldr r8, [r0]
- cmp lr, #0
- beq 2f @ skip filtering
+ transpose r12, r11, r10, r9, r6, r7, r8, lr
- sub r0, r0, r1, lsl #1 @ move r0 pointer down by 6 lines
+ uqsub8 r8, r12, r11 @ q3 - q2
+ uqsub8 lr, r11, r12 @ q2 - q3
+ uqsub8 r7, r9, r10 @ q0 - q1
+ uqsub8 r6, r10, r9 @ q1 - q0
+ uqsub8 r12, r11, r10 @ q2 - q1
+ uqsub8 r11, r10, r11 @ q1 - q2
+ orr r8, r8, lr @ abs(q3-q2)
+ orr r6, r7, r6 @ abs(q1-q0)
+ orr r11, r12, r11 @ abs(q2-q1)
+ ldr lr, [sp, #12] @ load back (f)limit accumulator
+ uqsub8 r8, r8, r2 @ compare to limit
+ uqsub8 r7, r6, r2 @ compare to limit
+ uqsub8 r11, r11, r2 @ compare to limit
+ orr lr, lr, r8
+ uqsub8 r8, r6, r3 @ compare to thresh
+ orr lr, lr, r7
+ ldr r12, [sp, #8] @ p1
+ orr lr, lr, r11
- @vp8_hevmask() function
- @calculate high edge variance
- orr r10, r6, r8 @ calculate vp8_hevmask
+ ldr r11, [sp, #4] @ p0
- usub8 r10, r12, r10 @ use usub8 instead of ssub8
- sel r6, r12, r11 @ obtain vp8_hevmask: r6
+ filter_mask_pq
- @vp8_filter() function
- ldr r8, [r0, r1] @ p0
- ldr_post r7, r0, r1, lsl #1 @ p1
- ldr r12, c0x80808080
- ldr r10,[r0, r1] @ q1
- ldr_post r9, r0, r1, lsl #1 @ q0
+ mov r10, #0
+ uqsub8 r12, r12, r4 @ compare to flimit
+ mvn r11, #0
+ orr lr, lr, r12
+ usub8 lr, r10, lr
+ sel lr, r11, r10 @ filter mask
+.endm
- eor r7, r7, r12 @ p1 offset to convert to a signed value
- eor r8, r8, r12 @ p0 offset to convert to a signed value
- eor r9, r9, r12 @ q0 offset to convert to a signed value
- eor r10, r10, r12 @ q1 offset to convert to a signed value
+.macro filter inner
+ mov32 r12, 0x80808080
+ eor r11, r7, r12 @ ps1
+ eor r8, r8, r12 @ ps0
+ eor r9, r9, r12 @ qs0
+ eor r10, r10, r12 @ qs1
- str r9, [sp] @ store qs0 temporarily
- str r8, [sp, #4] @ store ps0 temporarily
- str r10,[sp, #8] @ store qs1 temporarily
- str r7, [sp, #12] @ store ps1 temporarily
+ stm sp, {r8-r11}
- qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
+ qsub8 r7, r11, r10 @ vp8_signed_char_clamp(ps1-qs1)
qsub8 r8, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
-
- and r7, r7, r6 @ vp8_filter (r7) &= hev
-
+ .if \inner
+ and r7, r7, r6 @ vp8_filter &= hev
+ .endif
qadd8 r7, r7, r8
- ldr r9, c0x03030303 @ r9 = 3 --modified for vp8
-
+ lsr r10, r12, #5 @ 0x04040404
qadd8 r7, r7, r8
- ldr r10, c0x04040404
-
+ sub r9, r10, r12, lsr #7 @ 0x03030303
qadd8 r7, r7, r8
- and r7, r7, lr @ vp8_filter &= mask@
- qadd8 r8, r7, r9 @ Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
- qadd8 r7, r7, r10 @ vp8_filter = vp8_signed_char_clamp(vp8_filter+4)
+ and r7, r7, lr @ vp8_filter &= mask
+ .if !\inner
+ mov r12, r7 @ Filter2
+ and r7, r7, r6 @ Filter2 &= hev
+ .endif
+ qadd8 lr, r7, r9 @ Filter2 = vp8_signed_char_clamp(vp8_filter+3)
+ qadd8 r7, r7, r10 @ Filter1 = vp8_signed_char_clamp(vp8_filter+4)
mov r9, #0
- shadd8 r8, r8, r9 @ Filter2 >>= 3
- shadd8 r7, r7, r9 @ vp8_filter >>= 3
- shadd8 r8, r8, r9
+ shadd8 lr, lr, r9 @ Filter2 >>= 3
+ shadd8 r7, r7, r9 @ Filter1 >>= 3
+ shadd8 lr, lr, r9
shadd8 r7, r7, r9
- shadd8 lr, r8, r9 @ lr: Filter2
- shadd8 r7, r7, r9 @ r7: filter
+ shadd8 lr, lr, r9 @ Filter2
+ shadd8 r7, r7, r9 @ Filter1
+.endm
- @calculate output
+.macro filter_v inner
+ orr r10, r6, r8 @ calculate vp8_hevmask
+ ldr_nreg r7, r0, r1, lsl #1 @ p1
+ usub8 r10, r12, r10
+ ldr_nreg r8, r0, r1 @ p0
+ sel r6, r12, r11 @ obtain vp8_hevmask
+ ldr r9, [r0] @ q0
+ ldr r10, [r0, r1] @ q1
+ filter \inner
+.endm
- ldr r8, [sp] @ load qs0
- ldr r9, [sp, #4] @ load ps0
+.macro filter_h inner
+ orr r9, r6, r8
+ usub8 r9, r12, r9
+ sel r6, r12, r11 @ hev mask
+
+ stm sp, {r6, lr}
- ldr r10, c0x01010101
+ ldr_nreg r12, r0, r1, lsl #1
+ ldr_nreg r11, r0, r1
+ ldr r6, [r0]
+ ldr lr, [r0, r1]
- qsub8 r8, r8, r7 @ u = vp8_signed_char_clamp(qs0 - vp8_filter)
- qadd8 r9, r9, lr @ u = vp8_signed_char_clamp(ps0 + Filter2)
+ transpose r10, r9, r8, r7, r12, r11, r6, lr
+
+ ldm sp, {r6, lr}
+ filter \inner
+.endm
+.macro filter_inner
+ ldm sp, {r8, r9}
+ lsr r10, r10, #2 @ 0x01010101
+ qadd8 r8, r8, lr @ u = vp8_signed_char_clamp(ps0 + Filter2)
mov lr, #0
+ qsub8 r9, r9, r7 @ u = vp8_signed_char_clamp(qs0 - Filter1)
sadd8 r7, r7, r10 @ vp8_filter += 1
+ ldr r10, [sp, #8] @ qs1
shadd8 r7, r7, lr @ vp8_filter >>= 1
-
- ldr r11,[sp, #12] @ load ps1
- ldr r10,[sp, #8] @ load qs1
-
+ eor r8, r8, r12 @ *op0 = u ^ 0x80
bic r7, r7, r6 @ vp8_filter &= ~hev
- sub r0, r0, r1, lsl #2
-
qadd8 r11, r11, r7 @ u = vp8_signed_char_clamp(ps1 + vp8_filter)
+ eor r9, r9, r12 @ *oq0 = u ^ 0x80
qsub8 r10, r10, r7 @ u = vp8_signed_char_clamp(qs1 - vp8_filter)
+ eor r11, r11, r12 @ *op1 = u ^ 0x80
+ eor r10, r10, r12 @ *oq1 = u ^ 0x80
+.endm
- eor r11, r11, r12 @ *op1 = u^0x80
- eor r9, r9, r12 @ *op0 = u^0x80
- eor r8, r8, r12 @ *oq0 = u^0x80
- eor r10, r10, r12 @ *oq1 = u^0x80
- str r9, [r0, r1] @ store op0 result
- str_post r11, r0, r1, lsl #1 @ store op1
- str r10,[r0, r1] @ store oq1
- str_post r8, r0, r1, lsl #1 @ store oq0 result
-
- sub r0, r0, r1, lsl #1
-
-2:
- add r0, r0, #4
- sub r0, r0, r1, lsl #2
-
- subs r5, r5, #1
-T ittt ne
- ldrne r10,[r0, r1] @ p2
-A ldrne r9, [r0], r1, lsl #1 @ p3
-T ldrne r9, [r0] @ p3
-T addne r0, r0, r1, lsl #1
-T ittt ne
- ldrne r12,[r0, r1] @ p0
-A ldrne r11,[r0], r1, lsl #1 @ p1
-T ldrne r11,[r0] @ p3
-T addne r0, r0, r1, lsl #1
-
- bne 1b
-
- add sp, sp, #16
- pop {r4 - r11, pc}
-endfunc
-
-@ void vp8_v_loop_filter16(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh)
-@ and
-@ void vp8_v_loop_filter8uv(uint8_t *dstU, uint8_t *dstV, int stride,
-@ int fE, int fI, int hev_thresh)
-@ call:
-@ void vp8_v_loop_filter(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh, int count)
-function ff_vp8_v_loop_filter_armv6, export=1
- push {r4 - r11, lr}
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
- ldr r5, [sp, #40] @ counter
- ldr r6, [sp, #36] @ load thresh address
- sub sp, sp, #16 @ create temp buffer
-
- ldr r10,[r0, r1] @ p2
- ldr_post r9, r0, r1, lsl #1 @ p3
- ldr r12,[r0, r1] @ p0
- ldr_post r11, r0, r1, lsl #1 @ p1
-
- orr r2, r2, r2, lsl #16
- orr r3, r3, r3, lsl #16
- orr r6, r6, r6, lsl #16
- orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
- orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
- orr r3, r6, r6, lsl #8 @ thresh splat int -> byte
-
-1:
- @ vp8_filter_mask() function
- @ calculate breakout conditions
- uqsub8 r6, r9, r10 @ p3 - p2
- uqsub8 r7, r10, r9 @ p2 - p3
- uqsub8 r8, r10, r11 @ p2 - p1
- uqsub8 r10, r11, r10 @ p1 - p2
-
- orr r6, r6, r7 @ abs (p3-p2)
- orr r8, r8, r10 @ abs (p2-p1)
- uqsub8 lr, r6, r2 @ compare to limit. lr: vp8_filter_mask
- uqsub8 r8, r8, r2 @ compare to limit
-
- uqsub8 r6, r11, r12 @ p1 - p0
- orr lr, lr, r8
- uqsub8 r7, r12, r11 @ p0 - p1
- ldr r10,[r0, r1] @ q1
- ldr_post r9, r0, r1, lsl #1 @ q0
- orr r6, r6, r7 @ abs (p1-p0)
- uqsub8 r7, r6, r2 @ compare to limit
- uqsub8 r8, r6, r3 @ compare to thresh -- save r8 for later
- orr lr, lr, r7
-
- uqsub8 r6, r11, r10 @ p1 - q1
- uqsub8 r7, r10, r11 @ q1 - p1
- uqsub8 r11, r12, r9 @ p0 - q0
- uqsub8 r12, r9, r12 @ q0 - p0
- orr r6, r6, r7 @ abs (p1-q1)
- ldr r7, c0x7F7F7F7F
- orr r12, r11, r12 @ abs (p0-q0)
- ldr_post r11, r0, r1 @ q2
- uqadd8 r12, r12, r12 @ abs (p0-q0) * 2
- and r6, r7, r6, lsr #1 @ abs (p1-q1) / 2
- uqsub8 r7, r9, r10 @ q0 - q1
- uqadd8 r12, r12, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
- uqsub8 r6, r10, r9 @ q1 - q0
- uqsub8 r12, r12, r4 @ compare to flimit
- uqsub8 r9, r11, r10 @ q2 - q1
-
- orr lr, lr, r12
-
- ldr_post r12, r0, r1 @ q3
-
- uqsub8 r10, r10, r11 @ q1 - q2
- orr r6, r7, r6 @ abs (q1-q0)
- orr r10, r9, r10 @ abs (q2-q1)
- uqsub8 r7, r6, r2 @ compare to limit
- uqsub8 r10, r10, r2 @ compare to limit
- uqsub8 r6, r6, r3 @ compare to thresh -- save r6 for later
- orr lr, lr, r7
- orr lr, lr, r10
-
- uqsub8 r10, r12, r11 @ q3 - q2
- uqsub8 r9, r11, r12 @ q2 - q3
-
- mvn r11, #0 @ r11 == -1
-
- orr r10, r10, r9 @ abs (q3-q2)
- uqsub8 r10, r10, r2 @ compare to limit
-
- mov r12, #0
-
- orr lr, lr, r10
-
- usub8 lr, r12, lr @ use usub8 instead of ssub8
- sel lr, r11, r12 @ filter mask: lr
-
- cmp lr, #0
- beq 2f @ skip filtering
-
- @vp8_hevmask() function
- @calculate high edge variance
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 6 lines
- sub r0, r0, r1, lsl #1
-
- orr r10, r6, r8
-
- usub8 r10, r12, r10
- sel r6, r12, r11 @ hev mask: r6
-
- @vp8_mbfilter() function
- @p2, q2 are only needed at the end. Do not need to load them in now.
- ldr r8, [r0, r1] @ p0
- ldr_post r7, r0, r1, lsl #1 @ p1
- ldr r12, c0x80808080
- ldr_post r9, r0, r1 @ q0
- ldr r10,[r0] @ q1
-
- eor r7, r7, r12 @ ps1
- eor r8, r8, r12 @ ps0
- eor r9, r9, r12 @ qs0
- eor r10, r10, r12 @ qs1
-
- qsub8 r12, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
- str r7, [sp, #12] @ store ps1 temporarily
- qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
- str r10,[sp, #8] @ store qs1 temporarily
- qadd8 r7, r7, r12
- str r9, [sp] @ store qs0 temporarily
- qadd8 r7, r7, r12
- str r8, [sp, #4] @ store ps0 temporarily
- qadd8 r7, r7, r12 @ vp8_filter: r7
-
- ldr r10, c0x03030303 @ r10 = 3 --modified for vp8
- ldr r9, c0x04040404
-
- and r7, r7, lr @ vp8_filter &= mask (lr is free)
-
- mov r12, r7 @ Filter2: r12
- and r12, r12, r6 @ Filter2 &= hev
-
- @save bottom 3 bits so that we round one side +4 and the other +3
- qadd8 r8, r12, r9 @ Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
- qadd8 r12, r12, r10 @ Filter2 (r12) = vp8_signed_char_clamp(Filter2+3)
-
- mov r10, #0
- shadd8 r8, r8, r10 @ Filter1 >>= 3
- shadd8 r12, r12, r10 @ Filter2 >>= 3
- shadd8 r8, r8, r10
- shadd8 r12, r12, r10
- shadd8 r8, r8, r10 @ r8: Filter1
- shadd8 r12, r12, r10 @ r12: Filter2
-
- ldr r9, [sp] @ load qs0
- ldr r11,[sp, #4] @ load ps0
-
- qsub8 r9, r9, r8 @ qs0 = vp8_signed_char_clamp(qs0 - Filter1)
- qadd8 r11, r11, r12 @ ps0 = vp8_signed_char_clamp(ps0 + Filter2)
-
- bic r12, r7, r6 @ vp8_filter &= ~hev ( r6 is free)
-
- @roughly 3/7th difference across boundary
- mov lr, #0x1b @ 27
- mov r7, #0x3f @ 63
+.macro filter_x c0
+ mov lr, \c0
+ mov r7, #63
sxtb16 r6, r12
sxtb16 r10, r12, ror #8
@@ -849,975 +697,467 @@ function ff_vp8_v_loop_filter_armv6, export=1
ssat r7, #8, r7, asr #7
ssat r10, #8, r10, asr #7
- ldr lr, c0x80808080
-
pkhbt r6, r8, r6, lsl #16
pkhbt r10, r7, r10, lsl #16
uxtb16 r6, r6
uxtb16 r10, r10
- sub r0, r0, r1
+ mov32 lr, 0x80808080
orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7)
-
qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs0 - u)
qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps0 + u)
- eor r8, r8, lr @ *oq0 = s^0x80
- str r8, [r0] @ store *oq0
- sub r0, r0, r1
- eor r10, r10, lr @ *op0 = s^0x80
- str r10,[r0] @ store *op0
-
- @roughly 2/7th difference across boundary
- mov lr, #0x12 @ 18
- mov r7, #0x3f @ 63
-
- sxtb16 r6, r12
- sxtb16 r10, r12, ror #8
- smlabb r8, r6, lr, r7
- smlatb r6, r6, lr, r7
- smlabb r9, r10, lr, r7
- smlatb r10, r10, lr, r7
- ssat r8, #8, r8, asr #7
- ssat r6, #8, r6, asr #7
- ssat r9, #8, r9, asr #7
- ssat r10, #8, r10, asr #7
-
- ldr lr, c0x80808080
+ eor r8, r8, lr @ *oq0 = s ^ 0x80
+ eor r10, r10, lr @ *op0 = s ^ 0x80
+.endm
- pkhbt r6, r8, r6, lsl #16
- pkhbt r10, r9, r10, lsl #16
+.macro filter_1
+ ldm sp, {r8, r9}
+ qadd8 r11, r8, lr
+ qsub8 r9, r9, r7
+ bic r12, r12, r6 @ vp8_filter &= ~hev
+ filter_x #27
+.endm
- ldr r9, [sp, #8] @ load qs1
- ldr r11, [sp, #12] @ load ps1
+.macro filter_2
+ ldr r9, [sp, #8] @ qs1
+ ldr r11, [sp, #12] @ ps1
+ filter_x #18
+.endm
- uxtb16 r6, r6
- uxtb16 r10, r10
+.macro filter_3
+ eor r9, r9, lr
+ eor r11, r11, lr
+ filter_x #9
+.endm
- sub r0, r0, r1
+function vp8_v_loop_filter_inner_armv6
+ mov r5, #4
+ sub sp, sp, #16
- orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7)
+ orr r2, r2, r2, lsl #16
+ orr r3, r3, r3, lsl #16
+ orr r6, r6, r6, lsl #16
+ orr r4, r2, r2, lsl #8 @ flimE
+ orr r2, r3, r3, lsl #8 @ flimI
+ orr r3, r6, r6, lsl #8 @ thresh
+1:
+ sub r0, r0, r1, lsl #2
+ ldr r10, [r0, r1] @ p2
+ ldr_post r9, r0, r1, lsl #1 @ p3
+ ldr r12, [r0, r1] @ p0
+ ldr_post r11, r0, r1, lsl #1 @ p1
- qadd8 r11, r11, r10 @ s = vp8_signed_char_clamp(ps1 + u)
- qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs1 - u)
- eor r11, r11, lr @ *op1 = s^0x80
- str_post r11, r0, r1 @ store *op1
- eor r8, r8, lr @ *oq1 = s^0x80
- add r0, r0, r1, lsl #1
+ filter_mask_v
+ cmp lr, #0
+ beq 2f
+ filter_v inner=1
+ filter_inner
+
+A str r11, [r0, -r1, lsl #1] @ op1
+A str r8, [r0, -r1] @ op0
+T sub r0, r0, r1, lsl #1
+T str r8, [r0, r1]
+T str_post r11, r0, r1, lsl #1
+ str r9, [r0] @ oq0
+ str r10, [r0, r1] @ oq1
+2:
+ add r0, r0, #4
+ cmp r5, #3
+ it eq
+ ldreq r0, [sp, #16]
+ subs r5, r5, #1
+ bne 1b
- mov r7, #0x3f @ 63
+ add sp, sp, #16
+ pop {r0, r4-r11, pc}
+endfunc
- str_post r8, r0, r1 @ store *oq1
+function ff_vp8_v_loop_filter16_inner_armv6, export=1
+ push {r4-r11, lr}
+ add r12, r0, #8
+ push {r12}
+ ldr r6, [sp, #40]
+ orr r2, r2, r2, lsl #16
+ b vp8_v_loop_filter_inner_armv6
+endfunc
- @roughly 1/7th difference across boundary
- mov lr, #0x9 @ 9
- ldr r9, [r0] @ load q2
+function ff_vp8_v_loop_filter8uv_inner_armv6, export=1
+ push {r1, r4-r11, lr}
+ mov r1, r2
+ orr r2, r3, r3, lsl #16
+ ldr r3, [sp, #40]
+ ldr r6, [sp, #44]
+ b vp8_v_loop_filter_inner_armv6
+endfunc
- sxtb16 r6, r12
- sxtb16 r10, r12, ror #8
- smlabb r8, r6, lr, r7
- smlatb r6, r6, lr, r7
- smlabb r12, r10, lr, r7
- smlatb r10, r10, lr, r7
- ssat r8, #8, r8, asr #7
- ssat r6, #8, r6, asr #7
- ssat r12, #8, r12, asr #7
- ssat r10, #8, r10, asr #7
+function vp8_v_loop_filter_armv6
+ mov r5, #4
+ sub sp, sp, #16
+ orr r3, r3, r3, lsl #16
+ orr r6, r6, r6, lsl #16
+ orr r4, r2, r2, lsl #8 @ flimE
+ orr r2, r3, r3, lsl #8 @ flimI
+ orr r3, r6, r6, lsl #8 @ thresh
+1:
sub r0, r0, r1, lsl #2
+ ldr r10, [r0, r1] @ p2
+ ldr_post r9, r0, r1, lsl #1 @ p3
+ ldr r12, [r0, r1] @ p0
+ ldr_post r11, r0, r1, lsl #1 @ p1
- pkhbt r6, r8, r6, lsl #16
- pkhbt r10, r12, r10, lsl #16
-
- sub r0, r0, r1
- ldr lr, c0x80808080
+ filter_mask_v
+ cmp lr, #0
+ beq 2f
- ldr r11, [r0] @ load p2
+ filter_v inner=0
+ filter_1
- uxtb16 r6, r6
- uxtb16 r10, r10
+ str r8, [r0] @ *oq0
+A str r10, [r0, -r1] @ *op0
+T sub r0, r0, r1, lsl #1
+T str r10, [r0, r1]
- eor r9, r9, lr
- eor r11, r11, lr
+ filter_2
- orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7)
+A str r10, [r0, -r1, lsl #1] @ *op1
+T str_post r10, r0, r1, lsl #1
+ str r8, [r0, r1] @ *oq1
- qadd8 r8, r11, r10 @ s = vp8_signed_char_clamp(ps2 + u)
- qsub8 r10, r9, r10 @ s = vp8_signed_char_clamp(qs2 - u)
- eor r8, r8, lr @ *op2 = s^0x80
- str_post r8, r0, r1, lsl #2 @ store *op2
+ ldr r9, [r0, r1, lsl #1] @ q2
add r0, r0, r1
- eor r10, r10, lr @ *oq2 = s^0x80
- str_post r10, r0, r1, lsl #1 @ store *oq2
+A ldr r11, [r0, -r1, lsl #2] @ p2
+T ldr_dpre r11, r0, r1, lsl #2
+
+ filter_3
+A str r10, [r0, -r1, lsl #2] @ *op2
+T str_post r10, r0, r1, lsl #2
+ str r8, [r0, r1] @ *oq2
+ sub r0, r0, r1
2:
add r0, r0, #4
- sub r0, r0, r1, lsl #3
+ cmp r5, #3
+ it eq
+ ldreq r0, [sp, #16]
subs r5, r5, #1
-
-T ittt ne
- ldrne r10,[r0, r1] @ p2
-A ldrne r9, [r0], r1, lsl #1 @ p3
-T ldrne r9, [r0] @ p3
-T addne r0, r0, r1, lsl #1
-T ittt ne
- ldrne r12,[r0, r1] @ p0
-A ldrne r11,[r0], r1, lsl #1 @ p1
-T ldrne r11,[r0] @ p3
-T addne r0, r0, r1, lsl #1
-
bne 1b
add sp, sp, #16
- pop {r4 - r11, pc}
+ pop {r0, r4-r11, pc}
endfunc
-.macro TRANSPOSE_MATRIX i0, i1, i2, i3, o3, o2, o1, o0
- @ input: $0, $1, $2, $3
- @ output: $4, $5, $6, $7
- @ i0: 03 02 01 00
- @ i1: 13 12 11 10
- @ i2: 23 22 21 20
- @ i3: 33 32 31 30
- @ o3 o2 o1 o0
-
- uxtb16 \o1, \i1 @ xx 12 xx 10
- uxtb16 \o0, \i0 @ xx 02 xx 00
- uxtb16 \o3, \i3 @ xx 32 xx 30
- uxtb16 \o2, \i2 @ xx 22 xx 20
- orr \o1, \o0, \o1, lsl #8 @ 12 02 10 00
- orr \o3, \o2, \o3, lsl #8 @ 32 22 30 20
-
- uxtb16 \i1, \i1, ror #8 @ xx 13 xx 11
- uxtb16 \i3, \i3, ror #8 @ xx 33 xx 31
- uxtb16 \i0, \i0, ror #8 @ xx 03 xx 01
- uxtb16 \i2, \i2, ror #8 @ xx 23 xx 21
- orr \i0, \i0, \i1, lsl #8 @ 13 03 11 01
- orr \i2, \i2, \i3, lsl #8 @ 33 23 31 21
-
- pkhtb \o2, \o3, \o1, asr #16 @ 32 22 12 02 -- p1
- pkhbt \o0, \o1, \o3, lsl #16 @ 30 20 10 00 -- p3
+function ff_vp8_v_loop_filter16_armv6, export=1
+ push {r4-r11, lr}
+ add r12, r0, #8
+ push {r12}
+ ldr r6, [sp, #40]
+ orr r2, r2, r2, lsl #16
+ b vp8_v_loop_filter_armv6
+endfunc
- pkhtb \o3, \i2, \i0, asr #16 @ 33 23 13 03 -- p0
- pkhbt \o1, \i0, \i2, lsl #16 @ 31 21 11 01 -- p2
-.endm
+function ff_vp8_v_loop_filter8uv_armv6, export=1
+ push {r1, r4-r11, lr}
+ mov r1, r2
+ orr r2, r3, r3, lsl #16
+ ldr r3, [sp, #40]
+ ldr r6, [sp, #44]
+ b vp8_v_loop_filter_armv6
+endfunc
@ void vp8_h_loop_filter16_simple(uint8_t *dst, int stride, int flim)
function ff_vp8_h_loop_filter16_simple_armv6, export=1
- push {r4 - r11, lr}
+ push {r4-r11, lr}
orr r12, r2, r2, lsl #16
- ldr r2, c0x80808080
+ mov32 r2, 0x80808080
orr r12, r12, r12, lsl #8
- @ load soure data to r7, r8, r9, r10
+ mov lr, #0
+ mov r11, #4
+1:
sub r0, r0, #2
- ldr r8, [r0, r1]
+ ldr r8, [r0, r1]
ldr_post r7, r0, r1, lsl #1
- ldr r10,[r0, r1]
+ ldr r10, [r0, r1]
ldr_post r9, r0, r1, lsl #1
add r0, r0, #2
-
- mov r11, #4 @ count (r11) for 4-in-parallel
-1:
- @transpose r7, r8, r9, r10 to r3, r4, r5, r6
- TRANSPOSE_MATRIX r7, r8, r9, r10, r6, r5, r4, r3
-
- @ vp8_simple_filter_mask() function
- uqsub8 r7, r3, r6 @ p1 - q1
- uqsub8 r8, r6, r3 @ q1 - p1
- uqsub8 r9, r4, r5 @ p0 - q0
- uqsub8 r10, r5, r4 @ q0 - p0
- orr r7, r7, r8 @ abs(p1 - q1)
- orr r9, r9, r10 @ abs(p0 - q0)
- mov r8, #0
- uqadd8 r9, r9, r9 @ abs(p0 - q0) * 2
- uhadd8 r7, r7, r8 @ abs(p1 - q1) / 2
- uqadd8 r7, r7, r9 @ abs(p0 - q0)*2 + abs(p1 - q1)/2
- mvn r10, #0 @ r10 == -1
-
- usub8 r7, r12, r7 @ compare to flimit
- sel lr, r10, r8 @ filter mask
-
- cmp lr, #0
- beq 2f @ skip filtering
-
- @vp8_simple_filter() function
- eor r3, r3, r2 @ p1 offset to convert to a signed value
- eor r6, r6, r2 @ q1 offset to convert to a signed value
- eor r4, r4, r2 @ p0 offset to convert to a signed value
- eor r5, r5, r2 @ q0 offset to convert to a signed value
-
- qsub8 r3, r3, r6 @ vp8_filter = p1 - q1
- qsub8 r6, r5, r4 @ q0 - p0
-
- qadd8 r3, r3, r6 @ vp8_filter += q0 - p0
- ldr r9, c0x03030303 @ r9 = 3
-
- qadd8 r3, r3, r6 @ vp8_filter += q0 - p0
- ldr r7, c0x04040404
-
- qadd8 r3, r3, r6 @ vp8_filter = p1-q1 + 3*(q0-p0))
- @STALL
- and r3, r3, lr @ vp8_filter &= mask
-
- qadd8 r9, r3, r9 @ Filter2 = vp8_filter + 3
- qadd8 r3, r3, r7 @ Filter1 = vp8_filter + 4
-
- shadd8 r9, r9, r8
- shadd8 r3, r3, r8
- shadd8 r9, r9, r8
- shadd8 r3, r3, r8
- shadd8 r9, r9, r8 @ Filter2 >>= 3
- shadd8 r3, r3, r8 @ Filter1 >>= 3
-
- @calculate output
+ transpose r6, r5, r4, r3, r7, r8, r9, r10
+ simple_filter
sub r0, r0, r1, lsl #2
+ sub r0, r0, #1
- qadd8 r4, r4, r9 @ u = p0 + Filter2
- qsub8 r5, r5, r3 @ u = q0 - Filter1
- eor r4, r4, r2 @ *op0 = u^0x80
- eor r5, r5, r2 @ *oq0 = u^0x80
-
- strb r4, [r0, #-1] @ store the result
- mov r4, r4, lsr #8
- strb_post r5, r0, r1
- mov r5, r5, lsr #8
-
- strb r4, [r0, #-1]
- mov r4, r4, lsr #8
- strb_post r5, r0, r1
- mov r5, r5, lsr #8
-
- strb r4, [r0, #-1]
- mov r4, r4, lsr #8
- strb_post r5, r0, r1
- mov r5, r5, lsr #8
-
- strb r4, [r0, #-1]
- strb_post r5, r0, r1
+ uxtb16 r6, r4
+ uxtb16 r8, r5
+ uxtb16 r7, r4, ror #8
+ uxtb16 r9, r5, ror #8
+ orr r6, r6, r8, lsl #8
+ orr r7, r7, r9, lsl #8
+ lsr r4, r6, #16
+ lsr r5, r7, #16
+ strh_post r6, r0, r1
+ strh_post r7, r0, r1
+ strh_post r4, r0, r1
+ strh_post r5, r0, r1
+ add r0, r0, #1
2:
subs r11, r11, #1
-
- @ load soure data to r7, r8, r9, r10
- sub r0, r0, #2
-T ittt ne
- ldrne r8, [r0, r1]
-A ldrne r7, [r0], r1, lsl #1
-T ldrne r7, [r0]
-T addne r0, r0, r1, lsl #1
-T ittt ne
- ldrne r10,[r0, r1]
-A ldrne r9, [r0], r1, lsl #1
-T ldrne r9, [r0]
-T addne r0, r0, r1, lsl #1
- add r0, r0, #2
-
bne 1b
- pop {r4 - r11, pc}
+ pop {r4-r11, pc}
endfunc
-@ void vp8_h_loop_filter16_inner(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh)
-@ and
-@ void vp8_h_loop_filter8uv_inner(uint8_t *dstU, uint8_t *dstV, int stride,
-@ int fE, int fI, int hev_thresh)
-@ call:
-@ void vp8_h_loop_filter_inner(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh, int count)
-function ff_vp8_h_loop_filter_inner_armv6, export=1
- push {r4 - r11, lr}
-
- sub r0, r0, #4 @ move r0 pointer down by 4
- ldr r5, [sp, #40] @ counter
- ldr r9, [sp, #36] @ load thresh address
- sub sp, sp, #16 @ create temp buffer
-
- ldr r7, [r0, r1] @ transpose will make it into p3-p0
- ldr_post r6, r0, r1, lsl #1 @ load source data
- ldr lr, [r0, r1]
- ldr_post r8, r0, r1, lsl #1
+function vp8_h_loop_filter_inner_armv6
+ mov r5, #4
+ sub sp, sp, #16
- orr r2, r2, r2, lsl #16
orr r3, r3, r3, lsl #16
orr r9, r9, r9, lsl #16
- orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
- orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
- orr r3, r9, r9, lsl #8 @ thresh splat int -> byte
-
+ orr r4, r2, r2, lsl #8 @ flimE
+ orr r2, r3, r3, lsl #8 @ flimI
+ orr r3, r9, r9, lsl #8 @ thresh
+ sub r0, r0, #4
1:
- @ vp8_filter_mask() function
- @ calculate breakout conditions
- @ transpose the source data for 4-in-parallel operation
- TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
-
- uqsub8 r7, r9, r10 @ p3 - p2
- uqsub8 r8, r10, r9 @ p2 - p3
- uqsub8 r9, r10, r11 @ p2 - p1
- uqsub8 r10, r11, r10 @ p1 - p2
- orr r7, r7, r8 @ abs (p3-p2)
- orr r10, r9, r10 @ abs (p2-p1)
- uqsub8 lr, r7, r2 @ compare to limit. lr: vp8_filter_mask
- uqsub8 r10, r10, r2 @ compare to limit
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
-
- orr lr, lr, r10
-
- uqsub8 r6, r11, r12 @ p1 - p0
- uqsub8 r7, r12, r11 @ p0 - p1
- add r0, r0, #4 @ move r0 pointer up by 4
- orr r6, r6, r7 @ abs (p1-p0)
- str r11,[sp, #12] @ save p1
- uqsub8 r10, r6, r2 @ compare to limit
- uqsub8 r11, r6, r3 @ compare to thresh
- orr lr, lr, r10
-
- @ transpose uses 8 regs(r6 - r12 and lr). Need to save reg value now
- @ transpose the source data for 4-in-parallel operation
- str r11,[sp] @ push r11 to stack
- ldr r7, [r0, r1]
- ldr_post r6, r0, r1, lsl #1 @ load source data
- str r12,[sp, #4] @ save current reg before load q0 - q3 data
- str lr, [sp, #8]
- ldr lr, [r0, r1]
+ ldr r7, [r0, r1]
+ ldr_post r6, r0, r1, lsl #1
+ ldr lr, [r0, r1]
ldr_post r8, r0, r1, lsl #1
- TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
-
- ldr lr, [sp, #8] @ load back (f)limit accumulator
-
- uqsub8 r6, r12, r11 @ q3 - q2
- uqsub8 r7, r11, r12 @ q2 - q3
- uqsub8 r12, r11, r10 @ q2 - q1
- uqsub8 r11, r10, r11 @ q1 - q2
- orr r6, r6, r7 @ abs (q3-q2)
- orr r7, r12, r11 @ abs (q2-q1)
- uqsub8 r6, r6, r2 @ compare to limit
- uqsub8 r7, r7, r2 @ compare to limit
- ldr r11,[sp, #4] @ load back p0
- ldr r12,[sp, #12] @ load back p1
- orr lr, lr, r6
- orr lr, lr, r7
-
- uqsub8 r6, r11, r9 @ p0 - q0
- uqsub8 r7, r9, r11 @ q0 - p0
- uqsub8 r8, r12, r10 @ p1 - q1
- uqsub8 r11, r10, r12 @ q1 - p1
- orr r6, r6, r7 @ abs (p0-q0)
- ldr r7, c0x7F7F7F7F
- orr r8, r8, r11 @ abs (p1-q1)
- uqadd8 r6, r6, r6 @ abs (p0-q0) * 2
- and r8, r7, r8, lsr #1 @ abs (p1-q1) / 2
- uqsub8 r11, r10, r9 @ q1 - q0
- uqadd8 r6, r8, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
- uqsub8 r12, r9, r10 @ q0 - q1
- uqsub8 r6, r6, r4 @ compare to flimit
-
- orr r9, r11, r12 @ abs (q1-q0)
- uqsub8 r8, r9, r2 @ compare to limit
- uqsub8 r10, r9, r3 @ compare to thresh
- orr lr, lr, r6
- orr lr, lr, r8
-
- mvn r11, #0 @ r11 == -1
- mov r12, #0
-
- usub8 lr, r12, lr
- ldr r9, [sp] @ load the compared result
- sel lr, r11, r12 @ filter mask: lr
+ filter_mask_h
cmp lr, #0
- beq 2f @ skip filtering
-
- @vp8_hevmask() function
- @calculate high edge variance
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
-
- orr r9, r9, r10
-
- ldrh r7, [r0, #-2]
- ldrh_post r8, r0, r1
-
- usub8 r9, r12, r9
- sel r6, r12, r11 @ hev mask: r6
-
- @vp8_filter() function
- @ load soure data to r6, r11, r12, lr
- ldrh r9, [r0, #-2]
- ldrh_post r10, r0, r1
-
- pkhbt r12, r7, r8, lsl #16
-
- ldrh r7, [r0, #-2]
- ldrh_post r8, r0, r1
-
- pkhbt r11, r9, r10, lsl #16
-
- ldrh r9, [r0, #-2]
- ldrh_post r10, r0, r1
-
- @ Transpose needs 8 regs(r6 - r12, and lr). Save r6 and lr first
- str r6, [sp]
- str lr, [sp, #4]
-
- pkhbt r6, r7, r8, lsl #16
- pkhbt lr, r9, r10, lsl #16
-
- @transpose r12, r11, r6, lr to r7, r8, r9, r10
- TRANSPOSE_MATRIX r12, r11, r6, lr, r10, r9, r8, r7
-
- @load back hev_mask r6 and filter_mask lr
- ldr r12, c0x80808080
- ldr r6, [sp]
- ldr lr, [sp, #4]
-
- eor r7, r7, r12 @ p1 offset to convert to a signed value
- eor r8, r8, r12 @ p0 offset to convert to a signed value
- eor r9, r9, r12 @ q0 offset to convert to a signed value
- eor r10, r10, r12 @ q1 offset to convert to a signed value
-
- str r9, [sp] @ store qs0 temporarily
- str r8, [sp, #4] @ store ps0 temporarily
- str r10,[sp, #8] @ store qs1 temporarily
- str r7, [sp, #12] @ store ps1 temporarily
-
- qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
- qsub8 r8, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
-
- and r7, r7, r6 @ vp8_filter (r7) &= hev (r7 : filter)
-
- qadd8 r7, r7, r8
- ldr r9, c0x03030303 @ r9 = 3 --modified for vp8
-
- qadd8 r7, r7, r8
- ldr r10, c0x04040404
-
- qadd8 r7, r7, r8
-
- and r7, r7, lr @ vp8_filter &= mask
-
- qadd8 r8, r7, r9 @ Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
- qadd8 r7, r7, r10 @ vp8_filter = vp8_signed_char_clamp(vp8_filter+4)
-
- mov r9, #0
- shadd8 r8, r8, r9 @ Filter2 >>= 3
- shadd8 r7, r7, r9 @ vp8_filter >>= 3
- shadd8 r8, r8, r9
- shadd8 r7, r7, r9
- shadd8 lr, r8, r9 @ lr: filter2
- shadd8 r7, r7, r9 @ r7: filter
-
- @calculate output
- ldr r8, [sp] @ load qs0
- ldr r9, [sp, #4] @ load ps0
-
- ldr r10, c0x01010101
-
- qsub8 r8, r8, r7 @ u = vp8_signed_char_clamp(qs0 - vp8_filter)
- qadd8 r9, r9, lr @ u = vp8_signed_char_clamp(ps0 + Filter2)
-
- eor r8, r8, r12
- eor r9, r9, r12
-
- mov lr, #0
-
- sadd8 r7, r7, r10
- shadd8 r7, r7, lr
-
- ldr r10,[sp, #8] @ load qs1
- ldr r11,[sp, #12] @ load ps1
-
- bic r7, r7, r6 @ r7: vp8_filter
-
- qsub8 r10, r10, r7 @ u = vp8_signed_char_clamp(qs1 - vp8_filter)
- qadd8 r11, r11, r7 @ u = vp8_signed_char_clamp(ps1 + vp8_filter)
- eor r10, r10, r12
- eor r11, r11, r12
-
- sub r0, r0, r1, lsl #2
-
- @we can use TRANSPOSE_MATRIX macro to transpose output - input: q1, q0, p0, p1
- TRANSPOSE_MATRIX r11, r9, r8, r10, lr, r12, r7, r6
+ sub r0, r0, #2
+ beq 2f
- strh r6, [r0, #-2] @ store the result
- mov r6, r6, lsr #16
- strh_post r6, r0, r1
+ ldr r6, [sp]
- strh r7, [r0, #-2]
- mov r7, r7, lsr #16
- strh_post r7, r0, r1
+ filter_h inner=1
+ filter_inner
- strh r12, [r0, #-2]
- mov r12, r12, lsr #16
- strh_post r12, r0, r1
-
- strh lr, [r0, #-2]
- mov lr, lr, lsr #16
- strh_post lr, r0, r1
+ transpose lr, r12, r7, r6, r11, r8, r9, r10
+A str r6, [r0, -r1, lsl #1]
+A str r7, [r0, -r1]
+T sub r0, r0, r1, lsl #1
+T str r7, [r0, r1]
+T str_post r6, r0, r1, lsl #1
+ str r12, [r0]
+ str lr, [r0, r1]
2:
- sub r0, r0, #4
+ sub r0, r0, #2
+ add r0, r0, r1, lsl #1
+ cmp r5, #3
+ it eq
+ ldreq r0, [sp, #16]
subs r5, r5, #1
-
-T ittt ne
- ldrne r7, [r0, r1]
-A ldrne r6, [r0], r1, lsl #1 @ load source data
-T ldrne r6, [r0] @ load source data
-T addne r0, r0, r1, lsl #1
-T ittt ne
- ldrne lr, [r0, r1]
-A ldrne r8, [r0], r1, lsl #1
-T ldrne r8, [r0]
-T addne r0, r0, r1, lsl #1
-
bne 1b
add sp, sp, #16
- pop {r4 - r11, pc}
+ pop {r0, r4-r11, pc}
endfunc
-@ void vp8_h_loop_filter16(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh)
-@ and
-@ void vp8_h_loop_filter8uv(uint8_t *dstU, uint8_t *dstV, int stride,
-@ int fE, int fI, int hev_thresh)
-@ call:
-@ void vp8_h_loop_filter(uint8_t *dst, int stride,
-@ int fE, int fI, int hev_thresh, int count)
-function ff_vp8_h_loop_filter_armv6, export=1
- push {r4 - r11, lr}
-
- sub r0, r0, #4 @ move r0 pointer down by 4
- ldr r5, [sp, #40] @ counter
- ldr r9, [sp, #36] @ load thresh address
- sub sp, sp, #16 @ create temp buffer
-
- ldr r7, [r0, r1] @ transpose will make it into p3-p0
- ldr_post r6, r0, r1, lsl #1 @ load source data
- ldr lr, [r0, r1]
- ldr_post r8, r0, r1, lsl #1
-
+function ff_vp8_h_loop_filter16_inner_armv6, export=1
+ push {r4-r11, lr}
+ add r12, r0, r1, lsl #3
+ sub r12, r12, #4
+ push {r12}
+ ldr r9, [sp, #40]
orr r2, r2, r2, lsl #16
- orr r3, r3, r3, lsl #16
- orr r9, r9, r9, lsl #16
- orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
- orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
- orr r3, r9, r9, lsl #8 @ thresh splat int -> byte
-
-1:
- @ vp8_filter_mask() function
- @ calculate breakout conditions
- @ transpose the source data for 4-in-parallel operation
- TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
-
- uqsub8 r7, r9, r10 @ p3 - p2
- uqsub8 r8, r10, r9 @ p2 - p3
- uqsub8 r9, r10, r11 @ p2 - p1
- uqsub8 r10, r11, r10 @ p1 - p2
- orr r7, r7, r8 @ abs (p3-p2)
- orr r10, r9, r10 @ abs (p2-p1)
- uqsub8 lr, r7, r2 @ compare to limit. lr: vp8_filter_mask
- uqsub8 r10, r10, r2 @ compare to limit
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
+ b vp8_h_loop_filter_inner_armv6
+endfunc
- orr lr, lr, r10
+function ff_vp8_h_loop_filter8uv_inner_armv6, export=1
+ sub r1, r1, #4
+ push {r1, r4-r11, lr}
+ mov r1, r2
+ orr r2, r3, r3, lsl #16
+ ldr r3, [sp, #40]
+ ldr r9, [sp, #44]
+ b vp8_h_loop_filter_inner_armv6
+endfunc
- uqsub8 r6, r11, r12 @ p1 - p0
- uqsub8 r7, r12, r11 @ p0 - p1
- add r0, r0, #4 @ move r0 pointer up by 4
- orr r6, r6, r7 @ abs (p1-p0)
- str r11,[sp, #12] @ save p1
- uqsub8 r10, r6, r2 @ compare to limit
- uqsub8 r11, r6, r3 @ compare to thresh
- orr lr, lr, r10
+function vp8_h_loop_filter_armv6
+ mov r5, #4
+ sub sp, sp, #16
- @ transpose uses 8 regs(r6 - r12 and lr). Need to save reg value now
- @ transpose the source data for 4-in-parallel operation
- str r11,[sp] @ push r11 to stack
- ldr r7, [r0, r1]
- ldr_post r6, r0, r1, lsl #1 @ load source data
- str r12,[sp, #4] @ save current reg before load q0 - q3 data
- str lr, [sp, #8]
- ldr lr, [r0, r1]
+ orr r3, r3, r3, lsl #16
+ orr r9, r9, r9, lsl #16
+ orr r4, r2, r2, lsl #8 @ flimE
+ orr r2, r3, r3, lsl #8 @ flimI
+ orr r3, r9, r9, lsl #8 @ thresh
+1:
+ sub r0, r0, #4
+ ldr r7, [r0, r1]
+ ldr_post r6, r0, r1, lsl #1
+ ldr lr, [r0, r1]
ldr_post r8, r0, r1, lsl #1
- TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
-
- ldr lr, [sp, #8] @ load back (f)limit accumulator
-
- uqsub8 r6, r12, r11 @ q3 - q2
- uqsub8 r7, r11, r12 @ q2 - q3
- uqsub8 r12, r11, r10 @ q2 - q1
- uqsub8 r11, r10, r11 @ q1 - q2
- orr r6, r6, r7 @ abs (q3-q2)
- orr r7, r12, r11 @ abs (q2-q1)
- uqsub8 r6, r6, r2 @ compare to limit
- uqsub8 r7, r7, r2 @ compare to limit
- ldr r11,[sp, #4] @ load back p0
- ldr r12,[sp, #12] @ load back p1
- orr lr, lr, r6
- orr lr, lr, r7
-
- uqsub8 r6, r11, r9 @ p0 - q0
- uqsub8 r7, r9, r11 @ q0 - p0
- uqsub8 r8, r12, r10 @ p1 - q1
- uqsub8 r11, r10, r12 @ q1 - p1
- orr r6, r6, r7 @ abs (p0-q0)
- ldr r7, c0x7F7F7F7F
- orr r8, r8, r11 @ abs (p1-q1)
- uqadd8 r6, r6, r6 @ abs (p0-q0) * 2
- and r8, r7, r8, lsr #1 @ abs (p1-q1) / 2
- uqsub8 r11, r10, r9 @ q1 - q0
- uqadd8 r6, r8, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
- uqsub8 r12, r9, r10 @ q0 - q1
- uqsub8 r6, r6, r4 @ compare to flimit
-
- orr r9, r11, r12 @ abs (q1-q0)
- uqsub8 r8, r9, r2 @ compare to limit
- uqsub8 r10, r9, r3 @ compare to thresh
- orr lr, lr, r6
- orr lr, lr, r8
-
- mvn r11, #0 @ r11 == -1
- mov r12, #0
-
- usub8 lr, r12, lr
- ldr r9, [sp] @ load the compared result
- sel lr, r11, r12 @ filter mask: lr
-
+ filter_mask_h
cmp lr, #0
- beq 2f @ skip filtering
-
-
- @vp8_hevmask() function
- @calculate high edge variance
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
-
- orr r9, r9, r10
-
- ldrh r7, [r0, #-2]
- ldrh_post r8, r0, r1
-
- usub8 r9, r12, r9
- sel r6, r12, r11 @ hev mask: r6
-
-
- @ vp8_mbfilter() function
- @ p2, q2 are only needed at the end. do not need to load them in now.
- @ Transpose needs 8 regs(r6 - r12, and lr). Save r6 and lr first
- @ load soure data to r6, r11, r12, lr
- ldrh r9, [r0, #-2]
- ldrh_post r10, r0, r1
-
- pkhbt r12, r7, r8, lsl #16
-
- ldrh r7, [r0, #-2]
- ldrh_post r8, r0, r1
-
- pkhbt r11, r9, r10, lsl #16
-
- ldrh r9, [r0, #-2]
- ldrh_post r10, r0, r1
-
- str r6, [sp] @ save r6
- str lr, [sp, #4] @ save lr
-
- pkhbt r6, r7, r8, lsl #16
- pkhbt lr, r9, r10, lsl #16
-
- @transpose r12, r11, r6, lr to p1, p0, q0, q1
- TRANSPOSE_MATRIX r12, r11, r6, lr, r10, r9, r8, r7
-
- @load back hev_mask r6 and filter_mask lr
- ldr r12, c0x80808080
- ldr r6, [sp]
- ldr lr, [sp, #4]
-
- eor r7, r7, r12 @ ps1
- eor r8, r8, r12 @ ps0
- eor r9, r9, r12 @ qs0
- eor r10, r10, r12 @ qs1
-
- qsub8 r12, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
- str r7, [sp, #12] @ store ps1 temporarily
- qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
- str r10,[sp, #8] @ store qs1 temporarily
- qadd8 r7, r7, r12
- str r9, [sp] @ store qs0 temporarily
- qadd8 r7, r7, r12
- str r8, [sp, #4] @ store ps0 temporarily
- qadd8 r7, r7, r12 @ vp8_filter: r7
-
- ldr r10, c0x03030303 @ r10 = 3 --modified for vp8
- ldr r9, c0x04040404
-
- and r7, r7, lr @ vp8_filter &= mask (lr is free)
-
- mov r12, r7 @ Filter2: r12
- and r12, r12, r6 @ Filter2 &= hev
-
- @save bottom 3 bits so that we round one side +4 and the other +3
- qadd8 r8, r12, r9 @ Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
- qadd8 r12, r12, r10 @ Filter2 (r12) = vp8_signed_char_clamp(Filter2+3)
-
- mov r10, #0
- shadd8 r8, r8, r10 @ Filter1 >>= 3
- shadd8 r12, r12, r10 @ Filter2 >>= 3
- shadd8 r8, r8, r10
- shadd8 r12, r12, r10
- shadd8 r8, r8, r10 @ r8: Filter1
- shadd8 r12, r12, r10 @ r12: Filter2
-
- ldr r9, [sp] @ load qs0
- ldr r11,[sp, #4] @ load ps0
-
- qsub8 r9, r9, r8 @ qs0 = vp8_signed_char_clamp(qs0 - Filter1)
- qadd8 r11, r11, r12 @ ps0 = vp8_signed_char_clamp(ps0 + Filter2)
-
- bic r12, r7, r6 @vp8_filter &= ~hev ( r6 is free)
-
- @roughly 3/7th difference across boundary
- mov lr, #0x1b @ 27
- mov r7, #0x3f @ 63
-
- sxtb16 r6, r12
- sxtb16 r10, r12, ror #8
- smlabb r8, r6, lr, r7
- smlatb r6, r6, lr, r7
- smlabb r7, r10, lr, r7
- smultb r10, r10, lr
- ssat r8, #8, r8, asr #7
- ssat r6, #8, r6, asr #7
- add r10, r10, #63
- ssat r7, #8, r7, asr #7
- ssat r10, #8, r10, asr #7
-
- ldr lr, c0x80808080
-
- pkhbt r6, r8, r6, lsl #16
- pkhbt r10, r7, r10, lsl #16
- uxtb16 r6, r6
- uxtb16 r10, r10
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
-
- orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7)
-
- qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs0 - u)
- qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps0 + u)
- eor r8, r8, lr @ *oq0 = s^0x80
- eor r10, r10, lr @ *op0 = s^0x80
-
- strb r10,[r0, #-1] @ store op0 result
- strb_post r8, r0, r1 @ store oq0 result
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
- strb r10,[r0, #-1]
- strb_post r8, r0, r1
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
- strb r10,[r0, #-1]
- strb_post r8, r0, r1
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
- strb r10,[r0, #-1]
- strb_post r8, r0, r1
-
- @roughly 2/7th difference across boundary
- mov lr, #0x12 @ 18
- mov r7, #0x3f @ 63
-
- sxtb16 r6, r12
- sxtb16 r10, r12, ror #8
- smlabb r8, r6, lr, r7
- smlatb r6, r6, lr, r7
- smlabb r9, r10, lr, r7
- smlatb r10, r10, lr, r7
- ssat r8, #8, r8, asr #7
- ssat r6, #8, r6, asr #7
- ssat r9, #8, r9, asr #7
- ssat r10, #8, r10, asr #7
-
- sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
+ it eq
+ addeq r0, r0, r1, lsl #1
+ beq 2f
- pkhbt r6, r8, r6, lsl #16
- pkhbt r10, r9, r10, lsl #16
+ ldr r6, [sp]
+ sub r0, r0, #2
- ldr r9, [sp, #8] @ load qs1
- ldr r11,[sp, #12] @ load ps1
- ldr lr, c0x80808080
+ filter_h inner=0
+ filter_1
- uxtb16 r6, r6
- uxtb16 r10, r10
+ sub r0, r0, r1, lsl #1
+ uxtb16 r6, r10
+ uxtb16 r7, r8
+ uxtb16 r10, r10, ror #8
+ uxtb16 r8, r8, ror #8
+ orr r6, r6, r7, lsl #8
+ orr r10, r10, r8, lsl #8
+ lsr r7, r6, #16
+ lsr r8, r10, #16
- add r0, r0, #2
+ add r0, r0, #1
+ strh_post r6, r0, r1
+ strh_post r10, r0, r1
+ strh_post r7, r0, r1
+ strh_post r8, r0, r1
- orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7)
+ filter_2
- qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs1 - u)
- qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps1 + u)
- eor r8, r8, lr @ *oq1 = s^0x80
- eor r10, r10, lr @ *op1 = s^0x80
+ sub r0, r0, r1, lsl #2
+ add r0, r0, #3
- ldrb r11,[r0, #-5] @ load p2 for 1/7th difference across boundary
- strb r10,[r0, #-4] @ store op1
- strb r8, [r0, #-1] @ store oq1
- ldrb_post r9, r0, r1 @ load q2 for 1/7th difference across boundary
+ ldrb r11, [r0, #-5] @ p2 for 1/7th difference
+ strb r10, [r0, #-4] @ op1
+ strb r8, [r0, #-1] @ oq1
+ ldrb_post r9, r0, r1 @ q2 for 1/7th difference
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
+ lsr r10, r10, #8
+ lsr r8, r8, #8
- ldrb r6, [r0, #-5]
- strb r10,[r0, #-4]
- strb r8, [r0, #-1]
+ ldrb r6, [r0, #-5]
+ strb r10, [r0, #-4]
+ strb r8, [r0, #-1]
ldrb_post r7, r0, r1
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
+ lsr r10, r10, #8
+ lsr r8, r8, #8
orr r11, r11, r6, lsl #8
orr r9, r9, r7, lsl #8
- ldrb r6, [r0, #-5]
- strb r10,[r0, #-4]
- strb r8, [r0, #-1]
+ ldrb r6, [r0, #-5]
+ strb r10, [r0, #-4]
+ strb r8, [r0, #-1]
ldrb_post r7, r0, r1
- mov r10, r10, lsr #8
- mov r8, r8, lsr #8
+ lsr r10, r10, #8
+ lsr r8, r8, #8
orr r11, r11, r6, lsl #16
orr r9, r9, r7, lsl #16
- ldrb r6, [r0, #-5]
- strb r10,[r0, #-4]
- strb r8, [r0, #-1]
+ ldrb r6, [r0, #-5]
+ strb r10, [r0, #-4]
+ strb r8, [r0, #-1]
ldrb_post r7, r0, r1
orr r11, r11, r6, lsl #24
orr r9, r9, r7, lsl #24
- @roughly 1/7th difference across boundary
- eor r9, r9, lr
- eor r11, r11, lr
-
- mov lr, #0x9 @ 9
- mov r7, #0x3f @ 63
-
- sxtb16 r6, r12
- sxtb16 r10, r12, ror #8
- smlabb r8, r6, lr, r7
- smlatb r6, r6, lr, r7
- smlabb r12, r10, lr, r7
- smlatb r10, r10, lr, r7
- ssat r8, #8, r8, asr #7
- ssat r6, #8, r6, asr #7
- ssat r12, #8, r12, asr #7
- ssat r10, #8, r10, asr #7
+ filter_3
sub r0, r0, r1, lsl #2
+ strb r10, [r0, #-5]
+ strb_post r8, r0, r1
+ lsr r10, r10, #8
+ lsr r8, r8, #8
+ strb r10, [r0, #-5]
+ strb_post r8, r0, r1
+ lsr r10, r10, #8
+ lsr r8, r8, #8
+ strb r10, [r0, #-5]
+ strb_post r8, r0, r1
+ lsr r10, r10, #8
+ lsr r8, r8, #8
+ strb r10, [r0, #-5]
+ strb_post r8, r0, r1
- pkhbt r6, r8, r6, lsl #16
- pkhbt r10, r12, r10, lsl #16
-
- uxtb16 r6, r6
- uxtb16 r10, r10
-
- ldr lr, c0x80808080
-
- orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7)
-
- qadd8 r8, r11, r10 @ s = vp8_signed_char_clamp(ps2 + u)
- qsub8 r10, r9, r10 @ s = vp8_signed_char_clamp(qs2 - u)
- eor r8, r8, lr @ *op2 = s^0x80
- eor r10, r10, lr @ *oq2 = s^0x80
-
- strb r8, [r0, #-5] @ store *op2
- strb_post r10, r0, r1 @ store *oq2
- mov r8, r8, lsr #8
- mov r10, r10, lsr #8
- strb r8, [r0, #-5]
- strb_post r10, r0, r1
- mov r8, r8, lsr #8
- mov r10, r10, lsr #8
- strb r8, [r0, #-5]
- strb_post r10, r0, r1
- mov r8, r8, lsr #8
- mov r10, r10, lsr #8
- strb r8, [r0, #-5]
- strb_post r10, r0, r1
-
- @adjust r0 pointer for next loop
sub r0, r0, #2
-
2:
- sub r0, r0, #4
+ cmp r5, #3
+ it eq
+ ldreq r0, [sp, #16]
subs r5, r5, #1
-
-T ittt ne
- ldrne r7, [r0, r1]
-A ldrne r6, [r0], r1, lsl #1 @ load source data
-T ldrne r6, [r0]
-T addne r0, r0, r1, lsl #1
-T ittt ne
- ldrne lr, [r0, r1]
-A ldrne r8, [r0], r1, lsl #1
-T ldrne r8, [r0]
-T addne r0, r0, r1, lsl #1
-
bne 1b
add sp, sp, #16
- pop {r4 - r11, pc}
+ pop {r0, r4-r11, pc}
+endfunc
+
+function ff_vp8_h_loop_filter16_armv6, export=1
+ push {r4-r11, lr}
+ add r12, r0, r1, lsl #3
+ push {r12}
+ ldr r9, [sp, #40]
+ orr r2, r2, r2, lsl #16
+ b vp8_h_loop_filter_armv6
+endfunc
+
+function ff_vp8_h_loop_filter8uv_armv6, export=1
+ push {r1, r4-r11, lr}
+ mov r1, r2
+ orr r2, r3, r3, lsl #16
+ ldr r3, [sp, #40]
+ ldr r9, [sp, #44]
+ b vp8_h_loop_filter_armv6
endfunc
+.ltorg
+
@ MC
@ void put_vp8_pixels16(uint8_t *dst, int dststride, uint8_t *src,
@ int srcstride, int h, int mx, int my)
function ff_put_vp8_pixels16_armv6, export=1
- push {r4 - r11}
- ldr r12,[sp, #32] @ h
+ push {r4-r11}
+ ldr r12, [sp, #32] @ h
1:
subs r12, r12, #2
- ldr r5, [r2, #4]
- ldr r6, [r2, #8]
- ldr r7, [r2, #12]
+ ldr r5, [r2, #4]
+ ldr r6, [r2, #8]
+ ldr r7, [r2, #12]
ldr_post r4, r2, r3
- ldr r9, [r2, #4]
- ldr r10,[r2, #8]
- ldr r11,[r2, #12]
+ ldr r9, [r2, #4]
+ ldr r10, [r2, #8]
+ ldr r11, [r2, #12]
ldr_post r8, r2, r3
- strd r6, r7, [r0, #8]
+ strd r6, r7, [r0, #8]
strd_post r4, r5, r0, r1
- strd r10, r11,[r0, #8]
+ strd r10, r11, [r0, #8]
strd_post r8, r9, r0, r1
bgt 1b
- pop {r4 - r11}
+ pop {r4-r11}
bx lr
endfunc
@ void put_vp8_pixels8(uint8_t *dst, int dststride, uint8_t *src,
@ int srcstride, int h, int mx, int my)
function ff_put_vp8_pixels8_armv6, export=1
- push {r4 - r11}
- ldr r12,[sp, #32] @ h
+ push {r4-r11}
+ ldr r12, [sp, #32] @ h
1:
subs r12, r12, #4
- ldr r5, [r2, #4]
+ ldr r5, [r2, #4]
ldr_post r4, r2, r3
- ldr r7, [r2, #4]
+ ldr r7, [r2, #4]
ldr_post r6, r2, r3
- ldr r9, [r2, #4]
+ ldr r9, [r2, #4]
ldr_post r8, r2, r3
- ldr r11,[r2, #4]
+ ldr r11, [r2, #4]
ldr_post r10, r2, r3
strd_post r4, r5, r0, r1
strd_post r6, r7, r0, r1
strd_post r8, r9, r0, r1
strd_post r10, r11, r0, r1
bgt 1b
- pop {r4 - r11}
+ pop {r4-r11}
bx lr
endfunc
@@ -1825,66 +1165,145 @@ endfunc
@ int srcstride, int h, int mx, int my)
function ff_put_vp8_pixels4_armv6, export=1
ldr r12, [sp, #0] @ h
- push {r4 - r6, lr}
+ push {r4-r6,lr}
1:
subs r12, r12, #4
- ldr r5, [r2, r3]
- ldr_post r4, r2, r3, lsl #1
- ldr lr, [r2, r3]
- ldr_post r6, r2, r3, lsl #1
- str r5, [r0, r1]
- str_post r4, r0, r1, lsl #1
- str lr, [r0, r1]
- str_post r6, r0, r1, lsl #1
+ ldr_post r4, r2, r3
+ ldr_post r5, r2, r3
+ ldr_post r6, r2, r3
+ ldr_post lr, r2, r3
+ str_post r4, r0, r1
+ str_post r5, r0, r1
+ str_post r6, r0, r1
+ str_post lr, r0, r1
bgt 1b
- pop {r4 - r6, pc}
+ pop {r4-r6,pc}
endfunc
@ note: worst case sum of all 6-tap filter values * 255 is 0x7f80 so 16 bit
@ arithmatic can be used to apply filters
-const sixtap_filters_13245600, align=4
+const sixtap_filters_13245600, align=4
.short 2, 108, -11, 36, -8, 1, 0, 0
.short 3, 77, -16, 77, -16, 3, 0, 0
.short 1, 36, -8, 108, -11, 2, 0, 0
endconst
-const fourtap_filters_1324, align=4
+
+const fourtap_filters_1324, align=4
.short -6, 12, 123, -1
.short -9, 50, 93, -6
.short -6, 93, 50, -9
.short -1, 123, 12, -6
endconst
-@ void put_vp8_epel_h6(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int mx)
-function ff_put_vp8_epel_h6_armv6, export=1
- push {r4 - r11, lr}
+.macro vp8_mc_1 name, size, hv
+function ff_put_vp8_\name\size\()_\hv\()_armv6, export=1
+ sub r1, r1, #\size
+ mov r12, sp
+ push {r1, r4-r11, lr}
+ ldm r12, {r5-r7}
+ mov r4, #\size
+ stm r12, {r4, r5}
+ orr r12, r6, r7
+ b vp8_put_\name\()_\hv\()_armv6 + 4
+endfunc
+.endm
+
+vp8_mc_1 epel, 16, h6
+vp8_mc_1 epel, 16, v6
+vp8_mc_1 epel, 8, h6
+vp8_mc_1 epel, 8, v6
+vp8_mc_1 epel, 8, h4
+vp8_mc_1 epel, 8, v4
+vp8_mc_1 epel, 4, h6
+vp8_mc_1 epel, 4, v6
+vp8_mc_1 epel, 4, h4
+vp8_mc_1 epel, 4, v4
+
+vp8_mc_1 bilin, 16, h
+vp8_mc_1 bilin, 16, v
+vp8_mc_1 bilin, 8, h
+vp8_mc_1 bilin, 8, v
+vp8_mc_1 bilin, 4, h
+vp8_mc_1 bilin, 4, v
+
+#define TMPSIZE \size * (8 - 8*(\size > 4) + \ytaps - 1)
+
+.macro vp8_mc_hv name, size, h, v, ytaps
+function ff_put_vp8_\name\size\()_\h\v\()_armv6, export=1
+ push {r0, r1, r4, lr}
+ add r0, sp, #16
+ sub sp, sp, #TMPSIZE+16
+ ldm r0, {r0, r12}
+ mov r4, #\size
+ add lr, r0, #\ytaps-1
+ .if \ytaps > 2
+ sub r2, r2, r3, lsl #\ytaps >> 1 & 1
+ .endif
+ stm sp, {r4, lr}
+ add r0, sp, #16
+ mov r1, #0
+ bl vp8_put_\name\()_\h\()_armv6
+ add r0, sp, #TMPSIZE+16
+ ldr lr, [sp, #TMPSIZE+16+16]
+ ldm r0, {r0, r1}
+ mov r3, #\size
+ ldr r12, [sp, #TMPSIZE+16+16+8]
+ str lr, [sp, #4]
+ add r2, sp, #16 + \size * (\ytaps / 2 - 1)
+ sub r1, r1, #\size
+ bl vp8_put_\name\()_\v\()_armv6
+ add sp, sp, #TMPSIZE+16+8
+ pop {r4, pc}
+endfunc
+.endm
+
+vp8_mc_hv epel, 16, h6, v6, 6
+vp8_mc_hv epel, 8, h6, v6, 6
+vp8_mc_hv epel, 8, h4, v6, 6
+vp8_mc_hv epel, 8, h6, v4, 4
+vp8_mc_hv epel, 8, h4, v4, 4
+vp8_mc_hv epel, 4, h6, v6, 6
+vp8_mc_hv epel, 4, h4, v6, 6
+vp8_mc_hv epel, 4, h6, v4, 4
+vp8_mc_hv epel, 4, h4, v4, 4
+
+vp8_mc_hv bilin, 16, h, v, 2
+vp8_mc_hv bilin, 8, h, v, 2
+vp8_mc_hv bilin, 4, h, v, 2
+
+.macro sat4 r0, r1, r2, r3
+ asr \r0, \r0, #7
+ asr \r1, \r1, #7
+ pkhbt \r0, \r0, \r2, lsl #9
+ pkhbt \r1, \r1, \r3, lsl #9
+ usat16 \r0, #8, \r0
+ usat16 \r1, #8, \r1
+ orr \r0, \r0, \r1, lsl #8
+.endm
+@ Calling convention for the inner MC functions:
+@ r0 dst
+@ r1 dst_stride - block_width
+@ r2 src
+@ r3 src_stride
+@ r4 block_width
+@ r12 filter_index
+@ [sp] block_width
+@ [sp+4] height
+@ [sp+8] scratch
+
+function vp8_put_epel_h6_armv6
+ push {r1, r4-r11, lr}
sub r2, r2, #2
movrel lr, sixtap_filters_13245600 - 16
- ldr r12,[sp, #44] @ vp8_filter index
- ldr r4, [sp, #36] @ width
add lr, lr, r12, lsl #3
- sub r3, r3, r4 @ src_stride - block_width
- sub r1, r1, r4 @ dst_stride - block_width
- lsr r4, #2
-
- str r4, [sp, #36] @ "4-in-parallel" loop counter @40
- str r3, [sp, #44] @ src_stride - block_width @48
- push {r1} @ dst_stride - block_width @0
- @ height @44
-
- ldr r1, [lr], #4 @ coefficients
- ldr r3, [lr], #4
- ldr lr, [lr]
+ sub r3, r3, r4
+ str r3, [sp, #48]
+ ldm lr, {r1, r3, lr}
1:
- @ 3 loads, 10 shuffles and then mul/acc/add/shr
- @ o0: i0/i1/i2/i3/i4/i5 -> i0/i2 (ld1) | i1/i3 (ld1) | i4/i5 (ld2)
- @ o1: i1/i2/i3/i4/i5/i6 -> i1/i3 (ld1) | i2/i4 (ld2) | i5/i6 (ld2/3)
- @ o2: i2/i3/i4/i5/i6/i7 -> i2/i4 (ld2) | i3/i5 (ld2) | i6/i7 (ld3)
- @ o3: i3/i4/i5/i6/i7/i8 -> i3/i5 (ld2) | i4/i6 (ld2/3) | i7/i8 (ld3)
- ldr r7, [r2, #5] @ ld3 -> src[5-8]
- ldr r6, [r2, #2] @ ld2 -> src[2-5]
- ldr r5, [r2], #4 @ ld1 -> src[0-3]
+ ldr r7, [r2, #5] @ src[5-8]
+ ldr r6, [r2, #2] @ src[2-5]
+ ldr r5, [r2], #4 @ src[0-3]
pkhtb r7, r7, r7, asr #8 @ src[8,7,7,6]
uxtb16 r9, r6, ror #8 @ src[5] | src[3]
@@ -1892,189 +1311,136 @@ function ff_put_vp8_epel_h6_armv6, export=1
uxtb16 r8, r5, ror #8 @ src[3] | src[1]
uxtb16 r11, r7, ror #8 @ src[8] | src[7]
uxtb16 r7, r7 @ src[7] | src[6]
- pkhtb r10, r9, r6, asr #16 @ src[5] | src[4]
uxtb16 r5, r5 @ src[2] | src[0]
- smuad r11, r11, lr @ filter[3][2] -> r11
- subs r4, r4, #1
- pkhbt r12, r10, r7, lsl #16 @ src[6] | src[4]
- smuad r7, r7, lr @ filter[2][2] -> r7
- smuad r5, r5, r1 @ filter[0][0] -> r5
- smlad r11, r9, r1, r11 @ filter[3][0] -> r11
- smlad r7, r9, r3, r7 @ filter[2][1] -> r7
- smuad r9, r8, r1 @ filter[1][0] -> r9
- smlad r5, r8, r3, r5 @ filter[0][1] -> r5
- pkhtb r8, r12, r10, asr #16 @ src[6] | src[5]
- smlad r11, r12, r3, r11 @ filter[3][1] -> r11
- smlad r9, r6, r3, r9 @ filter[1][1] -> r9
- smlad r5, r10, lr, r5 @ filter[0][2] -> r5
- smlad r7, r6, r1, r7 @ filter[2][0] -> r7
- smlad r9, r8, lr, r9 @ filter[1][2] -> r9
-
- add r5, r5, #0x40 @ round_shift_and_clamp[0]
- add r9, r9, #0x40 @ round_shift_and_clamp[1]
- add r7, r7, #0x40 @ round_shift_and_clamp[2]
- add r11, r11, #0x40 @ round_shift_and_clamp[3]
-
- usat r5, #8, r5, asr #7
- usat r9, #8, r9, asr #7
- usat r7, #8, r7, asr #7
- usat r11, #8, r11, asr #7
-
- strb r5, [r0], #1 @ store res[0]
- strb r9, [r0], #1 @ store res[1]
- strb r7, [r0], #1 @ store res[2]
- strb r11,[r0], #1 @ store res[3]
+ mov r10, #0x40
+ smlad r5, r5, r1, r10 @ filter[0][0]
+ smlad r11, r11, lr, r10 @ filter[3][2]
+ smlad r12, r7, lr, r10 @ filter[2][2]
+ smlad r10, r8, r1, r10 @ filter[1][0]
+ smlad r5, r8, r3, r5 @ filter[0][1]
+ smlad r11, r9, r1, r11 @ filter[3][0]
+ smlad r12, r9, r3, r12 @ filter[2][1]
+ pkhtb r9, r9, r6, asr #16 @ src[5] | src[4]
+ smlad r10, r6, r3, r10 @ filter[1][1]
+ pkhbt r7, r9, r7, lsl #16 @ src[6] | src[4]
+ smlad r5, r9, lr, r5 @ filter[0][2]
+ pkhtb r8, r7, r9, asr #16 @ src[6] | src[5]
+ smlad r11, r7, r3, r11 @ filter[3][1]
+ smlad r9, r8, lr, r10 @ filter[1][2]
+ smlad r7, r6, r1, r12 @ filter[2][0]
+
+ subs r4, r4, #4
+
+ sat4 r5, r9, r7, r11
+ str r5, [r0], #4
bne 1b
- ldr r12,[sp, #44] @ height = outer-loop counter
- subs r12, r12, #1
-T itttt ne
- ldrne r4, [sp, #40] @ 4-in-parallel loop counter
- ldrne r5, [sp, #48]
- ldrne r6, [sp]
- strne r12,[sp, #44]
- add r2, r2, r5 @ move to next input/output lines
+ add r4, sp, #40
+ ldm r4, {r4, r5, r12}
+ ldr r6, [sp]
+ subs r5, r5, #1
+ add r2, r2, r12
+ str r5, [sp, #44]
add r0, r0, r6
bne 1b
- add sp, sp, #4 @ restore stack after push{r1} above
- pop {r4 - r11, pc}
+ pop {r1, r4-r11, pc}
endfunc
-@ void put_vp8_epel_v6(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int my)
-function ff_put_vp8_epel_v6_armv6, export=1
- push {r4 - r11, lr}
-
+function vp8_put_epel_v6_armv6
+ push {r1, r4-r11, lr}
movrel lr, sixtap_filters_13245600 - 16
- ldr r12,[sp, #44] @ vp8_filter index
- ldr r4, [sp, #36] @ width
add lr, lr, r12, lsl #3
- sub r1, r1, r4 @ dst_stride - block_width
- lsr r4, #2
-
- str r4, [sp, #36] @ "4-in-parallel" loop counter @40
- str r3, [sp, #44] @ src_stride - block_width @48
- push {r1} @ dst_stride - block_width @0
- @ height @44
+ str r3, [sp, #48]
1:
add r1, r3, r3, lsl #1 @ stride * 3
- ldr_dpren r5, r2, r3 @ src[0,1,2,3 + stride * 1]
- ldr r6, [r2, r3] @ src[0,1,2,3 + stride * 3]
- ldr r7, [r2, r3, lsl #1] @ src[0,1,2,3 + stride * 4]
- ldr r8, [r2, r1] @ src[0,1,2,3 + stride * 5]
-
- @ byte -> word and "transpose"
- uxtb16 r9, r5, ror #8 @ src[3 + stride*1] | src[1 + stride*1]
- uxtb16 r10, r6, ror #8 @ src[3 + stride*3] | src[1 + stride*3]
- uxtb16 r11, r7, ror #8 @ src[3 + stride*4] | src[1 + stride*4]
- uxtb16 r12, r8, ror #8 @ src[3 + stride*5] | src[1 + stride*5]
- uxtb16 r5, r5 @ src[2 + stride*1] | src[0 + stride*1]
- uxtb16 r6, r6 @ src[2 + stride*3] | src[0 + stride*3]
- uxtb16 r7, r7 @ src[2 + stride*4] | src[0 + stride*4]
- uxtb16 r8, r8 @ src[2 + stride*5] | src[0 + stride*5]
- pkhbt r1, r9, r10, lsl #16 @ src[1 + stride*3] | src[1 + stride*1]
- pkhtb r9, r10, r9, asr #16 @ src[3 + stride*3] | src[3 + stride*1]
- pkhbt r10, r11, r12, lsl #16 @ src[1 + stride*5] | src[1 + stride*4]
- pkhtb r11, r12, r11, asr #16 @ src[3 + stride*5] | src[3 + stride*4]
- pkhbt r12, r5, r6, lsl #16 @ src[0 + stride*3] | src[0 + stride*1]
- pkhtb r5, r6, r5, asr #16 @ src[2 + stride*3] | src[2 + stride*1]
- pkhbt r6, r7, r8, lsl #16 @ src[0 + stride*5] | src[0 + stride*4]
- pkhtb r7, r8, r7, asr #16 @ src[2 + stride*5] | src[2 + stride*4]
-
- ldr r8, [lr, #4] @ stall - if only I had more registers...
- smuad r12, r12, r8 @ filter[0][1]
- smuad r1, r1, r8 @ filter[1][1]
- smuad r5, r5, r8 @ filter[2][1]
- smuad r9, r9, r8 @ filter[3][1]
- ldr r8, [lr, #8] @ stall - if only I had more registers...
- smlad r12, r6, r8, r12 @ filter[0][2]
- smlad r1, r10, r8, r1 @ filter[1][2]
- ldr_dpren r6, r2, r3, lsl #1 @ src[0,1,2,3 + stride * 0]
- ldr r10,[r2], #4 @ src[0,1,2,3 + stride * 2]
- smlad r5, r7, r8, r5 @ filter[2][2]
- smlad r9, r11, r8, r9 @ filter[3][2]
-
- uxtb16 r7, r6, ror #8 @ src[3 + stride*0] | src[1 + stride*0]
- uxtb16 r11, r10, ror #8 @ src[3 + stride*2] | src[1 + stride*2]
- uxtb16 r6, r6 @ src[2 + stride*0] | src[0 + stride*0]
- uxtb16 r10, r10 @ src[2 + stride*2] | src[0 + stride*2]
-
- pkhbt r8, r7, r11, lsl #16 @ src[1 + stride*2] | src[1 + stride*0]
- pkhtb r7, r11, r7, asr #16 @ src[3 + stride*2] | src[3 + stride*0]
- pkhbt r11, r6, r10, lsl #16 @ src[0 + stride*2] | src[0 + stride*0]
- pkhtb r6, r10, r6, asr #16 @ src[2 + stride*2] | src[2 + stride*0]
-
- ldr r10,[lr] @ stall - if only I had more registers...
- subs r4, r4, #1 @ counter--
+ ldr_nreg r5, r2, r3 @ src[0,1,2,3 + stride * 1]
+ ldr r6, [r2, r3] @ src[0,1,2,3 + stride * 3]
+ ldr r7, [r2, r3, lsl #1] @ src[0,1,2,3 + stride * 4]
+ ldr r8, [r2, r1] @ src[0,1,2,3 + stride * 5]
+
+ uxtb16 r9, r5, ror #8 @ src[3 + s*1] | src[1 + s*1]
+ uxtb16 r10, r6, ror #8 @ src[3 + s*3] | src[1 + s*3]
+ uxtb16 r11, r7, ror #8 @ src[3 + s*4] | src[1 + s*4]
+ uxtb16 r12, r8, ror #8 @ src[3 + s*5] | src[1 + s*5]
+ uxtb16 r5, r5 @ src[2 + s*1] | src[0 + s*1]
+ uxtb16 r6, r6 @ src[2 + s*3] | src[0 + s*3]
+ uxtb16 r7, r7 @ src[2 + s*4] | src[0 + s*4]
+ uxtb16 r8, r8 @ src[2 + s*5] | src[0 + s*5]
+ pkhbt r1, r9, r10, lsl #16 @ src[1 + s*3] | src[1 + s*1]
+ pkhtb r9, r10, r9, asr #16 @ src[3 + s*3] | src[3 + s*1]
+ pkhbt r10, r11, r12, lsl #16 @ src[1 + s*5] | src[1 + s*4]
+ pkhtb r11, r12, r11, asr #16 @ src[3 + s*5] | src[3 + s*4]
+ pkhbt r12, r5, r6, lsl #16 @ src[0 + s*3] | src[0 + s*1]
+ pkhtb r5, r6, r5, asr #16 @ src[2 + s*3] | src[2 + s*1]
+ pkhbt r6, r7, r8, lsl #16 @ src[0 + s*5] | src[0 + s*4]
+ pkhtb r7, r8, r7, asr #16 @ src[2 + s*5] | src[2 + s*4]
+
+ ldr r8, [lr, #4]
+ mov r3, #0x40
+ smlad r12, r12, r8, r3 @ filter[0][1]
+ smlad r1, r1, r8, r3 @ filter[1][1]
+ smlad r5, r5, r8, r3 @ filter[2][1]
+ smlad r9, r9, r8, r3 @ filter[3][1]
+ ldr r8, [lr, #8]
+ ldr r3, [sp, #48]
+ smlad r12, r6, r8, r12 @ filter[0][2]
+ smlad r1, r10, r8, r1 @ filter[1][2]
+ ldr_nreg r6, r2, r3, lsl #1 @ src[0,1,2,3 + stride * 0]
+ ldr r10, [r2], #4 @ src[0,1,2,3 + stride * 2]
+ smlad r5, r7, r8, r5 @ filter[2][2]
+ smlad r9, r11, r8, r9 @ filter[3][2]
+
+ uxtb16 r7, r6, ror #8 @ src[3 + s*0] | src[1 + s*0]
+ uxtb16 r11, r10, ror #8 @ src[3 + s*2] | src[1 + s*2]
+ uxtb16 r6, r6 @ src[2 + s*0] | src[0 + s*0]
+ uxtb16 r10, r10 @ src[2 + s*2] | src[0 + s*2]
+
+ pkhbt r8, r7, r11, lsl #16 @ src[1 + s*2] | src[1 + s*0]
+ pkhtb r7, r11, r7, asr #16 @ src[3 + s*2] | src[3 + s*0]
+ pkhbt r11, r6, r10, lsl #16 @ src[0 + s*2] | src[0 + s*0]
+ pkhtb r6, r10, r6, asr #16 @ src[2 + s*2] | src[2 + s*0]
+
+ ldr r10, [lr]
+ subs r4, r4, #4
smlad r12, r11, r10, r12 @ filter[0][0]
smlad r1, r8, r10, r1 @ filter[1][0]
smlad r5, r6, r10, r5 @ filter[2][0]
smlad r9, r7, r10, r9 @ filter[3][0]
- add r12, r12, #0x40 @ round_shift_and_clamp[0]
- add r1, r1, #0x40 @ round_shift_and_clamp[1]
- add r5, r5, #0x40 @ round_shift_and_clamp[2]
- add r9, r9, #0x40 @ round_shift_and_clamp[3]
-
- usat r12, #8, r12, asr #7
- usat r1, #8, r1, asr #7
- usat r5, #8, r5, asr #7
- usat r9, #8, r9, asr #7
-
- strb r12,[r0], #1 @ store res[0]
- strb r1, [r0], #1 @ store res[1]
- strb r5, [r0], #1 @ store res[2]
- strb r9, [r0], #1 @ store res[3]
+ sat4 r12, r1, r5, r9
+ str r12, [r0], #4
bne 1b
- ldr r12,[sp, #44] @ height = outer-loop counter
- subs r12, r12, #1
-T itttt ne
- ldrne r4, [sp, #40] @ 4-in-parallel loop counter
- ldrne r6, [sp, #0]
- subne r2, r2, r4, lsl #2
- strne r12,[sp, #44]
+ ldrd r4, r5, [sp, #40]
+ ldr r6, [sp]
+ subs r5, r5, #1
+ sub r2, r2, r4
+ str r5, [sp, #44]
add r0, r0, r6
- add r2, r2, r3 @ move to next input/output lines
+ add r2, r2, r3
bne 1b
- add sp, sp, #4 @ restore stack after push{r1} above
- pop {r4 - r11, pc}
+ pop {r1, r4-r11, pc}
endfunc
-@ void put_vp8_epel_h4(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int mx)
-function ff_put_vp8_epel_h4_armv6, export=1
- push {r4 - r11, lr}
-
+function vp8_put_epel_h4_armv6
+ push {r1, r4-r11, lr}
subs r2, r2, #1
movrel lr, fourtap_filters_1324 - 4
- ldr r4, [sp, #36] @ width
- ldr r12,[sp, #44] @ vp8_filter index
add lr, lr, r12, lsl #2
- sub r3, r3, r4 @ src_stride - block_width
- sub r1, r1, r4 @ dst_stride - block_width
- ldr r5, [lr]
- ldr r6, [lr, #4]
- asr r4, #2
-
- ldr lr, [sp, #40] @ height = outer-loop counter
- str r4, [sp, #36] @ "4-in-parallel" inner loop counter
+ sub r3, r3, r4
+ ldm lr, {r5, r6}
+ ldr lr, [sp, #44]
1:
- @ 3 loads, 5 uxtb16s and then mul/acc/add/shr
- @ o0: i0/i1/i2/i3 -> i0/i2(ld1) + i1/i3(ld1)
- @ o1: i1/i2/i3/i4 -> i1/i3(ld1) + i2/i4(ld2)
- @ o2: i2/i3/i4/i5 -> i2/i4(ld2) + i3/i5(ld2)
- @ o3: i3/i4/i5/i6 -> i3/i5(ld2) + i4/i6(ld3)
- ldr r9, [r2, #3] @ load source data
- ldr r8, [r2, #2]
- ldr r7, [r2], #4
+ ldr r9, [r2, #3]
+ ldr r8, [r2, #2]
+ ldr r7, [r2], #4
uxtb16 r9, r9, ror #8 @ src[6] | src[4]
uxtb16 r10, r8, ror #8 @ src[5] | src[3]
@@ -2082,247 +1448,181 @@ function ff_put_vp8_epel_h4_armv6, export=1
uxtb16 r11, r7, ror #8 @ src[3] | src[1]
uxtb16 r7, r7 @ src[2] | src[0]
- smuad r9, r9, r6 @ filter[3][1] -> r9
- smuad r12, r10, r6 @ filter[2][1] -> r12
- smuad r7, r7, r5 @ filter[0][0] -> r7
- smlad r9, r10, r5, r9 @ filter[3][0] -> r9
- smuad r10, r11, r5 @ filter[1][0] -> r10
- smlad r12, r8, r5, r12 @ filter[2][0] -> r12
- smlad r7, r11, r6, r7 @ filter[0][1] -> r7
- smlad r10, r8, r6, r10 @ filter[1][1] -> r10
-
- subs r4, r4, #1 @ counter--
+ mov r12, #0x40
+ smlad r9, r9, r6, r12 @ filter[3][1]
+ smlad r7, r7, r5, r12 @ filter[0][0]
+ smlad r9, r10, r5, r9 @ filter[3][0]
+ smlad r10, r10, r6, r12 @ filter[2][1]
+ smlad r12, r11, r5, r12 @ filter[1][0]
+ smlad r7, r11, r6, r7 @ filter[0][1]
+ smlad r10, r8, r5, r10 @ filter[2][0]
+ smlad r12, r8, r6, r12 @ filter[1][1]
- add r7, r7, #0x40 @ round_shift_and_clamp[0]
- add r10, r10, #0x40 @ round_shift_and_clamp[1]
- add r12, r12, #0x40 @ round_shift_and_clamp[2]
- add r9, r9, #0x40 @ round_shift_and_clamp[3]
+ subs r4, r4, #4
- usat r7, #8, r7, asr #7
- usat r10, #8, r10, asr #7
- usat r12, #8, r12, asr #7
- usat r9, #8, r9, asr #7
-
- strb r7, [r0], #1 @ store res[0]
- strb r10,[r0], #1 @ store res[1]
- strb r12,[r0], #1 @ store res[2]
- strb r9, [r0], #1 @ store res[3]
+ sat4 r7, r12, r10, r9
+ str r7, [r0], #4
bne 1b
subs lr, lr, #1
-T it ne
- ldrne r4, [sp, #36] @ 4-in-parallel loop counter
- add r2, r2, r3 @ move to next input/output lines
+ ldr r4, [sp, #40]
+ add r2, r2, r3
add r0, r0, r1
bne 1b
- pop {r4 - r11, pc}
+ pop {r1, r4-r11, pc}
endfunc
-@ void put_vp8_epel_v4(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int my)
-function ff_put_vp8_epel_v4_armv6, export=1
- push {r4 - r11, lr}
-
+function vp8_put_epel_v4_armv6
+ push {r1, r4-r11, lr}
movrel lr, fourtap_filters_1324 - 4
- ldr r12,[sp, #44] @ vp8_filter index
- ldr r4, [sp, #36] @ width
add lr, lr, r12, lsl #2
- sub r1, r1, r4 @ dst_stride - block_width
- asr r4, #2
- ldr r5, [lr]
- ldr r6, [lr, #4]
-
- str r4, [sp, #36] @ "4-in-parallel" loop counter @40
- str r3, [sp, #44] @ src_stride @48
- push {r1} @ dst_stride - block_width @36
- @ height @44
+ ldm lr, {r5, r6}
+ str r3, [sp, #48]
1:
- ldr lr, [r2, r3, lsl #1] @ load source pixels
- ldr r12,[r2, r3]
- ldr_dpren r7, r2, r3
- ldr r11,[r2], #4
-
- @ byte -> word and "transpose"
- uxtb16 r8, lr, ror #8 @ src[3 + stride*3] | src[1 + stride*3]
- uxtb16 r9, r12, ror #8 @ src[3 + stride*2] | src[1 + stride*2]
- uxtb16 r3, r7, ror #8 @ src[3 + stride*0] | src[1 + stride*0]
- uxtb16 r1, r11, ror #8 @ src[3 + stride*1] | src[1 + stride*1]
- uxtb16 lr, lr @ src[2 + stride*3] | src[0 + stride*3]
- uxtb16 r12, r12 @ src[2 + stride*2] | src[0 + stride*2]
- uxtb16 r7, r7 @ src[2 + stride*0] | src[0 + stride*0]
- uxtb16 r11, r11 @ src[2 + stride*1] | src[0 + stride*1]
- pkhbt r10, r1, r8, lsl #16 @ src[1 + stride*3] | src[1 + stride*1]
- pkhtb r1, r8, r1, asr #16 @ src[3 + stride*3] | src[3 + stride*1]
- pkhbt r8, r3, r9, lsl #16 @ src[1 + stride*2] | src[1 + stride*0]
- pkhtb r3, r9, r3, asr #16 @ src[3 + stride*2] | src[3 + stride*0]
- pkhbt r9, r11, lr, lsl #16 @ src[0 + stride*3] | src[0 + stride*1]
- pkhtb r11, lr, r11, asr #16 @ src[2 + stride*3] | src[2 + stride*1]
- pkhbt lr, r7, r12, lsl #16 @ src[0 + stride*2] | src[0 + stride*0]
- pkhtb r7, r12, r7, asr #16 @ src[2 + stride*2] | src[2 + stride*0]
-
- smuad r9, r9, r6 @ filter[0][1]
- smuad r10, r10, r6 @ filter[1][1]
- smuad r11, r11, r6 @ filter[2][1]
- smuad r1, r1, r6 @ filter[3][1]
- smlad r9, lr, r5, r9 @ filter[0][0]
- smlad r10, r8, r5, r10 @ filter[1][0]
- smlad r11, r7, r5, r11 @ filter[2][0]
- smlad r1, r3, r5, r1 @ filter[3][0]
-
- subs r4, r4, #1 @ counter--
- ldr r3, [sp, #48] @ FIXME prevent clobber of r3 above?
-
- add r9, r9, #0x40 @ round_shift_and_clamp[0]
- add r10, r10, #0x40 @ round_shift_and_clamp[1]
- add r11, r11, #0x40 @ round_shift_and_clamp[2]
- add r1, r1, #0x40 @ round_shift_and_clamp[3]
-
- usat r9, #8, r9, asr #7
- usat r10, #8, r10, asr #7
- usat r11, #8, r11, asr #7
- usat r1, #8, r1, asr #7
-
- strb r9, [r0], #1 @ store result
- strb r10,[r0], #1
- strb r11,[r0], #1
- strb r1, [r0], #1
+ ldr lr, [r2, r3, lsl #1]
+ ldr r12, [r2, r3]
+ ldr_nreg r7, r2, r3
+ ldr r11, [r2], #4
+
+ uxtb16 r8, lr, ror #8 @ src[3 + s*3] | src[1 + s*3]
+ uxtb16 r9, r12, ror #8 @ src[3 + s*2] | src[1 + s*2]
+ uxtb16 r3, r7, ror #8 @ src[3 + s*0] | src[1 + s*0]
+ uxtb16 r1, r11, ror #8 @ src[3 + s*1] | src[1 + s*1]
+ uxtb16 lr, lr @ src[2 + s*3] | src[0 + s*3]
+ uxtb16 r12, r12 @ src[2 + s*2] | src[0 + s*2]
+ uxtb16 r7, r7 @ src[2 + s*0] | src[0 + s*0]
+ uxtb16 r11, r11 @ src[2 + s*1] | src[0 + s*1]
+ pkhbt r10, r1, r8, lsl #16 @ src[1 + s*3] | src[1 + s*1]
+ pkhtb r1, r8, r1, asr #16 @ src[3 + s*3] | src[3 + s*1]
+ pkhbt r8, r3, r9, lsl #16 @ src[1 + s*2] | src[1 + s*0]
+ pkhtb r3, r9, r3, asr #16 @ src[3 + s*2] | src[3 + s*0]
+ pkhbt r9, r11, lr, lsl #16 @ src[0 + s*3] | src[0 + s*1]
+ pkhtb r11, lr, r11, asr #16 @ src[2 + s*3] | src[2 + s*1]
+ pkhbt lr, r7, r12, lsl #16 @ src[0 + s*2] | src[0 + s*0]
+ pkhtb r7, r12, r7, asr #16 @ src[2 + s*2] | src[2 + s*0]
+
+ mov r12, #0x40
+ smlad r9, r9, r6, r12 @ filter[0][1]
+ smlad r10, r10, r6, r12 @ filter[1][1]
+ smlad r11, r11, r6, r12 @ filter[2][1]
+ smlad r1, r1, r6, r12 @ filter[3][1]
+ smlad r9, lr, r5, r9 @ filter[0][0]
+ smlad r10, r8, r5, r10 @ filter[1][0]
+ smlad r11, r7, r5, r11 @ filter[2][0]
+ smlad r1, r3, r5, r1 @ filter[3][0]
+
+ subs r4, r4, #4
+ ldr r3, [sp, #48]
+
+ sat4 r9, r10, r11, r1
+ str r9, [r0], #4
bne 1b
- ldr r12,[sp, #44] @ height = outer-loop counter
+ ldr r4, [sp, #40]
+ ldr r12, [sp, #44]
+ add r2, r2, r3
+ ldr r9, [sp, #0]
subs r12, r12, #1
-T ittt ne
- ldrne r4, [sp, #40] @ 4-in-parallel loop counter
- ldrne r9, [sp, #0]
- strne r12,[sp, #44]
- sub r2, r2, r4, lsl #2
+ sub r2, r2, r4
+ str r12, [sp, #44]
add r0, r0, r9
- add r2, r2, r3 @ move to next input/output lines
bne 1b
- add sp, sp, #4 @ restore stack after push{r1} above
- pop {r4 - r11, pc}
+ pop {r1, r4-r11, pc}
endfunc
-@ void put_vp8_bilin_h(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int mx)
-function ff_put_vp8_bilin_h_armv6, export=1
- push {r4 - r9, lr}
-
- ldr r8, [sp, #36] @ vp8_filter index
- ldr r12,[sp, #32] @ height = outer-loop counter
- ldr r4, [sp, #28] @ width
- lsl r5, r8, #16 @ mx << 16
- sub r3, r3, r4 @ src_stride - block_width
- sub r1, r1, r4 @ dst_stride - block_width
- asr r4, #2
- sub r5, r5, r8 @ (mx << 16) | (-mx)
- str r4, [sp, #28] @ "4-in-parallel" loop counter
- add r5, r5, #8 @ (8 - mx) | (mx << 16) = filter coefficients
+function vp8_put_bilin_h_armv6
+ push {r1, r4-r11, lr}
+ rsb r5, r12, r12, lsl #16
+ ldr r12, [sp, #44]
+ sub r3, r3, r4
+ add r5, r5, #8
1:
- ldrb r6, [r2], #1 @ load source data
- ldrb r7, [r2], #1
- ldrb r8, [r2], #1
- ldrb r9, [r2], #1
- ldrb lr, [r2]
+ ldrb r6, [r2], #1
+ ldrb r7, [r2], #1
+ ldrb r8, [r2], #1
+ ldrb r9, [r2], #1
+ ldrb lr, [r2]
pkhbt r6, r6, r7, lsl #16 @ src[1] | src[0]
pkhbt r7, r7, r8, lsl #16 @ src[2] | src[1]
pkhbt r8, r8, r9, lsl #16 @ src[3] | src[2]
pkhbt r9, r9, lr, lsl #16 @ src[4] | src[3]
- smuad r6, r6, r5 @ apply the filter
- smuad r7, r7, r5
- smuad r8, r8, r5
- smuad r9, r9, r5
+ mov r10, #4
+ smlad r6, r6, r5, r10
+ smlad r7, r7, r5, r10
+ smlad r8, r8, r5, r10
+ smlad r9, r9, r5, r10
- subs r4, r4, #1 @ counter--
-
- add r6, r6, #0x4 @ round_shift_and_clamp
- add r7, r7, #0x4
- add r8, r8, #0x4
- add r9, r9, #0x4
+ subs r4, r4, #4
asr r6, #3
asr r7, #3
pkhbt r6, r6, r8, lsl #13
pkhbt r7, r7, r9, lsl #13
orr r6, r6, r7, lsl #8
- str r6, [r0], #4 @ store result
+ str r6, [r0], #4
bne 1b
- ldr r4, [sp, #28] @ 4-in-parallel loop counter
+ ldr r4, [sp, #40]
subs r12, r12, #1
-
- add r2, r2, r3 @ move to next input/output lines
+ add r2, r2, r3
add r0, r0, r1
bne 1b
- pop {r4 - r9, pc}
+ pop {r1, r4-r11, pc}
endfunc
-@ void put_vp8_bilin_v(uint8_t *dst, int dststride, uint8_t *src,
-@ int srcstride, int w, int h, int my)
-function ff_put_vp8_bilin_v_armv6, export=1
- push {r4 - r11, lr}
-
- ldr r11,[sp, #44] @ vp8_filter index
- ldr r4, [sp, #36] @ width
- mov r5, r11, lsl #16 @ mx << 16
- ldr r12,[sp, #40] @ height = outer-loop counter
- sub r1, r1, r4
- sub r5, r5, r11 @ (mx << 16) | (-mx)
- asr r4, #2
- add r5, r5, #8 @ (8 - mx) | (mx << 16) = filter coefficients
- str r4, [sp, #36] @ "4-in-parallel" loop counter
+function vp8_put_bilin_v_armv6
+ push {r1, r4-r11, lr}
+ rsb r5, r12, r12, lsl #16
+ ldr r12, [sp, #44]
+ add r5, r5, #8
1:
- ldrb r10,[r2, r3] @ load the data
- ldrb r6, [r2], #1
- ldrb r11,[r2, r3]
- ldrb r7, [r2], #1
- ldrb lr, [r2, r3]
- ldrb r8, [r2], #1
- ldrb r9, [r2, r3]
+ ldrb r10, [r2, r3]
+ ldrb r6, [r2], #1
+ ldrb r11, [r2, r3]
+ ldrb r7, [r2], #1
+ ldrb lr, [r2, r3]
+ ldrb r8, [r2], #1
+ ldrb r9, [r2, r3]
pkhbt r6, r6, r10, lsl #16
- ldrb r10,[r2], #1
+ ldrb r10, [r2], #1
pkhbt r7, r7, r11, lsl #16
pkhbt r8, r8, lr, lsl #16
pkhbt r9, r10, r9, lsl #16
- smuad r6, r6, r5 @ apply the filter
- smuad r7, r7, r5
- smuad r8, r8, r5
- smuad r9, r9, r5
+ mov r10, #4
+ smlad r6, r6, r5, r10
+ smlad r7, r7, r5, r10
+ smlad r8, r8, r5, r10
+ smlad r9, r9, r5, r10
- subs r4, r4, #1 @ counter--
-
- add r6, r6, #0x4 @ round_shift_and_clamp
- add r7, r7, #0x4
- add r8, r8, #0x4
- add r9, r9, #0x4
+ subs r4, r4, #4
asr r6, #3
asr r7, #3
pkhbt r6, r6, r8, lsl #13
pkhbt r7, r7, r9, lsl #13
orr r6, r6, r7, lsl #8
- str r6, [r0], #4 @ store result
+ str r6, [r0], #4
bne 1b
- ldr r4, [sp, #36] @ 4-in-parallel loop counter
+ ldr r4, [sp, #40]
subs r12, r12, #1
-
- add r2, r2, r3 @ move to next input/output lines
+ add r2, r2, r3
add r0, r0, r1
- sub r2, r2, r4, lsl #2
+ sub r2, r2, r4
bne 1b
- pop {r4 - r11, pc}
+ pop {r1, r4-r11, pc}
endfunc
diff --git a/libavcodec/arm/vp8dsp_init_arm.c b/libavcodec/arm/vp8dsp_init_arm.c
index 2b105dbe1d..5c84ed3231 100644
--- a/libavcodec/arm/vp8dsp_init_arm.c
+++ b/libavcodec/arm/vp8dsp_init_arm.c
@@ -20,294 +20,14 @@
#include "libavutil/arm/cpu.h"
#include "libavcodec/vp8dsp.h"
-
-void ff_vp8_luma_dc_wht_dc_armv6(DCTELEM block[4][4][16], DCTELEM dc[16]);
-
-#define idct_funcs(opt) \
-void ff_vp8_luma_dc_wht_ ## opt(DCTELEM block[4][4][16], DCTELEM dc[16]); \
-void ff_vp8_idct_add_ ## opt(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride); \
-void ff_vp8_idct_dc_add_ ## opt(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride); \
-void ff_vp8_idct_dc_add4y_ ## opt(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride); \
-void ff_vp8_idct_dc_add4uv_ ## opt(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride)
-
-idct_funcs(neon);
-idct_funcs(armv6);
-
-void ff_vp8_v_loop_filter16_neon(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-void ff_vp8_h_loop_filter16_neon(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-void ff_vp8_v_loop_filter8uv_neon(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-void ff_vp8_h_loop_filter8uv_neon(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-
-void ff_vp8_v_loop_filter16_inner_neon(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-void ff_vp8_h_loop_filter16_inner_neon(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh);
-void ff_vp8_v_loop_filter8uv_inner_neon(uint8_t *dstU, uint8_t *dstV,
- ptrdiff_t stride, int flim_E, int flim_I,
- int hev_thresh);
-void ff_vp8_h_loop_filter8uv_inner_neon(uint8_t *dstU, uint8_t *dstV,
- ptrdiff_t stride, int flim_E, int flim_I,
- int hev_thresh);
-
-void ff_vp8_v_loop_filter_inner_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I,
- int hev_thresh, int count);
-void ff_vp8_h_loop_filter_inner_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I,
- int hev_thresh, int count);
-void ff_vp8_v_loop_filter_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I,
- int hev_thresh, int count);
-void ff_vp8_h_loop_filter_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I,
- int hev_thresh, int count);
-
-static void ff_vp8_v_loop_filter16_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_v_loop_filter_armv6(dst, stride, flim_E, flim_I, hev_thresh, 4);
-}
-
-static void ff_vp8_h_loop_filter16_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_h_loop_filter_armv6(dst, stride, flim_E, flim_I, hev_thresh, 4);
-}
-
-static void ff_vp8_v_loop_filter8uv_armv6(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_v_loop_filter_armv6(dstU, stride, flim_E, flim_I, hev_thresh, 2);
- ff_vp8_v_loop_filter_armv6(dstV, stride, flim_E, flim_I, hev_thresh, 2);
-}
-
-static void ff_vp8_h_loop_filter8uv_armv6(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_h_loop_filter_armv6(dstU, stride, flim_E, flim_I, hev_thresh, 2);
- ff_vp8_h_loop_filter_armv6(dstV, stride, flim_E, flim_I, hev_thresh, 2);
-}
-
-static void ff_vp8_v_loop_filter16_inner_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_v_loop_filter_inner_armv6(dst, stride, flim_E, flim_I, hev_thresh, 4);
-}
-
-static void ff_vp8_h_loop_filter16_inner_armv6(uint8_t *dst, ptrdiff_t stride,
- int flim_E, int flim_I, int hev_thresh)
-{
- ff_vp8_h_loop_filter_inner_armv6(dst, stride, flim_E, flim_I, hev_thresh, 4);
-}
-
-static void ff_vp8_v_loop_filter8uv_inner_armv6(uint8_t *dstU, uint8_t *dstV,
- ptrdiff_t stride, int flim_E, int flim_I,
- int hev_thresh)
-{
- ff_vp8_v_loop_filter_inner_armv6(dstU, stride, flim_E, flim_I, hev_thresh, 2);
- ff_vp8_v_loop_filter_inner_armv6(dstV, stride, flim_E, flim_I, hev_thresh, 2);
-}
-
-static void ff_vp8_h_loop_filter8uv_inner_armv6(uint8_t *dstU, uint8_t *dstV,
- ptrdiff_t stride, int flim_E, int flim_I,
- int hev_thresh)
-{
- ff_vp8_h_loop_filter_inner_armv6(dstU, stride, flim_E, flim_I, hev_thresh, 2);
- ff_vp8_h_loop_filter_inner_armv6(dstV, stride, flim_E, flim_I, hev_thresh, 2);
-}
-
-#define simple_lf_funcs(opt) \
-void ff_vp8_v_loop_filter16_simple_ ## opt(uint8_t *dst, ptrdiff_t stride, int flim); \
-void ff_vp8_h_loop_filter16_simple_ ## opt(uint8_t *dst, ptrdiff_t stride, int flim)
-
-simple_lf_funcs(neon);
-simple_lf_funcs(armv6);
-
-#define VP8_MC_OPT(n, opt) \
- void ff_put_vp8_##n##_##opt(uint8_t *dst, ptrdiff_t dststride, \
- uint8_t *src, ptrdiff_t srcstride, \
- int h, int x, int y)
-
-#define VP8_MC(n) \
- VP8_MC_OPT(n, neon)
-
-#define VP8_EPEL(w) \
- VP8_MC(epel ## w ## _h4); \
- VP8_MC(epel ## w ## _h6); \
- VP8_MC(epel ## w ## _h4v4); \
- VP8_MC(epel ## w ## _h6v4); \
- VP8_MC(epel ## w ## _v4); \
- VP8_MC(epel ## w ## _v6); \
- VP8_MC(epel ## w ## _h4v6); \
- VP8_MC(epel ## w ## _h6v6)
-
-VP8_EPEL(16);
-VP8_MC(pixels16);
-VP8_MC_OPT(pixels16, armv6);
-VP8_EPEL(8);
-VP8_MC(pixels8);
-VP8_MC_OPT(pixels8, armv6);
-VP8_EPEL(4);
-VP8_MC_OPT(pixels4, armv6);
-
-VP8_MC(bilin16_h);
-VP8_MC(bilin16_v);
-VP8_MC(bilin16_hv);
-VP8_MC(bilin8_h);
-VP8_MC(bilin8_v);
-VP8_MC(bilin8_hv);
-VP8_MC(bilin4_h);
-VP8_MC(bilin4_v);
-VP8_MC(bilin4_hv);
-
-#define VP8_V6_MC(n) \
-void ff_put_vp8_##n##_armv6(uint8_t *dst, int dststride, uint8_t *src, \
- int srcstride, int w, int h, int mxy)
-
-VP8_V6_MC(epel_v6);
-VP8_V6_MC(epel_h6);
-VP8_V6_MC(epel_v4);
-VP8_V6_MC(epel_h4);
-VP8_V6_MC(bilin_v);
-VP8_V6_MC(bilin_h);
-
-#define VP8_EPEL_HV(SIZE, TAPNUMX, TAPNUMY, NAME, HNAME, VNAME, MAXHEIGHT) \
-static void ff_put_vp8_##NAME##SIZE##_##HNAME##VNAME##_armv6( \
- uint8_t *dst, int dststride, uint8_t *src, \
- int srcstride, int h, int mx, int my) \
-{ \
- DECLARE_ALIGNED(4, uint8_t, tmp)[SIZE * (MAXHEIGHT + TAPNUMY - 1)]; \
- uint8_t *tmpptr = tmp + SIZE * (TAPNUMY / 2 - 1); \
- src -= srcstride * (TAPNUMY / 2 - 1); \
- ff_put_vp8_ ## NAME ## _ ## HNAME ## _armv6(tmp, SIZE, src, srcstride, \
- SIZE, h + TAPNUMY - 1, mx); \
- ff_put_vp8_ ## NAME ## _ ## VNAME ## _armv6(dst, dststride, tmpptr, SIZE, \
- SIZE, h, my); \
-}
-
-VP8_EPEL_HV(16, 6, 6, epel, h6, v6, 16);
-VP8_EPEL_HV(16, 2, 2, bilin, h, v, 16);
-VP8_EPEL_HV(8, 6, 6, epel, h6, v6, 16);
-VP8_EPEL_HV(8, 4, 6, epel, h4, v6, 16);
-VP8_EPEL_HV(8, 6, 4, epel, h6, v4, 16);
-VP8_EPEL_HV(8, 4, 4, epel, h4, v4, 16);
-VP8_EPEL_HV(8, 2, 2, bilin, h, v, 16);
-VP8_EPEL_HV(4, 6, 6, epel, h6, v6, 8);
-VP8_EPEL_HV(4, 4, 6, epel, h4, v6, 8);
-VP8_EPEL_HV(4, 6, 4, epel, h6, v4, 8);
-VP8_EPEL_HV(4, 4, 4, epel, h4, v4, 8);
-VP8_EPEL_HV(4, 2, 2, bilin, h, v, 8);
-
-extern void put_vp8_epel4_v6_c(uint8_t *dst, int d, uint8_t *src, int s, int h, int mx, int my);
-#undef printf
-#define VP8_EPEL_H_OR_V(SIZE, NAME, HV) \
-static void ff_put_vp8_##NAME##SIZE##_##HV##_armv6( \
- uint8_t *dst, int dststride, uint8_t *src, \
- int srcstride, int h, int mx, int my) \
-{ \
- ff_put_vp8_## NAME ## _ ## HV ## _armv6(dst, dststride, src, srcstride, \
- SIZE, h, mx | my); \
-}
-
-VP8_EPEL_H_OR_V(4, epel, h6);
-VP8_EPEL_H_OR_V(4, epel, h4);
-VP8_EPEL_H_OR_V(4, epel, v6);
-VP8_EPEL_H_OR_V(4, epel, v4);
-VP8_EPEL_H_OR_V(4, bilin, v);
-VP8_EPEL_H_OR_V(4, bilin, h);
-VP8_EPEL_H_OR_V(8, epel, h6);
-VP8_EPEL_H_OR_V(8, epel, h4);
-VP8_EPEL_H_OR_V(8, epel, v6);
-VP8_EPEL_H_OR_V(8, epel, v4);
-VP8_EPEL_H_OR_V(8, bilin, v);
-VP8_EPEL_H_OR_V(8, bilin, h);
-VP8_EPEL_H_OR_V(16, epel, h6);
-VP8_EPEL_H_OR_V(16, epel, v6);
-VP8_EPEL_H_OR_V(16, bilin, v);
-VP8_EPEL_H_OR_V(16, bilin, h);
+#include "vp8dsp.h"
av_cold void ff_vp8dsp_init_arm(VP8DSPContext *dsp)
{
-#define set_func_ptrs(opt) \
- dsp->vp8_luma_dc_wht = ff_vp8_luma_dc_wht_##opt; \
- dsp->vp8_luma_dc_wht_dc = ff_vp8_luma_dc_wht_dc_armv6; \
- \
- dsp->vp8_idct_add = ff_vp8_idct_add_##opt; \
- dsp->vp8_idct_dc_add = ff_vp8_idct_dc_add_##opt; \
- dsp->vp8_idct_dc_add4y = ff_vp8_idct_dc_add4y_##opt; \
- dsp->vp8_idct_dc_add4uv = ff_vp8_idct_dc_add4uv_##opt; \
- \
- dsp->vp8_v_loop_filter16y = ff_vp8_v_loop_filter16_##opt; \
- dsp->vp8_h_loop_filter16y = ff_vp8_h_loop_filter16_##opt; \
- dsp->vp8_v_loop_filter8uv = ff_vp8_v_loop_filter8uv_##opt; \
- dsp->vp8_h_loop_filter8uv = ff_vp8_h_loop_filter8uv_##opt; \
- \
- dsp->vp8_v_loop_filter16y_inner = ff_vp8_v_loop_filter16_inner_##opt; \
- dsp->vp8_h_loop_filter16y_inner = ff_vp8_h_loop_filter16_inner_##opt; \
- dsp->vp8_v_loop_filter8uv_inner = ff_vp8_v_loop_filter8uv_inner_##opt; \
- dsp->vp8_h_loop_filter8uv_inner = ff_vp8_h_loop_filter8uv_inner_##opt; \
- \
- dsp->vp8_v_loop_filter_simple = ff_vp8_v_loop_filter16_simple_##opt; \
- dsp->vp8_h_loop_filter_simple = ff_vp8_h_loop_filter16_simple_##opt; \
- \
- dsp->put_vp8_epel_pixels_tab[0][0][0] = ff_put_vp8_pixels16_##opt; \
- dsp->put_vp8_epel_pixels_tab[0][0][2] = ff_put_vp8_epel16_h6_##opt; \
- dsp->put_vp8_epel_pixels_tab[0][2][0] = ff_put_vp8_epel16_v6_##opt; \
- dsp->put_vp8_epel_pixels_tab[0][2][2] = ff_put_vp8_epel16_h6v6_##opt; \
- \
- dsp->put_vp8_epel_pixels_tab[1][0][0] = ff_put_vp8_pixels8_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][0][1] = ff_put_vp8_epel8_h4_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][0][2] = ff_put_vp8_epel8_h6_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][1][0] = ff_put_vp8_epel8_v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][1][1] = ff_put_vp8_epel8_h4v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][1][2] = ff_put_vp8_epel8_h6v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][2][0] = ff_put_vp8_epel8_v6_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][2][1] = ff_put_vp8_epel8_h4v6_##opt; \
- dsp->put_vp8_epel_pixels_tab[1][2][2] = ff_put_vp8_epel8_h6v6_##opt; \
- \
- dsp->put_vp8_epel_pixels_tab[2][0][0] = ff_put_vp8_pixels4_armv6; \
- dsp->put_vp8_epel_pixels_tab[2][0][1] = ff_put_vp8_epel4_h4_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][0][2] = ff_put_vp8_epel4_h6_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][1][0] = ff_put_vp8_epel4_v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][1][1] = ff_put_vp8_epel4_h4v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][1][2] = ff_put_vp8_epel4_h6v4_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][2][0] = ff_put_vp8_epel4_v6_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][2][1] = ff_put_vp8_epel4_h4v6_##opt; \
- dsp->put_vp8_epel_pixels_tab[2][2][2] = ff_put_vp8_epel4_h6v6_##opt; \
- \
- dsp->put_vp8_bilinear_pixels_tab[0][0][0] = ff_put_vp8_pixels16_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[0][0][2] = ff_put_vp8_bilin16_h_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[0][2][0] = ff_put_vp8_bilin16_v_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[0][2][2] = ff_put_vp8_bilin16_hv_##opt; \
- \
- dsp->put_vp8_bilinear_pixels_tab[1][0][0] = ff_put_vp8_pixels8_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][0][1] = ff_put_vp8_bilin8_h_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][0][2] = ff_put_vp8_bilin8_h_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][1][0] = ff_put_vp8_bilin8_v_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][1][1] = ff_put_vp8_bilin8_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][1][2] = ff_put_vp8_bilin8_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][2][0] = ff_put_vp8_bilin8_v_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][2][1] = ff_put_vp8_bilin8_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[1][2][2] = ff_put_vp8_bilin8_hv_##opt; \
- \
- dsp->put_vp8_bilinear_pixels_tab[2][0][0] = ff_put_vp8_pixels4_armv6; \
- dsp->put_vp8_bilinear_pixels_tab[2][0][1] = ff_put_vp8_bilin4_h_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][0][2] = ff_put_vp8_bilin4_h_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][1][0] = ff_put_vp8_bilin4_v_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][1][1] = ff_put_vp8_bilin4_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][1][2] = ff_put_vp8_bilin4_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][2][0] = ff_put_vp8_bilin4_v_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][2][1] = ff_put_vp8_bilin4_hv_##opt; \
- dsp->put_vp8_bilinear_pixels_tab[2][2][2] = ff_put_vp8_bilin4_hv_##opt
int cpu_flags = av_get_cpu_flags();
- if (have_neon(cpu_flags)) {
- set_func_ptrs(neon);
- } else if (HAVE_ARMV6) {
- set_func_ptrs(armv6);
- }
+ if (have_armv6(cpu_flags))
+ ff_vp8dsp_init_armv6(dsp);
+ if (have_neon(cpu_flags))
+ ff_vp8dsp_init_neon(dsp);
}
diff --git a/libavcodec/arm/vp8dsp_init_armv6.c b/libavcodec/arm/vp8dsp_init_armv6.c
new file mode 100644
index 0000000000..85a803af83
--- /dev/null
+++ b/libavcodec/arm/vp8dsp_init_armv6.c
@@ -0,0 +1,118 @@
+/*
+ * This file is part of Libav.
+ *
+ * Libav is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * Libav is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with Libav; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#include <stdint.h>
+#include "libavcodec/vp8dsp.h"
+#include "vp8dsp.h"
+
+void ff_vp8_luma_dc_wht_armv6(DCTELEM block[4][4][16], DCTELEM dc[16]);
+void ff_vp8_luma_dc_wht_dc_armv6(DCTELEM block[4][4][16], DCTELEM dc[16]);
+
+void ff_vp8_idct_add_armv6(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add_armv6(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add4y_armv6(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add4uv_armv6(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride);
+
+VP8_LF(armv6);
+
+VP8_EPEL(16, armv6);
+VP8_EPEL(8, armv6);
+VP8_EPEL(4, armv6);
+
+VP8_BILIN(16, armv6);
+VP8_BILIN(8, armv6);
+VP8_BILIN(4, armv6);
+
+av_cold void ff_vp8dsp_init_armv6(VP8DSPContext *dsp)
+{
+ dsp->vp8_luma_dc_wht = ff_vp8_luma_dc_wht_armv6;
+ dsp->vp8_luma_dc_wht_dc = ff_vp8_luma_dc_wht_dc_armv6;
+
+ dsp->vp8_idct_add = ff_vp8_idct_add_armv6;
+ dsp->vp8_idct_dc_add = ff_vp8_idct_dc_add_armv6;
+ dsp->vp8_idct_dc_add4y = ff_vp8_idct_dc_add4y_armv6;
+ dsp->vp8_idct_dc_add4uv = ff_vp8_idct_dc_add4uv_armv6;
+
+ dsp->vp8_v_loop_filter16y = ff_vp8_v_loop_filter16_armv6;
+ dsp->vp8_h_loop_filter16y = ff_vp8_h_loop_filter16_armv6;
+ dsp->vp8_v_loop_filter8uv = ff_vp8_v_loop_filter8uv_armv6;
+ dsp->vp8_h_loop_filter8uv = ff_vp8_h_loop_filter8uv_armv6;
+
+ dsp->vp8_v_loop_filter16y_inner = ff_vp8_v_loop_filter16_inner_armv6;
+ dsp->vp8_h_loop_filter16y_inner = ff_vp8_h_loop_filter16_inner_armv6;
+ dsp->vp8_v_loop_filter8uv_inner = ff_vp8_v_loop_filter8uv_inner_armv6;
+ dsp->vp8_h_loop_filter8uv_inner = ff_vp8_h_loop_filter8uv_inner_armv6;
+
+ dsp->vp8_v_loop_filter_simple = ff_vp8_v_loop_filter16_simple_armv6;
+ dsp->vp8_h_loop_filter_simple = ff_vp8_h_loop_filter16_simple_armv6;
+
+ dsp->put_vp8_epel_pixels_tab[0][0][0] = ff_put_vp8_pixels16_armv6;
+ dsp->put_vp8_epel_pixels_tab[0][0][2] = ff_put_vp8_epel16_h6_armv6;
+ dsp->put_vp8_epel_pixels_tab[0][2][0] = ff_put_vp8_epel16_v6_armv6;
+ dsp->put_vp8_epel_pixels_tab[0][2][2] = ff_put_vp8_epel16_h6v6_armv6;
+
+ dsp->put_vp8_epel_pixels_tab[1][0][0] = ff_put_vp8_pixels8_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][0][1] = ff_put_vp8_epel8_h4_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][0][2] = ff_put_vp8_epel8_h6_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][1][0] = ff_put_vp8_epel8_v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][1][1] = ff_put_vp8_epel8_h4v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][1][2] = ff_put_vp8_epel8_h6v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][2][0] = ff_put_vp8_epel8_v6_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][2][1] = ff_put_vp8_epel8_h4v6_armv6;
+ dsp->put_vp8_epel_pixels_tab[1][2][2] = ff_put_vp8_epel8_h6v6_armv6;
+
+ dsp->put_vp8_epel_pixels_tab[2][0][0] = ff_put_vp8_pixels4_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][0][1] = ff_put_vp8_epel4_h4_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][0][2] = ff_put_vp8_epel4_h6_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][1][0] = ff_put_vp8_epel4_v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][1][1] = ff_put_vp8_epel4_h4v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][1][2] = ff_put_vp8_epel4_h6v4_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][2][0] = ff_put_vp8_epel4_v6_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][2][1] = ff_put_vp8_epel4_h4v6_armv6;
+ dsp->put_vp8_epel_pixels_tab[2][2][2] = ff_put_vp8_epel4_h6v6_armv6;
+
+ dsp->put_vp8_bilinear_pixels_tab[0][0][0] = ff_put_vp8_pixels16_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][0][1] = ff_put_vp8_bilin16_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][0][2] = ff_put_vp8_bilin16_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][0] = ff_put_vp8_bilin16_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][1] = ff_put_vp8_bilin16_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][2] = ff_put_vp8_bilin16_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][0] = ff_put_vp8_bilin16_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][1] = ff_put_vp8_bilin16_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][2] = ff_put_vp8_bilin16_hv_armv6;
+
+ dsp->put_vp8_bilinear_pixels_tab[1][0][0] = ff_put_vp8_pixels8_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][0][1] = ff_put_vp8_bilin8_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][0][2] = ff_put_vp8_bilin8_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][0] = ff_put_vp8_bilin8_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][1] = ff_put_vp8_bilin8_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][2] = ff_put_vp8_bilin8_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][0] = ff_put_vp8_bilin8_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][1] = ff_put_vp8_bilin8_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][2] = ff_put_vp8_bilin8_hv_armv6;
+
+ dsp->put_vp8_bilinear_pixels_tab[2][0][0] = ff_put_vp8_pixels4_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][0][1] = ff_put_vp8_bilin4_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][0][2] = ff_put_vp8_bilin4_h_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][0] = ff_put_vp8_bilin4_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][1] = ff_put_vp8_bilin4_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][2] = ff_put_vp8_bilin4_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][0] = ff_put_vp8_bilin4_v_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][1] = ff_put_vp8_bilin4_hv_armv6;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][2] = ff_put_vp8_bilin4_hv_armv6;
+}
diff --git a/libavcodec/arm/vp8dsp_init_neon.c b/libavcodec/arm/vp8dsp_init_neon.c
new file mode 100644
index 0000000000..dbe5b9f961
--- /dev/null
+++ b/libavcodec/arm/vp8dsp_init_neon.c
@@ -0,0 +1,114 @@
+/*
+ * This file is part of Libav.
+ *
+ * Libav is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * Libav is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with Libav; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#include <stdint.h>
+#include "libavcodec/vp8dsp.h"
+#include "vp8dsp.h"
+
+void ff_vp8_luma_dc_wht_neon(DCTELEM block[4][4][16], DCTELEM dc[16]);
+
+void ff_vp8_idct_add_neon(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add_neon(uint8_t *dst, DCTELEM block[16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add4y_neon(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride);
+void ff_vp8_idct_dc_add4uv_neon(uint8_t *dst, DCTELEM block[4][16], ptrdiff_t stride);
+
+VP8_LF(neon);
+
+VP8_EPEL(16, neon);
+VP8_EPEL(8, neon);
+VP8_EPEL(4, neon);
+
+VP8_BILIN(16, neon);
+VP8_BILIN(8, neon);
+VP8_BILIN(4, neon);
+
+av_cold void ff_vp8dsp_init_neon(VP8DSPContext *dsp)
+{
+ dsp->vp8_luma_dc_wht = ff_vp8_luma_dc_wht_neon;
+
+ dsp->vp8_idct_add = ff_vp8_idct_add_neon;
+ dsp->vp8_idct_dc_add = ff_vp8_idct_dc_add_neon;
+ dsp->vp8_idct_dc_add4y = ff_vp8_idct_dc_add4y_neon;
+ dsp->vp8_idct_dc_add4uv = ff_vp8_idct_dc_add4uv_neon;
+
+ dsp->vp8_v_loop_filter16y = ff_vp8_v_loop_filter16_neon;
+ dsp->vp8_h_loop_filter16y = ff_vp8_h_loop_filter16_neon;
+ dsp->vp8_v_loop_filter8uv = ff_vp8_v_loop_filter8uv_neon;
+ dsp->vp8_h_loop_filter8uv = ff_vp8_h_loop_filter8uv_neon;
+
+ dsp->vp8_v_loop_filter16y_inner = ff_vp8_v_loop_filter16_inner_neon;
+ dsp->vp8_h_loop_filter16y_inner = ff_vp8_h_loop_filter16_inner_neon;
+ dsp->vp8_v_loop_filter8uv_inner = ff_vp8_v_loop_filter8uv_inner_neon;
+ dsp->vp8_h_loop_filter8uv_inner = ff_vp8_h_loop_filter8uv_inner_neon;
+
+ dsp->vp8_v_loop_filter_simple = ff_vp8_v_loop_filter16_simple_neon;
+ dsp->vp8_h_loop_filter_simple = ff_vp8_h_loop_filter16_simple_neon;
+
+ dsp->put_vp8_epel_pixels_tab[0][0][0] = ff_put_vp8_pixels16_neon;
+ dsp->put_vp8_epel_pixels_tab[0][0][2] = ff_put_vp8_epel16_h6_neon;
+ dsp->put_vp8_epel_pixels_tab[0][2][0] = ff_put_vp8_epel16_v6_neon;
+ dsp->put_vp8_epel_pixels_tab[0][2][2] = ff_put_vp8_epel16_h6v6_neon;
+
+ dsp->put_vp8_epel_pixels_tab[1][0][0] = ff_put_vp8_pixels8_neon;
+ dsp->put_vp8_epel_pixels_tab[1][0][1] = ff_put_vp8_epel8_h4_neon;
+ dsp->put_vp8_epel_pixels_tab[1][0][2] = ff_put_vp8_epel8_h6_neon;
+ dsp->put_vp8_epel_pixels_tab[1][1][0] = ff_put_vp8_epel8_v4_neon;
+ dsp->put_vp8_epel_pixels_tab[1][1][1] = ff_put_vp8_epel8_h4v4_neon;
+ dsp->put_vp8_epel_pixels_tab[1][1][2] = ff_put_vp8_epel8_h6v4_neon;
+ dsp->put_vp8_epel_pixels_tab[1][2][0] = ff_put_vp8_epel8_v6_neon;
+ dsp->put_vp8_epel_pixels_tab[1][2][1] = ff_put_vp8_epel8_h4v6_neon;
+ dsp->put_vp8_epel_pixels_tab[1][2][2] = ff_put_vp8_epel8_h6v6_neon;
+
+ dsp->put_vp8_epel_pixels_tab[2][0][1] = ff_put_vp8_epel4_h4_neon;
+ dsp->put_vp8_epel_pixels_tab[2][0][2] = ff_put_vp8_epel4_h6_neon;
+ dsp->put_vp8_epel_pixels_tab[2][1][0] = ff_put_vp8_epel4_v4_neon;
+ dsp->put_vp8_epel_pixels_tab[2][1][1] = ff_put_vp8_epel4_h4v4_neon;
+ dsp->put_vp8_epel_pixels_tab[2][1][2] = ff_put_vp8_epel4_h6v4_neon;
+ dsp->put_vp8_epel_pixels_tab[2][2][0] = ff_put_vp8_epel4_v6_neon;
+ dsp->put_vp8_epel_pixels_tab[2][2][1] = ff_put_vp8_epel4_h4v6_neon;
+ dsp->put_vp8_epel_pixels_tab[2][2][2] = ff_put_vp8_epel4_h6v6_neon;
+
+ dsp->put_vp8_bilinear_pixels_tab[0][0][0] = ff_put_vp8_pixels16_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][0][1] = ff_put_vp8_bilin16_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][0][2] = ff_put_vp8_bilin16_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][0] = ff_put_vp8_bilin16_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][1] = ff_put_vp8_bilin16_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][1][2] = ff_put_vp8_bilin16_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][0] = ff_put_vp8_bilin16_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][1] = ff_put_vp8_bilin16_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[0][2][2] = ff_put_vp8_bilin16_hv_neon;
+
+ dsp->put_vp8_bilinear_pixels_tab[1][0][0] = ff_put_vp8_pixels8_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][0][1] = ff_put_vp8_bilin8_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][0][2] = ff_put_vp8_bilin8_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][0] = ff_put_vp8_bilin8_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][1] = ff_put_vp8_bilin8_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][1][2] = ff_put_vp8_bilin8_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][0] = ff_put_vp8_bilin8_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][1] = ff_put_vp8_bilin8_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[1][2][2] = ff_put_vp8_bilin8_hv_neon;
+
+ dsp->put_vp8_bilinear_pixels_tab[2][0][1] = ff_put_vp8_bilin4_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][0][2] = ff_put_vp8_bilin4_h_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][0] = ff_put_vp8_bilin4_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][1] = ff_put_vp8_bilin4_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][1][2] = ff_put_vp8_bilin4_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][0] = ff_put_vp8_bilin4_v_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][1] = ff_put_vp8_bilin4_hv_neon;
+ dsp->put_vp8_bilinear_pixels_tab[2][2][2] = ff_put_vp8_bilin4_hv_neon;
+}