summaryrefslogtreecommitdiff
path: root/libavcodec/mips/h264chroma_mmi.c
diff options
context:
space:
mode:
authorZhouXiaoyong <zhouxiaoyong@loongson.cn>2016-05-17 13:02:41 +0800
committerMichael Niedermayer <michael@niedermayer.cc>2016-05-21 14:25:33 +0200
commita3eb5a3cdd5ce871c17185ad3f14fb775d4b59a3 (patch)
treec111048a00dd73ae1698c2c8a8c8345b784d9725 /libavcodec/mips/h264chroma_mmi.c
parent64fe1eebddbbf705d0c25d000c102bfca5558682 (diff)
avcodec/mips/h264chroma_mmi: Version 2 of the optimizations for loongson mmi
1. no longer use the register names directly and optimized code format 2. to be compatible with O32, specify type of address variable with mips_reg and handle the address variable with PTR_ operator 3. use uld and mtc1 to workaround cpu 3A2000 gslwlc1 bug (gslwlc1 instruction extension bug in O32 ABI) Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
Diffstat (limited to 'libavcodec/mips/h264chroma_mmi.c')
-rw-r--r--libavcodec/mips/h264chroma_mmi.c1123
1 files changed, 629 insertions, 494 deletions
diff --git a/libavcodec/mips/h264chroma_mmi.c b/libavcodec/mips/h264chroma_mmi.c
index ef294762b3..3dd123da36 100644
--- a/libavcodec/mips/h264chroma_mmi.c
+++ b/libavcodec/mips/h264chroma_mmi.c
@@ -23,6 +23,8 @@
*/
#include "h264chroma_mips.h"
+#include "constants.h"
+#include "libavutil/mips/asmdefs.h"
void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride,
int h, int x, int y)
@@ -32,171 +34,177 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride,
const int C = (8 - x) * y;
const int D = x * y;
const int E = B + C;
- int i;
-
- av_assert2(x<8 && y<8 && x>=0 && y>=0);
+ double ftmp[10];
+ uint64_t tmp[1];
+ mips_reg addr[1];
if (D) {
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "ldl $4, %6 \r\n"
- "ldr $4, %5 \r\n"
- "ldl $5, %8 \r\n"
- "ldr $5, %7 \r\n"
- "daddiu $6, $0, 32 \r\n"
- "mtc1 %9, $f6 \r\n"
- "mtc1 %10, $f8 \r\n"
- "mtc1 %11, $f10 \r\n"
- "mtc1 %12, $f12 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "mtc1 $6, $f14 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "pshufh $f12, $f12, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "mtc1 $6, $f22 \r\n"
- "dsrl32 $2, $2, 0 \r\n"
- "dsrl32 $3, $3, 0 \r\n"
- "dsrl32 $4, $4, 0 \r\n"
- "dsrl32 $5, $5, 0 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f24, $f2, $f22 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f24, $f2 \r\n"
- "sdc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*src),"m"(*(src+7)),"m"(*(src+1)),"m"(*(src+8)),
- "m"(*(src+stride)),"m"(*(src+stride+7)),
- "m"(*(src+stride+1)),"m"(*(src+stride+8)),
- "r"(A),"r"(B),"r"(C),"r"(D)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[B], %[B], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp9] \n\t"
+ "pshufh %[C], %[C], %[ftmp0] \n\t"
+ "pshufh %[D], %[D], %[ftmp0] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[stride] \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "gsldlc1 %[ftmp2], 0x08(%[src]) \n\t"
+ "gsldrc1 %[ftmp2], 0x01(%[src]) \n\t"
+ "gsldlc1 %[ftmp3], 0x07(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp3], 0x00(%[addr0]) \n\t"
+ "gsldlc1 %[ftmp4], 0x08(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp4], 0x01(%[addr0]) \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp8], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[A] \n\t"
+ "pmullh %[ftmp7], %[ftmp7], %[B] \n\t"
+ "paddh %[ftmp1], %[ftmp5], %[ftmp7] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[A] \n\t"
+ "pmullh %[ftmp8], %[ftmp8], %[B] \n\t"
+ "paddh %[ftmp2], %[ftmp6], %[ftmp8] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[C] \n\t"
+ "pmullh %[ftmp7], %[ftmp7], %[D] \n\t"
+ "paddh %[ftmp3], %[ftmp5], %[ftmp7] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[C] \n\t"
+ "pmullh %[ftmp8], %[ftmp8], %[D] \n\t"
+ "paddh %[ftmp4], %[ftmp6], %[ftmp8] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [B]"f"(B),
+ [C]"f"(C), [D]"f"(D)
+ : "memory"
+ );
} else if (E) {
const int step = C ? stride : 1;
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "daddiu $6, $0, 32 \r\n"
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "mtc1 $6, $f14 \r\n"
- "mtc1 %5, $f6 \r\n"
- "mtc1 %6, $f8 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "mtc1 $6, $f22 \r\n"
- "dsrl32 $2, $2, 0 \r\n"
- "dsrl32 $3, $3, 0 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f24, $f2, $f22 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f24, $f2 \r\n"
- "sdc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),
- "m"(*(src+step)),"m"(*(src+step+7)),
- "r"(A),"r"(E)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[E], %[E], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp7] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[step] \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "gsldlc1 %[ftmp2], 0x07(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp2], 0x00(%[addr0]) \n\t"
+
+ "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp3], %[ftmp3], %[A] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[E] \n\t"
+ "paddh %[ftmp1], %[ftmp3], %[ftmp5] \n\t"
+ "pmullh %[ftmp4], %[ftmp4], %[A] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[E] \n\t"
+ "paddh %[ftmp2], %[ftmp4], %[ftmp6] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step),
+ [ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [E]"f"(E)
+ : "memory"
+ );
} else {
- for (i = 0; i < h; i++) {
- __asm__ volatile (
- "daddiu $6, $0, 32 \r\n"
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "mtc1 $6, $f14 \r\n"
- "mtc1 %3, $f6 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "mtc1 $6, $f22 \r\n"
- "dsrl32 $2, $2, 0 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "psrah $f24, $f2, $f22 \r\n"
- "mtc1 $2, $f2 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f24, $f2 \r\n"
- "sdc1 $f2, %0 \r\n"
- :"=m"(*dst)
- :"m"(*src),"m"(*(src+7)),"r"(A)
- :"$2"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp4] \n\t"
+ "1: \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "pmullh %[ftmp2], %[ftmp3], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "pmullh %[ftmp2], %[ftmp3], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x02 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]),
+ [tmp0]"=&r"(tmp[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A)
+ : "memory"
+ );
}
}
@@ -208,140 +216,185 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride,
const int C = (8 - x) * y;
const int D = x * y;
const int E = B + C;
- int i;
-
- av_assert2(x<8 && y<8 && x>=0 && y>=0);
+ double ftmp[10];
+ uint64_t tmp[1];
+ mips_reg addr[1];
if (D) {
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "ldl $4, %6 \r\n"
- "ldr $4, %5 \r\n"
- "ldl $5, %8 \r\n"
- "ldr $5, %7 \r\n"
- "daddiu $6, $0, 32 \r\n"
- "mtc1 %9, $f6 \r\n"
- "mtc1 %10, $f8 \r\n"
- "mtc1 %11, $f10 \r\n"
- "mtc1 %12, $f12 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "mtc1 $6, $f14 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "pshufh $f12, $f12, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "mtc1 $6, $f22 \r\n"
- "dsrl32 $2, $2, 0 \r\n"
- "dsrl32 $3, $3, 0 \r\n"
- "dsrl32 $4, $4, 0 \r\n"
- "dsrl32 $5, $5, 0 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f24, $f2, $f22 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f24, $f2 \r\n"
- "ldc1 $f4, %0 \r\n"
- "pavgb $f2, $f2, $f4 \r\n"
- "sdc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),"m"(*(src+1)),"m"(*(src+8)),
- "m"(*(src+stride)),"m"(*(src+stride+7)),
- "m"(*(src+stride+1)),"m"(*(src+stride+8)),
- "r"(A),"r"(B),"r"(C),"r"(D)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
- } else {
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[B], %[B], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp9] \n\t"
+ "pshufh %[C], %[C], %[ftmp0] \n\t"
+ "pshufh %[D], %[D], %[ftmp0] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[stride] \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "gsldlc1 %[ftmp2], 0x08(%[src]) \n\t"
+ "gsldrc1 %[ftmp2], 0x01(%[src]) \n\t"
+ "gsldlc1 %[ftmp3], 0x07(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp3], 0x00(%[addr0]) \n\t"
+ "gsldlc1 %[ftmp4], 0x08(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp4], 0x01(%[addr0]) \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp8], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[A] \n\t"
+ "pmullh %[ftmp7], %[ftmp7], %[B] \n\t"
+ "paddh %[ftmp1], %[ftmp5], %[ftmp7] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[A] \n\t"
+ "pmullh %[ftmp8], %[ftmp8], %[B] \n\t"
+ "paddh %[ftmp2], %[ftmp6], %[ftmp8] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[C] \n\t"
+ "pmullh %[ftmp7], %[ftmp7], %[D] \n\t"
+ "paddh %[ftmp3], %[ftmp5], %[ftmp7] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[C] \n\t"
+ "pmullh %[ftmp8], %[ftmp8], %[D] \n\t"
+ "paddh %[ftmp4], %[ftmp6], %[ftmp8] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "ldc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [B]"f"(B),
+ [C]"f"(C), [D]"f"(D)
+ : "memory"
+ );
+ } else if (E) {
const int step = C ? stride : 1;
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "daddiu $6, $0, 32 \r\n"
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "mtc1 $6, $f14 \r\n"
- "mtc1 %5, $f6 \r\n"
- "mtc1 %6, $f8 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "mtc1 $6, $f22 \r\n"
- "dsrl32 $2, $2, 0 \r\n"
- "dsrl32 $3, $3, 0 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f24, $f2, $f22 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f24, $f2 \r\n"
- "ldc1 $f4, %0 \r\n"
- "pavgb $f2, $f2, $f4 \r\n"
- "sdc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),
- "m"(*(src+step)),"m"(*(src+step+7)),"r"(A),"r"(E)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[E], %[E], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp7] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[step] \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "gsldlc1 %[ftmp2], 0x07(%[addr0]) \n\t"
+ "gsldrc1 %[ftmp2], 0x00(%[addr0]) \n\t"
+
+ "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp3], %[ftmp3], %[A] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[E] \n\t"
+ "paddh %[ftmp1], %[ftmp3], %[ftmp5] \n\t"
+ "pmullh %[ftmp4], %[ftmp4], %[A] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[E] \n\t"
+ "paddh %[ftmp2], %[ftmp4], %[ftmp6] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "ldc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step),
+ [ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [E]"f"(E)
+ : "memory"
+ );
+ } else {
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp4] \n\t"
+ "1: \n\t"
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "pmullh %[ftmp2], %[ftmp3], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "ldc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+
+ "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t"
+ "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "pmullh %[ftmp2], %[ftmp3], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t"
+ "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "ldc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x02 \n\t"
+ "sdc1 %[ftmp1], 0x00(%[dst]) \n\t"
+
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]),
+ [tmp0]"=&r"(tmp[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A)
+ : "memory"
+ );
}
}
@@ -353,118 +406,150 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride,
const int C = (8 - x) * y;
const int D = x * y;
const int E = B + C;
- int i;
-
- av_assert2(x<8 && y<8 && x>=0 && y>=0);
+ double ftmp[8];
+ uint64_t tmp[1];
+ mips_reg addr[1];
+ uint64_t low32;
if (D) {
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "ldl $4, %6 \r\n"
- "ldr $4, %5 \r\n"
- "ldl $5, %8 \r\n"
- "ldr $5, %7 \r\n"
- "daddiu $6, $0, 32 \r\n"
- "mtc1 %9, $f6 \r\n"
- "mtc1 %10, $f8 \r\n"
- "mtc1 %11, $f10 \r\n"
- "mtc1 %12, $f12 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "mtc1 $6, $f14 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "pshufh $f12, $f12, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "mtc1 $6, $f22 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f2, $f2 \r\n"
- "swc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),"m"(*(src+1)),"m"(*(src+8)),
- "m"(*(src+stride)),"m"(*(src+stride+7)),
- "m"(*(src+stride+1)),"m"(*(src+stride+8)),
- "r"(A),"r"(B),"r"(C),"r"(D)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[B], %[B], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp7] \n\t"
+ "pshufh %[C], %[C], %[ftmp0] \n\t"
+ "pshufh %[D], %[D], %[ftmp0] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[stride] \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "uld %[low32], 0x01(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp2] \n\t"
+ "uld %[low32], 0x00(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp3] \n\t"
+ "uld %[low32], 0x01(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp4] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[A] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[B] \n\t"
+ "paddh %[ftmp1], %[ftmp5], %[ftmp6] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp6], %[ftmp4], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[C] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[D] \n\t"
+ "paddh %[ftmp2], %[ftmp5], %[ftmp6] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [B]"f"(B),
+ [C]"f"(C), [D]"f"(D)
+ : "memory"
+ );
} else if (E) {
const int step = C ? stride : 1;
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "daddiu $4, $0, 32 \r\n"
- "mtc1 %5, $f6 \r\n"
- "mtc1 %6, $f8 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f10 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "daddiu $4, $0, 6 \r\n"
- "mtc1 $4, $f22 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f10 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f2, $f20 \r\n"
- "swc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),"m"(*(src+step)),
- "m"(*(src+step+7)),"r"(A),"r"(E)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[E], %[E], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp5] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[step] \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "uld %[low32], 0x00(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp2] \n\t"
+
+ "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp3], %[ftmp3], %[A] \n\t"
+ "pmullh %[ftmp4], %[ftmp4], %[E] \n\t"
+ "paddh %[ftmp1], %[ftmp3], %[ftmp4] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step),
+ [ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [E]"f"(E)
+ : "memory"
+ );
} else {
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "lwl $2, %2 \r\n"
- "lwr $2, %1 \r\n"
- "sw $2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*src),"m"(*(src+3))
- : "$2"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp3] \n\t"
+ "1: \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "addi %[h], %[h], -0x02 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [tmp0]"=&r"(tmp[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A)
+ : "memory"
+ );
}
}
@@ -475,108 +560,158 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride,
const int B = x * (8 - y);
const int C = (8 - x) * y;
const int D = x * y;
- int i;
-
- av_assert2(x<8 && y<8 && x>=0 && y>=0);
+ const int E = B + C;
+ double ftmp[8];
+ uint64_t tmp[1];
+ mips_reg addr[1];
+ uint64_t low32;
if (D) {
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "ldl $4, %6 \r\n"
- "ldr $4, %5 \r\n"
- "ldl $5, %8 \r\n"
- "ldr $5, %7 \r\n"
- "daddiu $6, $0, 32 \r\n"
- "mtc1 %9, $f6 \r\n"
- "mtc1 %10, $f8 \r\n"
- "mtc1 %11, $f10 \r\n"
- "mtc1 %12, $f12 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f16 \r\n"
- "mtc1 $5, $f18 \r\n"
- "mtc1 $6, $f14 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "pshufh $f12, $f12, $f20 \r\n"
- "pshufh $f14, $f14, $f20 \r\n"
- "punpcklbh $f16, $f16, $f20 \r\n"
- "punpcklbh $f18, $f18, $f20 \r\n"
- "daddiu $6, $0, 6 \r\n"
- "mtc1 $6, $f22 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "pmullh $f16, $f10, $f16 \r\n"
- "pmullh $f18, $f12, $f18 \r\n"
- "paddh $f2, $f2, $f14 \r\n"
- "paddh $f4, $f4, $f16 \r\n"
- "paddh $f2, $f2, $f18 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f2, $f2 \r\n"
- "lwc1 $f4, %0 \r\n"
- "pavgb $f2, $f2, $f4 \r\n"
- "swc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),"m"(*(src+1)),"m"(*(src+8)),
- "m"(*(src+stride)),"m"(*(src+stride+7)),
- "m"(*(src+stride+1)),"m"(*(src+stride+8)),
- "r"(A),"r"(B),"r"(C),"r"(D)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
- } else {
- const int E = B + C;
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[B], %[B], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp7] \n\t"
+ "pshufh %[C], %[C], %[ftmp0] \n\t"
+ "pshufh %[D], %[D], %[ftmp0] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[stride] \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "uld %[low32], 0x01(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp2] \n\t"
+ "uld %[low32], 0x00(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp3] \n\t"
+ "uld %[low32], 0x01(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp4] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[A] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[B] \n\t"
+ "paddh %[ftmp1], %[ftmp5], %[ftmp6] \n\t"
+
+ "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp6], %[ftmp4], %[ftmp0] \n\t"
+ "pmullh %[ftmp5], %[ftmp5], %[C] \n\t"
+ "pmullh %[ftmp6], %[ftmp6], %[D] \n\t"
+ "paddh %[ftmp2], %[ftmp5], %[ftmp6] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "lwc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [B]"f"(B),
+ [C]"f"(C), [D]"f"(D)
+ : "memory"
+ );
+ } else if (E) {
const int step = C ? stride : 1;
- for (i=0; i<h; i++) {
- __asm__ volatile (
- "ldl $2, %2 \r\n"
- "ldr $2, %1 \r\n"
- "ldl $3, %4 \r\n"
- "ldr $3, %3 \r\n"
- "daddiu $4, $0, 32 \r\n"
- "mtc1 %5, $f6 \r\n"
- "mtc1 %6, $f8 \r\n"
- "mtc1 $0, $f20 \r\n"
- "mtc1 $2, $f2 \r\n"
- "mtc1 $3, $f4 \r\n"
- "mtc1 $4, $f10 \r\n"
- "punpcklbh $f2, $f2, $f20 \r\n"
- "punpcklbh $f4, $f4, $f20 \r\n"
- "pshufh $f6, $f6, $f20 \r\n"
- "pshufh $f8, $f8, $f20 \r\n"
- "pshufh $f10, $f10, $f20 \r\n"
- "daddiu $4, $0, 6 \r\n"
- "mtc1 $4, $f22 \r\n"
- "pmullh $f2, $f2, $f6 \r\n"
- "pmullh $f4, $f4, $f8 \r\n"
- "paddh $f2, $f2, $f10 \r\n"
- "paddh $f2, $f2, $f4 \r\n"
- "psrah $f2, $f2, $f22 \r\n"
- "packushb $f2, $f2, $f20 \r\n"
- "lwc1 $f4, %0 \r\n"
- "pavgb $f2, $f2, $f4 \r\n"
- "swc1 $f2, %0 \r\n"
- : "=m"(*dst)
- : "m"(*(src)),"m"(*(src+7)),"m"(*(src+step)),
- "m"(*(src+step+7)),"r"(A),"r"(E)
- : "$2","$3","$4","$5","$6"
- );
-
- dst += stride;
- src += stride;
- }
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "pshufh %[E], %[E], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp5] \n\t"
+ "1: \n\t"
+ PTR_ADDU "%[addr0], %[src], %[step] \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "uld %[low32], 0x00(%[addr0]) \n\t"
+ "mtc1 %[low32], %[ftmp2] \n\t"
+
+ "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t"
+ "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t"
+ "pmullh %[ftmp3], %[ftmp3], %[A] \n\t"
+ "pmullh %[ftmp4], %[ftmp4], %[E] \n\t"
+ "paddh %[ftmp1], %[ftmp3], %[ftmp4] \n\t"
+
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "lwc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x01 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]),
+ [tmp0]"=&r"(tmp[0]),
+ [addr0]"=&r"(addr[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step),
+ [ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A), [E]"f"(E)
+ : "memory"
+ );
+ } else {
+ __asm__ volatile (
+ "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t"
+ "dli %[tmp0], 0x06 \n\t"
+ "pshufh %[A], %[A], %[ftmp0] \n\t"
+ "mtc1 %[tmp0], %[ftmp3] \n\t"
+ "1: \n\t"
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "lwc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+
+ "uld %[low32], 0x00(%[src]) \n\t"
+ "mtc1 %[low32], %[ftmp1] \n\t"
+ "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t"
+ "pmullh %[ftmp1], %[ftmp2], %[A] \n\t"
+ "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t"
+ "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t"
+ "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t"
+ "lwc1 %[ftmp2], 0x00(%[dst]) \n\t"
+ "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t"
+ "addi %[h], %[h], -0x02 \n\t"
+ "swc1 %[ftmp1], 0x00(%[dst]) \n\t"
+
+ PTR_ADDU "%[src], %[src], %[stride] \n\t"
+ PTR_ADDU "%[dst], %[dst], %[stride] \n\t"
+ "bnez %[h], 1b \n\t"
+ : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]),
+ [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]),
+ [tmp0]"=&r"(tmp[0]),
+ [dst]"+&r"(dst), [src]"+&r"(src),
+ [h]"+&r"(h),
+ [low32]"=&r"(low32)
+ : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32),
+ [A]"f"(A)
+ : "memory"
+ );
}
}