summaryrefslogtreecommitdiff
path: root/libavcodec/x86/hevc_deblock.asm
diff options
context:
space:
mode:
authorCarl Eugen Hoyos <cehoyos@ag.or.at>2014-05-17 12:28:14 +0200
committerCarl Eugen Hoyos <cehoyos@ag.or.at>2014-05-17 12:50:55 +0200
commitef2713747f9db4b59afa844e62493c09cefaf11e (patch)
tree315ffffc5d9bdd076b2d21d4174158e72c911eac /libavcodec/x86/hevc_deblock.asm
parent4c49d0824a10b319c021a1df90abab5a45a644dc (diff)
Fix compilation of libavcodec/x86/hevc_deblock.asm with nasm.
Suggested-by: Reimar
Diffstat (limited to 'libavcodec/x86/hevc_deblock.asm')
-rw-r--r--libavcodec/x86/hevc_deblock.asm42
1 files changed, 21 insertions, 21 deletions
diff --git a/libavcodec/x86/hevc_deblock.asm b/libavcodec/x86/hevc_deblock.asm
index 1522d5e5b8..858ae55541 100644
--- a/libavcodec/x86/hevc_deblock.asm
+++ b/libavcodec/x86/hevc_deblock.asm
@@ -351,12 +351,12 @@ ALIGN 16
;beta calculations
mov r11, [betaq];
shl r11, %1 - 8
- movd m13, r11; beta0
+ movd m13, r11d; beta0
add betaq, 4;
punpcklwd m13, m13
mov r12, [betaq];
shl r12, %1 - 8
- movd m14, r12; beta1
+ movd m14, r12d; beta1
punpcklwd m14, m14
pshufd m13, m14, 0; beta0, beta1
;end beta calculations
@@ -388,28 +388,28 @@ ALIGN 16
pshufd m8, m10, 0x31
psrld m8, 16
paddw m8, m10
- movd r7, m8
+ movd r7d, m8
and r7, 0xffff; 1dp0 + 1dp3
pshufd m8, m8, 0x4E
- movd r8, m8
+ movd r8d, m8
and r8, 0xffff; 0dp0 + 0dp3
pshufd m8, m11, 0x31
psrld m8, 16
paddw m8, m11
- movd r9, m8
+ movd r9d, m8
and r9, 0xffff; 1dq0 + 1dq3
pshufd m8, m8, 0x4E
- movd r10, m8
+ movd r10d, m8
and r10, 0xffff; 0dq0 + 0dq3
; end calc for weak filter
; filtering mask
mov r2, r13
shr r2, 3
- movd m15, r2
+ movd m15, r2d
and r13, 1
- movd m11, r13
+ movd m11, r13d
shufps m11, m15, 0
shl r2, 1
or r13, r2
@@ -422,14 +422,14 @@ ALIGN 16
;tc25 calculations
mov r2d, [tcq];
shl r2, %1 - 8
- movd m8, r2; tc0
+ movd m8, r2d; tc0
add tcq, 4;
punpcklwd m8, m8
mov r3d, [tcq];
shl r3, %1 - 8
- movd m9, r3; tc0
+ movd m9, r3d; tc0
punpcklwd m9, m9
- movd m9, r3; tc1
+ movd m9, r3d; tc1
add r2d, r3d; tc0 + tc1
cmp r2d, 0;
je .bypassluma
@@ -483,10 +483,10 @@ ALIGN 16
and r14, 5; 0b101
mov r2, r14; strong mask
shr r14, 2;
- movd m12, r14; store to xmm for mask generation
+ movd m12, r14d; store to xmm for mask generation
shl r14, 1
and r2, 1
- movd m10, r2; store to xmm for mask generation
+ movd m10, r2d; store to xmm for mask generation
or r14, r2; final strong mask, bits 1 and 0
cmp r14, 0;
je .weakfilter
@@ -584,9 +584,9 @@ ALIGN 16
; weak filtering mask
mov r2, r14
shr r2, 1
- movd m12, r2
+ movd m12, r2d
and r14, 1
- movd m11, r14
+ movd m11, r14d
shufps m11, m12, 0
pcmpeqd m12, m12; set all bits to 1
@@ -646,14 +646,14 @@ ALIGN 16
paddw m15, m2; p1'
;beta calculations
- movd m10, r11; beta0
+ movd m10, r11d; beta0
punpcklwd m10, m10
- movd m13, r12; beta1
+ movd m13, r12d; beta1
punpcklwd m13, m13
shufps m10, m13, 0; betax0, betax1
- movd m13, r7; 1dp0 + 1dp3
- movd m8, r8; 0dp0 + 0dp3
+ movd m13, r7d; 1dp0 + 1dp3
+ movd m8, r8d; 0dp0 + 0dp3
punpcklwd m8, m8
punpcklwd m13, m13
shufps m13, m8, 0;
@@ -670,8 +670,8 @@ ALIGN 16
pminsw m8, m9; av_clip(deltaq1, -tc/2, tc/2)
paddw m8, m5; q1'
- movd m13, r9;
- movd m15, r10;
+ movd m13, r9d;
+ movd m15, r10d;
punpcklwd m15, m15
punpcklwd m13, m13
shufps m13, m15, 0; dq0 + dq3