aboutsummaryrefslogtreecommitdiff
path: root/arch/x86/math-emu/wm_shrx.S
diff options
context:
space:
mode:
authorGravatar Peter Zijlstra <peterz@infradead.org> 2021-12-04 14:43:40 +0100
committerGravatar Borislav Petkov <bp@suse.de> 2021-12-08 12:25:37 +0100
commitf94909ceb1ed4bfdb2ada72f93236305e6d6951f (patch)
treee7368c974b829530e82f34111f1f5ab183504921 /arch/x86/math-emu/wm_shrx.S
parentx86/lib/atomic64_386_32: Rename things (diff)
downloadlinux-f94909ceb1ed4bfdb2ada72f93236305e6d6951f.tar.gz
linux-f94909ceb1ed4bfdb2ada72f93236305e6d6951f.tar.bz2
linux-f94909ceb1ed4bfdb2ada72f93236305e6d6951f.zip
x86: Prepare asm files for straight-line-speculation
Replace all ret/retq instructions with RET in preparation of making RET a macro. Since AS is case insensitive it's a big no-op without RET defined. find arch/x86/ -name \*.S | while read file do sed -i 's/\<ret[q]*\>/RET/' $file done Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Borislav Petkov <bp@suse.de> Link: https://lore.kernel.org/r/20211204134907.905503893@infradead.org
Diffstat (limited to 'arch/x86/math-emu/wm_shrx.S')
-rw-r--r--arch/x86/math-emu/wm_shrx.S16
1 files changed, 8 insertions, 8 deletions
diff --git a/arch/x86/math-emu/wm_shrx.S b/arch/x86/math-emu/wm_shrx.S
index 4fc89174caf0..f608a28a4c43 100644
--- a/arch/x86/math-emu/wm_shrx.S
+++ b/arch/x86/math-emu/wm_shrx.S
@@ -55,7 +55,7 @@ SYM_FUNC_START(FPU_shrx)
popl %ebx
popl %esi
leave
- ret
+ RET
L_more_than_31:
cmpl $64,%ecx
@@ -70,7 +70,7 @@ L_more_than_31:
movl $0,4(%esi)
popl %esi
leave
- ret
+ RET
L_more_than_63:
cmpl $96,%ecx
@@ -84,7 +84,7 @@ L_more_than_63:
movl %edx,4(%esi)
popl %esi
leave
- ret
+ RET
L_more_than_95:
xorl %eax,%eax
@@ -92,7 +92,7 @@ L_more_than_95:
movl %eax,4(%esi)
popl %esi
leave
- ret
+ RET
SYM_FUNC_END(FPU_shrx)
@@ -146,7 +146,7 @@ SYM_FUNC_START(FPU_shrxs)
popl %ebx
popl %esi
leave
- ret
+ RET
/* Shift by [0..31] bits */
Ls_less_than_32:
@@ -163,7 +163,7 @@ Ls_less_than_32:
popl %ebx
popl %esi
leave
- ret
+ RET
/* Shift by [64..95] bits */
Ls_more_than_63:
@@ -189,7 +189,7 @@ Ls_more_than_63:
popl %ebx
popl %esi
leave
- ret
+ RET
Ls_more_than_95:
/* Shift by [96..inf) bits */
@@ -203,5 +203,5 @@ Ls_more_than_95:
popl %ebx
popl %esi
leave
- ret
+ RET
SYM_FUNC_END(FPU_shrxs)