projects
/
oota-llvm.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
[x86] Implement a faster vector population count based on the PSHUFB
[oota-llvm.git]
/
test
/
CodeGen
/
X86
/
byval2.ll
diff --git
a/test/CodeGen/X86/byval2.ll
b/test/CodeGen/X86/byval2.ll
index f85c8ffbe4fe7352149b68aa96fa8135b974bf84..cc72a8699a9c7c8454cac0667ac574dacb432f13 100644
(file)
--- a/
test/CodeGen/X86/byval2.ll
+++ b/
test/CodeGen/X86/byval2.ll
@@
-1,5
+1,28
@@
-; RUN: llvm-as < %s | llc -march=x86-64 | grep rep.movsq | count 2
-; RUN: llvm-as < %s | llc -march=x86 | grep rep.movsl | count 2
+; RUN: llc < %s -mtriple=x86_64-linux -mattr=-avx | FileCheck %s -check-prefix=X64
+; X64-NOT: movsq
+; X64: rep
+; X64-NOT: rep
+; X64: movsq
+; X64-NOT: movsq
+; X64: rep
+; X64-NOT: rep
+; X64: movsq
+; X64-NOT: rep
+; X64-NOT: movsq
+
+; Win64 has not supported byval yet.
+
+; RUN: llc < %s -march=x86 -mattr=-avx | FileCheck %s -check-prefix=X32
+; X32-NOT: movsl
+; X32: rep
+; X32-NOT: rep
+; X32: movsl
+; X32-NOT: movsl
+; X32: rep
+; X32-NOT: rep
+; X32: movsl
+; X32-NOT: rep
+; X32-NOT: movsl
%struct.s = type { i64, i64, i64, i64, i64, i64, i64, i64,
i64, i64, i64, i64, i64, i64, i64, i64,
%struct.s = type { i64, i64, i64, i64, i64, i64, i64, i64,
i64, i64, i64, i64, i64, i64, i64, i64,
@@
-8,14
+31,14
@@
define void @g(i64 %a, i64 %b, i64 %c) {
entry:
%d = alloca %struct.s, align 16
define void @g(i64 %a, i64 %b, i64 %c) {
entry:
%d = alloca %struct.s, align 16
- %tmp = getelementptr %struct.s* %d, i32 0, i32 0
+ %tmp = getelementptr %struct.s
, %struct.s
* %d, i32 0, i32 0
store i64 %a, i64* %tmp, align 16
store i64 %a, i64* %tmp, align 16
- %tmp2 = getelementptr %struct.s* %d, i32 0, i32 1
+ %tmp2 = getelementptr %struct.s
, %struct.s
* %d, i32 0, i32 1
store i64 %b, i64* %tmp2, align 16
store i64 %b, i64* %tmp2, align 16
- %tmp4 = getelementptr %struct.s* %d, i32 0, i32 2
+ %tmp4 = getelementptr %struct.s
, %struct.s
* %d, i32 0, i32 2
store i64 %c, i64* %tmp4, align 16
store i64 %c, i64* %tmp4, align 16
- call void @f( %struct.s*
%d byval
)
- call void @f( %struct.s*
%d byval
)
+ call void @f( %struct.s*
byval %d
)
+ call void @f( %struct.s*
byval %d
)
ret void
}
ret void
}