Add ARM mode tests for the NEON vector shift-accumulate tests.
authorJim Grosbach <grosbach@apple.com>
Mon, 23 Apr 2012 21:00:44 +0000 (21:00 +0000)
committerJim Grosbach <grosbach@apple.com>
Mon, 23 Apr 2012 21:00:44 +0000 (21:00 +0000)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@155390 91177308-0d34-0410-b5e6-96231b3b80d8

test/MC/ARM/neon-shiftaccum-encoding.s [new file with mode: 0644]

diff --git a/test/MC/ARM/neon-shiftaccum-encoding.s b/test/MC/ARM/neon-shiftaccum-encoding.s
new file mode 100644 (file)
index 0000000..a4fa358
--- /dev/null
@@ -0,0 +1,105 @@
+@ RUN: llvm-mc -mcpu=cortex-a8 -triple arm-unknown-unknown -show-encoding < %s | FileCheck %s
+
+       vsra.s8 d17, d16, #8
+       vsra.s16 d15, d14, #16
+       vsra.s32 d13, d12, #32
+       vsra.s64 d11, d10, #64
+       vsra.s8 q7, q2, #8
+       vsra.s16 q3, q6, #16
+       vsra.s32 q9, q5, #32
+       vsra.s64 q8, q4, #64
+       vsra.u8 d17, d16, #8
+       vsra.u16 d11, d14, #11
+       vsra.u32 d12, d15, #22
+       vsra.u64 d13, d16, #54
+       vsra.u8 q1, q7, #8
+       vsra.u16 q2, q7, #6
+       vsra.u32 q3, q6, #21
+       vsra.u64 q4, q5, #25
+
+@ CHECK: vsra.s8       d17, d16, #8    @ encoding: [0x30,0x11,0xc8,0xf2]
+@ CHECK: vsra.s16      d15, d14, #16   @ encoding: [0x1e,0xf1,0x90,0xf2]
+@ CHECK: vsra.s32      d13, d12, #32   @ encoding: [0x1c,0xd1,0xa0,0xf2]
+@ CHECK: vsra.s64      d11, d10, #64   @ encoding: [0x9a,0xb1,0x80,0xf2]
+@ CHECK: vsra.s8       q7, q2, #8      @ encoding: [0x54,0xe1,0x88,0xf2]
+@ CHECK: vsra.s16      q3, q6, #16     @ encoding: [0x5c,0x61,0x90,0xf2]
+@ CHECK: vsra.s32      q9, q5, #32     @ encoding: [0x5a,0x21,0xe0,0xf2]
+@ CHECK: vsra.s64      q8, q4, #64     @ encoding: [0xd8,0x01,0xc0,0xf2]
+@ CHECK: vsra.u8       d17, d16, #8    @ encoding: [0x30,0x11,0xc8,0xf3]
+@ CHECK: vsra.u16      d11, d14, #11   @ encoding: [0x1e,0xb1,0x95,0xf3]
+@ CHECK: vsra.u32      d12, d15, #22   @ encoding: [0x1f,0xc1,0xaa,0xf3]
+@ CHECK: vsra.u64      d13, d16, #54   @ encoding: [0xb0,0xd1,0x8a,0xf3]
+@ CHECK: vsra.u8       q1, q7, #8      @ encoding: [0x5e,0x21,0x88,0xf3]
+@ CHECK: vsra.u16      q2, q7, #6      @ encoding: [0x5e,0x41,0x9a,0xf3]
+@ CHECK: vsra.u32      q3, q6, #21     @ encoding: [0x5c,0x61,0xab,0xf3]
+@ CHECK: vsra.u64      q4, q5, #25     @ encoding: [0xda,0x81,0xa7,0xf3]
+
+
+       vrsra.s8 d5, d26, #8
+       vrsra.s16 d6, d25, #16
+       vrsra.s32 d7, d24, #32
+       vrsra.s64 d14, d23, #64
+       vrsra.u8 d15, d22, #8
+       vrsra.u16 d16, d21, #16
+       vrsra.u32 d17, d20, #32
+       vrsra.u64 d18, d19, #64
+       vrsra.s8 q1, q2, #8
+       vrsra.s16 q2, q3, #16
+       vrsra.s32 q3, q4, #32
+       vrsra.s64 q4, q5, #64
+       vrsra.u8 q5, q6, #8
+       vrsra.u16 q6, q7, #16
+       vrsra.u32 q7, q8, #32
+       vrsra.u64 q8, q9, #64
+
+@ CHECK: vrsra.s8      d5, d26, #8     @ encoding: [0x3a,0x53,0x88,0xf2]
+@ CHECK: vrsra.s16     d6, d25, #16    @ encoding: [0x39,0x63,0x90,0xf2]
+@ CHECK: vrsra.s32     d7, d24, #32    @ encoding: [0x38,0x73,0xa0,0xf2]
+@ CHECK: vrsra.s64     d14, d23, #64   @ encoding: [0xb7,0xe3,0x80,0xf2]
+@ CHECK: vrsra.u8      d15, d22, #8    @ encoding: [0x36,0xf3,0x88,0xf3]
+@ CHECK: vrsra.u16     d16, d21, #16   @ encoding: [0x35,0x03,0xd0,0xf3]
+@ CHECK: vrsra.u32     d17, d20, #32   @ encoding: [0x34,0x13,0xe0,0xf3]
+@ CHECK: vrsra.u64     d18, d19, #64   @ encoding: [0xb3,0x23,0xc0,0xf3]
+@ CHECK: vrsra.s8      q1, q2, #8      @ encoding: [0x54,0x23,0x88,0xf2]
+@ CHECK: vrsra.s16     q2, q3, #16     @ encoding: [0x56,0x43,0x90,0xf2]
+@ CHECK: vrsra.s32     q3, q4, #32     @ encoding: [0x58,0x63,0xa0,0xf2]
+@ CHECK: vrsra.s64     q4, q5, #64     @ encoding: [0xda,0x83,0x80,0xf2]
+@ CHECK: vrsra.u8      q5, q6, #8      @ encoding: [0x5c,0xa3,0x88,0xf3]
+@ CHECK: vrsra.u16     q6, q7, #16     @ encoding: [0x5e,0xc3,0x90,0xf3]
+@ CHECK: vrsra.u32     q7, q8, #32     @ encoding: [0x70,0xe3,0xa0,0xf3]
+@ CHECK: vrsra.u64     q8, q9, #64     @ encoding: [0xf2,0x03,0xc0,0xf3]
+
+
+       vsli.8 d11, d12, #7
+       vsli.16 d12, d13, #15
+       vsli.32 d13, d14, #31
+       vsli.64 d14, d15, #63
+       vsli.8 q1, q8, #7
+       vsli.16 q2, q7, #15
+       vsli.32 q3, q4, #31
+       vsli.64 q4, q5, #63
+       vsri.8 d28, d11, #8
+       vsri.16 d26, d12, #16
+       vsri.32 d24, d13, #32
+       vsri.64 d21, d14, #64
+       vsri.8 q1, q8, #8
+       vsri.16 q5, q2, #16
+       vsri.32 q7, q4, #32
+       vsri.64 q9, q6, #64
+
+@ CHECK: vsli.8        d11, d12, #7            @ encoding: [0x1c,0xb5,0x8f,0xf3]
+@ CHECK: vsli.16       d12, d13, #15   @ encoding: [0x1d,0xc5,0x9f,0xf3]
+@ CHECK: vsli.32       d13, d14, #31   @ encoding: [0x1e,0xd5,0xbf,0xf3]
+@ CHECK: vsli.64       d14, d15, #63   @ encoding: [0x9f,0xe5,0xbf,0xf3]
+@ CHECK: vsli.8        q1, q8, #7              @ encoding: [0x70,0x25,0x8f,0xf3]
+@ CHECK: vsli.16       q2, q7, #15     @ encoding: [0x5e,0x45,0x9f,0xf3]
+@ CHECK: vsli.32       q3, q4, #31     @ encoding: [0x58,0x65,0xbf,0xf3]
+@ CHECK: vsli.64       q4, q5, #63     @ encoding: [0xda,0x85,0xbf,0xf3]
+@ CHECK: vsri.8        d28, d11, #8            @ encoding: [0x1b,0xc4,0xc8,0xf3]
+@ CHECK: vsri.16       d26, d12, #16   @ encoding: [0x1c,0xa4,0xd0,0xf3]
+@ CHECK: vsri.32       d24, d13, #32   @ encoding: [0x1d,0x84,0xe0,0xf3]
+@ CHECK: vsri.64       d21, d14, #64   @ encoding: [0x9e,0x54,0xc0,0xf3]
+@ CHECK: vsri.8        q1, q8, #8              @ encoding: [0x70,0x24,0x88,0xf3]
+@ CHECK: vsri.16       q5, q2, #16     @ encoding: [0x54,0xa4,0x90,0xf3]
+@ CHECK: vsri.32       q7, q4, #32     @ encoding: [0x58,0xe4,0xa0,0xf3]
+@ CHECK: vsri.64       q9, q6, #64     @ encoding: [0xdc,0x24,0xc0,0xf3]