From e2a9376b1bd2204ea6f56a35b762e28e0ef4e35a Mon Sep 17 00:00:00 2001 From: Jack Carter Date: Thu, 15 Aug 2013 12:24:57 +0000 Subject: [PATCH] [Mips][msa] Added the simple builtins (add_a to dpsub[su], ilvev to ldi) Includes: add_a, adds_[asu], addv, addvi, andi.b, asub_[su].[bhwd], aver?_[su]_[bhwd], bclr, bclri, bins[lr], bins[lr]i, bmnzi, bmzi, bneg, bnegi, bseli, bset, bseti, c(eq|ne), c(eq|ne)i, cl[et]_[su], cl[et]i_[su], copy_[su].[bhw], div_[su], dotp_[su], dpadd_[su], dpsub_[su], ilvev, ilvl, ilvod, ilvr, insv, insve, ldi Patch by Daniel Sanders git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@188457 91177308-0d34-0410-b5e6-96231b3b80d8 --- include/llvm/IR/IntrinsicsMips.td | 443 ++++++++ lib/Target/Mips/MipsInstrInfo.td | 4 + lib/Target/Mips/MipsMSAInstrFormats.td | 77 ++ lib/Target/Mips/MipsMSAInstrInfo.td | 1137 ++++++++++++++++++++- lib/Target/Mips/MipsRegisterInfo.td | 5 +- lib/Target/Mips/MipsSEISelLowering.cpp | 33 +- lib/Target/Mips/MipsSEISelLowering.h | 2 + test/CodeGen/Mips/msa/2r_vector_scalar.ll | 59 ++ test/CodeGen/Mips/msa/3r-a.ll | 970 ++++++++++++++++++ test/CodeGen/Mips/msa/3r-b.ll | 442 ++++++++ test/CodeGen/Mips/msa/3r-c.ll | 442 ++++++++ test/CodeGen/Mips/msa/3r-d.ll | 354 +++++++ test/CodeGen/Mips/msa/3r-i.ll | 354 +++++++ test/CodeGen/Mips/msa/3r_4r_widen.ll | 302 ++++++ test/CodeGen/Mips/msa/elm_copy.ll | 116 +++ test/CodeGen/Mips/msa/elm_insv.ll | 68 ++ test/CodeGen/Mips/msa/i5-a.ll | 78 ++ test/CodeGen/Mips/msa/i5-b.ll | 382 +++++++ test/CodeGen/Mips/msa/i5-c.ll | 382 +++++++ test/CodeGen/Mips/msa/i8.ll | 78 ++ 20 files changed, 5708 insertions(+), 20 deletions(-) create mode 100644 test/CodeGen/Mips/msa/2r_vector_scalar.ll create mode 100644 test/CodeGen/Mips/msa/3r-a.ll create mode 100644 test/CodeGen/Mips/msa/3r-b.ll create mode 100644 test/CodeGen/Mips/msa/3r-c.ll create mode 100644 test/CodeGen/Mips/msa/3r-d.ll create mode 100644 test/CodeGen/Mips/msa/3r-i.ll create mode 100644 test/CodeGen/Mips/msa/3r_4r_widen.ll create mode 100644 test/CodeGen/Mips/msa/elm_copy.ll create mode 100644 test/CodeGen/Mips/msa/elm_insv.ll create mode 100644 test/CodeGen/Mips/msa/i5-a.ll create mode 100644 test/CodeGen/Mips/msa/i5-b.ll create mode 100644 test/CodeGen/Mips/msa/i5-c.ll create mode 100644 test/CodeGen/Mips/msa/i8.ll diff --git a/include/llvm/IR/IntrinsicsMips.td b/include/llvm/IR/IntrinsicsMips.td index a0987c815b5..f497c73a73b 100644 --- a/include/llvm/IR/IntrinsicsMips.td +++ b/include/llvm/IR/IntrinsicsMips.td @@ -386,4 +386,447 @@ def int_mips_subuh_qb: GCCBuiltin<"__builtin_mips_subuh_qb">, Intrinsic<[llvm_v4i8_ty], [llvm_v4i8_ty, llvm_v4i8_ty], [IntrNoMem]>; def int_mips_subuh_r_qb: GCCBuiltin<"__builtin_mips_subuh_r_qb">, Intrinsic<[llvm_v4i8_ty], [llvm_v4i8_ty, llvm_v4i8_ty], [IntrNoMem]>; + +//===----------------------------------------------------------------------===// +// MIPS MSA + +//===----------------------------------------------------------------------===// +// Addition/subtraction + +def int_mips_add_a_b : GCCBuiltin<"__builtin_msa_add_a_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_add_a_h : GCCBuiltin<"__builtin_msa_add_a_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_add_a_w : GCCBuiltin<"__builtin_msa_add_a_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_add_a_d : GCCBuiltin<"__builtin_msa_add_a_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_adds_a_b : GCCBuiltin<"__builtin_msa_adds_a_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_adds_a_h : GCCBuiltin<"__builtin_msa_adds_a_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_adds_a_w : GCCBuiltin<"__builtin_msa_adds_a_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_adds_a_d : GCCBuiltin<"__builtin_msa_adds_a_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_adds_s_b : GCCBuiltin<"__builtin_msa_adds_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_adds_s_h : GCCBuiltin<"__builtin_msa_adds_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_adds_s_w : GCCBuiltin<"__builtin_msa_adds_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_adds_s_d : GCCBuiltin<"__builtin_msa_adds_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_adds_u_b : GCCBuiltin<"__builtin_msa_adds_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_adds_u_h : GCCBuiltin<"__builtin_msa_adds_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_adds_u_w : GCCBuiltin<"__builtin_msa_adds_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_adds_u_d : GCCBuiltin<"__builtin_msa_adds_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_addv_b : GCCBuiltin<"__builtin_msa_addv_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_addv_h : GCCBuiltin<"__builtin_msa_addv_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_addv_w : GCCBuiltin<"__builtin_msa_addv_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_addv_d : GCCBuiltin<"__builtin_msa_addv_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_addvi_b : GCCBuiltin<"__builtin_msa_addvi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], [Commutative]>; +def int_mips_addvi_h : GCCBuiltin<"__builtin_msa_addvi_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], [Commutative]>; +def int_mips_addvi_w : GCCBuiltin<"__builtin_msa_addvi_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], [Commutative]>; +def int_mips_addvi_d : GCCBuiltin<"__builtin_msa_addvi_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], [Commutative]>; + +def int_mips_andi_b : GCCBuiltin<"__builtin_msa_andi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; + +def int_mips_asub_s_b : GCCBuiltin<"__builtin_msa_asub_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_asub_s_h : GCCBuiltin<"__builtin_msa_asub_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_asub_s_w : GCCBuiltin<"__builtin_msa_asub_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_asub_s_d : GCCBuiltin<"__builtin_msa_asub_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_asub_u_b : GCCBuiltin<"__builtin_msa_asub_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_asub_u_h : GCCBuiltin<"__builtin_msa_asub_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_asub_u_w : GCCBuiltin<"__builtin_msa_asub_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_asub_u_d : GCCBuiltin<"__builtin_msa_asub_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_ave_s_b : GCCBuiltin<"__builtin_msa_ave_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_ave_s_h : GCCBuiltin<"__builtin_msa_ave_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_ave_s_w : GCCBuiltin<"__builtin_msa_ave_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_ave_s_d : GCCBuiltin<"__builtin_msa_ave_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_ave_u_b : GCCBuiltin<"__builtin_msa_ave_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_ave_u_h : GCCBuiltin<"__builtin_msa_ave_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_ave_u_w : GCCBuiltin<"__builtin_msa_ave_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_ave_u_d : GCCBuiltin<"__builtin_msa_ave_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_aver_s_b : GCCBuiltin<"__builtin_msa_aver_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_aver_s_h : GCCBuiltin<"__builtin_msa_aver_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_aver_s_w : GCCBuiltin<"__builtin_msa_aver_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_aver_s_d : GCCBuiltin<"__builtin_msa_aver_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_aver_u_b : GCCBuiltin<"__builtin_msa_aver_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], [Commutative]>; +def int_mips_aver_u_h : GCCBuiltin<"__builtin_msa_aver_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], [Commutative]>; +def int_mips_aver_u_w : GCCBuiltin<"__builtin_msa_aver_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], [Commutative]>; +def int_mips_aver_u_d : GCCBuiltin<"__builtin_msa_aver_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], [Commutative]>; + +def int_mips_bclr_b : GCCBuiltin<"__builtin_msa_bclr_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_bclr_h : GCCBuiltin<"__builtin_msa_bclr_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_bclr_w : GCCBuiltin<"__builtin_msa_bclr_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_bclr_d : GCCBuiltin<"__builtin_msa_bclr_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_bclri_b : GCCBuiltin<"__builtin_msa_bclri_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_bclri_h : GCCBuiltin<"__builtin_msa_bclri_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_bclri_w : GCCBuiltin<"__builtin_msa_bclri_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_bclri_d : GCCBuiltin<"__builtin_msa_bclri_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_binsl_b : GCCBuiltin<"__builtin_msa_binsl_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_binsl_h : GCCBuiltin<"__builtin_msa_binsl_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_binsl_w : GCCBuiltin<"__builtin_msa_binsl_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_binsl_d : GCCBuiltin<"__builtin_msa_binsl_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_binsli_b : GCCBuiltin<"__builtin_msa_binsli_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_binsli_h : GCCBuiltin<"__builtin_msa_binsli_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_binsli_w : GCCBuiltin<"__builtin_msa_binsli_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_binsli_d : GCCBuiltin<"__builtin_msa_binsli_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_binsr_b : GCCBuiltin<"__builtin_msa_binsr_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_binsr_h : GCCBuiltin<"__builtin_msa_binsr_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_binsr_w : GCCBuiltin<"__builtin_msa_binsr_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_binsr_d : GCCBuiltin<"__builtin_msa_binsr_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_binsri_b : GCCBuiltin<"__builtin_msa_binsri_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_binsri_h : GCCBuiltin<"__builtin_msa_binsri_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_binsri_w : GCCBuiltin<"__builtin_msa_binsri_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_binsri_d : GCCBuiltin<"__builtin_msa_binsri_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_bmnzi_b : GCCBuiltin<"__builtin_msa_bmnzi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; + +def int_mips_bmzi_b : GCCBuiltin<"__builtin_msa_bmzi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; + +def int_mips_bneg_b : GCCBuiltin<"__builtin_msa_bneg_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_bneg_h : GCCBuiltin<"__builtin_msa_bneg_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_bneg_w : GCCBuiltin<"__builtin_msa_bneg_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_bneg_d : GCCBuiltin<"__builtin_msa_bneg_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_bnegi_b : GCCBuiltin<"__builtin_msa_bnegi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_bnegi_h : GCCBuiltin<"__builtin_msa_bnegi_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_bnegi_w : GCCBuiltin<"__builtin_msa_bnegi_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_bnegi_d : GCCBuiltin<"__builtin_msa_bnegi_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_bseli_b : GCCBuiltin<"__builtin_msa_bseli_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_bseli_h : GCCBuiltin<"__builtin_msa_bseli_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_bseli_w : GCCBuiltin<"__builtin_msa_bseli_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_bseli_d : GCCBuiltin<"__builtin_msa_bseli_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_bset_b : GCCBuiltin<"__builtin_msa_bset_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_bset_h : GCCBuiltin<"__builtin_msa_bset_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_bset_w : GCCBuiltin<"__builtin_msa_bset_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_bset_d : GCCBuiltin<"__builtin_msa_bset_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_bseti_b : GCCBuiltin<"__builtin_msa_bseti_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_bseti_h : GCCBuiltin<"__builtin_msa_bseti_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_bseti_w : GCCBuiltin<"__builtin_msa_bseti_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_bseti_d : GCCBuiltin<"__builtin_msa_bseti_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_ceq_b : GCCBuiltin<"__builtin_msa_ceq_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_ceq_h : GCCBuiltin<"__builtin_msa_ceq_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_ceq_w : GCCBuiltin<"__builtin_msa_ceq_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_ceq_d : GCCBuiltin<"__builtin_msa_ceq_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_ceqi_b : GCCBuiltin<"__builtin_msa_ceqi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_ceqi_h : GCCBuiltin<"__builtin_msa_ceqi_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_ceqi_w : GCCBuiltin<"__builtin_msa_ceqi_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_ceqi_d : GCCBuiltin<"__builtin_msa_ceqi_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_cle_s_b : GCCBuiltin<"__builtin_msa_cle_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_cle_s_h : GCCBuiltin<"__builtin_msa_cle_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_cle_s_w : GCCBuiltin<"__builtin_msa_cle_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_cle_s_d : GCCBuiltin<"__builtin_msa_cle_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_cle_u_b : GCCBuiltin<"__builtin_msa_cle_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_cle_u_h : GCCBuiltin<"__builtin_msa_cle_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_cle_u_w : GCCBuiltin<"__builtin_msa_cle_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_cle_u_d : GCCBuiltin<"__builtin_msa_cle_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_clei_s_b : GCCBuiltin<"__builtin_msa_clei_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_clei_s_h : GCCBuiltin<"__builtin_msa_clei_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_clei_s_w : GCCBuiltin<"__builtin_msa_clei_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_clei_s_d : GCCBuiltin<"__builtin_msa_clei_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_clei_u_b : GCCBuiltin<"__builtin_msa_clei_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_clei_u_h : GCCBuiltin<"__builtin_msa_clei_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_clei_u_w : GCCBuiltin<"__builtin_msa_clei_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_clei_u_d : GCCBuiltin<"__builtin_msa_clei_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_clt_s_b : GCCBuiltin<"__builtin_msa_clt_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_clt_s_h : GCCBuiltin<"__builtin_msa_clt_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_clt_s_w : GCCBuiltin<"__builtin_msa_clt_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_clt_s_d : GCCBuiltin<"__builtin_msa_clt_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_clt_u_b : GCCBuiltin<"__builtin_msa_clt_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_clt_u_h : GCCBuiltin<"__builtin_msa_clt_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_clt_u_w : GCCBuiltin<"__builtin_msa_clt_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_clt_u_d : GCCBuiltin<"__builtin_msa_clt_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_clti_s_b : GCCBuiltin<"__builtin_msa_clti_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_clti_s_h : GCCBuiltin<"__builtin_msa_clti_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_clti_s_w : GCCBuiltin<"__builtin_msa_clti_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_clti_s_d : GCCBuiltin<"__builtin_msa_clti_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_clti_u_b : GCCBuiltin<"__builtin_msa_clti_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_clti_u_h : GCCBuiltin<"__builtin_msa_clti_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_clti_u_w : GCCBuiltin<"__builtin_msa_clti_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; +def int_mips_clti_u_d : GCCBuiltin<"__builtin_msa_clti_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_i32_ty], []>; + +def int_mips_copy_s_b : GCCBuiltin<"__builtin_msa_copy_s_b">, + Intrinsic<[llvm_i32_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_copy_s_h : GCCBuiltin<"__builtin_msa_copy_s_h">, + Intrinsic<[llvm_i32_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_copy_s_w : GCCBuiltin<"__builtin_msa_copy_s_w">, + Intrinsic<[llvm_i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; + +def int_mips_copy_u_b : GCCBuiltin<"__builtin_msa_copy_u_b">, + Intrinsic<[llvm_i32_ty], [llvm_v16i8_ty, llvm_i32_ty], []>; +def int_mips_copy_u_h : GCCBuiltin<"__builtin_msa_copy_u_h">, + Intrinsic<[llvm_i32_ty], [llvm_v8i16_ty, llvm_i32_ty], []>; +def int_mips_copy_u_w : GCCBuiltin<"__builtin_msa_copy_u_w">, + Intrinsic<[llvm_i32_ty], [llvm_v4i32_ty, llvm_i32_ty], []>; + +def int_mips_div_s_b : GCCBuiltin<"__builtin_msa_div_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_div_s_h : GCCBuiltin<"__builtin_msa_div_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_div_s_w : GCCBuiltin<"__builtin_msa_div_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_div_s_d : GCCBuiltin<"__builtin_msa_div_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_div_u_b : GCCBuiltin<"__builtin_msa_div_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_div_u_h : GCCBuiltin<"__builtin_msa_div_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_div_u_w : GCCBuiltin<"__builtin_msa_div_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_div_u_d : GCCBuiltin<"__builtin_msa_div_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_dotp_s_b : GCCBuiltin<"__builtin_msa_dotp_s_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dotp_s_h : GCCBuiltin<"__builtin_msa_dotp_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dotp_s_w : GCCBuiltin<"__builtin_msa_dotp_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_dotp_s_d : GCCBuiltin<"__builtin_msa_dotp_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_dotp_u_b : GCCBuiltin<"__builtin_msa_dotp_u_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dotp_u_h : GCCBuiltin<"__builtin_msa_dotp_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dotp_u_w : GCCBuiltin<"__builtin_msa_dotp_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_dotp_u_d : GCCBuiltin<"__builtin_msa_dotp_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_dpadd_s_h : GCCBuiltin<"__builtin_msa_dpadd_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dpadd_s_w : GCCBuiltin<"__builtin_msa_dpadd_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dpadd_s_d : GCCBuiltin<"__builtin_msa_dpadd_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v4i32_ty, llvm_v4i32_ty], []>; + +def int_mips_dpadd_u_h : GCCBuiltin<"__builtin_msa_dpadd_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dpadd_u_w : GCCBuiltin<"__builtin_msa_dpadd_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dpadd_u_d : GCCBuiltin<"__builtin_msa_dpadd_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v4i32_ty, llvm_v4i32_ty], []>; + +def int_mips_dpsub_s_h : GCCBuiltin<"__builtin_msa_dpsub_s_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dpsub_s_w : GCCBuiltin<"__builtin_msa_dpsub_s_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dpsub_s_d : GCCBuiltin<"__builtin_msa_dpsub_s_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v4i32_ty, llvm_v4i32_ty], []>; + +def int_mips_dpsub_u_h : GCCBuiltin<"__builtin_msa_dpsub_u_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_dpsub_u_w : GCCBuiltin<"__builtin_msa_dpsub_u_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_dpsub_u_d : GCCBuiltin<"__builtin_msa_dpsub_u_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v4i32_ty, llvm_v4i32_ty], []>; + +def int_mips_ilvev_b : GCCBuiltin<"__builtin_msa_ilvev_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_ilvev_h : GCCBuiltin<"__builtin_msa_ilvev_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_ilvev_w : GCCBuiltin<"__builtin_msa_ilvev_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_ilvev_d : GCCBuiltin<"__builtin_msa_ilvev_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_ilvl_b : GCCBuiltin<"__builtin_msa_ilvl_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_ilvl_h : GCCBuiltin<"__builtin_msa_ilvl_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_ilvl_w : GCCBuiltin<"__builtin_msa_ilvl_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_ilvl_d : GCCBuiltin<"__builtin_msa_ilvl_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_ilvod_b : GCCBuiltin<"__builtin_msa_ilvod_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_ilvod_h : GCCBuiltin<"__builtin_msa_ilvod_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_ilvod_w : GCCBuiltin<"__builtin_msa_ilvod_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_ilvod_d : GCCBuiltin<"__builtin_msa_ilvod_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_ilvr_b : GCCBuiltin<"__builtin_msa_ilvr_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_v16i8_ty], []>; +def int_mips_ilvr_h : GCCBuiltin<"__builtin_msa_ilvr_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_v8i16_ty], []>; +def int_mips_ilvr_w : GCCBuiltin<"__builtin_msa_ilvr_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_v4i32_ty], []>; +def int_mips_ilvr_d : GCCBuiltin<"__builtin_msa_ilvr_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_v2i64_ty, llvm_v2i64_ty], []>; + +def int_mips_insert_b : GCCBuiltin<"__builtin_msa_insert_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_v16i8_ty, llvm_i32_ty, llvm_i32_ty], []>; +def int_mips_insert_h : GCCBuiltin<"__builtin_msa_insert_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_v8i16_ty, llvm_i32_ty, llvm_i32_ty], []>; +def int_mips_insert_w : GCCBuiltin<"__builtin_msa_insert_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_v4i32_ty, llvm_i32_ty, llvm_i32_ty], []>; + +def int_mips_ldi_b : GCCBuiltin<"__builtin_msa_ldi_b">, + Intrinsic<[llvm_v16i8_ty], [llvm_i32_ty], []>; +def int_mips_ldi_h : GCCBuiltin<"__builtin_msa_ldi_h">, + Intrinsic<[llvm_v8i16_ty], [llvm_i32_ty], []>; +def int_mips_ldi_w : GCCBuiltin<"__builtin_msa_ldi_w">, + Intrinsic<[llvm_v4i32_ty], [llvm_i32_ty], []>; +def int_mips_ldi_d : GCCBuiltin<"__builtin_msa_ldi_d">, + Intrinsic<[llvm_v2i64_ty], [llvm_i32_ty], []>; + } diff --git a/lib/Target/Mips/MipsInstrInfo.td b/lib/Target/Mips/MipsInstrInfo.td index 81dbd421f53..17532074203 100644 --- a/lib/Target/Mips/MipsInstrInfo.td +++ b/lib/Target/Mips/MipsInstrInfo.td @@ -264,6 +264,10 @@ def simm16_64 : Operand; def shamt : Operand; // Unsigned Operand +def uimm5 : Operand { + let PrintMethod = "printUnsignedImm"; +} + def uimm16 : Operand { let PrintMethod = "printUnsignedImm"; } diff --git a/lib/Target/Mips/MipsMSAInstrFormats.td b/lib/Target/Mips/MipsMSAInstrFormats.td index 7cc1a88bd67..6f1e58f2c69 100644 --- a/lib/Target/Mips/MipsMSAInstrFormats.td +++ b/lib/Target/Mips/MipsMSAInstrFormats.td @@ -21,14 +21,91 @@ class PseudoMSA pattern, let Predicates = [HasMSA]; } +class MSA_BIT_B_FMT major, bits<6> minor>: MSAInst { + let Inst{25-23} = major; + let Inst{22-19} = 0b1110; + let Inst{5-0} = minor; +} + +class MSA_BIT_H_FMT major, bits<6> minor>: MSAInst { + let Inst{25-23} = major; + let Inst{22-20} = 0b110; + let Inst{5-0} = minor; +} + +class MSA_BIT_W_FMT major, bits<6> minor>: MSAInst { + let Inst{25-23} = major; + let Inst{22-21} = 0b10; + let Inst{5-0} = minor; +} + +class MSA_BIT_D_FMT major, bits<6> minor>: MSAInst { + let Inst{25-23} = major; + let Inst{22} = 0b0; + let Inst{5-0} = minor; +} + +class MSA_2R_FMT major, bits<2> df, bits<6> minor>: MSAInst { + let Inst{25-18} = major; + let Inst{17-16} = df; + let Inst{5-0} = minor; +} + +class MSA_2RF_FMT major, bits<1> df, bits<6> minor>: MSAInst { + let Inst{25-17} = major; + let Inst{16} = df; + let Inst{5-0} = minor; +} + class MSA_3R_FMT major, bits<2> df, bits<6> minor>: MSAInst { let Inst{25-23} = major; let Inst{22-21} = df; let Inst{5-0} = minor; } +class MSA_3RF_FMT major, bits<1> df, bits<6> minor>: MSAInst { + let Inst{25-22} = major; + let Inst{21} = df; + let Inst{5-0} = minor; +} + +class MSA_ELM_B_FMT major, bits<6> minor>: MSAInst { + let Inst{25-22} = major; + let Inst{21-20} = 0b00; + let Inst{5-0} = minor; +} + +class MSA_ELM_H_FMT major, bits<6> minor>: MSAInst { + let Inst{25-22} = major; + let Inst{21-19} = 0b100; + let Inst{5-0} = minor; +} + +class MSA_ELM_W_FMT major, bits<6> minor>: MSAInst { + let Inst{25-22} = major; + let Inst{21-18} = 0b1100; + let Inst{5-0} = minor; +} + +class MSA_ELM_D_FMT major, bits<6> minor>: MSAInst { + let Inst{25-22} = major; + let Inst{21-17} = 0b11100; + let Inst{5-0} = minor; +} + class MSA_I5_FMT major, bits<2> df, bits<6> minor>: MSAInst { let Inst{25-23} = major; let Inst{22-21} = df; let Inst{5-0} = minor; } + +class MSA_I8_FMT major, bits<6> minor>: MSAInst { + let Inst{25-24} = major; + let Inst{5-0} = minor; +} + +class MSA_I10_FMT major, bits<2> df, bits<6> minor>: MSAInst { + let Inst{25-23} = major; + let Inst{22-21} = df; + let Inst{5-0} = minor; +} diff --git a/lib/Target/Mips/MipsMSAInstrInfo.td b/lib/Target/Mips/MipsMSAInstrInfo.td index 24482c07524..ea8938b8dcc 100644 --- a/lib/Target/Mips/MipsMSAInstrInfo.td +++ b/lib/Target/Mips/MipsMSAInstrInfo.td @@ -11,17 +11,887 @@ // //===----------------------------------------------------------------------===// +def immSExt5 : ImmLeaf(Imm);}]>; +def immSExt10: ImmLeaf(Imm);}]>; + +def uimm3 : Operand { + let PrintMethod = "printUnsignedImm"; +} + +def uimm4 : Operand { + let PrintMethod = "printUnsignedImm"; +} + +def uimm6 : Operand { + let PrintMethod = "printUnsignedImm"; +} + +def uimm8 : Operand { + let PrintMethod = "printUnsignedImm"; +} + +def simm5 : Operand; + +def simm10 : Operand; + // Instruction encoding. +class ADD_A_B_ENC : MSA_3R_FMT<0b000, 0b00, 0b010000>; +class ADD_A_H_ENC : MSA_3R_FMT<0b000, 0b01, 0b010000>; +class ADD_A_W_ENC : MSA_3R_FMT<0b000, 0b10, 0b010000>; +class ADD_A_D_ENC : MSA_3R_FMT<0b000, 0b11, 0b010000>; + +class ADDS_A_B_ENC : MSA_3R_FMT<0b001, 0b00, 0b010000>; +class ADDS_A_H_ENC : MSA_3R_FMT<0b001, 0b01, 0b010000>; +class ADDS_A_W_ENC : MSA_3R_FMT<0b001, 0b10, 0b010000>; +class ADDS_A_D_ENC : MSA_3R_FMT<0b001, 0b11, 0b010000>; + +class ADDS_S_B_ENC : MSA_3R_FMT<0b010, 0b00, 0b010000>; +class ADDS_S_H_ENC : MSA_3R_FMT<0b010, 0b01, 0b010000>; +class ADDS_S_W_ENC : MSA_3R_FMT<0b010, 0b10, 0b010000>; +class ADDS_S_D_ENC : MSA_3R_FMT<0b010, 0b11, 0b010000>; + +class ADDS_U_B_ENC : MSA_3R_FMT<0b011, 0b00, 0b010000>; +class ADDS_U_H_ENC : MSA_3R_FMT<0b011, 0b01, 0b010000>; +class ADDS_U_W_ENC : MSA_3R_FMT<0b011, 0b10, 0b010000>; +class ADDS_U_D_ENC : MSA_3R_FMT<0b011, 0b11, 0b010000>; + +class ADDV_B_ENC : MSA_3R_FMT<0b000, 0b00, 0b001110>; +class ADDV_H_ENC : MSA_3R_FMT<0b000, 0b01, 0b001110>; +class ADDV_W_ENC : MSA_3R_FMT<0b000, 0b10, 0b001110>; +class ADDV_D_ENC : MSA_3R_FMT<0b000, 0b11, 0b001110>; + +class ADDVI_B_ENC : MSA_I5_FMT<0b000, 0b00, 0b000110>; +class ADDVI_H_ENC : MSA_I5_FMT<0b000, 0b01, 0b000110>; +class ADDVI_W_ENC : MSA_I5_FMT<0b000, 0b10, 0b000110>; +class ADDVI_D_ENC : MSA_I5_FMT<0b000, 0b11, 0b000110>; + +class ANDI_B_ENC : MSA_I8_FMT<0b00, 0b000000>; + +class ASUB_S_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b010001>; +class ASUB_S_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b010001>; +class ASUB_S_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b010001>; +class ASUB_S_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b010001>; + +class ASUB_U_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b010001>; +class ASUB_U_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b010001>; +class ASUB_U_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b010001>; +class ASUB_U_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b010001>; + +class AVE_S_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b010000>; +class AVE_S_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b010000>; +class AVE_S_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b010000>; +class AVE_S_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b010000>; + +class AVE_U_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b010000>; +class AVE_U_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b010000>; +class AVE_U_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b010000>; +class AVE_U_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b010000>; + +class AVER_S_B_ENC : MSA_3R_FMT<0b110, 0b00, 0b010000>; +class AVER_S_H_ENC : MSA_3R_FMT<0b110, 0b01, 0b010000>; +class AVER_S_W_ENC : MSA_3R_FMT<0b110, 0b10, 0b010000>; +class AVER_S_D_ENC : MSA_3R_FMT<0b110, 0b11, 0b010000>; + +class AVER_U_B_ENC : MSA_3R_FMT<0b111, 0b00, 0b010000>; +class AVER_U_H_ENC : MSA_3R_FMT<0b111, 0b01, 0b010000>; +class AVER_U_W_ENC : MSA_3R_FMT<0b111, 0b10, 0b010000>; +class AVER_U_D_ENC : MSA_3R_FMT<0b111, 0b11, 0b010000>; + +class BCLR_B_ENC : MSA_3R_FMT<0b011, 0b00, 0b001101>; +class BCLR_H_ENC : MSA_3R_FMT<0b011, 0b01, 0b001101>; +class BCLR_W_ENC : MSA_3R_FMT<0b011, 0b10, 0b001101>; +class BCLR_D_ENC : MSA_3R_FMT<0b011, 0b11, 0b001101>; + +class BCLRI_B_ENC : MSA_BIT_B_FMT<0b011, 0b001001>; +class BCLRI_H_ENC : MSA_BIT_H_FMT<0b011, 0b001001>; +class BCLRI_W_ENC : MSA_BIT_W_FMT<0b011, 0b001001>; +class BCLRI_D_ENC : MSA_BIT_D_FMT<0b011, 0b001001>; + +class BINSL_B_ENC : MSA_3R_FMT<0b110, 0b00, 0b001101>; +class BINSL_H_ENC : MSA_3R_FMT<0b110, 0b01, 0b001101>; +class BINSL_W_ENC : MSA_3R_FMT<0b110, 0b10, 0b001101>; +class BINSL_D_ENC : MSA_3R_FMT<0b110, 0b11, 0b001101>; + +class BINSLI_B_ENC : MSA_BIT_B_FMT<0b110, 0b001001>; +class BINSLI_H_ENC : MSA_BIT_H_FMT<0b110, 0b001001>; +class BINSLI_W_ENC : MSA_BIT_W_FMT<0b110, 0b001001>; +class BINSLI_D_ENC : MSA_BIT_D_FMT<0b110, 0b001001>; + +class BINSR_B_ENC : MSA_3R_FMT<0b111, 0b00, 0b001101>; +class BINSR_H_ENC : MSA_3R_FMT<0b111, 0b01, 0b001101>; +class BINSR_W_ENC : MSA_3R_FMT<0b111, 0b10, 0b001101>; +class BINSR_D_ENC : MSA_3R_FMT<0b111, 0b11, 0b001101>; + +class BINSRI_B_ENC : MSA_BIT_B_FMT<0b111, 0b001001>; +class BINSRI_H_ENC : MSA_BIT_H_FMT<0b111, 0b001001>; +class BINSRI_W_ENC : MSA_BIT_W_FMT<0b111, 0b001001>; +class BINSRI_D_ENC : MSA_BIT_D_FMT<0b111, 0b001001>; + +class BMNZI_B_ENC : MSA_I8_FMT<0b00, 0b000001>; + +class BMZI_B_ENC : MSA_I8_FMT<0b01, 0b000001>; + +class BNEG_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b001101>; +class BNEG_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b001101>; +class BNEG_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b001101>; +class BNEG_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b001101>; + +class BNEGI_B_ENC : MSA_BIT_B_FMT<0b101, 0b001001>; +class BNEGI_H_ENC : MSA_BIT_H_FMT<0b101, 0b001001>; +class BNEGI_W_ENC : MSA_BIT_W_FMT<0b101, 0b001001>; +class BNEGI_D_ENC : MSA_BIT_D_FMT<0b101, 0b001001>; + +class BSELI_B_ENC : MSA_I8_FMT<0b10, 0b000001>; + +class BSET_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b001101>; +class BSET_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b001101>; +class BSET_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b001101>; +class BSET_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b001101>; + +class BSETI_B_ENC : MSA_BIT_B_FMT<0b100, 0b001001>; +class BSETI_H_ENC : MSA_BIT_H_FMT<0b100, 0b001001>; +class BSETI_W_ENC : MSA_BIT_W_FMT<0b100, 0b001001>; +class BSETI_D_ENC : MSA_BIT_D_FMT<0b100, 0b001001>; + +class CEQ_B_ENC : MSA_3R_FMT<0b000, 0b00, 0b001111>; +class CEQ_H_ENC : MSA_3R_FMT<0b000, 0b01, 0b001111>; +class CEQ_W_ENC : MSA_3R_FMT<0b000, 0b10, 0b001111>; +class CEQ_D_ENC : MSA_3R_FMT<0b000, 0b11, 0b001111>; + +class CEQI_B_ENC : MSA_I5_FMT<0b000, 0b00, 0b000111>; +class CEQI_H_ENC : MSA_I5_FMT<0b000, 0b01, 0b000111>; +class CEQI_W_ENC : MSA_I5_FMT<0b000, 0b10, 0b000111>; +class CEQI_D_ENC : MSA_I5_FMT<0b000, 0b11, 0b000111>; + +class CLE_S_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b001111>; +class CLE_S_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b001111>; +class CLE_S_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b001111>; +class CLE_S_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b001111>; + +class CLE_U_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b001111>; +class CLE_U_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b001111>; +class CLE_U_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b001111>; +class CLE_U_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b001111>; + +class CLEI_S_B_ENC : MSA_I5_FMT<0b100, 0b00, 0b000111>; +class CLEI_S_H_ENC : MSA_I5_FMT<0b100, 0b01, 0b000111>; +class CLEI_S_W_ENC : MSA_I5_FMT<0b100, 0b10, 0b000111>; +class CLEI_S_D_ENC : MSA_I5_FMT<0b100, 0b11, 0b000111>; + +class CLEI_U_B_ENC : MSA_I5_FMT<0b101, 0b00, 0b000111>; +class CLEI_U_H_ENC : MSA_I5_FMT<0b101, 0b01, 0b000111>; +class CLEI_U_W_ENC : MSA_I5_FMT<0b101, 0b10, 0b000111>; +class CLEI_U_D_ENC : MSA_I5_FMT<0b101, 0b11, 0b000111>; + +class CLT_S_B_ENC : MSA_3R_FMT<0b010, 0b00, 0b001111>; +class CLT_S_H_ENC : MSA_3R_FMT<0b010, 0b01, 0b001111>; +class CLT_S_W_ENC : MSA_3R_FMT<0b010, 0b10, 0b001111>; +class CLT_S_D_ENC : MSA_3R_FMT<0b010, 0b11, 0b001111>; + +class CLT_U_B_ENC : MSA_3R_FMT<0b011, 0b00, 0b001111>; +class CLT_U_H_ENC : MSA_3R_FMT<0b011, 0b01, 0b001111>; +class CLT_U_W_ENC : MSA_3R_FMT<0b011, 0b10, 0b001111>; +class CLT_U_D_ENC : MSA_3R_FMT<0b011, 0b11, 0b001111>; + +class CLTI_S_B_ENC : MSA_I5_FMT<0b010, 0b00, 0b000111>; +class CLTI_S_H_ENC : MSA_I5_FMT<0b010, 0b01, 0b000111>; +class CLTI_S_W_ENC : MSA_I5_FMT<0b010, 0b10, 0b000111>; +class CLTI_S_D_ENC : MSA_I5_FMT<0b010, 0b11, 0b000111>; + +class CLTI_U_B_ENC : MSA_I5_FMT<0b011, 0b00, 0b000111>; +class CLTI_U_H_ENC : MSA_I5_FMT<0b011, 0b01, 0b000111>; +class CLTI_U_W_ENC : MSA_I5_FMT<0b011, 0b10, 0b000111>; +class CLTI_U_D_ENC : MSA_I5_FMT<0b011, 0b11, 0b000111>; + +class COPY_S_B_ENC : MSA_ELM_B_FMT<0b0010, 0b011001>; +class COPY_S_H_ENC : MSA_ELM_H_FMT<0b0010, 0b011001>; +class COPY_S_W_ENC : MSA_ELM_W_FMT<0b0010, 0b011001>; + +class COPY_U_B_ENC : MSA_ELM_B_FMT<0b0011, 0b011001>; +class COPY_U_H_ENC : MSA_ELM_H_FMT<0b0011, 0b011001>; +class COPY_U_W_ENC : MSA_ELM_W_FMT<0b0011, 0b011001>; + +class DIV_S_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b010010>; +class DIV_S_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b010010>; +class DIV_S_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b010010>; +class DIV_S_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b010010>; + +class DIV_U_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b010010>; +class DIV_U_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b010010>; +class DIV_U_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b010010>; +class DIV_U_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b010010>; + +class DOTP_S_B_ENC : MSA_3R_FMT<0b000, 0b00, 0b010011>; +class DOTP_S_H_ENC : MSA_3R_FMT<0b000, 0b01, 0b010011>; +class DOTP_S_W_ENC : MSA_3R_FMT<0b000, 0b10, 0b010011>; +class DOTP_S_D_ENC : MSA_3R_FMT<0b000, 0b11, 0b010011>; + +class DOTP_U_B_ENC : MSA_3R_FMT<0b001, 0b00, 0b010011>; +class DOTP_U_H_ENC : MSA_3R_FMT<0b001, 0b01, 0b010011>; +class DOTP_U_W_ENC : MSA_3R_FMT<0b001, 0b10, 0b010011>; +class DOTP_U_D_ENC : MSA_3R_FMT<0b001, 0b11, 0b010011>; + +class DPADD_S_H_ENC : MSA_3R_FMT<0b010, 0b01, 0b010011>; +class DPADD_S_W_ENC : MSA_3R_FMT<0b010, 0b10, 0b010011>; +class DPADD_S_D_ENC : MSA_3R_FMT<0b010, 0b11, 0b010011>; + +class DPADD_U_H_ENC : MSA_3R_FMT<0b011, 0b01, 0b010011>; +class DPADD_U_W_ENC : MSA_3R_FMT<0b011, 0b10, 0b010011>; +class DPADD_U_D_ENC : MSA_3R_FMT<0b011, 0b11, 0b010011>; + +class DPSUB_S_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b010011>; +class DPSUB_S_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b010011>; +class DPSUB_S_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b010011>; + +class DPSUB_U_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b010011>; +class DPSUB_U_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b010011>; +class DPSUB_U_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b010011>; + +class ILVEV_B_ENC : MSA_3R_FMT<0b110, 0b00, 0b010100>; +class ILVEV_H_ENC : MSA_3R_FMT<0b110, 0b01, 0b010100>; +class ILVEV_W_ENC : MSA_3R_FMT<0b110, 0b10, 0b010100>; +class ILVEV_D_ENC : MSA_3R_FMT<0b110, 0b11, 0b010100>; + +class ILVL_B_ENC : MSA_3R_FMT<0b100, 0b00, 0b010100>; +class ILVL_H_ENC : MSA_3R_FMT<0b100, 0b01, 0b010100>; +class ILVL_W_ENC : MSA_3R_FMT<0b100, 0b10, 0b010100>; +class ILVL_D_ENC : MSA_3R_FMT<0b100, 0b11, 0b010100>; + +class ILVOD_B_ENC : MSA_3R_FMT<0b111, 0b00, 0b010100>; +class ILVOD_H_ENC : MSA_3R_FMT<0b111, 0b01, 0b010100>; +class ILVOD_W_ENC : MSA_3R_FMT<0b111, 0b10, 0b010100>; +class ILVOD_D_ENC : MSA_3R_FMT<0b111, 0b11, 0b010100>; + +class ILVR_B_ENC : MSA_3R_FMT<0b101, 0b00, 0b010100>; +class ILVR_H_ENC : MSA_3R_FMT<0b101, 0b01, 0b010100>; +class ILVR_W_ENC : MSA_3R_FMT<0b101, 0b10, 0b010100>; +class ILVR_D_ENC : MSA_3R_FMT<0b101, 0b11, 0b010100>; + +class INSERT_B_ENC : MSA_ELM_B_FMT<0b0100, 0b011001>; +class INSERT_H_ENC : MSA_ELM_H_FMT<0b0100, 0b011001>; +class INSERT_W_ENC : MSA_ELM_W_FMT<0b0100, 0b011001>; + class LD_B_ENC : MSA_I5_FMT<0b110, 0b00, 0b000111>; class LD_H_ENC : MSA_I5_FMT<0b110, 0b01, 0b000111>; class LD_W_ENC : MSA_I5_FMT<0b110, 0b10, 0b000111>; class LD_D_ENC : MSA_I5_FMT<0b110, 0b11, 0b000111>; + +class LDI_B_ENC : MSA_I10_FMT<0b010, 0b00, 0b001100>; +class LDI_H_ENC : MSA_I10_FMT<0b010, 0b01, 0b001100>; +class LDI_W_ENC : MSA_I10_FMT<0b010, 0b10, 0b001100>; +class LDI_D_ENC : MSA_I10_FMT<0b010, 0b11, 0b001100>; + class ST_B_ENC : MSA_I5_FMT<0b111, 0b00, 0b000111>; class ST_H_ENC : MSA_I5_FMT<0b111, 0b01, 0b000111>; class ST_W_ENC : MSA_I5_FMT<0b111, 0b10, 0b000111>; class ST_D_ENC : MSA_I5_FMT<0b111, 0b11, 0b000111>; // Instruction desc. +class MSA_BIT_D_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm6:$u6); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u6"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt6:$u6))]; + InstrItinClass Itinerary = itin; +} + +class MSA_BIT_W_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm5:$u5); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u5"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt5:$u5))]; + InstrItinClass Itinerary = itin; +} + +class MSA_BIT_H_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm4:$u4); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u4"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt4:$u4))]; + InstrItinClass Itinerary = itin; +} + +class MSA_BIT_B_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm3:$u3); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u3"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt3:$u3))]; + InstrItinClass Itinerary = itin; +} + +class MSA_COPY_DESC_BASE { + dag OutOperandList = (outs RCD:$rd); + dag InOperandList = (ins RCWS:$ws, uimm6:$n); + string AsmString = !strconcat(instr_asm, "\t$rd, $ws[$n]"); + list Pattern = [(set RCD:$rd, (OpNode RCWS:$ws, immZExt6:$n))]; + InstrItinClass Itinerary = itin; +} + +class MSA_I5_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm5:$u5); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u5"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt5:$u5))]; + InstrItinClass Itinerary = itin; +} + +class MSA_SI5_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, simm5:$s5); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $s5"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immSExt5:$s5))]; + InstrItinClass Itinerary = itin; +} + +class MSA_I8_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, uimm8:$u8); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $u8"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, immZExt8:$u8))]; + InstrItinClass Itinerary = itin; +} + +class MSA_I10_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins simm10:$i10); + string AsmString = !strconcat(instr_asm, "\t$wd, $i10"); + list Pattern = [(set RCWD:$wd, (OpNode immSExt10:$i10))]; + InstrItinClass Itinerary = itin; +} + +class MSA_2R_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws))]; + InstrItinClass Itinerary = itin; +} + +class MSA_2RF_DESC_BASE : + MSA_2R_DESC_BASE; + + +class MSA_3R_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWS:$ws, RCWT:$wt); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $wt"); + list Pattern = [(set RCWD:$wd, (OpNode RCWS:$ws, RCWT:$wt))]; + InstrItinClass Itinerary = itin; +} + +class MSA_3R_4R_DESC_BASE { + dag OutOperandList = (outs RCWD:$wd); + dag InOperandList = (ins RCWD:$wd_in, RCWS:$ws, RCWT:$wt); + string AsmString = !strconcat(instr_asm, "\t$wd, $ws, $wt"); + list Pattern = [(set RCWD:$wd, + (OpNode RCWD:$wd_in, RCWS:$ws, RCWT:$wt))]; + InstrItinClass Itinerary = itin; + string Constraints = "$wd = $wd_in"; +} + +class MSA_3RF_DESC_BASE : + MSA_3R_DESC_BASE; + +class MSA_3RF_4RF_DESC_BASE : + MSA_3R_4R_DESC_BASE; + +class MSA_INSERT_DESC_BASE { + dag OutOperandList = (outs RCD:$wd); + dag InOperandList = (ins RCD:$wd_in, uimm6:$n, RCWS:$rs); + string AsmString = !strconcat(instr_asm, "\t$wd[$n], $rs"); + list Pattern = [(set RCD:$wd, (OpNode RCD:$wd_in, + immZExt6:$n, + RCWS:$rs))]; + InstrItinClass Itinerary = itin; + string Constraints = "$wd = $wd_in"; +} + +class ADD_A_B_DESC : MSA_3R_DESC_BASE<"add_a.b", int_mips_add_a_b, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADD_A_H_DESC : MSA_3R_DESC_BASE<"add_a.h", int_mips_add_a_h, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADD_A_W_DESC : MSA_3R_DESC_BASE<"add_a.w", int_mips_add_a_w, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADD_A_D_DESC : MSA_3R_DESC_BASE<"add_a.d", int_mips_add_a_d, NoItinerary, + MSA128, MSA128>, IsCommutable; + +class ADDS_A_B_DESC : MSA_3R_DESC_BASE<"adds_a.b", int_mips_adds_a_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_A_H_DESC : MSA_3R_DESC_BASE<"adds_a.h", int_mips_adds_a_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_A_W_DESC : MSA_3R_DESC_BASE<"adds_a.w", int_mips_adds_a_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_A_D_DESC : MSA_3R_DESC_BASE<"adds_a.d", int_mips_adds_a_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class ADDS_S_B_DESC : MSA_3R_DESC_BASE<"adds_s.b", int_mips_adds_s_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_S_H_DESC : MSA_3R_DESC_BASE<"adds_s.h", int_mips_adds_s_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_S_W_DESC : MSA_3R_DESC_BASE<"adds_s.w", int_mips_adds_s_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_S_D_DESC : MSA_3R_DESC_BASE<"adds_s.d", int_mips_adds_s_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class ADDS_U_B_DESC : MSA_3R_DESC_BASE<"adds_u.b", int_mips_adds_u_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_U_H_DESC : MSA_3R_DESC_BASE<"adds_u.h", int_mips_adds_u_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_U_W_DESC : MSA_3R_DESC_BASE<"adds_u.w", int_mips_adds_u_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class ADDS_U_D_DESC : MSA_3R_DESC_BASE<"adds_u.d", int_mips_adds_u_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class ADDV_B_DESC : MSA_3R_DESC_BASE<"addv.b", int_mips_addv_b, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADDV_H_DESC : MSA_3R_DESC_BASE<"addv.h", int_mips_addv_h, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADDV_W_DESC : MSA_3R_DESC_BASE<"addv.w", int_mips_addv_w, NoItinerary, + MSA128, MSA128>, IsCommutable; +class ADDV_D_DESC : MSA_3R_DESC_BASE<"addv.d", int_mips_addv_d, NoItinerary, + MSA128, MSA128>, IsCommutable; + +class ADDVI_B_DESC : MSA_I5_DESC_BASE<"addvi.b", int_mips_addvi_b, NoItinerary, + MSA128, MSA128>; +class ADDVI_H_DESC : MSA_I5_DESC_BASE<"addvi.h", int_mips_addvi_h, NoItinerary, + MSA128, MSA128>; +class ADDVI_W_DESC : MSA_I5_DESC_BASE<"addvi.w", int_mips_addvi_w, NoItinerary, + MSA128, MSA128>; +class ADDVI_D_DESC : MSA_I5_DESC_BASE<"addvi.d", int_mips_addvi_d, NoItinerary, + MSA128, MSA128>; + +class ANDI_B_DESC : MSA_I8_DESC_BASE<"andi.b", int_mips_andi_b, NoItinerary, + MSA128, MSA128>; + +class ASUB_S_B_DESC : MSA_3R_DESC_BASE<"asub_s.b", int_mips_asub_s_b, + NoItinerary, MSA128, MSA128>; +class ASUB_S_H_DESC : MSA_3R_DESC_BASE<"asub_s.h", int_mips_asub_s_h, + NoItinerary, MSA128, MSA128>; +class ASUB_S_W_DESC : MSA_3R_DESC_BASE<"asub_s.w", int_mips_asub_s_w, + NoItinerary, MSA128, MSA128>; +class ASUB_S_D_DESC : MSA_3R_DESC_BASE<"asub_s.d", int_mips_asub_s_d, + NoItinerary, MSA128, MSA128>; + +class ASUB_U_B_DESC : MSA_3R_DESC_BASE<"asub_u.b", int_mips_asub_u_b, + NoItinerary, MSA128, MSA128>; +class ASUB_U_H_DESC : MSA_3R_DESC_BASE<"asub_u.h", int_mips_asub_u_h, + NoItinerary, MSA128, MSA128>; +class ASUB_U_W_DESC : MSA_3R_DESC_BASE<"asub_u.w", int_mips_asub_u_w, + NoItinerary, MSA128, MSA128>; +class ASUB_U_D_DESC : MSA_3R_DESC_BASE<"asub_u.d", int_mips_asub_u_d, + NoItinerary, MSA128, MSA128>; + +class AVE_S_B_DESC : MSA_3R_DESC_BASE<"ave_s.b", int_mips_ave_s_b, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_S_H_DESC : MSA_3R_DESC_BASE<"ave_s.h", int_mips_ave_s_h, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_S_W_DESC : MSA_3R_DESC_BASE<"ave_s.w", int_mips_ave_s_w, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_S_D_DESC : MSA_3R_DESC_BASE<"ave_s.d", int_mips_ave_s_d, NoItinerary, + MSA128, MSA128>, IsCommutable; + +class AVE_U_B_DESC : MSA_3R_DESC_BASE<"ave_u.b", int_mips_ave_u_b, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_U_H_DESC : MSA_3R_DESC_BASE<"ave_u.h", int_mips_ave_u_h, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_U_W_DESC : MSA_3R_DESC_BASE<"ave_u.w", int_mips_ave_u_w, NoItinerary, + MSA128, MSA128>, IsCommutable; +class AVE_U_D_DESC : MSA_3R_DESC_BASE<"ave_u.d", int_mips_ave_u_d, NoItinerary, + MSA128, MSA128>, IsCommutable; + +class AVER_S_B_DESC : MSA_3R_DESC_BASE<"aver_s.b", int_mips_aver_s_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_S_H_DESC : MSA_3R_DESC_BASE<"aver_s.h", int_mips_aver_s_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_S_W_DESC : MSA_3R_DESC_BASE<"aver_s.w", int_mips_aver_s_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_S_D_DESC : MSA_3R_DESC_BASE<"aver_s.d", int_mips_aver_s_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class AVER_U_B_DESC : MSA_3R_DESC_BASE<"aver_u.b", int_mips_aver_u_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_U_H_DESC : MSA_3R_DESC_BASE<"aver_u.h", int_mips_aver_u_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_U_W_DESC : MSA_3R_DESC_BASE<"aver_u.w", int_mips_aver_u_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class AVER_U_D_DESC : MSA_3R_DESC_BASE<"aver_u.d", int_mips_aver_u_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class BCLR_B_DESC : MSA_3R_DESC_BASE<"bclr.b", int_mips_bclr_b, NoItinerary, + MSA128, MSA128>; +class BCLR_H_DESC : MSA_3R_DESC_BASE<"bclr.h", int_mips_bclr_h, NoItinerary, + MSA128, MSA128>; +class BCLR_W_DESC : MSA_3R_DESC_BASE<"bclr.w", int_mips_bclr_w, NoItinerary, + MSA128, MSA128>; +class BCLR_D_DESC : MSA_3R_DESC_BASE<"bclr.d", int_mips_bclr_d, NoItinerary, + MSA128, MSA128>; + +class BCLRI_B_DESC : MSA_BIT_B_DESC_BASE<"bclri.b", int_mips_bclri_b, + NoItinerary, MSA128, MSA128>; +class BCLRI_H_DESC : MSA_BIT_H_DESC_BASE<"bclri.h", int_mips_bclri_h, + NoItinerary, MSA128, MSA128>; +class BCLRI_W_DESC : MSA_BIT_W_DESC_BASE<"bclri.w", int_mips_bclri_w, + NoItinerary, MSA128, MSA128>; +class BCLRI_D_DESC : MSA_BIT_D_DESC_BASE<"bclri.d", int_mips_bclri_d, + NoItinerary, MSA128, MSA128>; + +class BINSL_B_DESC : MSA_3R_DESC_BASE<"binsl.b", int_mips_binsl_b, NoItinerary, + MSA128, MSA128>; +class BINSL_H_DESC : MSA_3R_DESC_BASE<"binsl.h", int_mips_binsl_h, NoItinerary, + MSA128, MSA128>; +class BINSL_W_DESC : MSA_3R_DESC_BASE<"binsl.w", int_mips_binsl_w, NoItinerary, + MSA128, MSA128>; +class BINSL_D_DESC : MSA_3R_DESC_BASE<"binsl.d", int_mips_binsl_d, NoItinerary, + MSA128, MSA128>; + +class BINSLI_B_DESC : MSA_BIT_B_DESC_BASE<"binsli.b", int_mips_binsli_b, + NoItinerary, MSA128, MSA128>; +class BINSLI_H_DESC : MSA_BIT_H_DESC_BASE<"binsli.h", int_mips_binsli_h, + NoItinerary, MSA128, MSA128>; +class BINSLI_W_DESC : MSA_BIT_W_DESC_BASE<"binsli.w", int_mips_binsli_w, + NoItinerary, MSA128, MSA128>; +class BINSLI_D_DESC : MSA_BIT_D_DESC_BASE<"binsli.d", int_mips_binsli_d, + NoItinerary, MSA128, MSA128>; + +class BINSR_B_DESC : MSA_3R_DESC_BASE<"binsr.b", int_mips_binsr_b, NoItinerary, + MSA128, MSA128>; +class BINSR_H_DESC : MSA_3R_DESC_BASE<"binsr.h", int_mips_binsr_h, NoItinerary, + MSA128, MSA128>; +class BINSR_W_DESC : MSA_3R_DESC_BASE<"binsr.w", int_mips_binsr_w, NoItinerary, + MSA128, MSA128>; +class BINSR_D_DESC : MSA_3R_DESC_BASE<"binsr.d", int_mips_binsr_d, NoItinerary, + MSA128, MSA128>; + +class BINSRI_B_DESC : MSA_BIT_B_DESC_BASE<"binsri.b", int_mips_binsri_b, + NoItinerary, MSA128, MSA128>; +class BINSRI_H_DESC : MSA_BIT_H_DESC_BASE<"binsri.h", int_mips_binsri_h, + NoItinerary, MSA128, MSA128>; +class BINSRI_W_DESC : MSA_BIT_W_DESC_BASE<"binsri.w", int_mips_binsri_w, + NoItinerary, MSA128, MSA128>; +class BINSRI_D_DESC : MSA_BIT_D_DESC_BASE<"binsri.d", int_mips_binsri_d, + NoItinerary, MSA128, MSA128>; + +class BMNZI_B_DESC : MSA_I8_DESC_BASE<"bmnzi.b", int_mips_bmnzi_b, NoItinerary, + MSA128, MSA128>; + +class BMZI_B_DESC : MSA_I8_DESC_BASE<"bmzi.b", int_mips_bmzi_b, NoItinerary, + MSA128, MSA128>; + +class BNEG_B_DESC : MSA_3R_DESC_BASE<"bneg.b", int_mips_bneg_b, NoItinerary, + MSA128, MSA128>; +class BNEG_H_DESC : MSA_3R_DESC_BASE<"bneg.h", int_mips_bneg_h, NoItinerary, + MSA128, MSA128>; +class BNEG_W_DESC : MSA_3R_DESC_BASE<"bneg.w", int_mips_bneg_w, NoItinerary, + MSA128, MSA128>; +class BNEG_D_DESC : MSA_3R_DESC_BASE<"bneg.d", int_mips_bneg_d, NoItinerary, + MSA128, MSA128>; + +class BNEGI_B_DESC : MSA_BIT_B_DESC_BASE<"bnegi.b", int_mips_bnegi_b, + NoItinerary, MSA128, MSA128>; +class BNEGI_H_DESC : MSA_BIT_H_DESC_BASE<"bnegi.h", int_mips_bnegi_h, + NoItinerary, MSA128, MSA128>; +class BNEGI_W_DESC : MSA_BIT_W_DESC_BASE<"bnegi.w", int_mips_bnegi_w, + NoItinerary, MSA128, MSA128>; +class BNEGI_D_DESC : MSA_BIT_D_DESC_BASE<"bnegi.d", int_mips_bnegi_d, + NoItinerary, MSA128, MSA128>; + +class BSELI_B_DESC : MSA_I8_DESC_BASE<"bseli.b", int_mips_bseli_b, NoItinerary, + MSA128, MSA128>; + +class BSET_B_DESC : MSA_3R_DESC_BASE<"bset.b", int_mips_bset_b, NoItinerary, + MSA128, MSA128>; +class BSET_H_DESC : MSA_3R_DESC_BASE<"bset.h", int_mips_bset_h, NoItinerary, + MSA128, MSA128>; +class BSET_W_DESC : MSA_3R_DESC_BASE<"bset.w", int_mips_bset_w, NoItinerary, + MSA128, MSA128>; +class BSET_D_DESC : MSA_3R_DESC_BASE<"bset.d", int_mips_bset_d, NoItinerary, + MSA128, MSA128>; + +class BSETI_B_DESC : MSA_BIT_B_DESC_BASE<"bseti.b", int_mips_bseti_b, + NoItinerary, MSA128, MSA128>; +class BSETI_H_DESC : MSA_BIT_H_DESC_BASE<"bseti.h", int_mips_bseti_h, + NoItinerary, MSA128, MSA128>; +class BSETI_W_DESC : MSA_BIT_W_DESC_BASE<"bseti.w", int_mips_bseti_w, + NoItinerary, MSA128, MSA128>; +class BSETI_D_DESC : MSA_BIT_D_DESC_BASE<"bseti.d", int_mips_bseti_d, + NoItinerary, MSA128, MSA128>; + +class CEQ_B_DESC : MSA_3R_DESC_BASE<"ceq.b", int_mips_ceq_b, NoItinerary, + MSA128, MSA128>, IsCommutable; +class CEQ_H_DESC : MSA_3R_DESC_BASE<"ceq.h", int_mips_ceq_h, NoItinerary, + MSA128, MSA128>, IsCommutable; +class CEQ_W_DESC : MSA_3R_DESC_BASE<"ceq.w", int_mips_ceq_w, NoItinerary, + MSA128, MSA128>, IsCommutable; +class CEQ_D_DESC : MSA_3R_DESC_BASE<"ceq.d", int_mips_ceq_d, NoItinerary, + MSA128, MSA128>, IsCommutable; + +class CEQI_B_DESC : MSA_SI5_DESC_BASE<"ceqi.b", int_mips_ceqi_b, NoItinerary, + MSA128, MSA128>; +class CEQI_H_DESC : MSA_SI5_DESC_BASE<"ceqi.h", int_mips_ceqi_h, NoItinerary, + MSA128, MSA128>; +class CEQI_W_DESC : MSA_SI5_DESC_BASE<"ceqi.w", int_mips_ceqi_w, NoItinerary, + MSA128, MSA128>; +class CEQI_D_DESC : MSA_SI5_DESC_BASE<"ceqi.d", int_mips_ceqi_d, NoItinerary, + MSA128, MSA128>; + +class CLE_S_B_DESC : MSA_3R_DESC_BASE<"cle_s.b", int_mips_cle_s_b, NoItinerary, + MSA128, MSA128>; +class CLE_S_H_DESC : MSA_3R_DESC_BASE<"cle_s.h", int_mips_cle_s_h, NoItinerary, + MSA128, MSA128>; +class CLE_S_W_DESC : MSA_3R_DESC_BASE<"cle_s.w", int_mips_cle_s_w, NoItinerary, + MSA128, MSA128>; +class CLE_S_D_DESC : MSA_3R_DESC_BASE<"cle_s.d", int_mips_cle_s_d, NoItinerary, + MSA128, MSA128>; + +class CLE_U_B_DESC : MSA_3R_DESC_BASE<"cle_u.b", int_mips_cle_u_b, NoItinerary, + MSA128, MSA128>; +class CLE_U_H_DESC : MSA_3R_DESC_BASE<"cle_u.h", int_mips_cle_u_h, NoItinerary, + MSA128, MSA128>; +class CLE_U_W_DESC : MSA_3R_DESC_BASE<"cle_u.w", int_mips_cle_u_w, NoItinerary, + MSA128, MSA128>; +class CLE_U_D_DESC : MSA_3R_DESC_BASE<"cle_u.d", int_mips_cle_u_d, NoItinerary, + MSA128, MSA128>; + +class CLEI_S_B_DESC : MSA_SI5_DESC_BASE<"clei_s.b", int_mips_clei_s_b, + NoItinerary, MSA128, MSA128>; +class CLEI_S_H_DESC : MSA_SI5_DESC_BASE<"clei_s.h", int_mips_clei_s_h, + NoItinerary, MSA128, MSA128>; +class CLEI_S_W_DESC : MSA_SI5_DESC_BASE<"clei_s.w", int_mips_clei_s_w, + NoItinerary, MSA128, MSA128>; +class CLEI_S_D_DESC : MSA_SI5_DESC_BASE<"clei_s.d", int_mips_clei_s_d, + NoItinerary, MSA128, MSA128>; + +class CLEI_U_B_DESC : MSA_SI5_DESC_BASE<"clei_u.b", int_mips_clei_u_b, + NoItinerary, MSA128, MSA128>; +class CLEI_U_H_DESC : MSA_SI5_DESC_BASE<"clei_u.h", int_mips_clei_u_h, + NoItinerary, MSA128, MSA128>; +class CLEI_U_W_DESC : MSA_SI5_DESC_BASE<"clei_u.w", int_mips_clei_u_w, + NoItinerary, MSA128, MSA128>; +class CLEI_U_D_DESC : MSA_SI5_DESC_BASE<"clei_u.d", int_mips_clei_u_d, + NoItinerary, MSA128, MSA128>; + +class CLT_S_B_DESC : MSA_3R_DESC_BASE<"clt_s.b", int_mips_clt_s_b, NoItinerary, + MSA128, MSA128>; +class CLT_S_H_DESC : MSA_3R_DESC_BASE<"clt_s.h", int_mips_clt_s_h, NoItinerary, + MSA128, MSA128>; +class CLT_S_W_DESC : MSA_3R_DESC_BASE<"clt_s.w", int_mips_clt_s_w, NoItinerary, + MSA128, MSA128>; +class CLT_S_D_DESC : MSA_3R_DESC_BASE<"clt_s.d", int_mips_clt_s_d, NoItinerary, + MSA128, MSA128>; + +class CLT_U_B_DESC : MSA_3R_DESC_BASE<"clt_u.b", int_mips_clt_u_b, NoItinerary, + MSA128, MSA128>; +class CLT_U_H_DESC : MSA_3R_DESC_BASE<"clt_u.h", int_mips_clt_u_h, NoItinerary, + MSA128, MSA128>; +class CLT_U_W_DESC : MSA_3R_DESC_BASE<"clt_u.w", int_mips_clt_u_w, NoItinerary, + MSA128, MSA128>; +class CLT_U_D_DESC : MSA_3R_DESC_BASE<"clt_u.d", int_mips_clt_u_d, NoItinerary, + MSA128, MSA128>; + +class CLTI_S_B_DESC : MSA_SI5_DESC_BASE<"clti_s.b", int_mips_clti_s_b, + NoItinerary, MSA128, MSA128>; +class CLTI_S_H_DESC : MSA_SI5_DESC_BASE<"clti_s.h", int_mips_clti_s_h, + NoItinerary, MSA128, MSA128>; +class CLTI_S_W_DESC : MSA_SI5_DESC_BASE<"clti_s.w", int_mips_clti_s_w, + NoItinerary, MSA128, MSA128>; +class CLTI_S_D_DESC : MSA_SI5_DESC_BASE<"clti_s.d", int_mips_clti_s_d, + NoItinerary, MSA128, MSA128>; + +class CLTI_U_B_DESC : MSA_SI5_DESC_BASE<"clti_u.b", int_mips_clti_u_b, + NoItinerary, MSA128, MSA128>; +class CLTI_U_H_DESC : MSA_SI5_DESC_BASE<"clti_u.h", int_mips_clti_u_h, + NoItinerary, MSA128, MSA128>; +class CLTI_U_W_DESC : MSA_SI5_DESC_BASE<"clti_u.w", int_mips_clti_u_w, + NoItinerary, MSA128, MSA128>; +class CLTI_U_D_DESC : MSA_SI5_DESC_BASE<"clti_u.d", int_mips_clti_u_d, + NoItinerary, MSA128, MSA128>; + +class COPY_S_B_DESC : MSA_COPY_DESC_BASE<"copy_s.b", int_mips_copy_s_b, + NoItinerary, GPR32, MSA128>; +class COPY_S_H_DESC : MSA_COPY_DESC_BASE<"copy_s.h", int_mips_copy_s_h, + NoItinerary, GPR32, MSA128>; +class COPY_S_W_DESC : MSA_COPY_DESC_BASE<"copy_s.w", int_mips_copy_s_w, + NoItinerary, GPR32, MSA128>; + +class COPY_U_B_DESC : MSA_COPY_DESC_BASE<"copy_u.b", int_mips_copy_u_b, + NoItinerary, GPR32, MSA128>; +class COPY_U_H_DESC : MSA_COPY_DESC_BASE<"copy_u.h", int_mips_copy_u_h, + NoItinerary, GPR32, MSA128>; +class COPY_U_W_DESC : MSA_COPY_DESC_BASE<"copy_u.w", int_mips_copy_u_w, + NoItinerary, GPR32, MSA128>; + +class DIV_S_B_DESC : MSA_3R_DESC_BASE<"div_s.b", int_mips_div_s_b, NoItinerary, + MSA128, MSA128>; +class DIV_S_H_DESC : MSA_3R_DESC_BASE<"div_s.h", int_mips_div_s_h, NoItinerary, + MSA128, MSA128>; +class DIV_S_W_DESC : MSA_3R_DESC_BASE<"div_s.w", int_mips_div_s_w, NoItinerary, + MSA128, MSA128>; +class DIV_S_D_DESC : MSA_3R_DESC_BASE<"div_s.d", int_mips_div_s_d, NoItinerary, + MSA128, MSA128>; + +class DIV_U_B_DESC : MSA_3R_DESC_BASE<"div_u.b", int_mips_div_u_b, NoItinerary, + MSA128, MSA128>; +class DIV_U_H_DESC : MSA_3R_DESC_BASE<"div_u.h", int_mips_div_u_h, NoItinerary, + MSA128, MSA128>; +class DIV_U_W_DESC : MSA_3R_DESC_BASE<"div_u.w", int_mips_div_u_w, NoItinerary, + MSA128, MSA128>; +class DIV_U_D_DESC : MSA_3R_DESC_BASE<"div_u.d", int_mips_div_u_d, NoItinerary, + MSA128, MSA128>; + +class DOTP_S_B_DESC : MSA_3R_DESC_BASE<"dotp_s.b", int_mips_dotp_s_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_S_H_DESC : MSA_3R_DESC_BASE<"dotp_s.h", int_mips_dotp_s_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_S_W_DESC : MSA_3R_DESC_BASE<"dotp_s.w", int_mips_dotp_s_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_S_D_DESC : MSA_3R_DESC_BASE<"dotp_s.d", int_mips_dotp_s_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class DOTP_U_B_DESC : MSA_3R_DESC_BASE<"dotp_u.b", int_mips_dotp_u_b, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_U_H_DESC : MSA_3R_DESC_BASE<"dotp_u.h", int_mips_dotp_u_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_U_W_DESC : MSA_3R_DESC_BASE<"dotp_u.w", int_mips_dotp_u_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DOTP_U_D_DESC : MSA_3R_DESC_BASE<"dotp_u.d", int_mips_dotp_u_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class DPADD_S_H_DESC : MSA_3R_4R_DESC_BASE<"dpadd_s.h", int_mips_dpadd_s_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DPADD_S_W_DESC : MSA_3R_4R_DESC_BASE<"dpadd_s.w", int_mips_dpadd_s_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DPADD_S_D_DESC : MSA_3R_4R_DESC_BASE<"dpadd_s.d", int_mips_dpadd_s_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class DPADD_U_H_DESC : MSA_3R_4R_DESC_BASE<"dpadd_u.h", int_mips_dpadd_u_h, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DPADD_U_W_DESC : MSA_3R_4R_DESC_BASE<"dpadd_u.w", int_mips_dpadd_u_w, + NoItinerary, MSA128, MSA128>, + IsCommutable; +class DPADD_U_D_DESC : MSA_3R_4R_DESC_BASE<"dpadd_u.d", int_mips_dpadd_u_d, + NoItinerary, MSA128, MSA128>, + IsCommutable; + +class DPSUB_S_H_DESC : MSA_3R_4R_DESC_BASE<"dpsub_s.h", int_mips_dpsub_s_h, + NoItinerary, MSA128, MSA128>; +class DPSUB_S_W_DESC : MSA_3R_4R_DESC_BASE<"dpsub_s.w", int_mips_dpsub_s_w, + NoItinerary, MSA128, MSA128>; +class DPSUB_S_D_DESC : MSA_3R_4R_DESC_BASE<"dpsub_s.d", int_mips_dpsub_s_d, + NoItinerary, MSA128, MSA128>; + +class DPSUB_U_H_DESC : MSA_3R_4R_DESC_BASE<"dpsub_u.h", int_mips_dpsub_u_h, + NoItinerary, MSA128, MSA128>; +class DPSUB_U_W_DESC : MSA_3R_4R_DESC_BASE<"dpsub_u.w", int_mips_dpsub_u_w, + NoItinerary, MSA128, MSA128>; +class DPSUB_U_D_DESC : MSA_3R_4R_DESC_BASE<"dpsub_u.d", int_mips_dpsub_u_d, + NoItinerary, MSA128, MSA128>; + +class ILVEV_B_DESC : MSA_3R_DESC_BASE<"ilvev.b", int_mips_ilvev_b, NoItinerary, + MSA128, MSA128>; +class ILVEV_H_DESC : MSA_3R_DESC_BASE<"ilvev.h", int_mips_ilvev_h, NoItinerary, + MSA128, MSA128>; +class ILVEV_W_DESC : MSA_3R_DESC_BASE<"ilvev.w", int_mips_ilvev_w, NoItinerary, + MSA128, MSA128>; +class ILVEV_D_DESC : MSA_3R_DESC_BASE<"ilvev.d", int_mips_ilvev_d, NoItinerary, + MSA128, MSA128>; + +class ILVL_B_DESC : MSA_3R_DESC_BASE<"ilvl.b", int_mips_ilvl_b, NoItinerary, + MSA128, MSA128>; +class ILVL_H_DESC : MSA_3R_DESC_BASE<"ilvl.h", int_mips_ilvl_h, NoItinerary, + MSA128, MSA128>; +class ILVL_W_DESC : MSA_3R_DESC_BASE<"ilvl.w", int_mips_ilvl_w, NoItinerary, + MSA128, MSA128>; +class ILVL_D_DESC : MSA_3R_DESC_BASE<"ilvl.d", int_mips_ilvl_d, NoItinerary, + MSA128, MSA128>; + +class ILVOD_B_DESC : MSA_3R_DESC_BASE<"ilvod.b", int_mips_ilvod_b, NoItinerary, + MSA128, MSA128>; +class ILVOD_H_DESC : MSA_3R_DESC_BASE<"ilvod.h", int_mips_ilvod_h, NoItinerary, + MSA128, MSA128>; +class ILVOD_W_DESC : MSA_3R_DESC_BASE<"ilvod.w", int_mips_ilvod_w, NoItinerary, + MSA128, MSA128>; +class ILVOD_D_DESC : MSA_3R_DESC_BASE<"ilvod.d", int_mips_ilvod_d, NoItinerary, + MSA128, MSA128>; + +class ILVR_B_DESC : MSA_3R_DESC_BASE<"ilvr.b", int_mips_ilvr_b, NoItinerary, + MSA128, MSA128>; +class ILVR_H_DESC : MSA_3R_DESC_BASE<"ilvr.h", int_mips_ilvr_h, NoItinerary, + MSA128, MSA128>; +class ILVR_W_DESC : MSA_3R_DESC_BASE<"ilvr.w", int_mips_ilvr_w, NoItinerary, + MSA128, MSA128>; +class ILVR_D_DESC : MSA_3R_DESC_BASE<"ilvr.d", int_mips_ilvr_d, NoItinerary, + MSA128, MSA128>; + +class INSERT_B_DESC : MSA_INSERT_DESC_BASE<"insert.b", int_mips_insert_b, + NoItinerary, MSA128, GPR32>; +class INSERT_H_DESC : MSA_INSERT_DESC_BASE<"insert.h", int_mips_insert_h, + NoItinerary, MSA128, GPR32>; +class INSERT_W_DESC : MSA_INSERT_DESC_BASE<"insert.w", int_mips_insert_w, + NoItinerary, MSA128, GPR32>; + class LD_DESC_BASE { @@ -32,6 +902,20 @@ class LD_DESC_BASE; +class LD_H_DESC : LD_DESC_BASE<"ld.h", load, v8i16, NoItinerary, MSA128>; +class LD_W_DESC : LD_DESC_BASE<"ld.w", load, v4i32, NoItinerary, MSA128>; +class LD_D_DESC : LD_DESC_BASE<"ld.d", load, v2i64, NoItinerary, MSA128>; + +class LDI_B_DESC : MSA_I10_DESC_BASE<"ldi.b", int_mips_ldi_b, + NoItinerary, MSA128>; +class LDI_H_DESC : MSA_I10_DESC_BASE<"ldi.h", int_mips_ldi_h, + NoItinerary, MSA128>; +class LDI_W_DESC : MSA_I10_DESC_BASE<"ldi.w", int_mips_ldi_w, + NoItinerary, MSA128>; +class LDI_D_DESC : MSA_I10_DESC_BASE<"ldi.d", int_mips_ldi_d, + NoItinerary, MSA128>; + class ST_DESC_BASE { @@ -43,21 +927,257 @@ class ST_DESC_BASE; -class LD_H_DESC : LD_DESC_BASE<"ld.h", load, v8i16, NoItinerary, MSA128>; -class LD_W_DESC : LD_DESC_BASE<"ld.w", load, v4i32, NoItinerary, MSA128>; -class LD_D_DESC : LD_DESC_BASE<"ld.d", load, v2i64, NoItinerary, MSA128>; class ST_B_DESC : ST_DESC_BASE<"st.b", store, v16i8, NoItinerary, MSA128>; class ST_H_DESC : ST_DESC_BASE<"st.h", store, v8i16, NoItinerary, MSA128>; class ST_W_DESC : ST_DESC_BASE<"st.w", store, v4i32, NoItinerary, MSA128>; class ST_D_DESC : ST_DESC_BASE<"st.d", store, v2i64, NoItinerary, MSA128>; // Instruction defs. +def ADD_A_B : ADD_A_B_ENC, ADD_A_B_DESC, Requires<[HasMSA]>; +def ADD_A_H : ADD_A_H_ENC, ADD_A_H_DESC, Requires<[HasMSA]>; +def ADD_A_W : ADD_A_W_ENC, ADD_A_W_DESC, Requires<[HasMSA]>; +def ADD_A_D : ADD_A_D_ENC, ADD_A_D_DESC, Requires<[HasMSA]>; + +def ADDS_A_B : ADDS_A_B_ENC, ADDS_A_B_DESC, Requires<[HasMSA]>; +def ADDS_A_H : ADDS_A_H_ENC, ADDS_A_H_DESC, Requires<[HasMSA]>; +def ADDS_A_W : ADDS_A_W_ENC, ADDS_A_W_DESC, Requires<[HasMSA]>; +def ADDS_A_D : ADDS_A_D_ENC, ADDS_A_D_DESC, Requires<[HasMSA]>; + +def ADDS_S_B : ADDS_S_B_ENC, ADDS_S_B_DESC, Requires<[HasMSA]>; +def ADDS_S_H : ADDS_S_H_ENC, ADDS_S_H_DESC, Requires<[HasMSA]>; +def ADDS_S_W : ADDS_S_W_ENC, ADDS_S_W_DESC, Requires<[HasMSA]>; +def ADDS_S_D : ADDS_S_D_ENC, ADDS_S_D_DESC, Requires<[HasMSA]>; + +def ADDS_U_B : ADDS_U_B_ENC, ADDS_U_B_DESC, Requires<[HasMSA]>; +def ADDS_U_H : ADDS_U_H_ENC, ADDS_U_H_DESC, Requires<[HasMSA]>; +def ADDS_U_W : ADDS_U_W_ENC, ADDS_U_W_DESC, Requires<[HasMSA]>; +def ADDS_U_D : ADDS_U_D_ENC, ADDS_U_D_DESC, Requires<[HasMSA]>; + +def ADDV_B : ADDV_B_ENC, ADDV_B_DESC, Requires<[HasMSA]>; +def ADDV_H : ADDV_H_ENC, ADDV_H_DESC, Requires<[HasMSA]>; +def ADDV_W : ADDV_W_ENC, ADDV_W_DESC, Requires<[HasMSA]>; +def ADDV_D : ADDV_D_ENC, ADDV_D_DESC, Requires<[HasMSA]>; + +def ADDVI_B : ADDVI_B_ENC, ADDVI_B_DESC, Requires<[HasMSA]>; +def ADDVI_H : ADDVI_H_ENC, ADDVI_H_DESC, Requires<[HasMSA]>; +def ADDVI_W : ADDVI_W_ENC, ADDVI_W_DESC, Requires<[HasMSA]>; +def ADDVI_D : ADDVI_D_ENC, ADDVI_D_DESC, Requires<[HasMSA]>; + +def ANDI_B : ANDI_B_ENC, ANDI_B_DESC, Requires<[HasMSA]>; + +def ASUB_S_B : ASUB_S_B_ENC, ASUB_S_B_DESC, Requires<[HasMSA]>; +def ASUB_S_H : ASUB_S_H_ENC, ASUB_S_H_DESC, Requires<[HasMSA]>; +def ASUB_S_W : ASUB_S_W_ENC, ASUB_S_W_DESC, Requires<[HasMSA]>; +def ASUB_S_D : ASUB_S_D_ENC, ASUB_S_D_DESC, Requires<[HasMSA]>; + +def ASUB_U_B : ASUB_U_B_ENC, ASUB_U_B_DESC, Requires<[HasMSA]>; +def ASUB_U_H : ASUB_U_H_ENC, ASUB_U_H_DESC, Requires<[HasMSA]>; +def ASUB_U_W : ASUB_U_W_ENC, ASUB_U_W_DESC, Requires<[HasMSA]>; +def ASUB_U_D : ASUB_U_D_ENC, ASUB_U_D_DESC, Requires<[HasMSA]>; + +def AVE_S_B : AVE_S_B_ENC, AVE_S_B_DESC, Requires<[HasMSA]>; +def AVE_S_H : AVE_S_H_ENC, AVE_S_H_DESC, Requires<[HasMSA]>; +def AVE_S_W : AVE_S_W_ENC, AVE_S_W_DESC, Requires<[HasMSA]>; +def AVE_S_D : AVE_S_D_ENC, AVE_S_D_DESC, Requires<[HasMSA]>; + +def AVE_U_B : AVE_U_B_ENC, AVE_U_B_DESC, Requires<[HasMSA]>; +def AVE_U_H : AVE_U_H_ENC, AVE_U_H_DESC, Requires<[HasMSA]>; +def AVE_U_W : AVE_U_W_ENC, AVE_U_W_DESC, Requires<[HasMSA]>; +def AVE_U_D : AVE_U_D_ENC, AVE_U_D_DESC, Requires<[HasMSA]>; + +def AVER_S_B : AVER_S_B_ENC, AVER_S_B_DESC, Requires<[HasMSA]>; +def AVER_S_H : AVER_S_H_ENC, AVER_S_H_DESC, Requires<[HasMSA]>; +def AVER_S_W : AVER_S_W_ENC, AVER_S_W_DESC, Requires<[HasMSA]>; +def AVER_S_D : AVER_S_D_ENC, AVER_S_D_DESC, Requires<[HasMSA]>; + +def AVER_U_B : AVER_U_B_ENC, AVER_U_B_DESC, Requires<[HasMSA]>; +def AVER_U_H : AVER_U_H_ENC, AVER_U_H_DESC, Requires<[HasMSA]>; +def AVER_U_W : AVER_U_W_ENC, AVER_U_W_DESC, Requires<[HasMSA]>; +def AVER_U_D : AVER_U_D_ENC, AVER_U_D_DESC, Requires<[HasMSA]>; + +def BCLR_B : BCLR_B_ENC, BCLR_B_DESC, Requires<[HasMSA]>; +def BCLR_H : BCLR_H_ENC, BCLR_H_DESC, Requires<[HasMSA]>; +def BCLR_W : BCLR_W_ENC, BCLR_W_DESC, Requires<[HasMSA]>; +def BCLR_D : BCLR_D_ENC, BCLR_D_DESC, Requires<[HasMSA]>; + +def BCLRI_B : BCLRI_B_ENC, BCLRI_B_DESC, Requires<[HasMSA]>; +def BCLRI_H : BCLRI_H_ENC, BCLRI_H_DESC, Requires<[HasMSA]>; +def BCLRI_W : BCLRI_W_ENC, BCLRI_W_DESC, Requires<[HasMSA]>; +def BCLRI_D : BCLRI_D_ENC, BCLRI_D_DESC, Requires<[HasMSA]>; + +def BINSL_B : BINSL_B_ENC, BINSL_B_DESC, Requires<[HasMSA]>; +def BINSL_H : BINSL_H_ENC, BINSL_H_DESC, Requires<[HasMSA]>; +def BINSL_W : BINSL_W_ENC, BINSL_W_DESC, Requires<[HasMSA]>; +def BINSL_D : BINSL_D_ENC, BINSL_D_DESC, Requires<[HasMSA]>; + +def BINSLI_B : BINSLI_B_ENC, BINSLI_B_DESC, Requires<[HasMSA]>; +def BINSLI_H : BINSLI_H_ENC, BINSLI_H_DESC, Requires<[HasMSA]>; +def BINSLI_W : BINSLI_W_ENC, BINSLI_W_DESC, Requires<[HasMSA]>; +def BINSLI_D : BINSLI_D_ENC, BINSLI_D_DESC, Requires<[HasMSA]>; + +def BINSR_B : BINSR_B_ENC, BINSR_B_DESC, Requires<[HasMSA]>; +def BINSR_H : BINSR_H_ENC, BINSR_H_DESC, Requires<[HasMSA]>; +def BINSR_W : BINSR_W_ENC, BINSR_W_DESC, Requires<[HasMSA]>; +def BINSR_D : BINSR_D_ENC, BINSR_D_DESC, Requires<[HasMSA]>; + +def BINSRI_B : BINSRI_B_ENC, BINSRI_B_DESC, Requires<[HasMSA]>; +def BINSRI_H : BINSRI_H_ENC, BINSRI_H_DESC, Requires<[HasMSA]>; +def BINSRI_W : BINSRI_W_ENC, BINSRI_W_DESC, Requires<[HasMSA]>; +def BINSRI_D : BINSRI_D_ENC, BINSRI_D_DESC, Requires<[HasMSA]>; + +def BMNZI_B : BMNZI_B_ENC, BMNZI_B_DESC, Requires<[HasMSA]>; + +def BMZI_B : BMZI_B_ENC, BMZI_B_DESC, Requires<[HasMSA]>; + +def BNEG_B : BNEG_B_ENC, BNEG_B_DESC, Requires<[HasMSA]>; +def BNEG_H : BNEG_H_ENC, BNEG_H_DESC, Requires<[HasMSA]>; +def BNEG_W : BNEG_W_ENC, BNEG_W_DESC, Requires<[HasMSA]>; +def BNEG_D : BNEG_D_ENC, BNEG_D_DESC, Requires<[HasMSA]>; + +def BNEGI_B : BNEGI_B_ENC, BNEGI_B_DESC, Requires<[HasMSA]>; +def BNEGI_H : BNEGI_H_ENC, BNEGI_H_DESC, Requires<[HasMSA]>; +def BNEGI_W : BNEGI_W_ENC, BNEGI_W_DESC, Requires<[HasMSA]>; +def BNEGI_D : BNEGI_D_ENC, BNEGI_D_DESC, Requires<[HasMSA]>; + +def BSELI_B : BSELI_B_ENC, BSELI_B_DESC, Requires<[HasMSA]>; + +def BSET_B : BSET_B_ENC, BSET_B_DESC, Requires<[HasMSA]>; +def BSET_H : BSET_H_ENC, BSET_H_DESC, Requires<[HasMSA]>; +def BSET_W : BSET_W_ENC, BSET_W_DESC, Requires<[HasMSA]>; +def BSET_D : BSET_D_ENC, BSET_D_DESC, Requires<[HasMSA]>; + +def BSETI_B : BSETI_B_ENC, BSETI_B_DESC, Requires<[HasMSA]>; +def BSETI_H : BSETI_H_ENC, BSETI_H_DESC, Requires<[HasMSA]>; +def BSETI_W : BSETI_W_ENC, BSETI_W_DESC, Requires<[HasMSA]>; +def BSETI_D : BSETI_D_ENC, BSETI_D_DESC, Requires<[HasMSA]>; + +def CEQ_B : CEQ_B_ENC, CEQ_B_DESC, Requires<[HasMSA]>; +def CEQ_H : CEQ_H_ENC, CEQ_H_DESC, Requires<[HasMSA]>; +def CEQ_W : CEQ_W_ENC, CEQ_W_DESC, Requires<[HasMSA]>; +def CEQ_D : CEQ_D_ENC, CEQ_D_DESC, Requires<[HasMSA]>; + +def CEQI_B : CEQI_B_ENC, CEQI_B_DESC, Requires<[HasMSA]>; +def CEQI_H : CEQI_H_ENC, CEQI_H_DESC, Requires<[HasMSA]>; +def CEQI_W : CEQI_W_ENC, CEQI_W_DESC, Requires<[HasMSA]>; +def CEQI_D : CEQI_D_ENC, CEQI_D_DESC, Requires<[HasMSA]>; + +def CLE_S_B : CLE_S_B_ENC, CLE_S_B_DESC, Requires<[HasMSA]>; +def CLE_S_H : CLE_S_H_ENC, CLE_S_H_DESC, Requires<[HasMSA]>; +def CLE_S_W : CLE_S_W_ENC, CLE_S_W_DESC, Requires<[HasMSA]>; +def CLE_S_D : CLE_S_D_ENC, CLE_S_D_DESC, Requires<[HasMSA]>; + +def CLE_U_B : CLE_U_B_ENC, CLE_U_B_DESC, Requires<[HasMSA]>; +def CLE_U_H : CLE_U_H_ENC, CLE_U_H_DESC, Requires<[HasMSA]>; +def CLE_U_W : CLE_U_W_ENC, CLE_U_W_DESC, Requires<[HasMSA]>; +def CLE_U_D : CLE_U_D_ENC, CLE_U_D_DESC, Requires<[HasMSA]>; + +def CLEI_S_B : CLEI_S_B_ENC, CLEI_S_B_DESC, Requires<[HasMSA]>; +def CLEI_S_H : CLEI_S_H_ENC, CLEI_S_H_DESC, Requires<[HasMSA]>; +def CLEI_S_W : CLEI_S_W_ENC, CLEI_S_W_DESC, Requires<[HasMSA]>; +def CLEI_S_D : CLEI_S_D_ENC, CLEI_S_D_DESC, Requires<[HasMSA]>; + +def CLEI_U_B : CLEI_U_B_ENC, CLEI_U_B_DESC, Requires<[HasMSA]>; +def CLEI_U_H : CLEI_U_H_ENC, CLEI_U_H_DESC, Requires<[HasMSA]>; +def CLEI_U_W : CLEI_U_W_ENC, CLEI_U_W_DESC, Requires<[HasMSA]>; +def CLEI_U_D : CLEI_U_D_ENC, CLEI_U_D_DESC, Requires<[HasMSA]>; + +def CLT_S_B : CLT_S_B_ENC, CLT_S_B_DESC, Requires<[HasMSA]>; +def CLT_S_H : CLT_S_H_ENC, CLT_S_H_DESC, Requires<[HasMSA]>; +def CLT_S_W : CLT_S_W_ENC, CLT_S_W_DESC, Requires<[HasMSA]>; +def CLT_S_D : CLT_S_D_ENC, CLT_S_D_DESC, Requires<[HasMSA]>; + +def CLT_U_B : CLT_U_B_ENC, CLT_U_B_DESC, Requires<[HasMSA]>; +def CLT_U_H : CLT_U_H_ENC, CLT_U_H_DESC, Requires<[HasMSA]>; +def CLT_U_W : CLT_U_W_ENC, CLT_U_W_DESC, Requires<[HasMSA]>; +def CLT_U_D : CLT_U_D_ENC, CLT_U_D_DESC, Requires<[HasMSA]>; + +def CLTI_S_B : CLTI_S_B_ENC, CLTI_S_B_DESC, Requires<[HasMSA]>; +def CLTI_S_H : CLTI_S_H_ENC, CLTI_S_H_DESC, Requires<[HasMSA]>; +def CLTI_S_W : CLTI_S_W_ENC, CLTI_S_W_DESC, Requires<[HasMSA]>; +def CLTI_S_D : CLTI_S_D_ENC, CLTI_S_D_DESC, Requires<[HasMSA]>; + +def CLTI_U_B : CLTI_U_B_ENC, CLTI_U_B_DESC, Requires<[HasMSA]>; +def CLTI_U_H : CLTI_U_H_ENC, CLTI_U_H_DESC, Requires<[HasMSA]>; +def CLTI_U_W : CLTI_U_W_ENC, CLTI_U_W_DESC, Requires<[HasMSA]>; +def CLTI_U_D : CLTI_U_D_ENC, CLTI_U_D_DESC, Requires<[HasMSA]>; + +def COPY_S_B : COPY_S_B_ENC, COPY_S_B_DESC, Requires<[HasMSA]>; +def COPY_S_H : COPY_S_H_ENC, COPY_S_H_DESC, Requires<[HasMSA]>; +def COPY_S_W : COPY_S_W_ENC, COPY_S_W_DESC, Requires<[HasMSA]>; + +def COPY_U_B : COPY_U_B_ENC, COPY_U_B_DESC, Requires<[HasMSA]>; +def COPY_U_H : COPY_U_H_ENC, COPY_U_H_DESC, Requires<[HasMSA]>; +def COPY_U_W : COPY_U_W_ENC, COPY_U_W_DESC, Requires<[HasMSA]>; + +def DIV_S_B : DIV_S_B_ENC, DIV_S_B_DESC, Requires<[HasMSA]>; +def DIV_S_H : DIV_S_H_ENC, DIV_S_H_DESC, Requires<[HasMSA]>; +def DIV_S_W : DIV_S_W_ENC, DIV_S_W_DESC, Requires<[HasMSA]>; +def DIV_S_D : DIV_S_D_ENC, DIV_S_D_DESC, Requires<[HasMSA]>; + +def DIV_U_B : DIV_U_B_ENC, DIV_U_B_DESC, Requires<[HasMSA]>; +def DIV_U_H : DIV_U_H_ENC, DIV_U_H_DESC, Requires<[HasMSA]>; +def DIV_U_W : DIV_U_W_ENC, DIV_U_W_DESC, Requires<[HasMSA]>; +def DIV_U_D : DIV_U_D_ENC, DIV_U_D_DESC, Requires<[HasMSA]>; + +def DOTP_S_B : DOTP_S_B_ENC, DOTP_S_B_DESC, Requires<[HasMSA]>; +def DOTP_S_H : DOTP_S_H_ENC, DOTP_S_H_DESC, Requires<[HasMSA]>; +def DOTP_S_W : DOTP_S_W_ENC, DOTP_S_W_DESC, Requires<[HasMSA]>; +def DOTP_S_D : DOTP_S_D_ENC, DOTP_S_D_DESC, Requires<[HasMSA]>; + +def DOTP_U_B : DOTP_U_B_ENC, DOTP_U_B_DESC, Requires<[HasMSA]>; +def DOTP_U_H : DOTP_U_H_ENC, DOTP_U_H_DESC, Requires<[HasMSA]>; +def DOTP_U_W : DOTP_U_W_ENC, DOTP_U_W_DESC, Requires<[HasMSA]>; +def DOTP_U_D : DOTP_U_D_ENC, DOTP_U_D_DESC, Requires<[HasMSA]>; + +def DPADD_S_H : DPADD_S_H_ENC, DPADD_S_H_DESC, Requires<[HasMSA]>; +def DPADD_S_W : DPADD_S_W_ENC, DPADD_S_W_DESC, Requires<[HasMSA]>; +def DPADD_S_D : DPADD_S_D_ENC, DPADD_S_D_DESC, Requires<[HasMSA]>; + +def DPADD_U_H : DPADD_U_H_ENC, DPADD_U_H_DESC, Requires<[HasMSA]>; +def DPADD_U_W : DPADD_U_W_ENC, DPADD_U_W_DESC, Requires<[HasMSA]>; +def DPADD_U_D : DPADD_U_D_ENC, DPADD_U_D_DESC, Requires<[HasMSA]>; + +def DPSUB_S_H : DPSUB_S_H_ENC, DPSUB_S_H_DESC, Requires<[HasMSA]>; +def DPSUB_S_W : DPSUB_S_W_ENC, DPSUB_S_W_DESC, Requires<[HasMSA]>; +def DPSUB_S_D : DPSUB_S_D_ENC, DPSUB_S_D_DESC, Requires<[HasMSA]>; + +def DPSUB_U_H : DPSUB_U_H_ENC, DPSUB_U_H_DESC, Requires<[HasMSA]>; +def DPSUB_U_W : DPSUB_U_W_ENC, DPSUB_U_W_DESC, Requires<[HasMSA]>; +def DPSUB_U_D : DPSUB_U_D_ENC, DPSUB_U_D_DESC, Requires<[HasMSA]>; + +def ILVEV_B : ILVEV_B_ENC, ILVEV_B_DESC, Requires<[HasMSA]>; +def ILVEV_H : ILVEV_H_ENC, ILVEV_H_DESC, Requires<[HasMSA]>; +def ILVEV_W : ILVEV_W_ENC, ILVEV_W_DESC, Requires<[HasMSA]>; +def ILVEV_D : ILVEV_D_ENC, ILVEV_D_DESC, Requires<[HasMSA]>; + +def ILVL_B : ILVL_B_ENC, ILVL_B_DESC, Requires<[HasMSA]>; +def ILVL_H : ILVL_H_ENC, ILVL_H_DESC, Requires<[HasMSA]>; +def ILVL_W : ILVL_W_ENC, ILVL_W_DESC, Requires<[HasMSA]>; +def ILVL_D : ILVL_D_ENC, ILVL_D_DESC, Requires<[HasMSA]>; + +def ILVOD_B : ILVOD_B_ENC, ILVOD_B_DESC, Requires<[HasMSA]>; +def ILVOD_H : ILVOD_H_ENC, ILVOD_H_DESC, Requires<[HasMSA]>; +def ILVOD_W : ILVOD_W_ENC, ILVOD_W_DESC, Requires<[HasMSA]>; +def ILVOD_D : ILVOD_D_ENC, ILVOD_D_DESC, Requires<[HasMSA]>; + +def ILVR_B : ILVR_B_ENC, ILVR_B_DESC, Requires<[HasMSA]>; +def ILVR_H : ILVR_H_ENC, ILVR_H_DESC, Requires<[HasMSA]>; +def ILVR_W : ILVR_W_ENC, ILVR_W_DESC, Requires<[HasMSA]>; +def ILVR_D : ILVR_D_ENC, ILVR_D_DESC, Requires<[HasMSA]>; + +def INSERT_B : INSERT_B_ENC, INSERT_B_DESC, Requires<[HasMSA]>; +def INSERT_H : INSERT_H_ENC, INSERT_H_DESC, Requires<[HasMSA]>; +def INSERT_W : INSERT_W_ENC, INSERT_W_DESC, Requires<[HasMSA]>; + def LD_B: LD_B_ENC, LD_B_DESC, Requires<[HasMSA]>; def LD_H: LD_H_ENC, LD_H_DESC, Requires<[HasMSA]>; def LD_W: LD_W_ENC, LD_W_DESC, Requires<[HasMSA]>; def LD_D: LD_D_ENC, LD_D_DESC, Requires<[HasMSA]>; +def LDI_B : LDI_B_ENC, LDI_B_DESC, Requires<[HasMSA]>; +def LDI_H : LDI_H_ENC, LDI_H_DESC, Requires<[HasMSA]>; +def LDI_W : LDI_W_ENC, LDI_W_DESC, Requires<[HasMSA]>; + def ST_B: ST_B_ENC, ST_B_DESC, Requires<[HasMSA]>; def ST_H: ST_H_ENC, ST_H_DESC, Requires<[HasMSA]>; def ST_W: ST_W_ENC, ST_W_DESC, Requires<[HasMSA]>; @@ -67,3 +1187,12 @@ def ST_D: ST_D_ENC, ST_D_DESC, Requires<[HasMSA]>; class MSAPat : Pat, Requires<[pred]>; +def LD_FW : MSAPat<(v4f32 (load addr:$addr)), + (LD_W addr:$addr)>; +def LD_FD : MSAPat<(v2f64 (load addr:$addr)), + (LD_D addr:$addr)>; + +def ST_FW : MSAPat<(store (v4f32 MSA128:$ws), addr:$addr), + (ST_W MSA128:$ws, addr:$addr)>; +def ST_FD : MSAPat<(store (v2f64 MSA128:$ws), addr:$addr), + (ST_D MSA128:$ws, addr:$addr)>; diff --git a/lib/Target/Mips/MipsRegisterInfo.td b/lib/Target/Mips/MipsRegisterInfo.td index 176eb645bc5..b7c005f3b8b 100644 --- a/lib/Target/Mips/MipsRegisterInfo.td +++ b/lib/Target/Mips/MipsRegisterInfo.td @@ -343,8 +343,9 @@ def CCR : RegisterClass<"Mips", [i32], 32, (sequence "FCR%u", 0, 31)>, def FCC : RegisterClass<"Mips", [i32], 32, (sequence "FCC%u", 0, 7)>, Unallocatable; -def MSA128: RegisterClass<"Mips", [v16i8, v8i16, v4i32, v2i64], 128, - (sequence "W%u", 0, 31)>; +def MSA128: RegisterClass<"Mips", + [v16i8, v8i16, v4i32, v2i64, v8f16, v4f32, v2f64], + 128, (sequence "W%u", 0, 31)>; // Hi/Lo Registers def LO32 : RegisterClass<"Mips", [i32], 32, (add LO0)>; diff --git a/lib/Target/Mips/MipsSEISelLowering.cpp b/lib/Target/Mips/MipsSEISelLowering.cpp index c54c55cd143..750ec0e0d33 100644 --- a/lib/Target/Mips/MipsSEISelLowering.cpp +++ b/lib/Target/Mips/MipsSEISelLowering.cpp @@ -78,20 +78,13 @@ MipsSETargetLowering::MipsSETargetLowering(MipsTargetMachine &TM) setOperationAction(ISD::MUL, MVT::v2i16, Legal); if (Subtarget->hasMSA()) { - MVT::SimpleValueType VecTys[4] = {MVT::v16i8, MVT::v8i16, - MVT::v4i32, MVT::v2i64}; - - for (unsigned i = 0; i < array_lengthof(VecTys); ++i) { - addRegisterClass(VecTys[i], &Mips::MSA128RegClass); - - // Expand all builtin opcodes. - for (unsigned Opc = 0; Opc < ISD::BUILTIN_OP_END; ++Opc) - setOperationAction(Opc, VecTys[i], Expand); - - setOperationAction(ISD::LOAD, VecTys[i], Legal); - setOperationAction(ISD::STORE, VecTys[i], Legal); - setOperationAction(ISD::BITCAST, VecTys[i], Legal); - } + addMSAType(MVT::v16i8); + addMSAType(MVT::v8i16); + addMSAType(MVT::v4i32); + addMSAType(MVT::v2i64); + addMSAType(MVT::v8f16); + addMSAType(MVT::v4f32); + addMSAType(MVT::v2f64); } if (!TM.Options.UseSoftFloat) { @@ -140,6 +133,18 @@ llvm::createMipsSETargetLowering(MipsTargetMachine &TM) { return new MipsSETargetLowering(TM); } +void +MipsSETargetLowering::addMSAType(MVT::SimpleValueType Ty) { + addRegisterClass(Ty, &Mips::MSA128RegClass); + + // Expand all builtin opcodes. + for (unsigned Opc = 0; Opc < ISD::BUILTIN_OP_END; ++Opc) + setOperationAction(Opc, Ty, Expand); + + setOperationAction(ISD::LOAD, Ty, Legal); + setOperationAction(ISD::STORE, Ty, Legal); + setOperationAction(ISD::BITCAST, Ty, Legal); +} bool MipsSETargetLowering::allowsUnalignedMemoryAccesses(EVT VT, bool *Fast) const { diff --git a/lib/Target/Mips/MipsSEISelLowering.h b/lib/Target/Mips/MipsSEISelLowering.h index da1321becb0..b56036dc5a6 100644 --- a/lib/Target/Mips/MipsSEISelLowering.h +++ b/lib/Target/Mips/MipsSEISelLowering.h @@ -22,6 +22,8 @@ namespace llvm { public: explicit MipsSETargetLowering(MipsTargetMachine &TM); + void addMSAType(MVT::SimpleValueType Ty); + virtual bool allowsUnalignedMemoryAccesses(EVT VT, bool *Fast) const; virtual SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const; diff --git a/test/CodeGen/Mips/msa/2r_vector_scalar.ll b/test/CodeGen/Mips/msa/2r_vector_scalar.ll new file mode 100644 index 00000000000..9d0765a2523 --- /dev/null +++ b/test/CodeGen/Mips/msa/2r_vector_scalar.ll @@ -0,0 +1,59 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_fill_b_ARG1 = global i32 23, align 16 +@llvm_mips_fill_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_fill_b_test() nounwind { +entry: + %0 = load i32* @llvm_mips_fill_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.fill.b(i32 %0) + store <16 x i8> %1, <16 x i8>* @llvm_mips_fill_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.fill.b(i32) nounwind + +; CHECK: llvm_mips_fill_b_test: +; CHECK: lw +; CHECK: fill.b +; CHECK: st.b +; CHECK: .size llvm_mips_fill_b_test +; +@llvm_mips_fill_h_ARG1 = global i32 23, align 16 +@llvm_mips_fill_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_fill_h_test() nounwind { +entry: + %0 = load i32* @llvm_mips_fill_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.fill.h(i32 %0) + store <8 x i16> %1, <8 x i16>* @llvm_mips_fill_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.fill.h(i32) nounwind + +; CHECK: llvm_mips_fill_h_test: +; CHECK: lw +; CHECK: fill.h +; CHECK: st.h +; CHECK: .size llvm_mips_fill_h_test +; +@llvm_mips_fill_w_ARG1 = global i32 23, align 16 +@llvm_mips_fill_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_fill_w_test() nounwind { +entry: + %0 = load i32* @llvm_mips_fill_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.fill.w(i32 %0) + store <4 x i32> %1, <4 x i32>* @llvm_mips_fill_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.fill.w(i32) nounwind + +; CHECK: llvm_mips_fill_w_test: +; CHECK: lw +; CHECK: fill.w +; CHECK: st.w +; CHECK: .size llvm_mips_fill_w_test +; diff --git a/test/CodeGen/Mips/msa/3r-a.ll b/test/CodeGen/Mips/msa/3r-a.ll new file mode 100644 index 00000000000..3cdeae4af3a --- /dev/null +++ b/test/CodeGen/Mips/msa/3r-a.ll @@ -0,0 +1,970 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_add_a_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_add_a_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_add_a_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_add_a_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_add_a_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_add_a_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.add.a.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_add_a_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.add.a.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_add_a_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: add_a.b +; CHECK: st.b +; CHECK: .size llvm_mips_add_a_b_test +; +@llvm_mips_add_a_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_add_a_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_add_a_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_add_a_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_add_a_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_add_a_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.add.a.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_add_a_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.add.a.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_add_a_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: add_a.h +; CHECK: st.h +; CHECK: .size llvm_mips_add_a_h_test +; +@llvm_mips_add_a_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_add_a_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_add_a_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_add_a_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_add_a_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_add_a_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.add.a.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_add_a_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.add.a.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_add_a_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: add_a.w +; CHECK: st.w +; CHECK: .size llvm_mips_add_a_w_test +; +@llvm_mips_add_a_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_add_a_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_add_a_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_add_a_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_add_a_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_add_a_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.add.a.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_add_a_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.add.a.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_add_a_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: add_a.d +; CHECK: st.d +; CHECK: .size llvm_mips_add_a_d_test +; +@llvm_mips_adds_a_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_adds_a_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_adds_a_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_adds_a_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_adds_a_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_adds_a_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.adds.a.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_adds_a_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.adds.a.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_adds_a_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: adds_a.b +; CHECK: st.b +; CHECK: .size llvm_mips_adds_a_b_test +; +@llvm_mips_adds_a_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_adds_a_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_adds_a_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_adds_a_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_adds_a_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_adds_a_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.adds.a.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_adds_a_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.adds.a.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_adds_a_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: adds_a.h +; CHECK: st.h +; CHECK: .size llvm_mips_adds_a_h_test +; +@llvm_mips_adds_a_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_adds_a_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_adds_a_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_adds_a_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_adds_a_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_adds_a_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.adds.a.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_adds_a_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.adds.a.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_adds_a_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: adds_a.w +; CHECK: st.w +; CHECK: .size llvm_mips_adds_a_w_test +; +@llvm_mips_adds_a_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_adds_a_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_adds_a_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_adds_a_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_adds_a_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_adds_a_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.adds.a.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_adds_a_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.adds.a.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_adds_a_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: adds_a.d +; CHECK: st.d +; CHECK: .size llvm_mips_adds_a_d_test +; +@llvm_mips_adds_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_adds_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_adds_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_adds_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_adds_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_adds_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.adds.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_adds_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.adds.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_adds_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: adds_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_adds_s_b_test +; +@llvm_mips_adds_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_adds_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_adds_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_adds_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_adds_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_adds_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.adds.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_adds_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.adds.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_adds_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: adds_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_adds_s_h_test +; +@llvm_mips_adds_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_adds_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_adds_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_adds_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_adds_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_adds_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.adds.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_adds_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.adds.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_adds_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: adds_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_adds_s_w_test +; +@llvm_mips_adds_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_adds_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_adds_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_adds_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_adds_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_adds_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.adds.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_adds_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.adds.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_adds_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: adds_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_adds_s_d_test +; +@llvm_mips_adds_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_adds_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_adds_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_adds_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_adds_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_adds_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.adds.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_adds_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.adds.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_adds_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: adds_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_adds_u_b_test +; +@llvm_mips_adds_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_adds_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_adds_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_adds_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_adds_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_adds_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.adds.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_adds_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.adds.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_adds_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: adds_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_adds_u_h_test +; +@llvm_mips_adds_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_adds_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_adds_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_adds_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_adds_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_adds_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.adds.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_adds_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.adds.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_adds_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: adds_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_adds_u_w_test +; +@llvm_mips_adds_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_adds_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_adds_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_adds_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_adds_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_adds_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.adds.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_adds_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.adds.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_adds_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: adds_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_adds_u_d_test +; +@llvm_mips_addv_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_addv_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_addv_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_addv_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_addv_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_addv_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.addv.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_addv_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.addv.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_addv_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: addv.b +; CHECK: st.b +; CHECK: .size llvm_mips_addv_b_test +; +@llvm_mips_addv_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_addv_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_addv_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_addv_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_addv_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_addv_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.addv.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_addv_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.addv.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_addv_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: addv.h +; CHECK: st.h +; CHECK: .size llvm_mips_addv_h_test +; +@llvm_mips_addv_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_addv_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_addv_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_addv_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_addv_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_addv_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.addv.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_addv_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.addv.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_addv_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: addv.w +; CHECK: st.w +; CHECK: .size llvm_mips_addv_w_test +; +@llvm_mips_addv_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_addv_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_addv_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_addv_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_addv_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_addv_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.addv.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_addv_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.addv.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_addv_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: addv.d +; CHECK: st.d +; CHECK: .size llvm_mips_addv_d_test +; +@llvm_mips_asub_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_asub_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_asub_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_asub_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_asub_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_asub_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.asub.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_asub_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.asub.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_asub_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: asub_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_asub_s_b_test +; +@llvm_mips_asub_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_asub_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_asub_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_asub_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_asub_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_asub_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.asub.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_asub_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.asub.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_asub_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: asub_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_asub_s_h_test +; +@llvm_mips_asub_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_asub_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_asub_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_asub_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_asub_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_asub_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.asub.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_asub_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.asub.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_asub_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: asub_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_asub_s_w_test +; +@llvm_mips_asub_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_asub_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_asub_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_asub_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_asub_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_asub_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.asub.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_asub_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.asub.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_asub_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: asub_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_asub_s_d_test +; +@llvm_mips_asub_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_asub_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_asub_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_asub_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_asub_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_asub_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.asub.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_asub_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.asub.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_asub_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: asub_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_asub_u_b_test +; +@llvm_mips_asub_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_asub_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_asub_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_asub_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_asub_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_asub_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.asub.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_asub_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.asub.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_asub_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: asub_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_asub_u_h_test +; +@llvm_mips_asub_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_asub_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_asub_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_asub_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_asub_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_asub_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.asub.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_asub_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.asub.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_asub_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: asub_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_asub_u_w_test +; +@llvm_mips_asub_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_asub_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_asub_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_asub_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_asub_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_asub_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.asub.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_asub_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.asub.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_asub_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: asub_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_asub_u_d_test +; +@llvm_mips_ave_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ave_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ave_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ave_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ave_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ave_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ave.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ave_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ave.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ave_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ave_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_ave_s_b_test +; +@llvm_mips_ave_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ave_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ave_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ave_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ave_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ave_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ave.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ave_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ave.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ave_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ave_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_ave_s_h_test +; +@llvm_mips_ave_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ave_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ave_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ave_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ave_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ave_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ave.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ave_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ave.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ave_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ave_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_ave_s_w_test +; +@llvm_mips_ave_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ave_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ave_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ave_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ave_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ave_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ave.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ave_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ave.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ave_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ave_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_ave_s_d_test +; +@llvm_mips_ave_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ave_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ave_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ave_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ave_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ave_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ave.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ave_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ave.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ave_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ave_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_ave_u_b_test +; +@llvm_mips_ave_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ave_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ave_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ave_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ave_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ave_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ave.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ave_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ave.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ave_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ave_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_ave_u_h_test +; +@llvm_mips_ave_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ave_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ave_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ave_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ave_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ave_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ave.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ave_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ave.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ave_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ave_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_ave_u_w_test +; +@llvm_mips_ave_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ave_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ave_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ave_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ave_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ave_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ave.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ave_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ave.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ave_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ave_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_ave_u_d_test +; +@llvm_mips_aver_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_aver_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_aver_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_aver_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_aver_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_aver_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.aver.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_aver_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.aver.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_aver_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: aver_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_aver_s_b_test +; +@llvm_mips_aver_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_aver_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_aver_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_aver_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_aver_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_aver_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.aver.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_aver_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.aver.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_aver_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: aver_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_aver_s_h_test +; +@llvm_mips_aver_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_aver_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_aver_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_aver_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_aver_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_aver_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.aver.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_aver_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.aver.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_aver_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: aver_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_aver_s_w_test +; +@llvm_mips_aver_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_aver_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_aver_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_aver_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_aver_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_aver_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.aver.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_aver_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.aver.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_aver_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: aver_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_aver_s_d_test +; +@llvm_mips_aver_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_aver_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_aver_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_aver_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_aver_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_aver_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.aver.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_aver_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.aver.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_aver_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: aver_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_aver_u_b_test +; +@llvm_mips_aver_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_aver_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_aver_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_aver_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_aver_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_aver_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.aver.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_aver_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.aver.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_aver_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: aver_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_aver_u_h_test +; +@llvm_mips_aver_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_aver_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_aver_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_aver_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_aver_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_aver_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.aver.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_aver_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.aver.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_aver_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: aver_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_aver_u_w_test +; +@llvm_mips_aver_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_aver_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_aver_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_aver_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_aver_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_aver_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.aver.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_aver_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.aver.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_aver_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: aver_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_aver_u_d_test +; diff --git a/test/CodeGen/Mips/msa/3r-b.ll b/test/CodeGen/Mips/msa/3r-b.ll new file mode 100644 index 00000000000..afcc391e534 --- /dev/null +++ b/test/CodeGen/Mips/msa/3r-b.ll @@ -0,0 +1,442 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_bclr_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bclr_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_bclr_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bclr_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bclr_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_bclr_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.bclr.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_bclr_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bclr.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_bclr_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: bclr.b +; CHECK: st.b +; CHECK: .size llvm_mips_bclr_b_test +; +@llvm_mips_bclr_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bclr_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_bclr_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bclr_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bclr_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_bclr_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.bclr.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_bclr_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bclr.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_bclr_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: bclr.h +; CHECK: st.h +; CHECK: .size llvm_mips_bclr_h_test +; +@llvm_mips_bclr_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bclr_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_bclr_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bclr_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bclr_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_bclr_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.bclr.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_bclr_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bclr.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_bclr_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: bclr.w +; CHECK: st.w +; CHECK: .size llvm_mips_bclr_w_test +; +@llvm_mips_bclr_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bclr_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_bclr_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bclr_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bclr_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_bclr_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.bclr.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_bclr_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bclr.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_bclr_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: bclr.d +; CHECK: st.d +; CHECK: .size llvm_mips_bclr_d_test +; +@llvm_mips_binsl_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_binsl_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_binsl_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_binsl_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_binsl_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_binsl_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.binsl.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_binsl_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.binsl.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_binsl_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: binsl.b +; CHECK: st.b +; CHECK: .size llvm_mips_binsl_b_test +; +@llvm_mips_binsl_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_binsl_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_binsl_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_binsl_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_binsl_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_binsl_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.binsl.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_binsl_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.binsl.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_binsl_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: binsl.h +; CHECK: st.h +; CHECK: .size llvm_mips_binsl_h_test +; +@llvm_mips_binsl_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_binsl_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_binsl_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_binsl_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_binsl_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_binsl_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.binsl.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_binsl_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.binsl.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_binsl_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: binsl.w +; CHECK: st.w +; CHECK: .size llvm_mips_binsl_w_test +; +@llvm_mips_binsl_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_binsl_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_binsl_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_binsl_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_binsl_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_binsl_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.binsl.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_binsl_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.binsl.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_binsl_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: binsl.d +; CHECK: st.d +; CHECK: .size llvm_mips_binsl_d_test +; +@llvm_mips_binsr_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_binsr_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_binsr_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_binsr_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_binsr_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_binsr_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.binsr.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_binsr_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.binsr.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_binsr_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: binsr.b +; CHECK: st.b +; CHECK: .size llvm_mips_binsr_b_test +; +@llvm_mips_binsr_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_binsr_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_binsr_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_binsr_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_binsr_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_binsr_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.binsr.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_binsr_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.binsr.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_binsr_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: binsr.h +; CHECK: st.h +; CHECK: .size llvm_mips_binsr_h_test +; +@llvm_mips_binsr_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_binsr_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_binsr_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_binsr_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_binsr_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_binsr_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.binsr.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_binsr_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.binsr.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_binsr_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: binsr.w +; CHECK: st.w +; CHECK: .size llvm_mips_binsr_w_test +; +@llvm_mips_binsr_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_binsr_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_binsr_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_binsr_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_binsr_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_binsr_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.binsr.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_binsr_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.binsr.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_binsr_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: binsr.d +; CHECK: st.d +; CHECK: .size llvm_mips_binsr_d_test +; +@llvm_mips_bneg_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bneg_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_bneg_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bneg_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bneg_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_bneg_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.bneg.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_bneg_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bneg.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_bneg_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: bneg.b +; CHECK: st.b +; CHECK: .size llvm_mips_bneg_b_test +; +@llvm_mips_bneg_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bneg_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_bneg_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bneg_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bneg_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_bneg_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.bneg.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_bneg_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bneg.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_bneg_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: bneg.h +; CHECK: st.h +; CHECK: .size llvm_mips_bneg_h_test +; +@llvm_mips_bneg_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bneg_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_bneg_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bneg_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bneg_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_bneg_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.bneg.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_bneg_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bneg.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_bneg_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: bneg.w +; CHECK: st.w +; CHECK: .size llvm_mips_bneg_w_test +; +@llvm_mips_bneg_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bneg_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_bneg_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bneg_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bneg_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_bneg_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.bneg.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_bneg_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bneg.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_bneg_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: bneg.d +; CHECK: st.d +; CHECK: .size llvm_mips_bneg_d_test +; +@llvm_mips_bset_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bset_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_bset_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bset_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bset_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_bset_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.bset.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_bset_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bset.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_bset_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: bset.b +; CHECK: st.b +; CHECK: .size llvm_mips_bset_b_test +; +@llvm_mips_bset_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bset_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_bset_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bset_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bset_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_bset_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.bset.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_bset_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bset.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_bset_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: bset.h +; CHECK: st.h +; CHECK: .size llvm_mips_bset_h_test +; +@llvm_mips_bset_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bset_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_bset_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bset_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bset_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_bset_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.bset.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_bset_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bset.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_bset_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: bset.w +; CHECK: st.w +; CHECK: .size llvm_mips_bset_w_test +; +@llvm_mips_bset_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bset_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_bset_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bset_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bset_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_bset_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.bset.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_bset_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bset.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_bset_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: bset.d +; CHECK: st.d +; CHECK: .size llvm_mips_bset_d_test +; diff --git a/test/CodeGen/Mips/msa/3r-c.ll b/test/CodeGen/Mips/msa/3r-c.ll new file mode 100644 index 00000000000..56636947ac9 --- /dev/null +++ b/test/CodeGen/Mips/msa/3r-c.ll @@ -0,0 +1,442 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_ceq_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ceq_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ceq_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ceq_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ceq_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ceq_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ceq.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ceq_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ceq.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ceq_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ceq.b +; CHECK: st.b +; CHECK: .size llvm_mips_ceq_b_test +; +@llvm_mips_ceq_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ceq_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ceq_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ceq_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ceq_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ceq_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ceq.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ceq_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ceq.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ceq_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ceq.h +; CHECK: st.h +; CHECK: .size llvm_mips_ceq_h_test +; +@llvm_mips_ceq_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ceq_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ceq_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ceq_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ceq_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ceq_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ceq.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ceq_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ceq.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ceq_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ceq.w +; CHECK: st.w +; CHECK: .size llvm_mips_ceq_w_test +; +@llvm_mips_ceq_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ceq_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ceq_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ceq_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ceq_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ceq_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ceq.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ceq_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ceq.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ceq_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ceq.d +; CHECK: st.d +; CHECK: .size llvm_mips_ceq_d_test +; +@llvm_mips_cle_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_cle_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_cle_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_cle_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_cle_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_cle_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.cle.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_cle_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.cle.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_cle_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: cle_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_cle_s_b_test +; +@llvm_mips_cle_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_cle_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_cle_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_cle_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_cle_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_cle_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.cle.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_cle_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.cle.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_cle_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: cle_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_cle_s_h_test +; +@llvm_mips_cle_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_cle_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_cle_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_cle_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_cle_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_cle_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.cle.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_cle_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.cle.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_cle_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: cle_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_cle_s_w_test +; +@llvm_mips_cle_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_cle_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_cle_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_cle_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_cle_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_cle_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.cle.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_cle_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.cle.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_cle_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: cle_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_cle_s_d_test +; +@llvm_mips_cle_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_cle_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_cle_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_cle_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_cle_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_cle_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.cle.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_cle_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.cle.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_cle_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: cle_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_cle_u_b_test +; +@llvm_mips_cle_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_cle_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_cle_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_cle_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_cle_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_cle_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.cle.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_cle_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.cle.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_cle_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: cle_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_cle_u_h_test +; +@llvm_mips_cle_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_cle_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_cle_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_cle_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_cle_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_cle_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.cle.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_cle_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.cle.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_cle_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: cle_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_cle_u_w_test +; +@llvm_mips_cle_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_cle_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_cle_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_cle_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_cle_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_cle_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.cle.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_cle_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.cle.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_cle_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: cle_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_cle_u_d_test +; +@llvm_mips_clt_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clt_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_clt_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clt_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clt_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_clt_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.clt.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_clt_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clt.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_clt_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: clt_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_clt_s_b_test +; +@llvm_mips_clt_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clt_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_clt_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clt_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clt_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_clt_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.clt.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_clt_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clt.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_clt_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: clt_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_clt_s_h_test +; +@llvm_mips_clt_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clt_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_clt_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clt_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clt_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_clt_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.clt.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_clt_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clt.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_clt_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: clt_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_clt_s_w_test +; +@llvm_mips_clt_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clt_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_clt_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clt_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clt_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_clt_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.clt.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_clt_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clt.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_clt_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: clt_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_clt_s_d_test +; +@llvm_mips_clt_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clt_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_clt_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clt_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clt_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_clt_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.clt.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_clt_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clt.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_clt_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: clt_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_clt_u_b_test +; +@llvm_mips_clt_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clt_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_clt_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clt_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clt_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_clt_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.clt.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_clt_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clt.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_clt_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: clt_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_clt_u_h_test +; +@llvm_mips_clt_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clt_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_clt_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clt_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clt_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_clt_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.clt.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_clt_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clt.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_clt_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: clt_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_clt_u_w_test +; +@llvm_mips_clt_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clt_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_clt_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clt_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clt_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_clt_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.clt.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_clt_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clt.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_clt_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: clt_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_clt_u_d_test +; diff --git a/test/CodeGen/Mips/msa/3r-d.ll b/test/CodeGen/Mips/msa/3r-d.ll new file mode 100644 index 00000000000..3b20743a5da --- /dev/null +++ b/test/CodeGen/Mips/msa/3r-d.ll @@ -0,0 +1,354 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_div_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_div_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_div_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_div_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_div_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_div_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.div.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_div_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.div.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_div_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: div_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_div_s_b_test +; +@llvm_mips_div_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_div_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_div_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_div_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_div_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_div_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.div.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_div_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.div.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_div_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: div_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_div_s_h_test +; +@llvm_mips_div_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_div_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_div_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_div_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_div_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_div_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.div.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_div_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.div.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_div_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: div_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_div_s_w_test +; +@llvm_mips_div_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_div_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_div_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_div_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_div_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_div_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.div.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_div_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.div.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_div_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: div_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_div_s_d_test +; +@llvm_mips_div_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_div_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_div_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_div_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_div_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_div_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.div.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_div_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.div.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_div_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: div_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_div_u_b_test +; +@llvm_mips_div_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_div_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_div_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_div_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_div_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_div_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.div.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_div_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.div.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_div_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: div_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_div_u_h_test +; +@llvm_mips_div_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_div_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_div_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_div_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_div_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_div_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.div.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_div_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.div.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_div_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: div_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_div_u_w_test +; +@llvm_mips_div_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_div_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_div_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_div_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_div_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_div_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.div.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_div_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.div.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_div_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: div_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_div_u_d_test +; +@llvm_mips_dotp_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_dotp_s_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dotp_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_dotp_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_dotp_s_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_dotp_s_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.dotp.s.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_dotp_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.dotp.s.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dotp_s_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: dotp_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_dotp_s_b_test +; +@llvm_mips_dotp_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dotp_s_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dotp_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dotp_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dotp_s_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_dotp_s_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.dotp.s.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_dotp_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dotp.s.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dotp_s_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: dotp_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_dotp_s_h_test +; +@llvm_mips_dotp_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dotp_s_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dotp_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dotp_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dotp_s_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_dotp_s_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.dotp.s.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_dotp_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dotp.s.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dotp_s_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: dotp_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_dotp_s_w_test +; +@llvm_mips_dotp_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dotp_s_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_dotp_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dotp_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dotp_s_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_dotp_s_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.dotp.s.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_dotp_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dotp.s.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_dotp_s_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: dotp_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_dotp_s_d_test +; +@llvm_mips_dotp_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_dotp_u_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dotp_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_dotp_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_dotp_u_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_dotp_u_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.dotp.u.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_dotp_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.dotp.u.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dotp_u_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: dotp_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_dotp_u_b_test +; +@llvm_mips_dotp_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dotp_u_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dotp_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dotp_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dotp_u_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_dotp_u_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.dotp.u.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_dotp_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dotp.u.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dotp_u_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: dotp_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_dotp_u_h_test +; +@llvm_mips_dotp_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dotp_u_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dotp_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dotp_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dotp_u_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_dotp_u_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.dotp.u.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_dotp_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dotp.u.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dotp_u_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: dotp_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_dotp_u_w_test +; +@llvm_mips_dotp_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dotp_u_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_dotp_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dotp_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dotp_u_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_dotp_u_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.dotp.u.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_dotp_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dotp.u.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_dotp_u_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: dotp_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_dotp_u_d_test +; diff --git a/test/CodeGen/Mips/msa/3r-i.ll b/test/CodeGen/Mips/msa/3r-i.ll new file mode 100644 index 00000000000..b5c6a305d65 --- /dev/null +++ b/test/CodeGen/Mips/msa/3r-i.ll @@ -0,0 +1,354 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_ilvev_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ilvev_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ilvev_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ilvev_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ilvev_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ilvev_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ilvev.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ilvev_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ilvev.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ilvev_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ilvev.b +; CHECK: st.b +; CHECK: .size llvm_mips_ilvev_b_test +; +@llvm_mips_ilvev_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ilvev_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ilvev_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ilvev_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ilvev_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ilvev_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ilvev.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ilvev_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ilvev.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ilvev_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ilvev.h +; CHECK: st.h +; CHECK: .size llvm_mips_ilvev_h_test +; +@llvm_mips_ilvev_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ilvev_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ilvev_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ilvev_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ilvev_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ilvev_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ilvev.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ilvev_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ilvev.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ilvev_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ilvev.w +; CHECK: st.w +; CHECK: .size llvm_mips_ilvev_w_test +; +@llvm_mips_ilvev_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ilvev_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ilvev_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ilvev_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ilvev_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ilvev_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ilvev.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ilvev_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ilvev.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ilvev_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ilvev.d +; CHECK: st.d +; CHECK: .size llvm_mips_ilvev_d_test +; +@llvm_mips_ilvl_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ilvl_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ilvl_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ilvl_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ilvl_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ilvl_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ilvl.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ilvl_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ilvl.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ilvl_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ilvl.b +; CHECK: st.b +; CHECK: .size llvm_mips_ilvl_b_test +; +@llvm_mips_ilvl_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ilvl_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ilvl_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ilvl_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ilvl_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ilvl_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ilvl.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ilvl_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ilvl.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ilvl_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ilvl.h +; CHECK: st.h +; CHECK: .size llvm_mips_ilvl_h_test +; +@llvm_mips_ilvl_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ilvl_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ilvl_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ilvl_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ilvl_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ilvl_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ilvl.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ilvl_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ilvl.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ilvl_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ilvl.w +; CHECK: st.w +; CHECK: .size llvm_mips_ilvl_w_test +; +@llvm_mips_ilvl_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ilvl_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ilvl_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ilvl_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ilvl_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ilvl_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ilvl.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ilvl_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ilvl.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ilvl_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ilvl.d +; CHECK: st.d +; CHECK: .size llvm_mips_ilvl_d_test +; +@llvm_mips_ilvod_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ilvod_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ilvod_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ilvod_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ilvod_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ilvod_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ilvod.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ilvod_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ilvod.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ilvod_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ilvod.b +; CHECK: st.b +; CHECK: .size llvm_mips_ilvod_b_test +; +@llvm_mips_ilvod_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ilvod_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ilvod_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ilvod_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ilvod_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ilvod_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ilvod.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ilvod_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ilvod.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ilvod_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ilvod.h +; CHECK: st.h +; CHECK: .size llvm_mips_ilvod_h_test +; +@llvm_mips_ilvod_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ilvod_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ilvod_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ilvod_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ilvod_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ilvod_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ilvod.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ilvod_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ilvod.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ilvod_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ilvod.w +; CHECK: st.w +; CHECK: .size llvm_mips_ilvod_w_test +; +@llvm_mips_ilvod_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ilvod_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ilvod_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ilvod_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ilvod_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ilvod_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ilvod.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ilvod_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ilvod.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ilvod_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ilvod.d +; CHECK: st.d +; CHECK: .size llvm_mips_ilvod_d_test +; +@llvm_mips_ilvr_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ilvr_b_ARG2 = global <16 x i8> , align 16 +@llvm_mips_ilvr_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ilvr_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ilvr_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_ilvr_b_ARG2 + %2 = tail call <16 x i8> @llvm.mips.ilvr.b(<16 x i8> %0, <16 x i8> %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_ilvr_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ilvr.b(<16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_ilvr_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ilvr.b +; CHECK: st.b +; CHECK: .size llvm_mips_ilvr_b_test +; +@llvm_mips_ilvr_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ilvr_h_ARG2 = global <8 x i16> , align 16 +@llvm_mips_ilvr_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ilvr_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ilvr_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_ilvr_h_ARG2 + %2 = tail call <8 x i16> @llvm.mips.ilvr.h(<8 x i16> %0, <8 x i16> %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_ilvr_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ilvr.h(<8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_ilvr_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ilvr.h +; CHECK: st.h +; CHECK: .size llvm_mips_ilvr_h_test +; +@llvm_mips_ilvr_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ilvr_w_ARG2 = global <4 x i32> , align 16 +@llvm_mips_ilvr_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ilvr_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ilvr_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_ilvr_w_ARG2 + %2 = tail call <4 x i32> @llvm.mips.ilvr.w(<4 x i32> %0, <4 x i32> %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_ilvr_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ilvr.w(<4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_ilvr_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ilvr.w +; CHECK: st.w +; CHECK: .size llvm_mips_ilvr_w_test +; +@llvm_mips_ilvr_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ilvr_d_ARG2 = global <2 x i64> , align 16 +@llvm_mips_ilvr_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ilvr_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ilvr_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_ilvr_d_ARG2 + %2 = tail call <2 x i64> @llvm.mips.ilvr.d(<2 x i64> %0, <2 x i64> %1) + store <2 x i64> %2, <2 x i64>* @llvm_mips_ilvr_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ilvr.d(<2 x i64>, <2 x i64>) nounwind + +; CHECK: llvm_mips_ilvr_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: ilvr.d +; CHECK: st.d +; CHECK: .size llvm_mips_ilvr_d_test +; diff --git a/test/CodeGen/Mips/msa/3r_4r_widen.ll b/test/CodeGen/Mips/msa/3r_4r_widen.ll new file mode 100644 index 00000000000..891d11a2f4e --- /dev/null +++ b/test/CodeGen/Mips/msa/3r_4r_widen.ll @@ -0,0 +1,302 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_dpadd_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dpadd_s_h_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dpadd_s_h_ARG3 = global <16 x i8> , align 16 +@llvm_mips_dpadd_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dpadd_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dpadd_s_h_ARG1 + %1 = load <16 x i8>* @llvm_mips_dpadd_s_h_ARG2 + %2 = load <16 x i8>* @llvm_mips_dpadd_s_h_ARG3 + %3 = tail call <8 x i16> @llvm.mips.dpadd.s.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) + store <8 x i16> %3, <8 x i16>* @llvm_mips_dpadd_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dpadd.s.h(<8 x i16>, <16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dpadd_s_h_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ld.h +; CHECK: dpadd_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_dpadd_s_h_test +; +@llvm_mips_dpadd_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dpadd_s_w_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dpadd_s_w_ARG3 = global <8 x i16> , align 16 +@llvm_mips_dpadd_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dpadd_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dpadd_s_w_ARG1 + %1 = load <8 x i16>* @llvm_mips_dpadd_s_w_ARG2 + %2 = load <8 x i16>* @llvm_mips_dpadd_s_w_ARG3 + %3 = tail call <4 x i32> @llvm.mips.dpadd.s.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) + store <4 x i32> %3, <4 x i32>* @llvm_mips_dpadd_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dpadd.s.w(<4 x i32>, <8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dpadd_s_w_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ld.w +; CHECK: dpadd_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_dpadd_s_w_test +; +@llvm_mips_dpadd_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dpadd_s_d_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dpadd_s_d_ARG3 = global <4 x i32> , align 16 +@llvm_mips_dpadd_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dpadd_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dpadd_s_d_ARG1 + %1 = load <4 x i32>* @llvm_mips_dpadd_s_d_ARG2 + %2 = load <4 x i32>* @llvm_mips_dpadd_s_d_ARG3 + %3 = tail call <2 x i64> @llvm.mips.dpadd.s.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) + store <2 x i64> %3, <2 x i64>* @llvm_mips_dpadd_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dpadd.s.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dpadd_s_d_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ld.d +; CHECK: dpadd_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_dpadd_s_d_test +; +@llvm_mips_dpadd_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dpadd_u_h_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dpadd_u_h_ARG3 = global <16 x i8> , align 16 +@llvm_mips_dpadd_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dpadd_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dpadd_u_h_ARG1 + %1 = load <16 x i8>* @llvm_mips_dpadd_u_h_ARG2 + %2 = load <16 x i8>* @llvm_mips_dpadd_u_h_ARG3 + %3 = tail call <8 x i16> @llvm.mips.dpadd.u.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) + store <8 x i16> %3, <8 x i16>* @llvm_mips_dpadd_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dpadd.u.h(<8 x i16>, <16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dpadd_u_h_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ld.h +; CHECK: dpadd_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_dpadd_u_h_test +; +@llvm_mips_dpadd_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dpadd_u_w_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dpadd_u_w_ARG3 = global <8 x i16> , align 16 +@llvm_mips_dpadd_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dpadd_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dpadd_u_w_ARG1 + %1 = load <8 x i16>* @llvm_mips_dpadd_u_w_ARG2 + %2 = load <8 x i16>* @llvm_mips_dpadd_u_w_ARG3 + %3 = tail call <4 x i32> @llvm.mips.dpadd.u.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) + store <4 x i32> %3, <4 x i32>* @llvm_mips_dpadd_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dpadd.u.w(<4 x i32>, <8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dpadd_u_w_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ld.w +; CHECK: dpadd_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_dpadd_u_w_test +; +@llvm_mips_dpadd_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dpadd_u_d_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dpadd_u_d_ARG3 = global <4 x i32> , align 16 +@llvm_mips_dpadd_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dpadd_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dpadd_u_d_ARG1 + %1 = load <4 x i32>* @llvm_mips_dpadd_u_d_ARG2 + %2 = load <4 x i32>* @llvm_mips_dpadd_u_d_ARG3 + %3 = tail call <2 x i64> @llvm.mips.dpadd.u.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) + store <2 x i64> %3, <2 x i64>* @llvm_mips_dpadd_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dpadd.u.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dpadd_u_d_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ld.d +; CHECK: dpadd_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_dpadd_u_d_test +; +@llvm_mips_dpsub_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dpsub_s_h_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dpsub_s_h_ARG3 = global <16 x i8> , align 16 +@llvm_mips_dpsub_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dpsub_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dpsub_s_h_ARG1 + %1 = load <16 x i8>* @llvm_mips_dpsub_s_h_ARG2 + %2 = load <16 x i8>* @llvm_mips_dpsub_s_h_ARG3 + %3 = tail call <8 x i16> @llvm.mips.dpsub.s.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) + store <8 x i16> %3, <8 x i16>* @llvm_mips_dpsub_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dpsub.s.h(<8 x i16>, <16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dpsub_s_h_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ld.h +; CHECK: dpsub_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_dpsub_s_h_test +; +@llvm_mips_dpsub_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dpsub_s_w_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dpsub_s_w_ARG3 = global <8 x i16> , align 16 +@llvm_mips_dpsub_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dpsub_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dpsub_s_w_ARG1 + %1 = load <8 x i16>* @llvm_mips_dpsub_s_w_ARG2 + %2 = load <8 x i16>* @llvm_mips_dpsub_s_w_ARG3 + %3 = tail call <4 x i32> @llvm.mips.dpsub.s.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) + store <4 x i32> %3, <4 x i32>* @llvm_mips_dpsub_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dpsub.s.w(<4 x i32>, <8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dpsub_s_w_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ld.w +; CHECK: dpsub_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_dpsub_s_w_test +; +@llvm_mips_dpsub_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dpsub_s_d_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dpsub_s_d_ARG3 = global <4 x i32> , align 16 +@llvm_mips_dpsub_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dpsub_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dpsub_s_d_ARG1 + %1 = load <4 x i32>* @llvm_mips_dpsub_s_d_ARG2 + %2 = load <4 x i32>* @llvm_mips_dpsub_s_d_ARG3 + %3 = tail call <2 x i64> @llvm.mips.dpsub.s.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) + store <2 x i64> %3, <2 x i64>* @llvm_mips_dpsub_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dpsub.s.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dpsub_s_d_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ld.d +; CHECK: dpsub_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_dpsub_s_d_test +; +@llvm_mips_dpsub_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_dpsub_u_h_ARG2 = global <16 x i8> , align 16 +@llvm_mips_dpsub_u_h_ARG3 = global <16 x i8> , align 16 +@llvm_mips_dpsub_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_dpsub_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_dpsub_u_h_ARG1 + %1 = load <16 x i8>* @llvm_mips_dpsub_u_h_ARG2 + %2 = load <16 x i8>* @llvm_mips_dpsub_u_h_ARG3 + %3 = tail call <8 x i16> @llvm.mips.dpsub.u.h(<8 x i16> %0, <16 x i8> %1, <16 x i8> %2) + store <8 x i16> %3, <8 x i16>* @llvm_mips_dpsub_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.dpsub.u.h(<8 x i16>, <16 x i8>, <16 x i8>) nounwind + +; CHECK: llvm_mips_dpsub_u_h_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: ld.h +; CHECK: dpsub_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_dpsub_u_h_test +; +@llvm_mips_dpsub_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_dpsub_u_w_ARG2 = global <8 x i16> , align 16 +@llvm_mips_dpsub_u_w_ARG3 = global <8 x i16> , align 16 +@llvm_mips_dpsub_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_dpsub_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_dpsub_u_w_ARG1 + %1 = load <8 x i16>* @llvm_mips_dpsub_u_w_ARG2 + %2 = load <8 x i16>* @llvm_mips_dpsub_u_w_ARG3 + %3 = tail call <4 x i32> @llvm.mips.dpsub.u.w(<4 x i32> %0, <8 x i16> %1, <8 x i16> %2) + store <4 x i32> %3, <4 x i32>* @llvm_mips_dpsub_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.dpsub.u.w(<4 x i32>, <8 x i16>, <8 x i16>) nounwind + +; CHECK: llvm_mips_dpsub_u_w_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: ld.w +; CHECK: dpsub_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_dpsub_u_w_test +; +@llvm_mips_dpsub_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_dpsub_u_d_ARG2 = global <4 x i32> , align 16 +@llvm_mips_dpsub_u_d_ARG3 = global <4 x i32> , align 16 +@llvm_mips_dpsub_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_dpsub_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_dpsub_u_d_ARG1 + %1 = load <4 x i32>* @llvm_mips_dpsub_u_d_ARG2 + %2 = load <4 x i32>* @llvm_mips_dpsub_u_d_ARG3 + %3 = tail call <2 x i64> @llvm.mips.dpsub.u.d(<2 x i64> %0, <4 x i32> %1, <4 x i32> %2) + store <2 x i64> %3, <2 x i64>* @llvm_mips_dpsub_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.dpsub.u.d(<2 x i64>, <4 x i32>, <4 x i32>) nounwind + +; CHECK: llvm_mips_dpsub_u_d_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: ld.d +; CHECK: dpsub_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_dpsub_u_d_test +; diff --git a/test/CodeGen/Mips/msa/elm_copy.ll b/test/CodeGen/Mips/msa/elm_copy.ll new file mode 100644 index 00000000000..f31bb506db2 --- /dev/null +++ b/test/CodeGen/Mips/msa/elm_copy.ll @@ -0,0 +1,116 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_copy_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_copy_s_b_RES = global i32 0, align 16 + +define void @llvm_mips_copy_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_copy_s_b_ARG1 + %1 = tail call i32 @llvm.mips.copy.s.b(<16 x i8> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_s_b_RES + ret void +} + +declare i32 @llvm.mips.copy.s.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_copy_s_b_test: +; CHECK: ld.b +; CHECK: copy_s.b +; CHECK: sw +; CHECK: .size llvm_mips_copy_s_b_test +; +@llvm_mips_copy_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_copy_s_h_RES = global i32 0, align 16 + +define void @llvm_mips_copy_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_copy_s_h_ARG1 + %1 = tail call i32 @llvm.mips.copy.s.h(<8 x i16> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_s_h_RES + ret void +} + +declare i32 @llvm.mips.copy.s.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_copy_s_h_test: +; CHECK: ld.h +; CHECK: copy_s.h +; CHECK: sw +; CHECK: .size llvm_mips_copy_s_h_test +; +@llvm_mips_copy_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_copy_s_w_RES = global i32 0, align 16 + +define void @llvm_mips_copy_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_copy_s_w_ARG1 + %1 = tail call i32 @llvm.mips.copy.s.w(<4 x i32> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_s_w_RES + ret void +} + +declare i32 @llvm.mips.copy.s.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_copy_s_w_test: +; CHECK: ld.w +; CHECK: copy_s.w +; CHECK: sw +; CHECK: .size llvm_mips_copy_s_w_test +; +@llvm_mips_copy_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_copy_u_b_RES = global i32 0, align 16 + +define void @llvm_mips_copy_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_copy_u_b_ARG1 + %1 = tail call i32 @llvm.mips.copy.u.b(<16 x i8> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_u_b_RES + ret void +} + +declare i32 @llvm.mips.copy.u.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_copy_u_b_test: +; CHECK: ld.b +; CHECK: copy_u.b +; CHECK: sw +; CHECK: .size llvm_mips_copy_u_b_test +; +@llvm_mips_copy_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_copy_u_h_RES = global i32 0, align 16 + +define void @llvm_mips_copy_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_copy_u_h_ARG1 + %1 = tail call i32 @llvm.mips.copy.u.h(<8 x i16> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_u_h_RES + ret void +} + +declare i32 @llvm.mips.copy.u.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_copy_u_h_test: +; CHECK: ld.h +; CHECK: copy_u.h +; CHECK: sw +; CHECK: .size llvm_mips_copy_u_h_test +; +@llvm_mips_copy_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_copy_u_w_RES = global i32 0, align 16 + +define void @llvm_mips_copy_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_copy_u_w_ARG1 + %1 = tail call i32 @llvm.mips.copy.u.w(<4 x i32> %0, i32 1) + store i32 %1, i32* @llvm_mips_copy_u_w_RES + ret void +} + +declare i32 @llvm.mips.copy.u.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_copy_u_w_test: +; CHECK: ld.w +; CHECK: copy_u.w +; CHECK: sw +; CHECK: .size llvm_mips_copy_u_w_test +; diff --git a/test/CodeGen/Mips/msa/elm_insv.ll b/test/CodeGen/Mips/msa/elm_insv.ll new file mode 100644 index 00000000000..c5378eb1681 --- /dev/null +++ b/test/CodeGen/Mips/msa/elm_insv.ll @@ -0,0 +1,68 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_insert_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_insert_b_ARG3 = global i32 27, align 16 +@llvm_mips_insert_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_insert_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_insert_b_ARG1 + %1 = load i32* @llvm_mips_insert_b_ARG3 + %2 = tail call <16 x i8> @llvm.mips.insert.b(<16 x i8> %0, i32 1, i32 %1) + store <16 x i8> %2, <16 x i8>* @llvm_mips_insert_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.insert.b(<16 x i8>, i32, i32) nounwind + +; CHECK: llvm_mips_insert_b_test: +; CHECK: lw +; CHECK: ld.b +; CHECK: insert.b +; CHECK: st.b +; CHECK: .size llvm_mips_insert_b_test +; +@llvm_mips_insert_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_insert_h_ARG3 = global i32 27, align 16 +@llvm_mips_insert_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_insert_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_insert_h_ARG1 + %1 = load i32* @llvm_mips_insert_h_ARG3 + %2 = tail call <8 x i16> @llvm.mips.insert.h(<8 x i16> %0, i32 1, i32 %1) + store <8 x i16> %2, <8 x i16>* @llvm_mips_insert_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.insert.h(<8 x i16>, i32, i32) nounwind + +; CHECK: llvm_mips_insert_h_test: +; CHECK: lw +; CHECK: ld.h +; CHECK: insert.h +; CHECK: st.h +; CHECK: .size llvm_mips_insert_h_test +; +@llvm_mips_insert_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_insert_w_ARG3 = global i32 27, align 16 +@llvm_mips_insert_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_insert_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_insert_w_ARG1 + %1 = load i32* @llvm_mips_insert_w_ARG3 + %2 = tail call <4 x i32> @llvm.mips.insert.w(<4 x i32> %0, i32 1, i32 %1) + store <4 x i32> %2, <4 x i32>* @llvm_mips_insert_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.insert.w(<4 x i32>, i32, i32) nounwind + +; CHECK: llvm_mips_insert_w_test: +; CHECK: lw +; CHECK: ld.w +; CHECK: insert.w +; CHECK: st.w +; CHECK: .size llvm_mips_insert_w_test +; diff --git a/test/CodeGen/Mips/msa/i5-a.ll b/test/CodeGen/Mips/msa/i5-a.ll new file mode 100644 index 00000000000..fe93534b2ff --- /dev/null +++ b/test/CodeGen/Mips/msa/i5-a.ll @@ -0,0 +1,78 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_addvi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_addvi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_addvi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_addvi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.addvi.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_addvi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.addvi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_addvi_b_test: +; CHECK: ld.b +; CHECK: addvi.b +; CHECK: st.b +; CHECK: .size llvm_mips_addvi_b_test +; +@llvm_mips_addvi_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_addvi_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_addvi_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_addvi_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.addvi.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_addvi_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.addvi.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_addvi_h_test: +; CHECK: ld.h +; CHECK: addvi.h +; CHECK: st.h +; CHECK: .size llvm_mips_addvi_h_test +; +@llvm_mips_addvi_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_addvi_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_addvi_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_addvi_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.addvi.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_addvi_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.addvi.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_addvi_w_test: +; CHECK: ld.w +; CHECK: addvi.w +; CHECK: st.w +; CHECK: .size llvm_mips_addvi_w_test +; +@llvm_mips_addvi_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_addvi_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_addvi_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_addvi_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.addvi.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_addvi_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.addvi.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_addvi_d_test: +; CHECK: ld.d +; CHECK: addvi.d +; CHECK: st.d +; CHECK: .size llvm_mips_addvi_d_test +; diff --git a/test/CodeGen/Mips/msa/i5-b.ll b/test/CodeGen/Mips/msa/i5-b.ll new file mode 100644 index 00000000000..87d4527124a --- /dev/null +++ b/test/CodeGen/Mips/msa/i5-b.ll @@ -0,0 +1,382 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_bclri_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bclri_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bclri_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bclri_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bclri.b(<16 x i8> %0, i32 7) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bclri_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bclri.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bclri_b_test: +; CHECK: ld.b +; CHECK: bclri.b +; CHECK: st.b +; CHECK: .size llvm_mips_bclri_b_test +; +@llvm_mips_bclri_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bclri_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bclri_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bclri_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.bclri.h(<8 x i16> %0, i32 7) + store <8 x i16> %1, <8 x i16>* @llvm_mips_bclri_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bclri.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_bclri_h_test: +; CHECK: ld.h +; CHECK: bclri.h +; CHECK: st.h +; CHECK: .size llvm_mips_bclri_h_test +; +@llvm_mips_bclri_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bclri_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bclri_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bclri_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.bclri.w(<4 x i32> %0, i32 7) + store <4 x i32> %1, <4 x i32>* @llvm_mips_bclri_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bclri.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_bclri_w_test: +; CHECK: ld.w +; CHECK: bclri.w +; CHECK: st.w +; CHECK: .size llvm_mips_bclri_w_test +; +@llvm_mips_bclri_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bclri_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bclri_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bclri_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.bclri.d(<2 x i64> %0, i32 7) + store <2 x i64> %1, <2 x i64>* @llvm_mips_bclri_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bclri.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_bclri_d_test: +; CHECK: ld.d +; CHECK: bclri.d +; CHECK: st.d +; CHECK: .size llvm_mips_bclri_d_test +; +@llvm_mips_binsli_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_binsli_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_binsli_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_binsli_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.binsli.b(<16 x i8> %0, i32 7) + store <16 x i8> %1, <16 x i8>* @llvm_mips_binsli_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.binsli.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_binsli_b_test: +; CHECK: ld.b +; CHECK: binsli.b +; CHECK: st.b +; CHECK: .size llvm_mips_binsli_b_test +; +@llvm_mips_binsli_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_binsli_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_binsli_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_binsli_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.binsli.h(<8 x i16> %0, i32 7) + store <8 x i16> %1, <8 x i16>* @llvm_mips_binsli_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.binsli.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_binsli_h_test: +; CHECK: ld.h +; CHECK: binsli.h +; CHECK: st.h +; CHECK: .size llvm_mips_binsli_h_test +; +@llvm_mips_binsli_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_binsli_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_binsli_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_binsli_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.binsli.w(<4 x i32> %0, i32 7) + store <4 x i32> %1, <4 x i32>* @llvm_mips_binsli_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.binsli.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_binsli_w_test: +; CHECK: ld.w +; CHECK: binsli.w +; CHECK: st.w +; CHECK: .size llvm_mips_binsli_w_test +; +@llvm_mips_binsli_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_binsli_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_binsli_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_binsli_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.binsli.d(<2 x i64> %0, i32 7) + store <2 x i64> %1, <2 x i64>* @llvm_mips_binsli_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.binsli.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_binsli_d_test: +; CHECK: ld.d +; CHECK: binsli.d +; CHECK: st.d +; CHECK: .size llvm_mips_binsli_d_test +; +@llvm_mips_binsri_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_binsri_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_binsri_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_binsri_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.binsri.b(<16 x i8> %0, i32 7) + store <16 x i8> %1, <16 x i8>* @llvm_mips_binsri_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.binsri.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_binsri_b_test: +; CHECK: ld.b +; CHECK: binsri.b +; CHECK: st.b +; CHECK: .size llvm_mips_binsri_b_test +; +@llvm_mips_binsri_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_binsri_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_binsri_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_binsri_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.binsri.h(<8 x i16> %0, i32 7) + store <8 x i16> %1, <8 x i16>* @llvm_mips_binsri_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.binsri.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_binsri_h_test: +; CHECK: ld.h +; CHECK: binsri.h +; CHECK: st.h +; CHECK: .size llvm_mips_binsri_h_test +; +@llvm_mips_binsri_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_binsri_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_binsri_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_binsri_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.binsri.w(<4 x i32> %0, i32 7) + store <4 x i32> %1, <4 x i32>* @llvm_mips_binsri_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.binsri.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_binsri_w_test: +; CHECK: ld.w +; CHECK: binsri.w +; CHECK: st.w +; CHECK: .size llvm_mips_binsri_w_test +; +@llvm_mips_binsri_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_binsri_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_binsri_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_binsri_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.binsri.d(<2 x i64> %0, i32 7) + store <2 x i64> %1, <2 x i64>* @llvm_mips_binsri_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.binsri.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_binsri_d_test: +; CHECK: ld.d +; CHECK: binsri.d +; CHECK: st.d +; CHECK: .size llvm_mips_binsri_d_test +; +@llvm_mips_bnegi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bnegi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bnegi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bnegi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bnegi.b(<16 x i8> %0, i32 7) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bnegi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bnegi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bnegi_b_test: +; CHECK: ld.b +; CHECK: bnegi.b +; CHECK: st.b +; CHECK: .size llvm_mips_bnegi_b_test +; +@llvm_mips_bnegi_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bnegi_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bnegi_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bnegi_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.bnegi.h(<8 x i16> %0, i32 7) + store <8 x i16> %1, <8 x i16>* @llvm_mips_bnegi_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bnegi.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_bnegi_h_test: +; CHECK: ld.h +; CHECK: bnegi.h +; CHECK: st.h +; CHECK: .size llvm_mips_bnegi_h_test +; +@llvm_mips_bnegi_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bnegi_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bnegi_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bnegi_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.bnegi.w(<4 x i32> %0, i32 7) + store <4 x i32> %1, <4 x i32>* @llvm_mips_bnegi_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bnegi.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_bnegi_w_test: +; CHECK: ld.w +; CHECK: bnegi.w +; CHECK: st.w +; CHECK: .size llvm_mips_bnegi_w_test +; +@llvm_mips_bnegi_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bnegi_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bnegi_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bnegi_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.bnegi.d(<2 x i64> %0, i32 7) + store <2 x i64> %1, <2 x i64>* @llvm_mips_bnegi_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bnegi.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_bnegi_d_test: +; CHECK: ld.d +; CHECK: bnegi.d +; CHECK: st.d +; CHECK: .size llvm_mips_bnegi_d_test +; +@llvm_mips_bseti_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bseti_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bseti_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bseti_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bseti.b(<16 x i8> %0, i32 7) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bseti_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bseti.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bseti_b_test: +; CHECK: ld.b +; CHECK: bseti.b +; CHECK: st.b +; CHECK: .size llvm_mips_bseti_b_test +; +@llvm_mips_bseti_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_bseti_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_bseti_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_bseti_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.bseti.h(<8 x i16> %0, i32 7) + store <8 x i16> %1, <8 x i16>* @llvm_mips_bseti_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.bseti.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_bseti_h_test: +; CHECK: ld.h +; CHECK: bseti.h +; CHECK: st.h +; CHECK: .size llvm_mips_bseti_h_test +; +@llvm_mips_bseti_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_bseti_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_bseti_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_bseti_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.bseti.w(<4 x i32> %0, i32 7) + store <4 x i32> %1, <4 x i32>* @llvm_mips_bseti_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.bseti.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_bseti_w_test: +; CHECK: ld.w +; CHECK: bseti.w +; CHECK: st.w +; CHECK: .size llvm_mips_bseti_w_test +; +@llvm_mips_bseti_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_bseti_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_bseti_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_bseti_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.bseti.d(<2 x i64> %0, i32 7) + store <2 x i64> %1, <2 x i64>* @llvm_mips_bseti_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.bseti.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_bseti_d_test: +; CHECK: ld.d +; CHECK: bseti.d +; CHECK: st.d +; CHECK: .size llvm_mips_bseti_d_test +; diff --git a/test/CodeGen/Mips/msa/i5-c.ll b/test/CodeGen/Mips/msa/i5-c.ll new file mode 100644 index 00000000000..5e9eed8b2d0 --- /dev/null +++ b/test/CodeGen/Mips/msa/i5-c.ll @@ -0,0 +1,382 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_ceqi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_ceqi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_ceqi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_ceqi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.ceqi.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_ceqi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.ceqi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_ceqi_b_test: +; CHECK: ld.b +; CHECK: ceqi.b +; CHECK: st.b +; CHECK: .size llvm_mips_ceqi_b_test +; +@llvm_mips_ceqi_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_ceqi_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_ceqi_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_ceqi_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.ceqi.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_ceqi_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.ceqi.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_ceqi_h_test: +; CHECK: ld.h +; CHECK: ceqi.h +; CHECK: st.h +; CHECK: .size llvm_mips_ceqi_h_test +; +@llvm_mips_ceqi_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_ceqi_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_ceqi_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_ceqi_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.ceqi.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_ceqi_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.ceqi.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_ceqi_w_test: +; CHECK: ld.w +; CHECK: ceqi.w +; CHECK: st.w +; CHECK: .size llvm_mips_ceqi_w_test +; +@llvm_mips_ceqi_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_ceqi_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_ceqi_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_ceqi_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.ceqi.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_ceqi_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.ceqi.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_ceqi_d_test: +; CHECK: ld.d +; CHECK: ceqi.d +; CHECK: st.d +; CHECK: .size llvm_mips_ceqi_d_test +; +@llvm_mips_clei_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clei_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clei_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clei_s_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.clei.s.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_clei_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clei.s.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_clei_s_b_test: +; CHECK: ld.b +; CHECK: clei_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_clei_s_b_test +; +@llvm_mips_clei_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clei_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clei_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clei_s_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.clei.s.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_clei_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clei.s.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_clei_s_h_test: +; CHECK: ld.h +; CHECK: clei_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_clei_s_h_test +; +@llvm_mips_clei_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clei_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clei_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clei_s_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.clei.s.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_clei_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clei.s.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_clei_s_w_test: +; CHECK: ld.w +; CHECK: clei_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_clei_s_w_test +; +@llvm_mips_clei_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clei_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clei_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clei_s_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.clei.s.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_clei_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clei.s.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_clei_s_d_test: +; CHECK: ld.d +; CHECK: clei_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_clei_s_d_test +; +@llvm_mips_clei_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clei_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clei_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clei_u_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.clei.u.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_clei_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clei.u.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_clei_u_b_test: +; CHECK: ld.b +; CHECK: clei_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_clei_u_b_test +; +@llvm_mips_clei_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clei_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clei_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clei_u_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.clei.u.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_clei_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clei.u.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_clei_u_h_test: +; CHECK: ld.h +; CHECK: clei_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_clei_u_h_test +; +@llvm_mips_clei_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clei_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clei_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clei_u_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.clei.u.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_clei_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clei.u.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_clei_u_w_test: +; CHECK: ld.w +; CHECK: clei_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_clei_u_w_test +; +@llvm_mips_clei_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clei_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clei_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clei_u_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.clei.u.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_clei_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clei.u.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_clei_u_d_test: +; CHECK: ld.d +; CHECK: clei_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_clei_u_d_test +; +@llvm_mips_clti_s_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clti_s_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clti_s_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clti_s_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.clti.s.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_clti_s_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clti.s.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_clti_s_b_test: +; CHECK: ld.b +; CHECK: clti_s.b +; CHECK: st.b +; CHECK: .size llvm_mips_clti_s_b_test +; +@llvm_mips_clti_s_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clti_s_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clti_s_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clti_s_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.clti.s.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_clti_s_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clti.s.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_clti_s_h_test: +; CHECK: ld.h +; CHECK: clti_s.h +; CHECK: st.h +; CHECK: .size llvm_mips_clti_s_h_test +; +@llvm_mips_clti_s_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clti_s_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clti_s_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clti_s_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.clti.s.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_clti_s_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clti.s.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_clti_s_w_test: +; CHECK: ld.w +; CHECK: clti_s.w +; CHECK: st.w +; CHECK: .size llvm_mips_clti_s_w_test +; +@llvm_mips_clti_s_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clti_s_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clti_s_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clti_s_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.clti.s.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_clti_s_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clti.s.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_clti_s_d_test: +; CHECK: ld.d +; CHECK: clti_s.d +; CHECK: st.d +; CHECK: .size llvm_mips_clti_s_d_test +; +@llvm_mips_clti_u_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_clti_u_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_clti_u_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_clti_u_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.clti.u.b(<16 x i8> %0, i32 14) + store <16 x i8> %1, <16 x i8>* @llvm_mips_clti_u_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.clti.u.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_clti_u_b_test: +; CHECK: ld.b +; CHECK: clti_u.b +; CHECK: st.b +; CHECK: .size llvm_mips_clti_u_b_test +; +@llvm_mips_clti_u_h_ARG1 = global <8 x i16> , align 16 +@llvm_mips_clti_u_h_RES = global <8 x i16> , align 16 + +define void @llvm_mips_clti_u_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_clti_u_h_ARG1 + %1 = tail call <8 x i16> @llvm.mips.clti.u.h(<8 x i16> %0, i32 14) + store <8 x i16> %1, <8 x i16>* @llvm_mips_clti_u_h_RES + ret void +} + +declare <8 x i16> @llvm.mips.clti.u.h(<8 x i16>, i32) nounwind + +; CHECK: llvm_mips_clti_u_h_test: +; CHECK: ld.h +; CHECK: clti_u.h +; CHECK: st.h +; CHECK: .size llvm_mips_clti_u_h_test +; +@llvm_mips_clti_u_w_ARG1 = global <4 x i32> , align 16 +@llvm_mips_clti_u_w_RES = global <4 x i32> , align 16 + +define void @llvm_mips_clti_u_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_clti_u_w_ARG1 + %1 = tail call <4 x i32> @llvm.mips.clti.u.w(<4 x i32> %0, i32 14) + store <4 x i32> %1, <4 x i32>* @llvm_mips_clti_u_w_RES + ret void +} + +declare <4 x i32> @llvm.mips.clti.u.w(<4 x i32>, i32) nounwind + +; CHECK: llvm_mips_clti_u_w_test: +; CHECK: ld.w +; CHECK: clti_u.w +; CHECK: st.w +; CHECK: .size llvm_mips_clti_u_w_test +; +@llvm_mips_clti_u_d_ARG1 = global <2 x i64> , align 16 +@llvm_mips_clti_u_d_RES = global <2 x i64> , align 16 + +define void @llvm_mips_clti_u_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_clti_u_d_ARG1 + %1 = tail call <2 x i64> @llvm.mips.clti.u.d(<2 x i64> %0, i32 14) + store <2 x i64> %1, <2 x i64>* @llvm_mips_clti_u_d_RES + ret void +} + +declare <2 x i64> @llvm.mips.clti.u.d(<2 x i64>, i32) nounwind + +; CHECK: llvm_mips_clti_u_d_test: +; CHECK: ld.d +; CHECK: clti_u.d +; CHECK: st.d +; CHECK: .size llvm_mips_clti_u_d_test +; diff --git a/test/CodeGen/Mips/msa/i8.ll b/test/CodeGen/Mips/msa/i8.ll new file mode 100644 index 00000000000..4dc30e3b7fa --- /dev/null +++ b/test/CodeGen/Mips/msa/i8.ll @@ -0,0 +1,78 @@ +; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s + +@llvm_mips_andi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_andi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_andi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_andi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.andi.b(<16 x i8> %0, i32 25) + store <16 x i8> %1, <16 x i8>* @llvm_mips_andi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.andi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_andi_b_test: +; CHECK: ld.b +; CHECK: andi.b +; CHECK: st.b +; CHECK: .size llvm_mips_andi_b_test +; +@llvm_mips_bmnzi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bmnzi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bmnzi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bmnzi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bmnzi.b(<16 x i8> %0, i32 25) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bmnzi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bmnzi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bmnzi_b_test: +; CHECK: ld.b +; CHECK: bmnzi.b +; CHECK: st.b +; CHECK: .size llvm_mips_bmnzi_b_test +; +@llvm_mips_bmzi_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bmzi_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bmzi_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bmzi_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bmzi.b(<16 x i8> %0, i32 25) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bmzi_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bmzi.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bmzi_b_test: +; CHECK: ld.b +; CHECK: bmzi.b +; CHECK: st.b +; CHECK: .size llvm_mips_bmzi_b_test +; +@llvm_mips_bseli_b_ARG1 = global <16 x i8> , align 16 +@llvm_mips_bseli_b_RES = global <16 x i8> , align 16 + +define void @llvm_mips_bseli_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_bseli_b_ARG1 + %1 = tail call <16 x i8> @llvm.mips.bseli.b(<16 x i8> %0, i32 25) + store <16 x i8> %1, <16 x i8>* @llvm_mips_bseli_b_RES + ret void +} + +declare <16 x i8> @llvm.mips.bseli.b(<16 x i8>, i32) nounwind + +; CHECK: llvm_mips_bseli_b_test: +; CHECK: ld.b +; CHECK: bseli.b +; CHECK: st.b +; CHECK: .size llvm_mips_bseli_b_test +; -- 2.34.1