; RUN: opt < %s -instcombine -S | FileCheck %s target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128" define i16 @test1(float %f) { entry: ; CHECK-LABEL: @test1( ; CHECK: fmul float ; CHECK-NOT: insertelement {{.*}} 0.00 ; CHECK-NOT: call {{.*}} @llvm.x86.sse.mul ; CHECK-NOT: call {{.*}} @llvm.x86.sse.sub ; CHECK: ret %tmp = insertelement <4 x float> undef, float %f, i32 0 ; <<4 x float>> [#uses=1] %tmp10 = insertelement <4 x float> %tmp, float 0.000000e+00, i32 1 ; <<4 x float>> [#uses=1] %tmp11 = insertelement <4 x float> %tmp10, float 0.000000e+00, i32 2 ; <<4 x float>> [#uses=1] %tmp12 = insertelement <4 x float> %tmp11, float 0.000000e+00, i32 3 ; <<4 x float>> [#uses=1] %tmp28 = tail call <4 x float> @llvm.x86.sse.sub.ss( <4 x float> %tmp12, <4 x float> < float 1.000000e+00, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00 > ) ; <<4 x float>> [#uses=1] %tmp37 = tail call <4 x float> @llvm.x86.sse.mul.ss( <4 x float> %tmp28, <4 x float> < float 5.000000e-01, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00 > ) ; <<4 x float>> [#uses=1] %tmp48 = tail call <4 x float> @llvm.x86.sse.min.ss( <4 x float> %tmp37, <4 x float> < float 6.553500e+04, float 0.000000e+00, float 0.000000e+00, float 0.000000e+00 > ) ; <<4 x float>> [#uses=1] %tmp59 = tail call <4 x float> @llvm.x86.sse.max.ss( <4 x float> %tmp48, <4 x float> zeroinitializer ) ; <<4 x float>> [#uses=1] %tmp.upgrd.1 = tail call i32 @llvm.x86.sse.cvttss2si( <4 x float> %tmp59 ) ; [#uses=1] %tmp69 = trunc i32 %tmp.upgrd.1 to i16 ; [#uses=1] ret i16 %tmp69 } define i32 @test2(float %f) { ; CHECK-LABEL: @test2( ; CHECK-NOT: insertelement ; CHECK-NOT: extractelement ; CHECK: ret %tmp5 = fmul float %f, %f %tmp9 = insertelement <4 x float> undef, float %tmp5, i32 0 %tmp10 = insertelement <4 x float> %tmp9, float 0.000000e+00, i32 1 %tmp11 = insertelement <4 x float> %tmp10, float 0.000000e+00, i32 2 %tmp12 = insertelement <4 x float> %tmp11, float 0.000000e+00, i32 3 %tmp19 = bitcast <4 x float> %tmp12 to <4 x i32> %tmp21 = extractelement <4 x i32> %tmp19, i32 0 ret i32 %tmp21 } define i64 @test3(float %f, double %d) { ; CHECK-LABEL: @test3( ; CHECK-NOT: insertelement {{.*}} 0.00 ; CHECK: ret entry: %v00 = insertelement <4 x float> undef, float %f, i32 0 %v01 = insertelement <4 x float> %v00, float 0.000000e+00, i32 1 %v02 = insertelement <4 x float> %v01, float 0.000000e+00, i32 2 %v03 = insertelement <4 x float> %v02, float 0.000000e+00, i32 3 %tmp0 = tail call i32 @llvm.x86.sse.cvtss2si(<4 x float> %v03) %v10 = insertelement <4 x float> undef, float %f, i32 0 %v11 = insertelement <4 x float> %v10, float 0.000000e+00, i32 1 %v12 = insertelement <4 x float> %v11, float 0.000000e+00, i32 2 %v13 = insertelement <4 x float> %v12, float 0.000000e+00, i32 3 %tmp1 = tail call i64 @llvm.x86.sse.cvtss2si64(<4 x float> %v13) %v20 = insertelement <4 x float> undef, float %f, i32 0 %v21 = insertelement <4 x float> %v20, float 0.000000e+00, i32 1 %v22 = insertelement <4 x float> %v21, float 0.000000e+00, i32 2 %v23 = insertelement <4 x float> %v22, float 0.000000e+00, i32 3 %tmp2 = tail call i32 @llvm.x86.sse.cvttss2si(<4 x float> %v23) %v30 = insertelement <4 x float> undef, float %f, i32 0 %v31 = insertelement <4 x float> %v30, float 0.000000e+00, i32 1 %v32 = insertelement <4 x float> %v31, float 0.000000e+00, i32 2 %v33 = insertelement <4 x float> %v32, float 0.000000e+00, i32 3 %tmp3 = tail call i64 @llvm.x86.sse.cvttss2si64(<4 x float> %v33) %v40 = insertelement <2 x double> undef, double %d, i32 0 %v41 = insertelement <2 x double> %v40, double 0.000000e+00, i32 1 %tmp4 = tail call i32 @llvm.x86.sse2.cvtsd2si(<2 x double> %v41) %v50 = insertelement <2 x double> undef, double %d, i32 0 %v51 = insertelement <2 x double> %v50, double 0.000000e+00, i32 1 %tmp5 = tail call i64 @llvm.x86.sse2.cvtsd2si64(<2 x double> %v51) %v60 = insertelement <2 x double> undef, double %d, i32 0 %v61 = insertelement <2 x double> %v60, double 0.000000e+00, i32 1 %tmp6 = tail call i32 @llvm.x86.sse2.cvttsd2si(<2 x double> %v61) %v70 = insertelement <2 x double> undef, double %d, i32 0 %v71 = insertelement <2 x double> %v70, double 0.000000e+00, i32 1 %tmp7 = tail call i64 @llvm.x86.sse2.cvttsd2si64(<2 x double> %v71) %tmp8 = add i32 %tmp0, %tmp2 %tmp9 = add i32 %tmp4, %tmp6 %tmp10 = add i32 %tmp8, %tmp9 %tmp11 = sext i32 %tmp10 to i64 %tmp12 = add i64 %tmp1, %tmp3 %tmp13 = add i64 %tmp5, %tmp7 %tmp14 = add i64 %tmp12, %tmp13 %tmp15 = add i64 %tmp11, %tmp14 ret i64 %tmp15 } define void @get_image() nounwind { ; CHECK-LABEL: @get_image( ; CHECK-NOT: extractelement ; CHECK: unreachable entry: %0 = call i32 @fgetc(i8* null) nounwind ; [#uses=1] %1 = trunc i32 %0 to i8 ; [#uses=1] %tmp2 = insertelement <100 x i8> zeroinitializer, i8 %1, i32 1 ; <<100 x i8>> [#uses=1] %tmp1 = extractelement <100 x i8> %tmp2, i32 0 ; [#uses=1] %2 = icmp eq i8 %tmp1, 80 ; [#uses=1] br i1 %2, label %bb2, label %bb3 bb2: ; preds = %entry br label %bb3 bb3: ; preds = %bb2, %entry unreachable } ; PR4340 define void @vac(<4 x float>* nocapture %a) nounwind { ; CHECK-LABEL: @vac( ; CHECK-NOT: load ; CHECK: ret entry: %tmp1 = load <4 x float>, <4 x float>* %a ; <<4 x float>> [#uses=1] %vecins = insertelement <4 x float> %tmp1, float 0.000000e+00, i32 0 ; <<4 x float>> [#uses=1] %vecins4 = insertelement <4 x float> %vecins, float 0.000000e+00, i32 1; <<4 x float>> [#uses=1] %vecins6 = insertelement <4 x float> %vecins4, float 0.000000e+00, i32 2; <<4 x float>> [#uses=1] %vecins8 = insertelement <4 x float> %vecins6, float 0.000000e+00, i32 3; <<4 x float>> [#uses=1] store <4 x float> %vecins8, <4 x float>* %a ret void } declare i32 @fgetc(i8*) declare <4 x float> @llvm.x86.sse.sub.ss(<4 x float>, <4 x float>) declare <4 x float> @llvm.x86.sse.mul.ss(<4 x float>, <4 x float>) declare <4 x float> @llvm.x86.sse.min.ss(<4 x float>, <4 x float>) declare <4 x float> @llvm.x86.sse.max.ss(<4 x float>, <4 x float>) declare i32 @llvm.x86.sse.cvtss2si(<4 x float>) declare i64 @llvm.x86.sse.cvtss2si64(<4 x float>) declare i32 @llvm.x86.sse.cvttss2si(<4 x float>) declare i64 @llvm.x86.sse.cvttss2si64(<4 x float>) declare i32 @llvm.x86.sse2.cvtsd2si(<2 x double>) declare i64 @llvm.x86.sse2.cvtsd2si64(<2 x double>) declare i32 @llvm.x86.sse2.cvttsd2si(<2 x double>) declare i64 @llvm.x86.sse2.cvttsd2si64(<2 x double>) ; define <4 x i32> @kernel3_vertical(<4 x i16> * %src, <8 x i16> * %foo) nounwind { entry: %tmp = load <4 x i16>, <4 x i16>* %src %tmp1 = load <8 x i16>, <8 x i16>* %foo ; CHECK: %tmp2 = shufflevector %tmp2 = shufflevector <4 x i16> %tmp, <4 x i16> undef, <8 x i32> ; pmovzxwd ignores the upper 64-bits of its input; -instcombine should remove this shuffle: ; CHECK-NOT: shufflevector %tmp3 = shufflevector <8 x i16> %tmp1, <8 x i16> %tmp2, <8 x i32> ; CHECK-NEXT: pmovzxwd %0 = call <4 x i32> @llvm.x86.sse41.pmovzxwd(<8 x i16> %tmp3) ret <4 x i32> %0 } declare <4 x i32> @llvm.x86.sse41.pmovzxwd(<8 x i16>) nounwind readnone define <4 x float> @dead_shuffle_elt(<4 x float> %x, <2 x float> %y) nounwind { entry: ; CHECK-LABEL: define <4 x float> @dead_shuffle_elt( ; CHECK: shufflevector <2 x float> %y, <2 x float> undef, <4 x i32> %shuffle.i = shufflevector <2 x float> %y, <2 x float> %y, <4 x i32> %shuffle9.i = shufflevector <4 x float> %x, <4 x float> %shuffle.i, <4 x i32> ret <4 x float> %shuffle9.i } define <2 x float> @test_fptrunc(double %f) { ; CHECK-LABEL: @test_fptrunc( ; CHECK: insertelement ; CHECK: insertelement ; CHECK-NOT: insertelement %tmp9 = insertelement <4 x double> undef, double %f, i32 0 %tmp10 = insertelement <4 x double> %tmp9, double 0.000000e+00, i32 1 %tmp11 = insertelement <4 x double> %tmp10, double 0.000000e+00, i32 2 %tmp12 = insertelement <4 x double> %tmp11, double 0.000000e+00, i32 3 %tmp5 = fptrunc <4 x double> %tmp12 to <4 x float> %ret = shufflevector <4 x float> %tmp5, <4 x float> undef, <2 x i32> ret <2 x float> %ret } define <2 x double> @test_fpext(float %f) { ; CHECK-LABEL: @test_fpext( ; CHECK: insertelement ; CHECK: insertelement ; CHECK-NOT: insertelement %tmp9 = insertelement <4 x float> undef, float %f, i32 0 %tmp10 = insertelement <4 x float> %tmp9, float 0.000000e+00, i32 1 %tmp11 = insertelement <4 x float> %tmp10, float 0.000000e+00, i32 2 %tmp12 = insertelement <4 x float> %tmp11, float 0.000000e+00, i32 3 %tmp5 = fpext <4 x float> %tmp12 to <4 x double> %ret = shufflevector <4 x double> %tmp5, <4 x double> undef, <2 x i32> ret <2 x double> %ret } define <4 x float> @test_select(float %f, float %g) { ; CHECK-LABEL: @test_select( ; CHECK: %a0 = insertelement <4 x float> undef, float %f, i32 0 ; CHECK-NOT: insertelement ; CHECK: %a3 = insertelement <4 x float> %a0, float 3.000000e+00, i32 3 ; CHECK-NOT: insertelement ; CHECK: %ret = select <4 x i1> , <4 x float> %a3, <4 x float> %a0 = insertelement <4 x float> undef, float %f, i32 0 %a1 = insertelement <4 x float> %a0, float 1.000000e+00, i32 1 %a2 = insertelement <4 x float> %a1, float 2.000000e+00, i32 2 %a3 = insertelement <4 x float> %a2, float 3.000000e+00, i32 3 %b0 = insertelement <4 x float> undef, float %g, i32 0 %b1 = insertelement <4 x float> %b0, float 4.000000e+00, i32 1 %b2 = insertelement <4 x float> %b1, float 5.000000e+00, i32 2 %b3 = insertelement <4 x float> %b2, float 6.000000e+00, i32 3 %ret = select <4 x i1> , <4 x float> %a3, <4 x float> %b3 ret <4 x float> %ret } ; We should optimize these two redundant insertqi into one ; CHECK: define <2 x i64> @testInsertTwice(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertTwice(<2 x i64> %v, <2 x i64> %i) { ; CHECK: call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 32) ; CHECK-NOT: insertqi %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 32) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 32, i8 32) ret <2 x i64> %2 } ; The result of this insert is the second arg, since the top 64 bits of ; the result are undefined, and we copy the bottom 64 bits from the ; second arg ; CHECK: define <2 x i64> @testInsert64Bits(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsert64Bits(<2 x i64> %v, <2 x i64> %i) { ; CHECK: ret <2 x i64> %i %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 64, i8 0) ret <2 x i64> %1 } ; Test the several types of ranges and ordering that exist for two insertqi ; CHECK: define <2 x i64> @testInsertContainedRange(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertContainedRange(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 0) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 16) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertContainedRange_2(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertContainedRange_2(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 16) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 32, i8 0) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertOverlappingRange(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertOverlappingRange(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 48, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 0) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 32, i8 16) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertOverlappingRange_2(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertOverlappingRange_2(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 48, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 16) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 32, i8 0) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertAdjacentRange(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertAdjacentRange(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 48, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 32, i8 0) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 32) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertAdjacentRange_2(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertAdjacentRange_2(<2 x i64> %v, <2 x i64> %i) { ; CHECK: %[[RES:.*]] = call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 48, i8 0) ; CHECK: ret <2 x i64> %[[RES]] %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 32) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 32, i8 0) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertDisjointRange(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertDisjointRange(<2 x i64> %v, <2 x i64> %i) { ; CHECK: tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 0) ; CHECK: tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 32) %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 0) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 32) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testInsertDisjointRange_2(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testInsertDisjointRange_2(<2 x i64> %v, <2 x i64> %i) { ; CHECK: tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 0) ; CHECK: tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 32) %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 16, i8 0) %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %i, i8 16, i8 32) ret <2 x i64> %2 } ; CHECK: define <2 x i64> @testZeroLength(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testZeroLength(<2 x i64> %v, <2 x i64> %i) { ; CHECK: ret <2 x i64> %i %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 0, i8 0) ret <2 x i64> %1 } ; CHECK: define <2 x i64> @testUndefinedInsertq_1(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testUndefinedInsertq_1(<2 x i64> %v, <2 x i64> %i) { ; CHECK: ret <2 x i64> undef %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 0, i8 16) ret <2 x i64> %1 } ; CHECK: define <2 x i64> @testUndefinedInsertq_2(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testUndefinedInsertq_2(<2 x i64> %v, <2 x i64> %i) { ; CHECK: ret <2 x i64> undef %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 48, i8 32) ret <2 x i64> %1 } ; CHECK: define <2 x i64> @testUndefinedInsertq_3(<2 x i64> %v, <2 x i64> %i) define <2 x i64> @testUndefinedInsertq_3(<2 x i64> %v, <2 x i64> %i) { ; CHECK: ret <2 x i64> undef %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %v, <2 x i64> %i, i8 64, i8 16) ret <2 x i64> %1 } ; CHECK: declare <2 x i64> @llvm.x86.sse4a.insertqi declare <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64>, <2 x i64>, i8, i8) nounwind declare <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float>, <4 x i32>) define <4 x float> @test_vpermilvar_ps(<4 x float> %v) { ; CHECK-LABEL: @test_vpermilvar_ps( ; CHECK: shufflevector <4 x float> %v, <4 x float> undef, <4 x i32> %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> ) ret <4 x float> %a } declare <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float>, <8 x i32>) define <8 x float> @test_vpermilvar_ps_256(<8 x float> %v) { ; CHECK-LABEL: @test_vpermilvar_ps_256( ; CHECK: shufflevector <8 x float> %v, <8 x float> undef, <8 x i32> %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> ) ret <8 x float> %a } declare <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double>, <2 x i32>) define <2 x double> @test_vpermilvar_pd(<2 x double> %v) { ; CHECK-LABEL: @test_vpermilvar_pd( ; CHECK: shufflevector <2 x double> %v, <2 x double> undef, <2 x i32> %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i32> ) ret <2 x double> %a } declare <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double>, <4 x i32>) define <4 x double> @test_vpermilvar_pd_256(<4 x double> %v) { ; CHECK-LABEL: @test_vpermilvar_pd_256( ; CHECK: shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i32> ) ret <4 x double> %a } define <4 x float> @test_vpermilvar_ps_zero(<4 x float> %v) { ; CHECK-LABEL: @test_vpermilvar_ps_zero( ; CHECK: shufflevector <4 x float> %v, <4 x float> undef, <4 x i32> zeroinitializer %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> zeroinitializer) ret <4 x float> %a } define <8 x float> @test_vpermilvar_ps_256_zero(<8 x float> %v) { ; CHECK-LABEL: @test_vpermilvar_ps_256_zero( ; CHECK: shufflevector <8 x float> %v, <8 x float> undef, <8 x i32> %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> zeroinitializer) ret <8 x float> %a } define <2 x double> @test_vpermilvar_pd_zero(<2 x double> %v) { ; CHECK-LABEL: @test_vpermilvar_pd_zero( ; CHECK: shufflevector <2 x double> %v, <2 x double> undef, <2 x i32> zeroinitializer %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i32> zeroinitializer) ret <2 x double> %a } define <4 x double> @test_vpermilvar_pd_256_zero(<4 x double> %v) { ; CHECK-LABEL: @test_vpermilvar_pd_256_zero( ; CHECK: shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i32> zeroinitializer) ret <4 x double> %a } define <2 x i64> @test_sse2_1() nounwind readnone uwtable { %S = bitcast i32 1 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <8 x i16> @llvm.x86.sse2.psll.w(<8 x i16> , <8 x i16> %4) %6 = bitcast <8 x i16> %5 to <4 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <4 x i32> @llvm.x86.sse2.psll.d(<4 x i32> %6, <4 x i32> %7) %9 = bitcast <4 x i32> %8 to <2 x i64> %10 = tail call <2 x i64> @llvm.x86.sse2.psll.q(<2 x i64> %9, <2 x i64> %3) %11 = bitcast <2 x i64> %10 to <8 x i16> %12 = tail call <8 x i16> @llvm.x86.sse2.pslli.w(<8 x i16> %11, i32 %S) %13 = bitcast <8 x i16> %12 to <4 x i32> %14 = tail call <4 x i32> @llvm.x86.sse2.pslli.d(<4 x i32> %13, i32 %S) %15 = bitcast <4 x i32> %14 to <2 x i64> %16 = tail call <2 x i64> @llvm.x86.sse2.pslli.q(<2 x i64> %15, i32 %S) ret <2 x i64> %16 ; CHECK: test_sse2_1 ; CHECK: ret <2 x i64> } define <4 x i64> @test_avx2_1() nounwind readnone uwtable { %S = bitcast i32 1 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16> , <8 x i16> %4) %6 = bitcast <16 x i16> %5 to <8 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32> %6, <4 x i32> %7) %9 = bitcast <8 x i32> %8 to <4 x i64> %10 = tail call <4 x i64> @llvm.x86.avx2.psll.q(<4 x i64> %9, <2 x i64> %3) %11 = bitcast <4 x i64> %10 to <16 x i16> %12 = tail call <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16> %11, i32 %S) %13 = bitcast <16 x i16> %12 to <8 x i32> %14 = tail call <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32> %13, i32 %S) %15 = bitcast <8 x i32> %14 to <4 x i64> %16 = tail call <4 x i64> @llvm.x86.avx2.pslli.q(<4 x i64> %15, i32 %S) ret <4 x i64> %16 ; CHECK: test_avx2_1 ; CHECK: ret <4 x i64> } define <2 x i64> @test_sse2_0() nounwind readnone uwtable { %S = bitcast i32 128 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <8 x i16> @llvm.x86.sse2.psll.w(<8 x i16> , <8 x i16> %4) %6 = bitcast <8 x i16> %5 to <4 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <4 x i32> @llvm.x86.sse2.psll.d(<4 x i32> %6, <4 x i32> %7) %9 = bitcast <4 x i32> %8 to <2 x i64> %10 = tail call <2 x i64> @llvm.x86.sse2.psll.q(<2 x i64> %9, <2 x i64> %3) %11 = bitcast <2 x i64> %10 to <8 x i16> %12 = tail call <8 x i16> @llvm.x86.sse2.pslli.w(<8 x i16> %11, i32 %S) %13 = bitcast <8 x i16> %12 to <4 x i32> %14 = tail call <4 x i32> @llvm.x86.sse2.pslli.d(<4 x i32> %13, i32 %S) %15 = bitcast <4 x i32> %14 to <2 x i64> %16 = tail call <2 x i64> @llvm.x86.sse2.pslli.q(<2 x i64> %15, i32 %S) ret <2 x i64> %16 ; CHECK: test_sse2_0 ; CHECK: ret <2 x i64> zeroinitializer } define <4 x i64> @test_avx2_0() nounwind readnone uwtable { %S = bitcast i32 128 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16> , <8 x i16> %4) %6 = bitcast <16 x i16> %5 to <8 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32> %6, <4 x i32> %7) %9 = bitcast <8 x i32> %8 to <4 x i64> %10 = tail call <4 x i64> @llvm.x86.avx2.psll.q(<4 x i64> %9, <2 x i64> %3) %11 = bitcast <4 x i64> %10 to <16 x i16> %12 = tail call <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16> %11, i32 %S) %13 = bitcast <16 x i16> %12 to <8 x i32> %14 = tail call <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32> %13, i32 %S) %15 = bitcast <8 x i32> %14 to <4 x i64> %16 = tail call <4 x i64> @llvm.x86.avx2.pslli.q(<4 x i64> %15, i32 %S) ret <4 x i64> %16 ; CHECK: test_avx2_0 ; CHECK: ret <4 x i64> zeroinitializer } define <2 x i64> @test_sse2_psrl_1() nounwind readnone uwtable { %S = bitcast i32 1 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <8 x i16> @llvm.x86.sse2.psrl.w(<8 x i16> , <8 x i16> %4) %6 = bitcast <8 x i16> %5 to <4 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <4 x i32> @llvm.x86.sse2.psrl.d(<4 x i32> %6, <4 x i32> %7) %9 = bitcast <4 x i32> %8 to <2 x i64> %10 = tail call <2 x i64> @llvm.x86.sse2.psrl.q(<2 x i64> %9, <2 x i64> %3) %11 = bitcast <2 x i64> %10 to <8 x i16> %12 = tail call <8 x i16> @llvm.x86.sse2.psrli.w(<8 x i16> %11, i32 %S) %13 = bitcast <8 x i16> %12 to <4 x i32> %14 = tail call <4 x i32> @llvm.x86.sse2.psrli.d(<4 x i32> %13, i32 %S) %15 = bitcast <4 x i32> %14 to <2 x i64> %16 = tail call <2 x i64> @llvm.x86.sse2.psrli.q(<2 x i64> %15, i32 %S) ret <2 x i64> %16 ; CHECK: test_sse2_psrl_1 ; CHECK: ret <2 x i64> } define <4 x i64> @test_avx2_psrl_1() nounwind readnone uwtable { %S = bitcast i32 1 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16> , <8 x i16> %4) %6 = bitcast <16 x i16> %5 to <8 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32> %6, <4 x i32> %7) %9 = bitcast <8 x i32> %8 to <4 x i64> %10 = tail call <4 x i64> @llvm.x86.avx2.psrl.q(<4 x i64> %9, <2 x i64> %3) %11 = bitcast <4 x i64> %10 to <16 x i16> %12 = tail call <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16> %11, i32 %S) %13 = bitcast <16 x i16> %12 to <8 x i32> %14 = tail call <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32> %13, i32 %S) %15 = bitcast <8 x i32> %14 to <4 x i64> %16 = tail call <4 x i64> @llvm.x86.avx2.psrli.q(<4 x i64> %15, i32 %S) ret <4 x i64> %16 ; CHECK: test_avx2_psrl_1 ; CHECK: ret <4 x i64> } define <2 x i64> @test_sse2_psrl_0() nounwind readnone uwtable { %S = bitcast i32 128 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <8 x i16> @llvm.x86.sse2.psrl.w(<8 x i16> , <8 x i16> %4) %6 = bitcast <8 x i16> %5 to <4 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <4 x i32> @llvm.x86.sse2.psrl.d(<4 x i32> %6, <4 x i32> %7) %9 = bitcast <4 x i32> %8 to <2 x i64> %10 = tail call <2 x i64> @llvm.x86.sse2.psrl.q(<2 x i64> %9, <2 x i64> %3) %11 = bitcast <2 x i64> %10 to <8 x i16> %12 = tail call <8 x i16> @llvm.x86.sse2.psrli.w(<8 x i16> %11, i32 %S) %13 = bitcast <8 x i16> %12 to <4 x i32> %14 = tail call <4 x i32> @llvm.x86.sse2.psrli.d(<4 x i32> %13, i32 %S) %15 = bitcast <4 x i32> %14 to <2 x i64> %16 = tail call <2 x i64> @llvm.x86.sse2.psrli.q(<2 x i64> %15, i32 %S) ret <2 x i64> %16 ; CHECK: test_sse2_psrl_0 ; CHECK: ret <2 x i64> zeroinitializer } define <4 x i64> @test_avx2_psrl_0() nounwind readnone uwtable { %S = bitcast i32 128 to i32 %1 = zext i32 %S to i64 %2 = insertelement <2 x i64> undef, i64 %1, i32 0 %3 = insertelement <2 x i64> %2, i64 0, i32 1 %4 = bitcast <2 x i64> %3 to <8 x i16> %5 = tail call <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16> , <8 x i16> %4) %6 = bitcast <16 x i16> %5 to <8 x i32> %7 = bitcast <2 x i64> %3 to <4 x i32> %8 = tail call <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32> %6, <4 x i32> %7) %9 = bitcast <8 x i32> %8 to <4 x i64> %10 = tail call <4 x i64> @llvm.x86.avx2.psrl.q(<4 x i64> %9, <2 x i64> %3) %11 = bitcast <4 x i64> %10 to <16 x i16> %12 = tail call <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16> %11, i32 %S) %13 = bitcast <16 x i16> %12 to <8 x i32> %14 = tail call <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32> %13, i32 %S) %15 = bitcast <8 x i32> %14 to <4 x i64> %16 = tail call <4 x i64> @llvm.x86.avx2.psrli.q(<4 x i64> %15, i32 %S) ret <4 x i64> %16 ; CHECK: test_avx2_psrl_0 ; CHECK: ret <4 x i64> zeroinitializer } declare <4 x i64> @llvm.x86.avx2.pslli.q(<4 x i64>, i32) #1 declare <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32>, i32) #1 declare <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16>, i32) #1 declare <4 x i64> @llvm.x86.avx2.psll.q(<4 x i64>, <2 x i64>) #1 declare <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32>, <4 x i32>) #1 declare <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16>, <8 x i16>) #1 declare <2 x i64> @llvm.x86.sse2.pslli.q(<2 x i64>, i32) #1 declare <4 x i32> @llvm.x86.sse2.pslli.d(<4 x i32>, i32) #1 declare <8 x i16> @llvm.x86.sse2.pslli.w(<8 x i16>, i32) #1 declare <2 x i64> @llvm.x86.sse2.psll.q(<2 x i64>, <2 x i64>) #1 declare <4 x i32> @llvm.x86.sse2.psll.d(<4 x i32>, <4 x i32>) #1 declare <8 x i16> @llvm.x86.sse2.psll.w(<8 x i16>, <8 x i16>) #1 declare <4 x i64> @llvm.x86.avx2.psrli.q(<4 x i64>, i32) #1 declare <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32>, i32) #1 declare <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16>, i32) #1 declare <4 x i64> @llvm.x86.avx2.psrl.q(<4 x i64>, <2 x i64>) #1 declare <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32>, <4 x i32>) #1 declare <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16>, <8 x i16>) #1 declare <2 x i64> @llvm.x86.sse2.psrli.q(<2 x i64>, i32) #1 declare <4 x i32> @llvm.x86.sse2.psrli.d(<4 x i32>, i32) #1 declare <8 x i16> @llvm.x86.sse2.psrli.w(<8 x i16>, i32) #1 declare <2 x i64> @llvm.x86.sse2.psrl.q(<2 x i64>, <2 x i64>) #1 declare <4 x i32> @llvm.x86.sse2.psrl.d(<4 x i32>, <4 x i32>) #1 declare <8 x i16> @llvm.x86.sse2.psrl.w(<8 x i16>, <8 x i16>) #1 attributes #1 = { nounwind readnone }