Loading llvm/test/Transforms/SLPVectorizer/X86/pr47642.ll 0 → 100644 +44 −0 Original line number Diff line number Diff line ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt -slp-vectorizer -instcombine -S < %s | FileCheck %s ; These code should be fully vectorized by D57059 patch target triple = "x86_64-unknown-linux-gnu" define <4 x i32> @foo(<4 x i32> %x, i32 %f) { ; CHECK-LABEL: @foo( ; CHECK-NEXT: [[VECINIT:%.*]] = insertelement <4 x i32> undef, i32 [[F:%.*]], i32 0 ; CHECK-NEXT: [[ADD:%.*]] = add nsw i32 [[F]], 1 ; CHECK-NEXT: [[VECINIT1:%.*]] = insertelement <4 x i32> [[VECINIT]], i32 [[ADD]], i32 1 ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <2 x i32> poison, i32 [[F]], i32 0 ; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <2 x i32> [[TMP1]], <2 x i32> poison, <2 x i32> zeroinitializer ; CHECK-NEXT: [[TMP3:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 2, i32 3> ; CHECK-NEXT: [[TMP4:%.*]] = shufflevector <2 x i32> [[TMP3]], <2 x i32> poison, <4 x i32> <i32 0, i32 1, i32 undef, i32 undef> ; CHECK-NEXT: [[VECINIT51:%.*]] = shufflevector <4 x i32> [[VECINIT1]], <4 x i32> [[TMP4]], <4 x i32> <i32 0, i32 1, i32 4, i32 5> ; CHECK-NEXT: ret <4 x i32> [[VECINIT51]] ; %vecinit = insertelement <4 x i32> undef, i32 %f, i32 0 %add = add nsw i32 %f, 1 %vecinit1 = insertelement <4 x i32> %vecinit, i32 %add, i32 1 %add2 = add nsw i32 %f, 2 %vecinit3 = insertelement <4 x i32> %vecinit1, i32 %add2, i32 2 %add4 = add nsw i32 %f, 3 %vecinit5 = insertelement <4 x i32> %vecinit3, i32 %add4, i32 3 ret <4 x i32> %vecinit5 } define <4 x i32> @bar(<4 x i32> %x, i32 %f) { ; CHECK-LABEL: @bar( ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <2 x i32> poison, i32 [[F:%.*]], i32 0 ; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <2 x i32> [[TMP1]], <2 x i32> poison, <2 x i32> zeroinitializer ; CHECK-NEXT: [[TMP3:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 2, i32 3> ; CHECK-NEXT: [[SHUFFLE:%.*]] = shufflevector <2 x i32> [[TMP3]], <2 x i32> poison, <4 x i32> <i32 0, i32 0, i32 0, i32 1> ; CHECK-NEXT: ret <4 x i32> [[SHUFFLE]] ; %add = add nsw i32 %f, 2 %vecinit = insertelement <4 x i32> undef, i32 %add, i32 0 %vecinit2 = insertelement <4 x i32> %vecinit, i32 %add, i32 1 %vecinit4 = insertelement <4 x i32> %vecinit2, i32 %add, i32 2 %add5 = add nsw i32 %f, 3 %vecinit6 = insertelement <4 x i32> %vecinit4, i32 %add5, i32 3 ret <4 x i32> %vecinit6 } llvm/test/Transforms/SLPVectorizer/X86/pr49933.ll 0 → 100644 +122 −0 Original line number Diff line number Diff line ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt < %s -slp-vectorizer -S -mtriple=x86_64-- -mcpu=skylake-avx512 | FileCheck %s ; These code should be fully vectorized by D57059 patch define void @foo(i8* noalias nocapture %t0, i8* noalias nocapture readonly %t1) { ; CHECK-LABEL: @foo( ; CHECK-NEXT: [[T3:%.*]] = load i8, i8* [[T1:%.*]], align 1, !tbaa [[TBAA0:![0-9]+]] ; CHECK-NEXT: [[T4:%.*]] = icmp ult i8 [[T3]], 64 ; CHECK-NEXT: [[T5:%.*]] = sub i8 0, [[T3]] ; CHECK-NEXT: [[T6:%.*]] = select i1 [[T4]], i8 [[T3]], i8 [[T5]] ; CHECK-NEXT: store i8 [[T6]], i8* [[T0:%.*]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T7:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 1 ; CHECK-NEXT: [[T8:%.*]] = load i8, i8* [[T7]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T9:%.*]] = icmp ult i8 [[T8]], 64 ; CHECK-NEXT: [[T10:%.*]] = sub i8 0, [[T8]] ; CHECK-NEXT: [[T11:%.*]] = select i1 [[T9]], i8 [[T8]], i8 [[T10]] ; CHECK-NEXT: [[T12:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 1 ; CHECK-NEXT: store i8 [[T11]], i8* [[T12]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T13:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 2 ; CHECK-NEXT: [[T14:%.*]] = load i8, i8* [[T13]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T15:%.*]] = icmp ult i8 [[T14]], 64 ; CHECK-NEXT: [[T16:%.*]] = sub i8 0, [[T14]] ; CHECK-NEXT: [[T17:%.*]] = select i1 [[T15]], i8 [[T14]], i8 [[T16]] ; CHECK-NEXT: [[T18:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 2 ; CHECK-NEXT: store i8 [[T17]], i8* [[T18]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T19:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 3 ; CHECK-NEXT: [[T20:%.*]] = load i8, i8* [[T19]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T21:%.*]] = icmp ult i8 [[T20]], 64 ; CHECK-NEXT: [[T22:%.*]] = sub i8 0, [[T20]] ; CHECK-NEXT: [[T23:%.*]] = select i1 [[T21]], i8 [[T20]], i8 [[T22]] ; CHECK-NEXT: [[T24:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 3 ; CHECK-NEXT: store i8 [[T23]], i8* [[T24]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T25:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 4 ; CHECK-NEXT: [[T26:%.*]] = load i8, i8* [[T25]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T27:%.*]] = icmp ult i8 [[T26]], 64 ; CHECK-NEXT: [[T28:%.*]] = sub i8 0, [[T26]] ; CHECK-NEXT: [[T29:%.*]] = select i1 [[T27]], i8 [[T26]], i8 [[T28]] ; CHECK-NEXT: [[T30:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 4 ; CHECK-NEXT: store i8 [[T29]], i8* [[T30]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T31:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 5 ; CHECK-NEXT: [[T32:%.*]] = load i8, i8* [[T31]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T33:%.*]] = icmp ult i8 [[T32]], 64 ; CHECK-NEXT: [[T34:%.*]] = sub i8 0, [[T32]] ; CHECK-NEXT: [[T35:%.*]] = select i1 [[T33]], i8 [[T32]], i8 [[T34]] ; CHECK-NEXT: [[T36:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 5 ; CHECK-NEXT: store i8 [[T35]], i8* [[T36]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T37:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 6 ; CHECK-NEXT: [[T38:%.*]] = load i8, i8* [[T37]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T39:%.*]] = icmp ult i8 [[T38]], 64 ; CHECK-NEXT: [[T40:%.*]] = sub i8 0, [[T38]] ; CHECK-NEXT: [[T41:%.*]] = select i1 [[T39]], i8 [[T38]], i8 [[T40]] ; CHECK-NEXT: [[T42:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 6 ; CHECK-NEXT: store i8 [[T41]], i8* [[T42]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T43:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 7 ; CHECK-NEXT: [[T44:%.*]] = load i8, i8* [[T43]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T45:%.*]] = icmp ult i8 [[T44]], 64 ; CHECK-NEXT: [[T46:%.*]] = sub i8 0, [[T44]] ; CHECK-NEXT: [[T47:%.*]] = select i1 [[T45]], i8 [[T44]], i8 [[T46]] ; CHECK-NEXT: [[T48:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 7 ; CHECK-NEXT: store i8 [[T47]], i8* [[T48]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: ret void ; %t3 = load i8, i8* %t1, align 1, !tbaa !3 %t4 = icmp ult i8 %t3, 64 %t5 = sub i8 0, %t3 %t6 = select i1 %t4, i8 %t3, i8 %t5 store i8 %t6, i8* %t0, align 1, !tbaa !3 %t7 = getelementptr inbounds i8, i8* %t1, i64 1 %t8 = load i8, i8* %t7, align 1, !tbaa !3 %t9 = icmp ult i8 %t8, 64 %t10 = sub i8 0, %t8 %t11 = select i1 %t9, i8 %t8, i8 %t10 %t12 = getelementptr inbounds i8, i8* %t0, i64 1 store i8 %t11, i8* %t12, align 1, !tbaa !3 %t13 = getelementptr inbounds i8, i8* %t1, i64 2 %t14 = load i8, i8* %t13, align 1, !tbaa !3 %t15 = icmp ult i8 %t14, 64 %t16 = sub i8 0, %t14 %t17 = select i1 %t15, i8 %t14, i8 %t16 %t18 = getelementptr inbounds i8, i8* %t0, i64 2 store i8 %t17, i8* %t18, align 1, !tbaa !3 %t19 = getelementptr inbounds i8, i8* %t1, i64 3 %t20 = load i8, i8* %t19, align 1, !tbaa !3 %t21 = icmp ult i8 %t20, 64 %t22 = sub i8 0, %t20 %t23 = select i1 %t21, i8 %t20, i8 %t22 %t24 = getelementptr inbounds i8, i8* %t0, i64 3 store i8 %t23, i8* %t24, align 1, !tbaa !3 %t25 = getelementptr inbounds i8, i8* %t1, i64 4 %t26 = load i8, i8* %t25, align 1, !tbaa !3 %t27 = icmp ult i8 %t26, 64 %t28 = sub i8 0, %t26 %t29 = select i1 %t27, i8 %t26, i8 %t28 %t30 = getelementptr inbounds i8, i8* %t0, i64 4 store i8 %t29, i8* %t30, align 1, !tbaa !3 %t31 = getelementptr inbounds i8, i8* %t1, i64 5 %t32 = load i8, i8* %t31, align 1, !tbaa !3 %t33 = icmp ult i8 %t32, 64 %t34 = sub i8 0, %t32 %t35 = select i1 %t33, i8 %t32, i8 %t34 %t36 = getelementptr inbounds i8, i8* %t0, i64 5 store i8 %t35, i8* %t36, align 1, !tbaa !3 %t37 = getelementptr inbounds i8, i8* %t1, i64 6 %t38 = load i8, i8* %t37, align 1, !tbaa !3 %t39 = icmp ult i8 %t38, 64 %t40 = sub i8 0, %t38 %t41 = select i1 %t39, i8 %t38, i8 %t40 %t42 = getelementptr inbounds i8, i8* %t0, i64 6 store i8 %t41, i8* %t42, align 1, !tbaa !3 %t43 = getelementptr inbounds i8, i8* %t1, i64 7 %t44 = load i8, i8* %t43, align 1, !tbaa !3 %t45 = icmp ult i8 %t44, 64 %t46 = sub i8 0, %t44 %t47 = select i1 %t45, i8 %t44, i8 %t46 %t48 = getelementptr inbounds i8, i8* %t0, i64 7 store i8 %t47, i8* %t48, align 1, !tbaa !3 ret void } !3 = !{!4, !4, i64 0} !4 = !{!"omnipotent char", !5, i64 0} !5 = !{!"Simple C++ TBAA"} Loading
llvm/test/Transforms/SLPVectorizer/X86/pr47642.ll 0 → 100644 +44 −0 Original line number Diff line number Diff line ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt -slp-vectorizer -instcombine -S < %s | FileCheck %s ; These code should be fully vectorized by D57059 patch target triple = "x86_64-unknown-linux-gnu" define <4 x i32> @foo(<4 x i32> %x, i32 %f) { ; CHECK-LABEL: @foo( ; CHECK-NEXT: [[VECINIT:%.*]] = insertelement <4 x i32> undef, i32 [[F:%.*]], i32 0 ; CHECK-NEXT: [[ADD:%.*]] = add nsw i32 [[F]], 1 ; CHECK-NEXT: [[VECINIT1:%.*]] = insertelement <4 x i32> [[VECINIT]], i32 [[ADD]], i32 1 ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <2 x i32> poison, i32 [[F]], i32 0 ; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <2 x i32> [[TMP1]], <2 x i32> poison, <2 x i32> zeroinitializer ; CHECK-NEXT: [[TMP3:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 2, i32 3> ; CHECK-NEXT: [[TMP4:%.*]] = shufflevector <2 x i32> [[TMP3]], <2 x i32> poison, <4 x i32> <i32 0, i32 1, i32 undef, i32 undef> ; CHECK-NEXT: [[VECINIT51:%.*]] = shufflevector <4 x i32> [[VECINIT1]], <4 x i32> [[TMP4]], <4 x i32> <i32 0, i32 1, i32 4, i32 5> ; CHECK-NEXT: ret <4 x i32> [[VECINIT51]] ; %vecinit = insertelement <4 x i32> undef, i32 %f, i32 0 %add = add nsw i32 %f, 1 %vecinit1 = insertelement <4 x i32> %vecinit, i32 %add, i32 1 %add2 = add nsw i32 %f, 2 %vecinit3 = insertelement <4 x i32> %vecinit1, i32 %add2, i32 2 %add4 = add nsw i32 %f, 3 %vecinit5 = insertelement <4 x i32> %vecinit3, i32 %add4, i32 3 ret <4 x i32> %vecinit5 } define <4 x i32> @bar(<4 x i32> %x, i32 %f) { ; CHECK-LABEL: @bar( ; CHECK-NEXT: [[TMP1:%.*]] = insertelement <2 x i32> poison, i32 [[F:%.*]], i32 0 ; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <2 x i32> [[TMP1]], <2 x i32> poison, <2 x i32> zeroinitializer ; CHECK-NEXT: [[TMP3:%.*]] = add nsw <2 x i32> [[TMP2]], <i32 2, i32 3> ; CHECK-NEXT: [[SHUFFLE:%.*]] = shufflevector <2 x i32> [[TMP3]], <2 x i32> poison, <4 x i32> <i32 0, i32 0, i32 0, i32 1> ; CHECK-NEXT: ret <4 x i32> [[SHUFFLE]] ; %add = add nsw i32 %f, 2 %vecinit = insertelement <4 x i32> undef, i32 %add, i32 0 %vecinit2 = insertelement <4 x i32> %vecinit, i32 %add, i32 1 %vecinit4 = insertelement <4 x i32> %vecinit2, i32 %add, i32 2 %add5 = add nsw i32 %f, 3 %vecinit6 = insertelement <4 x i32> %vecinit4, i32 %add5, i32 3 ret <4 x i32> %vecinit6 }
llvm/test/Transforms/SLPVectorizer/X86/pr49933.ll 0 → 100644 +122 −0 Original line number Diff line number Diff line ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt < %s -slp-vectorizer -S -mtriple=x86_64-- -mcpu=skylake-avx512 | FileCheck %s ; These code should be fully vectorized by D57059 patch define void @foo(i8* noalias nocapture %t0, i8* noalias nocapture readonly %t1) { ; CHECK-LABEL: @foo( ; CHECK-NEXT: [[T3:%.*]] = load i8, i8* [[T1:%.*]], align 1, !tbaa [[TBAA0:![0-9]+]] ; CHECK-NEXT: [[T4:%.*]] = icmp ult i8 [[T3]], 64 ; CHECK-NEXT: [[T5:%.*]] = sub i8 0, [[T3]] ; CHECK-NEXT: [[T6:%.*]] = select i1 [[T4]], i8 [[T3]], i8 [[T5]] ; CHECK-NEXT: store i8 [[T6]], i8* [[T0:%.*]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T7:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 1 ; CHECK-NEXT: [[T8:%.*]] = load i8, i8* [[T7]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T9:%.*]] = icmp ult i8 [[T8]], 64 ; CHECK-NEXT: [[T10:%.*]] = sub i8 0, [[T8]] ; CHECK-NEXT: [[T11:%.*]] = select i1 [[T9]], i8 [[T8]], i8 [[T10]] ; CHECK-NEXT: [[T12:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 1 ; CHECK-NEXT: store i8 [[T11]], i8* [[T12]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T13:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 2 ; CHECK-NEXT: [[T14:%.*]] = load i8, i8* [[T13]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T15:%.*]] = icmp ult i8 [[T14]], 64 ; CHECK-NEXT: [[T16:%.*]] = sub i8 0, [[T14]] ; CHECK-NEXT: [[T17:%.*]] = select i1 [[T15]], i8 [[T14]], i8 [[T16]] ; CHECK-NEXT: [[T18:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 2 ; CHECK-NEXT: store i8 [[T17]], i8* [[T18]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T19:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 3 ; CHECK-NEXT: [[T20:%.*]] = load i8, i8* [[T19]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T21:%.*]] = icmp ult i8 [[T20]], 64 ; CHECK-NEXT: [[T22:%.*]] = sub i8 0, [[T20]] ; CHECK-NEXT: [[T23:%.*]] = select i1 [[T21]], i8 [[T20]], i8 [[T22]] ; CHECK-NEXT: [[T24:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 3 ; CHECK-NEXT: store i8 [[T23]], i8* [[T24]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T25:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 4 ; CHECK-NEXT: [[T26:%.*]] = load i8, i8* [[T25]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T27:%.*]] = icmp ult i8 [[T26]], 64 ; CHECK-NEXT: [[T28:%.*]] = sub i8 0, [[T26]] ; CHECK-NEXT: [[T29:%.*]] = select i1 [[T27]], i8 [[T26]], i8 [[T28]] ; CHECK-NEXT: [[T30:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 4 ; CHECK-NEXT: store i8 [[T29]], i8* [[T30]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T31:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 5 ; CHECK-NEXT: [[T32:%.*]] = load i8, i8* [[T31]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T33:%.*]] = icmp ult i8 [[T32]], 64 ; CHECK-NEXT: [[T34:%.*]] = sub i8 0, [[T32]] ; CHECK-NEXT: [[T35:%.*]] = select i1 [[T33]], i8 [[T32]], i8 [[T34]] ; CHECK-NEXT: [[T36:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 5 ; CHECK-NEXT: store i8 [[T35]], i8* [[T36]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T37:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 6 ; CHECK-NEXT: [[T38:%.*]] = load i8, i8* [[T37]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T39:%.*]] = icmp ult i8 [[T38]], 64 ; CHECK-NEXT: [[T40:%.*]] = sub i8 0, [[T38]] ; CHECK-NEXT: [[T41:%.*]] = select i1 [[T39]], i8 [[T38]], i8 [[T40]] ; CHECK-NEXT: [[T42:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 6 ; CHECK-NEXT: store i8 [[T41]], i8* [[T42]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T43:%.*]] = getelementptr inbounds i8, i8* [[T1]], i64 7 ; CHECK-NEXT: [[T44:%.*]] = load i8, i8* [[T43]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: [[T45:%.*]] = icmp ult i8 [[T44]], 64 ; CHECK-NEXT: [[T46:%.*]] = sub i8 0, [[T44]] ; CHECK-NEXT: [[T47:%.*]] = select i1 [[T45]], i8 [[T44]], i8 [[T46]] ; CHECK-NEXT: [[T48:%.*]] = getelementptr inbounds i8, i8* [[T0]], i64 7 ; CHECK-NEXT: store i8 [[T47]], i8* [[T48]], align 1, !tbaa [[TBAA0]] ; CHECK-NEXT: ret void ; %t3 = load i8, i8* %t1, align 1, !tbaa !3 %t4 = icmp ult i8 %t3, 64 %t5 = sub i8 0, %t3 %t6 = select i1 %t4, i8 %t3, i8 %t5 store i8 %t6, i8* %t0, align 1, !tbaa !3 %t7 = getelementptr inbounds i8, i8* %t1, i64 1 %t8 = load i8, i8* %t7, align 1, !tbaa !3 %t9 = icmp ult i8 %t8, 64 %t10 = sub i8 0, %t8 %t11 = select i1 %t9, i8 %t8, i8 %t10 %t12 = getelementptr inbounds i8, i8* %t0, i64 1 store i8 %t11, i8* %t12, align 1, !tbaa !3 %t13 = getelementptr inbounds i8, i8* %t1, i64 2 %t14 = load i8, i8* %t13, align 1, !tbaa !3 %t15 = icmp ult i8 %t14, 64 %t16 = sub i8 0, %t14 %t17 = select i1 %t15, i8 %t14, i8 %t16 %t18 = getelementptr inbounds i8, i8* %t0, i64 2 store i8 %t17, i8* %t18, align 1, !tbaa !3 %t19 = getelementptr inbounds i8, i8* %t1, i64 3 %t20 = load i8, i8* %t19, align 1, !tbaa !3 %t21 = icmp ult i8 %t20, 64 %t22 = sub i8 0, %t20 %t23 = select i1 %t21, i8 %t20, i8 %t22 %t24 = getelementptr inbounds i8, i8* %t0, i64 3 store i8 %t23, i8* %t24, align 1, !tbaa !3 %t25 = getelementptr inbounds i8, i8* %t1, i64 4 %t26 = load i8, i8* %t25, align 1, !tbaa !3 %t27 = icmp ult i8 %t26, 64 %t28 = sub i8 0, %t26 %t29 = select i1 %t27, i8 %t26, i8 %t28 %t30 = getelementptr inbounds i8, i8* %t0, i64 4 store i8 %t29, i8* %t30, align 1, !tbaa !3 %t31 = getelementptr inbounds i8, i8* %t1, i64 5 %t32 = load i8, i8* %t31, align 1, !tbaa !3 %t33 = icmp ult i8 %t32, 64 %t34 = sub i8 0, %t32 %t35 = select i1 %t33, i8 %t32, i8 %t34 %t36 = getelementptr inbounds i8, i8* %t0, i64 5 store i8 %t35, i8* %t36, align 1, !tbaa !3 %t37 = getelementptr inbounds i8, i8* %t1, i64 6 %t38 = load i8, i8* %t37, align 1, !tbaa !3 %t39 = icmp ult i8 %t38, 64 %t40 = sub i8 0, %t38 %t41 = select i1 %t39, i8 %t38, i8 %t40 %t42 = getelementptr inbounds i8, i8* %t0, i64 6 store i8 %t41, i8* %t42, align 1, !tbaa !3 %t43 = getelementptr inbounds i8, i8* %t1, i64 7 %t44 = load i8, i8* %t43, align 1, !tbaa !3 %t45 = icmp ult i8 %t44, 64 %t46 = sub i8 0, %t44 %t47 = select i1 %t45, i8 %t44, i8 %t46 %t48 = getelementptr inbounds i8, i8* %t0, i64 7 store i8 %t47, i8* %t48, align 1, !tbaa !3 ret void } !3 = !{!4, !4, i64 0} !4 = !{!"omnipotent char", !5, i64 0} !5 = !{!"Simple C++ TBAA"}