1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119
|
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
; RUN: opt < %s -passes=slp-vectorizer -S -mtriple=aarch64-unknown-unknown -mcpu=cortex-a53 | FileCheck %s
target datalayout = "e-m:e-i8:8:32-i16:16:32-i64:64-i128:128-n32:64-S128"
; This test is reduced from the matrix multiplication benchmark in the test-suite:
; https://github.com/llvm/llvm-test-suite/tree/main/SingleSource/Benchmarks/Misc/matmul_f64_4x4.c
; The operations here are expected to be vectorized to <2 x double>.
; Otherwise, performance will suffer on Cortex-A53.
define void @wrap_mul4(ptr nocapture %Out, ptr nocapture readonly %A, ptr nocapture readonly %B) {
; CHECK-LABEL: @wrap_mul4(
; CHECK-NEXT: [[TEMP:%.*]] = load double, ptr [[A:%.*]], align 8
; CHECK-NEXT: [[ARRAYIDX5_I:%.*]] = getelementptr inbounds [2 x double], ptr [[A]], i64 0, i64 1
; CHECK-NEXT: [[TEMP2:%.*]] = load double, ptr [[ARRAYIDX5_I]], align 8
; CHECK-NEXT: [[ARRAYIDX7_I:%.*]] = getelementptr inbounds [4 x double], ptr [[B:%.*]], i64 1, i64 0
; CHECK-NEXT: [[ARRAYIDX25_I:%.*]] = getelementptr inbounds [4 x double], ptr [[B]], i64 0, i64 2
; CHECK-NEXT: [[ARRAYIDX30_I:%.*]] = getelementptr inbounds [4 x double], ptr [[B]], i64 1, i64 2
; CHECK-NEXT: [[ARRAYIDX47_I:%.*]] = getelementptr inbounds [2 x double], ptr [[A]], i64 1, i64 0
; CHECK-NEXT: [[TEMP10:%.*]] = load double, ptr [[ARRAYIDX47_I]], align 8
; CHECK-NEXT: [[ARRAYIDX52_I:%.*]] = getelementptr inbounds [2 x double], ptr [[A]], i64 1, i64 1
; CHECK-NEXT: [[TEMP11:%.*]] = load double, ptr [[ARRAYIDX52_I]], align 8
; CHECK-NEXT: [[TMP2:%.*]] = load <2 x double>, ptr [[B]], align 8
; CHECK-NEXT: [[TMP3:%.*]] = insertelement <2 x double> poison, double [[TEMP]], i32 0
; CHECK-NEXT: [[SHUFFLE:%.*]] = shufflevector <2 x double> [[TMP3]], <2 x double> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: [[TMP4:%.*]] = fmul <2 x double> [[SHUFFLE]], [[TMP2]]
; CHECK-NEXT: [[TMP6:%.*]] = load <2 x double>, ptr [[ARRAYIDX7_I]], align 8
; CHECK-NEXT: [[TMP7:%.*]] = insertelement <2 x double> poison, double [[TEMP2]], i32 0
; CHECK-NEXT: [[SHUFFLE1:%.*]] = shufflevector <2 x double> [[TMP7]], <2 x double> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: [[TMP8:%.*]] = fmul <2 x double> [[SHUFFLE1]], [[TMP6]]
; CHECK-NEXT: [[TMP9:%.*]] = fadd <2 x double> [[TMP4]], [[TMP8]]
; CHECK-NEXT: [[RES_I_SROA_5_0_OUT2_I_SROA_IDX4:%.*]] = getelementptr inbounds double, ptr [[OUT:%.*]], i64 2
; CHECK-NEXT: [[TMP12:%.*]] = load <2 x double>, ptr [[ARRAYIDX25_I]], align 8
; CHECK-NEXT: [[TMP13:%.*]] = fmul <2 x double> [[SHUFFLE]], [[TMP12]]
; CHECK-NEXT: [[TMP15:%.*]] = load <2 x double>, ptr [[ARRAYIDX30_I]], align 8
; CHECK-NEXT: [[TMP16:%.*]] = fmul <2 x double> [[SHUFFLE1]], [[TMP15]]
; CHECK-NEXT: [[TMP17:%.*]] = fadd <2 x double> [[TMP13]], [[TMP16]]
; CHECK-NEXT: store <2 x double> [[TMP9]], ptr [[OUT]], align 8
; CHECK-NEXT: store <2 x double> [[TMP17]], ptr [[RES_I_SROA_5_0_OUT2_I_SROA_IDX4]], align 8
; CHECK-NEXT: [[RES_I_SROA_7_0_OUT2_I_SROA_IDX8:%.*]] = getelementptr inbounds double, ptr [[OUT]], i64 4
; CHECK-NEXT: [[TMP19:%.*]] = insertelement <2 x double> poison, double [[TEMP10]], i32 0
; CHECK-NEXT: [[SHUFFLE4:%.*]] = shufflevector <2 x double> [[TMP19]], <2 x double> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: [[TMP20:%.*]] = fmul <2 x double> [[TMP2]], [[SHUFFLE4]]
; CHECK-NEXT: [[TMP21:%.*]] = insertelement <2 x double> poison, double [[TEMP11]], i32 0
; CHECK-NEXT: [[SHUFFLE5:%.*]] = shufflevector <2 x double> [[TMP21]], <2 x double> poison, <2 x i32> zeroinitializer
; CHECK-NEXT: [[TMP22:%.*]] = fmul <2 x double> [[TMP6]], [[SHUFFLE5]]
; CHECK-NEXT: [[TMP23:%.*]] = fadd <2 x double> [[TMP20]], [[TMP22]]
; CHECK-NEXT: store <2 x double> [[TMP23]], ptr [[RES_I_SROA_7_0_OUT2_I_SROA_IDX8]], align 8
; CHECK-NEXT: [[RES_I_SROA_9_0_OUT2_I_SROA_IDX12:%.*]] = getelementptr inbounds double, ptr [[OUT]], i64 6
; CHECK-NEXT: [[TMP25:%.*]] = fmul <2 x double> [[TMP12]], [[SHUFFLE4]]
; CHECK-NEXT: [[TMP26:%.*]] = fmul <2 x double> [[TMP15]], [[SHUFFLE5]]
; CHECK-NEXT: [[TMP27:%.*]] = fadd <2 x double> [[TMP25]], [[TMP26]]
; CHECK-NEXT: store <2 x double> [[TMP27]], ptr [[RES_I_SROA_9_0_OUT2_I_SROA_IDX12]], align 8
; CHECK-NEXT: ret void
;
%temp = load double, ptr %A, align 8
%temp1 = load double, ptr %B, align 8
%mul.i = fmul double %temp, %temp1
%arrayidx5.i = getelementptr inbounds [2 x double], ptr %A, i64 0, i64 1
%temp2 = load double, ptr %arrayidx5.i, align 8
%arrayidx7.i = getelementptr inbounds [4 x double], ptr %B, i64 1, i64 0
%temp3 = load double, ptr %arrayidx7.i, align 8
%mul8.i = fmul double %temp2, %temp3
%add.i = fadd double %mul.i, %mul8.i
%arrayidx13.i = getelementptr inbounds [4 x double], ptr %B, i64 0, i64 1
%temp4 = load double, ptr %arrayidx13.i, align 8
%mul14.i = fmul double %temp, %temp4
%arrayidx18.i = getelementptr inbounds [4 x double], ptr %B, i64 1, i64 1
%temp5 = load double, ptr %arrayidx18.i, align 8
%mul19.i = fmul double %temp2, %temp5
%add20.i = fadd double %mul14.i, %mul19.i
%arrayidx25.i = getelementptr inbounds [4 x double], ptr %B, i64 0, i64 2
%temp6 = load double, ptr %arrayidx25.i, align 8
%mul26.i = fmul double %temp, %temp6
%arrayidx30.i = getelementptr inbounds [4 x double], ptr %B, i64 1, i64 2
%temp7 = load double, ptr %arrayidx30.i, align 8
%mul31.i = fmul double %temp2, %temp7
%add32.i = fadd double %mul26.i, %mul31.i
%arrayidx37.i = getelementptr inbounds [4 x double], ptr %B, i64 0, i64 3
%temp8 = load double, ptr %arrayidx37.i, align 8
%mul38.i = fmul double %temp, %temp8
%arrayidx42.i = getelementptr inbounds [4 x double], ptr %B, i64 1, i64 3
%temp9 = load double, ptr %arrayidx42.i, align 8
%mul43.i = fmul double %temp2, %temp9
%add44.i = fadd double %mul38.i, %mul43.i
%arrayidx47.i = getelementptr inbounds [2 x double], ptr %A, i64 1, i64 0
%temp10 = load double, ptr %arrayidx47.i, align 8
%mul50.i = fmul double %temp1, %temp10
%arrayidx52.i = getelementptr inbounds [2 x double], ptr %A, i64 1, i64 1
%temp11 = load double, ptr %arrayidx52.i, align 8
%mul55.i = fmul double %temp3, %temp11
%add56.i = fadd double %mul50.i, %mul55.i
%mul62.i = fmul double %temp4, %temp10
%mul67.i = fmul double %temp5, %temp11
%add68.i = fadd double %mul62.i, %mul67.i
%mul74.i = fmul double %temp6, %temp10
%mul79.i = fmul double %temp7, %temp11
%add80.i = fadd double %mul74.i, %mul79.i
%mul86.i = fmul double %temp8, %temp10
%mul91.i = fmul double %temp9, %temp11
%add92.i = fadd double %mul86.i, %mul91.i
store double %add.i, ptr %Out, align 8
%Res.i.sroa.4.0.Out2.i.sroa_idx2 = getelementptr inbounds double, ptr %Out, i64 1
store double %add20.i, ptr %Res.i.sroa.4.0.Out2.i.sroa_idx2, align 8
%Res.i.sroa.5.0.Out2.i.sroa_idx4 = getelementptr inbounds double, ptr %Out, i64 2
store double %add32.i, ptr %Res.i.sroa.5.0.Out2.i.sroa_idx4, align 8
%Res.i.sroa.6.0.Out2.i.sroa_idx6 = getelementptr inbounds double, ptr %Out, i64 3
store double %add44.i, ptr %Res.i.sroa.6.0.Out2.i.sroa_idx6, align 8
%Res.i.sroa.7.0.Out2.i.sroa_idx8 = getelementptr inbounds double, ptr %Out, i64 4
store double %add56.i, ptr %Res.i.sroa.7.0.Out2.i.sroa_idx8, align 8
%Res.i.sroa.8.0.Out2.i.sroa_idx10 = getelementptr inbounds double, ptr %Out, i64 5
store double %add68.i, ptr %Res.i.sroa.8.0.Out2.i.sroa_idx10, align 8
%Res.i.sroa.9.0.Out2.i.sroa_idx12 = getelementptr inbounds double, ptr %Out, i64 6
store double %add80.i, ptr %Res.i.sroa.9.0.Out2.i.sroa_idx12, align 8
%Res.i.sroa.10.0.Out2.i.sroa_idx14 = getelementptr inbounds double, ptr %Out, i64 7
store double %add92.i, ptr %Res.i.sroa.10.0.Out2.i.sroa_idx14, align 8
ret void
}
|