1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81
  
     | 
    
      ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple aarch64 -mcpu=cortex-a57 -mattr=+slow-paired-128 < %s | FileCheck %s --check-prefixes=DEFAULT
; RUN: llc -mtriple aarch64 -mcpu=cortex-a57 -mattr=+slow-paired-128 -mattr=+ascend-store-address < %s | FileCheck %s --check-prefixes=ASCEND
target triple = "aarch64-unknown-linux-gnu"
define dso_local void @memset_unroll2(ptr nocapture %array, i64 %size) {
; DEFAULT-LABEL: memset_unroll2:
; DEFAULT:       // %bb.0: // %entry
; DEFAULT-NEXT:    fmov v0.2d, #2.00000000
; DEFAULT-NEXT:    add x8, x0, #64
; DEFAULT-NEXT:    .p2align 4, , 8
; DEFAULT-NEXT:  .LBB0_1: // %vector.body
; DEFAULT-NEXT:    // =>This Inner Loop Header: Depth=1
; DEFAULT-NEXT:    stur q0, [x8, #-64]
; DEFAULT-NEXT:    subs x1, x1, #4
; DEFAULT-NEXT:    stur q0, [x8, #-48]
; DEFAULT-NEXT:    str q0, [x8]
; DEFAULT-NEXT:    str q0, [x8, #16]
; DEFAULT-NEXT:    str q0, [x8, #32]
; DEFAULT-NEXT:    str q0, [x8, #48]
; DEFAULT-NEXT:    stur q0, [x8, #-32]
; DEFAULT-NEXT:    stur q0, [x8, #-16]
; DEFAULT-NEXT:    add x8, x8, #128
; DEFAULT-NEXT:    b.ne .LBB0_1
; DEFAULT-NEXT:  // %bb.2: // %cleanup
; DEFAULT-NEXT:    ret
;
; ASCEND-LABEL: memset_unroll2:
; ASCEND:       // %bb.0: // %entry
; ASCEND-NEXT:    fmov v0.2d, #2.00000000
; ASCEND-NEXT:    add x8, x0, #64
; ASCEND-NEXT:    .p2align 4, , 8
; ASCEND-NEXT:  .LBB0_1: // %vector.body
; ASCEND-NEXT:    // =>This Inner Loop Header: Depth=1
; ASCEND-NEXT:    stur q0, [x8, #-64]
; ASCEND-NEXT:    subs x1, x1, #4
; ASCEND-NEXT:    stur q0, [x8, #-48]
; ASCEND-NEXT:    stur q0, [x8, #-32]
; ASCEND-NEXT:    stur q0, [x8, #-16]
; ASCEND-NEXT:    str q0, [x8]
; ASCEND-NEXT:    str q0, [x8, #16]
; ASCEND-NEXT:    str q0, [x8, #32]
; ASCEND-NEXT:    str q0, [x8, #48]
; ASCEND-NEXT:    add x8, x8, #128
; ASCEND-NEXT:    b.ne .LBB0_1
; ASCEND-NEXT:  // %bb.2: // %cleanup
; ASCEND-NEXT:    ret
entry:
  br label %vector.body
vector.body:                                      ; preds = %vector.body, %entry
  %index = phi i64 [ 0, %entry ], [ %index16, %vector.body ]
  %niter = phi i64 [ %size, %entry ], [ %niter.nsub.3, %vector.body ]
  %array0 = getelementptr inbounds double, ptr %array, i64 %index
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array0, align 8
  %array2 = getelementptr inbounds double, ptr %array0, i64 2
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array2, align 8
  %index4 = or i64 %index, 4
  %array4 = getelementptr inbounds double, ptr %array, i64 %index4
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array4, align 8
  %array6 = getelementptr inbounds double, ptr %array4, i64 2
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array6, align 8
  %index8 = or i64 %index, 8
  %array8 = getelementptr inbounds double, ptr %array, i64 %index8
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array8, align 8
  %array10 = getelementptr inbounds double, ptr %array8, i64 2
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array10, align 8
  %index12 = or i64 %index, 12
  %array12 = getelementptr inbounds double, ptr %array, i64 %index12
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array12, align 8
  %array14 = getelementptr inbounds double, ptr %array12, i64 2
  store <2 x double> <double 2.000000e+00, double 2.000000e+00>, ptr %array14, align 8
  %index16 = add i64 %index, 16
  %niter.nsub.3 = add i64 %niter, -4
  %niter.ncmp.3 = icmp eq i64 %niter.nsub.3, 0
  br i1 %niter.ncmp.3, label %cleanup, label %vector.body
cleanup:                           ; preds = %vector.body
  ret void
}
 
     |