| ; RUN: opt < %s -basicaa -slp-vectorizer -S -slp-schedule-budget=16 -mtriple=x86_64-apple-macosx10.8.0 -mcpu=corei7-avx | FileCheck %s |
| |
| target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128" |
| target triple = "x86_64-apple-macosx10.9.0" |
| |
| ; Test if the budget for the scheduling region size works. |
| ; We test with a reduced budget of 16 which should prevent vectorizing the loads. |
| |
| declare void @unknown() |
| |
| ; CHECK-LABEL: @test |
| ; CHECK: load float |
| ; CHECK: load float |
| ; CHECK: load float |
| ; CHECK: load float |
| ; CHECK: call void @unknown |
| ; CHECK: store float |
| ; CHECK: store float |
| ; CHECK: store float |
| ; CHECK: store float |
| ; CHECK: load <4 x float> |
| ; CHECK: store <4 x float> |
| define void @test(float * %a, float * %b, float * %c, float * %d) { |
| entry: |
| ; Don't vectorize these loads. |
| %l0 = load float, float* %a |
| %a1 = getelementptr inbounds float, float* %a, i64 1 |
| %l1 = load float, float* %a1 |
| %a2 = getelementptr inbounds float, float* %a, i64 2 |
| %l2 = load float, float* %a2 |
| %a3 = getelementptr inbounds float, float* %a, i64 3 |
| %l3 = load float, float* %a3 |
| |
| ; some unrelated instructions inbetween to enlarge the scheduling region |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| call void @unknown() |
| |
| ; Don't vectorize these stores because their operands are too far away. |
| store float %l0, float* %b |
| %b1 = getelementptr inbounds float, float* %b, i64 1 |
| store float %l1, float* %b1 |
| %b2 = getelementptr inbounds float, float* %b, i64 2 |
| store float %l2, float* %b2 |
| %b3 = getelementptr inbounds float, float* %b, i64 3 |
| store float %l3, float* %b3 |
| |
| ; But still vectorize the following instructions, because even if the budget |
| ; is exceeded there is a minimum region size. |
| %l4 = load float, float* %c |
| %c1 = getelementptr inbounds float, float* %c, i64 1 |
| %l5 = load float, float* %c1 |
| %c2 = getelementptr inbounds float, float* %c, i64 2 |
| %l6 = load float, float* %c2 |
| %c3 = getelementptr inbounds float, float* %c, i64 3 |
| %l7 = load float, float* %c3 |
| |
| store float %l4, float* %d |
| %d1 = getelementptr inbounds float, float* %d, i64 1 |
| store float %l5, float* %d1 |
| %d2 = getelementptr inbounds float, float* %d, i64 2 |
| store float %l6, float* %d2 |
| %d3 = getelementptr inbounds float, float* %d, i64 3 |
| store float %l7, float* %d3 |
| |
| ret void |
| } |
| |