; RUN: llc -mtriple=i386-unknown-freebsd -mcpu=core2 -stack-alignment=4 -relocation-model=pic < %s | FileCheck %s -check-prefix=UNALIGNED
; RUN: llc -mtriple=i386-unknown-freebsd -mcpu=core2 -stack-alignment=16 -relocation-model=pic < %s | FileCheck %s -check-prefix=ALIGNED
; RUN: llc -mtriple=i386-unknown-freebsd -mcpu=core2 -stack-alignment=4 -stackrealign -relocation-model=pic < %s | FileCheck %s -check-prefix=FORCEALIGNED

@arr = internal unnamed_addr global [32 x i32] zeroinitializer, align 16

; PR12250
define i32 @test1() {
vector.ph:
  br label %vector.body

vector.body:
  %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
  %0 = getelementptr inbounds [32 x i32], [32 x i32]* @arr, i32 0, i32 %index
  %1 = bitcast i32* %0 to <4 x i32>*
  %wide.load = load <4 x i32>, <4 x i32>* %1, align 16
  %2 = add nsw <4 x i32> %wide.load, <i32 10, i32 10, i32 10, i32 10>
  %3 = xor <4 x i32> %2, <i32 123345, i32 123345, i32 123345, i32 123345>
  %4 = add nsw <4 x i32> %3, <i32 112, i32 112, i32 112, i32 112>
  %5 = xor <4 x i32> %4, <i32 543345, i32 543345, i32 543345, i32 543345>
  %6 = add nsw <4 x i32> %5, <i32 73, i32 73, i32 73, i32 73>
  %7 = xor <4 x i32> %6, <i32 345987, i32 345987, i32 345987, i32 345987>
  %8 = add nsw <4 x i32> %7, <i32 48, i32 48, i32 48, i32 48>
  %9 = xor <4 x i32> %8, <i32 123987, i32 123987, i32 123987, i32 123987>
  store <4 x i32> %9, <4 x i32>* %1, align 16
  %index.next = add i32 %index, 4
  %10 = icmp eq i32 %index.next, 32
  br i1 %10, label %middle.block, label %vector.body

middle.block:
  ret i32 0

; We can't fold the spill into a padd unless the stack is aligned. Just spilling
; doesn't force stack realignment though
; UNALIGNED-LABEL: @test1
; UNALIGNED-NOT: andl $-{{..}}, %esp
; UNALIGNED: movdqu {{.*}} # 16-byte Folded Spill
; UNALIGNED-NOT: paddd {{.*}} # 16-byte Folded Reload

; ALIGNED-LABEL: @test1
; ALIGNED-NOT: andl $-{{..}}, %esp
; ALIGNED: movdqa {{.*}} # 16-byte Spill
; ALIGNED: paddd {{.*}} # 16-byte Folded Reload

; FORCEALIGNED-LABEL: @test1
; FORCEALIGNED: andl $-{{..}}, %esp
; FORCEALIGNED: movdqa {{.*}} # 16-byte Spill
; FORCEALIGNED: paddd {{.*}} # 16-byte Folded Reload
}