You've already forked linux-packaging-mono
Imported Upstream version 5.18.0.167
Former-commit-id: 289509151e0fee68a1b591a20c9f109c3c789d3a
This commit is contained in:
parent
e19d552987
commit
b084638f15
@ -1,2 +0,0 @@
|
||||
if not 'ARM' in config.root.targets:
|
||||
config.unsupported = True
|
@ -1,20 +0,0 @@
|
||||
; RUN: opt < %s -basicaa -slp-vectorizer -S -mtriple=thumbv7-apple-ios3.0.0 -mcpu=swift | FileCheck %s
|
||||
|
||||
target datalayout = "e-p:32:32:32-i1:8:32-i8:8:32-i16:16:32-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:32:64-v128:32:128-a0:0:32-n32-S32"
|
||||
|
||||
; On swift unaligned <2 x double> stores need 4uops and it is there for cheaper
|
||||
; to do this scalar.
|
||||
|
||||
; CHECK-LABEL: expensive_double_store
|
||||
; CHECK-NOT: load <2 x double>
|
||||
; CHECK-NOT: store <2 x double>
|
||||
define void @expensive_double_store(double* noalias %dst, double* noalias %src, i64 %count) {
|
||||
entry:
|
||||
%0 = load double, double* %src, align 8
|
||||
store double %0, double* %dst, align 8
|
||||
%arrayidx2 = getelementptr inbounds double, double* %src, i64 1
|
||||
%1 = load double, double* %arrayidx2, align 8
|
||||
%arrayidx3 = getelementptr inbounds double, double* %dst, i64 1
|
||||
store double %1, double* %arrayidx3, align 8
|
||||
ret void
|
||||
}
|
@ -1,52 +0,0 @@
|
||||
; RUN: opt -S -mcpu=swift -mtriple=thumbv7-apple-ios -basicaa -slp-vectorizer < %s | FileCheck %s
|
||||
|
||||
target datalayout = "e-p:32:32:32-i1:8:32-i8:8:32-i16:16:32-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:32:64-v128:32:128-a0:0:32-n32-S32"
|
||||
|
||||
%class.Complex = type { double, double }
|
||||
|
||||
; Code like this is the result of SROA. Make sure we don't vectorize this
|
||||
; because the scalar version of the shl/or are handled by the
|
||||
; backend and disappear, the vectorized code stays.
|
||||
|
||||
; CHECK-LABEL: SROAed
|
||||
; CHECK-NOT: shl nuw <2 x i64>
|
||||
; CHECK-NOT: or <2 x i64>
|
||||
|
||||
define void @SROAed(%class.Complex* noalias nocapture sret %agg.result, [4 x i32] %a.coerce, [4 x i32] %b.coerce) {
|
||||
entry:
|
||||
%a.coerce.fca.0.extract = extractvalue [4 x i32] %a.coerce, 0
|
||||
%a.sroa.0.0.insert.ext = zext i32 %a.coerce.fca.0.extract to i64
|
||||
%a.coerce.fca.1.extract = extractvalue [4 x i32] %a.coerce, 1
|
||||
%a.sroa.0.4.insert.ext = zext i32 %a.coerce.fca.1.extract to i64
|
||||
%a.sroa.0.4.insert.shift = shl nuw i64 %a.sroa.0.4.insert.ext, 32
|
||||
%a.sroa.0.4.insert.insert = or i64 %a.sroa.0.4.insert.shift, %a.sroa.0.0.insert.ext
|
||||
%0 = bitcast i64 %a.sroa.0.4.insert.insert to double
|
||||
%a.coerce.fca.2.extract = extractvalue [4 x i32] %a.coerce, 2
|
||||
%a.sroa.3.8.insert.ext = zext i32 %a.coerce.fca.2.extract to i64
|
||||
%a.coerce.fca.3.extract = extractvalue [4 x i32] %a.coerce, 3
|
||||
%a.sroa.3.12.insert.ext = zext i32 %a.coerce.fca.3.extract to i64
|
||||
%a.sroa.3.12.insert.shift = shl nuw i64 %a.sroa.3.12.insert.ext, 32
|
||||
%a.sroa.3.12.insert.insert = or i64 %a.sroa.3.12.insert.shift, %a.sroa.3.8.insert.ext
|
||||
%1 = bitcast i64 %a.sroa.3.12.insert.insert to double
|
||||
%b.coerce.fca.0.extract = extractvalue [4 x i32] %b.coerce, 0
|
||||
%b.sroa.0.0.insert.ext = zext i32 %b.coerce.fca.0.extract to i64
|
||||
%b.coerce.fca.1.extract = extractvalue [4 x i32] %b.coerce, 1
|
||||
%b.sroa.0.4.insert.ext = zext i32 %b.coerce.fca.1.extract to i64
|
||||
%b.sroa.0.4.insert.shift = shl nuw i64 %b.sroa.0.4.insert.ext, 32
|
||||
%b.sroa.0.4.insert.insert = or i64 %b.sroa.0.4.insert.shift, %b.sroa.0.0.insert.ext
|
||||
%2 = bitcast i64 %b.sroa.0.4.insert.insert to double
|
||||
%b.coerce.fca.2.extract = extractvalue [4 x i32] %b.coerce, 2
|
||||
%b.sroa.3.8.insert.ext = zext i32 %b.coerce.fca.2.extract to i64
|
||||
%b.coerce.fca.3.extract = extractvalue [4 x i32] %b.coerce, 3
|
||||
%b.sroa.3.12.insert.ext = zext i32 %b.coerce.fca.3.extract to i64
|
||||
%b.sroa.3.12.insert.shift = shl nuw i64 %b.sroa.3.12.insert.ext, 32
|
||||
%b.sroa.3.12.insert.insert = or i64 %b.sroa.3.12.insert.shift, %b.sroa.3.8.insert.ext
|
||||
%3 = bitcast i64 %b.sroa.3.12.insert.insert to double
|
||||
%add = fadd double %0, %2
|
||||
%add3 = fadd double %1, %3
|
||||
%re.i.i = getelementptr inbounds %class.Complex, %class.Complex* %agg.result, i32 0, i32 0
|
||||
store double %add, double* %re.i.i, align 4
|
||||
%im.i.i = getelementptr inbounds %class.Complex, %class.Complex* %agg.result, i32 0, i32 1
|
||||
store double %add3, double* %im.i.i, align 4
|
||||
ret void
|
||||
}
|
Reference in New Issue
Block a user