summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMatt Arsenault <Matthew.Arsenault@amd.com>2021-08-13 20:43:32 -0400
committerMatt Arsenault <Matthew.Arsenault@amd.com>2021-08-26 20:30:03 -0400
commit98d7aa435f5036b03048bdad24065c635959ac6b (patch)
tree250fbc9248c04cf6c39e1d4d989caf125ef4c0b8
parentf5cff292e297ad5094f35d6223a199e389adcd99 (diff)
downloadllvm-98d7aa435f5036b03048bdad24065c635959ac6b.tar.gz
AMDGPU: Stop inferring use of llvm.amdgcn.kernarg.segment.ptr
We no longer use this intrinsic outside of the backend and no longer support using it outside of kernels.
-rw-r--r--llvm/lib/Target/AMDGPU/AMDGPUAnnotateKernelFeatures.cpp16
-rw-r--r--llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp7
-rw-r--r--llvm/lib/Target/AMDGPU/AMDGPULegalizerInfo.cpp8
-rw-r--r--llvm/lib/Target/AMDGPU/SIISelLowering.cpp15
-rw-r--r--llvm/lib/Target/AMDGPU/SIMachineFunctionInfo.cpp5
-rw-r--r--llvm/test/CodeGen/AMDGPU/GlobalISel/irtranslator-call-return-values.ll3005
-rw-r--r--llvm/test/CodeGen/AMDGPU/GlobalISel/llvm.amdgcn.kernarg.segment.ptr.ll13
-rw-r--r--llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa-call.ll49
-rw-r--r--llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa.ll20
-rw-r--r--llvm/test/CodeGen/AMDGPU/llvm.amdgcn.kernarg.segment.ptr.ll13
10 files changed, 1560 insertions, 1591 deletions
diff --git a/llvm/lib/Target/AMDGPU/AMDGPUAnnotateKernelFeatures.cpp b/llvm/lib/Target/AMDGPU/AMDGPUAnnotateKernelFeatures.cpp
index af6dfc07eb50..d591be7052c9 100644
--- a/llvm/lib/Target/AMDGPU/AMDGPUAnnotateKernelFeatures.cpp
+++ b/llvm/lib/Target/AMDGPU/AMDGPUAnnotateKernelFeatures.cpp
@@ -173,8 +173,6 @@ static StringRef intrinsicToAttrName(Intrinsic::ID ID,
return "amdgpu-dispatch-ptr";
case Intrinsic::amdgcn_dispatch_id:
return "amdgpu-dispatch-id";
- case Intrinsic::amdgcn_kernarg_segment_ptr:
- return "amdgpu-kernarg-segment-ptr";
case Intrinsic::amdgcn_implicitarg_ptr:
return "amdgpu-implicitarg-ptr";
case Intrinsic::amdgcn_queue_ptr:
@@ -310,15 +308,11 @@ bool AMDGPUAnnotateKernelFeatures::addFeatureAttributes(Function &F) {
} else {
bool NonKernelOnly = false;
- if (!IsFunc && IID == Intrinsic::amdgcn_kernarg_segment_ptr) {
- F.addFnAttr("amdgpu-kernarg-segment-ptr");
- } else {
- StringRef AttrName = intrinsicToAttrName(IID, NonKernelOnly,
- NeedQueuePtr);
- if (!AttrName.empty() && (IsFunc || !NonKernelOnly)) {
- F.addFnAttr(AttrName);
- Changed = true;
- }
+ StringRef AttrName = intrinsicToAttrName(IID, NonKernelOnly,
+ NeedQueuePtr);
+ if (!AttrName.empty() && (IsFunc || !NonKernelOnly)) {
+ F.addFnAttr(AttrName);
+ Changed = true;
}
}
}
diff --git a/llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp b/llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp
index 0c24903490f0..82ae5b7edff5 100644
--- a/llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp
+++ b/llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp
@@ -60,8 +60,6 @@ static StringRef intrinsicToAttrName(Intrinsic::ID ID, bool &NonKernelOnly,
return "amdgpu-dispatch-ptr";
case Intrinsic::amdgcn_dispatch_id:
return "amdgpu-dispatch-id";
- case Intrinsic::amdgcn_kernarg_segment_ptr:
- return "amdgpu-kernarg-segment-ptr";
case Intrinsic::amdgcn_implicitarg_ptr:
return "amdgpu-implicitarg-ptr";
case Intrinsic::amdgcn_queue_ptr:
@@ -353,11 +351,6 @@ struct AAAMDAttributesFunction : public AAAMDAttributes {
for (Function *Callee : AAEdges.getOptimisticEdges()) {
Intrinsic::ID IID = Callee->getIntrinsicID();
if (IID != Intrinsic::not_intrinsic) {
- if (!IsNonEntryFunc && IID == Intrinsic::amdgcn_kernarg_segment_ptr) {
- AddAttribute("amdgpu-kernarg-segment-ptr");
- continue;
- }
-
bool NonKernelOnly = false;
StringRef AttrName =
intrinsicToAttrName(IID, NonKernelOnly, NeedsQueuePtr);
diff --git a/llvm/lib/Target/AMDGPU/AMDGPULegalizerInfo.cpp b/llvm/lib/Target/AMDGPU/AMDGPULegalizerInfo.cpp
index 71e120974c2e..5ce72d9c5957 100644
--- a/llvm/lib/Target/AMDGPU/AMDGPULegalizerInfo.cpp
+++ b/llvm/lib/Target/AMDGPU/AMDGPULegalizerInfo.cpp
@@ -2888,6 +2888,14 @@ bool AMDGPULegalizerInfo::loadInputValue(
LLT ArgTy;
std::tie(Arg, ArgRC, ArgTy) = MFI->getPreloadedValue(ArgType);
+ if (!Arg) {
+ assert(ArgType == AMDGPUFunctionArgInfo::KERNARG_SEGMENT_PTR);
+ // The intrinsic may appear when we have a 0 sized kernarg segment, in which
+ // case the pointer argument may be missing and we use null.
+ B.buildConstant(DstReg, 0);
+ return true;
+ }
+
if (!Arg->isRegister() || !Arg->getRegister().isValid())
return false; // TODO: Handle these
return loadInputValue(DstReg, B, Arg, ArgRC, ArgTy);
diff --git a/llvm/lib/Target/AMDGPU/SIISelLowering.cpp b/llvm/lib/Target/AMDGPU/SIISelLowering.cpp
index faf38bb0b914..5d6717e0490f 100644
--- a/llvm/lib/Target/AMDGPU/SIISelLowering.cpp
+++ b/llvm/lib/Target/AMDGPU/SIISelLowering.cpp
@@ -1663,12 +1663,17 @@ SDValue SITargetLowering::lowerKernArgParameterPtr(SelectionDAG &DAG,
const ArgDescriptor *InputPtrReg;
const TargetRegisterClass *RC;
LLT ArgTy;
+ MVT PtrVT = getPointerTy(DL, AMDGPUAS::CONSTANT_ADDRESS);
std::tie(InputPtrReg, RC, ArgTy) =
Info->getPreloadedValue(AMDGPUFunctionArgInfo::KERNARG_SEGMENT_PTR);
+ // We may not have the kernarg segment argument if we have no kernel
+ // arguments.
+ if (!InputPtrReg)
+ return DAG.getConstant(0, SL, PtrVT);
+
MachineRegisterInfo &MRI = DAG.getMachineFunction().getRegInfo();
- MVT PtrVT = getPointerTy(DL, AMDGPUAS::CONSTANT_ADDRESS);
SDValue BasePtr = DAG.getCopyFromReg(Chain, SL,
MRI.getLiveInVirtReg(InputPtrReg->getRegister()), PtrVT);
@@ -1814,6 +1819,14 @@ SDValue SITargetLowering::getPreloadedValue(SelectionDAG &DAG,
LLT Ty;
std::tie(Reg, RC, Ty) = MFI.getPreloadedValue(PVID);
+ if (!Reg) {
+ // It's possible for a kernarg intrinsic call to appear in a kernel with no
+ // allocated segment, in which case we do not add the user sgpr argument, so
+ // just return null.
+ assert(PVID == AMDGPUFunctionArgInfo::PreloadedValue::KERNARG_SEGMENT_PTR);
+ return DAG.getConstant(0, SDLoc(), VT);
+ }
+
return CreateLiveInRegister(DAG, RC, Reg->getRegister(), VT);
}
diff --git a/llvm/lib/Target/AMDGPU/SIMachineFunctionInfo.cpp b/llvm/lib/Target/AMDGPU/SIMachineFunctionInfo.cpp
index e59f825f5e88..71883411da28 100644
--- a/llvm/lib/Target/AMDGPU/SIMachineFunctionInfo.cpp
+++ b/llvm/lib/Target/AMDGPU/SIMachineFunctionInfo.cpp
@@ -69,7 +69,7 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
(!isEntryFunction() || HasCalls);
if (CC == CallingConv::AMDGPU_KERNEL || CC == CallingConv::SPIR_KERNEL) {
- if (!F.arg_empty())
+ if (!F.arg_empty() || ST.getImplicitArgNumBytes(F) != 0)
KernargSegmentPtr = true;
WorkGroupIDX = true;
WorkItemIDX = true;
@@ -156,9 +156,6 @@ SIMachineFunctionInfo::SIMachineFunctionInfo(const MachineFunction &MF)
else if (ST.isMesaGfxShader(F))
ImplicitBufferPtr = true;
- if (UseFixedABI || F.hasFnAttribute("amdgpu-kernarg-segment-ptr"))
- KernargSegmentPtr = true;
-
if (!AMDGPU::isGraphics(CC)) {
if (UseFixedABI) {
DispatchPtr = true;
diff --git a/llvm/test/CodeGen/AMDGPU/GlobalISel/irtranslator-call-return-values.ll b/llvm/test/CodeGen/AMDGPU/GlobalISel/irtranslator-call-return-values.ll
index 0b228cceebd4..2e8bfdf4e525 100644
--- a/llvm/test/CodeGen/AMDGPU/GlobalISel/irtranslator-call-return-values.ll
+++ b/llvm/test/CodeGen/AMDGPU/GlobalISel/irtranslator-call-return-values.ll
@@ -153,51 +153,50 @@ define amdgpu_gfx void @test_gfx_call_external_i32_func_i32_imm(i32 addrspace(1)
define amdgpu_kernel void @test_call_external_i1_func_void() #0 {
; GCN-LABEL: name: test_call_external_i1_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i1_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i1_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC]](s1), [[DEF]](p1) :: (volatile store (s1) into `i1 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -231,51 +230,50 @@ define amdgpu_gfx void @test_gfx_call_external_i1_func_void() #0 {
define amdgpu_kernel void @test_call_external_i1_zeroext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i1_zeroext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i1_zeroext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i1_zeroext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[ZEXT:%[0-9]+]]:_(s32) = G_ZEXT [[TRUNC]](s1)
; GCN: G_STORE [[ZEXT]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
@@ -289,51 +287,50 @@ define amdgpu_kernel void @test_call_external_i1_zeroext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i1_signext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i1_signext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i1_signext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i1_signext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s1) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[SEXT:%[0-9]+]]:_(s32) = G_SEXT [[TRUNC]](s1)
; GCN: G_STORE [[SEXT]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
@@ -347,51 +344,50 @@ define amdgpu_kernel void @test_call_external_i1_signext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i8_func_void() #0 {
; GCN-LABEL: name: test_call_external_i8_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i8_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i8_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: [[TRUNC1:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC]](s16)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC1]](s8), [[DEF]](p1) :: (volatile store (s8) into `i8 addrspace(1)* undef`, addrspace 1)
@@ -427,51 +423,50 @@ define amdgpu_gfx void @test_gfx_call_external_i8_func_void() #0 {
define amdgpu_kernel void @test_call_external_i8_zeroext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i8_zeroext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i8_zeroext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i8_zeroext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: [[TRUNC1:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC]](s16)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[ZEXT:%[0-9]+]]:_(s32) = G_ZEXT [[TRUNC1]](s8)
@@ -486,51 +481,50 @@ define amdgpu_kernel void @test_call_external_i8_zeroext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i8_signext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i8_signext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i8_signext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i8_signext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: [[TRUNC1:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC]](s16)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[SEXT:%[0-9]+]]:_(s32) = G_SEXT [[TRUNC1]](s8)
@@ -545,51 +539,50 @@ define amdgpu_kernel void @test_call_external_i8_signext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i16_func_void() #0 {
; GCN-LABEL: name: test_call_external_i16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC]](s16), [[DEF]](p1) :: (volatile store (s16) into `i16 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -601,51 +594,50 @@ define amdgpu_kernel void @test_call_external_i16_func_void() #0 {
define amdgpu_kernel void @test_call_external_i16_zeroext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i16_zeroext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i16_zeroext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i16_zeroext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[ZEXT:%[0-9]+]]:_(s32) = G_ZEXT [[TRUNC]](s16)
; GCN: G_STORE [[ZEXT]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
@@ -659,51 +651,50 @@ define amdgpu_kernel void @test_call_external_i16_zeroext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i16_signext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i16_signext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i16_signext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i16_signext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[SEXT:%[0-9]+]]:_(s32) = G_SEXT [[TRUNC]](s16)
; GCN: G_STORE [[SEXT]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
@@ -717,52 +708,51 @@ define amdgpu_kernel void @test_call_external_i16_signext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY19]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call i32 @external_i32_func_void()
store volatile i32 %val, i32 addrspace(1)* undef
@@ -793,52 +783,51 @@ define amdgpu_gfx void @test_gfx_call_external_i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_i48_func_void() #0 {
; GCN-LABEL: name: test_call_external_i48_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i48_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i48_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: [[TRUNC:%[0-9]+]]:_(s48) = G_TRUNC [[MV]](s64)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC]](s48), [[DEF]](p1) :: (volatile store (s48) into `i48 addrspace(1)* undef`, align 8, addrspace 1)
@@ -851,52 +840,51 @@ define amdgpu_kernel void @test_call_external_i48_func_void() #0 {
define amdgpu_kernel void @test_call_external_i48_zeroext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i48_zeroext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i48_zeroext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i48_zeroext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: [[TRUNC:%[0-9]+]]:_(s48) = G_TRUNC [[MV]](s64)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[ZEXT:%[0-9]+]]:_(s64) = G_ZEXT [[TRUNC]](s48)
@@ -911,52 +899,51 @@ define amdgpu_kernel void @test_call_external_i48_zeroext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i48_signext_func_void() #0 {
; GCN-LABEL: name: test_call_external_i48_signext_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i48_signext_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i48_signext_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: [[TRUNC:%[0-9]+]]:_(s48) = G_TRUNC [[MV]](s64)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[SEXT:%[0-9]+]]:_(s64) = G_SEXT [[TRUNC]](s48)
@@ -971,52 +958,51 @@ define amdgpu_kernel void @test_call_external_i48_signext_func_void() #0 {
define amdgpu_kernel void @test_call_external_i64_func_void() #0 {
; GCN-LABEL: name: test_call_external_i64_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i64_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i64_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[MV]](s64), [[DEF]](p1) :: (volatile store (s64) into `i64 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1028,52 +1014,51 @@ define amdgpu_kernel void @test_call_external_i64_func_void() #0 {
define amdgpu_kernel void @test_call_external_p1_func_void() #0 {
; GCN-LABEL: name: test_call_external_p1_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_p1_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_p1_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[MV]](p1), [[DEF]](p1) :: (volatile store (p1) into `i8 addrspace(1)* addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1085,55 +1070,54 @@ define amdgpu_kernel void @test_call_external_p1_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2p1_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2p1_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2p1_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2p1_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
- ; GCN: [[MV1:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY23]](s32), [[COPY24]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[MV:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
+ ; GCN: [[MV1:%[0-9]+]]:_(p1) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x p1>) = G_BUILD_VECTOR [[MV]](p1), [[MV1]](p1)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<2 x p1>), [[DEF]](p1) :: (volatile store (<2 x p1>) into `<2 x i8 addrspace(1)*> addrspace(1)* undef`, addrspace 1)
@@ -1146,52 +1130,51 @@ define amdgpu_kernel void @test_call_external_v2p1_func_void() #0 {
define amdgpu_kernel void @test_call_external_p3_func_void() #0 {
; GCN-LABEL: name: test_call_external_p3_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p3) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_p3_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_p3_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(p3) = COPY $vgpr0
+ ; GCN: [[COPY19:%[0-9]+]]:_(p3) = COPY $vgpr0
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](p3), [[DEF]](p3) :: (volatile store (p3) into `i8 addrspace(3)* addrspace(3)* undef`, addrspace 3)
+ ; GCN: G_STORE [[COPY19]](p3), [[DEF]](p3) :: (volatile store (p3) into `i8 addrspace(3)* addrspace(3)* undef`, addrspace 3)
; GCN: S_ENDPGM 0
%val = call i8 addrspace(3)* @external_p3_func_void()
store volatile i8 addrspace(3)* %val, i8 addrspace(3)* addrspace(3)* undef
@@ -1201,52 +1184,51 @@ define amdgpu_kernel void @test_call_external_p3_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2p3_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2p3_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p3) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2p3_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2p3_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(p3) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(p3) = COPY $vgpr1
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x p3>) = G_BUILD_VECTOR [[COPY21]](p3), [[COPY22]](p3)
+ ; GCN: [[COPY19:%[0-9]+]]:_(p3) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(p3) = COPY $vgpr1
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x p3>) = G_BUILD_VECTOR [[COPY19]](p3), [[COPY20]](p3)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<2 x p3>), [[DEF]](p3) :: (volatile store (<2 x p3>) into `<2 x i8 addrspace(3)*> addrspace(3)* undef`, addrspace 3)
; GCN: S_ENDPGM 0
@@ -1258,51 +1240,50 @@ define amdgpu_kernel void @test_call_external_v2p3_func_void() #0 {
define amdgpu_kernel void @test_call_external_f16_func_void() #0 {
; GCN-LABEL: name: test_call_external_f16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_f16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_f16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC]](s16), [[DEF]](p1) :: (volatile store (s16) into `half addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1314,52 +1295,51 @@ define amdgpu_kernel void @test_call_external_f16_func_void() #0 {
define amdgpu_kernel void @test_call_external_f32_func_void() #0 {
; GCN-LABEL: name: test_call_external_f32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_f32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_f32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](s32), [[DEF]](p1) :: (volatile store (s32) into `float addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY19]](s32), [[DEF]](p1) :: (volatile store (s32) into `float addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call float @external_f32_func_void()
store volatile float %val, float addrspace(1)* undef
@@ -1369,52 +1349,51 @@ define amdgpu_kernel void @test_call_external_f32_func_void() #0 {
define amdgpu_kernel void @test_call_external_f64_func_void() #0 {
; GCN-LABEL: name: test_call_external_f64_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_f64_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_f64_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[MV]](s64), [[DEF]](p1) :: (volatile store (s64) into `double addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1426,55 +1405,54 @@ define amdgpu_kernel void @test_call_external_f64_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2f64_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2f64_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2f64_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2f64_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
- ; GCN: [[MV1:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY23]](s32), [[COPY24]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY19]](s32), [[COPY20]](s32)
+ ; GCN: [[MV1:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x s64>) = G_BUILD_VECTOR [[MV]](s64), [[MV1]](s64)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<2 x s64>), [[DEF]](p1) :: (volatile store (<2 x s64>) into `<2 x double> addrspace(1)* undef`, addrspace 1)
@@ -1487,52 +1465,51 @@ define amdgpu_kernel void @test_call_external_v2f64_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<2 x s32>), [[DEF]](p1) :: (volatile store (<2 x s32>) into `<2 x i32> addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1544,53 +1521,52 @@ define amdgpu_kernel void @test_call_external_v2i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v3i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v3i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v3i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v3i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<3 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<3 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<3 x s32>), [[DEF]](p1) :: (volatile store (<3 x s32>) into `<3 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1602,54 +1578,53 @@ define amdgpu_kernel void @test_call_external_v3i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v4i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v4i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v4i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v4i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<4 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<4 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<4 x s32>), [[DEF]](p1) :: (volatile store (<4 x s32>) into `<4 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1661,55 +1636,54 @@ define amdgpu_kernel void @test_call_external_v4i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v5i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v5i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v5i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v5i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<5 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<5 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<5 x s32>), [[DEF]](p1) :: (volatile store (<5 x s32>) into `<5 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1721,58 +1695,57 @@ define amdgpu_kernel void @test_call_external_v5i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v8i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v8i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v8i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v8i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4, implicit-def $vgpr5, implicit-def $vgpr6, implicit-def $vgpr7
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr5
- ; GCN: [[COPY27:%[0-9]+]]:_(s32) = COPY $vgpr6
- ; GCN: [[COPY28:%[0-9]+]]:_(s32) = COPY $vgpr7
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<8 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32), [[COPY27]](s32), [[COPY28]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr5
+ ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr6
+ ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr7
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<8 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<8 x s32>), [[DEF]](p1) :: (volatile store (<8 x s32>) into `<8 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1784,66 +1757,65 @@ define amdgpu_kernel void @test_call_external_v8i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v16i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v16i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v16i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v16i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4, implicit-def $vgpr5, implicit-def $vgpr6, implicit-def $vgpr7, implicit-def $vgpr8, implicit-def $vgpr9, implicit-def $vgpr10, implicit-def $vgpr11, implicit-def $vgpr12, implicit-def $vgpr13, implicit-def $vgpr14, implicit-def $vgpr15
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr5
- ; GCN: [[COPY27:%[0-9]+]]:_(s32) = COPY $vgpr6
- ; GCN: [[COPY28:%[0-9]+]]:_(s32) = COPY $vgpr7
- ; GCN: [[COPY29:%[0-9]+]]:_(s32) = COPY $vgpr8
- ; GCN: [[COPY30:%[0-9]+]]:_(s32) = COPY $vgpr9
- ; GCN: [[COPY31:%[0-9]+]]:_(s32) = COPY $vgpr10
- ; GCN: [[COPY32:%[0-9]+]]:_(s32) = COPY $vgpr11
- ; GCN: [[COPY33:%[0-9]+]]:_(s32) = COPY $vgpr12
- ; GCN: [[COPY34:%[0-9]+]]:_(s32) = COPY $vgpr13
- ; GCN: [[COPY35:%[0-9]+]]:_(s32) = COPY $vgpr14
- ; GCN: [[COPY36:%[0-9]+]]:_(s32) = COPY $vgpr15
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<16 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32), [[COPY27]](s32), [[COPY28]](s32), [[COPY29]](s32), [[COPY30]](s32), [[COPY31]](s32), [[COPY32]](s32), [[COPY33]](s32), [[COPY34]](s32), [[COPY35]](s32), [[COPY36]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr5
+ ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr6
+ ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr7
+ ; GCN: [[COPY27:%[0-9]+]]:_(s32) = COPY $vgpr8
+ ; GCN: [[COPY28:%[0-9]+]]:_(s32) = COPY $vgpr9
+ ; GCN: [[COPY29:%[0-9]+]]:_(s32) = COPY $vgpr10
+ ; GCN: [[COPY30:%[0-9]+]]:_(s32) = COPY $vgpr11
+ ; GCN: [[COPY31:%[0-9]+]]:_(s32) = COPY $vgpr12
+ ; GCN: [[COPY32:%[0-9]+]]:_(s32) = COPY $vgpr13
+ ; GCN: [[COPY33:%[0-9]+]]:_(s32) = COPY $vgpr14
+ ; GCN: [[COPY34:%[0-9]+]]:_(s32) = COPY $vgpr15
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<16 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32), [[COPY27]](s32), [[COPY28]](s32), [[COPY29]](s32), [[COPY30]](s32), [[COPY31]](s32), [[COPY32]](s32), [[COPY33]](s32), [[COPY34]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<16 x s32>), [[DEF]](p1) :: (volatile store (<16 x s32>) into `<16 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1855,82 +1827,81 @@ define amdgpu_kernel void @test_call_external_v16i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v32i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v32i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v32i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v32i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4, implicit-def $vgpr5, implicit-def $vgpr6, implicit-def $vgpr7, implicit-def $vgpr8, implicit-def $vgpr9, implicit-def $vgpr10, implicit-def $vgpr11, implicit-def $vgpr12, implicit-def $vgpr13, implicit-def $vgpr14, implicit-def $vgpr15, implicit-def $vgpr16, implicit-def $vgpr17, implicit-def $vgpr18, implicit-def $vgpr19, implicit-def $vgpr20, implicit-def $vgpr21, implicit-def $vgpr22, implicit-def $vgpr23, implicit-def $vgpr24, implicit-def $vgpr25, implicit-def $vgpr26, implicit-def $vgpr27, implicit-def $vgpr28, implicit-def $vgpr29, implicit-def $vgpr30, implicit-def $vgpr31
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr5
- ; GCN: [[COPY27:%[0-9]+]]:_(s32) = COPY $vgpr6
- ; GCN: [[COPY28:%[0-9]+]]:_(s32) = COPY $vgpr7
- ; GCN: [[COPY29:%[0-9]+]]:_(s32) = COPY $vgpr8
- ; GCN: [[COPY30:%[0-9]+]]:_(s32) = COPY $vgpr9
- ; GCN: [[COPY31:%[0-9]+]]:_(s32) = COPY $vgpr10
- ; GCN: [[COPY32:%[0-9]+]]:_(s32) = COPY $vgpr11
- ; GCN: [[COPY33:%[0-9]+]]:_(s32) = COPY $vgpr12
- ; GCN: [[COPY34:%[0-9]+]]:_(s32) = COPY $vgpr13
- ; GCN: [[COPY35:%[0-9]+]]:_(s32) = COPY $vgpr14
- ; GCN: [[COPY36:%[0-9]+]]:_(s32) = COPY $vgpr15
- ; GCN: [[COPY37:%[0-9]+]]:_(s32) = COPY $vgpr16
- ; GCN: [[COPY38:%[0-9]+]]:_(s32) = COPY $vgpr17
- ; GCN: [[COPY39:%[0-9]+]]:_(s32) = COPY $vgpr18
- ; GCN: [[COPY40:%[0-9]+]]:_(s32) = COPY $vgpr19
- ; GCN: [[COPY41:%[0-9]+]]:_(s32) = COPY $vgpr20
- ; GCN: [[COPY42:%[0-9]+]]:_(s32) = COPY $vgpr21
- ; GCN: [[COPY43:%[0-9]+]]:_(s32) = COPY $vgpr22
- ; GCN: [[COPY44:%[0-9]+]]:_(s32) = COPY $vgpr23
- ; GCN: [[COPY45:%[0-9]+]]:_(s32) = COPY $vgpr24
- ; GCN: [[COPY46:%[0-9]+]]:_(s32) = COPY $vgpr25
- ; GCN: [[COPY47:%[0-9]+]]:_(s32) = COPY $vgpr26
- ; GCN: [[COPY48:%[0-9]+]]:_(s32) = COPY $vgpr27
- ; GCN: [[COPY49:%[0-9]+]]:_(s32) = COPY $vgpr28
- ; GCN: [[COPY50:%[0-9]+]]:_(s32) = COPY $vgpr29
- ; GCN: [[COPY51:%[0-9]+]]:_(s32) = COPY $vgpr30
- ; GCN: [[COPY52:%[0-9]+]]:_(s32) = COPY $vgpr31
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<32 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32), [[COPY27]](s32), [[COPY28]](s32), [[COPY29]](s32), [[COPY30]](s32), [[COPY31]](s32), [[COPY32]](s32), [[COPY33]](s32), [[COPY34]](s32), [[COPY35]](s32), [[COPY36]](s32), [[COPY37]](s32), [[COPY38]](s32), [[COPY39]](s32), [[COPY40]](s32), [[COPY41]](s32), [[COPY42]](s32), [[COPY43]](s32), [[COPY44]](s32), [[COPY45]](s32), [[COPY46]](s32), [[COPY47]](s32), [[COPY48]](s32), [[COPY49]](s32), [[COPY50]](s32), [[COPY51]](s32), [[COPY52]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr5
+ ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr6
+ ; GCN: [[COPY26:%[0-9]+]]:_(s32) = COPY $vgpr7
+ ; GCN: [[COPY27:%[0-9]+]]:_(s32) = COPY $vgpr8
+ ; GCN: [[COPY28:%[0-9]+]]:_(s32) = COPY $vgpr9
+ ; GCN: [[COPY29:%[0-9]+]]:_(s32) = COPY $vgpr10
+ ; GCN: [[COPY30:%[0-9]+]]:_(s32) = COPY $vgpr11
+ ; GCN: [[COPY31:%[0-9]+]]:_(s32) = COPY $vgpr12
+ ; GCN: [[COPY32:%[0-9]+]]:_(s32) = COPY $vgpr13
+ ; GCN: [[COPY33:%[0-9]+]]:_(s32) = COPY $vgpr14
+ ; GCN: [[COPY34:%[0-9]+]]:_(s32) = COPY $vgpr15
+ ; GCN: [[COPY35:%[0-9]+]]:_(s32) = COPY $vgpr16
+ ; GCN: [[COPY36:%[0-9]+]]:_(s32) = COPY $vgpr17
+ ; GCN: [[COPY37:%[0-9]+]]:_(s32) = COPY $vgpr18
+ ; GCN: [[COPY38:%[0-9]+]]:_(s32) = COPY $vgpr19
+ ; GCN: [[COPY39:%[0-9]+]]:_(s32) = COPY $vgpr20
+ ; GCN: [[COPY40:%[0-9]+]]:_(s32) = COPY $vgpr21
+ ; GCN: [[COPY41:%[0-9]+]]:_(s32) = COPY $vgpr22
+ ; GCN: [[COPY42:%[0-9]+]]:_(s32) = COPY $vgpr23
+ ; GCN: [[COPY43:%[0-9]+]]:_(s32) = COPY $vgpr24
+ ; GCN: [[COPY44:%[0-9]+]]:_(s32) = COPY $vgpr25
+ ; GCN: [[COPY45:%[0-9]+]]:_(s32) = COPY $vgpr26
+ ; GCN: [[COPY46:%[0-9]+]]:_(s32) = COPY $vgpr27
+ ; GCN: [[COPY47:%[0-9]+]]:_(s32) = COPY $vgpr28
+ ; GCN: [[COPY48:%[0-9]+]]:_(s32) = COPY $vgpr29
+ ; GCN: [[COPY49:%[0-9]+]]:_(s32) = COPY $vgpr30
+ ; GCN: [[COPY50:%[0-9]+]]:_(s32) = COPY $vgpr31
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<32 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32), [[COPY26]](s32), [[COPY27]](s32), [[COPY28]](s32), [[COPY29]](s32), [[COPY30]](s32), [[COPY31]](s32), [[COPY32]](s32), [[COPY33]](s32), [[COPY34]](s32), [[COPY35]](s32), [[COPY36]](s32), [[COPY37]](s32), [[COPY38]](s32), [[COPY39]](s32), [[COPY40]](s32), [[COPY41]](s32), [[COPY42]](s32), [[COPY43]](s32), [[COPY44]](s32), [[COPY45]](s32), [[COPY46]](s32), [[COPY47]](s32), [[COPY48]](s32), [[COPY49]](s32), [[COPY50]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<32 x s32>), [[DEF]](p1) :: (volatile store (<32 x s32>) into `<32 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
@@ -1942,52 +1913,51 @@ define amdgpu_kernel void @test_call_external_v32i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2i16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2i16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2i16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2i16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](<2 x s16>), [[DEF]](p1) :: (volatile store (<2 x s16>) into `<2 x i16> addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY19]](<2 x s16>), [[DEF]](p1) :: (volatile store (<2 x s16>) into `<2 x i16> addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call <2 x i16> @external_v2i16_func_void()
store volatile <2 x i16> %val, <2 x i16> addrspace(1)* undef
@@ -1997,53 +1967,52 @@ define amdgpu_kernel void @test_call_external_v2i16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v3i16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v3i16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v3i16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v3i16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
; GCN: [[DEF1:%[0-9]+]]:_(<2 x s16>) = G_IMPLICIT_DEF
- ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<6 x s16>) = G_CONCAT_VECTORS [[COPY21]](<2 x s16>), [[COPY22]](<2 x s16>), [[DEF1]](<2 x s16>)
+ ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<6 x s16>) = G_CONCAT_VECTORS [[COPY19]](<2 x s16>), [[COPY20]](<2 x s16>), [[DEF1]](<2 x s16>)
; GCN: [[UV:%[0-9]+]]:_(<3 x s16>), [[UV1:%[0-9]+]]:_(<3 x s16>) = G_UNMERGE_VALUES [[CONCAT_VECTORS]](<6 x s16>)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[UV]](<3 x s16>), [[DEF]](p1) :: (volatile store (<3 x s16>) into `<3 x i16> addrspace(1)* undef`, align 8, addrspace 1)
@@ -2056,52 +2025,51 @@ define amdgpu_kernel void @test_call_external_v3i16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v4i16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v4i16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v4i16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v4i16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
- ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<4 x s16>) = G_CONCAT_VECTORS [[COPY21]](<2 x s16>), [[COPY22]](<2 x s16>)
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
+ ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<4 x s16>) = G_CONCAT_VECTORS [[COPY19]](<2 x s16>), [[COPY20]](<2 x s16>)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[CONCAT_VECTORS]](<4 x s16>), [[DEF]](p1) :: (volatile store (<4 x s16>) into `<4 x i16> addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -2113,52 +2081,51 @@ define amdgpu_kernel void @test_call_external_v4i16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v2f16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v2f16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v2f16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v2f16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](<2 x s16>), [[DEF]](p1) :: (volatile store (<2 x s16>) into `<2 x half> addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY19]](<2 x s16>), [[DEF]](p1) :: (volatile store (<2 x s16>) into `<2 x half> addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call <2 x half> @external_v2f16_func_void()
store volatile <2 x half> %val, <2 x half> addrspace(1)* undef
@@ -2168,53 +2135,52 @@ define amdgpu_kernel void @test_call_external_v2f16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v3f16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v3f16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v3f16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v3f16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
; GCN: [[DEF1:%[0-9]+]]:_(<2 x s16>) = G_IMPLICIT_DEF
- ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<6 x s16>) = G_CONCAT_VECTORS [[COPY21]](<2 x s16>), [[COPY22]](<2 x s16>), [[DEF1]](<2 x s16>)
+ ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<6 x s16>) = G_CONCAT_VECTORS [[COPY19]](<2 x s16>), [[COPY20]](<2 x s16>), [[DEF1]](<2 x s16>)
; GCN: [[UV:%[0-9]+]]:_(<3 x s16>), [[UV1:%[0-9]+]]:_(<3 x s16>) = G_UNMERGE_VALUES [[CONCAT_VECTORS]](<6 x s16>)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[UV]](<3 x s16>), [[DEF]](p1) :: (volatile store (<3 x s16>) into `<3 x half> addrspace(1)* undef`, align 8, addrspace 1)
@@ -2227,52 +2193,51 @@ define amdgpu_kernel void @test_call_external_v3f16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v4f16_func_void() #0 {
; GCN-LABEL: name: test_call_external_v4f16_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v4f16_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v4f16_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
- ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<4 x s16>) = G_CONCAT_VECTORS [[COPY21]](<2 x s16>), [[COPY22]](<2 x s16>)
+ ; GCN: [[COPY19:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(<2 x s16>) = COPY $vgpr1
+ ; GCN: [[CONCAT_VECTORS:%[0-9]+]]:_(<4 x s16>) = G_CONCAT_VECTORS [[COPY19]](<2 x s16>), [[COPY20]](<2 x s16>)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[CONCAT_VECTORS]](<4 x s16>), [[DEF]](p1) :: (volatile store (<4 x s16>) into `<4 x half> addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
@@ -2284,53 +2249,52 @@ define amdgpu_kernel void @test_call_external_v4f16_func_void() #0 {
define amdgpu_kernel void @test_call_external_v3f32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v3f32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v3f32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v3f32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<3 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<3 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<3 x s32>), [[DEF]](p1) :: (volatile store (<3 x s32>) into `<3 x float> addrspace(1)* undef`, align 16, addrspace 1)
; GCN: S_ENDPGM 0
@@ -2342,55 +2306,54 @@ define amdgpu_kernel void @test_call_external_v3f32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v5f32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v5f32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v5f32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v5f32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<5 x s32>) = G_BUILD_VECTOR [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32), [[COPY24]](s32), [[COPY25]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[BUILD_VECTOR:%[0-9]+]]:_(<5 x s32>) = G_BUILD_VECTOR [[COPY19]](s32), [[COPY20]](s32), [[COPY21]](s32), [[COPY22]](s32), [[COPY23]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[BUILD_VECTOR]](<5 x s32>), [[DEF]](p1) :: (volatile store (<5 x s32>) into `<5 x float> addrspace(1)* undef`, align 32, addrspace 1)
; GCN: S_ENDPGM 0
@@ -2403,57 +2366,56 @@ define amdgpu_kernel void @test_call_external_v5f32_func_void() #0 {
define amdgpu_kernel void @test_call_external_i32_i64_func_void() #0 {
; GCN-LABEL: name: test_call_external_i32_i64_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
- ; GCN: [[COPY10:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i32_i64_func_void
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY13:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY13]], [[C]](s64)
- ; GCN: [[COPY14:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY18]], [[SHL]]
- ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY20]], [[C2]](s32)
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY12:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY16]], [[SHL]]
+ ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY21:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY21]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY11]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY12]](p4)
+ ; GCN: [[COPY19:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY19]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY14]](s64)
- ; GCN: $sgpr12 = COPY [[COPY15]](s32)
- ; GCN: $sgpr13 = COPY [[COPY16]](s32)
- ; GCN: $sgpr14 = COPY [[COPY17]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY12]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY14]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY15]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i32_i64_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY23]](s32), [[COPY24]](s32)
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[MV:%[0-9]+]]:_(s64) = G_MERGE_VALUES [[COPY21]](s32), [[COPY22]](s32)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY22]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
- ; GCN: G_STORE [[MV]](s64), [[COPY10]](p1) :: (volatile store (s64) into `i64 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY20]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[MV]](s64), [[COPY9]](p1) :: (volatile store (s64) into `i64 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call { i32, i64 } @external_i32_i64_func_void()
%val.0 = extractvalue { i32, i64 } %val, 0
@@ -2495,54 +2457,53 @@ define amdgpu_gfx void @test_gfx_call_external_i32_i64_func_void() #0 {
define amdgpu_kernel void @test_call_external_a2i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_a2i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_a2i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_a2i32_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
- ; GCN: G_STORE [[COPY21]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
- ; GCN: G_STORE [[COPY22]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY19]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[COPY20]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call [2 x i32] @external_a2i32_func_void()
%val.0 = extractvalue [2 x i32] %val, 0
@@ -2555,63 +2516,62 @@ define amdgpu_kernel void @test_call_external_a2i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_a5i8_func_void() #0 {
; GCN-LABEL: name: test_call_external_a5i8_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_a5i8_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_a5i8_func_void, csr_amdgpu_highregs, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31, implicit-def $vgpr0, implicit-def $vgpr1, implicit-def $vgpr2, implicit-def $vgpr3, implicit-def $vgpr4
- ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr0
- ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
+ ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY $vgpr0
+ ; GCN: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[COPY19]](s32)
; GCN: [[TRUNC1:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC]](s16)
- ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr1
- ; GCN: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY22]](s32)
+ ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY $vgpr1
+ ; GCN: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY20]](s32)
; GCN: [[TRUNC3:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC2]](s16)
- ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr2
- ; GCN: [[TRUNC4:%[0-9]+]]:_(s16) = G_TRUNC [[COPY23]](s32)
+ ; GCN: [[COPY21:%[0-9]+]]:_(s32) = COPY $vgpr2
+ ; GCN: [[TRUNC4:%[0-9]+]]:_(s16) = G_TRUNC [[COPY21]](s32)
; GCN: [[TRUNC5:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC4]](s16)
- ; GCN: [[COPY24:%[0-9]+]]:_(s32) = COPY $vgpr3
- ; GCN: [[TRUNC6:%[0-9]+]]:_(s16) = G_TRUNC [[COPY24]](s32)
+ ; GCN: [[COPY22:%[0-9]+]]:_(s32) = COPY $vgpr3
+ ; GCN: [[TRUNC6:%[0-9]+]]:_(s16) = G_TRUNC [[COPY22]](s32)
; GCN: [[TRUNC7:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC6]](s16)
- ; GCN: [[COPY25:%[0-9]+]]:_(s32) = COPY $vgpr4
- ; GCN: [[TRUNC8:%[0-9]+]]:_(s16) = G_TRUNC [[COPY25]](s32)
+ ; GCN: [[COPY23:%[0-9]+]]:_(s32) = COPY $vgpr4
+ ; GCN: [[TRUNC8:%[0-9]+]]:_(s16) = G_TRUNC [[COPY23]](s32)
; GCN: [[TRUNC9:%[0-9]+]]:_(s8) = G_TRUNC [[TRUNC8]](s16)
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: G_STORE [[TRUNC1]](s8), [[DEF]](p1) :: (volatile store (s8) into `i8 addrspace(1)* undef`, addrspace 1)
@@ -2637,59 +2597,58 @@ define amdgpu_kernel void @test_call_external_a5i8_func_void() #0 {
define amdgpu_kernel void @test_call_external_v32i32_i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v32i32_i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
- ; GCN: [[COPY10:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
; GCN: [[FRAME_INDEX:%[0-9]+]]:_(p5) = G_FRAME_INDEX %stack.0
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v32i32_i32_func_void
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY13:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY13]], [[C]](s64)
- ; GCN: [[COPY14:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY18]], [[SHL]]
- ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY20]], [[C2]](s32)
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY12:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY16]], [[SHL]]
+ ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
; GCN: $vgpr0 = COPY [[FRAME_INDEX]](p5)
- ; GCN: [[COPY21:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY21]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY11]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY12]](p4)
+ ; GCN: [[COPY19:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY19]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY14]](s64)
- ; GCN: $sgpr12 = COPY [[COPY15]](s32)
- ; GCN: $sgpr13 = COPY [[COPY16]](s32)
- ; GCN: $sgpr14 = COPY [[COPY17]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY12]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY14]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY15]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v32i32_i32_func_void, csr_amdgpu_highregs, implicit $vgpr0, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[LOAD:%[0-9]+]]:_(<32 x s32>) = G_LOAD [[FRAME_INDEX]](p5) :: (load (<32 x s32>) from %stack.0, addrspace 5)
- ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 128
- ; GCN: [[PTR_ADD1:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX]], [[C3]](s32)
+ ; GCN: [[C4:%[0-9]+]]:_(s32) = G_CONSTANT i32 128
+ ; GCN: [[PTR_ADD1:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX]], [[C4]](s32)
; GCN: [[LOAD1:%[0-9]+]]:_(s32) = G_LOAD [[PTR_ADD1]](p5) :: (load (s32) from %stack.0, align 128, addrspace 5)
; GCN: G_STORE [[LOAD]](<32 x s32>), [[DEF]](p1) :: (volatile store (<32 x s32>) into `<32 x i32> addrspace(1)* undef`, align 8, addrspace 1)
- ; GCN: G_STORE [[LOAD1]](s32), [[COPY10]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
+ ; GCN: G_STORE [[LOAD1]](s32), [[COPY9]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
; GCN: S_ENDPGM 0
%val = call { <32 x i32>, i32 } @external_v32i32_i32_func_void()
%val0 = extractvalue { <32 x i32>, i32 } %val, 0
@@ -2702,59 +2661,58 @@ define amdgpu_kernel void @test_call_external_v32i32_i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_i32_v32i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_i32_v32i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
- ; GCN: [[COPY10:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p1) = COPY [[DEF]](p1)
; GCN: [[FRAME_INDEX:%[0-9]+]]:_(p5) = G_FRAME_INDEX %stack.0
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_i32_v32i32_func_void
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY13:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY13]], [[C]](s64)
- ; GCN: [[COPY14:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY18]], [[SHL]]
- ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY20]], [[C2]](s32)
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY12:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY16]], [[SHL]]
+ ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
; GCN: $vgpr0 = COPY [[FRAME_INDEX]](p5)
- ; GCN: [[COPY21:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY21]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY11]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY12]](p4)
+ ; GCN: [[COPY19:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY19]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY14]](s64)
- ; GCN: $sgpr12 = COPY [[COPY15]](s32)
- ; GCN: $sgpr13 = COPY [[COPY16]](s32)
- ; GCN: $sgpr14 = COPY [[COPY17]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY12]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY14]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY15]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_i32_v32i32_func_void, csr_amdgpu_highregs, implicit $vgpr0, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
; GCN: [[LOAD:%[0-9]+]]:_(s32) = G_LOAD [[FRAME_INDEX]](p5) :: (load (s32) from %stack.0, align 128, addrspace 5)
- ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 128
- ; GCN: [[PTR_ADD1:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX]], [[C3]](s32)
+ ; GCN: [[C4:%[0-9]+]]:_(s32) = G_CONSTANT i32 128
+ ; GCN: [[PTR_ADD1:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX]], [[C4]](s32)
; GCN: [[LOAD1:%[0-9]+]]:_(<32 x s32>) = G_LOAD [[PTR_ADD1]](p5) :: (load (<32 x s32>) from %stack.0, addrspace 5)
; GCN: G_STORE [[LOAD]](s32), [[DEF]](p1) :: (volatile store (s32) into `i32 addrspace(1)* undef`, addrspace 1)
- ; GCN: G_STORE [[LOAD1]](<32 x s32>), [[COPY10]](p1) :: (volatile store (<32 x s32>) into `<32 x i32> addrspace(1)* undef`, align 8, addrspace 1)
+ ; GCN: G_STORE [[LOAD1]](<32 x s32>), [[COPY9]](p1) :: (volatile store (<32 x s32>) into `<32 x i32> addrspace(1)* undef`, align 8, addrspace 1)
; GCN: S_ENDPGM 0
%val = call { i32, <32 x i32> } @external_i32_v32i32_func_void()
%val0 = extractvalue { i32, <32 x i32> } %val, 0
@@ -2767,49 +2725,48 @@ define amdgpu_kernel void @test_call_external_i32_v32i32_func_void() #0 {
define amdgpu_kernel void @test_call_external_v33i32_func_void() #0 {
; GCN-LABEL: name: test_call_external_v33i32_func_void
; GCN: bb.1 (%ir-block.0):
- ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11
+ ; GCN: liveins: $sgpr12, $sgpr13, $sgpr14, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9
; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2
; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1
; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0
- ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16
- ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15
- ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14
- ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11
+ ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr14
+ ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr13
+ ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr12
+ ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr8_sgpr9
; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7
; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5
- ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9
; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF
; GCN: [[FRAME_INDEX:%[0-9]+]]:_(p5) = G_FRAME_INDEX %stack.0
; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc
; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_v33i32_func_void
- ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY8]]
- ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY7]]
- ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4)
- ; GCN: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
- ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY12]], [[C]](s64)
- ; GCN: [[COPY13:%[0-9]+]]:_(s64) = COPY [[COPY6]]
- ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY5]]
- ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY4]]
- ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY3]]
- ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
- ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
- ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
- ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY18]], [[C1]](s32)
- ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY17]], [[SHL]]
- ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
- ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
- ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C2]](s32)
+ ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY [[COPY8]]
+ ; GCN: [[COPY10:%[0-9]+]]:_(p4) = COPY [[COPY7]]
+ ; GCN: [[C:%[0-9]+]]:_(p4) = G_CONSTANT i64 0
+ ; GCN: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
+ ; GCN: [[PTR_ADD:%[0-9]+]]:_(p4) = G_PTR_ADD [[C]], [[C1]](s64)
+ ; GCN: [[COPY11:%[0-9]+]]:_(s64) = COPY [[COPY6]]
+ ; GCN: [[COPY12:%[0-9]+]]:_(s32) = COPY [[COPY5]]
+ ; GCN: [[COPY13:%[0-9]+]]:_(s32) = COPY [[COPY4]]
+ ; GCN: [[COPY14:%[0-9]+]]:_(s32) = COPY [[COPY3]]
+ ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32)
+ ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32)
+ ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 10
+ ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY16]], [[C2]](s32)
+ ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY15]], [[SHL]]
+ ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY]](s32)
+ ; GCN: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 20
+ ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY17]], [[C3]](s32)
; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]]
; GCN: $vgpr0 = COPY [[FRAME_INDEX]](p5)
- ; GCN: [[COPY20:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
- ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY20]](<4 x s32>)
- ; GCN: $sgpr4_sgpr5 = COPY [[COPY10]](p4)
- ; GCN: $sgpr6_sgpr7 = COPY [[COPY11]](p4)
+ ; GCN: [[COPY18:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg
+ ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY18]](<4 x s32>)
+ ; GCN: $sgpr4_sgpr5 = COPY [[COPY9]](p4)
+ ; GCN: $sgpr6_sgpr7 = COPY [[COPY10]](p4)
; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD]](p4)
- ; GCN: $sgpr10_sgpr11 = COPY [[COPY13]](s64)
- ; GCN: $sgpr12 = COPY [[COPY14]](s32)
- ; GCN: $sgpr13 = COPY [[COPY15]](s32)
- ; GCN: $sgpr14 = COPY [[COPY16]](s32)
+ ; GCN: $sgpr10_sgpr11 = COPY [[COPY11]](s64)
+ ; GCN: $sgpr12 = COPY [[COPY12]](s32)
+ ; GCN: $sgpr13 = COPY [[COPY13]](s32)
+ ; GCN: $sgpr14 = COPY [[COPY14]](s32)
; GCN: $vgpr31 = COPY [[OR1]](s32)
; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_v33i32_func_void, csr_amdgpu_highregs, implicit $vgpr0, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31
; GCN: ADJCALLSTACKDOWN 0, 0, implicit-def $scc
diff --git a/llvm/test/CodeGen/AMDGPU/GlobalISel/llvm.amdgcn.kernarg.segment.ptr.ll b/llvm/test/CodeGen/AMDGPU/GlobalISel/llvm.amdgcn.kernarg.segment.ptr.ll
index 111aa0d92e8a..18a4840242bc 100644
--- a/llvm/test/CodeGen/AMDGPU/GlobalISel/llvm.amdgcn.kernarg.segment.ptr.ll
+++ b/llvm/test/CodeGen/AMDGPU/GlobalISel/llvm.amdgcn.kernarg.segment.ptr.ll
@@ -74,13 +74,18 @@ define amdgpu_kernel void @opencl_test_implicit_alignment(i32 addrspace(1)* %out
ret void
}
+; Mesa implies 16-bytes are always allocated, hsa requires the
+; attribute for the additional space.
; ALL-LABEL: {{^}}test_no_kernargs:
-; CO-V2: enable_sgpr_kernarg_segment_ptr = 1
+; HSA: enable_sgpr_kernarg_segment_ptr = 0
; HSA: kernarg_segment_byte_size = 0
+
+; OS-MESA3D: enable_sgpr_kernarg_segment_ptr = 1
; OS-MESA3D: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4
-; HSA: s_load_dword s{{[0-9]+}}, s[4:5]
+; HSA: s_mov_b64 [[OFFSET_NULL:s\[[0-9]+:[0-9]+\]]], 40{{$}}
+; HSA: s_load_dword s{{[0-9]+}}, [[OFFSET_NULL]]
define amdgpu_kernel void @test_no_kernargs() #1 {
%kernarg.segment.ptr = call noalias i8 addrspace(4)* @llvm.amdgcn.kernarg.segment.ptr()
%header.ptr = bitcast i8 addrspace(4)* %kernarg.segment.ptr to i32 addrspace(4)*
@@ -90,7 +95,7 @@ define amdgpu_kernel void @test_no_kernargs() #1 {
ret void
}
-; GCN-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs:
+; ALL-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs:
; HSA: kernarg_segment_byte_size = 48
; OS-MESA3d: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4
@@ -102,7 +107,7 @@ define amdgpu_kernel void @opencl_test_implicit_alignment_no_explicit_kernargs()
ret void
}
-; GCN-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs_round_up:
+; ALL-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs_round_up:
; HSA: kernarg_segment_byte_size = 40
; OS-MESA3D: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4
diff --git a/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa-call.ll b/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa-call.ll
index d5d87289f4b9..d92a0085a0b9 100644
--- a/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa-call.ll
+++ b/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa-call.ll
@@ -542,7 +542,7 @@ define void @indirect_use_group_to_flat_addrspacecast_queue_ptr_gfx9() #1 {
define void @use_kernarg_segment_ptr() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_kernarg_segment_ptr
-; AKF_HSA-SAME: () #[[ATTR14:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR11]] {
; AKF_HSA-NEXT: [[KERNARG_SEGMENT_PTR:%.*]] = call i8 addrspace(4)* @llvm.amdgcn.kernarg.segment.ptr()
; AKF_HSA-NEXT: store volatile i8 addrspace(4)* [[KERNARG_SEGMENT_PTR]], i8 addrspace(4)* addrspace(1)* undef, align 8
; AKF_HSA-NEXT: ret void
@@ -573,7 +573,7 @@ define void @func_indirect_use_kernarg_segment_ptr() #1 {
define amdgpu_kernel void @kern_use_implicitarg_ptr() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@kern_use_implicitarg_ptr
-; AKF_HSA-SAME: () #[[ATTR15:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR14:[0-9]+]] {
; AKF_HSA-NEXT: [[IMPLICITARG_PTR:%.*]] = call i8 addrspace(4)* @llvm.amdgcn.implicitarg.ptr()
; AKF_HSA-NEXT: store volatile i8 addrspace(4)* [[IMPLICITARG_PTR]], i8 addrspace(4)* addrspace(1)* undef, align 8
; AKF_HSA-NEXT: ret void
@@ -590,7 +590,7 @@ define amdgpu_kernel void @kern_use_implicitarg_ptr() #1 {
define void @use_implicitarg_ptr() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_implicitarg_ptr
-; AKF_HSA-SAME: () #[[ATTR16:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR15:[0-9]+]] {
; AKF_HSA-NEXT: [[IMPLICITARG_PTR:%.*]] = call i8 addrspace(4)* @llvm.amdgcn.implicitarg.ptr()
; AKF_HSA-NEXT: store volatile i8 addrspace(4)* [[IMPLICITARG_PTR]], i8 addrspace(4)* addrspace(1)* undef, align 8
; AKF_HSA-NEXT: ret void
@@ -607,7 +607,7 @@ define void @use_implicitarg_ptr() #1 {
define void @func_indirect_use_implicitarg_ptr() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_indirect_use_implicitarg_ptr
-; AKF_HSA-SAME: () #[[ATTR16]] {
+; AKF_HSA-SAME: () #[[ATTR15]] {
; AKF_HSA-NEXT: call void @use_implicitarg_ptr()
; AKF_HSA-NEXT: ret void
;
@@ -625,7 +625,7 @@ declare void @external.func() #3
; This function gets deleted.
define internal void @defined.func() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@defined.func
-; AKF_HSA-SAME: () #[[ATTR17:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR16:[0-9]+]] {
; AKF_HSA-NEXT: ret void
;
ret void
@@ -633,7 +633,7 @@ define internal void @defined.func() #3 {
define void @func_call_external() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_call_external
-; AKF_HSA-SAME: () #[[ATTR17]] {
+; AKF_HSA-SAME: () #[[ATTR16]] {
; AKF_HSA-NEXT: call void @external.func()
; AKF_HSA-NEXT: ret void
;
@@ -648,7 +648,7 @@ define void @func_call_external() #3 {
define void @func_call_defined() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_call_defined
-; AKF_HSA-SAME: () #[[ATTR17]] {
+; AKF_HSA-SAME: () #[[ATTR16]] {
; AKF_HSA-NEXT: call void @defined.func()
; AKF_HSA-NEXT: ret void
;
@@ -661,8 +661,8 @@ define void @func_call_defined() #3 {
}
define void @func_call_asm() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_call_asm
-; AKF_HSA-SAME: () #[[ATTR18:[0-9]+]] {
-; AKF_HSA-NEXT: call void asm sideeffect "", ""() #[[ATTR18]]
+; AKF_HSA-SAME: () #[[ATTR17:[0-9]+]] {
+; AKF_HSA-NEXT: call void asm sideeffect "", ""() #[[ATTR17]]
; AKF_HSA-NEXT: ret void
;
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@func_call_asm
@@ -676,7 +676,7 @@ define void @func_call_asm() #3 {
define amdgpu_kernel void @kern_call_external() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@kern_call_external
-; AKF_HSA-SAME: () #[[ATTR19:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR18:[0-9]+]] {
; AKF_HSA-NEXT: call void @external.func()
; AKF_HSA-NEXT: ret void
;
@@ -691,7 +691,7 @@ define amdgpu_kernel void @kern_call_external() #3 {
define amdgpu_kernel void @func_kern_defined() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_kern_defined
-; AKF_HSA-SAME: () #[[ATTR19]] {
+; AKF_HSA-SAME: () #[[ATTR18]] {
; AKF_HSA-NEXT: call void @defined.func()
; AKF_HSA-NEXT: ret void
;
@@ -705,7 +705,7 @@ define amdgpu_kernel void @func_kern_defined() #3 {
define i32 @use_dispatch_ptr_ret_type() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_dispatch_ptr_ret_type
-; AKF_HSA-SAME: () #[[ATTR20:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR19:[0-9]+]] {
; AKF_HSA-NEXT: [[DISPATCH_PTR:%.*]] = call i8 addrspace(4)* @llvm.amdgcn.dispatch.ptr()
; AKF_HSA-NEXT: store volatile i8 addrspace(4)* [[DISPATCH_PTR]], i8 addrspace(4)* addrspace(1)* undef, align 8
; AKF_HSA-NEXT: ret i32 0
@@ -722,7 +722,7 @@ define i32 @use_dispatch_ptr_ret_type() #1 {
define float @func_indirect_use_dispatch_ptr_constexpr_cast_func() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_indirect_use_dispatch_ptr_constexpr_cast_func
-; AKF_HSA-SAME: () #[[ATTR20]] {
+; AKF_HSA-SAME: () #[[ATTR19]] {
; AKF_HSA-NEXT: [[F:%.*]] = call float bitcast (i32 ()* @use_dispatch_ptr_ret_type to float ()*)()
; AKF_HSA-NEXT: [[FADD:%.*]] = fadd float [[F]], 1.000000e+00
; AKF_HSA-NEXT: ret float [[FADD]]
@@ -740,7 +740,7 @@ define float @func_indirect_use_dispatch_ptr_constexpr_cast_func() #1 {
define float @func_indirect_call(float()* %fptr) #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_indirect_call
-; AKF_HSA-SAME: (float ()* [[FPTR:%.*]]) #[[ATTR21:[0-9]+]] {
+; AKF_HSA-SAME: (float ()* [[FPTR:%.*]]) #[[ATTR20:[0-9]+]] {
; AKF_HSA-NEXT: [[F:%.*]] = call float [[FPTR]]()
; AKF_HSA-NEXT: [[FADD:%.*]] = fadd float [[F]], 1.000000e+00
; AKF_HSA-NEXT: ret float [[FADD]]
@@ -759,7 +759,7 @@ define float @func_indirect_call(float()* %fptr) #3 {
declare float @extern() #3
define float @func_extern_call() #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_extern_call
-; AKF_HSA-SAME: () #[[ATTR17]] {
+; AKF_HSA-SAME: () #[[ATTR16]] {
; AKF_HSA-NEXT: [[F:%.*]] = call float @extern()
; AKF_HSA-NEXT: [[FADD:%.*]] = fadd float [[F]], 1.000000e+00
; AKF_HSA-NEXT: ret float [[FADD]]
@@ -777,7 +777,7 @@ define float @func_extern_call() #3 {
define float @func_null_call(float()* %fptr) #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_null_call
-; AKF_HSA-SAME: (float ()* [[FPTR:%.*]]) #[[ATTR21]] {
+; AKF_HSA-SAME: (float ()* [[FPTR:%.*]]) #[[ATTR20]] {
; AKF_HSA-NEXT: [[F:%.*]] = call float null()
; AKF_HSA-NEXT: [[FADD:%.*]] = fadd float [[F]], 1.000000e+00
; AKF_HSA-NEXT: ret float [[FADD]]
@@ -798,7 +798,7 @@ declare float @llvm.amdgcn.rcp.f32(float) #0
; Calls some other recognized intrinsic
define float @func_other_intrinsic_call(float %arg) #3 {
; AKF_HSA-LABEL: define {{[^@]+}}@func_other_intrinsic_call
-; AKF_HSA-SAME: (float [[ARG:%.*]]) #[[ATTR18]] {
+; AKF_HSA-SAME: (float [[ARG:%.*]]) #[[ATTR17]] {
; AKF_HSA-NEXT: [[F:%.*]] = call float @llvm.amdgcn.rcp.f32(float [[ARG]])
; AKF_HSA-NEXT: [[FADD:%.*]] = fadd float [[F]], 1.000000e+00
; AKF_HSA-NEXT: ret float [[FADD]]
@@ -834,14 +834,13 @@ attributes #3 = { nounwind }
; AKF_HSA: attributes #[[ATTR11]] = { nounwind "target-cpu"="fiji" "uniform-work-group-size"="false" }
; AKF_HSA: attributes #[[ATTR12]] = { nounwind "target-cpu"="gfx900" "uniform-work-group-size"="false" }
; AKF_HSA: attributes #[[ATTR13]] = { nounwind "amdgpu-queue-ptr" "target-cpu"="gfx900" "uniform-work-group-size"="false" }
-; AKF_HSA: attributes #[[ATTR14]] = { nounwind "amdgpu-kernarg-segment-ptr" "target-cpu"="fiji" "uniform-work-group-size"="false" }
-; AKF_HSA: attributes #[[ATTR15]] = { nounwind "amdgpu-implicitarg-ptr" "target-cpu"="fiji" }
-; AKF_HSA: attributes #[[ATTR16]] = { nounwind "amdgpu-implicitarg-ptr" "target-cpu"="fiji" "uniform-work-group-size"="false" }
-; AKF_HSA: attributes #[[ATTR17]] = { nounwind "uniform-work-group-size"="false" }
-; AKF_HSA: attributes #[[ATTR18]] = { nounwind }
-; AKF_HSA: attributes #[[ATTR19]] = { nounwind "amdgpu-calls" "uniform-work-group-size"="false" }
-; AKF_HSA: attributes #[[ATTR20]] = { nounwind "amdgpu-dispatch-id" "amdgpu-dispatch-ptr" "amdgpu-implicitarg-ptr" "amdgpu-queue-ptr" "amdgpu-work-group-id-x" "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-x" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" "target-cpu"="fiji" }
-; AKF_HSA: attributes #[[ATTR21]] = { nounwind "amdgpu-dispatch-id" "amdgpu-dispatch-ptr" "amdgpu-implicitarg-ptr" "amdgpu-queue-ptr" "amdgpu-work-group-id-x" "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-x" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" }
+; AKF_HSA: attributes #[[ATTR14]] = { nounwind "amdgpu-implicitarg-ptr" "target-cpu"="fiji" }
+; AKF_HSA: attributes #[[ATTR15]] = { nounwind "amdgpu-implicitarg-ptr" "target-cpu"="fiji" "uniform-work-group-size"="false" }
+; AKF_HSA: attributes #[[ATTR16]] = { nounwind "uniform-work-group-size"="false" }
+; AKF_HSA: attributes #[[ATTR17]] = { nounwind }
+; AKF_HSA: attributes #[[ATTR18]] = { nounwind "amdgpu-calls" "uniform-work-group-size"="false" }
+; AKF_HSA: attributes #[[ATTR19]] = { nounwind "amdgpu-dispatch-id" "amdgpu-dispatch-ptr" "amdgpu-implicitarg-ptr" "amdgpu-queue-ptr" "amdgpu-work-group-id-x" "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-x" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" "target-cpu"="fiji" }
+; AKF_HSA: attributes #[[ATTR20]] = { nounwind "amdgpu-dispatch-id" "amdgpu-dispatch-ptr" "amdgpu-implicitarg-ptr" "amdgpu-queue-ptr" "amdgpu-work-group-id-x" "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-x" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" }
;.
; ATTRIBUTOR_HSA: attributes #[[ATTR0:[0-9]+]] = { nounwind readnone speculatable willreturn }
; ATTRIBUTOR_HSA: attributes #[[ATTR1]] = { nounwind "target-cpu"="fiji" "uniform-work-group-size"="false" }
diff --git a/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa.ll b/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa.ll
index 99fab98422f7..b8cb070a0c15 100644
--- a/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa.ll
+++ b/llvm/test/CodeGen/AMDGPU/annotate-kernel-features-hsa.ll
@@ -293,7 +293,7 @@ define amdgpu_kernel void @use_queue_ptr(i32 addrspace(1)* %ptr) #1 {
define amdgpu_kernel void @use_kernarg_segment_ptr(i32 addrspace(1)* %ptr) #1 {
; HSA-LABEL: define {{[^@]+}}@use_kernarg_segment_ptr
-; HSA-SAME: (i32 addrspace(1)* [[PTR:%.*]]) #[[ATTR12:[0-9]+]] {
+; HSA-SAME: (i32 addrspace(1)* [[PTR:%.*]]) #[[ATTR1]] {
; HSA-NEXT: [[DISPATCH_PTR:%.*]] = call i8 addrspace(4)* @llvm.amdgcn.kernarg.segment.ptr()
; HSA-NEXT: [[BC:%.*]] = bitcast i8 addrspace(4)* [[DISPATCH_PTR]] to i32 addrspace(4)*
; HSA-NEXT: [[VAL:%.*]] = load i32, i32 addrspace(4)* [[BC]], align 4
@@ -442,13 +442,13 @@ define amdgpu_kernel void @use_is_private(i8* %ptr) #1 {
define amdgpu_kernel void @use_alloca() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_alloca
-; AKF_HSA-SAME: () #[[ATTR13:[0-9]+]] {
+; AKF_HSA-SAME: () #[[ATTR12:[0-9]+]] {
; AKF_HSA-NEXT: [[ALLOCA:%.*]] = alloca i32, align 4, addrspace(5)
; AKF_HSA-NEXT: store i32 0, i32 addrspace(5)* [[ALLOCA]], align 4
; AKF_HSA-NEXT: ret void
;
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_alloca
-; ATTRIBUTOR_HSA-SAME: () #[[ATTR13:[0-9]+]] {
+; ATTRIBUTOR_HSA-SAME: () #[[ATTR12:[0-9]+]] {
; ATTRIBUTOR_HSA-NEXT: [[ALLOCA:%.*]] = alloca i32, align 4, addrspace(5)
; ATTRIBUTOR_HSA-NEXT: ret void
;
@@ -459,7 +459,7 @@ define amdgpu_kernel void @use_alloca() #1 {
define amdgpu_kernel void @use_alloca_non_entry_block() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_alloca_non_entry_block
-; AKF_HSA-SAME: () #[[ATTR13]] {
+; AKF_HSA-SAME: () #[[ATTR12]] {
; AKF_HSA-NEXT: entry:
; AKF_HSA-NEXT: br label [[BB:%.*]]
; AKF_HSA: bb:
@@ -468,7 +468,7 @@ define amdgpu_kernel void @use_alloca_non_entry_block() #1 {
; AKF_HSA-NEXT: ret void
;
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_alloca_non_entry_block
-; ATTRIBUTOR_HSA-SAME: () #[[ATTR13]] {
+; ATTRIBUTOR_HSA-SAME: () #[[ATTR12]] {
; ATTRIBUTOR_HSA-NEXT: entry:
; ATTRIBUTOR_HSA-NEXT: br label [[BB:%.*]]
; ATTRIBUTOR_HSA: bb:
@@ -486,13 +486,13 @@ bb:
define void @use_alloca_func() #1 {
; AKF_HSA-LABEL: define {{[^@]+}}@use_alloca_func
-; AKF_HSA-SAME: () #[[ATTR13]] {
+; AKF_HSA-SAME: () #[[ATTR12]] {
; AKF_HSA-NEXT: [[ALLOCA:%.*]] = alloca i32, align 4, addrspace(5)
; AKF_HSA-NEXT: store i32 0, i32 addrspace(5)* [[ALLOCA]], align 4
; AKF_HSA-NEXT: ret void
;
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_alloca_func
-; ATTRIBUTOR_HSA-SAME: () #[[ATTR13]] {
+; ATTRIBUTOR_HSA-SAME: () #[[ATTR12]] {
; ATTRIBUTOR_HSA-NEXT: [[ALLOCA:%.*]] = alloca i32, align 4, addrspace(5)
; ATTRIBUTOR_HSA-NEXT: ret void
;
@@ -517,8 +517,7 @@ attributes #1 = { nounwind }
; AKF_HSA: attributes #[[ATTR9]] = { nounwind "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" }
; AKF_HSA: attributes #[[ATTR10]] = { nounwind "amdgpu-dispatch-ptr" }
; AKF_HSA: attributes #[[ATTR11]] = { nounwind "amdgpu-queue-ptr" }
-; AKF_HSA: attributes #[[ATTR12]] = { nounwind "amdgpu-kernarg-segment-ptr" }
-; AKF_HSA: attributes #[[ATTR13]] = { nounwind "amdgpu-stack-objects" }
+; AKF_HSA: attributes #[[ATTR12]] = { nounwind "amdgpu-stack-objects" }
;.
; ATTRIBUTOR_HSA: attributes #[[ATTR0:[0-9]+]] = { nounwind readnone speculatable willreturn }
; ATTRIBUTOR_HSA: attributes #[[ATTR1]] = { nounwind "uniform-work-group-size"="false" }
@@ -532,6 +531,5 @@ attributes #1 = { nounwind }
; ATTRIBUTOR_HSA: attributes #[[ATTR9]] = { nounwind "amdgpu-work-group-id-y" "amdgpu-work-group-id-z" "amdgpu-work-item-id-y" "amdgpu-work-item-id-z" "uniform-work-group-size"="false" }
; ATTRIBUTOR_HSA: attributes #[[ATTR10]] = { nounwind "amdgpu-dispatch-ptr" "uniform-work-group-size"="false" }
; ATTRIBUTOR_HSA: attributes #[[ATTR11]] = { nounwind "amdgpu-queue-ptr" "uniform-work-group-size"="false" }
-; ATTRIBUTOR_HSA: attributes #[[ATTR12]] = { nounwind "amdgpu-kernarg-segment-ptr" "uniform-work-group-size"="false" }
-; ATTRIBUTOR_HSA: attributes #[[ATTR13]] = { nounwind "amdgpu-stack-objects" "uniform-work-group-size"="false" }
+; ATTRIBUTOR_HSA: attributes #[[ATTR12]] = { nounwind "amdgpu-stack-objects" "uniform-work-group-size"="false" }
;.
diff --git a/llvm/test/CodeGen/AMDGPU/llvm.amdgcn.kernarg.segment.ptr.ll b/llvm/test/CodeGen/AMDGPU/llvm.amdgcn.kernarg.segment.ptr.ll
index 095394caf99a..7ed8159a9d7a 100644
--- a/llvm/test/CodeGen/AMDGPU/llvm.amdgcn.kernarg.segment.ptr.ll
+++ b/llvm/test/CodeGen/AMDGPU/llvm.amdgcn.kernarg.segment.ptr.ll
@@ -75,13 +75,18 @@ define amdgpu_kernel void @opencl_test_implicit_alignment(i32 addrspace(1)* %out
ret void
}
+; Mesa implies 16-bytes are always allocated, hsa requires the
+; attribute for the additional space.
; ALL-LABEL: {{^}}test_no_kernargs:
-; CO-V2: enable_sgpr_kernarg_segment_ptr = 1
+; HSA: enable_sgpr_kernarg_segment_ptr = 0
; HSA: kernarg_segment_byte_size = 0
+
+; OS-MESA3D: enable_sgpr_kernarg_segment_ptr = 1
; OS-MESA3D: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4
-; HSA: s_load_dword s{{[0-9]+}}, s[4:5]
+; HSA: s_mov_b64 [[NULL:s\[[0-9]+:[0-9]+\]]], 0{{$}}
+; HSA: s_load_dword s{{[0-9]+}}, [[NULL]], 0xa{{$}}
define amdgpu_kernel void @test_no_kernargs() #1 {
%kernarg.segment.ptr = call noalias i8 addrspace(4)* @llvm.amdgcn.kernarg.segment.ptr()
%header.ptr = bitcast i8 addrspace(4)* %kernarg.segment.ptr to i32 addrspace(4)*
@@ -91,7 +96,7 @@ define amdgpu_kernel void @test_no_kernargs() #1 {
ret void
}
-; GCN-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs:
+; ALL-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs:
; HSA: kernarg_segment_byte_size = 48
; OS-MESA3d: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4
@@ -103,7 +108,7 @@ define amdgpu_kernel void @opencl_test_implicit_alignment_no_explicit_kernargs()
ret void
}
-; GCN-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs_round_up:
+; ALL-LABEL: {{^}}opencl_test_implicit_alignment_no_explicit_kernargs_round_up:
; HSA: kernarg_segment_byte_size = 40
; OS-MESA3D: kernarg_segment_byte_size = 16
; CO-V2: kernarg_segment_alignment = 4