| ; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck %s |
| |
| ; CHECK-LABEL: {{^}}test_kill_depth_0_imm_pos: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_0_imm_pos() #0 { |
| call void @llvm.amdgcn.kill(i1 true) |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_depth_0_imm_neg: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: s_mov_b64 exec, 0 |
| ; CHECK-NEXT: s_cbranch_execz BB1_2 |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB1_2: |
| ; CHECK-NEXT: exp null off, off, off, off done vm |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_0_imm_neg() #0 { |
| call void @llvm.amdgcn.kill(i1 false) |
| ret void |
| } |
| |
| ; FIXME: Ideally only one would be emitted |
| ; CHECK-LABEL: {{^}}test_kill_depth_0_imm_neg_x2: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: s_mov_b64 exec, 0 |
| ; CHECK-NEXT: s_cbranch_execz BB2_3 |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK-NEXT: s_mov_b64 exec, 0 |
| ; CHECK-NEXT: s_cbranch_execz BB2_3 |
| ; CHECK-NEXT: ; %bb.2: |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB2_3: |
| ; CHECK: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_0_imm_neg_x2() #0 { |
| call void @llvm.amdgcn.kill(i1 false) |
| call void @llvm.amdgcn.kill(i1 false) |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_depth_var: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_cbranch_execz BB3_2 |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB3_2: |
| ; CHECK: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_var(float %x) #0 { |
| %cmp = fcmp olt float %x, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp) |
| ret void |
| } |
| |
| ; FIXME: Ideally only one would be emitted |
| ; CHECK-LABEL: {{^}}test_kill_depth_var_x2_same: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_cbranch_execz BB4_3 |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_cbranch_execz BB4_3 |
| ; CHECK-NEXT: ; %bb.2: |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB4_3: |
| ; CHECK: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_var_x2_same(float %x) #0 { |
| %cmp = fcmp olt float %x, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp) |
| call void @llvm.amdgcn.kill(i1 %cmp) |
| ret void |
| } |
| |
| ; FIXME: Ideally only one early-exit would be emitted |
| ; CHECK-LABEL: {{^}}test_kill_depth_var_x2: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_cbranch_execz BB5_3 |
| ; CHECK-NEXT: ; %bb.1 |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v1 |
| ; CHECK-NEXT: s_cbranch_execz BB5_3 |
| ; CHECK-NEXT: ; %bb.2 |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB5_3: |
| ; CHECK: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_var_x2(float %x, float %y) #0 { |
| %cmp.x = fcmp olt float %x, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.x) |
| %cmp.y = fcmp olt float %y, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.y) |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_depth_var_x2_instructions: |
| ; CHECK-NEXT: ; %bb.0: |
| ; CHECK-NEXT: v_cmpx_gt_f32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_cbranch_execz BB6_3 |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK: v_mov_b32_e64 v7, -1 |
| ; CHECK: v_cmpx_gt_f32_e32 vcc, 0, v7 |
| ; CHECK-NEXT: s_cbranch_execz BB6_3 |
| ; CHECK-NEXT: ; %bb.2: |
| ; CHECK-NEXT: s_endpgm |
| ; CHECK-NEXT: BB6_3: |
| ; CHECK-NEXT: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_depth_var_x2_instructions(float %x) #0 { |
| %cmp.x = fcmp olt float %x, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.x) |
| %y = call float asm sideeffect "v_mov_b32_e64 v7, -1", "={v7}"() |
| %cmp.y = fcmp olt float %y, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.y) |
| ret void |
| } |
| |
| ; FIXME: why does the skip depend on the asm length in the same block? |
| |
| ; CHECK-LABEL: {{^}}test_kill_control_flow: |
| ; CHECK: s_cmp_lg_u32 s{{[0-9]+}}, 0 |
| ; CHECK: s_cbranch_scc1 [[RETURN_BB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK-NEXT: ; %bb.1: |
| ; CHECK: v_mov_b32_e64 v7, -1 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| |
| ; CHECK: v_cmpx_gt_f32_e32 vcc, 0, v7 |
| |
| ; TODO: We could do an early-exit here (the branch above is uniform!) |
| ; CHECK-NOT: exp null |
| |
| ; CHECK: v_mov_b32_e32 v0, 1.0 |
| define amdgpu_ps float @test_kill_control_flow(i32 inreg %arg) #0 { |
| entry: |
| %cmp = icmp eq i32 %arg, 0 |
| br i1 %cmp, label %bb, label %exit |
| |
| bb: |
| %var = call float asm sideeffect " |
| v_mov_b32_e64 v7, -1 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64", "={v7}"() |
| %cmp.var = fcmp olt float %var, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.var) |
| br label %exit |
| |
| exit: |
| ret float 1.0 |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_control_flow_remainder: |
| ; CHECK: s_cmp_lg_u32 s{{[0-9]+}}, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v{{[0-9]+}}, 0 |
| ; CHECK-NEXT: s_cbranch_scc1 [[RETURN_BB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK-NEXT: ; %bb.1: ; %bb |
| ; CHECK: v_mov_b32_e64 v7, -1 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: ;;#ASMEND |
| ; CHECK: v_mov_b32_e64 v8, -1 |
| ; CHECK: ;;#ASMEND |
| ; CHECK: v_cmpx_gt_f32_e32 vcc, 0, v7 |
| |
| ; TODO: We could do an early-exit here (the branch above is uniform!) |
| ; CHECK-NOT: exp null |
| |
| ; CHECK: buffer_store_dword v8 |
| ; CHECK: v_mov_b32_e64 v9, -2 |
| |
| ; CHECK: {{^}}BB{{[0-9]+_[0-9]+}}: |
| ; CHECK: buffer_store_dword v9 |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @test_kill_control_flow_remainder(i32 inreg %arg) #0 { |
| entry: |
| %cmp = icmp eq i32 %arg, 0 |
| br i1 %cmp, label %bb, label %exit |
| |
| bb: |
| %var = call float asm sideeffect " |
| v_mov_b32_e64 v7, -1 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64", "={v7}"() |
| %live.across = call float asm sideeffect "v_mov_b32_e64 v8, -1", "={v8}"() |
| %cmp.var = fcmp olt float %var, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.var) |
| store volatile float %live.across, float addrspace(1)* undef |
| %live.out = call float asm sideeffect "v_mov_b32_e64 v9, -2", "={v9}"() |
| br label %exit |
| |
| exit: |
| %phi = phi float [ 0.0, %entry ], [ %live.out, %bb ] |
| store float %phi, float addrspace(1)* undef |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_control_flow_return: |
| |
| ; CHECK: v_cmp_eq_u32_e64 [[KILL_CC:s\[[0-9]+:[0-9]+\]]], s0, 1 |
| ; CHECK: s_and_b64 exec, exec, s[2:3] |
| ; CHECK-NEXT: s_cbranch_execz [[EXIT_BB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: s_cmp_lg_u32 s{{[0-9]+}}, 0 |
| ; CHECK: s_cbranch_scc0 [[COND_BB:BB[0-9]+_[0-9]+]] |
| ; CHECK: s_branch [[RETURN_BB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: [[COND_BB]]: |
| ; CHECK: v_mov_b32_e64 v7, -1 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_mov_b32_e32 v0, v7 |
| |
| ; CHECK: [[EXIT_BB]]: |
| ; CHECK-NEXT: exp null |
| ; CHECK-NEXT: s_endpgm |
| |
| ; CHECK: [[RETURN_BB]]: |
| define amdgpu_ps float @test_kill_control_flow_return(i32 inreg %arg) #0 { |
| entry: |
| %kill = icmp eq i32 %arg, 1 |
| %cmp = icmp eq i32 %arg, 0 |
| call void @llvm.amdgcn.kill(i1 %kill) |
| br i1 %cmp, label %bb, label %exit |
| |
| bb: |
| %var = call float asm sideeffect " |
| v_mov_b32_e64 v7, -1 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64", "={v7}"() |
| br label %exit |
| |
| exit: |
| %ret = phi float [ %var, %bb ], [ 0.0, %entry ] |
| ret float %ret |
| } |
| |
| ; CHECK-LABEL: {{^}}test_kill_divergent_loop: |
| ; CHECK: v_cmp_eq_u32_e32 vcc, 0, v0 |
| ; CHECK-NEXT: s_and_saveexec_b64 [[SAVEEXEC:s\[[0-9]+:[0-9]+\]]], vcc |
| ; CHECK-NEXT: s_xor_b64 [[SAVEEXEC]], exec, [[SAVEEXEC]] |
| ; CHECK-NEXT: s_cbranch_execz [[EXIT:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: ; %bb.{{[0-9]+}}: ; %bb.preheader |
| ; CHECK: s_mov_b32 |
| |
| ; CHECK: [[LOOP_BB:BB[0-9]+_[0-9]+]]: |
| |
| ; CHECK: v_mov_b32_e64 v7, -1 |
| ; CHECK: v_nop_e64 |
| ; CHECK: v_cmpx_gt_f32_e32 vcc, 0, v7 |
| |
| ; CHECK-NEXT: ; %bb.3: |
| ; CHECK: buffer_load_dword [[LOAD:v[0-9]+]] |
| ; CHECK: v_cmp_eq_u32_e32 vcc, 0, [[LOAD]] |
| ; CHECK-NEXT: s_and_b64 vcc, exec, vcc |
| ; CHECK-NEXT: s_cbranch_vccnz [[LOOP_BB]] |
| |
| ; CHECK-NEXT: {{^}}[[EXIT]]: |
| ; CHECK: s_or_b64 exec, exec, [[SAVEEXEC]] |
| ; CHECK: buffer_store_dword |
| ; CHECK: s_endpgm |
| define amdgpu_ps void @test_kill_divergent_loop(i32 %arg) #0 { |
| entry: |
| %cmp = icmp eq i32 %arg, 0 |
| br i1 %cmp, label %bb, label %exit |
| |
| bb: |
| %var = call float asm sideeffect " |
| v_mov_b32_e64 v7, -1 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64 |
| v_nop_e64", "={v7}"() |
| %cmp.var = fcmp olt float %var, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.var) |
| %vgpr = load volatile i32, i32 addrspace(1)* undef |
| %loop.cond = icmp eq i32 %vgpr, 0 |
| br i1 %loop.cond, label %bb, label %exit |
| |
| exit: |
| store volatile i32 8, i32 addrspace(1)* undef |
| ret void |
| } |
| |
| ; bug 28550 |
| ; CHECK-LABEL: {{^}}phi_use_def_before_kill: |
| ; CHECK: v_cndmask_b32_e64 [[PHIREG:v[0-9]+]], 0, -1.0, |
| ; CHECK: v_cmpx_lt_f32_e32 vcc, 0, |
| ; CHECK-NEXT: s_cbranch_execz [[EXITBB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: ; %[[KILLBB:bb.[0-9]+]]: |
| ; CHECK-NEXT: s_cbranch_scc0 [[PHIBB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: [[PHIBB]]: |
| ; CHECK: v_cmp_eq_f32_e32 vcc, 0, [[PHIREG]] |
| ; CHECK: s_cbranch_vccz [[ENDBB:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: ; %bb10 |
| ; CHECK: v_mov_b32_e32 v{{[0-9]+}}, 9 |
| ; CHECK: buffer_store_dword |
| |
| ; CHECK: [[ENDBB]]: |
| ; CHECK-NEXT: s_endpgm |
| |
| ; CHECK: [[EXITBB]]: |
| ; CHECK: exp null |
| ; CHECK-NEXT: s_endpgm |
| define amdgpu_ps void @phi_use_def_before_kill(float inreg %x) #0 { |
| bb: |
| %tmp = fadd float %x, 1.000000e+00 |
| %tmp1 = fcmp olt float 0.000000e+00, %tmp |
| %tmp2 = select i1 %tmp1, float -1.000000e+00, float 0.000000e+00 |
| %cmp.tmp2 = fcmp olt float %tmp2, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.tmp2) |
| br i1 undef, label %phibb, label %bb8 |
| |
| phibb: |
| %tmp5 = phi float [ %tmp2, %bb ], [ 4.0, %bb8 ] |
| %tmp6 = fcmp oeq float %tmp5, 0.000000e+00 |
| br i1 %tmp6, label %bb10, label %end |
| |
| bb8: |
| store volatile i32 8, i32 addrspace(1)* undef |
| br label %phibb |
| |
| bb10: |
| store volatile i32 9, i32 addrspace(1)* undef |
| br label %end |
| |
| end: |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}no_skip_no_successors: |
| ; CHECK: v_cmp_nge_f32 |
| ; CHECK: s_cbranch_vccz [[SKIPKILL:BB[0-9]+_[0-9]+]] |
| |
| ; CHECK: ; %bb6 |
| ; CHECK: s_mov_b64 exec, 0 |
| |
| ; CHECK: [[SKIPKILL]]: |
| ; CHECK: v_cmp_nge_f32_e32 vcc |
| ; CHECK: %bb.3: ; %bb5 |
| ; CHECK-NEXT: .Lfunc_end{{[0-9]+}} |
| define amdgpu_ps void @no_skip_no_successors(float inreg %arg, float inreg %arg1) #0 { |
| bb: |
| %tmp = fcmp ult float %arg1, 0.000000e+00 |
| %tmp2 = fcmp ult float %arg, 0x3FCF5C2900000000 |
| br i1 %tmp, label %bb6, label %bb3 |
| |
| bb3: ; preds = %bb |
| br i1 %tmp2, label %bb5, label %bb4 |
| |
| bb4: ; preds = %bb3 |
| br i1 true, label %bb5, label %bb7 |
| |
| bb5: ; preds = %bb4, %bb3 |
| unreachable |
| |
| bb6: ; preds = %bb |
| call void @llvm.amdgcn.kill(i1 false) |
| unreachable |
| |
| bb7: ; preds = %bb4 |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}if_after_kill_block: |
| ; CHECK: ; %bb.0: |
| ; CHECK: s_and_saveexec_b64 |
| ; CHECK: s_xor_b64 |
| |
| ; CHECK: v_cmpx_gt_f32_e32 vcc, 0, |
| ; CHECK: BB{{[0-9]+_[0-9]+}}: |
| ; CHECK: s_or_b64 exec, exec |
| ; CHECK: image_sample_c |
| |
| ; CHECK: v_cmp_neq_f32_e32 vcc, 0, |
| ; CHECK: s_and_saveexec_b64 s{{\[[0-9]+:[0-9]+\]}}, vcc |
| ; CHECK-NEXT: s_cbranch_execz [[END:BB[0-9]+_[0-9]+]] |
| ; CHECK-NOT: branch |
| |
| ; CHECK: ; %bb.{{[0-9]+}}: ; %bb8 |
| ; CHECK: buffer_store_dword |
| |
| ; CHECK: [[END]]: |
| ; CHECK: s_endpgm |
| define amdgpu_ps void @if_after_kill_block(float %arg, float %arg1, float %arg2, float %arg3) #0 { |
| bb: |
| %tmp = fcmp ult float %arg1, 0.000000e+00 |
| br i1 %tmp, label %bb3, label %bb4 |
| |
| bb3: ; preds = %bb |
| %cmp.arg = fcmp olt float %arg, 0.0 |
| call void @llvm.amdgcn.kill(i1 %cmp.arg) |
| br label %bb4 |
| |
| bb4: ; preds = %bb3, %bb |
| %tmp5 = call <4 x float> @llvm.amdgcn.image.sample.c.1d.v4f32.f32(i32 16, float %arg2, float %arg3, <8 x i32> undef, <4 x i32> undef, i1 0, i32 0, i32 0) |
| %tmp6 = extractelement <4 x float> %tmp5, i32 0 |
| %tmp7 = fcmp une float %tmp6, 0.000000e+00 |
| br i1 %tmp7, label %bb8, label %bb9 |
| |
| bb8: ; preds = %bb9, %bb4 |
| store volatile i32 9, i32 addrspace(1)* undef |
| ret void |
| |
| bb9: ; preds = %bb4 |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}cbranch_kill: |
| ; CHECK: ; %bb.{{[0-9]+}}: ; %export |
| ; CHECK-NEXT: s_or_b64 |
| ; CHECK-NEXT: s_cbranch_execz [[EXIT:BB[0-9]+_[0-9]+]] |
| ; CHECK: [[EXIT]]: |
| ; CHECK-NEXT: exp null off, off, off, off done vm |
| define amdgpu_ps void @cbranch_kill(i32 inreg %0, <2 x float> %1) { |
| .entry: |
| %val0 = extractelement <2 x float> %1, i32 0 |
| %val1 = extractelement <2 x float> %1, i32 1 |
| %p0 = call float @llvm.amdgcn.interp.p1(float %val0, i32 immarg 0, i32 immarg 1, i32 %0) #2 |
| %sample = call float @llvm.amdgcn.image.sample.l.2darray.f32.f32(i32 1, float %p0, float %p0, float %p0, float 0.000000e+00, <8 x i32> undef, <4 x i32> undef, i1 false, i32 0, i32 0) |
| %cond0 = fcmp ugt float %sample, 0.000000e+00 |
| br i1 %cond0, label %live, label %kill |
| |
| kill: |
| call void @llvm.amdgcn.kill(i1 false) |
| br label %export |
| |
| live: |
| %i0 = call float @llvm.amdgcn.interp.p1(float %val0, i32 immarg 0, i32 immarg 0, i32 %0) #2 |
| %i1 = call float @llvm.amdgcn.interp.p2(float %i0, float %val1, i32 immarg 0, i32 immarg 0, i32 %0) #2 |
| %i2 = call float @llvm.amdgcn.interp.p1(float %val0, i32 immarg 1, i32 immarg 0, i32 %0) #2 |
| %i3 = call float @llvm.amdgcn.interp.p2(float %i2, float %val1, i32 immarg 1, i32 immarg 0, i32 %0) #2 |
| %scale.i0 = fmul reassoc nnan nsz arcp contract float %i0, %sample |
| %scale.i1 = fmul reassoc nnan nsz arcp contract float %i1, %sample |
| %scale.i2 = fmul reassoc nnan nsz arcp contract float %i2, %sample |
| %scale.i3 = fmul reassoc nnan nsz arcp contract float %i3, %sample |
| br label %export |
| |
| export: |
| %proxy.0.0 = phi float [ undef, %kill ], [ %scale.i0, %live ] |
| %proxy.0.1 = phi float [ undef, %kill ], [ %scale.i1, %live ] |
| %proxy.0.2 = phi float [ undef, %kill ], [ %scale.i2, %live ] |
| %proxy.0.3 = phi float [ undef, %kill ], [ %scale.i3, %live ] |
| %out.0 = call <2 x half> @llvm.amdgcn.cvt.pkrtz(float %proxy.0.0, float %proxy.0.1) #2 |
| %out.1 = call <2 x half> @llvm.amdgcn.cvt.pkrtz(float %proxy.0.2, float %proxy.0.3) #2 |
| call void @llvm.amdgcn.exp.compr.v2f16(i32 immarg 0, i32 immarg 15, <2 x half> %out.0, <2 x half> %out.1, i1 immarg true, i1 immarg true) #3 |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}complex_loop: |
| ; CHECK: s_mov_b64 exec, 0 |
| ; CHECK-NOT: exp null |
| define amdgpu_ps void @complex_loop(i32 inreg %cmpa, i32 %cmpb, i32 %cmpc) { |
| .entry: |
| %flaga = icmp sgt i32 %cmpa, 0 |
| br i1 %flaga, label %.lr.ph, label %._crit_edge |
| |
| .lr.ph: |
| br label %hdr |
| |
| hdr: |
| %ctr = phi i32 [ 0, %.lr.ph ], [ %ctr.next, %latch ] |
| %flagb = icmp ugt i32 %ctr, %cmpb |
| br i1 %flagb, label %kill, label %latch |
| |
| kill: |
| call void @llvm.amdgcn.kill(i1 false) |
| br label %latch |
| |
| latch: |
| %ctr.next = add nuw nsw i32 %ctr, 1 |
| %flagc = icmp slt i32 %ctr.next, %cmpc |
| br i1 %flagc, label %hdr, label %._crit_edge |
| |
| ._crit_edge: |
| %tmp = phi i32 [ -1, %.entry ], [ %ctr.next, %latch ] |
| %out = bitcast i32 %tmp to <2 x half> |
| call void @llvm.amdgcn.exp.compr.v2f16(i32 immarg 0, i32 immarg 15, <2 x half> %out, <2 x half> undef, i1 immarg true, i1 immarg true) |
| ret void |
| } |
| |
| ; CHECK-LABEL: {{^}}skip_mode_switch: |
| ; CHECK: s_and_saveexec_b64 |
| ; CHECK-NEXT: s_cbranch_execz |
| ; CHECK: s_setreg_imm32 |
| ; CHECK: s_or_b64 exec, exec |
| define void @skip_mode_switch(i32 %arg) { |
| entry: |
| %cmp = icmp eq i32 %arg, 0 |
| br i1 %cmp, label %bb.0, label %bb.1 |
| |
| bb.0: |
| call void @llvm.amdgcn.s.setreg(i32 2049, i32 3) |
| br label %bb.1 |
| |
| bb.1: |
| ret void |
| } |
| |
| declare float @llvm.amdgcn.interp.p1(float, i32 immarg, i32 immarg, i32) #2 |
| declare float @llvm.amdgcn.interp.p2(float, float, i32 immarg, i32 immarg, i32) #2 |
| declare void @llvm.amdgcn.exp.compr.v2f16(i32 immarg, i32 immarg, <2 x half>, <2 x half>, i1 immarg, i1 immarg) #3 |
| declare <2 x half> @llvm.amdgcn.cvt.pkrtz(float, float) #2 |
| declare float @llvm.amdgcn.image.sample.l.2darray.f32.f32(i32 immarg, float, float, float, float, <8 x i32>, <4 x i32>, i1 immarg, i32 immarg, i32 immarg) #1 |
| declare <4 x float> @llvm.amdgcn.image.sample.c.1d.v4f32.f32(i32, float, float, <8 x i32>, <4 x i32>, i1, i32, i32) #1 |
| declare void @llvm.amdgcn.kill(i1) #0 |
| |
| declare void @llvm.amdgcn.s.setreg(i32 immarg, i32) |
| |
| attributes #0 = { nounwind } |
| attributes #1 = { nounwind readonly } |
| attributes #2 = { nounwind readnone speculatable } |
| attributes #3 = { inaccessiblememonly nounwind writeonly } |