| /* |
| * Copyright © 2020 Valve Corporation |
| * |
| * SPDX-License-Identifier: MIT |
| */ |
| #include "helpers.h" |
| |
| using namespace aco; |
| |
| BEGIN_TEST(optimize.neg) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, s1: %c, s1: %d = p_startpgm |
| if (!setup_cs("v1 v1 s1 s1", (amd_gfx_level)i)) |
| continue; |
| |
| //! v1: %res0 = v_mul_f32 %a, -%b |
| //! p_unit_test 0, %res0 |
| Temp neg_b = fneg(inputs[1]); |
| writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_b)); |
| |
| //~gfx9! v1: %neg_a = v_mul_f32 -1.0, %a |
| //~gfx9! v1: %res1 = v_max_f32 0x123456, %neg_a |
| //~gfx10! v1: %res1 = v_max_f32 0x123456, -%a |
| //! p_unit_test 1, %res1 |
| Temp neg_a = fneg(inputs[0]); |
| writeout(1, bld.vop2(aco_opcode::v_max_f32, bld.def(v1), Operand::c32(0x123456u), neg_a)); |
| |
| //! v1: %res2 = v_mul_f32 %a, %b |
| //! p_unit_test 2, %res2 |
| Temp neg_neg_a = fneg(neg_a); |
| writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_neg_a, inputs[1])); |
| |
| //! v1: %res3 = v_mul_f32 |%a|, %b |
| //! p_unit_test 3, %res3 |
| Temp abs_neg_a = fabs(neg_a); |
| writeout(3, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_a, inputs[1])); |
| |
| //! v1: %res4 = v_mul_f32 -|%a|, %b |
| //! p_unit_test 4, %res4 |
| Temp abs_a = fabs(inputs[0]); |
| Temp neg_abs_a = fneg(abs_a); |
| writeout(4, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_abs_a, inputs[1])); |
| |
| //~gfx9! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 |
| //~gfx10! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 fi |
| //! p_unit_test 5, %res5 |
| writeout(5, |
| bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), neg_a, inputs[1], dpp_row_sl(1))); |
| |
| //! v1: %res6 = v_subrev_f32 %a, %b |
| //! p_unit_test 6, %res6 |
| writeout(6, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), neg_a, inputs[1])); |
| |
| //! v1: %res7 = v_sub_f32 %b, %a |
| //! p_unit_test 7, %res7 |
| writeout(7, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[1], neg_a)); |
| |
| //! v1: %res8 = v_mul_f32 %a, -%c |
| //! p_unit_test 8, %res8 |
| Temp neg_c = fneg(bld.copy(bld.def(v1), inputs[2])); |
| writeout(8, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_c)); |
| |
| // //! v1: %res9 = v_mul_f32 |%neg_a|, %b |
| // //! p_unit_test 9, %res9 |
| Temp abs_neg_abs_a = fabs(neg_abs_a); |
| writeout(9, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_abs_a, inputs[1])); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.output_modifiers) |
| //>> v1: %a, v1: %b = p_startpgm |
| if (!setup_cs("v1 v1", GFX9)) |
| return; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| /* 32-bit modifiers */ |
| |
| //! v1: %res0 = v_add_f32 %a, %b *0.5 |
| //! p_unit_test 0, %res0 |
| Temp tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f000000u), tmp)); |
| |
| //! v1: %res1 = v_add_f32 %a, %b *2 |
| //! p_unit_test 1, %res1 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(1, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| |
| //! v1: %res2 = v_add_f32 %a, %b *4 |
| //! p_unit_test 2, %res2 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40800000u), tmp)); |
| |
| //! v1: %res3 = v_add_f32 %a, %b clamp |
| //! p_unit_test 3, %res3 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(3, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), |
| Operand::c32(0x3f800000u), tmp)); |
| |
| //! v1: %res4 = v_add_f32 %a, %b *2 clamp |
| //! p_unit_test 4, %res4 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp); |
| writeout(4, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), |
| Operand::c32(0x3f800000u), tmp)); |
| |
| /* 16-bit modifiers */ |
| |
| //! v2b: %res5 = v_add_f16 %a, %b *0.5 |
| //! p_unit_test 5, %res5 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(5, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x3800u), tmp)); |
| |
| //! v2b: %res6 = v_add_f16 %a, %b *2 |
| //! p_unit_test 6, %res6 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(6, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp)); |
| |
| //! v2b: %res7 = v_add_f16 %a, %b *4 |
| //! p_unit_test 7, %res7 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(7, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4400u), tmp)); |
| |
| //! v2b: %res8 = v_add_f16 %a, %b clamp |
| //! p_unit_test 8, %res8 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(8, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u), |
| Operand::c16(0x3c00u), tmp)); |
| |
| //! v2b: %res9 = v_add_f16 %a, %b *2 clamp |
| //! p_unit_test 9, %res9 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000), tmp); |
| writeout(9, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u), |
| Operand::c16(0x3c00u), tmp)); |
| |
| /* clamping is done after omod */ |
| |
| //! v1: %res10_tmp = v_add_f32 %a, %b clamp |
| //! v1: %res10 = v_mul_f32 2.0, %res10_tmp |
| //! p_unit_test 10, %res10 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| tmp = bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), Operand::c32(0x3f800000u), |
| tmp); |
| writeout(10, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| |
| /* unsupported instructions */ |
| |
| //! v1: %res11_tmp = v_xor_b32 %a, %b |
| //! v1: %res11 = v_mul_f32 2.0, %res11_tmp |
| //! p_unit_test 11, %res11 |
| tmp = bld.vop2(aco_opcode::v_xor_b32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(11, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| |
| /* several users */ |
| |
| //! v1: %res12_tmp = v_add_f32 %a, %b |
| //! p_unit_test %res12_tmp |
| //! v1: %res12 = v_mul_f32 2.0, %res12_tmp |
| //! p_unit_test 12, %res12 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| bld.pseudo(aco_opcode::p_unit_test, tmp); |
| writeout(12, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| |
| //! v1: %res13 = v_add_f32 %a, %b |
| //! p_unit_test 13, %res13 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp); |
| writeout(13, tmp); |
| |
| /* omod has no effect if denormals are enabled but clamp is fine */ |
| |
| //>> BB1 |
| //! /* logical preds: BB0, / linear preds: BB0, / kind: */ |
| program->next_fp_mode.denorm32 = fp_denorm_keep; |
| program->next_fp_mode.denorm16_64 = fp_denorm_flush; |
| bld.reset(program->create_and_insert_block()); |
| program->blocks[0].linear_succs.push_back(1); |
| program->blocks[0].logical_succs.push_back(1); |
| program->blocks[1].linear_preds.push_back(0); |
| program->blocks[1].logical_preds.push_back(0); |
| |
| //! v1: %res14_tmp = v_add_f32 %a, %b |
| //! v1: %res14 = v_mul_f32 2.0, %res13_tmp |
| //! p_unit_test 14, %res14 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(14, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| |
| //! v1: %res15 = v_add_f32 %a, %b clamp |
| //! p_unit_test 15, %res15 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(15, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), |
| Operand::c32(0x3f800000u), tmp)); |
| |
| //>> BB2 |
| //! /* logical preds: BB1, / linear preds: BB1, / kind: */ |
| program->next_fp_mode.denorm32 = fp_denorm_flush; |
| program->next_fp_mode.denorm16_64 = fp_denorm_keep; |
| bld.reset(program->create_and_insert_block()); |
| program->blocks[1].linear_succs.push_back(2); |
| program->blocks[1].logical_succs.push_back(2); |
| program->blocks[2].linear_preds.push_back(1); |
| program->blocks[2].logical_preds.push_back(1); |
| |
| //! v2b: %res16_tmp = v_add_f16 %a, %b |
| //! v2b: %res16 = v_mul_f16 2.0, %res15_tmp |
| //! p_unit_test 16, %res16 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(16, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp)); |
| |
| //! v2b: %res17 = v_add_f16 %a, %b clamp |
| //! p_unit_test 17, %res17 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(17, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u), |
| Operand::c16(0x3c00u), tmp)); |
| |
| /* omod flushes -0.0 to +0.0 */ |
| |
| //>> BB3 |
| //! /* logical preds: BB2, / linear preds: BB2, / kind: uniform, */ |
| program->next_fp_mode.denorm32 = fp_denorm_flush; |
| program->next_fp_mode.denorm16_64 = fp_denorm_flush; |
| bld.reset(program->create_and_insert_block()); |
| bld.is_sz_preserve = true; |
| program->blocks[2].linear_succs.push_back(3); |
| program->blocks[2].logical_succs.push_back(3); |
| program->blocks[3].linear_preds.push_back(2); |
| program->blocks[3].logical_preds.push_back(2); |
| |
| //! v1: (SzPreserve)%res18_tmp = v_add_f32 %a, %b |
| //! v1: (SzPreserve)%res18 = v_mul_f32 2.0, %res18_tmp |
| //! p_unit_test 18, %res18 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(18, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp)); |
| //! v1: (SzPreserve)%res19 = v_add_f32 %a, %b clamp |
| //! p_unit_test 19, %res19 |
| tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| writeout(19, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), |
| Operand::c32(0x3f800000u), tmp)); |
| |
| //! v2b: (SzPreserve)%res20_tmp = v_add_f16 %a, %b |
| //! v2b: (SzPreserve)%res20 = v_mul_f16 2.0, %res20_tmp |
| //! p_unit_test 20, %res20 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(20, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp)); |
| //! v2b: (SzPreserve)%res21 = v_add_f16 %a, %b clamp |
| //! p_unit_test 21, %res21 |
| tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]); |
| writeout(21, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u), |
| Operand::c16(0x3c00u), tmp)); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.add_lshl) |
| for (unsigned i = GFX8; i <= GFX10; i++) { |
| //>> s1: %a, v1: %b = p_startpgm |
| if (!setup_cs("s1 v1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp shift; |
| |
| //~gfx8! s1: %lshl0, s1: %_:scc = s_lshl_b32 %a, 3 |
| //~gfx8! s1: %res0, s1: %_:scc = s_add_u32 %lshl0, 4 |
| //~gfx(9|10)! s1: %res0, s1: %_:scc = s_lshl3_add_u32 %a, 4 |
| //! p_unit_test 0, %res0 |
| shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]), |
| Operand::c32(3u)); |
| writeout(0, bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift, |
| Operand::c32(4u))); |
| |
| //~gfx8! s1: %lshl1, s1: %_:scc = s_lshl_b32 %a, 3 |
| //~gfx8! s1: %add1, s1: %_:scc = s_add_u32 %lshl1, 4 |
| //~gfx8! v1: %add_co1, s2: %_ = v_add_co_u32 %lshl1, %b |
| //~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %add1, %add_co1 |
| //~gfx(9|10)! s1: %lshl1, s1: %_:scc = s_lshl3_add_u32 %a, 4 |
| //~gfx(9|10)! v1: %lshl_add = v_lshl_add_u32 %a, 3, %b |
| //~gfx(9|10)! v1: %res1 = v_add_u32 %lshl1, %lshl_add |
| //! p_unit_test 1, %res1 |
| shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]), |
| Operand::c32(3u)); |
| Temp sadd = |
| bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift, Operand::c32(4u)); |
| Temp vadd = bld.vadd32(bld.def(v1), shift, Operand(inputs[1])); |
| writeout(1, bld.vadd32(bld.def(v1), sadd, vadd)); |
| |
| //~gfx8! s1: %lshl2, s1: %_:scc = s_lshl_b32 %a, 3 |
| //~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b |
| //~gfx(9|10)! v1: %res2 = v_lshl_add_u32 %a, 3, %b |
| //! p_unit_test 2, %res2 |
| Temp lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), |
| Operand(inputs[0]), Operand::c32(3u)); |
| writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! s1: %lshl3, s1: %_:scc = s_lshl_b32 (is24bit)%a, 7 |
| //~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %lshl3, %b |
| //~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 7, %b |
| //! p_unit_test 3, %res3 |
| Operand a_24bit = Operand(inputs[0]); |
| a_24bit.set24bit(true); |
| lshl = |
| bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(7u)); |
| writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //! s1: %lshl4, s1: %_:scc = s_lshl_b32 (is24bit)%a, 3 |
| //~gfx(8|9)! v1: %res4, s2: %carry = v_add_co_u32 %lshl4, %b |
| //~gfx10! v1: %res4, s2: %carry = v_add_co_u32_e64 %lshl4, %b |
| //! p_unit_test 4, %carry |
| lshl = |
| bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u)); |
| Temp carry = bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]), true).def(1).getTemp(); |
| writeout(4, carry); |
| |
| //~gfx8! s1: %lshl5, s1: %_:scc = s_lshl_b32 (is24bit)%a, (is24bit)%a |
| //~gfx8! v1: %res5, s2: %_ = v_add_co_u32 %lshl5, %b |
| //~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%a, (is24bit)%a, %b |
| //! p_unit_test 5, %res5 |
| lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, a_24bit); |
| writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %res6 = v_mad_u32_u24 (is24bit)%a, 8, %b |
| //~gfx(9|10)! v1: %res6 = v_lshl_add_u32 (is24bit)%a, 3, %b |
| //! p_unit_test 6, %res6 |
| lshl = |
| bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u)); |
| writeout(6, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %res7 = v_mad_u32_u24 (is16bit)%a, 16, %b |
| //~gfx(9|10)! v1: %res7 = v_lshl_add_u32 (is16bit)%a, 4, %b |
| //! p_unit_test 7, %res7 |
| Operand a_16bit = Operand(inputs[0]); |
| a_16bit.set16bit(true); |
| lshl = |
| bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_16bit, Operand::c32(4u)); |
| writeout(7, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.bcnt) |
| for (unsigned i = GFX8; i <= GFX10; i++) { |
| //>> v1: %a, s1: %b = p_startpgm |
| if (!setup_cs("v1 s1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp bcnt; |
| |
| //! v1: %res0 = v_bcnt_u32_b32 %a, %a |
| //! p_unit_test 0, %res0 |
| bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero()); |
| writeout(0, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0]))); |
| |
| //! v1: %res1 = v_bcnt_u32_b32 %a, %b |
| //! p_unit_test 1, %res1 |
| bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero()); |
| writeout(1, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[1]))); |
| |
| //! v1: %res2 = v_bcnt_u32_b32 %a, 42 |
| //! p_unit_test 2, %res2 |
| bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero()); |
| writeout(2, bld.vadd32(bld.def(v1), bcnt, Operand::c32(42u))); |
| |
| //! v1: %bnct3 = v_bcnt_u32_b32 %b, 0 |
| //~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %bcnt3, %a |
| //~gfx(9|10)! v1: %res3 = v_add_u32 %bcnt3, %a |
| //! p_unit_test 3, %res3 |
| bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[1]), Operand::zero()); |
| writeout(3, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0]))); |
| |
| //! v1: %bnct4 = v_bcnt_u32_b32 %a, 0 |
| //~gfx(8|9)! v1: %add4, s2: %carry = v_add_co_u32 %bcnt4, %a |
| //~gfx10! v1: %add4, s2: %carry = v_add_co_u32_e64 %bcnt4, %a |
| //! p_unit_test 4, %carry |
| bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero()); |
| Temp carry = bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0]), true).def(1).getTemp(); |
| writeout(4, carry); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| struct clamp_config { |
| const char* name; |
| aco_opcode min, max, med3; |
| Operand lb, ub; |
| }; |
| |
| static const clamp_config clamp_configs[] = { |
| /* 0.0, 4.0 */ |
| {"_0,4f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32, |
| Operand::zero(), Operand::c32(0x40800000u)}, |
| {"_0,4f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16, |
| Operand::c16(0u), Operand::c16(0x4400)}, |
| /* -1.0, 0.0 */ |
| {"_-1,0f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32, |
| Operand::c32(0xbf800000u), Operand::zero()}, |
| {"_-1,0f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16, |
| Operand::c16(0xBC00), Operand::c16(0u)}, |
| /* 0, 3 */ |
| {"_0,3u32", aco_opcode::v_min_u32, aco_opcode::v_max_u32, aco_opcode::v_med3_u32, |
| Operand::zero(), Operand::c32(3u)}, |
| {"_0,3u16", aco_opcode::v_min_u16, aco_opcode::v_max_u16, aco_opcode::v_med3_u16, |
| Operand::c16(0u), Operand::c16(3u)}, |
| {"_0,3i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32, |
| Operand::zero(), Operand::c32(3u)}, |
| {"_0,3i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16, |
| Operand::c16(0u), Operand::c16(3u)}, |
| /* -5, 0 */ |
| {"_-5,0i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32, |
| Operand::c32(0xfffffffbu), Operand::zero()}, |
| {"_-5,0i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16, |
| Operand::c16(0xfffbu), Operand::c16(0u)}, |
| }; |
| |
| BEGIN_TEST(optimize.clamp) |
| for (clamp_config cfg : clamp_configs) { |
| if (!setup_cs("v1 v1 v1", GFX9, CHIP_UNKNOWN, cfg.name)) |
| continue; |
| |
| //! cfg: @match_func(min max med3 lb ub) |
| fprintf(output, "cfg: %s ", instr_info.name[(int)cfg.min]); |
| fprintf(output, "%s ", instr_info.name[(int)cfg.max]); |
| fprintf(output, "%s ", instr_info.name[(int)cfg.med3]); |
| aco_print_operand(&cfg.lb, output); |
| fprintf(output, " "); |
| aco_print_operand(&cfg.ub, output); |
| fprintf(output, "\n"); |
| |
| //>> v1: %a, v1: %b, v1: %c = p_startpgm |
| |
| //! v1: %res0 = @med3 @ub, @lb, %a |
| //! p_unit_test 0, %res0 |
| writeout(0, bld.vop2(cfg.min, bld.def(v1), cfg.ub, |
| bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]))); |
| |
| //! v1: %res1 = @med3 @lb, @ub, %a |
| //! p_unit_test 1, %res1 |
| writeout(1, bld.vop2(cfg.max, bld.def(v1), cfg.lb, |
| bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]))); |
| |
| /* min constant must be greater than max constant */ |
| //! v1: %res2_tmp = @min @lb, %a |
| //! v1: %res2 = @max @ub, %res2_tmp |
| //! p_unit_test 2, %res2 |
| writeout(2, bld.vop2(cfg.max, bld.def(v1), cfg.ub, |
| bld.vop2(cfg.min, bld.def(v1), cfg.lb, inputs[0]))); |
| |
| //! v1: %res3_tmp = @max @ub, %a |
| //! v1: %res3 = @min @lb, %res3_tmp |
| //! p_unit_test 3, %res3 |
| writeout(3, bld.vop2(cfg.min, bld.def(v1), cfg.lb, |
| bld.vop2(cfg.max, bld.def(v1), cfg.ub, inputs[0]))); |
| |
| /* needs two constants */ |
| |
| //! v1: %res4_tmp = @max @lb, %a |
| //! v1: %res4 = @min %b, %res4_tmp |
| //! p_unit_test 4, %res4 |
| writeout(4, bld.vop2(cfg.min, bld.def(v1), inputs[1], |
| bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]))); |
| |
| //! v1: %res5_tmp = @max %b, %a |
| //! v1: %res5 = @min @ub, %res5_tmp |
| //! p_unit_test 5, %res5 |
| writeout(5, bld.vop2(cfg.min, bld.def(v1), cfg.ub, |
| bld.vop2(cfg.max, bld.def(v1), inputs[1], inputs[0]))); |
| |
| //! v1: %res6_tmp = @max %c, %a |
| //! v1: %res6 = @min %b, %res6_tmp |
| //! p_unit_test 6, %res6 |
| writeout(6, bld.vop2(cfg.min, bld.def(v1), inputs[1], |
| bld.vop2(cfg.max, bld.def(v1), inputs[2], inputs[0]))); |
| |
| /* correct NaN behaviour with precise */ |
| if (cfg.min == aco_opcode::v_min_f16 || cfg.min == aco_opcode::v_min_f32) { |
| //~f(16|32)! v1: %res7 = @med3 @ub, @lb, %a |
| //~f(16|32)! p_unit_test 7, %res7 |
| Builder::Result max = bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]); |
| max.def(0).setPrecise(true); |
| Builder::Result min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, max); |
| max.def(0).setPrecise(true); |
| writeout(7, min); |
| |
| //~f(16|32)! v1: (precise)%res8_tmp = @min @ub, %a |
| //~f(16|32)! v1: %res8 = @max @lb, %res8_tmp |
| //~f(16|32)! p_unit_test 8, %res8 |
| min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]); |
| min.def(0).setPrecise(true); |
| writeout(8, bld.vop2(cfg.max, bld.def(v1), cfg.lb, min)); |
| } |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.add3) |
| //>> v1: %a, v1: %b, v1: %c = p_startpgm |
| if (!setup_cs("v1 v1 v1", GFX9)) |
| return; |
| |
| //! v1: %res0 = v_add3_u32 %a, %b, %c |
| //! p_unit_test 0, %res0 |
| Builder::Result tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]); |
| writeout(0, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp)); |
| |
| //! v1: %tmp1 = v_add_u32 %b, %c clamp |
| //! v1: %res1 = v_add_u32 %a, %tmp1 |
| //! p_unit_test 1, %res1 |
| tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]); |
| tmp->valu().clamp = true; |
| writeout(1, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp)); |
| |
| //! v1: %tmp2 = v_add_u32 %b, %c |
| //! v1: %res2 = v_add_u32 %a, %tmp2 clamp |
| //! p_unit_test 2, %res2 |
| tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]); |
| tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp); |
| tmp->valu().clamp = true; |
| writeout(2, tmp); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.minmax) |
| for (unsigned i = GFX10_3; i <= GFX11; i++) { |
| //>> v1: %a, v1: %b, v1: %c = p_startpgm |
| if (!setup_cs("v1 v1 v1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| |
| //! v1: %res0 = v_min3_f32 %a, %b, %c |
| //! p_unit_test 0, %res0 |
| writeout(0, fmin(c, fmin(a, b))); |
| |
| //! v1: %res1 = v_max3_f32 %a, %b, %c |
| //! p_unit_test 1, %res1 |
| writeout(1, fmax(c, fmax(a, b))); |
| |
| //! v1: %res2 = v_min3_f32 -%a, -%b, %c |
| //! p_unit_test 2, %res2 |
| writeout(2, fmin(c, fneg(fmax(a, b)))); |
| |
| //! v1: %res3 = v_max3_f32 -%a, -%b, %c |
| //! p_unit_test 3, %res3 |
| writeout(3, fmax(c, fneg(fmin(a, b)))); |
| |
| //! v1: %res4 = v_max3_f32 -%a, %b, %c |
| //! p_unit_test 4, %res4 |
| writeout(4, fmax(c, fneg(fmin(a, fneg(b))))); |
| |
| //~gfx10_3! v1: %res5_tmp = v_max_f32 %a, %b |
| //~gfx10_3! v1: %res5 = v_min_f32 %c, %res5_tmp |
| //~gfx11! v1: %res5 = v_maxmin_f32 %a, %b, %c |
| //! p_unit_test 5, %res5 |
| writeout(5, fmin(c, fmax(a, b))); |
| |
| //~gfx10_3! v1: %res6_tmp = v_min_f32 %a, %b |
| //~gfx10_3! v1: %res6 = v_max_f32 %c, %res6_tmp |
| //~gfx11! v1: %res6 = v_minmax_f32 %a, %b, %c |
| //! p_unit_test 6, %res6 |
| writeout(6, fmax(c, fmin(a, b))); |
| |
| //~gfx10_3! v1: %res7_tmp = v_min_f32 %a, %b |
| //~gfx10_3! v1: %res7 = v_min_f32 %c, -%res7_tmp |
| //~gfx11! v1: %res7 = v_maxmin_f32 -%a, -%b, %c |
| //! p_unit_test 7, %res7 |
| writeout(7, fmin(c, fneg(fmin(a, b)))); |
| |
| //~gfx10_3! v1: %res8_tmp = v_max_f32 %a, %b |
| //~gfx10_3! v1: %res8 = v_max_f32 %c, -%res8_tmp |
| //~gfx11! v1: %res8 = v_minmax_f32 -%a, -%b, %c |
| //! p_unit_test 8, %res8 |
| writeout(8, fmax(c, fneg(fmax(a, b)))); |
| |
| //~gfx10_3! v1: %res9_tmp = v_max_f32 %a, -%b |
| //~gfx10_3! v1: %res9 = v_max_f32 %c, -%res9_tmp |
| //~gfx11! v1: %res9 = v_minmax_f32 -%a, %b, %c |
| //! p_unit_test 9, %res9 |
| writeout(9, fmax(c, fneg(fmax(a, fneg(b))))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_32_24) |
| for (unsigned i = GFX8; i <= GFX9; i++) { |
| //>> v1: %a, v1: %b, v1: %c = p_startpgm |
| if (!setup_cs("v1 v1 v1", (amd_gfx_level)i)) |
| continue; |
| |
| //! v1: %res0 = v_mad_u32_u24 %b, %c, %a |
| //! p_unit_test 0, %res0 |
| Temp mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]); |
| writeout(0, bld.vadd32(bld.def(v1), inputs[0], mul)); |
| |
| //! v1: %res1_tmp = v_mul_u32_u24 %b, %c |
| //! v1: %_, s2: %res1 = v_add_co_u32 %a, %res1_tmp |
| //! p_unit_test 1, %res1 |
| mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]); |
| writeout(1, bld.vadd32(bld.def(v1), inputs[0], mul, true).def(1).getTemp()); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.add_lshlrev) |
| for (unsigned i = GFX8; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, s1: %c = p_startpgm |
| if (!setup_cs("v1 v1 s1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp lshl; |
| |
| //~gfx8! v1: %lshl0 = v_lshlrev_b32 3, %a |
| //~gfx8! v1: %res0, s2: %_ = v_add_co_u32 %lshl0, %b |
| //~gfx(9|10)! v1: %res0 = v_lshl_add_u32 %a, 3, %b |
| //! p_unit_test 0, %res0 |
| lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), Operand(inputs[0])); |
| writeout(0, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %lshl1 = v_lshlrev_b32 7, (is24bit)%a |
| //~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %lshl1, %b |
| //~gfx(9|10)! v1: %res1 = v_lshl_add_u32 (is24bit)%a, 7, %b |
| //! p_unit_test 1, %res1 |
| Operand a_24bit = Operand(inputs[0]); |
| a_24bit.set24bit(true); |
| lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(7u), a_24bit); |
| writeout(1, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %lshl2 = v_lshlrev_b32 (is24bit)%a, (is24bit)%b |
| //~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b |
| //~gfx(9|10)! v1: %res2 = v_lshl_add_u32 (is24bit)%b, (is24bit)%a, %b |
| //! p_unit_test 2, %res2 |
| Operand b_24bit = Operand(inputs[1]); |
| b_24bit.set24bit(true); |
| lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), a_24bit, b_24bit); |
| writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %res3 = v_mad_u32_u24 (is24bit)%a, 8, %b |
| //~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 3, %b |
| //! p_unit_test 3, %res3 |
| lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), a_24bit); |
| writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %res4 = v_mad_u32_u24 (is16bit)%a, 16, %b |
| //~gfx(9|10)! v1: %res4 = v_lshl_add_u32 (is16bit)%a, 4, %b |
| //! p_unit_test 4, %res4 |
| Operand a_16bit = Operand(inputs[0]); |
| a_16bit.set16bit(true); |
| lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), a_16bit); |
| writeout(4, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]))); |
| |
| //~gfx8! v1: %res5 = v_mad_u32_u24 (is24bit)%c, 16, %c |
| //~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%c, 4, %c |
| //! p_unit_test 5, %res5 |
| Operand c_24bit = Operand(inputs[2]); |
| c_24bit.set24bit(true); |
| lshl = bld.vop2_e64(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), c_24bit); |
| writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[2]))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| enum denorm_op { |
| denorm_mul1 = 0, |
| denorm_fneg = 1, |
| denorm_fabs = 2, |
| denorm_fnegabs = 3, |
| }; |
| |
| static const char* denorm_op_names[] = { |
| "mul1", |
| "fneg", |
| "fabs", |
| "fnegabs", |
| }; |
| |
| struct denorm_config { |
| bool flush; |
| unsigned op; |
| aco_opcode src; |
| aco_opcode dest; |
| }; |
| |
| static const char* |
| srcdest_op_name(aco_opcode op) |
| { |
| switch (op) { |
| case aco_opcode::v_cndmask_b32: return "cndmask"; |
| case aco_opcode::v_min_f32: return "min"; |
| case aco_opcode::v_rcp_f32: return "rcp"; |
| default: return "none"; |
| } |
| } |
| |
| static Temp |
| emit_denorm_srcdest(aco_opcode op, Temp val) |
| { |
| switch (op) { |
| case aco_opcode::v_cndmask_b32: |
| return bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]); |
| case aco_opcode::v_min_f32: |
| return bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), val); |
| case aco_opcode::v_rcp_f32: return bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), val); |
| default: return val; |
| } |
| } |
| |
| BEGIN_TEST(optimize.denorm_propagation) |
| for (unsigned i = GFX8; i <= GFX9; i++) { |
| std::vector<denorm_config> configs; |
| for (bool flush : {false, true}) { |
| for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs}) |
| configs.push_back({flush, op, aco_opcode::num_opcodes, aco_opcode::num_opcodes}); |
| |
| for (aco_opcode dest : {aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) { |
| for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs}) |
| configs.push_back({flush, op, aco_opcode::num_opcodes, dest}); |
| } |
| |
| for (aco_opcode src : |
| {aco_opcode::v_cndmask_b32, aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) { |
| for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs}) |
| configs.push_back({flush, op, src, aco_opcode::num_opcodes}); |
| } |
| } |
| |
| for (denorm_config cfg : configs) { |
| char subvariant[128]; |
| sprintf(subvariant, "_%s_%s_%s_%s", cfg.flush ? "flush" : "keep", srcdest_op_name(cfg.src), |
| denorm_op_names[(int)cfg.op], srcdest_op_name(cfg.dest)); |
| if (!setup_cs("v1 s2", (amd_gfx_level)i, CHIP_UNKNOWN, subvariant)) |
| continue; |
| |
| bool can_propagate = cfg.src == aco_opcode::v_rcp_f32 || |
| (i >= GFX9 && cfg.src == aco_opcode::v_min_f32) || |
| cfg.dest == aco_opcode::v_rcp_f32 || |
| (i >= GFX9 && cfg.dest == aco_opcode::v_min_f32) || !cfg.flush; |
| |
| fprintf(output, "src, dest, op: %s %s %s\n", srcdest_op_name(cfg.src), |
| srcdest_op_name(cfg.dest), denorm_op_names[(int)cfg.op]); |
| fprintf(output, "can_propagate: %u\n", can_propagate); |
| //! src, dest, op: $src $dest $op |
| //! can_propagate: #can_propagate |
| //>> v1: %a, s2: %b = p_startpgm |
| |
| //; patterns = {'cndmask': 'v1: %{} = v_cndmask_b32 0, {}, %b', |
| //; 'min': 'v1: %{} = v_min_f32 0, {}', |
| //; 'rcp': 'v1: %{} = v_rcp_f32 {}'} |
| //; ops = {'mul1': 'v1: %{} = v_mul_f32 1.0, %{}', |
| //; 'fneg': 'v1: %{} = v_mul_f32 -1.0, %{}', |
| //; 'fabs': 'v1: %{} = v_mul_f32 1.0, |%{}|', |
| //; 'fnegabs': 'v1: %{} = v_mul_f32 -1.0, |%{}|'} |
| //; inline_ops = {'mul1': '%{}', 'fneg': '-%{}', 'fabs': '|%{}|', 'fnegabs': '-|%{}|'} |
| |
| //; name = 'a' |
| //; if src != 'none': |
| //; insert_pattern(patterns[src].format('src_res', '%'+name)) |
| //; name = 'src_res' |
| |
| //; if can_propagate: |
| //; name = inline_ops[op].format(name) |
| //; else: |
| //; insert_pattern(ops[op].format('op_res', name)) |
| //; name = '%op_res' |
| |
| //; if dest != 'none': |
| //; insert_pattern(patterns[dest].format('dest_res', name)) |
| //; name = '%dest_res' |
| |
| //; insert_pattern('v1: %res = v_cndmask_b32 0, {}, %b'.format(name)) |
| //! p_unit_test 0, %res |
| |
| program->blocks[0].fp_mode.denorm32 = cfg.flush ? fp_denorm_flush : fp_denorm_keep; |
| |
| Temp val = emit_denorm_srcdest(cfg.src, inputs[0]); |
| switch (cfg.op) { |
| case denorm_mul1: |
| val = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f800000u), val); |
| break; |
| case denorm_fneg: val = fneg(val); break; |
| case denorm_fabs: val = fabs(val); break; |
| case denorm_fnegabs: val = fneg(fabs(val)); break; |
| } |
| val = emit_denorm_srcdest(cfg.dest, val); |
| writeout( |
| 0, bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1])); |
| |
| finish_opt_test(); |
| } |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimizer.dpp) |
| //>> v1: %a, v1: %b, s2: %c, s1: %d = p_startpgm |
| if (!setup_cs("v1 v1 s2 s1", GFX10_3)) |
| return; |
| |
| Operand a(inputs[0]); |
| Operand b(inputs[1]); |
| Operand c(inputs[2]); |
| Operand d(inputs[3]); |
| |
| /* basic optimization */ |
| //! v1: %res0 = v_add_f32 %a, %b row_mirror bound_ctrl:1 fi |
| //! p_unit_test 0, %res0 |
| Temp tmp0 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp0, b); |
| writeout(0, res0); |
| |
| /* operand swapping */ |
| //! v1: %res1 = v_subrev_f32 %a, %b row_mirror bound_ctrl:1 fi |
| //! p_unit_test 1, %res1 |
| Temp tmp1 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res1 = bld.vop2(aco_opcode::v_sub_f32, bld.def(v1), b, tmp1); |
| writeout(1, res1); |
| |
| //! v1: %tmp2 = v_mov_b32 %a row_mirror bound_ctrl:1 fi |
| //! v1: %res2 = v_sub_f32 %b, %tmp2 row_half_mirror bound_ctrl:1 fi |
| //! p_unit_test 2, %res2 |
| Temp tmp2 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res2 = bld.vop2_dpp(aco_opcode::v_sub_f32, bld.def(v1), b, tmp2, dpp_row_half_mirror); |
| writeout(2, res2); |
| |
| /* modifiers */ |
| //! v1: %res3 = v_max_f32 -%a, %b row_mirror bound_ctrl:1 fi |
| //! p_unit_test 3, %res3 |
| auto tmp3 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| tmp3->dpp16().neg[0] = true; |
| Temp res3 = bld.vop2(aco_opcode::v_max_f32, bld.def(v1), tmp3, b); |
| writeout(3, res3); |
| |
| //! v1: %res4 = v_max_f32 -%a, %b row_mirror bound_ctrl:1 fi |
| //! p_unit_test 4, %res4 |
| Temp tmp4 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| auto res4 = bld.vop2_e64(aco_opcode::v_max_f32, bld.def(v1), tmp4, b); |
| res4->valu().neg[0] = true; |
| writeout(4, res4); |
| |
| //! v1: %tmp5 = v_mov_b32 %a row_mirror bound_ctrl:1 fi |
| //! v1: %res5 = v_add_f32 %tmp5, %b clamp |
| //! p_unit_test 5, %res5 |
| Temp tmp5 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| auto res5 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp5, b); |
| res5->valu().clamp = true; |
| writeout(5, res5); |
| |
| //! v1: %res6 = v_add_f32 |%a|, %b row_mirror bound_ctrl:1 fi |
| //! p_unit_test 6, %res6 |
| auto tmp6 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| tmp6->dpp16().neg[0] = true; |
| auto res6 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp6, b); |
| res6->valu().abs[0] = true; |
| writeout(6, res6); |
| |
| //! v1: %res7 = v_subrev_f32 %a, |%b| row_mirror bound_ctrl:1 fi |
| //! p_unit_test 7, %res7 |
| Temp tmp7 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| auto res7 = bld.vop2_e64(aco_opcode::v_sub_f32, bld.def(v1), b, tmp7); |
| res7->valu().abs[0] = true; |
| writeout(7, res7); |
| |
| //! v1: %tmp11 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi |
| //! v1: %res11 = v_add_u32 %tmp11, %b |
| //! p_unit_test 11, %res11 |
| auto tmp11 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| tmp11->dpp16().neg[0] = true; |
| Temp res11 = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), tmp11, b); |
| writeout(11, res11); |
| |
| //! v1: %tmp12 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi |
| //! v1: %res12 = v_add_f16 %tmp12, %b |
| //! p_unit_test 12, %res12 |
| auto tmp12 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| tmp12->dpp16().neg[0] = true; |
| Temp res12 = bld.vop2(aco_opcode::v_add_f16, bld.def(v1), tmp12, b); |
| writeout(12, res12); |
| |
| /* vcc */ |
| //! v1: %res8 = v_cndmask_b32 %a, %b, %c:vcc row_mirror bound_ctrl:1 fi |
| //! p_unit_test 8, %res8 |
| Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), tmp8, b, c); |
| writeout(8, res8); |
| |
| /* sgprs */ |
| //! v1: %tmp9 = v_mov_b32 %a row_mirror bound_ctrl:1 fi |
| //! v1: %res9 = v_add_f32 %tmp9, %d |
| //! p_unit_test 9, %res9 |
| Temp tmp9 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res9 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp9, d); |
| writeout(9, res9); |
| |
| //! v1: %tmp10 = v_mov_b32 %a row_mirror bound_ctrl:1 fi |
| //! v1: %res10 = v_add_f32 %d, %tmp10 |
| //! p_unit_test 10, %res10 |
| Temp tmp10 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp res10 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), d, tmp10); |
| writeout(10, res10); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.dpp_prop) |
| //>> v1: %a, s1: %b = p_startpgm |
| if (!setup_cs("v1 s1", GFX10)) |
| return; |
| |
| //! v1: %one = p_parallelcopy 1 |
| //! v1: %res0 = v_mul_f32 1, %a |
| //! p_unit_test 0, %res0 |
| Temp one = bld.copy(bld.def(v1), Operand::c32(1)); |
| writeout(0, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), one, inputs[0], dpp_row_rr(1))); |
| |
| //! v1: %res1 = v_mul_f32 %a, %one row_ror:1 bound_ctrl:1 fi |
| //! p_unit_test 1, %res1 |
| writeout(1, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], one, dpp_row_rr(1))); |
| |
| //! v1: %res2 = v_mul_f32 0x12345678, %a |
| //! p_unit_test 2, %res2 |
| Temp literal1 = bld.copy(bld.def(v1), Operand::c32(0x12345678u)); |
| writeout(2, |
| bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), literal1, inputs[0], dpp_row_rr(1))); |
| |
| //! v1: %literal2 = p_parallelcopy 0x12345679 |
| //! v1: %res3 = v_mul_f32 %a, %literal row_ror:1 bound_ctrl:1 fi |
| //! p_unit_test 3, %res3 |
| Temp literal2 = bld.copy(bld.def(v1), Operand::c32(0x12345679u)); |
| writeout(3, |
| bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], literal2, dpp_row_rr(1))); |
| |
| //! v1: %b_v = p_parallelcopy %b |
| //! v1: %res4 = v_mul_f32 %b, %a |
| //! p_unit_test 4, %res4 |
| Temp b_v = bld.copy(bld.def(v1), inputs[1]); |
| writeout(4, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), b_v, inputs[0], dpp_row_rr(1))); |
| |
| //! v1: %res5 = v_mul_f32 %a, %b_v row_ror:1 bound_ctrl:1 fi |
| //! p_unit_test 5, %res5 |
| writeout(5, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], b_v, dpp_row_rr(1))); |
| |
| //! v1: %res6 = v_rcp_f32 %b |
| //! p_unit_test 6, %res6 |
| writeout(6, bld.vop1_dpp(aco_opcode::v_rcp_f32, bld.def(v1), b_v, dpp_row_rr(1))); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.casts) |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", GFX10_3)) |
| return; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| //! v1: %res0_tmp = v_mul_f32 -1.0, %a |
| //! v2b: %res0 = v_mul_f16 %res0_tmp, %a16 |
| //! p_unit_test 0, %res0 |
| writeout(0, fmul(u2u16(fneg(a)), a16)); |
| |
| //! v2b: %res1_tmp = v_mul_f16 -1.0, %a16 |
| //! v1: %res1 = v_mul_f32 %res1_tmp, %a |
| //! p_unit_test 1, %res1 |
| writeout(1, fmul(bld.as_uniform(fneg(a16)), a)); |
| |
| //! v1: %res2_tmp = v_mul_f32 -1.0, %a16 |
| //! v2b: %res2 = v_mul_f16 %res2_tmp, %a16 |
| //! p_unit_test 2, %res2 |
| writeout(2, fmul(u2u16(bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), |
| Operand::c32(0xbf800000u), bld.as_uniform(a16))), |
| a16)); |
| |
| //! v1: %res3_tmp = v_mul_f32 %a, %a |
| //! v2b: %res3 = v_add_f16 %res3_tmp, 0 clamp |
| //! p_unit_test 3, %res3 |
| writeout(3, fsat(u2u16(fmul(a, a)))); |
| |
| //! v2b: %res4_tmp = v_mul_f16 %a16, %a16 |
| //! v1: %res4 = v_add_f32 %res4_tmp, 0 clamp |
| //! p_unit_test 4, %res4 |
| writeout(4, fsat(bld.as_uniform(fmul(a16, a16)))); |
| |
| //! v1: %res5_tmp = v_mul_f32 %a, %a |
| //! v2b: %res5 = v_mul_f16 2.0, %res5_tmp |
| //! p_unit_test 5, %res5 |
| writeout(5, fmul(u2u16(fmul(a, a)), bld.copy(bld.def(v2b), Operand::c16(0x4000)))); |
| |
| //! v2b: %res6_tmp = v_mul_f16 %a16, %a16 |
| //! v1: %res6 = v_mul_f32 2.0, %res6_tmp |
| //! p_unit_test 6, %res6 |
| writeout(6, |
| fmul(bld.as_uniform(fmul(a16, a16)), bld.copy(bld.def(v1), Operand::c32(0x40000000)))); |
| |
| //! v1: %res7_tmp = v_mul_f32 %a, %a |
| //! v2b: %res7 = v_add_f16 %res7_tmp, %a16 |
| //! p_unit_test 7, %res7 |
| writeout(7, fadd(u2u16(fmul(a, a)), a16)); |
| |
| //! v2b: %res8_tmp = v_mul_f16 %a16, %a16 |
| //! v1: %res8 = v_add_f32 %res8_tmp, %a |
| //! p_unit_test 8, %res8 |
| writeout(8, fadd(bld.as_uniform(fmul(a16, a16)), a)); |
| |
| //! v1: %res9_tmp = v_mul_f32 %a, %a |
| //! v2b: %res9 = v_mul_f16 -1.0, %res9_tmp |
| //! p_unit_test 9, %res9 |
| writeout(9, fneg(u2u16(fmul(a, a)))); |
| |
| //! v2b: %res10_tmp = v_mul_f16 %a16, %a16 |
| //! v1: %res10 = v_mul_f32 -1.0, %res10_tmp |
| //! p_unit_test 10, %res10 |
| writeout(10, bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0xbf800000u), |
| bld.as_uniform(fmul(a16, a16)))); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.input_conv.basic) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| //! v1: %res0 = v_fma_mix_f32 %a, lo(%a16), neg(0) |
| //! p_unit_test 0, %res0 |
| writeout(0, fmul(a, f2f32(a16))); |
| |
| //! v1: %res1 = v_fma_mix_f32 1.0, lo(%a16), %a |
| //! p_unit_test 1, %res1 |
| writeout(1, fadd(a, f2f32(a16))); |
| |
| //! v1: %res2 = v_fma_mix_f32 1.0, lo(%a16), %a |
| //! p_unit_test 2, %res2 |
| writeout(2, fadd(f2f32(a16), a)); |
| |
| //! v1: %res3 = v_fma_mix_f32 %a, %a, lo(%a16) |
| //! p_unit_test 3, %res3 |
| writeout(3, fma(a, a, f2f32(a16))); |
| |
| //! v1: %res4 = v_fma_mix_f32 %a, %a, lo(%a16) |
| //! p_unit_test 4, %res4 |
| writeout(4, fma(a, a, f2f32(a16))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.input_conv.precision) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| /* precise arithmetic */ |
| //~gfx9! v1: %res0_cvt = v_cvt_f32_f16 %a16 |
| //~gfx9! v1: (precise)%res0 = v_fma_f32 %a, %a, %res0_cvt |
| //~gfx10! v1: (precise)%res0 = v_fma_mix_f32 %a, %a, lo(%a16) |
| //! p_unit_test 0, %res0 |
| writeout(0, fma(a, a, f2f32(a16), bld.precise())); |
| |
| //! v2b: %res1_cvt = v_cvt_f16_f32 %a |
| //! v2b: (precise)%res1 = v_mul_f16 %a16, %res1_cvt |
| //! p_unit_test 1, %res1 |
| writeout(1, fmul(a16, f2f16(a), bld.precise())); |
| |
| //! v2b: %res2_cvt = v_cvt_f16_f32 %a |
| //! v2b: (precise)%res2 = v_add_f16 %a16, %res2_cvt |
| //! p_unit_test 2, %res2 |
| writeout(2, fadd(a16, f2f16(a), bld.precise())); |
| |
| //! v2b: %res3_cvt = v_cvt_f16_f32 %a |
| //! v2b: (precise)%res3 = v_fma_f16 %a16, %a16, %res3_cvt |
| //! p_unit_test 3, %res3 |
| writeout(3, fma(a16, a16, f2f16(a), bld.precise())); |
| |
| /* precise conversions */ |
| //! v2b: (precise)%res4_cvt = v_cvt_f16_f32 %a |
| //! v2b: %res4 = v_mul_f16 %a16, %res4_cvt |
| //! p_unit_test 4, %res4 |
| writeout(4, fmul(a16, f2f16(a, bld.precise()))); |
| |
| //! v2b: (precise)%res5_cvt = v_cvt_f16_f32 %a |
| //! v2b: %res5 = v_add_f16 %a16, %res5_cvt |
| //! p_unit_test 5, %res5 |
| writeout(5, fadd(a16, f2f16(a, bld.precise()))); |
| |
| //! v2b: (precise)%res6_cvt = v_cvt_f16_f32 %a |
| //! v2b: %res6 = v_fma_f16 %a16, %a16, %res6_cvt |
| //! p_unit_test 6, %res6 |
| writeout(6, fma(a16, a16, f2f16(a, bld.precise()))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.input_conv.modifiers) |
| for (unsigned i = GFX9; i <= GFX11; i++) { |
| if (i == GFX10_3) |
| continue; |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| /* check whether modifiers are preserved when converting to VOP3P */ |
| //! v1: %res0 = v_fma_mix_f32 -%a, lo(%a16), neg(0) |
| //! p_unit_test 0, %res0 |
| writeout(0, fmul(fneg(a), f2f32(a16))); |
| |
| //! v1: %res1 = v_fma_mix_f32 |%a|, lo(%a16), neg(0) |
| //! p_unit_test 1, %res1 |
| writeout(1, fmul(fabs(a), f2f32(a16))); |
| |
| /* fneg modifiers */ |
| //! v1: %res2 = v_fma_mix_f32 %a, -lo(%a16), neg(0) |
| //! p_unit_test 2, %res2 |
| writeout(2, fmul(a, fneg(f2f32(a16)))); |
| |
| //! v1: %res3 = v_fma_mix_f32 %a, -lo(%a16), neg(0) |
| //! p_unit_test 3, %res3 |
| writeout(3, fmul(a, f2f32(fneg(a16)))); |
| |
| /* fabs modifiers */ |
| //! v1: %res4 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0) |
| //! p_unit_test 4, %res4 |
| writeout(4, fmul(a, fabs(f2f32(a16)))); |
| |
| //! v1: %res5 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0) |
| //! p_unit_test 5, %res5 |
| writeout(5, fmul(a, f2f32(fabs(a16)))); |
| |
| /* both fabs and fneg modifiers */ |
| //! v1: %res6 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0) |
| //! p_unit_test 6, %res6 |
| writeout(6, fmul(a, fneg(f2f32(fabs(a16))))); |
| |
| //! v1: %res7 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0) |
| //! p_unit_test 7, %res7 |
| writeout(7, fmul(a, fabs(f2f32(fabs(a16))))); |
| |
| //! v1: %res8 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0) |
| //! p_unit_test 8, %res8 |
| writeout(8, fmul(a, fneg(fabs(f2f32(fabs(a16)))))); |
| |
| //! v1: %res9 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0) |
| //! p_unit_test 9, %res9 |
| writeout(9, fmul(a, f2f32(fneg(fabs(a16))))); |
| |
| //! v1: %res10 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0) |
| //! p_unit_test 10, %res10 |
| writeout(10, fmul(a, fneg(f2f32(fneg(fabs(a16)))))); |
| |
| //! v1: %res11 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0) |
| //! p_unit_test 11, %res11 |
| writeout(11, fmul(a, fabs(f2f32(fneg(fabs(a16)))))); |
| |
| //! v1: %res12 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0) |
| //! p_unit_test 12, %res12 |
| writeout(12, fmul(a, fneg(fabs(f2f32(fneg(fabs(a16))))))); |
| |
| /* sdwa */ |
| //! v1: %res13 = v_fma_mix_f32 lo(%a), %a, neg(0) |
| //! p_unit_test 13, %res13 |
| writeout(13, fmul(f2f32(ext_ushort(a, 0)), a)); |
| |
| //! v1: %res14 = v_fma_mix_f32 hi(%a), %a, neg(0) |
| //! p_unit_test 14, %res14 |
| writeout(14, fmul(f2f32(ext_ushort(a, 1)), a)); |
| |
| //~gfx(9|10)! v1: %res15_cvt = v_cvt_f32_f16 %a dst_sel:uword0 src0_sel:dword |
| //~gfx11! v1: %res16_cvt1 = v_fma_mix_f32 lo(%a), 1.0, neg(0) |
| //~gfx11! v1: %res15_cvt = p_extract %res16_cvt1, 0, 16, 0 |
| //! v1: %res15 = v_mul_f32 %res15_cvt, %a |
| //! p_unit_test 15, %res15 |
| writeout(15, fmul(ext_ushort(f2f32(a), 0), a)); |
| |
| //~gfx(9|10)! v1: %res16_cvt = v_cvt_f32_f16 %a |
| //~gfx(9|10)! v1: %res16 = v_mul_f32 %res16_cvt, %a dst_sel:dword src0_sel:uword1 src1_sel:dword |
| //~gfx11! v1: %res16_cvt = v_fma_mix_f32 lo(%a), 1.0, neg(0) |
| //~gfx11! v1: %res16_ext = p_extract %res16_cvt, 1, 16, 0 |
| //~gfx11! v1: %res16 = v_mul_f32 %res16_ext, %a |
| //! p_unit_test 16, %res16 |
| writeout(16, fmul(ext_ushort(f2f32(a), 1), a)); |
| |
| //~gfx(9|10)! v1: %res17_cvt = v_cvt_f32_f16 %a dst_sel:dword src0_sel:ubyte2 |
| //~gfx(9|10)! v1: %res17 = v_mul_f32 %res17_cvt, %a |
| //~gfx11! v1: %res17_ext = p_extract %a, 2, 8, 0 |
| //~gfx11! v1: %res17 = v_fma_mix_f32 lo(%res17_ext), %a, neg(0) |
| //! p_unit_test 17, %res17 |
| writeout(17, fmul(f2f32(ext_ubyte(a, 2)), a)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.output_conv.basic) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm |
| if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| Temp a16 = inputs[3]; |
| Temp b16 = inputs[4]; |
| |
| //! v2b: %res0 = v_fma_mixlo_f16 %a, %b, neg(lo(0)) |
| //! p_unit_test 0, %res0 |
| writeout(0, f2f16(fmul(a, b))); |
| |
| //! v2b: %res1 = v_fma_mixlo_f16 1.0, %a, %b |
| //! p_unit_test 1, %res1 |
| writeout(1, f2f16(fadd(a, b))); |
| |
| //! v2b: %res2 = v_fma_mixlo_f16 %a, %b, %c |
| //! p_unit_test 2, %res2 |
| writeout(2, f2f16(fma(a, b, c))); |
| |
| //! v2b: %res3 = v_fma_mixlo_f16 lo(%a16), %b, neg(lo(0)) |
| //! p_unit_test 3, %res3 |
| writeout(3, f2f16(fmul(f2f32(a16), b))); |
| |
| //! v2b: %res4 = v_fma_mixlo_f16 1.0, %a, lo(%b16) |
| //! p_unit_test 4, %res4 |
| writeout(4, f2f16(fadd(a, f2f32(b16)))); |
| |
| //! v2b: %res5 = v_fma_mixlo_f16 %a, lo(%b16), %c |
| //! p_unit_test 5, %res5 |
| writeout(5, f2f16(fma(a, f2f32(b16), c))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.output_conv.precision) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v2b: %a16 = p_startpgm |
| if (!setup_cs("v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a16 = inputs[0]; |
| |
| //! v2b: %res0_tmp = v_mul_f16 %a16, %a16 |
| //! v1: (precise)%res0 = v_cvt_f32_f16 %res0_tmp |
| //! p_unit_test 0, %res0 |
| writeout(0, f2f32(fmul(a16, a16), bld.precise())); |
| |
| //! v2b: (precise)%res1_tmp = v_mul_f16 %a16, %a16 |
| //! v1: %res1 = v_cvt_f32_f16 %res1_tmp |
| //! p_unit_test 1, %res1 |
| writeout(1, f2f32(fmul(a16, a16, bld.precise()))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.output_conv.modifiers) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, v2b: %a16, v2b: %b16 = p_startpgm |
| if (!setup_cs("v1 v1 v2b v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp a16 = inputs[2]; |
| Temp b16 = inputs[3]; |
| |
| /* fneg/fabs */ |
| //! v1: %res0_add = v_add_f32 %1, %2 |
| //! v2b: %res0 = v_cvt_f16_f32 |%res0_add| |
| //! p_unit_test 0, %res0 |
| writeout(0, f2f16(fabs(fadd(a, b)))); |
| |
| //! v1: %res1_add = v_add_f32 %1, %2 |
| //! v2b: %res1 = v_cvt_f16_f32 -%res1_add |
| //! p_unit_test 1, %res1 |
| writeout(1, f2f16(fneg(fadd(a, b)))); |
| |
| //! v2b: %res2_add = v_add_f16 %3, %4 |
| //! v1: %res2 = v_cvt_f32_f16 |%res2_add| |
| //! p_unit_test 2, %res2 |
| writeout(2, f2f32(fabs(fadd(a16, b16)))); |
| |
| //! v2b: %res3_add = v_add_f16 %3, %4 |
| //! v1: %res3 = v_cvt_f32_f16 -%res3_add |
| //! p_unit_test 3, %res3 |
| writeout(3, f2f32(fneg(fadd(a16, b16)))); |
| |
| /* sdwa */ |
| //! v2b: %res4_add = v_fma_mixlo_f16 1.0, %a, %b |
| //! v2b: %res4 = p_extract %res4_add, 0, 8, 0 |
| //! p_unit_test 4, %res4 |
| writeout(4, ext_ubyte(f2f16(fadd(a, b)), 0)); |
| |
| //! v1: %res5_mul = v_add_f32 %a, %b dst_sel:uword0 src0_sel:dword src1_sel:dword |
| //! v2b: %res5 = v_cvt_f16_f32 %res5_mul |
| //! p_unit_test 5, %res5 |
| writeout(5, f2f16(ext_ushort(fadd(a, b), 0))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.fma.basic) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %c16 = p_startpgm |
| if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| Temp a16 = inputs[3]; |
| Temp c16 = inputs[4]; |
| |
| //! v1: %res0 = v_fma_mix_f32 lo(%a16), %b, %c |
| //! p_unit_test 0, %res0 |
| writeout(0, fadd(fmul(f2f32(a16), b), c)); |
| |
| //! v1: %res1 = v_fma_mix_f32 %a, %b, lo(%c16) |
| //! p_unit_test 1, %res1 |
| writeout(1, fadd(fmul(a, b), f2f32(c16))); |
| |
| /* omod/clamp check */ |
| //! v1: %res2_mul = v_fma_mix_f32 lo(%a16), %b, neg(0) |
| //! v1: %res2 = v_add_f32 %res2_mul, %c *2 |
| //! p_unit_test 2, %res2 |
| writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000), |
| fadd(fmul(f2f32(a16), b), c))); |
| |
| /* neg/abs modifiers */ |
| //! v1: %res3 = v_fma_mix_f32 -lo(%a16), %b, |lo(%c16)| |
| //! p_unit_test 3, %res3 |
| writeout(3, fadd(fmul(fneg(f2f32(a16)), b), fabs(f2f32(c16)))); |
| |
| //! v1: %res4 = v_fma_mix_f32 |%a|, |%b|, lo(%c16) |
| //! p_unit_test 4, %res4 |
| writeout(4, fadd(fabs(fmul(fneg(a), fneg(b))), f2f32(c16))); |
| |
| //! v1: %res5 = v_fma_mix_f32 %a, -%b, lo(%c16) |
| //! p_unit_test 5, %res5 |
| writeout(5, fadd(fneg(fmul(a, b)), f2f32(c16))); |
| |
| //! v1: %res6 = v_fma_mix_f32 |%a|, -|%b|, lo(%c16) |
| //! p_unit_test 6, %res6 |
| writeout(6, fadd(fneg(fabs(fmul(fneg(a), fneg(b)))), f2f32(c16))); |
| |
| /* output conversions */ |
| //! v2b: %res7 = v_fma_mixlo_f16 %a, %b, %c |
| //! p_unit_test 7, %res7 |
| writeout(7, f2f16(fadd(fmul(a, b), c))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.fma.precision) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm |
| if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| Temp a16 = inputs[3]; |
| Temp b16 = inputs[4]; |
| |
| /* the optimization is precise for 32-bit on GFX9 */ |
| //~gfx9! v1: (precise)%res0 = v_fma_mix_f32 lo(%a16), %b, %c |
| //~gfx10! v1: (precise)%res0_tmp = v_fma_mix_f32 lo(%a16), %b, neg(0) |
| //~gfx10! v1: %res0 = v_add_f32 %res0_tmp, %c |
| //! p_unit_test 0, %res0 |
| writeout(0, fadd(fmul(f2f32(a16), b, bld.precise()), c)); |
| |
| //~gfx9! v1: (precise)%res1 = v_fma_mix_f32 lo(%a16), %b, %c |
| //~gfx10! v1: %res1_tmp = v_fma_mix_f32 lo(%a16), %b, neg(0) |
| //~gfx10! v1: (precise)%res1 = v_add_f32 %res1_tmp, %c |
| //! p_unit_test 1, %res1 |
| writeout(1, fadd(fmul(f2f32(a16), b), c, bld.precise())); |
| |
| /* never promote 16-bit arithmetic to 32-bit */ |
| //! v2b: %res2_tmp = v_cvt_f16_f32 %a |
| //! v2b: %res2 = v_add_f16 %res2_tmp, %b16 |
| //! p_unit_test 2, %res2 |
| writeout(2, fadd(f2f16(a), b16)); |
| |
| //! v2b: %res3_tmp = v_cvt_f16_f32 %a |
| //! v2b: %res3 = v_mul_f16 %res3_tmp, %b16 |
| //! p_unit_test 3, %res3 |
| writeout(3, fmul(f2f16(a), b16)); |
| |
| //! v2b: %res4_tmp = v_mul_f16 %a16, %b16 |
| //! v1: %res4 = v_cvt_f32_f16 %res4_tmp |
| //! p_unit_test 4, %res4 |
| writeout(4, f2f32(fmul(a16, b16))); |
| |
| //! v2b: %res5_tmp = v_add_f16 %a16, %b16 |
| //! v1: %res5 = v_cvt_f32_f16 %res5_tmp |
| //! p_unit_test 5, %res5 |
| writeout(5, f2f32(fadd(a16, b16))); |
| |
| //! v2b: %res6_tmp = v_fma_mixlo_f16 %a, %b, neg(lo(0)) |
| //! v2b: %res6 = v_add_f16 %res6_tmp, %a16 |
| //! p_unit_test 6, %res6 |
| writeout(6, fadd(f2f16(fmul(a, b)), a16)); |
| |
| //! v2b: %res7_tmp = v_mul_f16 %a16, %b16 |
| //! v1: %res7 = v_fma_mix_f32 1.0, lo(%res7_tmp), %c |
| //! p_unit_test 7, %res7 |
| writeout(7, fadd(f2f32(fmul(a16, b16)), c)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.clamp) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| //! v1: %res0 = v_fma_mix_f32 lo(%a16), %a, neg(0) clamp |
| //! p_unit_test 0, %res0 |
| writeout(0, fsat(fmul(f2f32(a16), a))); |
| |
| //! v2b: %res1 = v_fma_mixlo_f16 %a, %a, neg(lo(0)) clamp |
| //! p_unit_test 1, %res1 |
| writeout(1, f2f16(fsat(fmul(a, a)))); |
| |
| //! v2b: %res2 = v_fma_mixlo_f16 %a, %a, neg(lo(0)) clamp |
| //! p_unit_test 2, %res2 |
| writeout(2, fsat(f2f16(fmul(a, a)))); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.mad_mix.cast) |
| for (unsigned i = GFX9; i <= GFX10; i++) { |
| //>> v1: %a, v2b: %a16 = p_startpgm |
| if (!setup_cs("v1 v2b", (amd_gfx_level)i)) |
| continue; |
| |
| program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush; |
| |
| Temp a = inputs[0]; |
| Temp a16 = inputs[1]; |
| |
| /* The optimizer copy-propagates v2b=p_extract_vector(v1, 0) and p_as_uniform, so the |
| * optimizer has to check compatibility. |
| */ |
| |
| //! v1: %res0_cvt = v_cvt_f32_f16 %a16 |
| //! v2b: %res0 = v_mul_f16 %res0_cvt, %a16 |
| //! p_unit_test 0, %res0 |
| writeout(0, fmul(u2u16(f2f32(a16)), a16)); |
| |
| //! v2b: %res1_cvt = v_cvt_f16_f32 %a |
| //! v1: %res1 = v_mul_f32 %res1_cvt, %a |
| //! p_unit_test 1, %res1 |
| writeout(1, fmul(bld.as_uniform(f2f16(a)), a)); |
| |
| //! v2b: %res2_mul = v_mul_f16 %a16, %a16 |
| //! v2b: %res2 = v_cvt_f16_f32 %res2_mul |
| //! p_unit_test 2, %res2 |
| writeout(2, f2f16(bld.as_uniform(fmul(a16, a16)))); |
| |
| //! v1: %res3_mul = v_mul_f32 %a, %a |
| //! v1: %res3 = v_cvt_f32_f16 %res3_mul |
| //! p_unit_test 3, %res3 |
| writeout(3, f2f32(u2u16(fmul(a, a)))); |
| |
| //! v1: %res4_mul = v_fma_mix_f32 lo(%a16), %a, neg(0) |
| //! v2b: %res4 = v_add_f16 %res4_mul, 0 clamp |
| //! p_unit_test 4, %res4 |
| writeout(4, fsat(u2u16(fmul(f2f32(a16), a)))); |
| |
| //! v2b: %res5_mul = v_fma_mixlo_f16 %a, %a, neg(lo(0)) |
| //! v1: %res5 = v_add_f32 %res5_mul, 0 clamp |
| //! p_unit_test 5, %res5 |
| writeout(5, fsat(bld.as_uniform(f2f16(fmul(a, a))))); |
| |
| //! v1: %res6_mul = v_mul_f32 %a, %a |
| //! v1: %res6 = v_fma_mix_f32 1.0, lo(%res6_mul), %a |
| //! p_unit_test 6, %res6 |
| writeout(6, fadd(f2f32(u2u16(fmul(a, a))), a)); |
| |
| //! v2b: %res7_mul = v_mul_f16 %a16, %a16 |
| //! v1: %res7 = v_fma_mix_f32 1.0, lo(%a16), %res7_mul |
| //! p_unit_test 7, %res7 |
| writeout(7, fadd(bld.as_uniform(fmul(a16, a16)), f2f32(a16))); |
| |
| /* opsel_hi should be obtained from the original opcode, not the operand regclass */ |
| //! v1: %res8 = v_fma_mix_f32 lo(%a16), %a16, neg(0) |
| //! p_unit_test 8, %res8 |
| writeout(8, fmul(f2f32(a16), a16)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| static void |
| vop3p_constant(unsigned* idx, aco_opcode op, const char* swizzle, uint32_t val) |
| { |
| uint32_t halves[2] = {val & 0xffff, val >> 16}; |
| uint32_t expected = halves[swizzle[0] - 'x'] | (halves[swizzle[1] - 'x'] << 16); |
| fprintf(output, "Expected for %u: 0x%.8x / %u\n", *idx, expected, expected); |
| |
| unsigned opsel_lo = swizzle[0] == 'x' ? 0x0 : 0x1; |
| unsigned opsel_hi = swizzle[1] == 'x' ? 0x2 : 0x3; |
| writeout((*idx)++, bld.vop3p(op, bld.def(v1), bld.copy(bld.def(v1), Operand::c32(val)), |
| inputs[0], opsel_lo, opsel_hi)); |
| } |
| |
| BEGIN_TEST(optimize.vop3p_constants) |
| for (aco_opcode op : {aco_opcode::v_pk_add_f16, aco_opcode::v_pk_add_u16}) { |
| for (const char* swizzle : {"xx", "yy", "xy", "yx"}) { |
| char variant[16]; |
| strcpy(variant, op == aco_opcode::v_pk_add_f16 ? "_f16" : "_u16"); |
| strcat(variant, "_"); |
| strcat(variant, swizzle); |
| |
| //; for i in range(36): |
| //; insert_pattern('Expected for %u: $_ / #expected%u' % (i, i)) |
| |
| //>> v1: %a = p_startpgm |
| if (!setup_cs("v1", GFX10_3, CHIP_UNKNOWN, variant)) |
| continue; |
| |
| //; opcode = 'v_pk_add_u16' if 'u16' in variant else 'v_pk_add_f16' |
| //; for i in range(36): |
| //; insert_pattern('v1: %%res%u = %s $got%u %%a' % (i, opcode, i)) |
| //; insert_pattern('p_unit_test %u, %%res%u' % (i, i)) |
| //! s_endpgm |
| |
| //; def parse_op(op): |
| //; is_int = opcode == 'v_pk_add_u16' |
| //; op = op.rstrip(',') |
| //; |
| //; mods = lambda v: v |
| //; if op.endswith('*[1,-1]'): |
| //; mods = lambda v: v ^ 0x80000000 |
| //; assert(not is_int) |
| //; elif op.endswith('*[-1,1]'): |
| //; mods = lambda v: v ^ 0x00008000 |
| //; assert(not is_int) |
| //; elif op.startswith('neg('): |
| //; mods = lambda v: v ^ 0x80008000 |
| //; assert(not is_int) |
| //; op = op[4:-1] |
| //; op = op.split('*')[0] |
| //; |
| //; swizzle = lambda v: v |
| //; if op.endswith('.xx'): |
| //; swizzle = lambda v: ((v & 0xffff) | (v << 16)) & 0xffffffff; |
| //; elif op.endswith('.yy'): |
| //; swizzle = lambda v: (v >> 16) | (v & 0xffff0000); |
| //; elif op.endswith('.yx'): |
| //; swizzle = lambda v: ((v >> 16) | (v << 16)) & 0xffffffff; |
| //; op = op.rstrip('xy.') |
| //; |
| //; val = None |
| //; if op.startswith('0x'): |
| //; val = int(op[2:], 16) |
| //; elif op == '-1.0': |
| //; val = 0xbf800000 if is_int else 0xbC00 |
| //; elif op == '1.0': |
| //; val = 0x3f800000 if is_int else 0x3c00 |
| //; else: |
| //; val = int(op) & 0xffffffff |
| //; |
| //; return mods(swizzle(val)) |
| |
| //; # Check correctness |
| //; for i in range(36): |
| //; expected = globals()['expected%u' % i] |
| //; got = globals()['got%u' % i] |
| //; got_parsed = parse_op(got) |
| //; if got_parsed != expected: |
| //; raise Exception('Check %u failed: expected 0x%.8x, got 0x%.8x ("%s")' % (i, expected, got_parsed, got)) |
| |
| //; # Check that all literals are ones that cannot be encoded as inline constants |
| //; allowed_literals = [0x00004242, 0x0000fffe, 0x00308030, 0x0030ffff, 0x3c00ffff, |
| //; 0x42420000, 0x42424242, 0x4242c242, 0x4242ffff, 0x7ffefffe, |
| //; 0x80300030, 0xbeefdead, 0xc2424242, 0xdeadbeef, 0xfffe0000, |
| //; 0xfffe7ffe, 0xffff0030, 0xffff3c00, 0xffff4242, 0xc242c242, |
| //; 0x80308030, 0xdeaddead, 0xbeefbeef, 0x7ffe7ffe] |
| //; if opcode == 'v_pk_add_u16': |
| //; allowed_literals.extend([0x00003c00, 0x3c000000, 0x3c003c00, 0x3c00bc00, 0xbc003c00, 0xbc00bc00]) |
| //; else: |
| //; allowed_literals.extend([0x00003f80, 0x3f800000, 0x3f803f80]) |
| //; |
| //; for i in range(36): |
| //; got = globals()['got%u' % i].removeprefix('neg(') |
| //; if not got.startswith('0x'): |
| //; continue; |
| //; got = int(got[2:].rstrip(',)').split('*')[0].split('.')[0], 16) |
| //; if got not in allowed_literals: |
| //; raise Exception('Literal check %u failed: 0x%.8x not in allowed literals' % (i, got)) |
| |
| unsigned idx = 0; |
| for (uint32_t constant : {0x3C00, 0x0030, 0xfffe, 0x4242}) { |
| vop3p_constant(&idx, op, swizzle, constant); |
| vop3p_constant(&idx, op, swizzle, constant | 0xffff0000); |
| vop3p_constant(&idx, op, swizzle, constant | (constant << 16)); |
| vop3p_constant(&idx, op, swizzle, constant << 16); |
| vop3p_constant(&idx, op, swizzle, (constant << 16) | 0x0000ffff); |
| vop3p_constant(&idx, op, swizzle, constant | ((constant ^ 0x8000) << 16)); |
| vop3p_constant(&idx, op, swizzle, (constant ^ 0x8000) | (constant << 16)); |
| } |
| |
| for (uint32_t constant : {0x3f800000u, 0xfffffffeu, 0x00000030u, 0xdeadbeefu}) { |
| uint32_t lo = constant & 0xffff; |
| uint32_t hi = constant >> 16; |
| vop3p_constant(&idx, op, swizzle, constant); |
| vop3p_constant(&idx, op, swizzle, hi | (lo << 16)); |
| } |
| |
| finish_opt_test(); |
| } |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.fmamix_two_literals) |
| /* This test has to recreate literals sometimes because we don't combine them at all if there's |
| * at least one uncombined use. |
| */ |
| for (unsigned i = GFX10; i <= GFX10_3; i++) { |
| //>> v1: %a, v1: %b = p_startpgm |
| if (!setup_cs("v1 v1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| |
| Temp c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); |
| Temp c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); |
| Temp c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000)); |
| |
| //! v1: %res0 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00) |
| //! p_unit_test 0, %res0 |
| writeout(0, fma(a, c15, c30)); |
| |
| /* No need to use v_fma_mix_f32. */ |
| //! v1: %res1 = v_fmaak_f32 %a, %b, 0x40400000 |
| //! p_unit_test 1, %res1 |
| writeout(1, fma(a, b, c30)); |
| |
| /* Separate mul/add can become v_fma_mix_f32 if it's not precise. */ |
| //! v1: %res2 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00) |
| //! p_unit_test 2, %res2 |
| writeout(2, fadd(fmul(a, c15), c30)); |
| |
| //~gfx10! v1: %c15 = p_parallelcopy 0x3fc00000 |
| c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); |
| c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); |
| |
| /* v_fma_mix_f32 is a fused mul/add, so it can't be used for precise separate mul/add. */ |
| //~gfx10! v1: (precise)%res3 = v_madak_f32 %a, %c15, 0x40400000 |
| //~gfx10_3! v1: (precise)%res3_tmp = v_mul_f32 %a, 0x3fc00000 |
| //~gfx10_3! v1: %res3 = v_add_f32 %res3_tmp, 0x40400000 |
| //! p_unit_test 3, %res3 |
| writeout(3, fadd(bld.precise().vop2(aco_opcode::v_mul_f32, bld.def(v1), a, c15), c30)); |
| |
| //~gfx10! v1: (precise)%res4 = v_madak_f32 %1, %c16, 0x40400000 |
| //~gfx10_3! v1: %res4_tmp = v_mul_f32 %a, 0x3fc00000 |
| //~gfx10_3! v1: (precise)%res4 = v_add_f32 %res4_tmp, 0x40400000 |
| //! p_unit_test 4, %res4 |
| writeout(4, bld.precise().vop2(aco_opcode::v_add_f32, bld.def(v1), fmul(a, c15), c30)); |
| |
| /* Can't convert to fp16 if it will be flushed as a denormal. */ |
| //! v1: %res5 = v_fma_mix_f32 %1, lo(0x3ff3e00), hi(0x3ff3e00) |
| //! p_unit_test 5, %res5 |
| c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); |
| writeout(5, fma(a, c15, c_denorm)); |
| |
| //>> BB1 |
| //! /* logical preds: BB0, / linear preds: BB0, / kind: uniform, */ |
| program->next_fp_mode.denorm16_64 = fp_denorm_flush; |
| bld.reset(program->create_and_insert_block()); |
| program->blocks[0].linear_succs.push_back(1); |
| program->blocks[0].logical_succs.push_back(1); |
| program->blocks[1].linear_preds.push_back(0); |
| program->blocks[1].logical_preds.push_back(0); |
| |
| //~gfx10; del c15 |
| //! v1: %c15 = p_parallelcopy 0x3fc00000 |
| //! v1: %res6 = v_fmaak_f32 %a, %c15, 0x387fc000 |
| //! p_unit_test 6, %res6 |
| c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f))); |
| c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000)); |
| writeout(6, fma(a, c15, c_denorm)); |
| |
| /* Can't accept more than 3 unique fp16 literals. */ |
| //! v1: %c45 = p_parallelcopy 0x40900000 |
| //! v1: %res7 = v_fma_mix_f32 lo(0x42003e00), hi(0x42003e00), %c45 |
| //! p_unit_test 7, %res7 |
| Temp c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f))); |
| writeout(7, fma(c15, c30, c45)); |
| |
| /* Modifiers must be preserved. */ |
| //! v1: %res8 = v_fma_mix_f32 -%a, lo(0x44804200), hi(0x44804200) |
| //! p_unit_test 8, %res8 |
| writeout(8, fma(fneg(a), c30, c45)); |
| |
| //! v1: %res9 = v_fma_mix_f32 lo(0x44804200), |%a|, hi(0x44804200) |
| //! p_unit_test 9, %res9 |
| writeout(9, fma(c30, fabs(a), c45)); |
| |
| //! v1: %res10 = v_fma_mix_f32 %a, lo(0x44804200), hi(0x44804200) clamp |
| //! p_unit_test 10, %res10 |
| writeout(10, fsat(fma(a, c30, c45))); |
| |
| /* Output modifiers are not supported by v_fma_mix_f32. */ |
| c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f))); |
| //; del c45 |
| //! v1: %c45 = p_parallelcopy 0x40900000 |
| //! v1: %res11 = v_fma_f32 %a, 0x40400000, %c45 *0.5 |
| //! p_unit_test 11, %res11 |
| c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f))); |
| writeout(11, fmul(fma(a, c30, c45), bld.copy(bld.def(v1), Operand::c32(0x3f000000)))); |
| |
| /* Has a literal which can't be represented as fp16. */ |
| //! v1: %c03 = p_parallelcopy 0x3e99999a |
| //! v1: %res12 = v_fmaak_f32 %a, %c03, 0x40400000 |
| //! p_unit_test 12, %res12 |
| Temp c03 = bld.copy(bld.def(v1), Operand::c32(fui(0.3f))); |
| writeout(12, fma(a, c03, c30)); |
| |
| /* We should still use fmaak/fmamk if the two literals are identical. */ |
| //! v1: %res13 = v_fmaak_f32 0x40400000, %a, 0x40400000 |
| //! p_unit_test 13, %res13 |
| writeout(13, fma(a, c30, c30)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.fma_opsel) |
| /* TODO make these work before GFX11 using SDWA. */ |
| for (unsigned i = GFX11; i <= GFX11; i++) { |
| //>> v2b: %a, v2b: %b, v1: %c, v1: %d, v1: %e = p_startpgm |
| if (!setup_cs("v2b v2b v1 v1 v1", (amd_gfx_level)i)) |
| continue; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| Temp d = inputs[3]; |
| Temp e = inputs[4]; |
| Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1)); |
| Temp d_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), d, Operand::c32(1)); |
| Temp e_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), e, Operand::c32(1)); |
| |
| //! v2b: %res0 = v_fma_f16 %b, hi(%c), %a |
| //! p_unit_test 0, %res0 |
| writeout(0, fadd(fmul(b, c_hi), a)); |
| |
| //! v2b: %res1 = v_fma_f16 %a, %b, hi(%d) |
| //! p_unit_test 1, %res1 |
| writeout(1, fadd(fmul(a, b), d_hi)); |
| |
| //! v2b: %res2 = v_fma_f16 %a, %b, hi(%e) |
| //! p_unit_test 2, %res2 |
| writeout(2, fma(a, b, e_hi)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.dpp_opsel) |
| //>> v1: %a, v1: %b = p_startpgm |
| if (!setup_cs("v1 v1", GFX11)) |
| return; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| |
| Temp dpp16 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror); |
| Temp dpp16_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp16, Operand::c32(1)); |
| Temp dpp8 = bld.vop1_dpp8(aco_opcode::v_mov_b32, bld.def(v1), a); |
| Temp dpp8_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp8, Operand::c32(1)); |
| |
| Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1)); |
| Temp b_lo = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(0)); |
| |
| //! v2b: %res0 = v_add_f16 hi(%a), hi(%b) row_mirror bound_ctrl:1 fi |
| //! p_unit_test 0, %res0 |
| writeout(0, fadd(dpp16_hi, b_hi)); |
| |
| //! v2b: %res1 = v_add_f16 hi(%a), %b dpp8:[0,0,0,0,0,0,0,0] fi |
| //! p_unit_test 1, %res1 |
| writeout(1, fadd(b_lo, dpp8_hi)); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.apply_sgpr_swap_opsel) |
| //>> v1: %a, s1: %b = p_startpgm |
| if (!setup_cs("v1 s1", GFX11)) |
| return; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| |
| Temp b_vgpr = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), bld.copy(bld.def(v1), b), |
| Operand::c32(0)); |
| |
| Temp res0 = bld.tmp(v2b); |
| VALU_instruction& valu = bld.vop2(aco_opcode::v_sub_f16, Definition(res0), a, b_vgpr)->valu(); |
| valu.opsel[0] = true; |
| |
| //! v2b: %res0 = v_subrev_f16 %b, hi(%a) |
| //! p_unit_test 0, %res0 |
| writeout(0, res0); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.max3_opsel) |
| /* TODO make these work before GFX11 using SDWA. */ |
| for (unsigned i = GFX11; i <= GFX11; i++) { |
| //>> v1: %a, v1: %b, v2b: %c = p_startpgm |
| if (!setup_cs("v1 v1 v2b", GFX11)) |
| continue; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| Temp c = inputs[2]; |
| |
| Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); |
| Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1)); |
| |
| //! v2b: %res0 = v_max3_f16 hi(%a), %c, hi(%b) |
| //! p_unit_test 0, %res0 |
| writeout(0, bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), |
| bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), a_hi, c), b_hi)); |
| |
| finish_opt_test(); |
| } |
| END_TEST |
| |
| BEGIN_TEST(optimize.neg_mul_opsel) |
| //>> v1: %a, v2b: %b = p_startpgm |
| if (!setup_cs("v1 v2b", GFX11)) |
| return; |
| |
| Temp a = inputs[0]; |
| Temp b = inputs[1]; |
| |
| Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1)); |
| |
| //! v2b: %res0 = v_mul_f16 -hi(%a), %b |
| //! p_unit_test 0, %res0 |
| writeout(0, fneg(fmul(a_hi, b))); |
| |
| //! v1: %res1 = v_fma_mix_f32 -hi(%a), lo(%b), neg(0) |
| //! p_unit_test 1, %res1 |
| writeout(1, fneg(fmul(f2f32(a_hi), f2f32(b)))); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.vinterp_inreg_output_modifiers) |
| //>> v1: %a, v1: %b, v1: %c = p_startpgm |
| if (!setup_cs("v1 v1 v1", GFX11)) |
| return; |
| |
| //! v1: %res0 = v_interp_p2_f32_inreg %a, %b, %c clamp |
| //! p_unit_test 0, %res0 |
| Temp tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[0], |
| inputs[1], inputs[2]); |
| writeout(0, fsat(tmp)); |
| |
| //! v1: %res1 = v_fma_f32 %b, %a, %c *2 quad_perm:[2,2,2,2] fi |
| //! p_unit_test 1, %res1 |
| tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[1], inputs[0], |
| inputs[2]); |
| tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp); |
| writeout(1, tmp); |
| |
| //! v2b: %res2 = v_interp_p2_f16_f32_inreg %a, %b, %c clamp |
| //! p_unit_test 2, %res2 |
| tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[0], |
| inputs[1], inputs[2]); |
| writeout(2, fsat(tmp)); |
| |
| //! v2b: %tmp3 = v_interp_p2_f16_f32_inreg %b, %a, %c |
| //! v2b: %res3 = v_mul_f16 2.0, %tmp3 |
| //! p_unit_test 3, %res3 |
| tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[1], |
| inputs[0], inputs[2]); |
| tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp); |
| writeout(3, tmp); |
| |
| //! v2b: %res4 = v_fma_mixlo_f16 %c, %b, %a quad_perm:[2,2,2,2] fi |
| //! p_unit_test 4, %res4 |
| tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[2], inputs[1], |
| inputs[0]); |
| writeout(4, f2f16(tmp)); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimize.s_pack) |
| //>> s1: %a, s1: %b, s1: %c = p_startpgm |
| if (!setup_cs("s1 s1 s1", GFX11)) |
| return; |
| |
| Temp lo = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[1], |
| Operand::c32(0), Operand::c32(16u), Operand::c32(false)); |
| Temp hi = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[2], |
| Operand::c32(1), Operand::c32(16u), Operand::c32(false)); |
| |
| //! s1: %res0 = s_pack_lh_b32_b16 %b, %c |
| //! p_unit_test 0, %res0 |
| writeout(0, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, hi)); |
| |
| //! s1: %res1 = s_pack_ll_b32_b16 %b, %b |
| //! p_unit_test 1, %res1 |
| writeout(1, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, lo)); |
| |
| //! s1: %res2 = s_pack_hl_b32_b16 %c, %b |
| //! p_unit_test 2, %res2 |
| writeout(2, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, lo)); |
| |
| //! s1: %res3 = s_pack_hh_b32_b16 %c, %c |
| //! p_unit_test 3, %res3 |
| writeout(3, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, hi)); |
| |
| lo = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[1], Operand::c32(0), |
| Operand::c32(16u), Operand::c32(false)); |
| hi = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[2], Operand::c32(1), |
| Operand::c32(16u), Operand::c32(false)); |
| |
| //! s1: %res4 = s_pack_ll_b32_b16 %a, %b |
| //! p_unit_test 4, %res4 |
| writeout(4, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), inputs[0], lo)); |
| |
| //! s1: %res5 = s_pack_lh_b32_b16 %a, %c |
| //! p_unit_test 5, %res5 |
| writeout(5, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), inputs[0], hi)); |
| |
| //! s1: %res6 = s_pack_ll_b32_b16 %b, %a |
| //! p_unit_test 6, %res6 |
| writeout(6, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, inputs[0])); |
| |
| //! s1: %res7 = s_pack_hl_b32_b16 %c, %a |
| //! p_unit_test 7, %res7 |
| writeout(7, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, inputs[0])); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimizer.trans_inline_constant) |
| if (!setup_cs("", GFX12)) |
| return; |
| |
| //>> s1: %res0 = v_s_rcp_f32 1.0 |
| //! p_unit_test 0, %res0 |
| writeout(0, bld.vop3(aco_opcode::v_s_rcp_f32, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3f800000)))); |
| |
| //! s1: %res1 = v_s_rcp_f32 0x3c00 |
| //! p_unit_test 1, %res1 |
| writeout(1, bld.vop3(aco_opcode::v_s_rcp_f32, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3c00)))); |
| |
| //! s1: %tmp2 = p_parallelcopy 1.0 |
| //! s1: %res2 = v_s_rcp_f16 %tmp2 |
| //! p_unit_test 2, %res2 |
| writeout(2, bld.vop3(aco_opcode::v_s_rcp_f16, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3f800000)))); |
| |
| //! s1: %tmp3 = p_parallelcopy 0x3c00 |
| //! s1: %res3 = v_s_rcp_f16 %tmp3 |
| //! p_unit_test 3, %res3 |
| writeout(3, bld.vop3(aco_opcode::v_s_rcp_f16, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3c00)))); |
| |
| //! v1: %res4 = v_rcp_f32 1.0 |
| //! p_unit_test 4, %res4 |
| writeout(4, bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), bld.copy(bld.def(s1), Operand::c32(0x3f800000)))); |
| |
| //! v1: %res5 = v_rcp_f32 0x3c00 |
| //! p_unit_test 5, %res5 |
| writeout(5, bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), bld.copy(bld.def(s1), Operand::c32(0x3c00)))); |
| |
| //! v2b: %res6 = v_rcp_f16 0x3f800000 |
| //! p_unit_test 6, %res6 |
| writeout(6, bld.vop1(aco_opcode::v_rcp_f16, bld.def(v2b), bld.copy(bld.def(s1), Operand::c32(0x3f800000)))); |
| |
| //! v2b: %res7 = v_rcp_f16 1.0 |
| //! p_unit_test 7, %res7 |
| writeout(7, bld.vop1(aco_opcode::v_rcp_f16, bld.def(v2b), bld.copy(bld.def(s1), Operand::c32(0x3c00)))); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimizer.trans_no_omod) |
| //>> s1: %a = p_startpgm |
| if (!setup_cs("s1", GFX12)) |
| return; |
| |
| //! s1: %tmp0 = v_s_log_f32 %a |
| //! v1: %res = v_mul_legacy_f32 %tmp0, 0.5 |
| //! p_unit_test 0, %res |
| Temp dst = bld.vop3(aco_opcode::v_s_log_f32, bld.def(s1), inputs[0]); |
| writeout(0, bld.vop2(aco_opcode::v_mul_legacy_f32, bld.def(v1), dst, |
| bld.copy(bld.def(v1), Operand::c32(0x3f000000)))); |
| |
| finish_opt_test(); |
| END_TEST |
| |
| BEGIN_TEST(optimizer.pk_clamp_fma_mix) |
| //>> v1: %a, v1: %b, v2b: %c = p_startpgm |
| if (!setup_cs("v1 v1 v2b", GFX12)) |
| return; |
| |
| //! v2b: %cvt0 = v_fma_mixlo_f16 1.0, %a, %b |
| //! v1: %clamp0 = v_pk_mul_f16 %cvt0.xx, 1.0.xx clamp |
| //! p_unit_test 0, %clamp0 |
| Temp add0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]); |
| Temp cvt0 = bld.vop1(aco_opcode::v_cvt_f16_f32, bld.def(v2b), add0); |
| Builder::Result clamp0 = |
| bld.vop3p(aco_opcode::v_pk_mul_f16, bld.def(v1), cvt0, Operand::c16(0x3c00), 0, 0); |
| clamp0->valu().clamp = true; |
| |
| writeout(0, clamp0); |
| |
| //! v1: %add1 = v_fma_mix_f32 1.0, lo(%c), %a |
| //! v1: %clamp1 = v_pk_mul_f16 %add1, 1.0.xx clamp |
| //! p_unit_test 1, %clamp1 |
| Temp cvt1 = bld.vop1(aco_opcode::v_cvt_f32_f16, bld.def(v1), inputs[2]); |
| Temp add1 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], cvt1); |
| Builder::Result clamp1 = |
| bld.vop3p(aco_opcode::v_pk_mul_f16, bld.def(v1), add1, Operand::c16(0x3c00), 0, 0x1); |
| clamp1->valu().clamp = true; |
| |
| writeout(1, clamp1); |
| |
| finish_opt_test(); |
| END_TEST |