blob: 745d0eb46e0862529416c7cc896ce5b707d9e4c8 [file] [log] [blame]
/*
* Copyright © 2020 Valve Corporation
*
* SPDX-License-Identifier: MIT
*/
#include "helpers.h"
using namespace aco;
BEGIN_TEST(optimize.neg)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v1: %b, s1: %c, s1: %d = p_startpgm
if (!setup_cs("v1 v1 s1 s1", (amd_gfx_level)i))
continue;
//! v1: %res0 = v_mul_f32 %a, -%b
//! p_unit_test 0, %res0
Temp neg_b = fneg(inputs[1]);
writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_b));
//~gfx9! v1: %neg_a = v_mul_f32 -1.0, %a
//~gfx9! v1: %res1 = v_max_f32 0x123456, %neg_a
//~gfx10! v1: %res1 = v_max_f32 0x123456, -%a
//! p_unit_test 1, %res1
Temp neg_a = fneg(inputs[0]);
writeout(1, bld.vop2(aco_opcode::v_max_f32, bld.def(v1), Operand::c32(0x123456u), neg_a));
//! v1: %res2 = v_mul_f32 %a, %b
//! p_unit_test 2, %res2
Temp neg_neg_a = fneg(neg_a);
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_neg_a, inputs[1]));
//! v1: %res3 = v_mul_f32 |%a|, %b
//! p_unit_test 3, %res3
Temp abs_neg_a = fabs(neg_a);
writeout(3, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_a, inputs[1]));
//! v1: %res4 = v_mul_f32 -|%a|, %b
//! p_unit_test 4, %res4
Temp abs_a = fabs(inputs[0]);
Temp neg_abs_a = fneg(abs_a);
writeout(4, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), neg_abs_a, inputs[1]));
//~gfx9! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1
//~gfx10! v1: %res5 = v_mul_f32 -%a, %b row_shl:1 bound_ctrl:1 fi
//! p_unit_test 5, %res5
writeout(5,
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), neg_a, inputs[1], dpp_row_sl(1)));
//! v1: %res6 = v_subrev_f32 %a, %b
//! p_unit_test 6, %res6
writeout(6, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), neg_a, inputs[1]));
//! v1: %res7 = v_sub_f32 %b, %a
//! p_unit_test 7, %res7
writeout(7, bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[1], neg_a));
//! v1: %res8 = v_mul_f32 %a, -%c
//! p_unit_test 8, %res8
Temp neg_c = fneg(bld.copy(bld.def(v1), inputs[2]));
writeout(8, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], neg_c));
// //! v1: %res9 = v_mul_f32 |%neg_a|, %b
// //! p_unit_test 9, %res9
Temp abs_neg_abs_a = fabs(neg_abs_a);
writeout(9, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), abs_neg_abs_a, inputs[1]));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.output_modifiers)
//>> v1: %a, v1: %b = p_startpgm
if (!setup_cs("v1 v1", GFX9))
return;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
/* 32-bit modifiers */
//! v1: %res0 = v_add_f32 %a, %b *0.5
//! p_unit_test 0, %res0
Temp tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(0, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f000000u), tmp));
//! v1: %res1 = v_add_f32 %a, %b *2
//! p_unit_test 1, %res1
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(1, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
//! v1: %res2 = v_add_f32 %a, %b *4
//! p_unit_test 2, %res2
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40800000u), tmp));
//! v1: %res3 = v_add_f32 %a, %b clamp
//! p_unit_test 3, %res3
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(3, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
Operand::c32(0x3f800000u), tmp));
//! v1: %res4 = v_add_f32 %a, %b *2 clamp
//! p_unit_test 4, %res4
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp);
writeout(4, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
Operand::c32(0x3f800000u), tmp));
/* 16-bit modifiers */
//! v2b: %res5 = v_add_f16 %a, %b *0.5
//! p_unit_test 5, %res5
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(5, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x3800u), tmp));
//! v2b: %res6 = v_add_f16 %a, %b *2
//! p_unit_test 6, %res6
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(6, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
//! v2b: %res7 = v_add_f16 %a, %b *4
//! p_unit_test 7, %res7
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(7, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4400u), tmp));
//! v2b: %res8 = v_add_f16 %a, %b clamp
//! p_unit_test 8, %res8
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(8, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
Operand::c16(0x3c00u), tmp));
//! v2b: %res9 = v_add_f16 %a, %b *2 clamp
//! p_unit_test 9, %res9
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000), tmp);
writeout(9, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
Operand::c16(0x3c00u), tmp));
/* clamping is done after omod */
//! v1: %res10_tmp = v_add_f32 %a, %b clamp
//! v1: %res10 = v_mul_f32 2.0, %res10_tmp
//! p_unit_test 10, %res10
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
tmp = bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(), Operand::c32(0x3f800000u),
tmp);
writeout(10, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
/* unsupported instructions */
//! v1: %res11_tmp = v_xor_b32 %a, %b
//! v1: %res11 = v_mul_f32 2.0, %res11_tmp
//! p_unit_test 11, %res11
tmp = bld.vop2(aco_opcode::v_xor_b32, bld.def(v1), inputs[0], inputs[1]);
writeout(11, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
/* several users */
//! v1: %res12_tmp = v_add_f32 %a, %b
//! p_unit_test %res12_tmp
//! v1: %res12 = v_mul_f32 2.0, %res12_tmp
//! p_unit_test 12, %res12
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
bld.pseudo(aco_opcode::p_unit_test, tmp);
writeout(12, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
//! v1: %res13 = v_add_f32 %a, %b
//! p_unit_test 13, %res13
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp);
writeout(13, tmp);
/* omod has no effect if denormals are enabled but clamp is fine */
//>> BB1
//! /* logical preds: BB0, / linear preds: BB0, / kind: */
program->next_fp_mode.denorm32 = fp_denorm_keep;
program->next_fp_mode.denorm16_64 = fp_denorm_flush;
bld.reset(program->create_and_insert_block());
program->blocks[0].linear_succs.push_back(1);
program->blocks[0].logical_succs.push_back(1);
program->blocks[1].linear_preds.push_back(0);
program->blocks[1].logical_preds.push_back(0);
//! v1: %res14_tmp = v_add_f32 %a, %b
//! v1: %res14 = v_mul_f32 2.0, %res13_tmp
//! p_unit_test 14, %res14
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(14, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
//! v1: %res15 = v_add_f32 %a, %b clamp
//! p_unit_test 15, %res15
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(15, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
Operand::c32(0x3f800000u), tmp));
//>> BB2
//! /* logical preds: BB1, / linear preds: BB1, / kind: */
program->next_fp_mode.denorm32 = fp_denorm_flush;
program->next_fp_mode.denorm16_64 = fp_denorm_keep;
bld.reset(program->create_and_insert_block());
program->blocks[1].linear_succs.push_back(2);
program->blocks[1].logical_succs.push_back(2);
program->blocks[2].linear_preds.push_back(1);
program->blocks[2].logical_preds.push_back(1);
//! v2b: %res16_tmp = v_add_f16 %a, %b
//! v2b: %res16 = v_mul_f16 2.0, %res15_tmp
//! p_unit_test 16, %res16
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(16, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
//! v2b: %res17 = v_add_f16 %a, %b clamp
//! p_unit_test 17, %res17
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(17, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
Operand::c16(0x3c00u), tmp));
/* omod flushes -0.0 to +0.0 */
//>> BB3
//! /* logical preds: BB2, / linear preds: BB2, / kind: uniform, */
program->next_fp_mode.denorm32 = fp_denorm_flush;
program->next_fp_mode.denorm16_64 = fp_denorm_flush;
bld.reset(program->create_and_insert_block());
bld.is_sz_preserve = true;
program->blocks[2].linear_succs.push_back(3);
program->blocks[2].logical_succs.push_back(3);
program->blocks[3].linear_preds.push_back(2);
program->blocks[3].logical_preds.push_back(2);
//! v1: (SzPreserve)%res18_tmp = v_add_f32 %a, %b
//! v1: (SzPreserve)%res18 = v_mul_f32 2.0, %res18_tmp
//! p_unit_test 18, %res18
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(18, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp));
//! v1: (SzPreserve)%res19 = v_add_f32 %a, %b clamp
//! p_unit_test 19, %res19
tmp = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
writeout(19, bld.vop3(aco_opcode::v_med3_f32, bld.def(v1), Operand::zero(),
Operand::c32(0x3f800000u), tmp));
//! v2b: (SzPreserve)%res20_tmp = v_add_f16 %a, %b
//! v2b: (SzPreserve)%res20 = v_mul_f16 2.0, %res20_tmp
//! p_unit_test 20, %res20
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(20, bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp));
//! v2b: (SzPreserve)%res21 = v_add_f16 %a, %b clamp
//! p_unit_test 21, %res21
tmp = bld.vop2(aco_opcode::v_add_f16, bld.def(v2b), inputs[0], inputs[1]);
writeout(21, bld.vop3(aco_opcode::v_med3_f16, bld.def(v2b), Operand::c16(0u),
Operand::c16(0x3c00u), tmp));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.add_lshl)
for (unsigned i = GFX8; i <= GFX10; i++) {
//>> s1: %a, v1: %b = p_startpgm
if (!setup_cs("s1 v1", (amd_gfx_level)i))
continue;
Temp shift;
//~gfx8! s1: %lshl0, s1: %_:scc = s_lshl_b32 %a, 3
//~gfx8! s1: %res0, s1: %_:scc = s_add_u32 %lshl0, 4
//~gfx(9|10)! s1: %res0, s1: %_:scc = s_lshl3_add_u32 %a, 4
//! p_unit_test 0, %res0
shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]),
Operand::c32(3u));
writeout(0, bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift,
Operand::c32(4u)));
//~gfx8! s1: %lshl1, s1: %_:scc = s_lshl_b32 %a, 3
//~gfx8! s1: %add1, s1: %_:scc = s_add_u32 %lshl1, 4
//~gfx8! v1: %add_co1, s2: %_ = v_add_co_u32 %lshl1, %b
//~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %add1, %add_co1
//~gfx(9|10)! s1: %lshl1, s1: %_:scc = s_lshl3_add_u32 %a, 4
//~gfx(9|10)! v1: %lshl_add = v_lshl_add_u32 %a, 3, %b
//~gfx(9|10)! v1: %res1 = v_add_u32 %lshl1, %lshl_add
//! p_unit_test 1, %res1
shift = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), Operand(inputs[0]),
Operand::c32(3u));
Temp sadd =
bld.sop2(aco_opcode::s_add_u32, bld.def(s1), bld.def(s1, scc), shift, Operand::c32(4u));
Temp vadd = bld.vadd32(bld.def(v1), shift, Operand(inputs[1]));
writeout(1, bld.vadd32(bld.def(v1), sadd, vadd));
//~gfx8! s1: %lshl2, s1: %_:scc = s_lshl_b32 %a, 3
//~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b
//~gfx(9|10)! v1: %res2 = v_lshl_add_u32 %a, 3, %b
//! p_unit_test 2, %res2
Temp lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc),
Operand(inputs[0]), Operand::c32(3u));
writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! s1: %lshl3, s1: %_:scc = s_lshl_b32 (is24bit)%a, 7
//~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %lshl3, %b
//~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 7, %b
//! p_unit_test 3, %res3
Operand a_24bit = Operand(inputs[0]);
a_24bit.set24bit(true);
lshl =
bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(7u));
writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//! s1: %lshl4, s1: %_:scc = s_lshl_b32 (is24bit)%a, 3
//~gfx(8|9)! v1: %res4, s2: %carry = v_add_co_u32 %lshl4, %b
//~gfx10! v1: %res4, s2: %carry = v_add_co_u32_e64 %lshl4, %b
//! p_unit_test 4, %carry
lshl =
bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u));
Temp carry = bld.vadd32(bld.def(v1), lshl, Operand(inputs[1]), true).def(1).getTemp();
writeout(4, carry);
//~gfx8! s1: %lshl5, s1: %_:scc = s_lshl_b32 (is24bit)%a, (is24bit)%a
//~gfx8! v1: %res5, s2: %_ = v_add_co_u32 %lshl5, %b
//~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%a, (is24bit)%a, %b
//! p_unit_test 5, %res5
lshl = bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, a_24bit);
writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %res6 = v_mad_u32_u24 (is24bit)%a, 8, %b
//~gfx(9|10)! v1: %res6 = v_lshl_add_u32 (is24bit)%a, 3, %b
//! p_unit_test 6, %res6
lshl =
bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_24bit, Operand::c32(3u));
writeout(6, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %res7 = v_mad_u32_u24 (is16bit)%a, 16, %b
//~gfx(9|10)! v1: %res7 = v_lshl_add_u32 (is16bit)%a, 4, %b
//! p_unit_test 7, %res7
Operand a_16bit = Operand(inputs[0]);
a_16bit.set16bit(true);
lshl =
bld.sop2(aco_opcode::s_lshl_b32, bld.def(s1), bld.def(s1, scc), a_16bit, Operand::c32(4u));
writeout(7, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.bcnt)
for (unsigned i = GFX8; i <= GFX10; i++) {
//>> v1: %a, s1: %b = p_startpgm
if (!setup_cs("v1 s1", (amd_gfx_level)i))
continue;
Temp bcnt;
//! v1: %res0 = v_bcnt_u32_b32 %a, %a
//! p_unit_test 0, %res0
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
writeout(0, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0])));
//! v1: %res1 = v_bcnt_u32_b32 %a, %b
//! p_unit_test 1, %res1
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
writeout(1, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[1])));
//! v1: %res2 = v_bcnt_u32_b32 %a, 42
//! p_unit_test 2, %res2
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
writeout(2, bld.vadd32(bld.def(v1), bcnt, Operand::c32(42u)));
//! v1: %bnct3 = v_bcnt_u32_b32 %b, 0
//~gfx8! v1: %res3, s2: %_ = v_add_co_u32 %bcnt3, %a
//~gfx(9|10)! v1: %res3 = v_add_u32 %bcnt3, %a
//! p_unit_test 3, %res3
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[1]), Operand::zero());
writeout(3, bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0])));
//! v1: %bnct4 = v_bcnt_u32_b32 %a, 0
//~gfx(8|9)! v1: %add4, s2: %carry = v_add_co_u32 %bcnt4, %a
//~gfx10! v1: %add4, s2: %carry = v_add_co_u32_e64 %bcnt4, %a
//! p_unit_test 4, %carry
bcnt = bld.vop3(aco_opcode::v_bcnt_u32_b32, bld.def(v1), Operand(inputs[0]), Operand::zero());
Temp carry = bld.vadd32(bld.def(v1), bcnt, Operand(inputs[0]), true).def(1).getTemp();
writeout(4, carry);
finish_opt_test();
}
END_TEST
struct clamp_config {
const char* name;
aco_opcode min, max, med3;
Operand lb, ub;
};
static const clamp_config clamp_configs[] = {
/* 0.0, 4.0 */
{"_0,4f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32,
Operand::zero(), Operand::c32(0x40800000u)},
{"_0,4f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16,
Operand::c16(0u), Operand::c16(0x4400)},
/* -1.0, 0.0 */
{"_-1,0f32", aco_opcode::v_min_f32, aco_opcode::v_max_f32, aco_opcode::v_med3_f32,
Operand::c32(0xbf800000u), Operand::zero()},
{"_-1,0f16", aco_opcode::v_min_f16, aco_opcode::v_max_f16, aco_opcode::v_med3_f16,
Operand::c16(0xBC00), Operand::c16(0u)},
/* 0, 3 */
{"_0,3u32", aco_opcode::v_min_u32, aco_opcode::v_max_u32, aco_opcode::v_med3_u32,
Operand::zero(), Operand::c32(3u)},
{"_0,3u16", aco_opcode::v_min_u16, aco_opcode::v_max_u16, aco_opcode::v_med3_u16,
Operand::c16(0u), Operand::c16(3u)},
{"_0,3i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32,
Operand::zero(), Operand::c32(3u)},
{"_0,3i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16,
Operand::c16(0u), Operand::c16(3u)},
/* -5, 0 */
{"_-5,0i32", aco_opcode::v_min_i32, aco_opcode::v_max_i32, aco_opcode::v_med3_i32,
Operand::c32(0xfffffffbu), Operand::zero()},
{"_-5,0i16", aco_opcode::v_min_i16, aco_opcode::v_max_i16, aco_opcode::v_med3_i16,
Operand::c16(0xfffbu), Operand::c16(0u)},
};
BEGIN_TEST(optimize.clamp)
for (clamp_config cfg : clamp_configs) {
if (!setup_cs("v1 v1 v1", GFX9, CHIP_UNKNOWN, cfg.name))
continue;
//! cfg: @match_func(min max med3 lb ub)
fprintf(output, "cfg: %s ", instr_info.name[(int)cfg.min]);
fprintf(output, "%s ", instr_info.name[(int)cfg.max]);
fprintf(output, "%s ", instr_info.name[(int)cfg.med3]);
aco_print_operand(&cfg.lb, output);
fprintf(output, " ");
aco_print_operand(&cfg.ub, output);
fprintf(output, "\n");
//>> v1: %a, v1: %b, v1: %c = p_startpgm
//! v1: %res0 = @med3 @ub, @lb, %a
//! p_unit_test 0, %res0
writeout(0, bld.vop2(cfg.min, bld.def(v1), cfg.ub,
bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0])));
//! v1: %res1 = @med3 @lb, @ub, %a
//! p_unit_test 1, %res1
writeout(1, bld.vop2(cfg.max, bld.def(v1), cfg.lb,
bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0])));
/* min constant must be greater than max constant */
//! v1: %res2_tmp = @min @lb, %a
//! v1: %res2 = @max @ub, %res2_tmp
//! p_unit_test 2, %res2
writeout(2, bld.vop2(cfg.max, bld.def(v1), cfg.ub,
bld.vop2(cfg.min, bld.def(v1), cfg.lb, inputs[0])));
//! v1: %res3_tmp = @max @ub, %a
//! v1: %res3 = @min @lb, %res3_tmp
//! p_unit_test 3, %res3
writeout(3, bld.vop2(cfg.min, bld.def(v1), cfg.lb,
bld.vop2(cfg.max, bld.def(v1), cfg.ub, inputs[0])));
/* needs two constants */
//! v1: %res4_tmp = @max @lb, %a
//! v1: %res4 = @min %b, %res4_tmp
//! p_unit_test 4, %res4
writeout(4, bld.vop2(cfg.min, bld.def(v1), inputs[1],
bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0])));
//! v1: %res5_tmp = @max %b, %a
//! v1: %res5 = @min @ub, %res5_tmp
//! p_unit_test 5, %res5
writeout(5, bld.vop2(cfg.min, bld.def(v1), cfg.ub,
bld.vop2(cfg.max, bld.def(v1), inputs[1], inputs[0])));
//! v1: %res6_tmp = @max %c, %a
//! v1: %res6 = @min %b, %res6_tmp
//! p_unit_test 6, %res6
writeout(6, bld.vop2(cfg.min, bld.def(v1), inputs[1],
bld.vop2(cfg.max, bld.def(v1), inputs[2], inputs[0])));
/* correct NaN behaviour with precise */
if (cfg.min == aco_opcode::v_min_f16 || cfg.min == aco_opcode::v_min_f32) {
//~f(16|32)! v1: %res7 = @med3 @ub, @lb, %a
//~f(16|32)! p_unit_test 7, %res7
Builder::Result max = bld.vop2(cfg.max, bld.def(v1), cfg.lb, inputs[0]);
max.def(0).setPrecise(true);
Builder::Result min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, max);
max.def(0).setPrecise(true);
writeout(7, min);
//~f(16|32)! v1: (precise)%res8_tmp = @min @ub, %a
//~f(16|32)! v1: %res8 = @max @lb, %res8_tmp
//~f(16|32)! p_unit_test 8, %res8
min = bld.vop2(cfg.min, bld.def(v1), cfg.ub, inputs[0]);
min.def(0).setPrecise(true);
writeout(8, bld.vop2(cfg.max, bld.def(v1), cfg.lb, min));
}
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.add3)
//>> v1: %a, v1: %b, v1: %c = p_startpgm
if (!setup_cs("v1 v1 v1", GFX9))
return;
//! v1: %res0 = v_add3_u32 %a, %b, %c
//! p_unit_test 0, %res0
Builder::Result tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
writeout(0, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp));
//! v1: %tmp1 = v_add_u32 %b, %c clamp
//! v1: %res1 = v_add_u32 %a, %tmp1
//! p_unit_test 1, %res1
tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
tmp->valu().clamp = true;
writeout(1, bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp));
//! v1: %tmp2 = v_add_u32 %b, %c
//! v1: %res2 = v_add_u32 %a, %tmp2 clamp
//! p_unit_test 2, %res2
tmp = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), inputs[1], inputs[2]);
tmp = bld.vop2_e64(aco_opcode::v_add_u32, bld.def(v1), inputs[0], tmp);
tmp->valu().clamp = true;
writeout(2, tmp);
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.minmax)
for (unsigned i = GFX10_3; i <= GFX11; i++) {
//>> v1: %a, v1: %b, v1: %c = p_startpgm
if (!setup_cs("v1 v1 v1", (amd_gfx_level)i))
continue;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
//! v1: %res0 = v_min3_f32 %a, %b, %c
//! p_unit_test 0, %res0
writeout(0, fmin(c, fmin(a, b)));
//! v1: %res1 = v_max3_f32 %a, %b, %c
//! p_unit_test 1, %res1
writeout(1, fmax(c, fmax(a, b)));
//! v1: %res2 = v_min3_f32 -%a, -%b, %c
//! p_unit_test 2, %res2
writeout(2, fmin(c, fneg(fmax(a, b))));
//! v1: %res3 = v_max3_f32 -%a, -%b, %c
//! p_unit_test 3, %res3
writeout(3, fmax(c, fneg(fmin(a, b))));
//! v1: %res4 = v_max3_f32 -%a, %b, %c
//! p_unit_test 4, %res4
writeout(4, fmax(c, fneg(fmin(a, fneg(b)))));
//~gfx10_3! v1: %res5_tmp = v_max_f32 %a, %b
//~gfx10_3! v1: %res5 = v_min_f32 %c, %res5_tmp
//~gfx11! v1: %res5 = v_maxmin_f32 %a, %b, %c
//! p_unit_test 5, %res5
writeout(5, fmin(c, fmax(a, b)));
//~gfx10_3! v1: %res6_tmp = v_min_f32 %a, %b
//~gfx10_3! v1: %res6 = v_max_f32 %c, %res6_tmp
//~gfx11! v1: %res6 = v_minmax_f32 %a, %b, %c
//! p_unit_test 6, %res6
writeout(6, fmax(c, fmin(a, b)));
//~gfx10_3! v1: %res7_tmp = v_min_f32 %a, %b
//~gfx10_3! v1: %res7 = v_min_f32 %c, -%res7_tmp
//~gfx11! v1: %res7 = v_maxmin_f32 -%a, -%b, %c
//! p_unit_test 7, %res7
writeout(7, fmin(c, fneg(fmin(a, b))));
//~gfx10_3! v1: %res8_tmp = v_max_f32 %a, %b
//~gfx10_3! v1: %res8 = v_max_f32 %c, -%res8_tmp
//~gfx11! v1: %res8 = v_minmax_f32 -%a, -%b, %c
//! p_unit_test 8, %res8
writeout(8, fmax(c, fneg(fmax(a, b))));
//~gfx10_3! v1: %res9_tmp = v_max_f32 %a, -%b
//~gfx10_3! v1: %res9 = v_max_f32 %c, -%res9_tmp
//~gfx11! v1: %res9 = v_minmax_f32 -%a, %b, %c
//! p_unit_test 9, %res9
writeout(9, fmax(c, fneg(fmax(a, fneg(b)))));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_32_24)
for (unsigned i = GFX8; i <= GFX9; i++) {
//>> v1: %a, v1: %b, v1: %c = p_startpgm
if (!setup_cs("v1 v1 v1", (amd_gfx_level)i))
continue;
//! v1: %res0 = v_mad_u32_u24 %b, %c, %a
//! p_unit_test 0, %res0
Temp mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]);
writeout(0, bld.vadd32(bld.def(v1), inputs[0], mul));
//! v1: %res1_tmp = v_mul_u32_u24 %b, %c
//! v1: %_, s2: %res1 = v_add_co_u32 %a, %res1_tmp
//! p_unit_test 1, %res1
mul = bld.vop2(aco_opcode::v_mul_u32_u24, bld.def(v1), inputs[1], inputs[2]);
writeout(1, bld.vadd32(bld.def(v1), inputs[0], mul, true).def(1).getTemp());
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.add_lshlrev)
for (unsigned i = GFX8; i <= GFX10; i++) {
//>> v1: %a, v1: %b, s1: %c = p_startpgm
if (!setup_cs("v1 v1 s1", (amd_gfx_level)i))
continue;
Temp lshl;
//~gfx8! v1: %lshl0 = v_lshlrev_b32 3, %a
//~gfx8! v1: %res0, s2: %_ = v_add_co_u32 %lshl0, %b
//~gfx(9|10)! v1: %res0 = v_lshl_add_u32 %a, 3, %b
//! p_unit_test 0, %res0
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), Operand(inputs[0]));
writeout(0, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %lshl1 = v_lshlrev_b32 7, (is24bit)%a
//~gfx8! v1: %res1, s2: %_ = v_add_co_u32 %lshl1, %b
//~gfx(9|10)! v1: %res1 = v_lshl_add_u32 (is24bit)%a, 7, %b
//! p_unit_test 1, %res1
Operand a_24bit = Operand(inputs[0]);
a_24bit.set24bit(true);
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(7u), a_24bit);
writeout(1, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %lshl2 = v_lshlrev_b32 (is24bit)%a, (is24bit)%b
//~gfx8! v1: %res2, s2: %_ = v_add_co_u32 %lshl2, %b
//~gfx(9|10)! v1: %res2 = v_lshl_add_u32 (is24bit)%b, (is24bit)%a, %b
//! p_unit_test 2, %res2
Operand b_24bit = Operand(inputs[1]);
b_24bit.set24bit(true);
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), a_24bit, b_24bit);
writeout(2, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %res3 = v_mad_u32_u24 (is24bit)%a, 8, %b
//~gfx(9|10)! v1: %res3 = v_lshl_add_u32 (is24bit)%a, 3, %b
//! p_unit_test 3, %res3
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(3u), a_24bit);
writeout(3, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %res4 = v_mad_u32_u24 (is16bit)%a, 16, %b
//~gfx(9|10)! v1: %res4 = v_lshl_add_u32 (is16bit)%a, 4, %b
//! p_unit_test 4, %res4
Operand a_16bit = Operand(inputs[0]);
a_16bit.set16bit(true);
lshl = bld.vop2(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), a_16bit);
writeout(4, bld.vadd32(bld.def(v1), lshl, Operand(inputs[1])));
//~gfx8! v1: %res5 = v_mad_u32_u24 (is24bit)%c, 16, %c
//~gfx(9|10)! v1: %res5 = v_lshl_add_u32 (is24bit)%c, 4, %c
//! p_unit_test 5, %res5
Operand c_24bit = Operand(inputs[2]);
c_24bit.set24bit(true);
lshl = bld.vop2_e64(aco_opcode::v_lshlrev_b32, bld.def(v1), Operand::c32(4u), c_24bit);
writeout(5, bld.vadd32(bld.def(v1), lshl, Operand(inputs[2])));
finish_opt_test();
}
END_TEST
enum denorm_op {
denorm_mul1 = 0,
denorm_fneg = 1,
denorm_fabs = 2,
denorm_fnegabs = 3,
};
static const char* denorm_op_names[] = {
"mul1",
"fneg",
"fabs",
"fnegabs",
};
struct denorm_config {
bool flush;
unsigned op;
aco_opcode src;
aco_opcode dest;
};
static const char*
srcdest_op_name(aco_opcode op)
{
switch (op) {
case aco_opcode::v_cndmask_b32: return "cndmask";
case aco_opcode::v_min_f32: return "min";
case aco_opcode::v_rcp_f32: return "rcp";
default: return "none";
}
}
static Temp
emit_denorm_srcdest(aco_opcode op, Temp val)
{
switch (op) {
case aco_opcode::v_cndmask_b32:
return bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]);
case aco_opcode::v_min_f32:
return bld.vop2(aco_opcode::v_min_f32, bld.def(v1), Operand::zero(), val);
case aco_opcode::v_rcp_f32: return bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), val);
default: return val;
}
}
BEGIN_TEST(optimize.denorm_propagation)
for (unsigned i = GFX8; i <= GFX9; i++) {
std::vector<denorm_config> configs;
for (bool flush : {false, true}) {
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
configs.push_back({flush, op, aco_opcode::num_opcodes, aco_opcode::num_opcodes});
for (aco_opcode dest : {aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) {
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
configs.push_back({flush, op, aco_opcode::num_opcodes, dest});
}
for (aco_opcode src :
{aco_opcode::v_cndmask_b32, aco_opcode::v_min_f32, aco_opcode::v_rcp_f32}) {
for (denorm_op op : {denorm_mul1, denorm_fneg, denorm_fabs, denorm_fnegabs})
configs.push_back({flush, op, src, aco_opcode::num_opcodes});
}
}
for (denorm_config cfg : configs) {
char subvariant[128];
sprintf(subvariant, "_%s_%s_%s_%s", cfg.flush ? "flush" : "keep", srcdest_op_name(cfg.src),
denorm_op_names[(int)cfg.op], srcdest_op_name(cfg.dest));
if (!setup_cs("v1 s2", (amd_gfx_level)i, CHIP_UNKNOWN, subvariant))
continue;
bool can_propagate = cfg.src == aco_opcode::v_rcp_f32 ||
(i >= GFX9 && cfg.src == aco_opcode::v_min_f32) ||
cfg.dest == aco_opcode::v_rcp_f32 ||
(i >= GFX9 && cfg.dest == aco_opcode::v_min_f32) || !cfg.flush;
fprintf(output, "src, dest, op: %s %s %s\n", srcdest_op_name(cfg.src),
srcdest_op_name(cfg.dest), denorm_op_names[(int)cfg.op]);
fprintf(output, "can_propagate: %u\n", can_propagate);
//! src, dest, op: $src $dest $op
//! can_propagate: #can_propagate
//>> v1: %a, s2: %b = p_startpgm
//; patterns = {'cndmask': 'v1: %{} = v_cndmask_b32 0, {}, %b',
//; 'min': 'v1: %{} = v_min_f32 0, {}',
//; 'rcp': 'v1: %{} = v_rcp_f32 {}'}
//; ops = {'mul1': 'v1: %{} = v_mul_f32 1.0, %{}',
//; 'fneg': 'v1: %{} = v_mul_f32 -1.0, %{}',
//; 'fabs': 'v1: %{} = v_mul_f32 1.0, |%{}|',
//; 'fnegabs': 'v1: %{} = v_mul_f32 -1.0, |%{}|'}
//; inline_ops = {'mul1': '%{}', 'fneg': '-%{}', 'fabs': '|%{}|', 'fnegabs': '-|%{}|'}
//; name = 'a'
//; if src != 'none':
//; insert_pattern(patterns[src].format('src_res', '%'+name))
//; name = 'src_res'
//; if can_propagate:
//; name = inline_ops[op].format(name)
//; else:
//; insert_pattern(ops[op].format('op_res', name))
//; name = '%op_res'
//; if dest != 'none':
//; insert_pattern(patterns[dest].format('dest_res', name))
//; name = '%dest_res'
//; insert_pattern('v1: %res = v_cndmask_b32 0, {}, %b'.format(name))
//! p_unit_test 0, %res
program->blocks[0].fp_mode.denorm32 = cfg.flush ? fp_denorm_flush : fp_denorm_keep;
Temp val = emit_denorm_srcdest(cfg.src, inputs[0]);
switch (cfg.op) {
case denorm_mul1:
val = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x3f800000u), val);
break;
case denorm_fneg: val = fneg(val); break;
case denorm_fabs: val = fabs(val); break;
case denorm_fnegabs: val = fneg(fabs(val)); break;
}
val = emit_denorm_srcdest(cfg.dest, val);
writeout(
0, bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), Operand::zero(), val, inputs[1]));
finish_opt_test();
}
}
END_TEST
BEGIN_TEST(optimizer.dpp)
//>> v1: %a, v1: %b, s2: %c, s1: %d = p_startpgm
if (!setup_cs("v1 v1 s2 s1", GFX10_3))
return;
Operand a(inputs[0]);
Operand b(inputs[1]);
Operand c(inputs[2]);
Operand d(inputs[3]);
/* basic optimization */
//! v1: %res0 = v_add_f32 %a, %b row_mirror bound_ctrl:1 fi
//! p_unit_test 0, %res0
Temp tmp0 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), tmp0, b);
writeout(0, res0);
/* operand swapping */
//! v1: %res1 = v_subrev_f32 %a, %b row_mirror bound_ctrl:1 fi
//! p_unit_test 1, %res1
Temp tmp1 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res1 = bld.vop2(aco_opcode::v_sub_f32, bld.def(v1), b, tmp1);
writeout(1, res1);
//! v1: %tmp2 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
//! v1: %res2 = v_sub_f32 %b, %tmp2 row_half_mirror bound_ctrl:1 fi
//! p_unit_test 2, %res2
Temp tmp2 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res2 = bld.vop2_dpp(aco_opcode::v_sub_f32, bld.def(v1), b, tmp2, dpp_row_half_mirror);
writeout(2, res2);
/* modifiers */
//! v1: %res3 = v_max_f32 -%a, %b row_mirror bound_ctrl:1 fi
//! p_unit_test 3, %res3
auto tmp3 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
tmp3->dpp16().neg[0] = true;
Temp res3 = bld.vop2(aco_opcode::v_max_f32, bld.def(v1), tmp3, b);
writeout(3, res3);
//! v1: %res4 = v_max_f32 -%a, %b row_mirror bound_ctrl:1 fi
//! p_unit_test 4, %res4
Temp tmp4 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
auto res4 = bld.vop2_e64(aco_opcode::v_max_f32, bld.def(v1), tmp4, b);
res4->valu().neg[0] = true;
writeout(4, res4);
//! v1: %tmp5 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
//! v1: %res5 = v_add_f32 %tmp5, %b clamp
//! p_unit_test 5, %res5
Temp tmp5 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
auto res5 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp5, b);
res5->valu().clamp = true;
writeout(5, res5);
//! v1: %res6 = v_add_f32 |%a|, %b row_mirror bound_ctrl:1 fi
//! p_unit_test 6, %res6
auto tmp6 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
tmp6->dpp16().neg[0] = true;
auto res6 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp6, b);
res6->valu().abs[0] = true;
writeout(6, res6);
//! v1: %res7 = v_subrev_f32 %a, |%b| row_mirror bound_ctrl:1 fi
//! p_unit_test 7, %res7
Temp tmp7 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
auto res7 = bld.vop2_e64(aco_opcode::v_sub_f32, bld.def(v1), b, tmp7);
res7->valu().abs[0] = true;
writeout(7, res7);
//! v1: %tmp11 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi
//! v1: %res11 = v_add_u32 %tmp11, %b
//! p_unit_test 11, %res11
auto tmp11 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
tmp11->dpp16().neg[0] = true;
Temp res11 = bld.vop2(aco_opcode::v_add_u32, bld.def(v1), tmp11, b);
writeout(11, res11);
//! v1: %tmp12 = v_mov_b32 -%a row_mirror bound_ctrl:1 fi
//! v1: %res12 = v_add_f16 %tmp12, %b
//! p_unit_test 12, %res12
auto tmp12 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
tmp12->dpp16().neg[0] = true;
Temp res12 = bld.vop2(aco_opcode::v_add_f16, bld.def(v1), tmp12, b);
writeout(12, res12);
/* vcc */
//! v1: %res8 = v_cndmask_b32 %a, %b, %c:vcc row_mirror bound_ctrl:1 fi
//! p_unit_test 8, %res8
Temp tmp8 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res8 = bld.vop2(aco_opcode::v_cndmask_b32, bld.def(v1), tmp8, b, c);
writeout(8, res8);
/* sgprs */
//! v1: %tmp9 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
//! v1: %res9 = v_add_f32 %tmp9, %d
//! p_unit_test 9, %res9
Temp tmp9 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res9 = bld.vop2_e64(aco_opcode::v_add_f32, bld.def(v1), tmp9, d);
writeout(9, res9);
//! v1: %tmp10 = v_mov_b32 %a row_mirror bound_ctrl:1 fi
//! v1: %res10 = v_add_f32 %d, %tmp10
//! p_unit_test 10, %res10
Temp tmp10 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp res10 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), d, tmp10);
writeout(10, res10);
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.dpp_prop)
//>> v1: %a, s1: %b = p_startpgm
if (!setup_cs("v1 s1", GFX10))
return;
//! v1: %one = p_parallelcopy 1
//! v1: %res0 = v_mul_f32 1, %a
//! p_unit_test 0, %res0
Temp one = bld.copy(bld.def(v1), Operand::c32(1));
writeout(0, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), one, inputs[0], dpp_row_rr(1)));
//! v1: %res1 = v_mul_f32 %a, %one row_ror:1 bound_ctrl:1 fi
//! p_unit_test 1, %res1
writeout(1, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], one, dpp_row_rr(1)));
//! v1: %res2 = v_mul_f32 0x12345678, %a
//! p_unit_test 2, %res2
Temp literal1 = bld.copy(bld.def(v1), Operand::c32(0x12345678u));
writeout(2,
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), literal1, inputs[0], dpp_row_rr(1)));
//! v1: %literal2 = p_parallelcopy 0x12345679
//! v1: %res3 = v_mul_f32 %a, %literal row_ror:1 bound_ctrl:1 fi
//! p_unit_test 3, %res3
Temp literal2 = bld.copy(bld.def(v1), Operand::c32(0x12345679u));
writeout(3,
bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], literal2, dpp_row_rr(1)));
//! v1: %b_v = p_parallelcopy %b
//! v1: %res4 = v_mul_f32 %b, %a
//! p_unit_test 4, %res4
Temp b_v = bld.copy(bld.def(v1), inputs[1]);
writeout(4, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), b_v, inputs[0], dpp_row_rr(1)));
//! v1: %res5 = v_mul_f32 %a, %b_v row_ror:1 bound_ctrl:1 fi
//! p_unit_test 5, %res5
writeout(5, bld.vop2_dpp(aco_opcode::v_mul_f32, bld.def(v1), inputs[0], b_v, dpp_row_rr(1)));
//! v1: %res6 = v_rcp_f32 %b
//! p_unit_test 6, %res6
writeout(6, bld.vop1_dpp(aco_opcode::v_rcp_f32, bld.def(v1), b_v, dpp_row_rr(1)));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.casts)
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", GFX10_3))
return;
Temp a = inputs[0];
Temp a16 = inputs[1];
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
//! v1: %res0_tmp = v_mul_f32 -1.0, %a
//! v2b: %res0 = v_mul_f16 %res0_tmp, %a16
//! p_unit_test 0, %res0
writeout(0, fmul(u2u16(fneg(a)), a16));
//! v2b: %res1_tmp = v_mul_f16 -1.0, %a16
//! v1: %res1 = v_mul_f32 %res1_tmp, %a
//! p_unit_test 1, %res1
writeout(1, fmul(bld.as_uniform(fneg(a16)), a));
//! v1: %res2_tmp = v_mul_f32 -1.0, %a16
//! v2b: %res2 = v_mul_f16 %res2_tmp, %a16
//! p_unit_test 2, %res2
writeout(2, fmul(u2u16(bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1),
Operand::c32(0xbf800000u), bld.as_uniform(a16))),
a16));
//! v1: %res3_tmp = v_mul_f32 %a, %a
//! v2b: %res3 = v_add_f16 %res3_tmp, 0 clamp
//! p_unit_test 3, %res3
writeout(3, fsat(u2u16(fmul(a, a))));
//! v2b: %res4_tmp = v_mul_f16 %a16, %a16
//! v1: %res4 = v_add_f32 %res4_tmp, 0 clamp
//! p_unit_test 4, %res4
writeout(4, fsat(bld.as_uniform(fmul(a16, a16))));
//! v1: %res5_tmp = v_mul_f32 %a, %a
//! v2b: %res5 = v_mul_f16 2.0, %res5_tmp
//! p_unit_test 5, %res5
writeout(5, fmul(u2u16(fmul(a, a)), bld.copy(bld.def(v2b), Operand::c16(0x4000))));
//! v2b: %res6_tmp = v_mul_f16 %a16, %a16
//! v1: %res6 = v_mul_f32 2.0, %res6_tmp
//! p_unit_test 6, %res6
writeout(6,
fmul(bld.as_uniform(fmul(a16, a16)), bld.copy(bld.def(v1), Operand::c32(0x40000000))));
//! v1: %res7_tmp = v_mul_f32 %a, %a
//! v2b: %res7 = v_add_f16 %res7_tmp, %a16
//! p_unit_test 7, %res7
writeout(7, fadd(u2u16(fmul(a, a)), a16));
//! v2b: %res8_tmp = v_mul_f16 %a16, %a16
//! v1: %res8 = v_add_f32 %res8_tmp, %a
//! p_unit_test 8, %res8
writeout(8, fadd(bld.as_uniform(fmul(a16, a16)), a));
//! v1: %res9_tmp = v_mul_f32 %a, %a
//! v2b: %res9 = v_mul_f16 -1.0, %res9_tmp
//! p_unit_test 9, %res9
writeout(9, fneg(u2u16(fmul(a, a))));
//! v2b: %res10_tmp = v_mul_f16 %a16, %a16
//! v1: %res10 = v_mul_f32 -1.0, %res10_tmp
//! p_unit_test 10, %res10
writeout(10, bld.vop2_e64(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0xbf800000u),
bld.as_uniform(fmul(a16, a16))));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.mad_mix.input_conv.basic)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp a16 = inputs[1];
//! v1: %res0 = v_fma_mix_f32 %a, lo(%a16), neg(0)
//! p_unit_test 0, %res0
writeout(0, fmul(a, f2f32(a16)));
//! v1: %res1 = v_fma_mix_f32 1.0, lo(%a16), %a
//! p_unit_test 1, %res1
writeout(1, fadd(a, f2f32(a16)));
//! v1: %res2 = v_fma_mix_f32 1.0, lo(%a16), %a
//! p_unit_test 2, %res2
writeout(2, fadd(f2f32(a16), a));
//! v1: %res3 = v_fma_mix_f32 %a, %a, lo(%a16)
//! p_unit_test 3, %res3
writeout(3, fma(a, a, f2f32(a16)));
//! v1: %res4 = v_fma_mix_f32 %a, %a, lo(%a16)
//! p_unit_test 4, %res4
writeout(4, fma(a, a, f2f32(a16)));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.input_conv.precision)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp a16 = inputs[1];
/* precise arithmetic */
//~gfx9! v1: %res0_cvt = v_cvt_f32_f16 %a16
//~gfx9! v1: (precise)%res0 = v_fma_f32 %a, %a, %res0_cvt
//~gfx10! v1: (precise)%res0 = v_fma_mix_f32 %a, %a, lo(%a16)
//! p_unit_test 0, %res0
writeout(0, fma(a, a, f2f32(a16), bld.precise()));
//! v2b: %res1_cvt = v_cvt_f16_f32 %a
//! v2b: (precise)%res1 = v_mul_f16 %a16, %res1_cvt
//! p_unit_test 1, %res1
writeout(1, fmul(a16, f2f16(a), bld.precise()));
//! v2b: %res2_cvt = v_cvt_f16_f32 %a
//! v2b: (precise)%res2 = v_add_f16 %a16, %res2_cvt
//! p_unit_test 2, %res2
writeout(2, fadd(a16, f2f16(a), bld.precise()));
//! v2b: %res3_cvt = v_cvt_f16_f32 %a
//! v2b: (precise)%res3 = v_fma_f16 %a16, %a16, %res3_cvt
//! p_unit_test 3, %res3
writeout(3, fma(a16, a16, f2f16(a), bld.precise()));
/* precise conversions */
//! v2b: (precise)%res4_cvt = v_cvt_f16_f32 %a
//! v2b: %res4 = v_mul_f16 %a16, %res4_cvt
//! p_unit_test 4, %res4
writeout(4, fmul(a16, f2f16(a, bld.precise())));
//! v2b: (precise)%res5_cvt = v_cvt_f16_f32 %a
//! v2b: %res5 = v_add_f16 %a16, %res5_cvt
//! p_unit_test 5, %res5
writeout(5, fadd(a16, f2f16(a, bld.precise())));
//! v2b: (precise)%res6_cvt = v_cvt_f16_f32 %a
//! v2b: %res6 = v_fma_f16 %a16, %a16, %res6_cvt
//! p_unit_test 6, %res6
writeout(6, fma(a16, a16, f2f16(a, bld.precise())));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.input_conv.modifiers)
for (unsigned i = GFX9; i <= GFX11; i++) {
if (i == GFX10_3)
continue;
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp a16 = inputs[1];
/* check whether modifiers are preserved when converting to VOP3P */
//! v1: %res0 = v_fma_mix_f32 -%a, lo(%a16), neg(0)
//! p_unit_test 0, %res0
writeout(0, fmul(fneg(a), f2f32(a16)));
//! v1: %res1 = v_fma_mix_f32 |%a|, lo(%a16), neg(0)
//! p_unit_test 1, %res1
writeout(1, fmul(fabs(a), f2f32(a16)));
/* fneg modifiers */
//! v1: %res2 = v_fma_mix_f32 %a, -lo(%a16), neg(0)
//! p_unit_test 2, %res2
writeout(2, fmul(a, fneg(f2f32(a16))));
//! v1: %res3 = v_fma_mix_f32 %a, -lo(%a16), neg(0)
//! p_unit_test 3, %res3
writeout(3, fmul(a, f2f32(fneg(a16))));
/* fabs modifiers */
//! v1: %res4 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0)
//! p_unit_test 4, %res4
writeout(4, fmul(a, fabs(f2f32(a16))));
//! v1: %res5 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0)
//! p_unit_test 5, %res5
writeout(5, fmul(a, f2f32(fabs(a16))));
/* both fabs and fneg modifiers */
//! v1: %res6 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0)
//! p_unit_test 6, %res6
writeout(6, fmul(a, fneg(f2f32(fabs(a16)))));
//! v1: %res7 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0)
//! p_unit_test 7, %res7
writeout(7, fmul(a, fabs(f2f32(fabs(a16)))));
//! v1: %res8 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0)
//! p_unit_test 8, %res8
writeout(8, fmul(a, fneg(fabs(f2f32(fabs(a16))))));
//! v1: %res9 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0)
//! p_unit_test 9, %res9
writeout(9, fmul(a, f2f32(fneg(fabs(a16)))));
//! v1: %res10 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0)
//! p_unit_test 10, %res10
writeout(10, fmul(a, fneg(f2f32(fneg(fabs(a16))))));
//! v1: %res11 = v_fma_mix_f32 %a, |lo(%a16)|, neg(0)
//! p_unit_test 11, %res11
writeout(11, fmul(a, fabs(f2f32(fneg(fabs(a16))))));
//! v1: %res12 = v_fma_mix_f32 %a, -|lo(%a16)|, neg(0)
//! p_unit_test 12, %res12
writeout(12, fmul(a, fneg(fabs(f2f32(fneg(fabs(a16)))))));
/* sdwa */
//! v1: %res13 = v_fma_mix_f32 lo(%a), %a, neg(0)
//! p_unit_test 13, %res13
writeout(13, fmul(f2f32(ext_ushort(a, 0)), a));
//! v1: %res14 = v_fma_mix_f32 hi(%a), %a, neg(0)
//! p_unit_test 14, %res14
writeout(14, fmul(f2f32(ext_ushort(a, 1)), a));
//~gfx(9|10)! v1: %res15_cvt = v_cvt_f32_f16 %a dst_sel:uword0 src0_sel:dword
//~gfx11! v1: %res16_cvt1 = v_fma_mix_f32 lo(%a), 1.0, neg(0)
//~gfx11! v1: %res15_cvt = p_extract %res16_cvt1, 0, 16, 0
//! v1: %res15 = v_mul_f32 %res15_cvt, %a
//! p_unit_test 15, %res15
writeout(15, fmul(ext_ushort(f2f32(a), 0), a));
//~gfx(9|10)! v1: %res16_cvt = v_cvt_f32_f16 %a
//~gfx(9|10)! v1: %res16 = v_mul_f32 %res16_cvt, %a dst_sel:dword src0_sel:uword1 src1_sel:dword
//~gfx11! v1: %res16_cvt = v_fma_mix_f32 lo(%a), 1.0, neg(0)
//~gfx11! v1: %res16_ext = p_extract %res16_cvt, 1, 16, 0
//~gfx11! v1: %res16 = v_mul_f32 %res16_ext, %a
//! p_unit_test 16, %res16
writeout(16, fmul(ext_ushort(f2f32(a), 1), a));
//~gfx(9|10)! v1: %res17_cvt = v_cvt_f32_f16 %a dst_sel:dword src0_sel:ubyte2
//~gfx(9|10)! v1: %res17 = v_mul_f32 %res17_cvt, %a
//~gfx11! v1: %res17_ext = p_extract %a, 2, 8, 0
//~gfx11! v1: %res17 = v_fma_mix_f32 lo(%res17_ext), %a, neg(0)
//! p_unit_test 17, %res17
writeout(17, fmul(f2f32(ext_ubyte(a, 2)), a));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.output_conv.basic)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
Temp a16 = inputs[3];
Temp b16 = inputs[4];
//! v2b: %res0 = v_fma_mixlo_f16 %a, %b, neg(lo(0))
//! p_unit_test 0, %res0
writeout(0, f2f16(fmul(a, b)));
//! v2b: %res1 = v_fma_mixlo_f16 1.0, %a, %b
//! p_unit_test 1, %res1
writeout(1, f2f16(fadd(a, b)));
//! v2b: %res2 = v_fma_mixlo_f16 %a, %b, %c
//! p_unit_test 2, %res2
writeout(2, f2f16(fma(a, b, c)));
//! v2b: %res3 = v_fma_mixlo_f16 lo(%a16), %b, neg(lo(0))
//! p_unit_test 3, %res3
writeout(3, f2f16(fmul(f2f32(a16), b)));
//! v2b: %res4 = v_fma_mixlo_f16 1.0, %a, lo(%b16)
//! p_unit_test 4, %res4
writeout(4, f2f16(fadd(a, f2f32(b16))));
//! v2b: %res5 = v_fma_mixlo_f16 %a, lo(%b16), %c
//! p_unit_test 5, %res5
writeout(5, f2f16(fma(a, f2f32(b16), c)));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.output_conv.precision)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v2b: %a16 = p_startpgm
if (!setup_cs("v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a16 = inputs[0];
//! v2b: %res0_tmp = v_mul_f16 %a16, %a16
//! v1: (precise)%res0 = v_cvt_f32_f16 %res0_tmp
//! p_unit_test 0, %res0
writeout(0, f2f32(fmul(a16, a16), bld.precise()));
//! v2b: (precise)%res1_tmp = v_mul_f16 %a16, %a16
//! v1: %res1 = v_cvt_f32_f16 %res1_tmp
//! p_unit_test 1, %res1
writeout(1, f2f32(fmul(a16, a16, bld.precise())));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.output_conv.modifiers)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v1: %b, v2b: %a16, v2b: %b16 = p_startpgm
if (!setup_cs("v1 v1 v2b v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp b = inputs[1];
Temp a16 = inputs[2];
Temp b16 = inputs[3];
/* fneg/fabs */
//! v1: %res0_add = v_add_f32 %1, %2
//! v2b: %res0 = v_cvt_f16_f32 |%res0_add|
//! p_unit_test 0, %res0
writeout(0, f2f16(fabs(fadd(a, b))));
//! v1: %res1_add = v_add_f32 %1, %2
//! v2b: %res1 = v_cvt_f16_f32 -%res1_add
//! p_unit_test 1, %res1
writeout(1, f2f16(fneg(fadd(a, b))));
//! v2b: %res2_add = v_add_f16 %3, %4
//! v1: %res2 = v_cvt_f32_f16 |%res2_add|
//! p_unit_test 2, %res2
writeout(2, f2f32(fabs(fadd(a16, b16))));
//! v2b: %res3_add = v_add_f16 %3, %4
//! v1: %res3 = v_cvt_f32_f16 -%res3_add
//! p_unit_test 3, %res3
writeout(3, f2f32(fneg(fadd(a16, b16))));
/* sdwa */
//! v2b: %res4_add = v_fma_mixlo_f16 1.0, %a, %b
//! v2b: %res4 = p_extract %res4_add, 0, 8, 0
//! p_unit_test 4, %res4
writeout(4, ext_ubyte(f2f16(fadd(a, b)), 0));
//! v1: %res5_mul = v_add_f32 %a, %b dst_sel:uword0 src0_sel:dword src1_sel:dword
//! v2b: %res5 = v_cvt_f16_f32 %res5_mul
//! p_unit_test 5, %res5
writeout(5, f2f16(ext_ushort(fadd(a, b), 0)));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.fma.basic)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %c16 = p_startpgm
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
Temp a16 = inputs[3];
Temp c16 = inputs[4];
//! v1: %res0 = v_fma_mix_f32 lo(%a16), %b, %c
//! p_unit_test 0, %res0
writeout(0, fadd(fmul(f2f32(a16), b), c));
//! v1: %res1 = v_fma_mix_f32 %a, %b, lo(%c16)
//! p_unit_test 1, %res1
writeout(1, fadd(fmul(a, b), f2f32(c16)));
/* omod/clamp check */
//! v1: %res2_mul = v_fma_mix_f32 lo(%a16), %b, neg(0)
//! v1: %res2 = v_add_f32 %res2_mul, %c *2
//! p_unit_test 2, %res2
writeout(2, bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000),
fadd(fmul(f2f32(a16), b), c)));
/* neg/abs modifiers */
//! v1: %res3 = v_fma_mix_f32 -lo(%a16), %b, |lo(%c16)|
//! p_unit_test 3, %res3
writeout(3, fadd(fmul(fneg(f2f32(a16)), b), fabs(f2f32(c16))));
//! v1: %res4 = v_fma_mix_f32 |%a|, |%b|, lo(%c16)
//! p_unit_test 4, %res4
writeout(4, fadd(fabs(fmul(fneg(a), fneg(b))), f2f32(c16)));
//! v1: %res5 = v_fma_mix_f32 %a, -%b, lo(%c16)
//! p_unit_test 5, %res5
writeout(5, fadd(fneg(fmul(a, b)), f2f32(c16)));
//! v1: %res6 = v_fma_mix_f32 |%a|, -|%b|, lo(%c16)
//! p_unit_test 6, %res6
writeout(6, fadd(fneg(fabs(fmul(fneg(a), fneg(b)))), f2f32(c16)));
/* output conversions */
//! v2b: %res7 = v_fma_mixlo_f16 %a, %b, %c
//! p_unit_test 7, %res7
writeout(7, f2f16(fadd(fmul(a, b), c)));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.fma.precision)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v1: %b, v1: %c, v2b: %a16, v2b: %b16 = p_startpgm
if (!setup_cs("v1 v1 v1 v2b v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
Temp a16 = inputs[3];
Temp b16 = inputs[4];
/* the optimization is precise for 32-bit on GFX9 */
//~gfx9! v1: (precise)%res0 = v_fma_mix_f32 lo(%a16), %b, %c
//~gfx10! v1: (precise)%res0_tmp = v_fma_mix_f32 lo(%a16), %b, neg(0)
//~gfx10! v1: %res0 = v_add_f32 %res0_tmp, %c
//! p_unit_test 0, %res0
writeout(0, fadd(fmul(f2f32(a16), b, bld.precise()), c));
//~gfx9! v1: (precise)%res1 = v_fma_mix_f32 lo(%a16), %b, %c
//~gfx10! v1: %res1_tmp = v_fma_mix_f32 lo(%a16), %b, neg(0)
//~gfx10! v1: (precise)%res1 = v_add_f32 %res1_tmp, %c
//! p_unit_test 1, %res1
writeout(1, fadd(fmul(f2f32(a16), b), c, bld.precise()));
/* never promote 16-bit arithmetic to 32-bit */
//! v2b: %res2_tmp = v_cvt_f16_f32 %a
//! v2b: %res2 = v_add_f16 %res2_tmp, %b16
//! p_unit_test 2, %res2
writeout(2, fadd(f2f16(a), b16));
//! v2b: %res3_tmp = v_cvt_f16_f32 %a
//! v2b: %res3 = v_mul_f16 %res3_tmp, %b16
//! p_unit_test 3, %res3
writeout(3, fmul(f2f16(a), b16));
//! v2b: %res4_tmp = v_mul_f16 %a16, %b16
//! v1: %res4 = v_cvt_f32_f16 %res4_tmp
//! p_unit_test 4, %res4
writeout(4, f2f32(fmul(a16, b16)));
//! v2b: %res5_tmp = v_add_f16 %a16, %b16
//! v1: %res5 = v_cvt_f32_f16 %res5_tmp
//! p_unit_test 5, %res5
writeout(5, f2f32(fadd(a16, b16)));
//! v2b: %res6_tmp = v_fma_mixlo_f16 %a, %b, neg(lo(0))
//! v2b: %res6 = v_add_f16 %res6_tmp, %a16
//! p_unit_test 6, %res6
writeout(6, fadd(f2f16(fmul(a, b)), a16));
//! v2b: %res7_tmp = v_mul_f16 %a16, %b16
//! v1: %res7 = v_fma_mix_f32 1.0, lo(%res7_tmp), %c
//! p_unit_test 7, %res7
writeout(7, fadd(f2f32(fmul(a16, b16)), c));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.clamp)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp a16 = inputs[1];
//! v1: %res0 = v_fma_mix_f32 lo(%a16), %a, neg(0) clamp
//! p_unit_test 0, %res0
writeout(0, fsat(fmul(f2f32(a16), a)));
//! v2b: %res1 = v_fma_mixlo_f16 %a, %a, neg(lo(0)) clamp
//! p_unit_test 1, %res1
writeout(1, f2f16(fsat(fmul(a, a))));
//! v2b: %res2 = v_fma_mixlo_f16 %a, %a, neg(lo(0)) clamp
//! p_unit_test 2, %res2
writeout(2, fsat(f2f16(fmul(a, a))));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.mad_mix.cast)
for (unsigned i = GFX9; i <= GFX10; i++) {
//>> v1: %a, v2b: %a16 = p_startpgm
if (!setup_cs("v1 v2b", (amd_gfx_level)i))
continue;
program->blocks[0].fp_mode.denorm16_64 = fp_denorm_flush;
Temp a = inputs[0];
Temp a16 = inputs[1];
/* The optimizer copy-propagates v2b=p_extract_vector(v1, 0) and p_as_uniform, so the
* optimizer has to check compatibility.
*/
//! v1: %res0_cvt = v_cvt_f32_f16 %a16
//! v2b: %res0 = v_mul_f16 %res0_cvt, %a16
//! p_unit_test 0, %res0
writeout(0, fmul(u2u16(f2f32(a16)), a16));
//! v2b: %res1_cvt = v_cvt_f16_f32 %a
//! v1: %res1 = v_mul_f32 %res1_cvt, %a
//! p_unit_test 1, %res1
writeout(1, fmul(bld.as_uniform(f2f16(a)), a));
//! v2b: %res2_mul = v_mul_f16 %a16, %a16
//! v2b: %res2 = v_cvt_f16_f32 %res2_mul
//! p_unit_test 2, %res2
writeout(2, f2f16(bld.as_uniform(fmul(a16, a16))));
//! v1: %res3_mul = v_mul_f32 %a, %a
//! v1: %res3 = v_cvt_f32_f16 %res3_mul
//! p_unit_test 3, %res3
writeout(3, f2f32(u2u16(fmul(a, a))));
//! v1: %res4_mul = v_fma_mix_f32 lo(%a16), %a, neg(0)
//! v2b: %res4 = v_add_f16 %res4_mul, 0 clamp
//! p_unit_test 4, %res4
writeout(4, fsat(u2u16(fmul(f2f32(a16), a))));
//! v2b: %res5_mul = v_fma_mixlo_f16 %a, %a, neg(lo(0))
//! v1: %res5 = v_add_f32 %res5_mul, 0 clamp
//! p_unit_test 5, %res5
writeout(5, fsat(bld.as_uniform(f2f16(fmul(a, a)))));
//! v1: %res6_mul = v_mul_f32 %a, %a
//! v1: %res6 = v_fma_mix_f32 1.0, lo(%res6_mul), %a
//! p_unit_test 6, %res6
writeout(6, fadd(f2f32(u2u16(fmul(a, a))), a));
//! v2b: %res7_mul = v_mul_f16 %a16, %a16
//! v1: %res7 = v_fma_mix_f32 1.0, lo(%a16), %res7_mul
//! p_unit_test 7, %res7
writeout(7, fadd(bld.as_uniform(fmul(a16, a16)), f2f32(a16)));
/* opsel_hi should be obtained from the original opcode, not the operand regclass */
//! v1: %res8 = v_fma_mix_f32 lo(%a16), %a16, neg(0)
//! p_unit_test 8, %res8
writeout(8, fmul(f2f32(a16), a16));
finish_opt_test();
}
END_TEST
static void
vop3p_constant(unsigned* idx, aco_opcode op, const char* swizzle, uint32_t val)
{
uint32_t halves[2] = {val & 0xffff, val >> 16};
uint32_t expected = halves[swizzle[0] - 'x'] | (halves[swizzle[1] - 'x'] << 16);
fprintf(output, "Expected for %u: 0x%.8x / %u\n", *idx, expected, expected);
unsigned opsel_lo = swizzle[0] == 'x' ? 0x0 : 0x1;
unsigned opsel_hi = swizzle[1] == 'x' ? 0x2 : 0x3;
writeout((*idx)++, bld.vop3p(op, bld.def(v1), bld.copy(bld.def(v1), Operand::c32(val)),
inputs[0], opsel_lo, opsel_hi));
}
BEGIN_TEST(optimize.vop3p_constants)
for (aco_opcode op : {aco_opcode::v_pk_add_f16, aco_opcode::v_pk_add_u16}) {
for (const char* swizzle : {"xx", "yy", "xy", "yx"}) {
char variant[16];
strcpy(variant, op == aco_opcode::v_pk_add_f16 ? "_f16" : "_u16");
strcat(variant, "_");
strcat(variant, swizzle);
//; for i in range(36):
//; insert_pattern('Expected for %u: $_ / #expected%u' % (i, i))
//>> v1: %a = p_startpgm
if (!setup_cs("v1", GFX10_3, CHIP_UNKNOWN, variant))
continue;
//; opcode = 'v_pk_add_u16' if 'u16' in variant else 'v_pk_add_f16'
//; for i in range(36):
//; insert_pattern('v1: %%res%u = %s $got%u %%a' % (i, opcode, i))
//; insert_pattern('p_unit_test %u, %%res%u' % (i, i))
//! s_endpgm
//; def parse_op(op):
//; is_int = opcode == 'v_pk_add_u16'
//; op = op.rstrip(',')
//;
//; mods = lambda v: v
//; if op.endswith('*[1,-1]'):
//; mods = lambda v: v ^ 0x80000000
//; assert(not is_int)
//; elif op.endswith('*[-1,1]'):
//; mods = lambda v: v ^ 0x00008000
//; assert(not is_int)
//; elif op.startswith('neg('):
//; mods = lambda v: v ^ 0x80008000
//; assert(not is_int)
//; op = op[4:-1]
//; op = op.split('*')[0]
//;
//; swizzle = lambda v: v
//; if op.endswith('.xx'):
//; swizzle = lambda v: ((v & 0xffff) | (v << 16)) & 0xffffffff;
//; elif op.endswith('.yy'):
//; swizzle = lambda v: (v >> 16) | (v & 0xffff0000);
//; elif op.endswith('.yx'):
//; swizzle = lambda v: ((v >> 16) | (v << 16)) & 0xffffffff;
//; op = op.rstrip('xy.')
//;
//; val = None
//; if op.startswith('0x'):
//; val = int(op[2:], 16)
//; elif op == '-1.0':
//; val = 0xbf800000 if is_int else 0xbC00
//; elif op == '1.0':
//; val = 0x3f800000 if is_int else 0x3c00
//; else:
//; val = int(op) & 0xffffffff
//;
//; return mods(swizzle(val))
//; # Check correctness
//; for i in range(36):
//; expected = globals()['expected%u' % i]
//; got = globals()['got%u' % i]
//; got_parsed = parse_op(got)
//; if got_parsed != expected:
//; raise Exception('Check %u failed: expected 0x%.8x, got 0x%.8x ("%s")' % (i, expected, got_parsed, got))
//; # Check that all literals are ones that cannot be encoded as inline constants
//; allowed_literals = [0x00004242, 0x0000fffe, 0x00308030, 0x0030ffff, 0x3c00ffff,
//; 0x42420000, 0x42424242, 0x4242c242, 0x4242ffff, 0x7ffefffe,
//; 0x80300030, 0xbeefdead, 0xc2424242, 0xdeadbeef, 0xfffe0000,
//; 0xfffe7ffe, 0xffff0030, 0xffff3c00, 0xffff4242, 0xc242c242,
//; 0x80308030, 0xdeaddead, 0xbeefbeef, 0x7ffe7ffe]
//; if opcode == 'v_pk_add_u16':
//; allowed_literals.extend([0x00003c00, 0x3c000000, 0x3c003c00, 0x3c00bc00, 0xbc003c00, 0xbc00bc00])
//; else:
//; allowed_literals.extend([0x00003f80, 0x3f800000, 0x3f803f80])
//;
//; for i in range(36):
//; got = globals()['got%u' % i].removeprefix('neg(')
//; if not got.startswith('0x'):
//; continue;
//; got = int(got[2:].rstrip(',)').split('*')[0].split('.')[0], 16)
//; if got not in allowed_literals:
//; raise Exception('Literal check %u failed: 0x%.8x not in allowed literals' % (i, got))
unsigned idx = 0;
for (uint32_t constant : {0x3C00, 0x0030, 0xfffe, 0x4242}) {
vop3p_constant(&idx, op, swizzle, constant);
vop3p_constant(&idx, op, swizzle, constant | 0xffff0000);
vop3p_constant(&idx, op, swizzle, constant | (constant << 16));
vop3p_constant(&idx, op, swizzle, constant << 16);
vop3p_constant(&idx, op, swizzle, (constant << 16) | 0x0000ffff);
vop3p_constant(&idx, op, swizzle, constant | ((constant ^ 0x8000) << 16));
vop3p_constant(&idx, op, swizzle, (constant ^ 0x8000) | (constant << 16));
}
for (uint32_t constant : {0x3f800000u, 0xfffffffeu, 0x00000030u, 0xdeadbeefu}) {
uint32_t lo = constant & 0xffff;
uint32_t hi = constant >> 16;
vop3p_constant(&idx, op, swizzle, constant);
vop3p_constant(&idx, op, swizzle, hi | (lo << 16));
}
finish_opt_test();
}
}
END_TEST
BEGIN_TEST(optimize.fmamix_two_literals)
/* This test has to recreate literals sometimes because we don't combine them at all if there's
* at least one uncombined use.
*/
for (unsigned i = GFX10; i <= GFX10_3; i++) {
//>> v1: %a, v1: %b = p_startpgm
if (!setup_cs("v1 v1", (amd_gfx_level)i))
continue;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
Temp c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
Temp c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000));
//! v1: %res0 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00)
//! p_unit_test 0, %res0
writeout(0, fma(a, c15, c30));
/* No need to use v_fma_mix_f32. */
//! v1: %res1 = v_fmaak_f32 %a, %b, 0x40400000
//! p_unit_test 1, %res1
writeout(1, fma(a, b, c30));
/* Separate mul/add can become v_fma_mix_f32 if it's not precise. */
//! v1: %res2 = v_fma_mix_f32 %a, lo(0x42003e00), hi(0x42003e00)
//! p_unit_test 2, %res2
writeout(2, fadd(fmul(a, c15), c30));
//~gfx10! v1: %c15 = p_parallelcopy 0x3fc00000
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
/* v_fma_mix_f32 is a fused mul/add, so it can't be used for precise separate mul/add. */
//~gfx10! v1: (precise)%res3 = v_madak_f32 %a, %c15, 0x40400000
//~gfx10_3! v1: (precise)%res3_tmp = v_mul_f32 %a, 0x3fc00000
//~gfx10_3! v1: %res3 = v_add_f32 %res3_tmp, 0x40400000
//! p_unit_test 3, %res3
writeout(3, fadd(bld.precise().vop2(aco_opcode::v_mul_f32, bld.def(v1), a, c15), c30));
//~gfx10! v1: (precise)%res4 = v_madak_f32 %1, %c16, 0x40400000
//~gfx10_3! v1: %res4_tmp = v_mul_f32 %a, 0x3fc00000
//~gfx10_3! v1: (precise)%res4 = v_add_f32 %res4_tmp, 0x40400000
//! p_unit_test 4, %res4
writeout(4, bld.precise().vop2(aco_opcode::v_add_f32, bld.def(v1), fmul(a, c15), c30));
/* Can't convert to fp16 if it will be flushed as a denormal. */
//! v1: %res5 = v_fma_mix_f32 %1, lo(0x3ff3e00), hi(0x3ff3e00)
//! p_unit_test 5, %res5
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
writeout(5, fma(a, c15, c_denorm));
//>> BB1
//! /* logical preds: BB0, / linear preds: BB0, / kind: uniform, */
program->next_fp_mode.denorm16_64 = fp_denorm_flush;
bld.reset(program->create_and_insert_block());
program->blocks[0].linear_succs.push_back(1);
program->blocks[0].logical_succs.push_back(1);
program->blocks[1].linear_preds.push_back(0);
program->blocks[1].logical_preds.push_back(0);
//~gfx10; del c15
//! v1: %c15 = p_parallelcopy 0x3fc00000
//! v1: %res6 = v_fmaak_f32 %a, %c15, 0x387fc000
//! p_unit_test 6, %res6
c15 = bld.copy(bld.def(v1), Operand::c32(fui(1.5f)));
c_denorm = bld.copy(bld.def(v1), Operand::c32(0x387fc000));
writeout(6, fma(a, c15, c_denorm));
/* Can't accept more than 3 unique fp16 literals. */
//! v1: %c45 = p_parallelcopy 0x40900000
//! v1: %res7 = v_fma_mix_f32 lo(0x42003e00), hi(0x42003e00), %c45
//! p_unit_test 7, %res7
Temp c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f)));
writeout(7, fma(c15, c30, c45));
/* Modifiers must be preserved. */
//! v1: %res8 = v_fma_mix_f32 -%a, lo(0x44804200), hi(0x44804200)
//! p_unit_test 8, %res8
writeout(8, fma(fneg(a), c30, c45));
//! v1: %res9 = v_fma_mix_f32 lo(0x44804200), |%a|, hi(0x44804200)
//! p_unit_test 9, %res9
writeout(9, fma(c30, fabs(a), c45));
//! v1: %res10 = v_fma_mix_f32 %a, lo(0x44804200), hi(0x44804200) clamp
//! p_unit_test 10, %res10
writeout(10, fsat(fma(a, c30, c45)));
/* Output modifiers are not supported by v_fma_mix_f32. */
c30 = bld.copy(bld.def(v1), Operand::c32(fui(3.0f)));
//; del c45
//! v1: %c45 = p_parallelcopy 0x40900000
//! v1: %res11 = v_fma_f32 %a, 0x40400000, %c45 *0.5
//! p_unit_test 11, %res11
c45 = bld.copy(bld.def(v1), Operand::c32(fui(4.5f)));
writeout(11, fmul(fma(a, c30, c45), bld.copy(bld.def(v1), Operand::c32(0x3f000000))));
/* Has a literal which can't be represented as fp16. */
//! v1: %c03 = p_parallelcopy 0x3e99999a
//! v1: %res12 = v_fmaak_f32 %a, %c03, 0x40400000
//! p_unit_test 12, %res12
Temp c03 = bld.copy(bld.def(v1), Operand::c32(fui(0.3f)));
writeout(12, fma(a, c03, c30));
/* We should still use fmaak/fmamk if the two literals are identical. */
//! v1: %res13 = v_fmaak_f32 0x40400000, %a, 0x40400000
//! p_unit_test 13, %res13
writeout(13, fma(a, c30, c30));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.fma_opsel)
/* TODO make these work before GFX11 using SDWA. */
for (unsigned i = GFX11; i <= GFX11; i++) {
//>> v2b: %a, v2b: %b, v1: %c, v1: %d, v1: %e = p_startpgm
if (!setup_cs("v2b v2b v1 v1 v1", (amd_gfx_level)i))
continue;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
Temp d = inputs[3];
Temp e = inputs[4];
Temp c_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), c, Operand::c32(1));
Temp d_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), d, Operand::c32(1));
Temp e_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), e, Operand::c32(1));
//! v2b: %res0 = v_fma_f16 %b, hi(%c), %a
//! p_unit_test 0, %res0
writeout(0, fadd(fmul(b, c_hi), a));
//! v2b: %res1 = v_fma_f16 %a, %b, hi(%d)
//! p_unit_test 1, %res1
writeout(1, fadd(fmul(a, b), d_hi));
//! v2b: %res2 = v_fma_f16 %a, %b, hi(%e)
//! p_unit_test 2, %res2
writeout(2, fma(a, b, e_hi));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.dpp_opsel)
//>> v1: %a, v1: %b = p_startpgm
if (!setup_cs("v1 v1", GFX11))
return;
Temp a = inputs[0];
Temp b = inputs[1];
Temp dpp16 = bld.vop1_dpp(aco_opcode::v_mov_b32, bld.def(v1), a, dpp_row_mirror);
Temp dpp16_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp16, Operand::c32(1));
Temp dpp8 = bld.vop1_dpp8(aco_opcode::v_mov_b32, bld.def(v1), a);
Temp dpp8_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), dpp8, Operand::c32(1));
Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1));
Temp b_lo = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(0));
//! v2b: %res0 = v_add_f16 hi(%a), hi(%b) row_mirror bound_ctrl:1 fi
//! p_unit_test 0, %res0
writeout(0, fadd(dpp16_hi, b_hi));
//! v2b: %res1 = v_add_f16 hi(%a), %b dpp8:[0,0,0,0,0,0,0,0] fi
//! p_unit_test 1, %res1
writeout(1, fadd(b_lo, dpp8_hi));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.apply_sgpr_swap_opsel)
//>> v1: %a, s1: %b = p_startpgm
if (!setup_cs("v1 s1", GFX11))
return;
Temp a = inputs[0];
Temp b = inputs[1];
Temp b_vgpr = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), bld.copy(bld.def(v1), b),
Operand::c32(0));
Temp res0 = bld.tmp(v2b);
VALU_instruction& valu = bld.vop2(aco_opcode::v_sub_f16, Definition(res0), a, b_vgpr)->valu();
valu.opsel[0] = true;
//! v2b: %res0 = v_subrev_f16 %b, hi(%a)
//! p_unit_test 0, %res0
writeout(0, res0);
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.max3_opsel)
/* TODO make these work before GFX11 using SDWA. */
for (unsigned i = GFX11; i <= GFX11; i++) {
//>> v1: %a, v1: %b, v2b: %c = p_startpgm
if (!setup_cs("v1 v1 v2b", GFX11))
continue;
Temp a = inputs[0];
Temp b = inputs[1];
Temp c = inputs[2];
Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
Temp b_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), b, Operand::c32(1));
//! v2b: %res0 = v_max3_f16 hi(%a), %c, hi(%b)
//! p_unit_test 0, %res0
writeout(0, bld.vop2(aco_opcode::v_max_f16, bld.def(v2b),
bld.vop2(aco_opcode::v_max_f16, bld.def(v2b), a_hi, c), b_hi));
finish_opt_test();
}
END_TEST
BEGIN_TEST(optimize.neg_mul_opsel)
//>> v1: %a, v2b: %b = p_startpgm
if (!setup_cs("v1 v2b", GFX11))
return;
Temp a = inputs[0];
Temp b = inputs[1];
Temp a_hi = bld.pseudo(aco_opcode::p_extract_vector, bld.def(v2b), a, Operand::c32(1));
//! v2b: %res0 = v_mul_f16 -hi(%a), %b
//! p_unit_test 0, %res0
writeout(0, fneg(fmul(a_hi, b)));
//! v1: %res1 = v_fma_mix_f32 -hi(%a), lo(%b), neg(0)
//! p_unit_test 1, %res1
writeout(1, fneg(fmul(f2f32(a_hi), f2f32(b))));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.vinterp_inreg_output_modifiers)
//>> v1: %a, v1: %b, v1: %c = p_startpgm
if (!setup_cs("v1 v1 v1", GFX11))
return;
//! v1: %res0 = v_interp_p2_f32_inreg %a, %b, %c clamp
//! p_unit_test 0, %res0
Temp tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[0],
inputs[1], inputs[2]);
writeout(0, fsat(tmp));
//! v1: %res1 = v_fma_f32 %b, %a, %c *2 quad_perm:[2,2,2,2] fi
//! p_unit_test 1, %res1
tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[1], inputs[0],
inputs[2]);
tmp = bld.vop2(aco_opcode::v_mul_f32, bld.def(v1), Operand::c32(0x40000000u), tmp);
writeout(1, tmp);
//! v2b: %res2 = v_interp_p2_f16_f32_inreg %a, %b, %c clamp
//! p_unit_test 2, %res2
tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[0],
inputs[1], inputs[2]);
writeout(2, fsat(tmp));
//! v2b: %tmp3 = v_interp_p2_f16_f32_inreg %b, %a, %c
//! v2b: %res3 = v_mul_f16 2.0, %tmp3
//! p_unit_test 3, %res3
tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f16_f32_inreg, bld.def(v2b), inputs[1],
inputs[0], inputs[2]);
tmp = bld.vop2(aco_opcode::v_mul_f16, bld.def(v2b), Operand::c16(0x4000u), tmp);
writeout(3, tmp);
//! v2b: %res4 = v_fma_mixlo_f16 %c, %b, %a quad_perm:[2,2,2,2] fi
//! p_unit_test 4, %res4
tmp = bld.vinterp_inreg(aco_opcode::v_interp_p2_f32_inreg, bld.def(v1), inputs[2], inputs[1],
inputs[0]);
writeout(4, f2f16(tmp));
finish_opt_test();
END_TEST
BEGIN_TEST(optimize.s_pack)
//>> s1: %a, s1: %b, s1: %c = p_startpgm
if (!setup_cs("s1 s1 s1", GFX11))
return;
Temp lo = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[1],
Operand::c32(0), Operand::c32(16u), Operand::c32(false));
Temp hi = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[2],
Operand::c32(1), Operand::c32(16u), Operand::c32(false));
//! s1: %res0 = s_pack_lh_b32_b16 %b, %c
//! p_unit_test 0, %res0
writeout(0, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, hi));
//! s1: %res1 = s_pack_ll_b32_b16 %b, %b
//! p_unit_test 1, %res1
writeout(1, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, lo));
//! s1: %res2 = s_pack_hl_b32_b16 %c, %b
//! p_unit_test 2, %res2
writeout(2, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, lo));
//! s1: %res3 = s_pack_hh_b32_b16 %c, %c
//! p_unit_test 3, %res3
writeout(3, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, hi));
lo = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[1], Operand::c32(0),
Operand::c32(16u), Operand::c32(false));
hi = bld.pseudo(aco_opcode::p_extract, bld.def(s1), bld.def(s1, scc), inputs[2], Operand::c32(1),
Operand::c32(16u), Operand::c32(false));
//! s1: %res4 = s_pack_ll_b32_b16 %a, %b
//! p_unit_test 4, %res4
writeout(4, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), inputs[0], lo));
//! s1: %res5 = s_pack_lh_b32_b16 %a, %c
//! p_unit_test 5, %res5
writeout(5, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), inputs[0], hi));
//! s1: %res6 = s_pack_ll_b32_b16 %b, %a
//! p_unit_test 6, %res6
writeout(6, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), lo, inputs[0]));
//! s1: %res7 = s_pack_hl_b32_b16 %c, %a
//! p_unit_test 7, %res7
writeout(7, bld.sop2(aco_opcode::s_pack_ll_b32_b16, bld.def(s1), hi, inputs[0]));
finish_opt_test();
END_TEST
BEGIN_TEST(optimizer.trans_inline_constant)
if (!setup_cs("", GFX12))
return;
//>> s1: %res0 = v_s_rcp_f32 1.0
//! p_unit_test 0, %res0
writeout(0, bld.vop3(aco_opcode::v_s_rcp_f32, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3f800000))));
//! s1: %res1 = v_s_rcp_f32 0x3c00
//! p_unit_test 1, %res1
writeout(1, bld.vop3(aco_opcode::v_s_rcp_f32, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3c00))));
//! s1: %tmp2 = p_parallelcopy 1.0
//! s1: %res2 = v_s_rcp_f16 %tmp2
//! p_unit_test 2, %res2
writeout(2, bld.vop3(aco_opcode::v_s_rcp_f16, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3f800000))));
//! s1: %tmp3 = p_parallelcopy 0x3c00
//! s1: %res3 = v_s_rcp_f16 %tmp3
//! p_unit_test 3, %res3
writeout(3, bld.vop3(aco_opcode::v_s_rcp_f16, bld.def(s1), bld.copy(bld.def(s1), Operand::c32(0x3c00))));
//! v1: %res4 = v_rcp_f32 1.0
//! p_unit_test 4, %res4
writeout(4, bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), bld.copy(bld.def(s1), Operand::c32(0x3f800000))));
//! v1: %res5 = v_rcp_f32 0x3c00
//! p_unit_test 5, %res5
writeout(5, bld.vop1(aco_opcode::v_rcp_f32, bld.def(v1), bld.copy(bld.def(s1), Operand::c32(0x3c00))));
//! v2b: %res6 = v_rcp_f16 0x3f800000
//! p_unit_test 6, %res6
writeout(6, bld.vop1(aco_opcode::v_rcp_f16, bld.def(v2b), bld.copy(bld.def(s1), Operand::c32(0x3f800000))));
//! v2b: %res7 = v_rcp_f16 1.0
//! p_unit_test 7, %res7
writeout(7, bld.vop1(aco_opcode::v_rcp_f16, bld.def(v2b), bld.copy(bld.def(s1), Operand::c32(0x3c00))));
finish_opt_test();
END_TEST
BEGIN_TEST(optimizer.trans_no_omod)
//>> s1: %a = p_startpgm
if (!setup_cs("s1", GFX12))
return;
//! s1: %tmp0 = v_s_log_f32 %a
//! v1: %res = v_mul_legacy_f32 %tmp0, 0.5
//! p_unit_test 0, %res
Temp dst = bld.vop3(aco_opcode::v_s_log_f32, bld.def(s1), inputs[0]);
writeout(0, bld.vop2(aco_opcode::v_mul_legacy_f32, bld.def(v1), dst,
bld.copy(bld.def(v1), Operand::c32(0x3f000000))));
finish_opt_test();
END_TEST
BEGIN_TEST(optimizer.pk_clamp_fma_mix)
//>> v1: %a, v1: %b, v2b: %c = p_startpgm
if (!setup_cs("v1 v1 v2b", GFX12))
return;
//! v2b: %cvt0 = v_fma_mixlo_f16 1.0, %a, %b
//! v1: %clamp0 = v_pk_mul_f16 %cvt0.xx, 1.0.xx clamp
//! p_unit_test 0, %clamp0
Temp add0 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], inputs[1]);
Temp cvt0 = bld.vop1(aco_opcode::v_cvt_f16_f32, bld.def(v2b), add0);
Builder::Result clamp0 =
bld.vop3p(aco_opcode::v_pk_mul_f16, bld.def(v1), cvt0, Operand::c16(0x3c00), 0, 0);
clamp0->valu().clamp = true;
writeout(0, clamp0);
//! v1: %add1 = v_fma_mix_f32 1.0, lo(%c), %a
//! v1: %clamp1 = v_pk_mul_f16 %add1, 1.0.xx clamp
//! p_unit_test 1, %clamp1
Temp cvt1 = bld.vop1(aco_opcode::v_cvt_f32_f16, bld.def(v1), inputs[2]);
Temp add1 = bld.vop2(aco_opcode::v_add_f32, bld.def(v1), inputs[0], cvt1);
Builder::Result clamp1 =
bld.vop3p(aco_opcode::v_pk_mul_f16, bld.def(v1), add1, Operand::c16(0x3c00), 0, 0x1);
clamp1->valu().clamp = true;
writeout(1, clamp1);
finish_opt_test();
END_TEST