mirror of
https://github.com/yuzu-emu/yuzu-mainline.git
synced 2024-12-24 12:25:28 +00:00
shader: Better constant folding
This commit is contained in:
parent
da8096e6e3
commit
6dafb08f52
|
@ -87,7 +87,7 @@ static std::string ArgToIndex(const std::map<const Block*, size_t>& block_to_ind
|
||||||
}
|
}
|
||||||
switch (arg.Type()) {
|
switch (arg.Type()) {
|
||||||
case Type::U1:
|
case Type::U1:
|
||||||
return fmt::format("#{}", arg.U1() ? '1' : '0');
|
return fmt::format("#{}", arg.U1() ? "true" : "false");
|
||||||
case Type::U8:
|
case Type::U8:
|
||||||
return fmt::format("#{}", arg.U8());
|
return fmt::format("#{}", arg.U8());
|
||||||
case Type::U16:
|
case Type::U16:
|
||||||
|
|
|
@ -3,6 +3,7 @@
|
||||||
// Refer to the license.txt file included.
|
// Refer to the license.txt file included.
|
||||||
|
|
||||||
#include <algorithm>
|
#include <algorithm>
|
||||||
|
#include <tuple>
|
||||||
#include <type_traits>
|
#include <type_traits>
|
||||||
|
|
||||||
#include "common/bit_cast.h"
|
#include "common/bit_cast.h"
|
||||||
|
@ -13,12 +14,17 @@
|
||||||
|
|
||||||
namespace Shader::Optimization {
|
namespace Shader::Optimization {
|
||||||
namespace {
|
namespace {
|
||||||
[[nodiscard]] u32 BitFieldUExtract(u32 base, u32 shift, u32 count) {
|
// Metaprogramming stuff to get arguments information out of a lambda
|
||||||
if (static_cast<size_t>(shift) + static_cast<size_t>(count) > Common::BitSize<u32>()) {
|
template <typename Func>
|
||||||
throw LogicError("Undefined result in BitFieldUExtract({}, {}, {})", base, shift, count);
|
struct LambdaTraits : LambdaTraits<decltype(&std::remove_reference_t<Func>::operator())> {};
|
||||||
}
|
|
||||||
return (base >> shift) & ((1U << count) - 1);
|
template <typename ReturnType, typename LambdaType, typename... Args>
|
||||||
}
|
struct LambdaTraits<ReturnType (LambdaType::*)(Args...) const> {
|
||||||
|
template <size_t I>
|
||||||
|
using ArgType = std::tuple_element_t<I, std::tuple<Args...>>;
|
||||||
|
|
||||||
|
static constexpr size_t NUM_ARGS{sizeof...(Args)};
|
||||||
|
};
|
||||||
|
|
||||||
template <typename T>
|
template <typename T>
|
||||||
[[nodiscard]] T Arg(const IR::Value& value) {
|
[[nodiscard]] T Arg(const IR::Value& value) {
|
||||||
|
@ -104,6 +110,14 @@ void FoldAdd(IR::Inst& inst) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
void FoldSelect(IR::Inst& inst) {
|
||||||
|
const IR::Value cond{inst.Arg(0)};
|
||||||
|
if (cond.IsImmediate()) {
|
||||||
|
inst.ReplaceUsesWith(cond.U1() ? inst.Arg(1) : inst.Arg(2));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
void FoldLogicalAnd(IR::Inst& inst) {
|
void FoldLogicalAnd(IR::Inst& inst) {
|
||||||
if (!FoldCommutative(inst, [](bool a, bool b) { return a && b; })) {
|
if (!FoldCommutative(inst, [](bool a, bool b) { return a && b; })) {
|
||||||
return;
|
return;
|
||||||
|
@ -131,6 +145,21 @@ void FoldBitCast(IR::Inst& inst, IR::Opcode reverse) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
template <typename Func, size_t... I>
|
||||||
|
IR::Value EvalImmediates(const IR::Inst& inst, Func&& func, std::index_sequence<I...>) {
|
||||||
|
using Traits = LambdaTraits<decltype(func)>;
|
||||||
|
return IR::Value{func(Arg<Traits::ArgType<I>>(inst.Arg(I))...)};
|
||||||
|
}
|
||||||
|
|
||||||
|
template <typename Func>
|
||||||
|
void FoldWhenAllImmediates(IR::Inst& inst, Func&& func) {
|
||||||
|
if (!inst.AreAllArgsImmediates() || inst.HasAssociatedPseudoOperation()) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
using Indices = std::make_index_sequence<LambdaTraits<decltype(func)>::NUM_ARGS>;
|
||||||
|
inst.ReplaceUsesWith(EvalImmediates(inst, func, Indices{}));
|
||||||
|
}
|
||||||
|
|
||||||
void ConstantPropagation(IR::Inst& inst) {
|
void ConstantPropagation(IR::Inst& inst) {
|
||||||
switch (inst.Opcode()) {
|
switch (inst.Opcode()) {
|
||||||
case IR::Opcode::GetRegister:
|
case IR::Opcode::GetRegister:
|
||||||
|
@ -145,14 +174,20 @@ void ConstantPropagation(IR::Inst& inst) {
|
||||||
return FoldBitCast<u32, f32>(inst, IR::Opcode::BitCastF32U32);
|
return FoldBitCast<u32, f32>(inst, IR::Opcode::BitCastF32U32);
|
||||||
case IR::Opcode::IAdd64:
|
case IR::Opcode::IAdd64:
|
||||||
return FoldAdd<u64>(inst);
|
return FoldAdd<u64>(inst);
|
||||||
case IR::Opcode::BitFieldUExtract:
|
case IR::Opcode::Select32:
|
||||||
if (inst.AreAllArgsImmediates() && !inst.HasAssociatedPseudoOperation()) {
|
return FoldSelect<u32>(inst);
|
||||||
inst.ReplaceUsesWith(IR::Value{
|
|
||||||
BitFieldUExtract(inst.Arg(0).U32(), inst.Arg(1).U32(), inst.Arg(2).U32())});
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
case IR::Opcode::LogicalAnd:
|
case IR::Opcode::LogicalAnd:
|
||||||
return FoldLogicalAnd(inst);
|
return FoldLogicalAnd(inst);
|
||||||
|
case IR::Opcode::ULessThan:
|
||||||
|
return FoldWhenAllImmediates(inst, [](u32 a, u32 b) { return a < b; });
|
||||||
|
case IR::Opcode::BitFieldUExtract:
|
||||||
|
return FoldWhenAllImmediates(inst, [](u32 base, u32 shift, u32 count) {
|
||||||
|
if (static_cast<size_t>(shift) + static_cast<size_t>(count) > Common::BitSize<u32>()) {
|
||||||
|
throw LogicError("Undefined result in {}({}, {}, {})", IR::Opcode::BitFieldUExtract,
|
||||||
|
base, shift, count);
|
||||||
|
}
|
||||||
|
return (base >> shift) & ((1U << count) - 1);
|
||||||
|
});
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue