Compare commits

...

10 commits

14 changed files with 152 additions and 222 deletions

View file

@ -1,139 +0,0 @@
// SPDX-FileCopyrightText: Copyright 2021 yuzu Emulator Project
// SPDX-License-Identifier: GPL-2.0-or-later
#pragma once
#include <deque>
#include <memory>
#include <type_traits>
#include "common/common_types.h"
namespace Common {
template <class Traits>
class LeastRecentlyUsedCache {
using ObjectType = typename Traits::ObjectType;
using TickType = typename Traits::TickType;
struct Item {
ObjectType obj;
TickType tick;
Item* next{};
Item* prev{};
};
public:
LeastRecentlyUsedCache() : first_item{}, last_item{} {}
~LeastRecentlyUsedCache() = default;
size_t Insert(ObjectType obj, TickType tick) {
const auto new_id = Build();
auto& item = item_pool[new_id];
item.obj = obj;
item.tick = tick;
Attach(item);
return new_id;
}
void Touch(size_t id, TickType tick) {
auto& item = item_pool[id];
if (item.tick >= tick) {
return;
}
item.tick = tick;
if (&item == last_item) {
return;
}
Detach(item);
Attach(item);
}
void Free(size_t id) {
auto& item = item_pool[id];
Detach(item);
item.prev = nullptr;
item.next = nullptr;
free_items.push_back(id);
}
template <typename Func>
void ForEachItemBelow(TickType tick, Func&& func) {
static constexpr bool RETURNS_BOOL =
std::is_same_v<std::invoke_result<Func, ObjectType>, bool>;
Item* iterator = first_item;
while (iterator) {
if (static_cast<s64>(tick) - static_cast<s64>(iterator->tick) < 0) {
return;
}
Item* next = iterator->next;
if constexpr (RETURNS_BOOL) {
if (func(iterator->obj)) {
return;
}
} else {
func(iterator->obj);
}
iterator = next;
}
}
private:
size_t Build() {
if (free_items.empty()) {
const size_t item_id = item_pool.size();
auto& item = item_pool.emplace_back();
item.next = nullptr;
item.prev = nullptr;
return item_id;
}
const size_t item_id = free_items.front();
free_items.pop_front();
auto& item = item_pool[item_id];
item.next = nullptr;
item.prev = nullptr;
return item_id;
}
void Attach(Item& item) {
if (!first_item) {
first_item = &item;
}
if (!last_item) {
last_item = &item;
} else {
item.prev = last_item;
last_item->next = &item;
item.next = nullptr;
last_item = &item;
}
}
void Detach(Item& item) {
if (item.prev) {
item.prev->next = item.next;
}
if (item.next) {
item.next->prev = item.prev;
}
if (&item == first_item) {
first_item = item.next;
if (first_item) {
first_item->prev = nullptr;
}
}
if (&item == last_item) {
last_item = item.prev;
if (last_item) {
last_item->next = nullptr;
}
}
}
std::deque<Item> item_pool;
std::deque<size_t> free_items;
Item* first_item{};
Item* last_item{};
};
} // namespace Common

View file

@ -109,12 +109,12 @@ public:
return static_cast<u32>(other_cpu_addr - cpu_addr); return static_cast<u32>(other_cpu_addr - cpu_addr);
} }
size_t getLRUID() const noexcept { u64 GetFrameTick() const noexcept {
return lru_id; return frame_tick;
} }
void setLRUID(size_t lru_id_) { void SetFrameTick(u64 tick) noexcept {
lru_id = lru_id_; frame_tick = tick;
} }
size_t SizeBytes() const { size_t SizeBytes() const {
@ -125,7 +125,7 @@ private:
VAddr cpu_addr = 0; VAddr cpu_addr = 0;
BufferFlagBits flags{}; BufferFlagBits flags{};
int stream_score = 0; int stream_score = 0;
size_t lru_id = SIZE_MAX; u64 frame_tick = 0;
size_t size_bytes = 0; size_t size_bytes = 0;
}; };

View file

@ -58,17 +58,22 @@ void BufferCache<P>::RunGarbageCollector() {
const bool aggressive_gc = total_used_memory >= critical_memory; const bool aggressive_gc = total_used_memory >= critical_memory;
const u64 ticks_to_destroy = aggressive_gc ? 60 : 120; const u64 ticks_to_destroy = aggressive_gc ? 60 : 120;
int num_iterations = aggressive_gc ? 64 : 32; int num_iterations = aggressive_gc ? 64 : 32;
const auto clean_up = [this, &num_iterations](BufferId buffer_id) { const u64 threshold = frame_tick - ticks_to_destroy;
boost::container::small_vector<BufferId, 64> expired;
for (auto [id, buffer] : slot_buffers) {
if (buffer->GetFrameTick() < threshold) {
expired.push_back(id);
}
}
for (const auto buffer_id : expired) {
if (num_iterations == 0) { if (num_iterations == 0) {
return true; break;
} }
--num_iterations; --num_iterations;
auto& buffer = slot_buffers[buffer_id]; auto& buffer = slot_buffers[buffer_id];
DownloadBufferMemory(buffer); DownloadBufferMemory(buffer);
DeleteBuffer(buffer_id); DeleteBuffer(buffer_id);
return false; }
};
lru_cache.ForEachItemBelow(frame_tick - ticks_to_destroy, clean_up);
} }
template <class P> template <class P>
@ -1595,10 +1600,9 @@ void BufferCache<P>::ChangeRegister(BufferId buffer_id) {
const auto size = buffer.SizeBytes(); const auto size = buffer.SizeBytes();
if (insert) { if (insert) {
total_used_memory += Common::AlignUp(size, 1024); total_used_memory += Common::AlignUp(size, 1024);
buffer.setLRUID(lru_cache.Insert(buffer_id, frame_tick)); buffer.SetFrameTick(frame_tick);
} else { } else {
total_used_memory -= Common::AlignUp(size, 1024); total_used_memory -= Common::AlignUp(size, 1024);
lru_cache.Free(buffer.getLRUID());
} }
const DAddr device_addr_begin = buffer.CpuAddr(); const DAddr device_addr_begin = buffer.CpuAddr();
const DAddr device_addr_end = device_addr_begin + size; const DAddr device_addr_end = device_addr_begin + size;
@ -1616,7 +1620,7 @@ void BufferCache<P>::ChangeRegister(BufferId buffer_id) {
template <class P> template <class P>
void BufferCache<P>::TouchBuffer(Buffer& buffer, BufferId buffer_id) noexcept { void BufferCache<P>::TouchBuffer(Buffer& buffer, BufferId buffer_id) noexcept {
if (buffer_id != NULL_BUFFER_ID) { if (buffer_id != NULL_BUFFER_ID) {
lru_cache.Touch(buffer.getLRUID(), frame_tick); buffer.SetFrameTick(frame_tick);
} }
} }

View file

@ -23,7 +23,6 @@
#include "common/common_types.h" #include "common/common_types.h"
#include "common/div_ceil.h" #include "common/div_ceil.h"
#include "common/literals.h" #include "common/literals.h"
#include "common/lru_cache.h"
#include "common/range_sets.h" #include "common/range_sets.h"
#include "common/scope_exit.h" #include "common/scope_exit.h"
#include "common/settings.h" #include "common/settings.h"
@ -506,11 +505,6 @@ private:
size_t immediate_buffer_capacity = 0; size_t immediate_buffer_capacity = 0;
Common::ScratchBuffer<u8> immediate_buffer_alloc; Common::ScratchBuffer<u8> immediate_buffer_alloc;
struct LRUItemParams {
using ObjectType = BufferId;
using TickType = u64;
};
Common::LeastRecentlyUsedCache<LRUItemParams> lru_cache;
u64 frame_tick = 0; u64 frame_tick = 0;
u64 total_used_memory = 0; u64 total_used_memory = 0;
u64 minimum_memory = 0; u64 minimum_memory = 0;

View file

@ -1,4 +1,4 @@
// SPDX-FileCopyrightText: Copyright 2025 Eden Emulator Project // SPDX-FileCopyrightText: Copyright 2026 Eden Emulator Project
// SPDX-License-Identifier: GPL-3.0-or-later // SPDX-License-Identifier: GPL-3.0-or-later
// SPDX-FileCopyrightText: Copyright 2021 yuzu Emulator Project // SPDX-FileCopyrightText: Copyright 2021 yuzu Emulator Project
@ -122,7 +122,35 @@ void DmaPusher::ProcessCommands(std::span<const CommandHeader> commands) {
dma_state.is_last_call = true; dma_state.is_last_call = true;
index += max_write; index += max_write;
} else if (dma_state.method_count) { } else if (dma_state.method_count) {
auto const command_header = commands[index]; //can copy if (!dma_state.non_incrementing && !dma_increment_once &&
dma_state.method >= non_puller_methods) {
auto subchannel = subchannels[dma_state.subchannel];
const u32 available = u32(std::min<size_t>(
index + dma_state.method_count, commands.size()) - index);
u32 batch = 0;
u32 method = dma_state.method;
while (batch < available) {
const bool needs_exec =
(method < Engines::EngineInterface::EXECUTION_MASK_TABLE_SIZE)
? subchannel->execution_mask[method]
: subchannel->execution_mask_default;
if (needs_exec) break;
batch++;
method++;
}
if (batch > 0) {
auto& sink = subchannel->method_sink;
sink.reserve(sink.size() + batch);
for (u32 j = 0; j < batch; j++) {
sink.emplace_back(dma_state.method + j, commands[index + j].argument);
}
dma_state.method += batch;
dma_state.method_count -= batch;
index += batch;
continue;
}
}
auto const command_header = commands[index];
dma_state.dma_word_offset = u32(index * sizeof(u32)); dma_state.dma_word_offset = u32(index * sizeof(u32));
dma_state.is_last_call = dma_state.method_count <= 1; dma_state.is_last_call = dma_state.method_count <= 1;
CallMethod(command_header.argument); CallMethod(command_header.argument);
@ -181,7 +209,11 @@ void DmaPusher::CallMethod(u32 argument) const {
}); });
} else { } else {
auto subchannel = subchannels[dma_state.subchannel]; auto subchannel = subchannels[dma_state.subchannel];
if (!subchannel->execution_mask[dma_state.method]) { const bool needs_execution =
(dma_state.method < Engines::EngineInterface::EXECUTION_MASK_TABLE_SIZE)
? subchannel->execution_mask[dma_state.method]
: subchannel->execution_mask_default;
if (!needs_execution) {
subchannel->method_sink.emplace_back(dma_state.method, argument); subchannel->method_sink.emplace_back(dma_state.method, argument);
} else { } else {
subchannel->ConsumeSink(); subchannel->ConsumeSink();

View file

@ -6,9 +6,9 @@
#pragma once #pragma once
#include <bitset> #include <array>
#include <limits>
#include <vector> #include <boost/container/small_vector.hpp>
#include "common/common_types.h" #include "common/common_types.h"
@ -41,8 +41,11 @@ public:
ConsumeSinkImpl(); ConsumeSinkImpl();
} }
std::bitset<(std::numeric_limits<u16>::max)()> execution_mask{}; static constexpr size_t EXECUTION_MASK_TABLE_SIZE = 0xE00;
std::vector<std::pair<u32, u32>> method_sink{};
std::array<u8, EXECUTION_MASK_TABLE_SIZE> execution_mask{};
bool execution_mask_default{};
boost::container::small_vector<std::pair<u32, u32>, 64> method_sink{};
bool current_dirty{}; bool current_dirty{};
GPUVAddr current_dma_segment; GPUVAddr current_dma_segment;

View file

@ -26,7 +26,7 @@ Fermi2D::Fermi2D(MemoryManager& memory_manager_) : memory_manager{memory_manager
regs.src.depth = 1; regs.src.depth = 1;
regs.dst.depth = 1; regs.dst.depth = 1;
execution_mask.reset(); execution_mask.fill(0);
execution_mask[FERMI2D_REG_INDEX(pixels_from_memory.src_y0) + 1] = true; execution_mask[FERMI2D_REG_INDEX(pixels_from_memory.src_y0) + 1] = true;
} }

View file

@ -18,7 +18,7 @@ namespace Tegra::Engines {
KeplerCompute::KeplerCompute(Core::System& system_, MemoryManager& memory_manager_) KeplerCompute::KeplerCompute(Core::System& system_, MemoryManager& memory_manager_)
: system{system_}, memory_manager{memory_manager_}, upload_state{memory_manager, regs.upload} { : system{system_}, memory_manager{memory_manager_}, upload_state{memory_manager, regs.upload} {
execution_mask.reset(); execution_mask.fill(0);
execution_mask[KEPLER_COMPUTE_REG_INDEX(exec_upload)] = true; execution_mask[KEPLER_COMPUTE_REG_INDEX(exec_upload)] = true;
execution_mask[KEPLER_COMPUTE_REG_INDEX(data_upload)] = true; execution_mask[KEPLER_COMPUTE_REG_INDEX(data_upload)] = true;
execution_mask[KEPLER_COMPUTE_REG_INDEX(launch)] = true; execution_mask[KEPLER_COMPUTE_REG_INDEX(launch)] = true;

View file

@ -22,7 +22,7 @@ KeplerMemory::~KeplerMemory() = default;
void KeplerMemory::BindRasterizer(VideoCore::RasterizerInterface* rasterizer_) { void KeplerMemory::BindRasterizer(VideoCore::RasterizerInterface* rasterizer_) {
upload_state.BindRasterizer(rasterizer_); upload_state.BindRasterizer(rasterizer_);
execution_mask.reset(); execution_mask.fill(0);
execution_mask[KEPLERMEMORY_REG_INDEX(exec)] = true; execution_mask[KEPLERMEMORY_REG_INDEX(exec)] = true;
execution_mask[KEPLERMEMORY_REG_INDEX(data)] = true; execution_mask[KEPLERMEMORY_REG_INDEX(data)] = true;
} }

View file

@ -4,8 +4,10 @@
// SPDX-FileCopyrightText: Copyright 2018 yuzu Emulator Project // SPDX-FileCopyrightText: Copyright 2018 yuzu Emulator Project
// SPDX-License-Identifier: GPL-2.0-or-later // SPDX-License-Identifier: GPL-2.0-or-later
#include <algorithm>
#include <cstring> #include <cstring>
#include <optional> #include <optional>
#include "common/assert.h" #include "common/assert.h"
#include "common/bit_util.h" #include "common/bit_util.h"
#include "common/scope_exit.h" #include "common/scope_exit.h"
@ -37,9 +39,10 @@ Maxwell3D::Maxwell3D(Core::System& system_, MemoryManager& memory_manager_)
{ {
dirty.flags.flip(); dirty.flags.flip();
InitializeRegisterDefaults(); InitializeRegisterDefaults();
execution_mask.reset(); execution_mask.fill(0);
for (size_t i = 0; i < execution_mask.size(); i++) for (size_t i = 0; i < EXECUTION_MASK_TABLE_SIZE; i++)
execution_mask[i] = IsMethodExecutable(u32(i)); execution_mask[i] = IsMethodExecutable(u32(i));
execution_mask_default = true;
} }
Maxwell3D::~Maxwell3D() = default; Maxwell3D::~Maxwell3D() = default;
@ -298,18 +301,27 @@ u32 Maxwell3D::ProcessShadowRam(u32 method, u32 argument) {
} }
void Maxwell3D::ConsumeSinkImpl() { void Maxwell3D::ConsumeSinkImpl() {
std::stable_sort(method_sink.begin(), method_sink.end(),
[](const auto& a, const auto& b) { return a.first < b.first; });
const auto sink_size = method_sink.size();
const auto control = shadow_state.shadow_ram_control; const auto control = shadow_state.shadow_ram_control;
if (control == Regs::ShadowRamControl::Track || control == Regs::ShadowRamControl::TrackWithFilter) { if (control == Regs::ShadowRamControl::Track || control == Regs::ShadowRamControl::TrackWithFilter) {
for (auto [method, value] : method_sink) { for (size_t i = 0; i < sink_size; ++i) {
const auto [method, value] = method_sink[i];
shadow_state.reg_array[method] = value; shadow_state.reg_array[method] = value;
ProcessDirtyRegisters(method, value); ProcessDirtyRegisters(method, value);
} }
} else if (control == Regs::ShadowRamControl::Replay) { } else if (control == Regs::ShadowRamControl::Replay) {
for (auto [method, value] : method_sink) for (size_t i = 0; i < sink_size; ++i) {
const auto [method, value] = method_sink[i];
ProcessDirtyRegisters(method, shadow_state.reg_array[method]); ProcessDirtyRegisters(method, shadow_state.reg_array[method]);
}
} else { } else {
for (auto [method, value] : method_sink) for (size_t i = 0; i < sink_size; ++i) {
const auto [method, value] = method_sink[i];
ProcessDirtyRegisters(method, value); ProcessDirtyRegisters(method, value);
}
} }
method_sink.clear(); method_sink.clear();
} }

View file

@ -23,7 +23,7 @@ using namespace Texture;
MaxwellDMA::MaxwellDMA(Core::System& system_, MemoryManager& memory_manager_) MaxwellDMA::MaxwellDMA(Core::System& system_, MemoryManager& memory_manager_)
: system{system_}, memory_manager{memory_manager_} { : system{system_}, memory_manager{memory_manager_} {
execution_mask.reset(); execution_mask.fill(0);
execution_mask[offsetof(Regs, launch_dma) / sizeof(u32)] = true; execution_mask[offsetof(Regs, launch_dma) / sizeof(u32)] = true;
} }

View file

@ -1,3 +1,6 @@
// SPDX-FileCopyrightText: Copyright 2026 Eden Emulator Project
// SPDX-License-Identifier: GPL-3.0-or-later
// SPDX-FileCopyrightText: Copyright 2020 yuzu Emulator Project // SPDX-FileCopyrightText: Copyright 2020 yuzu Emulator Project
// SPDX-License-Identifier: GPL-2.0-or-later // SPDX-License-Identifier: GPL-2.0-or-later
@ -102,7 +105,7 @@ struct ImageBase {
VAddr cpu_addr_end = 0; VAddr cpu_addr_end = 0;
u64 modification_tick = 0; u64 modification_tick = 0;
size_t lru_index = SIZE_MAX; u64 last_use_tick = 0;
std::array<u32, MAX_MIP_LEVELS> mip_level_offsets{}; std::array<u32, MAX_MIP_LEVELS> mip_level_offsets{};

View file

@ -6,6 +6,7 @@
#pragma once #pragma once
#include <algorithm>
#include <limits> #include <limits>
#include <optional> #include <optional>
#include <bit> #include <bit>
@ -118,24 +119,17 @@ TextureCache<P>::TextureCache(Runtime& runtime_, Tegra::MaxwellDeviceMemoryManag
template <class P> template <class P>
void TextureCache<P>::RunGarbageCollector() { void TextureCache<P>::RunGarbageCollector() {
bool high_priority_mode = false; bool high_priority_mode = total_used_memory >= expected_memory;
bool aggressive_mode = false; bool aggressive_mode = false;
u64 ticks_to_destroy = 0; u64 ticks_to_destroy = high_priority_mode ? 25ULL : 50ULL;
size_t num_iterations = 0; size_t num_iterations = high_priority_mode ? 20 : 10;
const auto Configure = [&](bool allow_aggressive) {
high_priority_mode = total_used_memory >= expected_memory;
aggressive_mode = allow_aggressive && total_used_memory >= critical_memory;
ticks_to_destroy = aggressive_mode ? 10ULL : high_priority_mode ? 25ULL : 50ULL;
num_iterations = aggressive_mode ? 40 : (high_priority_mode ? 20 : 10);
};
const auto Cleanup = [this, &num_iterations, &high_priority_mode, const auto Cleanup = [this, &num_iterations, &high_priority_mode,
&aggressive_mode](ImageId image_id) { &aggressive_mode](ImageId image_id) {
if (num_iterations == 0) { if (num_iterations == 0) {
return true; return true;
} }
--num_iterations;
auto& image = slot_images[image_id]; auto& image = slot_images[image_id];
// Never delete recently allocated sparse textures (within 3 frames) // Never delete recently allocated sparse textures (within 3 frames)
@ -145,12 +139,9 @@ void TextureCache<P>::RunGarbageCollector() {
} }
if (True(image.flags & ImageFlagBits::IsDecoding)) { if (True(image.flags & ImageFlagBits::IsDecoding)) {
// This image is still being decoded, deleting it will invalidate the slot
// used by the async decoder thread.
return false; return false;
} }
// Prioritize large sparse textures for cleanup
const bool is_large_sparse = lowmemorydevice && const bool is_large_sparse = lowmemorydevice &&
image.info.is_sparse && image.info.is_sparse &&
image.guest_size_bytes >= 256_MiB; image.guest_size_bytes >= 256_MiB;
@ -166,6 +157,8 @@ void TextureCache<P>::RunGarbageCollector() {
return false; return false;
} }
--num_iterations;
if (must_download && !is_large_sparse) { if (must_download && !is_large_sparse) {
auto map = runtime.DownloadStagingBuffer(image.unswizzled_size_bytes); auto map = runtime.DownloadStagingBuffer(image.unswizzled_size_bytes);
const auto copies = FixSmallVectorADL(FullDownloadCopies(image.info)); const auto copies = FixSmallVectorADL(FullDownloadCopies(image.info));
@ -183,7 +176,6 @@ void TextureCache<P>::RunGarbageCollector() {
if (total_used_memory < critical_memory) { if (total_used_memory < critical_memory) {
if (aggressive_mode) { if (aggressive_mode) {
// Sink the aggresiveness.
num_iterations >>= 2; num_iterations >>= 2;
aggressive_mode = false; aggressive_mode = false;
return false; return false;
@ -196,31 +188,64 @@ void TextureCache<P>::RunGarbageCollector() {
return false; return false;
}; };
// Aggressively clear massive sparse textures const auto SortByAge = [this](auto& vec) {
if (total_used_memory >= expected_memory) { std::sort(vec.begin(), vec.end(), [this](ImageId a, ImageId b) {
lru_cache.ForEachItemBelow(frame_tick, [&](ImageId image_id) { return slot_images[a].last_use_tick < slot_images[b].last_use_tick;
auto& image = slot_images[image_id];
// Only target sparse textures that are old enough
if (lowmemorydevice &&
image.info.is_sparse &&
image.guest_size_bytes >= 256_MiB &&
image.allocation_tick < frame_tick - 3) {
LOG_DEBUG(HW_GPU, "GC targeting old sparse texture at 0x{:X} ({} MiB, age: {} frames)",
image.gpu_addr, image.guest_size_bytes / (1024 * 1024),
frame_tick - image.allocation_tick);
return Cleanup(image_id);
}
return false;
}); });
};
// Single pass: collect all candidates, classified by tier
const u64 normal_threshold = frame_tick > ticks_to_destroy ? frame_tick - ticks_to_destroy : 0;
const u64 aggressive_threshold = frame_tick > 10 ? frame_tick - 10 : 0;
boost::container::small_vector<ImageId, 64> sparse_candidates;
boost::container::small_vector<ImageId, 64> expired;
boost::container::small_vector<ImageId, 64> aggressive_expired;
for (auto [id, image] : slot_images) {
if (False(image->flags & ImageFlagBits::Registered)) {
continue;
}
const u64 tick = image->last_use_tick;
if (tick < normal_threshold) {
expired.push_back(id);
} else if (tick < aggressive_threshold) {
aggressive_expired.push_back(id);
} else if (high_priority_mode && tick < frame_tick &&
lowmemorydevice && image->info.is_sparse &&
image->guest_size_bytes >= 256_MiB) {
sparse_candidates.push_back(id);
}
} }
Configure(false); SortByAge(expired);
lru_cache.ForEachItemBelow(frame_tick - ticks_to_destroy, Cleanup); SortByAge(aggressive_expired);
// If pressure is still too high, prune aggressively. // Tier 1: large sparse textures under memory pressure
for (const auto image_id : sparse_candidates) {
auto& image = slot_images[image_id];
if (image.allocation_tick < frame_tick - 3) {
if (Cleanup(image_id)) {
break;
}
}
}
// Tier 2: normal expiration
for (const auto image_id : expired) {
if (Cleanup(image_id)) {
break;
}
}
// Tier 3: if still critical, use aggressive threshold with more iterations
if (total_used_memory >= critical_memory) { if (total_used_memory >= critical_memory) {
Configure(true); aggressive_mode = true;
lru_cache.ForEachItemBelow(frame_tick - ticks_to_destroy, Cleanup); num_iterations = 40;
for (const auto image_id : aggressive_expired) {
if (Cleanup(image_id)) {
break;
}
}
} }
} }
@ -2027,8 +2052,8 @@ std::pair<u32, u32> TextureCache<P>::PrepareDmaImage(ImageId dst_id, GPUVAddr ba
const auto& image = slot_images[dst_id]; const auto& image = slot_images[dst_id];
const auto base = image.TryFindBase(base_addr); const auto base = image.TryFindBase(base_addr);
PrepareImage(dst_id, mark_as_modified, false); PrepareImage(dst_id, mark_as_modified, false);
const auto& new_image = slot_images[dst_id]; auto& new_image = slot_images[dst_id];
lru_cache.Touch(new_image.lru_index, frame_tick); new_image.last_use_tick = frame_tick;
return std::make_pair(base->level, base->layer); return std::make_pair(base->level, base->layer);
} }
@ -2377,7 +2402,7 @@ void TextureCache<P>::RegisterImage(ImageId image_id) {
tentative_size = TranscodedAstcSize(tentative_size, image.info.format); tentative_size = TranscodedAstcSize(tentative_size, image.info.format);
} }
total_used_memory += Common::AlignUp(tentative_size, 1024); total_used_memory += Common::AlignUp(tentative_size, 1024);
image.lru_index = lru_cache.Insert(image_id, frame_tick); image.last_use_tick = frame_tick;
ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, image_id](u64 page) { ForEachGPUPage(image.gpu_addr, image.guest_size_bytes, [this, image_id](u64 page) {
(*channel_state->gpu_page_table)[page].push_back(image_id); (*channel_state->gpu_page_table)[page].push_back(image_id);
@ -2411,7 +2436,7 @@ void TextureCache<P>::UnregisterImage(ImageId image_id) {
"Trying to unregister an already registered image"); "Trying to unregister an already registered image");
image.flags &= ~ImageFlagBits::Registered; image.flags &= ~ImageFlagBits::Registered;
image.flags &= ~ImageFlagBits::BadOverlap; image.flags &= ~ImageFlagBits::BadOverlap;
lru_cache.Free(image.lru_index);
const auto& clear_page_table = const auto& clear_page_table =
[image_id](u64 page, ankerl::unordered_dense::map<u64, std::vector<ImageId>, Common::IdentityHash<u64>>& selected_page_table) { [image_id](u64 page, ankerl::unordered_dense::map<u64, std::vector<ImageId>, Common::IdentityHash<u64>>& selected_page_table) {
const auto page_it = selected_page_table.find(page); const auto page_it = selected_page_table.find(page);
@ -2738,7 +2763,7 @@ void TextureCache<P>::PrepareImage(ImageId image_id, bool is_modification, bool
if (is_modification) { if (is_modification) {
MarkModification(image); MarkModification(image);
} }
lru_cache.Touch(image.lru_index, frame_tick); image.last_use_tick = frame_tick;
} }
template <class P> template <class P>

View file

@ -22,7 +22,7 @@
#include "common/common_types.h" #include "common/common_types.h"
#include "common/hash.h" #include "common/hash.h"
#include "common/literals.h" #include "common/literals.h"
#include "common/lru_cache.h"
#include <ranges> #include <ranges>
#include "common/scratch_buffer.h" #include "common/scratch_buffer.h"
#include "common/slot_vector.h" #include "common/slot_vector.h"
@ -510,11 +510,7 @@ private:
std::deque<std::vector<AsyncBuffer>> async_buffers; std::deque<std::vector<AsyncBuffer>> async_buffers;
std::deque<AsyncBuffer> async_buffers_death_ring; std::deque<AsyncBuffer> async_buffers_death_ring;
struct LRUItemParams {
using ObjectType = ImageId;
using TickType = u64;
};
Common::LeastRecentlyUsedCache<LRUItemParams> lru_cache;
#ifdef YUZU_LEGACY #ifdef YUZU_LEGACY
static constexpr size_t TICKS_TO_DESTROY = 6; static constexpr size_t TICKS_TO_DESTROY = 6;