|
|
|
@ -76,8 +76,9 @@ class BufferCache { |
|
|
|
|
|
|
|
static constexpr BufferId NULL_BUFFER_ID{0}; |
|
|
|
|
|
|
|
static constexpr u64 EXPECTED_MEMORY = 512_MiB; |
|
|
|
static constexpr u64 CRITICAL_MEMORY = 1_GiB; |
|
|
|
static constexpr s64 DEFAULT_EXPECTED_MEMORY = 512_MiB; |
|
|
|
static constexpr s64 DEFAULT_CRITICAL_MEMORY = 1_GiB; |
|
|
|
static constexpr s64 TARGET_THRESHOLD = 4_GiB; |
|
|
|
|
|
|
|
using Maxwell = Tegra::Engines::Maxwell3D::Regs; |
|
|
|
|
|
|
|
@ -436,6 +437,8 @@ private: |
|
|
|
Common::LeastRecentlyUsedCache<LRUItemParams> lru_cache; |
|
|
|
u64 frame_tick = 0; |
|
|
|
u64 total_used_memory = 0; |
|
|
|
u64 minimum_memory = 0; |
|
|
|
u64 critical_memory = 0; |
|
|
|
|
|
|
|
std::array<BufferId, ((1ULL << 39) >> PAGE_BITS)> page_table; |
|
|
|
}; |
|
|
|
@ -451,11 +454,30 @@ BufferCache<P>::BufferCache(VideoCore::RasterizerInterface& rasterizer_, |
|
|
|
// Ensure the first slot is used for the null buffer |
|
|
|
void(slot_buffers.insert(runtime, NullBufferParams{})); |
|
|
|
common_ranges.clear(); |
|
|
|
|
|
|
|
if (!runtime.CanReportMemoryUsage()) { |
|
|
|
minimum_memory = DEFAULT_EXPECTED_MEMORY; |
|
|
|
critical_memory = DEFAULT_CRITICAL_MEMORY; |
|
|
|
return; |
|
|
|
} |
|
|
|
|
|
|
|
const s64 device_memory = static_cast<s64>(runtime.GetDeviceLocalMemory()); |
|
|
|
const s64 min_spacing_expected = device_memory - 1_GiB - 512_MiB; |
|
|
|
const s64 min_spacing_critical = device_memory - 1_GiB; |
|
|
|
const s64 mem_tresshold = std::min(device_memory, TARGET_THRESHOLD); |
|
|
|
const s64 min_vacancy_expected = (6 * mem_tresshold) / 10; |
|
|
|
const s64 min_vacancy_critical = (3 * mem_tresshold) / 10; |
|
|
|
minimum_memory = static_cast<u64>( |
|
|
|
std::max(std::min(device_memory - min_vacancy_expected, min_spacing_expected), |
|
|
|
DEFAULT_EXPECTED_MEMORY)); |
|
|
|
critical_memory = static_cast<u64>( |
|
|
|
std::max(std::min(device_memory - min_vacancy_critical, min_spacing_critical), |
|
|
|
DEFAULT_CRITICAL_MEMORY)); |
|
|
|
} |
|
|
|
|
|
|
|
template <class P> |
|
|
|
void BufferCache<P>::RunGarbageCollector() { |
|
|
|
const bool aggressive_gc = total_used_memory >= CRITICAL_MEMORY; |
|
|
|
const bool aggressive_gc = total_used_memory >= critical_memory; |
|
|
|
const u64 ticks_to_destroy = aggressive_gc ? 60 : 120; |
|
|
|
int num_iterations = aggressive_gc ? 64 : 32; |
|
|
|
const auto clean_up = [this, &num_iterations](BufferId buffer_id) { |
|
|
|
@ -486,7 +508,11 @@ void BufferCache<P>::TickFrame() { |
|
|
|
const bool skip_preferred = hits * 256 < shots * 251; |
|
|
|
uniform_buffer_skip_cache_size = skip_preferred ? DEFAULT_SKIP_CACHE_SIZE : 0; |
|
|
|
|
|
|
|
if (total_used_memory >= EXPECTED_MEMORY) { |
|
|
|
// If we can obtain the memory info, use it instead of the estimate. |
|
|
|
if (runtime.CanReportMemoryUsage()) { |
|
|
|
total_used_memory = runtime.GetDeviceMemoryUsage(); |
|
|
|
} |
|
|
|
if (total_used_memory >= minimum_memory) { |
|
|
|
RunGarbageCollector(); |
|
|
|
} |
|
|
|
++frame_tick; |
|
|
|
|