|
|
@ -133,11 +133,11 @@ public: |
|
|
regs.zeta.memory_layout.block_depth, regs.zeta.memory_layout.type)}; |
|
|
regs.zeta.memory_layout.block_depth, regs.zeta.memory_layout.type)}; |
|
|
auto surface_view = GetSurface(gpu_addr, depth_params, preserve_contents, true); |
|
|
auto surface_view = GetSurface(gpu_addr, depth_params, preserve_contents, true); |
|
|
if (depth_buffer.target) |
|
|
if (depth_buffer.target) |
|
|
depth_buffer.target->MarkAsRenderTarget(false, -1); |
|
|
|
|
|
|
|
|
depth_buffer.target->MarkAsRenderTarget(false, NO_RT); |
|
|
depth_buffer.target = surface_view.first; |
|
|
depth_buffer.target = surface_view.first; |
|
|
depth_buffer.view = surface_view.second; |
|
|
depth_buffer.view = surface_view.second; |
|
|
if (depth_buffer.target) |
|
|
if (depth_buffer.target) |
|
|
depth_buffer.target->MarkAsRenderTarget(true, 8); |
|
|
|
|
|
|
|
|
depth_buffer.target->MarkAsRenderTarget(true, DEPTH_RT); |
|
|
return surface_view.second; |
|
|
return surface_view.second; |
|
|
} |
|
|
} |
|
|
|
|
|
|
|
|
@ -167,7 +167,7 @@ public: |
|
|
auto surface_view = GetSurface(gpu_addr, SurfaceParams::CreateForFramebuffer(system, index), |
|
|
auto surface_view = GetSurface(gpu_addr, SurfaceParams::CreateForFramebuffer(system, index), |
|
|
preserve_contents, true); |
|
|
preserve_contents, true); |
|
|
if (render_targets[index].target) |
|
|
if (render_targets[index].target) |
|
|
render_targets[index].target->MarkAsRenderTarget(false, -1); |
|
|
|
|
|
|
|
|
render_targets[index].target->MarkAsRenderTarget(false, NO_RT); |
|
|
render_targets[index].target = surface_view.first; |
|
|
render_targets[index].target = surface_view.first; |
|
|
render_targets[index].view = surface_view.second; |
|
|
render_targets[index].view = surface_view.second; |
|
|
if (render_targets[index].target) |
|
|
if (render_targets[index].target) |
|
|
@ -191,7 +191,7 @@ public: |
|
|
if (depth_buffer.target == nullptr) { |
|
|
if (depth_buffer.target == nullptr) { |
|
|
return; |
|
|
return; |
|
|
} |
|
|
} |
|
|
depth_buffer.target->MarkAsRenderTarget(false, -1); |
|
|
|
|
|
|
|
|
depth_buffer.target->MarkAsRenderTarget(false, NO_RT); |
|
|
depth_buffer.target = nullptr; |
|
|
depth_buffer.target = nullptr; |
|
|
depth_buffer.view = nullptr; |
|
|
depth_buffer.view = nullptr; |
|
|
} |
|
|
} |
|
|
@ -200,7 +200,7 @@ public: |
|
|
if (render_targets[index].target == nullptr) { |
|
|
if (render_targets[index].target == nullptr) { |
|
|
return; |
|
|
return; |
|
|
} |
|
|
} |
|
|
render_targets[index].target->MarkAsRenderTarget(false, -1); |
|
|
|
|
|
|
|
|
render_targets[index].target->MarkAsRenderTarget(false, NO_RT); |
|
|
render_targets[index].target = nullptr; |
|
|
render_targets[index].target = nullptr; |
|
|
render_targets[index].view = nullptr; |
|
|
render_targets[index].view = nullptr; |
|
|
} |
|
|
} |
|
|
@ -272,8 +272,8 @@ protected: |
|
|
|
|
|
|
|
|
void ManageRenderTargetUnregister(TSurface& surface) { |
|
|
void ManageRenderTargetUnregister(TSurface& surface) { |
|
|
auto& maxwell3d = system.GPU().Maxwell3D(); |
|
|
auto& maxwell3d = system.GPU().Maxwell3D(); |
|
|
u32 index = surface->GetRenderTarget(); |
|
|
|
|
|
if (index == 8) { |
|
|
|
|
|
|
|
|
const u32 index = surface->GetRenderTarget(); |
|
|
|
|
|
if (index == DEPTH_RT) { |
|
|
maxwell3d.dirty_flags.zeta_buffer = true; |
|
|
maxwell3d.dirty_flags.zeta_buffer = true; |
|
|
} else { |
|
|
} else { |
|
|
maxwell3d.dirty_flags.color_buffer.set(index, true); |
|
|
maxwell3d.dirty_flags.color_buffer.set(index, true); |
|
|
@ -797,6 +797,9 @@ private: |
|
|
static constexpr u64 registry_page_size{1 << registry_page_bits}; |
|
|
static constexpr u64 registry_page_size{1 << registry_page_bits}; |
|
|
std::unordered_map<CacheAddr, std::vector<TSurface>> registry; |
|
|
std::unordered_map<CacheAddr, std::vector<TSurface>> registry; |
|
|
|
|
|
|
|
|
|
|
|
static constexpr u32 DEPTH_RT = 8; |
|
|
|
|
|
static constexpr u32 NO_RT = 0xFFFFFFFF; |
|
|
|
|
|
|
|
|
// The L1 Cache is used for fast texture lookup before checking the overlaps |
|
|
// The L1 Cache is used for fast texture lookup before checking the overlaps |
|
|
// This avoids calculating size and other stuffs. |
|
|
// This avoids calculating size and other stuffs. |
|
|
std::unordered_map<CacheAddr, TSurface> l1_cache; |
|
|
std::unordered_map<CacheAddr, TSurface> l1_cache; |
|
|
|