|
|
|
@ -11,18 +11,39 @@ |
|
|
|
#include "common/assert.h"
|
|
|
|
#include "common/common_types.h"
|
|
|
|
#include "common/div_ceil.h"
|
|
|
|
#include "video_core/host_shaders/astc_decoder_comp_spv.h"
|
|
|
|
#include "video_core/host_shaders/vulkan_quad_indexed_comp_spv.h"
|
|
|
|
#include "video_core/host_shaders/vulkan_uint8_comp_spv.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_compute_pass.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_descriptor_pool.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_scheduler.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_staging_buffer_pool.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_texture_cache.h"
|
|
|
|
#include "video_core/renderer_vulkan/vk_update_descriptor.h"
|
|
|
|
#include "video_core/texture_cache/accelerated_swizzle.h"
|
|
|
|
#include "video_core/texture_cache/types.h"
|
|
|
|
#include "video_core/textures/astc.h"
|
|
|
|
#include "video_core/textures/decoders.h"
|
|
|
|
#include "video_core/vulkan_common/vulkan_device.h"
|
|
|
|
#include "video_core/vulkan_common/vulkan_wrapper.h"
|
|
|
|
|
|
|
|
namespace Vulkan { |
|
|
|
|
|
|
|
using Tegra::Texture::SWIZZLE_TABLE; |
|
|
|
using Tegra::Texture::ASTC::EncodingsValues; |
|
|
|
using namespace Tegra::Texture::ASTC; |
|
|
|
|
|
|
|
namespace { |
|
|
|
|
|
|
|
constexpr u32 ASTC_BINDING_INPUT_BUFFER = 0; |
|
|
|
constexpr u32 ASTC_BINDING_ENC_BUFFER = 1; |
|
|
|
constexpr u32 ASTC_BINDING_6_TO_8_BUFFER = 2; |
|
|
|
constexpr u32 ASTC_BINDING_7_TO_8_BUFFER = 3; |
|
|
|
constexpr u32 ASTC_BINDING_8_TO_8_BUFFER = 4; |
|
|
|
constexpr u32 ASTC_BINDING_BYTE_TO_16_BUFFER = 5; |
|
|
|
constexpr u32 ASTC_BINDING_SWIZZLE_BUFFER = 6; |
|
|
|
constexpr u32 ASTC_BINDING_OUTPUT_IMAGE = 7; |
|
|
|
|
|
|
|
VkPushConstantRange BuildComputePushConstantRange(std::size_t size) { |
|
|
|
return { |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
@ -50,6 +71,67 @@ std::array<VkDescriptorSetLayoutBinding, 2> BuildInputOutputDescriptorSetBinding |
|
|
|
}}; |
|
|
|
} |
|
|
|
|
|
|
|
std::array<VkDescriptorSetLayoutBinding, 8> BuildASTCDescriptorSetBindings() { |
|
|
|
return {{ |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_INPUT_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_ENC_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_6_TO_8_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_7_TO_8_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_8_TO_8_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_BYTE_TO_16_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_SWIZZLE_BUFFER, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
{ |
|
|
|
.binding = ASTC_BINDING_OUTPUT_IMAGE, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, |
|
|
|
.descriptorCount = 1, |
|
|
|
.stageFlags = VK_SHADER_STAGE_COMPUTE_BIT, |
|
|
|
.pImmutableSamplers = nullptr, |
|
|
|
}, |
|
|
|
}}; |
|
|
|
} |
|
|
|
|
|
|
|
VkDescriptorUpdateTemplateEntryKHR BuildInputOutputDescriptorUpdateTemplate() { |
|
|
|
return { |
|
|
|
.dstBinding = 0, |
|
|
|
@ -61,6 +143,94 @@ VkDescriptorUpdateTemplateEntryKHR BuildInputOutputDescriptorUpdateTemplate() { |
|
|
|
}; |
|
|
|
} |
|
|
|
|
|
|
|
std::array<VkDescriptorUpdateTemplateEntryKHR, 8> BuildASTCPassDescriptorUpdateTemplateEntry() { |
|
|
|
return {{ |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_INPUT_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_INPUT_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_ENC_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_ENC_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_6_TO_8_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_6_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_7_TO_8_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_7_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_8_TO_8_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_8_TO_8_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_BYTE_TO_16_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_BYTE_TO_16_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_SWIZZLE_BUFFER, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, |
|
|
|
.offset = ASTC_BINDING_SWIZZLE_BUFFER * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
{ |
|
|
|
.dstBinding = ASTC_BINDING_OUTPUT_IMAGE, |
|
|
|
.dstArrayElement = 0, |
|
|
|
.descriptorCount = 1, |
|
|
|
.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, |
|
|
|
.offset = ASTC_BINDING_OUTPUT_IMAGE * sizeof(DescriptorUpdateEntry), |
|
|
|
.stride = sizeof(DescriptorUpdateEntry), |
|
|
|
}, |
|
|
|
}}; |
|
|
|
} |
|
|
|
|
|
|
|
struct AstcPushConstants { |
|
|
|
std::array<u32, 2> blocks_dims; |
|
|
|
u32 bytes_per_block_log2; |
|
|
|
u32 layer_stride; |
|
|
|
u32 block_size; |
|
|
|
u32 x_shift; |
|
|
|
u32 block_height; |
|
|
|
u32 block_height_mask; |
|
|
|
}; |
|
|
|
|
|
|
|
struct AstcBufferData { |
|
|
|
decltype(SWIZZLE_TABLE) swizzle_table_buffer = SWIZZLE_TABLE; |
|
|
|
decltype(EncodingsValues) encoding_values = EncodingsValues; |
|
|
|
decltype(REPLICATE_6_BIT_TO_8_TABLE) replicate_6_to_8 = REPLICATE_6_BIT_TO_8_TABLE; |
|
|
|
decltype(REPLICATE_7_BIT_TO_8_TABLE) replicate_7_to_8 = REPLICATE_7_BIT_TO_8_TABLE; |
|
|
|
decltype(REPLICATE_8_BIT_TO_8_TABLE) replicate_8_to_8 = REPLICATE_8_BIT_TO_8_TABLE; |
|
|
|
decltype(REPLICATE_BYTE_TO_16_TABLE) replicate_byte_to_16 = REPLICATE_BYTE_TO_16_TABLE; |
|
|
|
} constexpr ASTC_BUFFER_DATA; |
|
|
|
|
|
|
|
} // Anonymous namespace
|
|
|
|
|
|
|
|
VKComputePass::VKComputePass(const Device& device, VKDescriptorPool& descriptor_pool, |
|
|
|
@ -238,4 +408,167 @@ std::pair<VkBuffer, VkDeviceSize> QuadIndexedPass::Assemble( |
|
|
|
return {staging.buffer, staging.offset}; |
|
|
|
} |
|
|
|
|
|
|
|
ASTCDecoderPass::ASTCDecoderPass(const Device& device_, VKScheduler& scheduler_, |
|
|
|
VKDescriptorPool& descriptor_pool_, |
|
|
|
StagingBufferPool& staging_buffer_pool_, |
|
|
|
VKUpdateDescriptorQueue& update_descriptor_queue_, |
|
|
|
MemoryAllocator& memory_allocator_) |
|
|
|
: VKComputePass(device_, descriptor_pool_, BuildASTCDescriptorSetBindings(), |
|
|
|
BuildASTCPassDescriptorUpdateTemplateEntry(), |
|
|
|
BuildComputePushConstantRange(sizeof(AstcPushConstants)), |
|
|
|
ASTC_DECODER_COMP_SPV), |
|
|
|
device{device_}, scheduler{scheduler_}, staging_buffer_pool{staging_buffer_pool_}, |
|
|
|
update_descriptor_queue{update_descriptor_queue_}, memory_allocator{memory_allocator_} {} |
|
|
|
|
|
|
|
ASTCDecoderPass::~ASTCDecoderPass() = default; |
|
|
|
|
|
|
|
void ASTCDecoderPass::MakeDataBuffer() { |
|
|
|
constexpr size_t TOTAL_BUFFER_SIZE = sizeof(ASTC_BUFFER_DATA) + sizeof(SWIZZLE_TABLE); |
|
|
|
data_buffer = device.GetLogical().CreateBuffer(VkBufferCreateInfo{ |
|
|
|
.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO, |
|
|
|
.pNext = nullptr, |
|
|
|
.flags = 0, |
|
|
|
.size = TOTAL_BUFFER_SIZE, |
|
|
|
.usage = VK_BUFFER_USAGE_STORAGE_BUFFER_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT, |
|
|
|
.sharingMode = VK_SHARING_MODE_EXCLUSIVE, |
|
|
|
.queueFamilyIndexCount = 0, |
|
|
|
.pQueueFamilyIndices = nullptr, |
|
|
|
}); |
|
|
|
data_buffer_commit = memory_allocator.Commit(data_buffer, MemoryUsage::Upload); |
|
|
|
|
|
|
|
const auto staging_ref = staging_buffer_pool.Request(TOTAL_BUFFER_SIZE, MemoryUsage::Upload); |
|
|
|
std::memcpy(staging_ref.mapped_span.data(), &ASTC_BUFFER_DATA, sizeof(ASTC_BUFFER_DATA)); |
|
|
|
// Tack on the swizzle table at the end of the buffer
|
|
|
|
std::memcpy(staging_ref.mapped_span.data() + sizeof(ASTC_BUFFER_DATA), &SWIZZLE_TABLE, |
|
|
|
sizeof(SWIZZLE_TABLE)); |
|
|
|
|
|
|
|
scheduler.Record([src = staging_ref.buffer, offset = staging_ref.offset, dst = *data_buffer, |
|
|
|
TOTAL_BUFFER_SIZE](vk::CommandBuffer cmdbuf) { |
|
|
|
cmdbuf.CopyBuffer(src, dst, |
|
|
|
VkBufferCopy{ |
|
|
|
.srcOffset = offset, |
|
|
|
.dstOffset = 0, |
|
|
|
.size = TOTAL_BUFFER_SIZE, |
|
|
|
}); |
|
|
|
cmdbuf.PipelineBarrier( |
|
|
|
VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, 0, |
|
|
|
VkMemoryBarrier{ |
|
|
|
.sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER, |
|
|
|
.pNext = nullptr, |
|
|
|
.srcAccessMask = 0, |
|
|
|
.dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT, |
|
|
|
}); |
|
|
|
}); |
|
|
|
} |
|
|
|
|
|
|
|
void ASTCDecoderPass::Assemble(Image& image, const StagingBufferRef& map, |
|
|
|
std::span<const VideoCommon::SwizzleParameters> swizzles) { |
|
|
|
using namespace VideoCommon::Accelerated; |
|
|
|
const std::array<u32, 2> block_dims{ |
|
|
|
VideoCore::Surface::DefaultBlockWidth(image.info.format), |
|
|
|
VideoCore::Surface::DefaultBlockHeight(image.info.format), |
|
|
|
}; |
|
|
|
scheduler.RequestOutsideRenderPassOperationContext(); |
|
|
|
if (!data_buffer) { |
|
|
|
MakeDataBuffer(); |
|
|
|
} |
|
|
|
const VkPipeline vk_pipeline = *pipeline; |
|
|
|
const VkImageAspectFlags aspect_mask = image.AspectMask(); |
|
|
|
const VkImage vk_image = image.Handle(); |
|
|
|
const bool is_initialized = image.ExchangeInitialization(); |
|
|
|
scheduler.Record( |
|
|
|
[vk_pipeline, vk_image, aspect_mask, is_initialized](vk::CommandBuffer cmdbuf) { |
|
|
|
const VkImageMemoryBarrier image_barrier{ |
|
|
|
.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, |
|
|
|
.pNext = nullptr, |
|
|
|
.srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT, |
|
|
|
.dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT, |
|
|
|
.oldLayout = is_initialized ? VK_IMAGE_LAYOUT_GENERAL : VK_IMAGE_LAYOUT_UNDEFINED, |
|
|
|
.newLayout = VK_IMAGE_LAYOUT_GENERAL, |
|
|
|
.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, |
|
|
|
.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, |
|
|
|
.image = vk_image, |
|
|
|
.subresourceRange{ |
|
|
|
.aspectMask = aspect_mask, |
|
|
|
.baseMipLevel = 0, |
|
|
|
.levelCount = VK_REMAINING_MIP_LEVELS, |
|
|
|
.baseArrayLayer = 0, |
|
|
|
.layerCount = VK_REMAINING_ARRAY_LAYERS, |
|
|
|
}, |
|
|
|
}; |
|
|
|
cmdbuf.PipelineBarrier(is_initialized ? VK_PIPELINE_STAGE_ALL_COMMANDS_BIT : 0, |
|
|
|
VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, 0, image_barrier); |
|
|
|
cmdbuf.BindPipeline(VK_PIPELINE_BIND_POINT_COMPUTE, vk_pipeline); |
|
|
|
}); |
|
|
|
for (const VideoCommon::SwizzleParameters& swizzle : swizzles) { |
|
|
|
const size_t input_offset = swizzle.buffer_offset + map.offset; |
|
|
|
const u32 num_dispatches_x = Common::DivCeil(swizzle.num_tiles.width, 32U); |
|
|
|
const u32 num_dispatches_y = Common::DivCeil(swizzle.num_tiles.height, 32U); |
|
|
|
const u32 num_dispatches_z = image.info.resources.layers; |
|
|
|
|
|
|
|
update_descriptor_queue.Acquire(); |
|
|
|
update_descriptor_queue.AddBuffer(map.buffer, input_offset, |
|
|
|
image.guest_size_bytes - swizzle.buffer_offset); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, encoding_values), |
|
|
|
sizeof(AstcBufferData::encoding_values)); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_6_to_8), |
|
|
|
sizeof(AstcBufferData::replicate_6_to_8)); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_7_to_8), |
|
|
|
sizeof(AstcBufferData::replicate_7_to_8)); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, offsetof(AstcBufferData, replicate_8_to_8), |
|
|
|
sizeof(AstcBufferData::replicate_8_to_8)); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, |
|
|
|
offsetof(AstcBufferData, replicate_byte_to_16), |
|
|
|
sizeof(AstcBufferData::replicate_byte_to_16)); |
|
|
|
update_descriptor_queue.AddBuffer(*data_buffer, sizeof(AstcBufferData), |
|
|
|
sizeof(SWIZZLE_TABLE)); |
|
|
|
update_descriptor_queue.AddImage(image.StorageImageView(swizzle.level)); |
|
|
|
|
|
|
|
const VkDescriptorSet set = CommitDescriptorSet(update_descriptor_queue); |
|
|
|
const VkPipelineLayout vk_layout = *layout; |
|
|
|
|
|
|
|
// To unswizzle the ASTC data
|
|
|
|
const auto params = MakeBlockLinearSwizzle2DParams(swizzle, image.info); |
|
|
|
ASSERT(params.origin == (std::array<u32, 3>{0, 0, 0})); |
|
|
|
ASSERT(params.destination == (std::array<s32, 3>{0, 0, 0})); |
|
|
|
scheduler.Record([vk_layout, num_dispatches_x, num_dispatches_y, num_dispatches_z, |
|
|
|
block_dims, params, set](vk::CommandBuffer cmdbuf) { |
|
|
|
const AstcPushConstants uniforms{ |
|
|
|
.blocks_dims = block_dims, |
|
|
|
.bytes_per_block_log2 = params.bytes_per_block_log2, |
|
|
|
.layer_stride = params.layer_stride, |
|
|
|
.block_size = params.block_size, |
|
|
|
.x_shift = params.x_shift, |
|
|
|
.block_height = params.block_height, |
|
|
|
.block_height_mask = params.block_height_mask, |
|
|
|
}; |
|
|
|
cmdbuf.BindDescriptorSets(VK_PIPELINE_BIND_POINT_COMPUTE, vk_layout, 0, set, {}); |
|
|
|
cmdbuf.PushConstants(vk_layout, VK_SHADER_STAGE_COMPUTE_BIT, uniforms); |
|
|
|
cmdbuf.Dispatch(num_dispatches_x, num_dispatches_y, num_dispatches_z); |
|
|
|
}); |
|
|
|
} |
|
|
|
scheduler.Record([vk_image, aspect_mask](vk::CommandBuffer cmdbuf) { |
|
|
|
const VkImageMemoryBarrier image_barrier{ |
|
|
|
.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, |
|
|
|
.pNext = nullptr, |
|
|
|
.srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT, |
|
|
|
.dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT, |
|
|
|
.oldLayout = VK_IMAGE_LAYOUT_GENERAL, |
|
|
|
.newLayout = VK_IMAGE_LAYOUT_GENERAL, |
|
|
|
.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, |
|
|
|
.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, |
|
|
|
.image = vk_image, |
|
|
|
.subresourceRange{ |
|
|
|
.aspectMask = aspect_mask, |
|
|
|
.baseMipLevel = 0, |
|
|
|
.levelCount = VK_REMAINING_MIP_LEVELS, |
|
|
|
.baseArrayLayer = 0, |
|
|
|
.layerCount = VK_REMAINING_ARRAY_LAYERS, |
|
|
|
}, |
|
|
|
}; |
|
|
|
cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, |
|
|
|
VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, 0, image_barrier); |
|
|
|
}); |
|
|
|
} |
|
|
|
|
|
|
|
} // namespace Vulkan
|