|
|
|
@ -6,11 +6,12 @@ |
|
|
|
|
|
|
|
namespace Shader::Backend::SPIRV { |
|
|
|
namespace { |
|
|
|
|
|
|
|
Id GetSharedPointer(EmitContext& ctx, Id offset, u32 index_offset = 0) { |
|
|
|
Id SharedPointer(EmitContext& ctx, Id offset, u32 index_offset = 0) { |
|
|
|
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)}; |
|
|
|
const Id shifted_value{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)}; |
|
|
|
const Id index{ctx.OpIAdd(ctx.U32[1], shifted_value, ctx.Constant(ctx.U32[1], index_offset))}; |
|
|
|
Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)}; |
|
|
|
if (index_offset > 0) { |
|
|
|
index = ctx.OpIAdd(ctx.U32[1], index, ctx.Constant(ctx.U32[1], index_offset)); |
|
|
|
} |
|
|
|
return ctx.profile.support_explicit_workgroup_layout |
|
|
|
? ctx.OpAccessChain(ctx.shared_u32, ctx.shared_memory_u32, ctx.u32_zero_value, index) |
|
|
|
: ctx.OpAccessChain(ctx.shared_u32, ctx.shared_memory_u32, index); |
|
|
|
@ -30,340 +31,258 @@ Id StorageIndex(EmitContext& ctx, const IR::Value& offset, size_t element_size) |
|
|
|
return ctx.OpShiftRightLogical(ctx.U32[1], index, shift_id); |
|
|
|
} |
|
|
|
|
|
|
|
Id GetStoragePointer(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
u32 index_offset = 0) { |
|
|
|
// TODO: Support reinterpreting bindings, guaranteed to be aligned
|
|
|
|
Id StoragePointer(EmitContext& ctx, const StorageTypeDefinition& type_def, |
|
|
|
Id StorageDefinitions::*member_ptr, const IR::Value& binding, |
|
|
|
const IR::Value& offset, size_t element_size) { |
|
|
|
if (!binding.IsImmediate()) { |
|
|
|
throw NotImplementedException("Dynamic storage buffer indexing"); |
|
|
|
} |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id index{ctx.OpIAdd(ctx.U32[1], base_index, ctx.Constant(ctx.U32[1], index_offset))}; |
|
|
|
return ctx.OpAccessChain(ctx.storage_u32, ssbo, ctx.u32_zero_value, index); |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].*member_ptr}; |
|
|
|
const Id index{StorageIndex(ctx, offset, element_size)}; |
|
|
|
return ctx.OpAccessChain(type_def.element, ssbo, ctx.u32_zero_value, index); |
|
|
|
} |
|
|
|
|
|
|
|
std::pair<Id, Id> GetAtomicArgs(EmitContext& ctx) { |
|
|
|
std::pair<Id, Id> AtomicArgs(EmitContext& ctx) { |
|
|
|
const Id scope{ctx.Constant(ctx.U32[1], static_cast<u32>(spv::Scope::Device))}; |
|
|
|
const Id semantics{ctx.u32_zero_value}; |
|
|
|
return {scope, semantics}; |
|
|
|
} |
|
|
|
|
|
|
|
Id LoadU64(EmitContext& ctx, Id pointer_1, Id pointer_2) { |
|
|
|
const Id value_1{ctx.OpLoad(ctx.U32[1], pointer_1)}; |
|
|
|
const Id value_2{ctx.OpLoad(ctx.U32[1], pointer_2)}; |
|
|
|
const Id original_composite{ctx.OpCompositeConstruct(ctx.U32[2], value_1, value_2)}; |
|
|
|
return ctx.OpBitcast(ctx.U64, original_composite); |
|
|
|
Id SharedAtomicU32(EmitContext& ctx, Id offset, Id value, |
|
|
|
Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id)) { |
|
|
|
const Id pointer{SharedPointer(ctx, offset)}; |
|
|
|
const auto [scope, semantics]{AtomicArgs(ctx)}; |
|
|
|
return (ctx.*atomic_func)(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
} |
|
|
|
|
|
|
|
void StoreResult(EmitContext& ctx, Id pointer_1, Id pointer_2, Id result) { |
|
|
|
const Id composite{ctx.OpBitcast(ctx.U32[2], result)}; |
|
|
|
ctx.OpStore(pointer_1, ctx.OpCompositeExtract(ctx.U32[1], composite, 0)); |
|
|
|
ctx.OpStore(pointer_2, ctx.OpCompositeExtract(ctx.U32[1], composite, 1)); |
|
|
|
Id StorageAtomicU32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, Id value, |
|
|
|
Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id)) { |
|
|
|
const Id pointer{StoragePointer(ctx, ctx.storage_types.U32, &StorageDefinitions::U32, binding, |
|
|
|
offset, sizeof(u32))}; |
|
|
|
const auto [scope, semantics]{AtomicArgs(ctx)}; |
|
|
|
return (ctx.*atomic_func)(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
} |
|
|
|
|
|
|
|
Id StorageAtomicU64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, Id value, |
|
|
|
Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id), |
|
|
|
Id (Sirit::Module::*non_atomic_func)(Id, Id, Id)) { |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const Id pointer{StoragePointer(ctx, ctx.storage_types.U64, &StorageDefinitions::U64, |
|
|
|
binding, offset, sizeof(u64))}; |
|
|
|
const auto [scope, semantics]{AtomicArgs(ctx)}; |
|
|
|
return (ctx.*atomic_func)(ctx.U64, pointer, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(..., "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer{StoragePointer(ctx, ctx.storage_types.U32x2, &StorageDefinitions::U32x2, |
|
|
|
binding, offset, sizeof(u32[2]))}; |
|
|
|
const Id original_value{ctx.OpBitcast(ctx.U64, ctx.OpLoad(ctx.U32[2], pointer))}; |
|
|
|
const Id result{(ctx.*non_atomic_func)(ctx.U64, value, original_value)}; |
|
|
|
ctx.OpStore(pointer, result); |
|
|
|
return original_value; |
|
|
|
} |
|
|
|
} // Anonymous namespace
|
|
|
|
|
|
|
|
Id EmitSharedAtomicIAdd32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicIAdd(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicIAdd32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicIAdd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicSMin32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMin(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicSMin32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicSMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicUMin32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMin(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicUMin32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicUMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicSMax32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMax(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicSMax32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicSMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicUMax32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMax(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicUMax32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicUMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicInc32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
Id EmitSharedAtomicInc32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)}; |
|
|
|
const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], pointer_offset, shift_id)}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_shared, index, value, |
|
|
|
ctx.shared_memory_u32); |
|
|
|
const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_shared, index, value); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicDec32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
Id EmitSharedAtomicDec32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)}; |
|
|
|
const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], pointer_offset, shift_id)}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_shared, index, value, |
|
|
|
ctx.shared_memory_u32); |
|
|
|
const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_shared, index, value); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicAnd32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicAnd(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicAnd32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicAnd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicOr32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicOr(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicOr32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicOr); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicXor32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicXor(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicXor32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicXor); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicExchange32(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicExchange32(EmitContext& ctx, Id offset, Id value) { |
|
|
|
return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicExchange); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitSharedAtomicExchange64(EmitContext& ctx, Id pointer_offset, Id value) { |
|
|
|
const Id pointer_1{GetSharedPointer(ctx, pointer_offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
Id EmitSharedAtomicExchange64(EmitContext& ctx, Id offset, Id value) { |
|
|
|
if (ctx.profile.support_int64_atomics && ctx.profile.support_explicit_workgroup_layout) { |
|
|
|
const Id shift_id{ctx.Constant(ctx.U32[1], 3U)}; |
|
|
|
const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)}; |
|
|
|
const Id pointer{ |
|
|
|
ctx.OpAccessChain(ctx.shared_u64, ctx.shared_memory_u64, ctx.u32_zero_value, index)}; |
|
|
|
const auto [scope, semantics]{AtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U64, pointer, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetSharedPointer(ctx, pointer_offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, value); |
|
|
|
return original_value; |
|
|
|
// LOG_WARNING("Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_1{SharedPointer(ctx, offset, 0)}; |
|
|
|
const Id pointer_2{SharedPointer(ctx, offset, 1)}; |
|
|
|
const Id value_1{ctx.OpLoad(ctx.U32[1], pointer_1)}; |
|
|
|
const Id value_2{ctx.OpLoad(ctx.U32[1], pointer_2)}; |
|
|
|
const Id new_vector{ctx.OpBitcast(ctx.U32[2], value)}; |
|
|
|
ctx.OpStore(pointer_1, ctx.OpCompositeExtract(ctx.U32[1], new_vector, 0U)); |
|
|
|
ctx.OpStore(pointer_2, ctx.OpCompositeExtract(ctx.U32[1], new_vector, 1U)); |
|
|
|
return ctx.OpBitcast(ctx.U64, ctx.OpCompositeConstruct(ctx.U32[2], value_1, value_2)); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicIAdd32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicIAdd(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicIAdd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicSMin32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMin(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicUMin32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMin(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicSMax32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMax(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicUMax32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMax(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicInc32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_ssbo, base_index, value, ssbo); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicDec32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_ssbo, base_index, value, ssbo); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicAnd32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicAnd(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicAnd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicOr32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicOr(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicOr); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicXor32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicXor(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicXor); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicExchange32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U32[1], pointer, scope, semantics, value); |
|
|
|
return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicExchange); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicIAdd64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicIAdd(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpIAdd(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicIAdd, |
|
|
|
&Sirit::Module::OpIAdd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicSMin64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMin(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpSMin(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMin, |
|
|
|
&Sirit::Module::OpSMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicUMin64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMin(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpUMin(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMin, |
|
|
|
&Sirit::Module::OpUMin); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicSMax64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicSMax(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpSMax(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMax, |
|
|
|
&Sirit::Module::OpSMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicUMax64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicUMax(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpUMax(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMax, |
|
|
|
&Sirit::Module::OpUMax); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicAnd64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicAnd(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpBitwiseAnd(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicAnd, |
|
|
|
&Sirit::Module::OpBitwiseAnd); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicOr64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicOr(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpBitwiseOr(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicOr, |
|
|
|
&Sirit::Module::OpBitwiseOr); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicXor64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicXor(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
const Id result{ctx.OpBitwiseXor(ctx.U64, value, original_value)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, result); |
|
|
|
return original_value; |
|
|
|
return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicXor, |
|
|
|
&Sirit::Module::OpBitwiseXor); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicExchange64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id pointer_1{GetStoragePointer(ctx, binding, offset)}; |
|
|
|
if (ctx.profile.support_int64_atomics) { |
|
|
|
const auto [scope, semantics]{GetAtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U64, pointer_1, scope, semantics, value); |
|
|
|
const Id pointer{StoragePointer(ctx, ctx.storage_types.U64, &StorageDefinitions::U64, |
|
|
|
binding, offset, sizeof(u64))}; |
|
|
|
const auto [scope, semantics]{AtomicArgs(ctx)}; |
|
|
|
return ctx.OpAtomicExchange(ctx.U64, pointer, scope, semantics, value); |
|
|
|
} |
|
|
|
// LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)}; |
|
|
|
const Id original_value{LoadU64(ctx, pointer_1, pointer_2)}; |
|
|
|
StoreResult(ctx, pointer_1, pointer_2, value); |
|
|
|
return original_value; |
|
|
|
// LOG_WARNING(..., "Int64 Atomics not supported, fallback to non-atomic");
|
|
|
|
const Id pointer{StoragePointer(ctx, ctx.storage_types.U32x2, &StorageDefinitions::U32x2, |
|
|
|
binding, offset, sizeof(u32[2]))}; |
|
|
|
const Id original{ctx.OpBitcast(ctx.U64, ctx.OpLoad(ctx.U32[2], pointer))}; |
|
|
|
ctx.OpStore(pointer, value); |
|
|
|
return original; |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicAddF32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
return ctx.OpFunctionCall(ctx.F32[1], ctx.f32_add_cas, base_index, value, ssbo); |
|
|
|
} |
|
|
|
|
|
|
|
Id EmitStorageAtomicAddF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_add_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpBitcast(ctx.U32[1], result); |
|
|
|
@ -371,7 +290,7 @@ Id EmitStorageAtomicAddF16x2(EmitContext& ctx, const IR::Value& binding, const I |
|
|
|
|
|
|
|
Id EmitStorageAtomicAddF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_add_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpPackHalf2x16(ctx.U32[1], result); |
|
|
|
@ -379,7 +298,7 @@ Id EmitStorageAtomicAddF32x2(EmitContext& ctx, const IR::Value& binding, const I |
|
|
|
|
|
|
|
Id EmitStorageAtomicMinF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_min_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpBitcast(ctx.U32[1], result); |
|
|
|
@ -387,7 +306,7 @@ Id EmitStorageAtomicMinF16x2(EmitContext& ctx, const IR::Value& binding, const I |
|
|
|
|
|
|
|
Id EmitStorageAtomicMinF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_min_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpPackHalf2x16(ctx.U32[1], result); |
|
|
|
@ -395,7 +314,7 @@ Id EmitStorageAtomicMinF32x2(EmitContext& ctx, const IR::Value& binding, const I |
|
|
|
|
|
|
|
Id EmitStorageAtomicMaxF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_max_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpBitcast(ctx.U32[1], result); |
|
|
|
@ -403,7 +322,7 @@ Id EmitStorageAtomicMaxF16x2(EmitContext& ctx, const IR::Value& binding, const I |
|
|
|
|
|
|
|
Id EmitStorageAtomicMaxF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, |
|
|
|
Id value) { |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()]}; |
|
|
|
const Id ssbo{ctx.ssbos[binding.U32()].U32}; |
|
|
|
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))}; |
|
|
|
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_max_cas, base_index, value, ssbo)}; |
|
|
|
return ctx.OpPackHalf2x16(ctx.U32[1], result); |
|
|
|
|