|
@@ -40,7 +40,6 @@
|
|
#include "drivers/vulkan/vulkan_context.h"
|
|
#include "drivers/vulkan/vulkan_context.h"
|
|
|
|
|
|
#include "thirdparty/misc/smolv.h"
|
|
#include "thirdparty/misc/smolv.h"
|
|
-#include "thirdparty/spirv-reflect/spirv_reflect.h"
|
|
|
|
|
|
|
|
//#define FORCE_FULL_BARRIER
|
|
//#define FORCE_FULL_BARRIER
|
|
|
|
|
|
@@ -4524,14 +4523,6 @@ RID RenderingDeviceVulkan::index_array_create(RID p_index_buffer, uint32_t p_ind
|
|
/**** SHADER ****/
|
|
/**** SHADER ****/
|
|
/****************/
|
|
/****************/
|
|
|
|
|
|
-static const char *shader_stage_names[RenderingDevice::SHADER_STAGE_MAX] = {
|
|
|
|
- "Vertex",
|
|
|
|
- "Fragment",
|
|
|
|
- "TesselationControl",
|
|
|
|
- "TesselationEvaluation",
|
|
|
|
- "Compute"
|
|
|
|
-};
|
|
|
|
-
|
|
|
|
static const char *shader_uniform_names[RenderingDevice::UNIFORM_TYPE_MAX] = {
|
|
static const char *shader_uniform_names[RenderingDevice::UNIFORM_TYPE_MAX] = {
|
|
"Sampler", "CombinedSampler", "Texture", "Image", "TextureBuffer", "SamplerTextureBuffer", "ImageBuffer", "UniformBuffer", "StorageBuffer", "InputAttachment"
|
|
"Sampler", "CombinedSampler", "Texture", "Image", "TextureBuffer", "SamplerTextureBuffer", "ImageBuffer", "UniformBuffer", "StorageBuffer", "InputAttachment"
|
|
};
|
|
};
|
|
@@ -4562,198 +4553,6 @@ String RenderingDeviceVulkan::_shader_uniform_debug(RID p_shader, int p_set) {
|
|
}
|
|
}
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
-#if 0
|
|
|
|
-bool RenderingDeviceVulkan::_uniform_add_binding(Vector<Vector<VkDescriptorSetLayoutBinding> > &bindings, Vector<Vector<UniformInfo> > &uniform_infos, const glslang::TObjectReflection &reflection, RenderingDevice::ShaderStage p_stage, Shader::PushConstant &push_constant, String *r_error) {
|
|
|
|
- VkDescriptorSetLayoutBinding layout_binding;
|
|
|
|
- UniformInfo info;
|
|
|
|
-
|
|
|
|
- switch (reflection.getType()->getBasicType()) {
|
|
|
|
- case glslang::EbtSampler: {
|
|
|
|
- //print_line("DEBUG: IsSampler");
|
|
|
|
- if (reflection.getType()->getSampler().dim == glslang::EsdBuffer) {
|
|
|
|
- // Texture buffers.
|
|
|
|
- if (reflection.getType()->getSampler().isCombined()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER;
|
|
|
|
- info.type = UNIFORM_TYPE_SAMPLER_WITH_TEXTURE_BUFFER;
|
|
|
|
- //print_line("DEBUG: SAMPLER: texel combined");
|
|
|
|
- } else if (reflection.getType()->getSampler().isTexture()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER;
|
|
|
|
- info.type = UNIFORM_TYPE_TEXTURE_BUFFER;
|
|
|
|
- //print_line("DEBUG: SAMPLER: texel alone");
|
|
|
|
- } else if (reflection.getType()->getSampler().isImage()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER;
|
|
|
|
- info.type = UNIFORM_TYPE_IMAGE_BUFFER;
|
|
|
|
- //print_line("DEBUG: SAMPLER: texel buffer");
|
|
|
|
- } else {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' is of unsupported buffer type.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
- } else if (reflection.getType()->getSampler().isCombined()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
|
|
|
|
- info.type = UNIFORM_TYPE_SAMPLER_WITH_TEXTURE;
|
|
|
|
- //print_line("DEBUG: SAMPLER: combined");
|
|
|
|
- } else if (reflection.getType()->getSampler().isPureSampler()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
|
|
|
|
- info.type = UNIFORM_TYPE_SAMPLER;
|
|
|
|
- //print_line("DEBUG: SAMPLER: sampler");
|
|
|
|
- } else if (reflection.getType()->getSampler().isTexture()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
|
|
|
|
- info.type = UNIFORM_TYPE_TEXTURE;
|
|
|
|
- //print_line("DEBUG: SAMPLER: image");
|
|
|
|
- } else if (reflection.getType()->getSampler().isImage()) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
|
|
|
|
- info.type = UNIFORM_TYPE_IMAGE;
|
|
|
|
- //print_line("DEBUG: SAMPLER: storage image");
|
|
|
|
- } else {
|
|
|
|
- //print_line("DEBUG: sampler unknown");
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' is of unsupported sampler type.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (reflection.getType()->isArray()) {
|
|
|
|
- layout_binding.descriptorCount = reflection.getType()->getArraySizes()->getCumulativeSize();
|
|
|
|
- //print_line("DEBUG: array of size: " + itos(layout_binding.descriptorCount));
|
|
|
|
- } else {
|
|
|
|
- layout_binding.descriptorCount = 1;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- info.length = layout_binding.descriptorCount;
|
|
|
|
-
|
|
|
|
- } break;
|
|
|
|
- /*case glslang::EbtStruct: {
|
|
|
|
- print_line("DEBUG: Struct");
|
|
|
|
-
|
|
|
|
- } break;*/
|
|
|
|
- case glslang::EbtBlock: {
|
|
|
|
- //print_line("DEBUG: Block");
|
|
|
|
- if (reflection.getType()->getQualifier().storage == glslang::EvqUniform) {
|
|
|
|
- if (reflection.getType()->getQualifier().layoutPushConstant) {
|
|
|
|
- uint32_t len = reflection.size;
|
|
|
|
- if (push_constant.push_constant_size != 0 && push_constant.push_constant_size != len) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' push constants for different stages should all be the same size.";
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
- push_constant.push_constant_size = len;
|
|
|
|
- push_constant.push_constants_vk_stage |= shader_stage_masks[p_stage];
|
|
|
|
- return true;
|
|
|
|
- }
|
|
|
|
- //print_line("DEBUG: Uniform buffer");
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER;
|
|
|
|
- info.type = UNIFORM_TYPE_UNIFORM_BUFFER;
|
|
|
|
- } else if (reflection.getType()->getQualifier().storage == glslang::EvqBuffer) {
|
|
|
|
- layout_binding.descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
|
|
|
|
- info.type = UNIFORM_TYPE_STORAGE_BUFFER;
|
|
|
|
- //print_line("DEBUG: Storage buffer");
|
|
|
|
- } else {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' is of unsupported block type: (" + itos(reflection.getType()->getQualifier().storage) + ").";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (reflection.getType()->isArray()) {
|
|
|
|
- layout_binding.descriptorCount = reflection.getType()->getArraySizes()->getCumulativeSize();
|
|
|
|
- //print_line("DEBUG: array of size: " + itos(layout_binding.descriptorCount));
|
|
|
|
- } else {
|
|
|
|
- layout_binding.descriptorCount = 1;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- info.length = reflection.size;
|
|
|
|
-
|
|
|
|
- } break;
|
|
|
|
- /*case glslang::EbtReference: {
|
|
|
|
- } break;*/
|
|
|
|
- /*case glslang::EbtAtomicUint: {
|
|
|
|
- } break;*/
|
|
|
|
- default: {
|
|
|
|
- if (reflection.getType()->getQualifier().hasOffset() || reflection.name.find(".") != std::string::npos) {
|
|
|
|
- // Member of uniform block?
|
|
|
|
- return true;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' unsupported uniform type.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (!reflection.getType()->getQualifier().hasBinding()) {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' lacks a binding number.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- uint32_t set = reflection.getType()->getQualifier().hasSet() ? reflection.getType()->getQualifier().layoutSet : 0;
|
|
|
|
-
|
|
|
|
- if (set >= MAX_UNIFORM_SETS) {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' uses a set (" + itos(set) + ") index larger than what is supported (" + itos(MAX_UNIFORM_SETS) + ").";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (set >= limits.maxBoundDescriptorSets) {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' uses a set (" + itos(set) + ") index larger than what is supported by the hardware (" + itos(limits.maxBoundDescriptorSets) + ").";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- uint32_t binding = reflection.getType()->getQualifier().layoutBinding;
|
|
|
|
-
|
|
|
|
- if (set < (uint32_t)bindings.size()) {
|
|
|
|
- // Check if this already exists.
|
|
|
|
- for (int i = 0; i < bindings[set].size(); i++) {
|
|
|
|
- if (bindings[set][i].binding == binding) {
|
|
|
|
- // Already exists, verify that it's the same type.
|
|
|
|
- if (bindings[set][i].descriptorType != layout_binding.descriptorType) {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' trying to re-use location for set=" + itos(set) + ", binding=" + itos(binding) + " with different uniform type.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- // Also, verify that it's the same size.
|
|
|
|
- if (bindings[set][i].descriptorCount != layout_binding.descriptorCount || uniform_infos[set][i].length != info.length) {
|
|
|
|
- if (r_error) {
|
|
|
|
- *r_error = "On shader stage '" + String(shader_stage_names[p_stage]) + "', uniform '" + reflection.name + "' trying to re-use location for set=" + itos(set) + ", binding=" + itos(binding) + " with different uniform size.";
|
|
|
|
- }
|
|
|
|
- return false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- // Just append stage mask and return.
|
|
|
|
- bindings.write[set].write[i].stageFlags |= shader_stage_masks[p_stage];
|
|
|
|
- uniform_infos.write[set].write[i].stages |= 1 << p_stage;
|
|
|
|
- return true;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- layout_binding.binding = binding;
|
|
|
|
- layout_binding.stageFlags = shader_stage_masks[p_stage];
|
|
|
|
- layout_binding.pImmutableSamplers = nullptr; // No support for this yet.
|
|
|
|
-
|
|
|
|
- info.stages = 1 << p_stage;
|
|
|
|
- info.binding = binding;
|
|
|
|
-
|
|
|
|
- if (set >= (uint32_t)bindings.size()) {
|
|
|
|
- bindings.resize(set + 1);
|
|
|
|
- uniform_infos.resize(set + 1);
|
|
|
|
- }
|
|
|
|
-#if 0
|
|
|
|
- print_line("stage: " + String(shader_stage_names[p_stage]) + " set: " + itos(set) + " binding: " + itos(info.binding) + " type:" + shader_uniform_names[info.type] + " length: " + itos(info.length));
|
|
|
|
-#endif
|
|
|
|
- bindings.write[set].push_back(layout_binding);
|
|
|
|
- uniform_infos.write[set].push_back(info);
|
|
|
|
-
|
|
|
|
- return true;
|
|
|
|
-}
|
|
|
|
-#endif
|
|
|
|
|
|
|
|
// Version 1: initial.
|
|
// Version 1: initial.
|
|
// Version 2: Added shader name.
|
|
// Version 2: Added shader name.
|
|
@@ -4786,346 +4585,68 @@ struct RenderingDeviceVulkanShaderBinarySpecializationConstant {
|
|
|
|
|
|
struct RenderingDeviceVulkanShaderBinaryData {
|
|
struct RenderingDeviceVulkanShaderBinaryData {
|
|
uint32_t vertex_input_mask;
|
|
uint32_t vertex_input_mask;
|
|
- uint32_t fragment_outputs;
|
|
|
|
- uint32_t specialization_constant_count;
|
|
|
|
|
|
+ uint32_t fragment_output_mask;
|
|
|
|
+ uint32_t specialization_constants_count;
|
|
uint32_t is_compute;
|
|
uint32_t is_compute;
|
|
uint32_t compute_local_size[3];
|
|
uint32_t compute_local_size[3];
|
|
uint32_t set_count;
|
|
uint32_t set_count;
|
|
uint32_t push_constant_size;
|
|
uint32_t push_constant_size;
|
|
- uint32_t push_constants_vk_stage;
|
|
|
|
|
|
+ uint32_t push_constant_vk_stages_mask;
|
|
uint32_t stage_count;
|
|
uint32_t stage_count;
|
|
uint32_t shader_name_len;
|
|
uint32_t shader_name_len;
|
|
};
|
|
};
|
|
|
|
|
|
Vector<uint8_t> RenderingDeviceVulkan::shader_compile_binary_from_spirv(const Vector<ShaderStageSPIRVData> &p_spirv, const String &p_shader_name) {
|
|
Vector<uint8_t> RenderingDeviceVulkan::shader_compile_binary_from_spirv(const Vector<ShaderStageSPIRVData> &p_spirv, const String &p_shader_name) {
|
|
- RenderingDeviceVulkanShaderBinaryData binary_data;
|
|
|
|
- binary_data.vertex_input_mask = 0;
|
|
|
|
- binary_data.fragment_outputs = 0;
|
|
|
|
- binary_data.specialization_constant_count = 0;
|
|
|
|
- binary_data.is_compute = 0;
|
|
|
|
- binary_data.compute_local_size[0] = 0;
|
|
|
|
- binary_data.compute_local_size[1] = 0;
|
|
|
|
- binary_data.compute_local_size[2] = 0;
|
|
|
|
- binary_data.set_count = 0;
|
|
|
|
- binary_data.push_constant_size = 0;
|
|
|
|
- binary_data.push_constants_vk_stage = 0;
|
|
|
|
|
|
+ SpirvReflectionData spirv_data;
|
|
|
|
+ if (_reflect_spirv(p_spirv, spirv_data) != OK) {
|
|
|
|
+ return Vector<uint8_t>();
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ ERR_FAIL_COND_V_MSG((uint32_t)spirv_data.uniforms.size() > limits.maxBoundDescriptorSets, Vector<uint8_t>(),
|
|
|
|
+ "Number of uniform sets is larger than what is supported by the hardware (" + itos(limits.maxBoundDescriptorSets) + ").");
|
|
|
|
|
|
|
|
+ // Collect reflection data into binary data.
|
|
|
|
+ RenderingDeviceVulkanShaderBinaryData binary_data;
|
|
Vector<Vector<RenderingDeviceVulkanShaderBinaryDataBinding>> uniform_info; // Set bindings.
|
|
Vector<Vector<RenderingDeviceVulkanShaderBinaryDataBinding>> uniform_info; // Set bindings.
|
|
Vector<RenderingDeviceVulkanShaderBinarySpecializationConstant> specialization_constants;
|
|
Vector<RenderingDeviceVulkanShaderBinarySpecializationConstant> specialization_constants;
|
|
-
|
|
|
|
- uint32_t stages_processed = 0;
|
|
|
|
-
|
|
|
|
- for (int i = 0; i < p_spirv.size(); i++) {
|
|
|
|
- if (p_spirv[i].shader_stage == SHADER_STAGE_COMPUTE) {
|
|
|
|
- binary_data.is_compute = true;
|
|
|
|
- ERR_FAIL_COND_V_MSG(p_spirv.size() != 1, Vector<uint8_t>(),
|
|
|
|
- "Compute shaders can only receive one stage, dedicated to compute.");
|
|
|
|
- }
|
|
|
|
- ERR_FAIL_COND_V_MSG(stages_processed & (1 << p_spirv[i].shader_stage), Vector<uint8_t>(),
|
|
|
|
- "Stage " + String(shader_stage_names[p_spirv[i].shader_stage]) + " submitted more than once.");
|
|
|
|
-
|
|
|
|
- {
|
|
|
|
- SpvReflectShaderModule module;
|
|
|
|
- const uint8_t *spirv = p_spirv[i].spir_v.ptr();
|
|
|
|
- SpvReflectResult result = spvReflectCreateShaderModule(p_spirv[i].spir_v.size(), spirv, &module);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed parsing shader.");
|
|
|
|
-
|
|
|
|
- if (binary_data.is_compute) {
|
|
|
|
- binary_data.compute_local_size[0] = module.entry_points->local_size.x;
|
|
|
|
- binary_data.compute_local_size[1] = module.entry_points->local_size.y;
|
|
|
|
- binary_data.compute_local_size[2] = module.entry_points->local_size.z;
|
|
|
|
- }
|
|
|
|
- uint32_t binding_count = 0;
|
|
|
|
- result = spvReflectEnumerateDescriptorBindings(&module, &binding_count, nullptr);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed enumerating descriptor bindings.");
|
|
|
|
-
|
|
|
|
- uint32_t stage = p_spirv[i].shader_stage;
|
|
|
|
-
|
|
|
|
- if (binding_count > 0) {
|
|
|
|
- // Parse bindings.
|
|
|
|
-
|
|
|
|
- Vector<SpvReflectDescriptorBinding *> bindings;
|
|
|
|
- bindings.resize(binding_count);
|
|
|
|
- result = spvReflectEnumerateDescriptorBindings(&module, &binding_count, bindings.ptrw());
|
|
|
|
-
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed getting descriptor bindings.");
|
|
|
|
-
|
|
|
|
- for (uint32_t j = 0; j < binding_count; j++) {
|
|
|
|
- const SpvReflectDescriptorBinding &binding = *bindings[j];
|
|
|
|
-
|
|
|
|
- RenderingDeviceVulkanShaderBinaryDataBinding info{};
|
|
|
|
-
|
|
|
|
- bool need_array_dimensions = false;
|
|
|
|
- bool need_block_size = false;
|
|
|
|
- bool may_be_writable = false;
|
|
|
|
-
|
|
|
|
- switch (binding.descriptor_type) {
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_SAMPLER: {
|
|
|
|
- info.type = UNIFORM_TYPE_SAMPLER;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER: {
|
|
|
|
- info.type = UNIFORM_TYPE_SAMPLER_WITH_TEXTURE;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_SAMPLED_IMAGE: {
|
|
|
|
- info.type = UNIFORM_TYPE_TEXTURE;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_STORAGE_IMAGE: {
|
|
|
|
- info.type = UNIFORM_TYPE_IMAGE;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- may_be_writable = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER: {
|
|
|
|
- info.type = UNIFORM_TYPE_TEXTURE_BUFFER;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER: {
|
|
|
|
- info.type = UNIFORM_TYPE_IMAGE_BUFFER;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- may_be_writable = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_UNIFORM_BUFFER: {
|
|
|
|
- info.type = UNIFORM_TYPE_UNIFORM_BUFFER;
|
|
|
|
- need_block_size = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_STORAGE_BUFFER: {
|
|
|
|
- info.type = UNIFORM_TYPE_STORAGE_BUFFER;
|
|
|
|
- need_block_size = true;
|
|
|
|
- may_be_writable = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC: {
|
|
|
|
- ERR_PRINT("Dynamic uniform buffer not supported.");
|
|
|
|
- continue;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC: {
|
|
|
|
- ERR_PRINT("Dynamic storage buffer not supported.");
|
|
|
|
- continue;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_INPUT_ATTACHMENT: {
|
|
|
|
- info.type = UNIFORM_TYPE_INPUT_ATTACHMENT;
|
|
|
|
- need_array_dimensions = true;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_DESCRIPTOR_TYPE_ACCELERATION_STRUCTURE_KHR: {
|
|
|
|
- ERR_PRINT("Acceleration structure not supported.");
|
|
|
|
- continue;
|
|
|
|
- } break;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (need_array_dimensions) {
|
|
|
|
- if (binding.array.dims_count == 0) {
|
|
|
|
- info.length = 1;
|
|
|
|
- } else {
|
|
|
|
- for (uint32_t k = 0; k < binding.array.dims_count; k++) {
|
|
|
|
- if (k == 0) {
|
|
|
|
- info.length = binding.array.dims[0];
|
|
|
|
- } else {
|
|
|
|
- info.length *= binding.array.dims[k];
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- } else if (need_block_size) {
|
|
|
|
- info.length = binding.block.size;
|
|
|
|
- } else {
|
|
|
|
- info.length = 0;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (may_be_writable) {
|
|
|
|
- info.writable = !(binding.type_description->decoration_flags & SPV_REFLECT_DECORATION_NON_WRITABLE) && !(binding.block.decoration_flags & SPV_REFLECT_DECORATION_NON_WRITABLE);
|
|
|
|
- } else {
|
|
|
|
- info.writable = false;
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- info.binding = binding.binding;
|
|
|
|
- uint32_t set = binding.set;
|
|
|
|
-
|
|
|
|
- ERR_FAIL_COND_V_MSG(set >= MAX_UNIFORM_SETS, Vector<uint8_t>(),
|
|
|
|
- "On shader stage '" + String(shader_stage_names[stage]) + "', uniform '" + binding.name + "' uses a set (" + itos(set) + ") index larger than what is supported (" + itos(MAX_UNIFORM_SETS) + ").");
|
|
|
|
-
|
|
|
|
- ERR_FAIL_COND_V_MSG(set >= limits.maxBoundDescriptorSets, Vector<uint8_t>(),
|
|
|
|
- "On shader stage '" + String(shader_stage_names[stage]) + "', uniform '" + binding.name + "' uses a set (" + itos(set) + ") index larger than what is supported by the hardware (" + itos(limits.maxBoundDescriptorSets) + ").");
|
|
|
|
-
|
|
|
|
- if (set < (uint32_t)uniform_info.size()) {
|
|
|
|
- // Check if this already exists.
|
|
|
|
- bool exists = false;
|
|
|
|
- for (int k = 0; k < uniform_info[set].size(); k++) {
|
|
|
|
- if (uniform_info[set][k].binding == (uint32_t)info.binding) {
|
|
|
|
- // Already exists, verify that it's the same type.
|
|
|
|
- ERR_FAIL_COND_V_MSG(uniform_info[set][k].type != info.type, Vector<uint8_t>(),
|
|
|
|
- "On shader stage '" + String(shader_stage_names[stage]) + "', uniform '" + binding.name + "' trying to re-use location for set=" + itos(set) + ", binding=" + itos(info.binding) + " with different uniform type.");
|
|
|
|
-
|
|
|
|
- // Also, verify that it's the same size.
|
|
|
|
- ERR_FAIL_COND_V_MSG(uniform_info[set][k].length != info.length, Vector<uint8_t>(),
|
|
|
|
- "On shader stage '" + String(shader_stage_names[stage]) + "', uniform '" + binding.name + "' trying to re-use location for set=" + itos(set) + ", binding=" + itos(info.binding) + " with different uniform size.");
|
|
|
|
-
|
|
|
|
- // Also, verify that it has the same writability.
|
|
|
|
- ERR_FAIL_COND_V_MSG(uniform_info[set][k].writable != info.writable, Vector<uint8_t>(),
|
|
|
|
- "On shader stage '" + String(shader_stage_names[stage]) + "', uniform '" + binding.name + "' trying to re-use location for set=" + itos(set) + ", binding=" + itos(info.binding) + " with different writability.");
|
|
|
|
-
|
|
|
|
- // Just append stage mask and return.
|
|
|
|
- uniform_info.write[set].write[k].stages |= 1 << stage;
|
|
|
|
- exists = true;
|
|
|
|
- break;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (exists) {
|
|
|
|
- continue; // Merged.
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- info.stages = 1 << stage;
|
|
|
|
-
|
|
|
|
- if (set >= (uint32_t)uniform_info.size()) {
|
|
|
|
- uniform_info.resize(set + 1);
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- uniform_info.write[set].push_back(info);
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- {
|
|
|
|
- // Specialization constants.
|
|
|
|
-
|
|
|
|
- uint32_t sc_count = 0;
|
|
|
|
- result = spvReflectEnumerateSpecializationConstants(&module, &sc_count, nullptr);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed enumerating specialization constants.");
|
|
|
|
-
|
|
|
|
- if (sc_count) {
|
|
|
|
- Vector<SpvReflectSpecializationConstant *> spec_constants;
|
|
|
|
- spec_constants.resize(sc_count);
|
|
|
|
-
|
|
|
|
- result = spvReflectEnumerateSpecializationConstants(&module, &sc_count, spec_constants.ptrw());
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed obtaining specialization constants.");
|
|
|
|
-
|
|
|
|
- for (uint32_t j = 0; j < sc_count; j++) {
|
|
|
|
- int32_t existing = -1;
|
|
|
|
- RenderingDeviceVulkanShaderBinarySpecializationConstant sconst{};
|
|
|
|
- SpvReflectSpecializationConstant *spc = spec_constants[j];
|
|
|
|
-
|
|
|
|
- sconst.constant_id = spc->constant_id;
|
|
|
|
- sconst.int_value = 0.0; // Clear previous value JIC.
|
|
|
|
- switch (spc->constant_type) {
|
|
|
|
- case SPV_REFLECT_SPECIALIZATION_CONSTANT_BOOL: {
|
|
|
|
- sconst.type = PIPELINE_SPECIALIZATION_CONSTANT_TYPE_BOOL;
|
|
|
|
- sconst.bool_value = spc->default_value.int_bool_value != 0;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_SPECIALIZATION_CONSTANT_INT: {
|
|
|
|
- sconst.type = PIPELINE_SPECIALIZATION_CONSTANT_TYPE_INT;
|
|
|
|
- sconst.int_value = spc->default_value.int_bool_value;
|
|
|
|
- } break;
|
|
|
|
- case SPV_REFLECT_SPECIALIZATION_CONSTANT_FLOAT: {
|
|
|
|
- sconst.type = PIPELINE_SPECIALIZATION_CONSTANT_TYPE_FLOAT;
|
|
|
|
- sconst.float_value = spc->default_value.float_value;
|
|
|
|
- } break;
|
|
|
|
- }
|
|
|
|
- sconst.stage_flags = 1 << p_spirv[i].shader_stage;
|
|
|
|
-
|
|
|
|
- for (int k = 0; k < specialization_constants.size(); k++) {
|
|
|
|
- if (specialization_constants[k].constant_id == sconst.constant_id) {
|
|
|
|
- ERR_FAIL_COND_V_MSG(specialization_constants[k].type != sconst.type, Vector<uint8_t>(), "More than one specialization constant used for id (" + itos(sconst.constant_id) + "), but their types differ.");
|
|
|
|
- ERR_FAIL_COND_V_MSG(specialization_constants[k].int_value != sconst.int_value, Vector<uint8_t>(), "More than one specialization constant used for id (" + itos(sconst.constant_id) + "), but their default values differ.");
|
|
|
|
- existing = k;
|
|
|
|
- break;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (existing > 0) {
|
|
|
|
- specialization_constants.write[existing].stage_flags |= sconst.stage_flags;
|
|
|
|
- } else {
|
|
|
|
- specialization_constants.push_back(sconst);
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (stage == SHADER_STAGE_VERTEX) {
|
|
|
|
- uint32_t iv_count = 0;
|
|
|
|
- result = spvReflectEnumerateInputVariables(&module, &iv_count, nullptr);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed enumerating input variables.");
|
|
|
|
-
|
|
|
|
- if (iv_count) {
|
|
|
|
- Vector<SpvReflectInterfaceVariable *> input_vars;
|
|
|
|
- input_vars.resize(iv_count);
|
|
|
|
-
|
|
|
|
- result = spvReflectEnumerateInputVariables(&module, &iv_count, input_vars.ptrw());
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed obtaining input variables.");
|
|
|
|
-
|
|
|
|
- for (uint32_t j = 0; j < iv_count; j++) {
|
|
|
|
- if (input_vars[j] && input_vars[j]->decoration_flags == 0) { // Regular input.
|
|
|
|
- binary_data.vertex_input_mask |= (1 << uint32_t(input_vars[j]->location));
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
- if (stage == SHADER_STAGE_FRAGMENT) {
|
|
|
|
- uint32_t ov_count = 0;
|
|
|
|
- result = spvReflectEnumerateOutputVariables(&module, &ov_count, nullptr);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed enumerating output variables.");
|
|
|
|
-
|
|
|
|
- if (ov_count) {
|
|
|
|
- Vector<SpvReflectInterfaceVariable *> output_vars;
|
|
|
|
- output_vars.resize(ov_count);
|
|
|
|
-
|
|
|
|
- result = spvReflectEnumerateOutputVariables(&module, &ov_count, output_vars.ptrw());
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed obtaining output variables.");
|
|
|
|
-
|
|
|
|
- for (uint32_t j = 0; j < ov_count; j++) {
|
|
|
|
- const SpvReflectInterfaceVariable *refvar = output_vars[j];
|
|
|
|
- if (refvar != nullptr && refvar->built_in != SpvBuiltInFragDepth) {
|
|
|
|
- binary_data.fragment_outputs |= 1 << refvar->location;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
|
|
+ {
|
|
|
|
+ binary_data.vertex_input_mask = spirv_data.vertex_input_mask;
|
|
|
|
+ binary_data.fragment_output_mask = spirv_data.fragment_output_mask;
|
|
|
|
+ binary_data.specialization_constants_count = spirv_data.specialization_constants.size();
|
|
|
|
+ binary_data.is_compute = spirv_data.is_compute;
|
|
|
|
+ binary_data.compute_local_size[0] = spirv_data.compute_local_size[0];
|
|
|
|
+ binary_data.compute_local_size[1] = spirv_data.compute_local_size[1];
|
|
|
|
+ binary_data.compute_local_size[2] = spirv_data.compute_local_size[2];
|
|
|
|
+ binary_data.set_count = spirv_data.uniforms.size();
|
|
|
|
+ binary_data.push_constant_size = spirv_data.push_constant_size;
|
|
|
|
+ for (uint32_t i = 0; i < SHADER_STAGE_MAX; i++) {
|
|
|
|
+ if (spirv_data.push_constant_stages_mask.has_flag((ShaderStage)(1 << i))) {
|
|
|
|
+ binary_data.push_constant_vk_stages_mask |= shader_stage_masks[i];
|
|
}
|
|
}
|
|
|
|
+ }
|
|
|
|
|
|
- uint32_t pc_count = 0;
|
|
|
|
- result = spvReflectEnumeratePushConstantBlocks(&module, &pc_count, nullptr);
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed enumerating push constants.");
|
|
|
|
-
|
|
|
|
- if (pc_count) {
|
|
|
|
- ERR_FAIL_COND_V_MSG(pc_count > 1, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "': Only one push constant is supported, which should be the same across shader stages.");
|
|
|
|
-
|
|
|
|
- Vector<SpvReflectBlockVariable *> pconstants;
|
|
|
|
- pconstants.resize(pc_count);
|
|
|
|
- result = spvReflectEnumeratePushConstantBlocks(&module, &pc_count, pconstants.ptrw());
|
|
|
|
- ERR_FAIL_COND_V_MSG(result != SPV_REFLECT_RESULT_SUCCESS, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "' failed obtaining push constants.");
|
|
|
|
-#if 0
|
|
|
|
- if (pconstants[0] == nullptr) {
|
|
|
|
- Ref<FileAccess> f = FileAccess::open("res://popo.spv", FileAccess::WRITE);
|
|
|
|
- f->store_buffer((const uint8_t *)&SpirV[0], SpirV.size() * sizeof(uint32_t));
|
|
|
|
- }
|
|
|
|
-#endif
|
|
|
|
-
|
|
|
|
- ERR_FAIL_COND_V_MSG(binary_data.push_constant_size && binary_data.push_constant_size != pconstants[0]->size, Vector<uint8_t>(),
|
|
|
|
- "Reflection of SPIR-V shader stage '" + String(shader_stage_names[p_spirv[i].shader_stage]) + "': Push constant block must be the same across shader stages.");
|
|
|
|
-
|
|
|
|
- binary_data.push_constant_size = pconstants[0]->size;
|
|
|
|
- binary_data.push_constants_vk_stage |= shader_stage_masks[stage];
|
|
|
|
-
|
|
|
|
- //print_line("Stage: " + String(shader_stage_names[stage]) + " push constant of size=" + itos(push_constant.push_constant_size));
|
|
|
|
|
|
+ for (const Vector<SpirvReflectionData::Uniform> &spirv_set : spirv_data.uniforms) {
|
|
|
|
+ Vector<RenderingDeviceVulkanShaderBinaryDataBinding> set_bindings;
|
|
|
|
+ for (const SpirvReflectionData::Uniform &spirv_uniform : spirv_set) {
|
|
|
|
+ RenderingDeviceVulkanShaderBinaryDataBinding binding{};
|
|
|
|
+ binding.type = (uint32_t)spirv_uniform.type;
|
|
|
|
+ binding.binding = spirv_uniform.binding;
|
|
|
|
+ binding.stages = (uint32_t)spirv_uniform.stages_mask;
|
|
|
|
+ binding.length = spirv_uniform.length;
|
|
|
|
+ binding.writable = (uint32_t)spirv_uniform.writable;
|
|
|
|
+ set_bindings.push_back(binding);
|
|
}
|
|
}
|
|
-
|
|
|
|
- // Destroy the reflection data when no longer required.
|
|
|
|
- spvReflectDestroyShaderModule(&module);
|
|
|
|
|
|
+ uniform_info.push_back(set_bindings);
|
|
}
|
|
}
|
|
|
|
|
|
- stages_processed |= (1 << p_spirv[i].shader_stage);
|
|
|
|
|
|
+ for (const SpirvReflectionData::SpecializationConstant &spirv_sc : spirv_data.specialization_constants) {
|
|
|
|
+ RenderingDeviceVulkanShaderBinarySpecializationConstant spec_constant{};
|
|
|
|
+ spec_constant.type = (uint32_t)spirv_sc.type;
|
|
|
|
+ spec_constant.constant_id = spirv_sc.constant_id;
|
|
|
|
+ spec_constant.int_value = spirv_sc.int_value;
|
|
|
|
+ spec_constant.stage_flags = (uint32_t)spirv_sc.stages_mask;
|
|
|
|
+ specialization_constants.push_back(spec_constant);
|
|
|
|
+ }
|
|
}
|
|
}
|
|
|
|
|
|
Vector<Vector<uint8_t>> compressed_stages;
|
|
Vector<Vector<uint8_t>> compressed_stages;
|
|
@@ -5167,7 +4688,7 @@ Vector<uint8_t> RenderingDeviceVulkan::shader_compile_binary_from_spirv(const Ve
|
|
stages_binary_size += s;
|
|
stages_binary_size += s;
|
|
}
|
|
}
|
|
|
|
|
|
- binary_data.specialization_constant_count = specialization_constants.size();
|
|
|
|
|
|
+ binary_data.specialization_constants_count = specialization_constants.size();
|
|
binary_data.set_count = uniform_info.size();
|
|
binary_data.set_count = uniform_info.size();
|
|
binary_data.stage_count = p_spirv.size();
|
|
binary_data.stage_count = p_spirv.size();
|
|
|
|
|
|
@@ -5273,12 +4794,12 @@ RID RenderingDeviceVulkan::shader_create_from_bytecode(const Vector<uint8_t> &p_
|
|
const RenderingDeviceVulkanShaderBinaryData &binary_data = *(reinterpret_cast<const RenderingDeviceVulkanShaderBinaryData *>(binptr + 12));
|
|
const RenderingDeviceVulkanShaderBinaryData &binary_data = *(reinterpret_cast<const RenderingDeviceVulkanShaderBinaryData *>(binptr + 12));
|
|
|
|
|
|
Shader::PushConstant push_constant;
|
|
Shader::PushConstant push_constant;
|
|
- push_constant.push_constant_size = binary_data.push_constant_size;
|
|
|
|
- push_constant.push_constants_vk_stage = binary_data.push_constants_vk_stage;
|
|
|
|
|
|
+ push_constant.size = binary_data.push_constant_size;
|
|
|
|
+ push_constant.vk_stages_mask = binary_data.push_constant_vk_stages_mask;
|
|
|
|
|
|
uint32_t vertex_input_mask = binary_data.vertex_input_mask;
|
|
uint32_t vertex_input_mask = binary_data.vertex_input_mask;
|
|
|
|
|
|
- uint32_t fragment_outputs = binary_data.fragment_outputs;
|
|
|
|
|
|
+ uint32_t fragment_output_mask = binary_data.fragment_output_mask;
|
|
|
|
|
|
bool is_compute = binary_data.is_compute;
|
|
bool is_compute = binary_data.is_compute;
|
|
|
|
|
|
@@ -5374,11 +4895,11 @@ RID RenderingDeviceVulkan::shader_create_from_bytecode(const Vector<uint8_t> &p_
|
|
read_offset += set_size;
|
|
read_offset += set_size;
|
|
}
|
|
}
|
|
|
|
|
|
- ERR_FAIL_COND_V(read_offset + binary_data.specialization_constant_count * sizeof(RenderingDeviceVulkanShaderBinarySpecializationConstant) >= binsize, RID());
|
|
|
|
|
|
+ ERR_FAIL_COND_V(read_offset + binary_data.specialization_constants_count * sizeof(RenderingDeviceVulkanShaderBinarySpecializationConstant) >= binsize, RID());
|
|
|
|
|
|
Vector<Shader::SpecializationConstant> specialization_constants;
|
|
Vector<Shader::SpecializationConstant> specialization_constants;
|
|
|
|
|
|
- for (uint32_t i = 0; i < binary_data.specialization_constant_count; i++) {
|
|
|
|
|
|
+ for (uint32_t i = 0; i < binary_data.specialization_constants_count; i++) {
|
|
const RenderingDeviceVulkanShaderBinarySpecializationConstant &src_sc = *(reinterpret_cast<const RenderingDeviceVulkanShaderBinarySpecializationConstant *>(binptr + read_offset));
|
|
const RenderingDeviceVulkanShaderBinarySpecializationConstant &src_sc = *(reinterpret_cast<const RenderingDeviceVulkanShaderBinarySpecializationConstant *>(binptr + read_offset));
|
|
Shader::SpecializationConstant sc;
|
|
Shader::SpecializationConstant sc;
|
|
sc.constant.int_value = src_sc.int_value;
|
|
sc.constant.int_value = src_sc.int_value;
|
|
@@ -5444,7 +4965,7 @@ RID RenderingDeviceVulkan::shader_create_from_bytecode(const Vector<uint8_t> &p_
|
|
Shader shader;
|
|
Shader shader;
|
|
|
|
|
|
shader.vertex_input_mask = vertex_input_mask;
|
|
shader.vertex_input_mask = vertex_input_mask;
|
|
- shader.fragment_output_mask = fragment_outputs;
|
|
|
|
|
|
+ shader.fragment_output_mask = fragment_output_mask;
|
|
shader.push_constant = push_constant;
|
|
shader.push_constant = push_constant;
|
|
shader.is_compute = is_compute;
|
|
shader.is_compute = is_compute;
|
|
shader.compute_local_size[0] = compute_local_size[0];
|
|
shader.compute_local_size[0] = compute_local_size[0];
|
|
@@ -5474,19 +4995,11 @@ RID RenderingDeviceVulkan::shader_create_from_bytecode(const Vector<uint8_t> &p_
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
|
|
|
|
- const VkShaderStageFlagBits shader_stage_bits[SHADER_STAGE_MAX] = {
|
|
|
|
- VK_SHADER_STAGE_VERTEX_BIT,
|
|
|
|
- VK_SHADER_STAGE_FRAGMENT_BIT,
|
|
|
|
- VK_SHADER_STAGE_TESSELLATION_CONTROL_BIT,
|
|
|
|
- VK_SHADER_STAGE_TESSELLATION_EVALUATION_BIT,
|
|
|
|
- VK_SHADER_STAGE_COMPUTE_BIT,
|
|
|
|
- };
|
|
|
|
-
|
|
|
|
VkPipelineShaderStageCreateInfo shader_stage;
|
|
VkPipelineShaderStageCreateInfo shader_stage;
|
|
shader_stage.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
|
|
shader_stage.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
|
|
shader_stage.pNext = nullptr;
|
|
shader_stage.pNext = nullptr;
|
|
shader_stage.flags = 0;
|
|
shader_stage.flags = 0;
|
|
- shader_stage.stage = shader_stage_bits[stage_type[i]];
|
|
|
|
|
|
+ shader_stage.stage = shader_stage_masks[stage_type[i]];
|
|
shader_stage.module = module;
|
|
shader_stage.module = module;
|
|
shader_stage.pName = "main";
|
|
shader_stage.pName = "main";
|
|
shader_stage.pSpecializationInfo = nullptr;
|
|
shader_stage.pSpecializationInfo = nullptr;
|
|
@@ -5558,10 +5071,10 @@ RID RenderingDeviceVulkan::shader_create_from_bytecode(const Vector<uint8_t> &p_
|
|
// Needs to be declared in this outer scope, otherwise it may not outlive its assignment
|
|
// Needs to be declared in this outer scope, otherwise it may not outlive its assignment
|
|
// to pipeline_layout_create_info.
|
|
// to pipeline_layout_create_info.
|
|
VkPushConstantRange push_constant_range;
|
|
VkPushConstantRange push_constant_range;
|
|
- if (push_constant.push_constant_size) {
|
|
|
|
- push_constant_range.stageFlags = push_constant.push_constants_vk_stage;
|
|
|
|
|
|
+ if (push_constant.size) {
|
|
|
|
+ push_constant_range.stageFlags = push_constant.vk_stages_mask;
|
|
push_constant_range.offset = 0;
|
|
push_constant_range.offset = 0;
|
|
- push_constant_range.size = push_constant.push_constant_size;
|
|
|
|
|
|
+ push_constant_range.size = push_constant.size;
|
|
|
|
|
|
pipeline_layout_create_info.pushConstantRangeCount = 1;
|
|
pipeline_layout_create_info.pushConstantRangeCount = 1;
|
|
pipeline_layout_create_info.pPushConstantRanges = &push_constant_range;
|
|
pipeline_layout_create_info.pPushConstantRanges = &push_constant_range;
|
|
@@ -6878,10 +6391,10 @@ RID RenderingDeviceVulkan::render_pipeline_create(RID p_shader, FramebufferForma
|
|
ERR_FAIL_COND_V_MSG(err, RID(), "vkCreateGraphicsPipelines failed with error " + itos(err) + " for shader '" + shader->name + "'.");
|
|
ERR_FAIL_COND_V_MSG(err, RID(), "vkCreateGraphicsPipelines failed with error " + itos(err) + " for shader '" + shader->name + "'.");
|
|
|
|
|
|
pipeline.set_formats = shader->set_formats;
|
|
pipeline.set_formats = shader->set_formats;
|
|
- pipeline.push_constant_stages = shader->push_constant.push_constants_vk_stage;
|
|
|
|
|
|
+ pipeline.push_constant_stages_mask = shader->push_constant.vk_stages_mask;
|
|
pipeline.pipeline_layout = shader->pipeline_layout;
|
|
pipeline.pipeline_layout = shader->pipeline_layout;
|
|
pipeline.shader = p_shader;
|
|
pipeline.shader = p_shader;
|
|
- pipeline.push_constant_size = shader->push_constant.push_constant_size;
|
|
|
|
|
|
+ pipeline.push_constant_size = shader->push_constant.size;
|
|
|
|
|
|
#ifdef DEBUG_ENABLED
|
|
#ifdef DEBUG_ENABLED
|
|
pipeline.validation.dynamic_state = p_dynamic_state_flags;
|
|
pipeline.validation.dynamic_state = p_dynamic_state_flags;
|
|
@@ -6993,10 +6506,10 @@ RID RenderingDeviceVulkan::compute_pipeline_create(RID p_shader, const Vector<Pi
|
|
ERR_FAIL_COND_V_MSG(err, RID(), "vkCreateComputePipelines failed with error " + itos(err) + ".");
|
|
ERR_FAIL_COND_V_MSG(err, RID(), "vkCreateComputePipelines failed with error " + itos(err) + ".");
|
|
|
|
|
|
pipeline.set_formats = shader->set_formats;
|
|
pipeline.set_formats = shader->set_formats;
|
|
- pipeline.push_constant_stages = shader->push_constant.push_constants_vk_stage;
|
|
|
|
|
|
+ pipeline.push_constant_stages_mask = shader->push_constant.vk_stages_mask;
|
|
pipeline.pipeline_layout = shader->pipeline_layout;
|
|
pipeline.pipeline_layout = shader->pipeline_layout;
|
|
pipeline.shader = p_shader;
|
|
pipeline.shader = p_shader;
|
|
- pipeline.push_constant_size = shader->push_constant.push_constant_size;
|
|
|
|
|
|
+ pipeline.push_constant_size = shader->push_constant.size;
|
|
pipeline.local_group_size[0] = shader->compute_local_size[0];
|
|
pipeline.local_group_size[0] = shader->compute_local_size[0];
|
|
pipeline.local_group_size[1] = shader->compute_local_size[1];
|
|
pipeline.local_group_size[1] = shader->compute_local_size[1];
|
|
pipeline.local_group_size[2] = shader->compute_local_size[2];
|
|
pipeline.local_group_size[2] = shader->compute_local_size[2];
|
|
@@ -7638,7 +7151,7 @@ void RenderingDeviceVulkan::draw_list_bind_render_pipeline(DrawListID p_list, RI
|
|
dl->state.set_count = pcount; // Update set count.
|
|
dl->state.set_count = pcount; // Update set count.
|
|
|
|
|
|
if (pipeline->push_constant_size) {
|
|
if (pipeline->push_constant_size) {
|
|
- dl->state.pipeline_push_constant_stages = pipeline->push_constant_stages;
|
|
|
|
|
|
+ dl->state.pipeline_push_constant_stages = pipeline->push_constant_stages_mask;
|
|
#ifdef DEBUG_ENABLED
|
|
#ifdef DEBUG_ENABLED
|
|
dl->validation.pipeline_push_constant_supplied = false;
|
|
dl->validation.pipeline_push_constant_supplied = false;
|
|
#endif
|
|
#endif
|
|
@@ -8254,7 +7767,7 @@ void RenderingDeviceVulkan::compute_list_bind_compute_pipeline(ComputeListID p_l
|
|
cl->state.set_count = pcount; // Update set count.
|
|
cl->state.set_count = pcount; // Update set count.
|
|
|
|
|
|
if (pipeline->push_constant_size) {
|
|
if (pipeline->push_constant_size) {
|
|
- cl->state.pipeline_push_constant_stages = pipeline->push_constant_stages;
|
|
|
|
|
|
+ cl->state.pipeline_push_constant_stages = pipeline->push_constant_stages_mask;
|
|
#ifdef DEBUG_ENABLED
|
|
#ifdef DEBUG_ENABLED
|
|
cl->validation.pipeline_push_constant_supplied = false;
|
|
cl->validation.pipeline_push_constant_supplied = false;
|
|
#endif
|
|
#endif
|