VkAccelerationStructure.cpp 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334
  1. // Copyright (C) 2009-present, Panagiotis Christopoulos Charitos and contributors.
  2. // All rights reserved.
  3. // Code licensed under the BSD License.
  4. // http://www.anki3d.org/LICENSE
  5. #include <AnKi/Gr/Vulkan/VkAccelerationStructure.h>
  6. #include <AnKi/Gr/Vulkan/VkGrManager.h>
  7. #include <AnKi/Gr/Vulkan/VkBuffer.h>
  8. namespace anki {
  9. AccelerationStructure* AccelerationStructure::newInstance(const AccelerationStructureInitInfo& init)
  10. {
  11. AccelerationStructureImpl* impl = anki::newInstance<AccelerationStructureImpl>(GrMemoryPool::getSingleton(), init.getName());
  12. const Error err = impl->init(init);
  13. if(err)
  14. {
  15. deleteInstance(GrMemoryPool::getSingleton(), impl);
  16. impl = nullptr;
  17. }
  18. return impl;
  19. }
  20. U64 AccelerationStructure::getGpuAddress() const
  21. {
  22. ANKI_VK_SELF_CONST(AccelerationStructureImpl);
  23. return self.m_deviceAddress;
  24. }
  25. AccelerationStructureImpl::~AccelerationStructureImpl()
  26. {
  27. if(m_handle)
  28. {
  29. vkDestroyAccelerationStructureKHR(getVkDevice(), m_handle, nullptr);
  30. }
  31. }
  32. Error AccelerationStructureImpl::init(const AccelerationStructureInitInfo& inf)
  33. {
  34. ANKI_TRACE_SCOPED_EVENT(VkInitAccStruct);
  35. ANKI_ASSERT(inf.isValid());
  36. m_type = inf.m_type;
  37. const VkDevice vkdev = getGrManagerImpl().getDevice();
  38. PtrSize asBufferSize;
  39. getMemoryRequirement(inf, asBufferSize, m_scratchBufferSize);
  40. m_scratchBufferSize += getGrManagerImpl().getVulkanCapabilities().m_asBuildScratchAlignment;
  41. // Allocate AS buffer
  42. BufferView asBuff = inf.m_accelerationStructureBuffer;
  43. if(!asBuff.isValid())
  44. {
  45. BufferInitInfo bufferInit(inf.getName());
  46. bufferInit.m_usage = BufferUsageBit::kAccelerationStructure;
  47. bufferInit.m_size = asBufferSize;
  48. m_asBuffer = getGrManagerImpl().newBuffer(bufferInit);
  49. m_asBufferOffset = 0;
  50. }
  51. else
  52. {
  53. const PtrSize alignedOffset = getAlignedRoundUp(getGrManagerImpl().getVulkanCapabilities().m_asBufferAlignment, asBuff.getOffset());
  54. asBuff = asBuff.incrementOffset(alignedOffset - asBuff.getOffset());
  55. ANKI_ASSERT(asBuff.getRange() >= asBufferSize);
  56. m_asBuffer.reset(&asBuff.getBuffer());
  57. m_asBufferOffset = asBuff.getOffset();
  58. }
  59. // Create the AS
  60. if(m_type == AccelerationStructureType::kBottomLevel)
  61. {
  62. // Geom
  63. VkAccelerationStructureGeometryKHR& geom = m_geometry;
  64. geom.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_KHR;
  65. geom.geometryType = VK_GEOMETRY_TYPE_TRIANGLES_KHR;
  66. geom.geometry.triangles.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_TRIANGLES_DATA_KHR;
  67. geom.geometry.triangles.vertexFormat = convertFormat(inf.m_bottomLevel.m_positionsFormat);
  68. geom.geometry.triangles.vertexData.deviceAddress =
  69. inf.m_bottomLevel.m_positionBuffer.getBuffer().getGpuAddress() + inf.m_bottomLevel.m_positionBuffer.getOffset();
  70. geom.geometry.triangles.vertexStride = inf.m_bottomLevel.m_positionStride;
  71. geom.geometry.triangles.maxVertex = inf.m_bottomLevel.m_positionCount - 1;
  72. geom.geometry.triangles.indexType = convertIndexType(inf.m_bottomLevel.m_indexType);
  73. geom.geometry.triangles.indexData.deviceAddress =
  74. inf.m_bottomLevel.m_indexBuffer.getBuffer().getGpuAddress() + inf.m_bottomLevel.m_indexBuffer.getOffset();
  75. geom.flags = 0;
  76. // Geom build info
  77. VkAccelerationStructureBuildGeometryInfoKHR& buildInfo = m_buildInfo;
  78. buildInfo.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_BUILD_GEOMETRY_INFO_KHR;
  79. buildInfo.type = VK_ACCELERATION_STRUCTURE_TYPE_BOTTOM_LEVEL_KHR;
  80. buildInfo.flags = VK_BUILD_ACCELERATION_STRUCTURE_PREFER_FAST_TRACE_BIT_KHR;
  81. buildInfo.mode = VK_BUILD_ACCELERATION_STRUCTURE_MODE_BUILD_KHR;
  82. buildInfo.geometryCount = 1;
  83. buildInfo.pGeometries = &geom;
  84. // Create the AS
  85. VkAccelerationStructureCreateInfoKHR asCi = {};
  86. asCi.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_CREATE_INFO_KHR;
  87. asCi.createFlags = 0;
  88. asCi.buffer = static_cast<const BufferImpl&>(*m_asBuffer).getHandle();
  89. asCi.offset = m_asBufferOffset;
  90. asCi.size = asBufferSize;
  91. asCi.type = VK_ACCELERATION_STRUCTURE_TYPE_BOTTOM_LEVEL_KHR;
  92. ANKI_VK_CHECK(vkCreateAccelerationStructureKHR(vkdev, &asCi, nullptr, &m_handle));
  93. // Get its address
  94. VkAccelerationStructureDeviceAddressInfoKHR addressInfo = {};
  95. addressInfo.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_DEVICE_ADDRESS_INFO_KHR;
  96. addressInfo.accelerationStructure = m_handle;
  97. m_deviceAddress = vkGetAccelerationStructureDeviceAddressKHR(vkdev, &addressInfo);
  98. // Almost finalize the build info
  99. buildInfo.dstAccelerationStructure = m_handle;
  100. // Range info
  101. m_rangeInfo.primitiveCount = inf.m_bottomLevel.m_indexCount / 3;
  102. m_blas.m_positionsBuffer.reset(&inf.m_bottomLevel.m_positionBuffer.getBuffer());
  103. m_blas.m_indexBuffer.reset(&inf.m_bottomLevel.m_positionBuffer.getBuffer());
  104. }
  105. else
  106. {
  107. const Bool isEmpty = inf.m_topLevel.m_instanceCount == 0;
  108. if(!isEmpty)
  109. {
  110. ANKI_ASSERT(sizeof(VkAccelerationStructureInstanceKHR) * inf.m_topLevel.m_instanceCount <= inf.m_topLevel.m_instancesBuffer.getRange());
  111. m_tlas.m_instancesBuffer.reset(&inf.m_topLevel.m_instancesBuffer.getBuffer());
  112. }
  113. // Geom
  114. VkAccelerationStructureGeometryKHR& geom = m_geometry;
  115. geom.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_KHR;
  116. geom.geometryType = VK_GEOMETRY_TYPE_INSTANCES_KHR;
  117. geom.geometry.instances.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_INSTANCES_DATA_KHR;
  118. geom.geometry.instances.data.deviceAddress =
  119. (isEmpty) ? 0 : (m_tlas.m_instancesBuffer->getGpuAddress() + inf.m_topLevel.m_instancesBuffer.getOffset());
  120. geom.geometry.instances.arrayOfPointers = false;
  121. geom.flags = 0;
  122. // Geom build info
  123. VkAccelerationStructureBuildGeometryInfoKHR& buildInfo = m_buildInfo;
  124. buildInfo.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_BUILD_GEOMETRY_INFO_KHR;
  125. buildInfo.type = VK_ACCELERATION_STRUCTURE_TYPE_TOP_LEVEL_KHR;
  126. buildInfo.flags = VK_BUILD_ACCELERATION_STRUCTURE_PREFER_FAST_BUILD_BIT_KHR;
  127. buildInfo.mode = VK_BUILD_ACCELERATION_STRUCTURE_MODE_BUILD_KHR;
  128. buildInfo.geometryCount = 1;
  129. buildInfo.pGeometries = &geom;
  130. // Create the AS
  131. VkAccelerationStructureCreateInfoKHR asCi = {};
  132. asCi.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_CREATE_INFO_KHR;
  133. asCi.createFlags = 0;
  134. asCi.buffer = static_cast<const BufferImpl&>(*m_asBuffer).getHandle();
  135. asCi.offset = m_asBufferOffset;
  136. asCi.size = asBufferSize;
  137. asCi.type = VK_ACCELERATION_STRUCTURE_TYPE_TOP_LEVEL_KHR;
  138. ANKI_VK_CHECK(vkCreateAccelerationStructureKHR(vkdev, &asCi, nullptr, &m_handle));
  139. // Almost finalize the build info
  140. buildInfo.dstAccelerationStructure = m_handle;
  141. // Range info
  142. m_rangeInfo.primitiveCount = inf.m_topLevel.m_instanceCount;
  143. }
  144. return Error::kNone;
  145. }
  146. VkMemoryBarrier AccelerationStructureImpl::computeBarrierInfo(AccelerationStructureUsageBit before, AccelerationStructureUsageBit after,
  147. VkPipelineStageFlags& srcStages_, VkPipelineStageFlags& dstStages_)
  148. {
  149. VkMemoryBarrier barrier = {};
  150. barrier.sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER;
  151. // Before
  152. VkPipelineStageFlags srcStages = 0;
  153. if(before == AccelerationStructureUsageBit::kNone)
  154. {
  155. srcStages |= VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  156. barrier.srcAccessMask |= 0;
  157. }
  158. if(!!(before & AccelerationStructureUsageBit::kBuild))
  159. {
  160. srcStages |= VK_PIPELINE_STAGE_ACCELERATION_STRUCTURE_BUILD_BIT_KHR;
  161. barrier.srcAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_WRITE_BIT_KHR;
  162. }
  163. if(!!(before & AccelerationStructureUsageBit::kAttach))
  164. {
  165. srcStages |= VK_PIPELINE_STAGE_ACCELERATION_STRUCTURE_BUILD_BIT_KHR;
  166. barrier.srcAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_READ_BIT_KHR;
  167. }
  168. if(!!(before & AccelerationStructureUsageBit::kSrvGeometry))
  169. {
  170. srcStages |= VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT
  171. | VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT | VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT;
  172. barrier.srcAccessMask |= VK_ACCESS_MEMORY_READ_BIT; // READ_BIT is the only viable solution by elimination
  173. }
  174. if(!!(before & AccelerationStructureUsageBit::kSrvPixel))
  175. {
  176. srcStages |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
  177. barrier.srcAccessMask |= VK_ACCESS_MEMORY_READ_BIT;
  178. }
  179. if(!!(before & AccelerationStructureUsageBit::kSrvCompute))
  180. {
  181. srcStages |= VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  182. barrier.srcAccessMask |= VK_ACCESS_MEMORY_READ_BIT;
  183. }
  184. if(!!(before & AccelerationStructureUsageBit::kSrvDispatchRays))
  185. {
  186. srcStages |= VK_PIPELINE_STAGE_RAY_TRACING_SHADER_BIT_KHR;
  187. barrier.srcAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_READ_BIT_KHR;
  188. }
  189. // After
  190. VkPipelineStageFlags dstStages = 0;
  191. if(!!(after & AccelerationStructureUsageBit::kBuild))
  192. {
  193. dstStages |= VK_PIPELINE_STAGE_ACCELERATION_STRUCTURE_BUILD_BIT_KHR;
  194. barrier.dstAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_WRITE_BIT_KHR;
  195. }
  196. if(!!(after & AccelerationStructureUsageBit::kAttach))
  197. {
  198. dstStages |= VK_PIPELINE_STAGE_ACCELERATION_STRUCTURE_BUILD_BIT_KHR;
  199. barrier.dstAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_READ_BIT_KHR;
  200. }
  201. if(!!(after & AccelerationStructureUsageBit::kSrvGeometry))
  202. {
  203. dstStages |= VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT
  204. | VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT | VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT;
  205. barrier.dstAccessMask |= VK_ACCESS_MEMORY_READ_BIT; // READ_BIT is the only viable solution by elimination
  206. }
  207. if(!!(after & AccelerationStructureUsageBit::kSrvPixel))
  208. {
  209. dstStages |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
  210. barrier.dstAccessMask |= VK_ACCESS_MEMORY_READ_BIT;
  211. }
  212. if(!!(after & AccelerationStructureUsageBit::kSrvCompute))
  213. {
  214. dstStages |= VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  215. barrier.dstAccessMask |= VK_ACCESS_MEMORY_READ_BIT;
  216. }
  217. if(!!(after & AccelerationStructureUsageBit::kSrvDispatchRays))
  218. {
  219. dstStages |= VK_PIPELINE_STAGE_RAY_TRACING_SHADER_BIT_KHR;
  220. barrier.dstAccessMask |= VK_ACCESS_ACCELERATION_STRUCTURE_READ_BIT_KHR;
  221. }
  222. ANKI_ASSERT(srcStages && dstStages);
  223. srcStages_ |= srcStages;
  224. dstStages_ |= dstStages;
  225. return barrier;
  226. }
  227. void AccelerationStructureImpl::generateBuildInfo(U64 scratchBufferAddress, VkAccelerationStructureBuildGeometryInfoKHR& buildInfo,
  228. VkAccelerationStructureBuildRangeInfoKHR& rangeInfo) const
  229. {
  230. buildInfo = m_buildInfo;
  231. buildInfo.scratchData.deviceAddress =
  232. getAlignedRoundUp(getGrManagerImpl().getVulkanCapabilities().m_asBuildScratchAlignment, scratchBufferAddress);
  233. rangeInfo = m_rangeInfo;
  234. }
  235. void AccelerationStructureImpl::getMemoryRequirement(const AccelerationStructureInitInfo& inf, PtrSize& asBufferSize, PtrSize& buildScratchBufferSize)
  236. {
  237. ANKI_ASSERT(inf.isValidForGettingMemoryRequirements());
  238. VkAccelerationStructureGeometryKHR geom = {};
  239. geom.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_KHR;
  240. VkAccelerationStructureBuildGeometryInfoKHR buildInfo = {};
  241. buildInfo.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_BUILD_GEOMETRY_INFO_KHR;
  242. VkAccelerationStructureBuildSizesInfoKHR buildSizes = {};
  243. buildSizes.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_BUILD_SIZES_INFO_KHR;
  244. if(inf.m_type == AccelerationStructureType::kBottomLevel)
  245. {
  246. geom.geometryType = VK_GEOMETRY_TYPE_TRIANGLES_KHR;
  247. geom.geometry.triangles.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_TRIANGLES_DATA_KHR;
  248. geom.geometry.triangles.vertexFormat = convertFormat(inf.m_bottomLevel.m_positionsFormat);
  249. geom.geometry.triangles.vertexStride = inf.m_bottomLevel.m_positionStride;
  250. geom.geometry.triangles.maxVertex = inf.m_bottomLevel.m_positionCount - 1;
  251. geom.geometry.triangles.indexType = convertIndexType(inf.m_bottomLevel.m_indexType);
  252. geom.flags = 0;
  253. buildInfo.type = VK_ACCELERATION_STRUCTURE_TYPE_BOTTOM_LEVEL_KHR;
  254. buildInfo.flags = VK_BUILD_ACCELERATION_STRUCTURE_PREFER_FAST_TRACE_BIT_KHR;
  255. buildInfo.mode = VK_BUILD_ACCELERATION_STRUCTURE_MODE_BUILD_KHR;
  256. buildInfo.geometryCount = 1;
  257. buildInfo.pGeometries = &geom;
  258. const U32 primitiveCount = inf.m_bottomLevel.m_indexCount / 3;
  259. vkGetAccelerationStructureBuildSizesKHR(getVkDevice(), VK_ACCELERATION_STRUCTURE_BUILD_TYPE_DEVICE_KHR, &buildInfo, &primitiveCount,
  260. &buildSizes);
  261. }
  262. else
  263. {
  264. geom.geometryType = VK_GEOMETRY_TYPE_INSTANCES_KHR;
  265. geom.geometry.instances.sType = VK_STRUCTURE_TYPE_ACCELERATION_STRUCTURE_GEOMETRY_INSTANCES_DATA_KHR;
  266. geom.geometry.instances.arrayOfPointers = false;
  267. geom.flags = 0;
  268. buildInfo.type = VK_ACCELERATION_STRUCTURE_TYPE_TOP_LEVEL_KHR;
  269. buildInfo.flags = VK_BUILD_ACCELERATION_STRUCTURE_PREFER_FAST_BUILD_BIT_KHR;
  270. buildInfo.mode = VK_BUILD_ACCELERATION_STRUCTURE_MODE_BUILD_KHR;
  271. buildInfo.geometryCount = 1;
  272. buildInfo.pGeometries = &geom;
  273. vkGetAccelerationStructureBuildSizesKHR(getVkDevice(), VK_ACCELERATION_STRUCTURE_BUILD_TYPE_DEVICE_KHR, &buildInfo,
  274. &inf.m_topLevel.m_instanceCount, &buildSizes);
  275. }
  276. asBufferSize = buildSizes.accelerationStructureSize;
  277. buildScratchBufferSize = buildSizes.buildScratchSize;
  278. }
  279. } // end namespace anki