BsVulkanGpuParams.cpp 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanGpuParams.h"
  4. #include "BsVulkanUtility.h"
  5. #include "BsVulkanRenderAPI.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParamBlockBuffer.h"
  8. #include "BsVulkanGpuBuffer.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanHardwareBuffer.h"
  11. #include "BsVulkanDescriptorSet.h"
  12. #include "BsVulkanDescriptorLayout.h"
  13. #include "BsVulkanSamplerState.h"
  14. #include "BsVulkanGpuPipelineParamInfo.h"
  15. #include "BsVulkanCommandBuffer.h"
  16. #include "BsGpuParamDesc.h"
  17. namespace BansheeEngine
  18. {
  19. VulkanGpuParams::VulkanGpuParams(const SPtr<GpuPipelineParamInfoCore>& paramInfo, GpuDeviceFlags deviceMask)
  20. : GpuParamsCore(paramInfo, deviceMask), mPerDeviceData(), mDeviceMask(deviceMask), mData(nullptr), mSetsDirty(nullptr)
  21. {
  22. }
  23. VulkanGpuParams::~VulkanGpuParams()
  24. {
  25. {
  26. Lock lock(mMutex);
  27. UINT32 numSets = mParamInfo->getNumSets();
  28. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  29. {
  30. for (UINT32 j = 0; j < numSets; j++)
  31. {
  32. for (auto& entry : mPerDeviceData[i].perSetData[j].sets)
  33. entry->destroy();
  34. }
  35. }
  36. }
  37. bs_free(mData); // Everything allocated under a single buffer to a single free is enough
  38. }
  39. void VulkanGpuParams::initialize()
  40. {
  41. VulkanGpuPipelineParamInfo& vkParamInfo = static_cast<VulkanGpuPipelineParamInfo&>(*mParamInfo);
  42. VulkanRenderAPI& rapi = static_cast<VulkanRenderAPI&>(RenderAPICore::instance());
  43. VulkanDevice* devices[BS_MAX_DEVICES];
  44. VulkanUtility::getDevices(rapi, mDeviceMask, devices);
  45. UINT32 numDevices = 0;
  46. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  47. {
  48. if (devices != nullptr)
  49. numDevices++;
  50. }
  51. UINT32 numSets = vkParamInfo.getNumSets();
  52. UINT32 numBindings = vkParamInfo.getNumElements();
  53. // Note: I'm assuming a single WriteInfo per binding, but if arrays sizes larger than 1 are eventually supported
  54. // I'll need to adjust the code.
  55. UINT32 setsDirtyBytes = sizeof(bool) * numSets;
  56. UINT32 perSetBytes = sizeof(PerSetData) * numSets;
  57. UINT32 writeSetInfosBytes = sizeof(VkWriteDescriptorSet) * numBindings;
  58. UINT32 writeInfosBytes = sizeof(WriteInfo) * numBindings;
  59. mData = (UINT8*)bs_alloc(setsDirtyBytes + (perSetBytes + writeSetInfosBytes + writeInfosBytes) * numDevices);
  60. UINT8* dataIter = mData;
  61. Lock lock(mMutex); // Set write operations need to be thread safe
  62. mSetsDirty = (bool*)dataIter;
  63. memset(mSetsDirty, 1, setsDirtyBytes);
  64. dataIter += setsDirtyBytes;
  65. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  66. {
  67. if (devices[i] == nullptr)
  68. {
  69. mPerDeviceData[i].perSetData = nullptr;
  70. continue;
  71. }
  72. mPerDeviceData[i].perSetData = (PerSetData*)dataIter;
  73. dataIter += sizeof(perSetBytes);
  74. VulkanDescriptorManager& descManager = devices[i]->getDescriptorManager();
  75. for (UINT32 j = 0; j < numSets; j++)
  76. {
  77. UINT32 numBindingsPerSet = vkParamInfo.getNumBindings(j);
  78. PerSetData& perSetData = mPerDeviceData[i].perSetData[j];
  79. perSetData.writeSetInfos = (VkWriteDescriptorSet*)dataIter;
  80. dataIter += sizeof(VkWriteDescriptorSet) * numBindingsPerSet;
  81. perSetData.writeInfos = (WriteInfo*)dataIter;
  82. dataIter += sizeof(WriteInfo) * numBindingsPerSet;
  83. VulkanDescriptorLayout* layout = vkParamInfo.getLayout(i, j);
  84. perSetData.numElements = numBindingsPerSet;
  85. perSetData.latestSet = descManager.createSet(layout);
  86. perSetData.sets.push_back(perSetData.latestSet);
  87. VkDescriptorSetLayoutBinding* perSetBindings = vkParamInfo.getBindings(j);
  88. for (UINT32 k = 0; k < numBindingsPerSet; k++)
  89. {
  90. // Note: Instead of using one structure per binding, it's possible to update multiple at once
  91. // by specifying larger descriptorCount, if they all share type and shader stages.
  92. VkWriteDescriptorSet& writeSetInfo = perSetData.writeSetInfos[k];
  93. writeSetInfo.sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  94. writeSetInfo.pNext = nullptr;
  95. writeSetInfo.dstSet = VK_NULL_HANDLE;
  96. writeSetInfo.dstBinding = perSetBindings[k].binding;
  97. writeSetInfo.dstArrayElement = 0;
  98. writeSetInfo.descriptorCount = perSetBindings[k].descriptorCount;
  99. writeSetInfo.descriptorType = perSetBindings[k].descriptorType;
  100. bool isImage = writeSetInfo.descriptorType == VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER ||
  101. writeSetInfo.descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_IMAGE ||
  102. writeSetInfo.descriptorType == VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
  103. if (isImage)
  104. {
  105. bool isLoadStore = writeSetInfo.descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_IMAGE;
  106. VkDescriptorImageInfo& imageInfo = perSetData.writeInfos[k].image;
  107. imageInfo.sampler = VK_NULL_HANDLE;
  108. imageInfo.imageView = VK_NULL_HANDLE;
  109. imageInfo.imageLayout = isLoadStore ? VK_IMAGE_LAYOUT_GENERAL : VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
  110. writeSetInfo.pImageInfo = &imageInfo;
  111. writeSetInfo.pBufferInfo = nullptr;
  112. writeSetInfo.pTexelBufferView = nullptr;
  113. }
  114. else
  115. {
  116. bool isLoadStore = writeSetInfo.descriptorType == VK_DESCRIPTOR_TYPE_STORAGE_TEXEL_BUFFER;
  117. if (!isLoadStore)
  118. {
  119. VkDescriptorBufferInfo& bufferInfo = perSetData.writeInfos[k].buffer;
  120. bufferInfo.buffer = VK_NULL_HANDLE;
  121. bufferInfo.offset = 0;
  122. bufferInfo.range = VK_WHOLE_SIZE;
  123. writeSetInfo.pBufferInfo = &bufferInfo;
  124. }
  125. else
  126. writeSetInfo.pBufferInfo = nullptr;
  127. writeSetInfo.pTexelBufferView = nullptr;
  128. writeSetInfo.pImageInfo = nullptr;
  129. }
  130. }
  131. }
  132. }
  133. GpuParamsCore::initialize();
  134. }
  135. void VulkanGpuParams::setParamBlockBuffer(UINT32 set, UINT32 slot, const SPtr<GpuParamBlockBufferCore>& paramBlockBuffer)
  136. {
  137. GpuParamsCore::setParamBlockBuffer(set, slot, paramBlockBuffer);
  138. Lock(mMutex);
  139. VulkanGpuParamBlockBufferCore* vulkanParamBlockBuffer =
  140. static_cast<VulkanGpuParamBlockBufferCore*>(paramBlockBuffer.get());
  141. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  142. {
  143. if (mPerDeviceData[i].perSetData == nullptr)
  144. continue;
  145. VulkanBuffer* bufferRes = vulkanParamBlockBuffer->getResource(i);
  146. if (bufferRes != nullptr)
  147. mPerDeviceData[i].perSetData[set].writeInfos[slot].buffer.buffer = bufferRes->getHandle();
  148. else
  149. mPerDeviceData[i].perSetData[set].writeInfos[slot].buffer.buffer = VK_NULL_HANDLE;
  150. }
  151. mSetsDirty[set] = true;
  152. }
  153. void VulkanGpuParams::setTexture(UINT32 set, UINT32 slot, const SPtr<TextureCore>& texture)
  154. {
  155. GpuParamsCore::setTexture(set, slot, texture);
  156. Lock(mMutex);
  157. VulkanTextureCore* vulkanTexture = static_cast<VulkanTextureCore*>(texture.get());
  158. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  159. {
  160. if (mPerDeviceData[i].perSetData == nullptr)
  161. continue;
  162. mPerDeviceData[i].perSetData[set].writeInfos[slot].image.imageView = vulkanTexture->getView(i);
  163. }
  164. mSetsDirty[set] = true;
  165. }
  166. void VulkanGpuParams::setLoadStoreTexture(UINT32 set, UINT32 slot, const SPtr<TextureCore>& texture,
  167. const TextureSurface& surface)
  168. {
  169. GpuParamsCore::setLoadStoreTexture(set, slot, texture, surface);
  170. Lock(mMutex);
  171. VulkanTextureCore* vulkanTexture = static_cast<VulkanTextureCore*>(texture.get());
  172. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  173. {
  174. if (mPerDeviceData[i].perSetData == nullptr)
  175. continue;
  176. mPerDeviceData[i].perSetData[set].writeInfos[slot].image.imageView = vulkanTexture->getView(i, surface);
  177. }
  178. mSetsDirty[set] = true;
  179. }
  180. void VulkanGpuParams::setBuffer(UINT32 set, UINT32 slot, const SPtr<GpuBufferCore>& buffer)
  181. {
  182. GpuParamsCore::setBuffer(set, slot, buffer);
  183. Lock(mMutex);
  184. VulkanGpuBufferCore* vulkanBuffer = static_cast<VulkanGpuBufferCore*>(buffer.get());
  185. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  186. {
  187. if (mPerDeviceData[i].perSetData == nullptr)
  188. continue;
  189. VulkanBuffer* bufferRes = vulkanBuffer->getResource(i);
  190. if (bufferRes != nullptr)
  191. mPerDeviceData[i].perSetData[set].writeInfos[slot].bufferView = bufferRes->getView();
  192. else
  193. mPerDeviceData[i].perSetData[set].writeInfos[slot].bufferView = VK_NULL_HANDLE;
  194. }
  195. mSetsDirty[set] = true;
  196. }
  197. void VulkanGpuParams::setSamplerState(UINT32 set, UINT32 slot, const SPtr<SamplerStateCore>& sampler)
  198. {
  199. GpuParamsCore::setSamplerState(set, slot, sampler);
  200. Lock(mMutex);
  201. VulkanSamplerStateCore* vulkanSampler = static_cast<VulkanSamplerStateCore*>(sampler.get());
  202. for(UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  203. {
  204. if (mPerDeviceData[i].perSetData == nullptr)
  205. continue;
  206. VulkanSampler* samplerRes = vulkanSampler->getResource(i);
  207. if (samplerRes != nullptr)
  208. mPerDeviceData[i].perSetData[set].writeInfos[slot].image.sampler = samplerRes->getHandle();
  209. else
  210. mPerDeviceData[i].perSetData[set].writeInfos[slot].image.sampler = VK_NULL_HANDLE;
  211. }
  212. mSetsDirty[set] = true;
  213. }
  214. void VulkanGpuParams::setLoadStoreSurface(UINT32 set, UINT32 slot, const TextureSurface& surface)
  215. {
  216. GpuParamsCore::setLoadStoreSurface(set, slot, surface);
  217. SPtr<TextureCore> texture = getLoadStoreTexture(set, slot);
  218. if (texture == nullptr)
  219. return;
  220. Lock(mMutex);
  221. VulkanTextureCore* vulkanTexture = static_cast<VulkanTextureCore*>(texture.get());
  222. for (UINT32 i = 0; i < BS_MAX_DEVICES; i++)
  223. {
  224. if (mPerDeviceData[i].perSetData == nullptr)
  225. continue;
  226. mPerDeviceData[i].perSetData[set].writeInfos[slot].image.imageView = vulkanTexture->getView(i, surface);
  227. }
  228. mSetsDirty[set] = true;
  229. }
  230. void VulkanGpuParams::bind(VulkanCommandBuffer& buffer)
  231. {
  232. UINT32 deviceIdx = buffer.getDeviceIdx();
  233. PerDeviceData& perDeviceData = mPerDeviceData[deviceIdx];
  234. if (perDeviceData.perSetData == nullptr)
  235. return;
  236. UINT32 numParamBlocks = mParamInfo->getNumElements(GpuPipelineParamInfo::ParamType::ParamBlock);
  237. UINT32 numTextures = mParamInfo->getNumElements(GpuPipelineParamInfo::ParamType::Texture);
  238. UINT32 numStorageTextures = mParamInfo->getNumElements(GpuPipelineParamInfo::ParamType::LoadStoreTexture);
  239. UINT32 numBuffers = mParamInfo->getNumElements(GpuPipelineParamInfo::ParamType::Buffer);
  240. UINT32 numSamplers = mParamInfo->getNumElements(GpuPipelineParamInfo::ParamType::SamplerState);
  241. UINT32 numSets = mParamInfo->getNumSets();
  242. // Registers resources with the command buffer
  243. // Note: Makes the assumption that this object (and all of the resources it holds) are externally locked, and will
  244. // not be modified on another thread while being bound.
  245. VulkanCmdBuffer* internalCB = buffer.getInternal();
  246. for (UINT32 i = 0; i < numParamBlocks; i++)
  247. {
  248. if (mParamBlockBuffers[i] == nullptr)
  249. continue;
  250. VulkanGpuParamBlockBufferCore* element = static_cast<VulkanGpuParamBlockBufferCore*>(mParamBlockBuffers[i].get());
  251. VulkanBuffer* resource = element->getResource(deviceIdx);
  252. internalCB->registerResource(resource, VK_ACCESS_UNIFORM_READ_BIT, VulkanUseFlag::Read);
  253. }
  254. for (UINT32 i = 0; i < numBuffers; i++)
  255. {
  256. if (mBuffers[i] == nullptr)
  257. continue;
  258. VulkanGpuBufferCore* element = static_cast<VulkanGpuBufferCore*>(mBuffers[i].get());
  259. VkAccessFlags accessFlags = VK_ACCESS_SHADER_READ_BIT;
  260. VulkanUseFlags useFlags = VulkanUseFlag::Read;
  261. if (element->getProperties().getRandomGpuWrite())
  262. {
  263. accessFlags |= VK_ACCESS_SHADER_WRITE_BIT;
  264. useFlags |= VulkanUseFlag::Write;
  265. }
  266. VulkanBuffer* resource = element->getResource(deviceIdx);
  267. internalCB->registerResource(resource, accessFlags, useFlags);
  268. }
  269. for (UINT32 i = 0; i < numSamplers; i++)
  270. {
  271. if (mSamplerStates[i] == nullptr)
  272. continue;
  273. VulkanSamplerStateCore* element = static_cast<VulkanSamplerStateCore*>(mSamplerStates[i].get());
  274. VulkanSampler* resource = element->getResource(deviceIdx);
  275. if (resource == nullptr)
  276. continue;
  277. internalCB->registerResource(resource, VulkanUseFlag::Read);
  278. }
  279. for (UINT32 i = 0; i < numStorageTextures; i++)
  280. {
  281. if (mLoadStoreTextures[i] == nullptr)
  282. continue;
  283. VulkanTextureCore* element = static_cast<VulkanTextureCore*>(mLoadStoreTextures[i].get());
  284. VulkanImage* resource = element->getResource(deviceIdx);
  285. if (resource == nullptr)
  286. continue;
  287. VkAccessFlags accessFlags = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT;
  288. VulkanUseFlags useFlags = VulkanUseFlag::Read | VulkanUseFlag::Write;
  289. const TextureSurface& surface = mLoadStoreSurfaces[i];
  290. VkImageSubresourceRange range;
  291. range.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  292. range.baseArrayLayer = surface.arraySlice;
  293. range.layerCount = surface.numArraySlices;
  294. range.baseMipLevel = surface.mipLevel;
  295. range.levelCount = surface.numMipLevels;
  296. internalCB->registerResource(resource, accessFlags, VK_IMAGE_LAYOUT_GENERAL, range, useFlags);
  297. }
  298. for (UINT32 i = 0; i < numTextures; i++)
  299. {
  300. if (mTextures[i] == nullptr)
  301. continue;
  302. VulkanTextureCore* element = static_cast<VulkanTextureCore*>(mTextures[i].get());
  303. VulkanImage* resource = element->getResource(deviceIdx);
  304. if (resource == nullptr)
  305. continue;
  306. const TextureProperties& props = element->getProperties();
  307. bool isDepthStencil = (props.getUsage() & TU_DEPTHSTENCIL) != 0;
  308. VkImageSubresourceRange range;
  309. range.aspectMask = isDepthStencil ? VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT
  310. : VK_IMAGE_ASPECT_COLOR_BIT;
  311. range.baseArrayLayer = 0;
  312. range.layerCount = props.getNumFaces();
  313. range.baseMipLevel = 0;
  314. range.levelCount = props.getNumMipmaps();
  315. internalCB->registerResource(resource, VK_ACCESS_SHADER_READ_BIT, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL,
  316. range, VulkanUseFlag::Read);
  317. }
  318. // Acquire sets as needed, and updated their contents if dirty
  319. VulkanRenderAPI& rapi = static_cast<VulkanRenderAPI&>(RenderAPICore::instance());
  320. VulkanDevice& device = *rapi._getDevice(deviceIdx);
  321. VulkanDescriptorManager& descManager = device.getDescriptorManager();
  322. VulkanGpuPipelineParamInfo& vkParamInfo = static_cast<VulkanGpuPipelineParamInfo&>(*mParamInfo);
  323. Lock(mMutex);
  324. for (UINT32 i = 0; i < numSets; i++)
  325. {
  326. PerSetData& perSetData = perDeviceData.perSetData[i];
  327. if (!mSetsDirty[i]) // Set not dirty, just use the last one we wrote (this is fine even across multiple command buffers)
  328. continue;
  329. // Set is dirty, we need to update
  330. //// Use latest unless already used, otherwise try to find an unused one
  331. if(perSetData.latestSet->isBound()) // Checking this is okay, because it's only modified below when we call registerResource, which is under the same lock as this
  332. {
  333. perSetData.latestSet = nullptr;
  334. for(auto& entry : perSetData.sets)
  335. {
  336. if(!entry->isBound())
  337. {
  338. perSetData.latestSet = entry;
  339. break;
  340. }
  341. }
  342. // Cannot find an empty set, allocate a new one
  343. VulkanDescriptorLayout* layout = vkParamInfo.getLayout(deviceIdx, i);
  344. perSetData.latestSet = descManager.createSet(layout);
  345. perSetData.sets.push_back(perSetData.latestSet);
  346. }
  347. // Note: Currently I write to the entire set at once, but it might be beneficial to remember only the exact
  348. // entries that were updated, and only write to them individually.
  349. perSetData.latestSet->write(perSetData.writeSetInfos, perSetData.numElements);
  350. mSetsDirty[i] = false;
  351. }
  352. // Actually bind the sets to the command buffer
  353. VkCommandBuffer vkCB = internalCB->getHandle();
  354. VkPipelineBindPoint bindPoint;
  355. GpuQueueType queueType = buffer.getType();
  356. switch(queueType)
  357. {
  358. case GQT_GRAPHICS:
  359. bindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS;
  360. break;
  361. case GQT_COMPUTE:
  362. bindPoint = VK_PIPELINE_BIND_POINT_COMPUTE;
  363. break;
  364. case GQT_UPLOAD:
  365. default:
  366. LOGERR("Cannot bind GpuParams on the upload queue. Ignoring.");
  367. return;
  368. }
  369. VkDescriptorSet* sets = bs_stack_alloc<VkDescriptorSet>(numSets);
  370. for (UINT32 i = 0; i < numSets; i++)
  371. {
  372. VulkanDescriptorSet* set = perDeviceData.perSetData[i].latestSet;
  373. internalCB->registerResource(set, VulkanUseFlag::Read);
  374. sets[i] = set->getHandle();
  375. }
  376. vkCmdBindDescriptorSets(vkCB, bindPoint, perDeviceData.pipelineLayout, 0,
  377. numSets, sets, 0, nullptr);
  378. bs_stack_free(sets);
  379. }
  380. }