BsVulkanCommandBuffer.cpp 33 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanCommandBuffer.h"
  4. #include "BsVulkanCommandBufferManager.h"
  5. #include "BsVulkanUtility.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParams.h"
  8. #include "BsVulkanQueue.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanIndexBuffer.h"
  11. #include "BsVulkanVertexBuffer.h"
  12. #include "BsVulkanHardwareBuffer.h"
  13. #include "BsVulkanFramebuffer.h"
  14. #include "BsVulkanVertexInputManager.h"
  15. namespace BansheeEngine
  16. {
  17. VulkanCmdBufferPool::VulkanCmdBufferPool(VulkanDevice& device)
  18. :mDevice(device), mNextId(1)
  19. {
  20. for (UINT32 i = 0; i < GQT_COUNT; i++)
  21. {
  22. UINT32 familyIdx = device.getQueueFamily((GpuQueueType)i);
  23. if (familyIdx == (UINT32)-1)
  24. continue;
  25. VkCommandPoolCreateInfo poolCI;
  26. poolCI.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  27. poolCI.pNext = nullptr;
  28. poolCI.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
  29. poolCI.queueFamilyIndex = familyIdx;
  30. PoolInfo& poolInfo = mPools[familyIdx];
  31. poolInfo.queueFamily = familyIdx;
  32. memset(poolInfo.buffers, 0, sizeof(poolInfo.buffers));
  33. vkCreateCommandPool(device.getLogical(), &poolCI, gVulkanAllocator, &poolInfo.pool);
  34. }
  35. }
  36. VulkanCmdBufferPool::~VulkanCmdBufferPool()
  37. {
  38. // Note: Shutdown should be the only place command buffers are destroyed at, as the system relies on the fact that
  39. // they won't be destroyed during normal operation.
  40. for(auto& entry : mPools)
  41. {
  42. PoolInfo& poolInfo = entry.second;
  43. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  44. {
  45. VulkanCmdBuffer* buffer = poolInfo.buffers[i];
  46. if (buffer == nullptr)
  47. break;
  48. bs_delete(buffer);
  49. }
  50. vkDestroyCommandPool(mDevice.getLogical(), poolInfo.pool, gVulkanAllocator);
  51. }
  52. }
  53. VulkanCmdBuffer* VulkanCmdBufferPool::getBuffer(UINT32 queueFamily, bool secondary)
  54. {
  55. auto iterFind = mPools.find(queueFamily);
  56. if (iterFind != mPools.end())
  57. return nullptr;
  58. VulkanCmdBuffer** buffers = iterFind->second.buffers;
  59. UINT32 i = 0;
  60. for(; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  61. {
  62. if (buffers[i] == nullptr)
  63. break;
  64. if(buffers[i]->mState == VulkanCmdBuffer::State::Ready)
  65. {
  66. buffers[i]->begin();
  67. return buffers[i];
  68. }
  69. }
  70. assert(i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY &&
  71. "Too many command buffers allocated. Increment BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY to a higher value. ");
  72. buffers[i] = createBuffer(queueFamily, secondary);
  73. buffers[i]->begin();
  74. return buffers[i];
  75. }
  76. VulkanCmdBuffer* VulkanCmdBufferPool::createBuffer(UINT32 queueFamily, bool secondary)
  77. {
  78. auto iterFind = mPools.find(queueFamily);
  79. if (iterFind != mPools.end())
  80. return nullptr;
  81. const PoolInfo& poolInfo = iterFind->second;
  82. return bs_new<VulkanCmdBuffer>(mDevice, mNextId++, poolInfo.pool, poolInfo.queueFamily, secondary);
  83. }
  84. VulkanCmdBuffer::VulkanCmdBuffer(VulkanDevice& device, UINT32 id, VkCommandPool pool, UINT32 queueFamily, bool secondary)
  85. : mId(id), mQueueFamily(queueFamily), mState(State::Ready), mDevice(device), mPool(pool), mFenceCounter(0)
  86. , mFramebuffer(nullptr), mPresentSemaphore(VK_NULL_HANDLE), mRenderTargetWidth(0), mRenderTargetHeight(0)
  87. , mRenderTargetDepthReadOnly(false), mRenderTargetPreserveContents(false), mGlobalQueueIdx(-1)
  88. , mViewport(0.0f, 0.0f, 1.0f, 1.0f), mScissor(0, 0, 0, 0), mStencilRef(0), mDrawOp(DOT_TRIANGLE_LIST)
  89. , mNumBoundDescriptorSets(0), mGfxPipelineRequiresBind(true), mCmpPipelineRequiresBind(true)
  90. , mViewportRequiresBind(true), mStencilRefRequiresBind(true), mScissorRequiresBind(true), mVertexBuffersTemp()
  91. , mVertexBufferOffsetsTemp()
  92. {
  93. UINT32 maxBoundDescriptorSets = device.getDeviceProperties().limits.maxBoundDescriptorSets;
  94. mDescriptorSetsTemp = (VkDescriptorSet*)bs_alloc(sizeof(VkDescriptorSet) * maxBoundDescriptorSets);
  95. VkCommandBufferAllocateInfo cmdBufferAllocInfo;
  96. cmdBufferAllocInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  97. cmdBufferAllocInfo.pNext = nullptr;
  98. cmdBufferAllocInfo.commandPool = pool;
  99. cmdBufferAllocInfo.level = secondary ? VK_COMMAND_BUFFER_LEVEL_SECONDARY : VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  100. cmdBufferAllocInfo.commandBufferCount = 1;
  101. VkResult result = vkAllocateCommandBuffers(mDevice.getLogical(), &cmdBufferAllocInfo, &mCmdBuffer);
  102. assert(result == VK_SUCCESS);
  103. VkFenceCreateInfo fenceCI;
  104. fenceCI.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  105. fenceCI.pNext = nullptr;
  106. fenceCI.flags = 0;
  107. result = vkCreateFence(mDevice.getLogical(), &fenceCI, gVulkanAllocator, &mFence);
  108. assert(result == VK_SUCCESS);
  109. VkSemaphoreCreateInfo semaphoreCI;
  110. semaphoreCI.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  111. semaphoreCI.pNext = nullptr;
  112. semaphoreCI.flags = 0;
  113. result = vkCreateSemaphore(mDevice.getLogical(), &semaphoreCI, gVulkanAllocator, &mSemaphore);
  114. assert(result == VK_SUCCESS);
  115. }
  116. VulkanCmdBuffer::~VulkanCmdBuffer()
  117. {
  118. VkDevice device = mDevice.getLogical();
  119. if(mState == State::Submitted)
  120. {
  121. // Wait 1s
  122. UINT64 waitTime = 1000 * 1000 * 1000;
  123. VkResult result = vkWaitForFences(device, 1, &mFence, true, waitTime);
  124. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  125. if (result == VK_TIMEOUT)
  126. LOGWRN("Freeing a command buffer before done executing because fence wait expired!");
  127. // Resources have been marked as used, make sure to notify them we're done with them
  128. refreshFenceStatus();
  129. }
  130. else if(mState != State::Ready)
  131. {
  132. // Notify any resources that they are no longer bound
  133. for (auto& entry : mResources)
  134. {
  135. ResourceUseHandle& useHandle = entry.second;
  136. assert(useHandle.used);
  137. entry.first->notifyUnbound();
  138. }
  139. for (auto& entry : mImages)
  140. {
  141. ResourceUseHandle& useHandle = entry.second.useHandle;
  142. assert(useHandle.used);
  143. entry.first->notifyUnbound();
  144. }
  145. for (auto& entry : mBuffers)
  146. {
  147. ResourceUseHandle& useHandle = entry.second.useHandle;
  148. assert(useHandle.used);
  149. entry.first->notifyUnbound();
  150. }
  151. }
  152. vkDestroyFence(device, mFence, gVulkanAllocator);
  153. vkDestroySemaphore(device, mSemaphore, gVulkanAllocator);
  154. vkFreeCommandBuffers(device, mPool, 1, &mCmdBuffer);
  155. bs_free(mDescriptorSetsTemp);
  156. }
  157. UINT32 VulkanCmdBuffer::getDeviceIdx() const
  158. {
  159. return mDevice.getIndex();
  160. }
  161. void VulkanCmdBuffer::begin()
  162. {
  163. assert(mState == State::Ready);
  164. VkCommandBufferBeginInfo beginInfo;
  165. beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  166. beginInfo.pNext = nullptr;
  167. beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  168. beginInfo.pInheritanceInfo = nullptr;
  169. VkResult result = vkBeginCommandBuffer(mCmdBuffer, &beginInfo);
  170. assert(result == VK_SUCCESS);
  171. mState = State::Recording;
  172. }
  173. void VulkanCmdBuffer::end()
  174. {
  175. assert(mState == State::Recording);
  176. VkResult result = vkEndCommandBuffer(mCmdBuffer);
  177. assert(result == VK_SUCCESS);
  178. mState = State::RecordingDone;
  179. }
  180. void VulkanCmdBuffer::beginRenderPass()
  181. {
  182. assert(mState == State::Recording);
  183. if (mFramebuffer == nullptr)
  184. {
  185. LOGWRN("Attempting to begin a render pass but no render target is bound to the command buffer.");
  186. return;
  187. }
  188. VkRenderPassBeginInfo renderPassBeginInfo;
  189. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  190. renderPassBeginInfo.pNext = nullptr;
  191. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(mRenderTargetPreserveContents);
  192. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(mRenderTargetPreserveContents);
  193. renderPassBeginInfo.renderArea.offset.x = 0;
  194. renderPassBeginInfo.renderArea.offset.y = 0;
  195. renderPassBeginInfo.renderArea.extent.width = mRenderTargetWidth;
  196. renderPassBeginInfo.renderArea.extent.height = mRenderTargetHeight;
  197. // TODO: Handle clears (if provided) here. See VulkanRenderAPI::clearRenderTarget.
  198. // - Potential problem is that we might need different framebuffers depending on whether we use load or clear
  199. // ops during render pass start.
  200. renderPassBeginInfo.clearValueCount = 0; // TODO
  201. renderPassBeginInfo.pClearValues = nullptr; // TODO
  202. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  203. mState = State::RecordingRenderPass;
  204. }
  205. void VulkanCmdBuffer::endRenderPass()
  206. {
  207. assert(mState == State::RecordingRenderPass);
  208. vkCmdEndRenderPass(mCmdBuffer);
  209. mState = State::Recording;
  210. }
  211. void VulkanCmdBuffer::submit(VulkanQueue* queue, UINT32 queueIdx, UINT32 syncMask)
  212. {
  213. assert(isReadyForSubmit());
  214. // Issue pipeline barriers for queue transitions (need to happen on original queue first, then on new queue)
  215. for (auto& entry : mBuffers)
  216. {
  217. VulkanBuffer* resource = static_cast<VulkanBuffer*>(entry.first);
  218. if (!resource->isExclusive())
  219. continue;
  220. UINT32 currentQueueFamily = resource->getQueueFamily();
  221. if (currentQueueFamily != mQueueFamily)
  222. {
  223. Vector<VkBufferMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].bufferBarriers;
  224. barriers.push_back(VkBufferMemoryBarrier());
  225. VkBufferMemoryBarrier& barrier = barriers.back();
  226. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  227. barrier.pNext = nullptr;
  228. barrier.srcAccessMask = entry.second.accessFlags;
  229. barrier.dstAccessMask = entry.second.accessFlags;
  230. barrier.srcQueueFamilyIndex = currentQueueFamily;
  231. barrier.dstQueueFamilyIndex = mQueueFamily;
  232. barrier.buffer = resource->getHandle();
  233. barrier.offset = 0;
  234. barrier.size = VK_WHOLE_SIZE;
  235. }
  236. }
  237. for (auto& entry : mImages)
  238. {
  239. VulkanImage* resource = static_cast<VulkanImage*>(entry.first);
  240. UINT32 currentQueueFamily = resource->getQueueFamily();
  241. bool queueMismatch = resource->isExclusive() && currentQueueFamily != mQueueFamily;
  242. if (queueMismatch || resource->getLayout() != entry.second.layout)
  243. {
  244. Vector<VkImageMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].imageBarriers;
  245. barriers.push_back(VkImageMemoryBarrier());
  246. VkImageMemoryBarrier& barrier = barriers.back();
  247. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  248. barrier.pNext = nullptr;
  249. barrier.srcAccessMask = entry.second.accessFlags;
  250. barrier.dstAccessMask = entry.second.accessFlags;
  251. barrier.srcQueueFamilyIndex = currentQueueFamily;
  252. barrier.dstQueueFamilyIndex = mQueueFamily;
  253. barrier.oldLayout = resource->getLayout();
  254. barrier.newLayout = entry.second.layout;
  255. barrier.image = resource->getHandle();
  256. barrier.subresourceRange = entry.second.range;
  257. resource->setLayout(entry.second.layout);
  258. }
  259. }
  260. VulkanDevice& device = queue->getDevice();
  261. for (auto& entry : mTransitionInfoTemp)
  262. {
  263. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  264. if (empty)
  265. continue;
  266. UINT32 entryQueueFamily = entry.first;
  267. // No queue transition needed for entries on this queue (this entry is most likely an in image layout transition)
  268. if (entryQueueFamily == mQueueFamily)
  269. continue;
  270. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(entryQueueFamily, false);
  271. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  272. TransitionInfo& barriers = entry.second;
  273. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  274. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  275. vkCmdPipelineBarrier(vkCmdBuffer,
  276. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // Note: VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT might be more correct here, according to the spec
  277. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // The main idea is that the barrier executes before the semaphore triggers, no actual stage dependencies are needed.
  278. 0, 0, nullptr,
  279. numBufferBarriers, barriers.bufferBarriers.data(),
  280. numImgBarriers, barriers.imageBarriers.data());
  281. // Find an appropriate queue to execute on
  282. UINT32 otherQueueIdx = 0;
  283. VulkanQueue* otherQueue = nullptr;
  284. GpuQueueType otherQueueType = GQT_GRAPHICS;
  285. for (UINT32 i = 0; i < GQT_COUNT; i++)
  286. {
  287. if (device.getQueueFamily((GpuQueueType)i) != entryQueueFamily)
  288. continue;
  289. UINT32 numQueues = device.getNumQueues(otherQueueType);
  290. for (UINT32 j = 0; j < numQueues; j++)
  291. {
  292. // Try to find a queue not currently executing
  293. VulkanQueue* curQueue = device.getQueue(otherQueueType, j);
  294. if (!curQueue->isExecuting())
  295. {
  296. otherQueue = curQueue;
  297. otherQueueIdx = j;
  298. }
  299. }
  300. // Can't find empty one, use the first one then
  301. if (otherQueue == nullptr)
  302. {
  303. otherQueue = device.getQueue(otherQueueType, 0);
  304. otherQueueIdx = 0;
  305. }
  306. otherQueueType = (GpuQueueType)i;
  307. break;
  308. }
  309. syncMask |= CommandSyncMask::getGlobalQueueMask(otherQueueType, otherQueueIdx);
  310. cmdBuffer->end();
  311. cmdBuffer->submit(otherQueue, otherQueueIdx, 0);
  312. // If there are any layout transitions, reset them as we don't need them for the second pipeline barrier
  313. for (auto& barrierEntry : barriers.imageBarriers)
  314. barrierEntry.oldLayout = barrierEntry.newLayout;
  315. }
  316. UINT32 deviceIdx = device.getIndex();
  317. VulkanCommandBufferManager& cbm = static_cast<VulkanCommandBufferManager&>(CommandBufferManager::instance());
  318. UINT32 numSemaphores;
  319. cbm.getSyncSemaphores(deviceIdx, syncMask, mSemaphoresTemp, numSemaphores);
  320. // Wait on present (i.e. until the back buffer becomes available), if we're rendering to a window
  321. if (mPresentSemaphore != VK_NULL_HANDLE)
  322. {
  323. mSemaphoresTemp[numSemaphores] = mPresentSemaphore;
  324. numSemaphores++;
  325. }
  326. // Issue second part of transition pipeline barriers (on this queue)
  327. for (auto& entry : mTransitionInfoTemp)
  328. {
  329. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  330. if (empty)
  331. continue;
  332. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  333. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  334. TransitionInfo& barriers = entry.second;
  335. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  336. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  337. vkCmdPipelineBarrier(vkCmdBuffer,
  338. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // Note: VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT might be more correct here, according to the spec
  339. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT,
  340. 0, 0, nullptr,
  341. numBufferBarriers, barriers.bufferBarriers.data(),
  342. numImgBarriers, barriers.imageBarriers.data());
  343. cmdBuffer->end();
  344. queue->submit(cmdBuffer, mSemaphoresTemp, numSemaphores);
  345. numSemaphores = 0; // Semaphores are only needed the first time, since we're adding the buffers on the same queue
  346. }
  347. queue->submit(this, mSemaphoresTemp, numSemaphores);
  348. mGlobalQueueIdx = CommandSyncMask::getGlobalQueueIdx(queue->getType(), queueIdx);
  349. for (auto& entry : mResources)
  350. {
  351. ResourceUseHandle& useHandle = entry.second;
  352. assert(!useHandle.used);
  353. useHandle.used = true;
  354. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  355. }
  356. for (auto& entry : mImages)
  357. {
  358. ResourceUseHandle& useHandle = entry.second.useHandle;
  359. assert(!useHandle.used);
  360. useHandle.used = true;
  361. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  362. }
  363. for (auto& entry : mBuffers)
  364. {
  365. ResourceUseHandle& useHandle = entry.second.useHandle;
  366. assert(!useHandle.used);
  367. useHandle.used = true;
  368. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  369. }
  370. // Note: Uncommented for debugging only, prevents any device concurrency issues.
  371. // vkQueueWaitIdle(queue->getHandle());
  372. mState = State::Submitted;
  373. cbm.setActiveBuffer(queue->getType(), deviceIdx, queueIdx, this);
  374. // Clear vectors but don't clear the actual map, as we want to re-use the memory since we expect queue family
  375. // indices to be the same
  376. for (auto& entry : mTransitionInfoTemp)
  377. {
  378. entry.second.imageBarriers.clear();
  379. entry.second.bufferBarriers.clear();
  380. }
  381. mGraphicsPipeline = nullptr;
  382. mComputePipeline = nullptr;
  383. mGfxPipelineRequiresBind = true;
  384. mCmpPipelineRequiresBind = true;
  385. }
  386. void VulkanCmdBuffer::refreshFenceStatus()
  387. {
  388. VkResult result = vkGetFenceStatus(mDevice.getLogical(), mFence);
  389. assert(result == VK_SUCCESS || result == VK_NOT_READY);
  390. bool signaled = result == VK_SUCCESS;
  391. if (mState == State::Submitted)
  392. {
  393. if(signaled)
  394. {
  395. mState = State::Ready;
  396. vkResetCommandBuffer(mCmdBuffer, VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT); // Note: Maybe better not to release resources?
  397. result = vkResetFences(mDevice.getLogical(), 1, &mFence);
  398. assert(result == VK_SUCCESS);
  399. mFenceCounter++;
  400. for (auto& entry : mResources)
  401. {
  402. ResourceUseHandle& useHandle = entry.second;
  403. assert(useHandle.used);
  404. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  405. }
  406. for (auto& entry : mImages)
  407. {
  408. ResourceUseHandle& useHandle = entry.second.useHandle;
  409. assert(useHandle.used);
  410. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  411. }
  412. for (auto& entry : mBuffers)
  413. {
  414. ResourceUseHandle& useHandle = entry.second.useHandle;
  415. assert(useHandle.used);
  416. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  417. }
  418. mResources.clear();
  419. mImages.clear();
  420. mBuffers.clear();
  421. }
  422. }
  423. else
  424. assert(!signaled); // We reset the fence along with mState so this shouldn't be possible
  425. }
  426. void VulkanCmdBuffer::setRenderTarget(const SPtr<RenderTargetCore>& rt, bool readOnlyDepthStencil,
  427. bool preserveContents)
  428. {
  429. assert(mState != State::RecordingRenderPass && mState != State::Submitted);
  430. if(rt == nullptr)
  431. {
  432. mFramebuffer = nullptr;
  433. mPresentSemaphore = VK_NULL_HANDLE;
  434. mRenderTargetWidth = 0;
  435. mRenderTargetHeight = 0;
  436. mRenderTargetDepthReadOnly = false;
  437. mRenderTargetPreserveContents = false;
  438. }
  439. else
  440. {
  441. rt->getCustomAttribute("FB", &mFramebuffer);
  442. if (rt->getProperties().isWindow())
  443. rt->getCustomAttribute("PS", &mPresentSemaphore);
  444. else
  445. mPresentSemaphore = VK_NULL_HANDLE;
  446. mRenderTargetWidth = rt->getProperties().getWidth();
  447. mRenderTargetHeight = rt->getProperties().getHeight();
  448. mRenderTargetDepthReadOnly = readOnlyDepthStencil;
  449. mRenderTargetPreserveContents = preserveContents;
  450. registerResource(mFramebuffer, VulkanUseFlag::Write);
  451. }
  452. mGfxPipelineRequiresBind = true;
  453. }
  454. void VulkanCmdBuffer::clearViewport(const Rect2I& area, UINT32 buffers, const Color& color, float depth, UINT16 stencil,
  455. UINT8 targetMask)
  456. {
  457. if (buffers == 0 || mFramebuffer == nullptr)
  458. return;
  459. VkClearAttachment attachments[BS_MAX_MULTIPLE_RENDER_TARGETS + 1];
  460. UINT32 baseLayer = 0;
  461. UINT32 attachmentIdx = 0;
  462. if ((buffers & FBT_COLOR) != 0)
  463. {
  464. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  465. for (UINT32 i = 0; i < numColorAttachments; i++)
  466. {
  467. if (((1 << i) & targetMask) == 0)
  468. continue;
  469. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  470. attachments[attachmentIdx].colorAttachment = i;
  471. VkClearColorValue& colorValue = attachments[attachmentIdx].clearValue.color;
  472. colorValue.float32[0] = color.r;
  473. colorValue.float32[1] = color.g;
  474. colorValue.float32[2] = color.b;
  475. colorValue.float32[3] = color.a;
  476. UINT32 curBaseLayer = mFramebuffer->getColorBaseLayer(i);
  477. if (attachmentIdx == 0)
  478. baseLayer = curBaseLayer;
  479. else
  480. {
  481. if(baseLayer != curBaseLayer)
  482. {
  483. // Note: This could be supported relatively easily: we would need to issue multiple separate
  484. // clear commands for such framebuffers.
  485. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  486. "starting layers. This is currently not supported.");
  487. }
  488. }
  489. attachmentIdx++;
  490. }
  491. }
  492. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  493. {
  494. if (mFramebuffer->hasDepthAttachment())
  495. {
  496. if ((buffers & FBT_DEPTH) != 0)
  497. {
  498. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_DEPTH_BIT;
  499. attachments[attachmentIdx].clearValue.depthStencil.depth = depth;
  500. }
  501. if ((buffers & FBT_STENCIL) != 0)
  502. {
  503. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_STENCIL_BIT;
  504. attachments[attachmentIdx].clearValue.depthStencil.stencil = stencil;
  505. }
  506. attachments[attachmentIdx].colorAttachment = 0;
  507. UINT32 curBaseLayer = mFramebuffer->getDepthStencilBaseLayer();
  508. if (attachmentIdx == 0)
  509. baseLayer = curBaseLayer;
  510. else
  511. {
  512. if (baseLayer != curBaseLayer)
  513. {
  514. // Note: This could be supported relatively easily: we would need to issue multiple separate
  515. // clear commands for such framebuffers.
  516. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  517. "starting layers. This is currently not supported.");
  518. }
  519. }
  520. attachmentIdx++;
  521. }
  522. }
  523. VkClearRect clearRect;
  524. clearRect.baseArrayLayer = baseLayer;
  525. clearRect.layerCount = mFramebuffer->getNumLayers();
  526. clearRect.rect.offset.x = area.x;
  527. clearRect.rect.offset.y = area.y;
  528. clearRect.rect.extent.width = area.width;
  529. clearRect.rect.extent.height = area.height;
  530. UINT32 numAttachments = attachmentIdx;
  531. vkCmdClearAttachments(mCmdBuffer, numAttachments, attachments, 1, &clearRect);
  532. }
  533. void VulkanCmdBuffer::clearRenderTarget(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  534. {
  535. Rect2I area(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  536. clearViewport(area, buffers, color, depth, stencil, targetMask);
  537. }
  538. void VulkanCmdBuffer::clearViewport(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  539. {
  540. Rect2I area;
  541. area.x = (UINT32)(mViewport.x * mRenderTargetWidth);
  542. area.y = (UINT32)(mViewport.y * mRenderTargetHeight);
  543. area.width = (UINT32)(mViewport.width * mRenderTargetWidth);
  544. area.height = (UINT32)(mViewport.height * mRenderTargetHeight);
  545. clearViewport(area, buffers, color, depth, stencil, targetMask);
  546. }
  547. void VulkanCmdBuffer::setPipelineState(const SPtr<GraphicsPipelineStateCore>& state)
  548. {
  549. if (mGraphicsPipeline == state)
  550. return;
  551. mGraphicsPipeline = std::static_pointer_cast<VulkanGraphicsPipelineStateCore>(state);
  552. mGfxPipelineRequiresBind = true;
  553. }
  554. void VulkanCmdBuffer::setPipelineState(const SPtr<ComputePipelineStateCore>& state)
  555. {
  556. if (mComputePipeline == state)
  557. return;
  558. mComputePipeline = std::static_pointer_cast<VulkanComputePipelineStateCore>(state);
  559. mCmpPipelineRequiresBind = true;
  560. }
  561. void VulkanCmdBuffer::setGpuParams(const SPtr<GpuParamsCore>& gpuParams)
  562. {
  563. SPtr<VulkanGpuParams> vulkanGpuParams = std::static_pointer_cast<VulkanGpuParams>(gpuParams);
  564. if(vulkanGpuParams != nullptr)
  565. {
  566. mNumBoundDescriptorSets = vulkanGpuParams->getNumSets();
  567. vulkanGpuParams->prepareForBind(*this, mDescriptorSetsTemp);
  568. }
  569. else
  570. {
  571. mNumBoundDescriptorSets = 0;
  572. }
  573. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  574. }
  575. void VulkanCmdBuffer::setViewport(const Rect2& area)
  576. {
  577. if (mViewport == area)
  578. return;
  579. mViewport = area;
  580. mViewportRequiresBind = true;
  581. }
  582. void VulkanCmdBuffer::setScissorRect(const Rect2I& value)
  583. {
  584. if (mScissor == value)
  585. return;
  586. mScissor = value;
  587. mScissorRequiresBind = true;
  588. }
  589. void VulkanCmdBuffer::setStencilRef(UINT32 value)
  590. {
  591. if (mStencilRef == value)
  592. return;
  593. mStencilRef = value;
  594. mStencilRefRequiresBind = true;
  595. }
  596. void VulkanCmdBuffer::setDrawOp(DrawOperationType drawOp)
  597. {
  598. if (mDrawOp == drawOp)
  599. return;
  600. mDrawOp = drawOp;
  601. mGfxPipelineRequiresBind = true;
  602. }
  603. void VulkanCmdBuffer::setVertexBuffers(UINT32 index, SPtr<VertexBufferCore>* buffers, UINT32 numBuffers)
  604. {
  605. if (numBuffers == 0)
  606. return;
  607. for(UINT32 i = 0; i < numBuffers; i++)
  608. {
  609. VulkanVertexBufferCore* vertexBuffer = static_cast<VulkanVertexBufferCore*>(buffers[i].get());
  610. if (vertexBuffer != nullptr)
  611. {
  612. VulkanBuffer* resource = vertexBuffer->getResource(mDevice.getIndex());
  613. if (resource != nullptr)
  614. {
  615. mVertexBuffersTemp[i] = resource->getHandle();
  616. registerResource(resource, VulkanUseFlag::Read);
  617. }
  618. else
  619. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  620. }
  621. else
  622. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  623. }
  624. vkCmdBindVertexBuffers(mCmdBuffer, index, numBuffers, mVertexBuffersTemp, mVertexBufferOffsetsTemp);
  625. }
  626. void VulkanCmdBuffer::setIndexBuffer(const SPtr<IndexBufferCore>& buffer)
  627. {
  628. VulkanIndexBufferCore* indexBuffer = static_cast<VulkanIndexBufferCore*>(buffer.get());
  629. VkBuffer vkBuffer = VK_NULL_HANDLE;
  630. VkIndexType indexType = VK_INDEX_TYPE_UINT32;
  631. if (indexBuffer != nullptr)
  632. {
  633. VulkanBuffer* resource = indexBuffer->getResource(mDevice.getIndex());
  634. if (resource != nullptr)
  635. {
  636. vkBuffer = resource->getHandle();
  637. indexType = VulkanUtility::getIndexType(buffer->getProperties().getType());
  638. registerResource(resource, VulkanUseFlag::Read);
  639. }
  640. }
  641. vkCmdBindIndexBuffer(mCmdBuffer, vkBuffer, 0, indexType);
  642. }
  643. void VulkanCmdBuffer::setVertexDeclaration(const SPtr<VertexDeclarationCore>& decl)
  644. {
  645. if (mVertexDecl == decl)
  646. return;
  647. mVertexDecl = decl;
  648. mGfxPipelineRequiresBind = true;
  649. }
  650. bool VulkanCmdBuffer::isReadyForRender()
  651. {
  652. if (mGraphicsPipeline == nullptr)
  653. return false;
  654. SPtr<VertexDeclarationCore> inputDecl = mGraphicsPipeline->getInputDeclaration();
  655. if (inputDecl == nullptr)
  656. return false;
  657. return mFramebuffer != nullptr && mVertexDecl != nullptr;
  658. }
  659. bool VulkanCmdBuffer::bindGraphicsPipeline()
  660. {
  661. SPtr<VertexDeclarationCore> inputDecl = mGraphicsPipeline->getInputDeclaration();
  662. SPtr<VulkanVertexInput> vertexInput = VulkanVertexInputManager::instance().getVertexInfo(mVertexDecl, inputDecl);
  663. VulkanPipeline* pipeline = mGraphicsPipeline->getPipeline(mDevice.getIndex(), mFramebuffer,
  664. mRenderTargetDepthReadOnly, mRenderTargetPreserveContents,
  665. mDrawOp, vertexInput);
  666. if (pipeline == nullptr)
  667. return false;
  668. registerResource(pipeline, VulkanUseFlag::Read);
  669. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline->getHandle());
  670. bindDynamicStates(true);
  671. mGfxPipelineRequiresBind = false;
  672. return true;
  673. }
  674. void VulkanCmdBuffer::bindDynamicStates(bool forceAll)
  675. {
  676. if (mViewportRequiresBind || forceAll)
  677. {
  678. VkViewport viewport;
  679. viewport.x = mViewport.x * mRenderTargetWidth;
  680. viewport.y = mViewport.y * mRenderTargetHeight;
  681. viewport.width = mViewport.width * mRenderTargetWidth;
  682. viewport.height = mViewport.height * mRenderTargetHeight;
  683. viewport.minDepth = 0.0f;
  684. viewport.maxDepth = 1.0f;
  685. vkCmdSetViewport(mCmdBuffer, 0, 1, &viewport);
  686. mViewportRequiresBind = false;
  687. }
  688. if(mStencilRefRequiresBind || forceAll)
  689. {
  690. vkCmdSetStencilReference(mCmdBuffer, VK_STENCIL_FRONT_AND_BACK, mStencilRef);
  691. mStencilRefRequiresBind = false;
  692. }
  693. if(mScissorRequiresBind || forceAll)
  694. {
  695. VkRect2D scissorRect;
  696. if(mGraphicsPipeline->isScissorEnabled())
  697. {
  698. scissorRect.offset.x = mScissor.x;
  699. scissorRect.offset.y = mScissor.y;
  700. scissorRect.extent.width = mScissor.width;
  701. scissorRect.extent.height = mScissor.height;
  702. }
  703. else
  704. {
  705. scissorRect.offset.x = 0;
  706. scissorRect.offset.y = 0;
  707. scissorRect.extent.width = mRenderTargetWidth;
  708. scissorRect.extent.height = mRenderTargetHeight;
  709. }
  710. vkCmdSetScissor(mCmdBuffer, 0, 1, &scissorRect);
  711. mScissorRequiresBind = false;
  712. }
  713. }
  714. void VulkanCmdBuffer::draw(UINT32 vertexOffset, UINT32 vertexCount, UINT32 instanceCount)
  715. {
  716. if (!isReadyForRender())
  717. return;
  718. if (!isInRenderPass())
  719. beginRenderPass();
  720. if (mGfxPipelineRequiresBind)
  721. {
  722. if (!bindGraphicsPipeline())
  723. return;
  724. }
  725. else
  726. bindDynamicStates(false);
  727. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  728. {
  729. UINT32 deviceIdx = mDevice.getIndex();
  730. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  731. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  732. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  733. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  734. }
  735. vkCmdDraw(mCmdBuffer, vertexCount, instanceCount, vertexOffset, 0);
  736. }
  737. void VulkanCmdBuffer::drawIndexed(UINT32 startIndex, UINT32 indexCount, UINT32 vertexOffset, UINT32 instanceCount)
  738. {
  739. if (!isReadyForRender())
  740. return;
  741. if (!isInRenderPass())
  742. beginRenderPass();
  743. if (mGfxPipelineRequiresBind)
  744. {
  745. if (!bindGraphicsPipeline())
  746. return;
  747. }
  748. else
  749. bindDynamicStates(false);
  750. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  751. {
  752. UINT32 deviceIdx = mDevice.getIndex();
  753. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  754. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  755. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  756. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  757. }
  758. vkCmdDrawIndexed(mCmdBuffer, indexCount, instanceCount, startIndex, vertexOffset, 0);
  759. }
  760. void VulkanCmdBuffer::dispatch(UINT32 numGroupsX, UINT32 numGroupsY, UINT32 numGroupsZ)
  761. {
  762. if (mComputePipeline == nullptr)
  763. return;
  764. if (isInRenderPass())
  765. endRenderPass();
  766. UINT32 deviceIdx = mDevice.getIndex();
  767. if(mCmpPipelineRequiresBind)
  768. {
  769. VulkanPipeline* pipeline = mComputePipeline->getPipeline(deviceIdx);
  770. if (pipeline == nullptr)
  771. return;
  772. registerResource(pipeline, VulkanUseFlag::Read);
  773. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline->getHandle());
  774. mCmpPipelineRequiresBind = false;
  775. }
  776. if(mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Compute))
  777. {
  778. VkPipelineLayout pipelineLayout = mComputePipeline->getPipelineLayout(deviceIdx);
  779. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipelineLayout, 0,
  780. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  781. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Compute);
  782. }
  783. vkCmdDispatch(mCmdBuffer, numGroupsX, numGroupsY, numGroupsZ);
  784. }
  785. void VulkanCmdBuffer::registerResource(VulkanResource* res, VulkanUseFlags flags)
  786. {
  787. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  788. if(insertResult.second) // New element
  789. {
  790. ResourceUseHandle& useHandle = insertResult.first->second;
  791. useHandle.used = false;
  792. useHandle.flags = flags;
  793. res->notifyBound();
  794. }
  795. else // Existing element
  796. {
  797. ResourceUseHandle& useHandle = insertResult.first->second;
  798. assert(!useHandle.used);
  799. useHandle.flags |= flags;
  800. }
  801. }
  802. void VulkanCmdBuffer::registerResource(VulkanImage* res, VkAccessFlags accessFlags, VkImageLayout layout,
  803. const VkImageSubresourceRange& range, VulkanUseFlags flags)
  804. {
  805. auto insertResult = mImages.insert(std::make_pair(res, ImageInfo()));
  806. if (insertResult.second) // New element
  807. {
  808. ImageInfo& imageInfo = insertResult.first->second;
  809. imageInfo.accessFlags = accessFlags;
  810. imageInfo.layout = layout;
  811. imageInfo.range = range;
  812. imageInfo.useHandle.used = false;
  813. imageInfo.useHandle.flags = flags;
  814. res->notifyBound();
  815. }
  816. else // Existing element
  817. {
  818. ImageInfo& imageInfo = insertResult.first->second;
  819. assert(!imageInfo.useHandle.used);
  820. imageInfo.useHandle.flags |= flags;
  821. assert(imageInfo.layout == layout && "Cannot bind the same image with two different layouts on the same command buffer.");
  822. imageInfo.accessFlags |= accessFlags;
  823. imageInfo.range = range;
  824. }
  825. }
  826. void VulkanCmdBuffer::registerResource(VulkanBuffer* res, VkAccessFlags accessFlags, VulkanUseFlags flags)
  827. {
  828. auto insertResult = mBuffers.insert(std::make_pair(res, BufferInfo()));
  829. if (insertResult.second) // New element
  830. {
  831. BufferInfo& bufferInfo = insertResult.first->second;
  832. bufferInfo.accessFlags = accessFlags;
  833. bufferInfo.useHandle.used = false;
  834. bufferInfo.useHandle.flags = flags;
  835. res->notifyBound();
  836. }
  837. else // Existing element
  838. {
  839. BufferInfo& bufferInfo = insertResult.first->second;
  840. assert(!bufferInfo.useHandle.used);
  841. bufferInfo.useHandle.flags |= flags;
  842. bufferInfo.accessFlags |= accessFlags;
  843. }
  844. }
  845. VulkanCommandBuffer::VulkanCommandBuffer(VulkanDevice& device, GpuQueueType type, UINT32 deviceIdx,
  846. UINT32 queueIdx, bool secondary)
  847. : CommandBuffer(type, deviceIdx, queueIdx, secondary), mBuffer(nullptr)
  848. , mDevice(device), mQueue(nullptr), mIdMask(0)
  849. {
  850. UINT32 numQueues = device.getNumQueues(mType);
  851. if (numQueues == 0) // Fall back to graphics queue
  852. {
  853. mType = GQT_GRAPHICS;
  854. numQueues = device.getNumQueues(GQT_GRAPHICS);
  855. }
  856. mQueue = device.getQueue(mType, mQueueIdx % numQueues);
  857. mIdMask = device.getQueueMask(mType, mQueueIdx);
  858. acquireNewBuffer();
  859. }
  860. void VulkanCommandBuffer::acquireNewBuffer()
  861. {
  862. VulkanCmdBufferPool& pool = mDevice.getCmdBufferPool();
  863. if (mBuffer != nullptr)
  864. assert(mBuffer->isSubmitted());
  865. UINT32 queueFamily = mDevice.getQueueFamily(mType);
  866. mBuffer = pool.getBuffer(queueFamily, mIsSecondary);
  867. }
  868. void VulkanCommandBuffer::submit(UINT32 syncMask)
  869. {
  870. // Ignore myself
  871. syncMask &= ~mIdMask;
  872. if (mBuffer->isInRenderPass())
  873. mBuffer->endRenderPass();
  874. if (mBuffer->isRecording())
  875. mBuffer->end();
  876. mBuffer->submit(mQueue, mQueueIdx, syncMask);
  877. gVulkanCBManager().refreshStates(mDeviceIdx);
  878. acquireNewBuffer();
  879. }
  880. }