BsVulkanCommandBuffer.cpp 78 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309231023112312231323142315231623172318231923202321232223232324232523262327232823292330233123322333233423352336233723382339234023412342234323442345234623472348234923502351235223532354235523562357235823592360236123622363236423652366236723682369237023712372237323742375237623772378237923802381238223832384238523862387238823892390239123922393239423952396239723982399240024012402240324042405240624072408240924102411241224132414241524162417241824192420
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanCommandBuffer.h"
  4. #include "Managers/BsVulkanCommandBufferManager.h"
  5. #include "BsVulkanUtility.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParams.h"
  8. #include "BsVulkanQueue.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanIndexBuffer.h"
  11. #include "BsVulkanVertexBuffer.h"
  12. #include "BsVulkanHardwareBuffer.h"
  13. #include "BsVulkanFramebuffer.h"
  14. #include "Managers/BsVulkanVertexInputManager.h"
  15. #include "BsVulkanEventQuery.h"
  16. #include "Managers/BsVulkanQueryManager.h"
  17. #include "BsVulkanSwapChain.h"
  18. #include "BsVulkanTimerQuery.h"
  19. #include "BsVulkanOcclusionQuery.h"
  20. #if BS_PLATFORM == BS_PLATFORM_WIN32
  21. #include "Win32/BsWin32RenderWindow.h"
  22. #elif BS_PLATFORM == BS_PLATFORM_LINUX
  23. #include "Linux/BsLinuxRenderWindow.h"
  24. #else
  25. static_assert(false, "Other platforms go here");
  26. #endif
  27. namespace bs { namespace ct
  28. {
  29. VulkanSemaphore::VulkanSemaphore(VulkanResourceManager* owner)
  30. :VulkanResource(owner, true)
  31. {
  32. VkSemaphoreCreateInfo semaphoreCI;
  33. semaphoreCI.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  34. semaphoreCI.pNext = nullptr;
  35. semaphoreCI.flags = 0;
  36. VkResult result = vkCreateSemaphore(owner->getDevice().getLogical(), &semaphoreCI, gVulkanAllocator, &mSemaphore);
  37. assert(result == VK_SUCCESS);
  38. }
  39. VulkanSemaphore::~VulkanSemaphore()
  40. {
  41. vkDestroySemaphore(mOwner->getDevice().getLogical(), mSemaphore, gVulkanAllocator);
  42. }
  43. VulkanCmdBufferPool::VulkanCmdBufferPool(VulkanDevice& device)
  44. :mDevice(device), mNextId(1)
  45. {
  46. for (UINT32 i = 0; i < GQT_COUNT; i++)
  47. {
  48. UINT32 familyIdx = device.getQueueFamily((GpuQueueType)i);
  49. if (familyIdx == (UINT32)-1)
  50. continue;
  51. VkCommandPoolCreateInfo poolCI;
  52. poolCI.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  53. poolCI.pNext = nullptr;
  54. poolCI.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
  55. poolCI.queueFamilyIndex = familyIdx;
  56. PoolInfo& poolInfo = mPools[familyIdx];
  57. poolInfo.queueFamily = familyIdx;
  58. memset(poolInfo.buffers, 0, sizeof(poolInfo.buffers));
  59. vkCreateCommandPool(device.getLogical(), &poolCI, gVulkanAllocator, &poolInfo.pool);
  60. }
  61. }
  62. VulkanCmdBufferPool::~VulkanCmdBufferPool()
  63. {
  64. // Note: Shutdown should be the only place command buffers are destroyed at, as the system relies on the fact that
  65. // they won't be destroyed during normal operation.
  66. for(auto& entry : mPools)
  67. {
  68. PoolInfo& poolInfo = entry.second;
  69. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  70. {
  71. VulkanCmdBuffer* buffer = poolInfo.buffers[i];
  72. if (buffer == nullptr)
  73. break;
  74. bs_delete(buffer);
  75. }
  76. vkDestroyCommandPool(mDevice.getLogical(), poolInfo.pool, gVulkanAllocator);
  77. }
  78. }
  79. VulkanCmdBuffer* VulkanCmdBufferPool::getBuffer(UINT32 queueFamily, bool secondary)
  80. {
  81. auto iterFind = mPools.find(queueFamily);
  82. if (iterFind == mPools.end())
  83. return nullptr;
  84. VulkanCmdBuffer** buffers = iterFind->second.buffers;
  85. UINT32 i = 0;
  86. for(; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  87. {
  88. if (buffers[i] == nullptr)
  89. break;
  90. if(buffers[i]->mState == VulkanCmdBuffer::State::Ready)
  91. {
  92. buffers[i]->begin();
  93. return buffers[i];
  94. }
  95. }
  96. assert(i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY &&
  97. "Too many command buffers allocated. Increment BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY to a higher value. ");
  98. buffers[i] = createBuffer(queueFamily, secondary);
  99. buffers[i]->begin();
  100. return buffers[i];
  101. }
  102. VulkanCmdBuffer* VulkanCmdBufferPool::createBuffer(UINT32 queueFamily, bool secondary)
  103. {
  104. auto iterFind = mPools.find(queueFamily);
  105. if (iterFind == mPools.end())
  106. return nullptr;
  107. const PoolInfo& poolInfo = iterFind->second;
  108. return bs_new<VulkanCmdBuffer>(mDevice, mNextId++, poolInfo.pool, poolInfo.queueFamily, secondary);
  109. }
  110. /** Returns a set of pipeline stages that can are allowed to be used for the specified set of access flags. */
  111. VkPipelineStageFlags getPipelineStageFlags(VkAccessFlags accessFlags)
  112. {
  113. VkPipelineStageFlags flags = 0;
  114. if ((accessFlags & VK_ACCESS_INDIRECT_COMMAND_READ_BIT) != 0)
  115. flags |= VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT;
  116. if ((accessFlags & (VK_ACCESS_INDEX_READ_BIT | VK_ACCESS_VERTEX_ATTRIBUTE_READ_BIT)) != 0)
  117. flags |= VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
  118. if ((accessFlags & (VK_ACCESS_UNIFORM_READ_BIT | VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT)) != 0)
  119. {
  120. flags |= VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT;
  121. flags |= VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT | VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT;
  122. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  123. }
  124. if ((accessFlags & VK_ACCESS_INPUT_ATTACHMENT_READ_BIT) != 0)
  125. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
  126. if ((accessFlags & (VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT)) != 0)
  127. flags |= VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  128. if ((accessFlags & (VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT)) != 0)
  129. flags |= VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT;
  130. if ((accessFlags & (VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT)) != 0)
  131. flags |= VK_PIPELINE_STAGE_TRANSFER_BIT;
  132. if ((accessFlags & (VK_ACCESS_HOST_READ_BIT | VK_ACCESS_HOST_WRITE_BIT)) != 0)
  133. flags |= VK_PIPELINE_STAGE_HOST_BIT;
  134. if (flags == 0)
  135. flags = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  136. return flags;
  137. }
  138. template<class T>
  139. void getPipelineStageFlags(const Vector<T>& barriers, VkPipelineStageFlags& src, VkPipelineStageFlags& dst)
  140. {
  141. for(auto& entry : barriers)
  142. {
  143. src |= getPipelineStageFlags(entry.srcAccessMask);
  144. dst |= getPipelineStageFlags(entry.dstAccessMask);
  145. }
  146. if (src == 0)
  147. src = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  148. if (dst == 0)
  149. dst = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  150. }
  151. VulkanCmdBuffer::VulkanCmdBuffer(VulkanDevice& device, UINT32 id, VkCommandPool pool, UINT32 queueFamily, bool secondary)
  152. : mId(id), mQueueFamily(queueFamily), mState(State::Ready), mDevice(device), mPool(pool)
  153. , mIntraQueueSemaphore(nullptr), mInterQueueSemaphores(), mNumUsedInterQueueSemaphores(0)
  154. , mFramebuffer(nullptr), mRenderTargetWidth(0)
  155. , mRenderTargetHeight(0), mRenderTargetReadOnlyFlags(0), mRenderTargetLoadMask(RT_NONE), mGlobalQueueIdx(-1)
  156. , mViewport(0.0f, 0.0f, 1.0f, 1.0f), mScissor(0, 0, 0, 0), mStencilRef(0), mDrawOp(DOT_TRIANGLE_LIST)
  157. , mNumBoundDescriptorSets(0), mGfxPipelineRequiresBind(true), mCmpPipelineRequiresBind(true)
  158. , mViewportRequiresBind(true), mStencilRefRequiresBind(true), mScissorRequiresBind(true), mBoundParamsDirty(false)
  159. , mClearValues(), mClearMask(), mSemaphoresTemp(BS_MAX_UNIQUE_QUEUES), mVertexBuffersTemp()
  160. , mVertexBufferOffsetsTemp()
  161. {
  162. UINT32 maxBoundDescriptorSets = device.getDeviceProperties().limits.maxBoundDescriptorSets;
  163. mDescriptorSetsTemp = (VkDescriptorSet*)bs_alloc(sizeof(VkDescriptorSet) * maxBoundDescriptorSets);
  164. VkCommandBufferAllocateInfo cmdBufferAllocInfo;
  165. cmdBufferAllocInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  166. cmdBufferAllocInfo.pNext = nullptr;
  167. cmdBufferAllocInfo.commandPool = pool;
  168. cmdBufferAllocInfo.level = secondary ? VK_COMMAND_BUFFER_LEVEL_SECONDARY : VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  169. cmdBufferAllocInfo.commandBufferCount = 1;
  170. VkResult result = vkAllocateCommandBuffers(mDevice.getLogical(), &cmdBufferAllocInfo, &mCmdBuffer);
  171. assert(result == VK_SUCCESS);
  172. VkFenceCreateInfo fenceCI;
  173. fenceCI.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  174. fenceCI.pNext = nullptr;
  175. fenceCI.flags = 0;
  176. result = vkCreateFence(mDevice.getLogical(), &fenceCI, gVulkanAllocator, &mFence);
  177. assert(result == VK_SUCCESS);
  178. }
  179. VulkanCmdBuffer::~VulkanCmdBuffer()
  180. {
  181. VkDevice device = mDevice.getLogical();
  182. if(mState == State::Submitted)
  183. {
  184. // Wait 1s
  185. UINT64 waitTime = 1000 * 1000 * 1000;
  186. VkResult result = vkWaitForFences(device, 1, &mFence, true, waitTime);
  187. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  188. if (result == VK_TIMEOUT)
  189. LOGWRN("Freeing a command buffer before done executing because fence wait expired!");
  190. // Resources have been marked as used, make sure to notify them we're done with them
  191. reset();
  192. }
  193. else if(mState != State::Ready)
  194. {
  195. // Notify any resources that they are no longer bound
  196. for (auto& entry : mResources)
  197. {
  198. ResourceUseHandle& useHandle = entry.second;
  199. assert(!useHandle.used);
  200. entry.first->notifyUnbound();
  201. }
  202. for (auto& entry : mImages)
  203. {
  204. UINT32 imageInfoIdx = entry.second;
  205. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  206. ResourceUseHandle& useHandle = imageInfo.useHandle;
  207. assert(!useHandle.used);
  208. entry.first->notifyUnbound();
  209. }
  210. for (auto& entry : mBuffers)
  211. {
  212. ResourceUseHandle& useHandle = entry.second.useHandle;
  213. assert(!useHandle.used);
  214. entry.first->notifyUnbound();
  215. }
  216. }
  217. if (mIntraQueueSemaphore != nullptr)
  218. mIntraQueueSemaphore->destroy();
  219. for(UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  220. {
  221. if (mInterQueueSemaphores[i] != nullptr)
  222. mInterQueueSemaphores[i]->destroy();
  223. }
  224. vkDestroyFence(device, mFence, gVulkanAllocator);
  225. vkFreeCommandBuffers(device, mPool, 1, &mCmdBuffer);
  226. bs_free(mDescriptorSetsTemp);
  227. }
  228. UINT32 VulkanCmdBuffer::getDeviceIdx() const
  229. {
  230. return mDevice.getIndex();
  231. }
  232. void VulkanCmdBuffer::begin()
  233. {
  234. assert(mState == State::Ready);
  235. VkCommandBufferBeginInfo beginInfo;
  236. beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  237. beginInfo.pNext = nullptr;
  238. beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  239. beginInfo.pInheritanceInfo = nullptr;
  240. VkResult result = vkBeginCommandBuffer(mCmdBuffer, &beginInfo);
  241. assert(result == VK_SUCCESS);
  242. mState = State::Recording;
  243. }
  244. void VulkanCmdBuffer::end()
  245. {
  246. assert(mState == State::Recording);
  247. // If a clear is queued, execute the render pass with no additional instructions
  248. if (mClearMask)
  249. executeClearPass();
  250. VkResult result = vkEndCommandBuffer(mCmdBuffer);
  251. assert(result == VK_SUCCESS);
  252. mState = State::RecordingDone;
  253. }
  254. void VulkanCmdBuffer::beginRenderPass()
  255. {
  256. assert(mState == State::Recording);
  257. if (mFramebuffer == nullptr)
  258. {
  259. LOGWRN("Attempting to begin a render pass but no render target is bound to the command buffer.");
  260. return;
  261. }
  262. if(mClearMask != CLEAR_NONE)
  263. {
  264. // If a previous clear is queued, but it doesn't match the rendered area, need to execute a separate pass
  265. // just for it
  266. Rect2I rtArea(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  267. if (mClearArea != rtArea)
  268. executeClearPass();
  269. }
  270. executeLayoutTransitions();
  271. RenderSurfaceMask readMask = getFBReadMask();
  272. VkRenderPassBeginInfo renderPassBeginInfo;
  273. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  274. renderPassBeginInfo.pNext = nullptr;
  275. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(mRenderTargetLoadMask, readMask, mClearMask);
  276. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(mRenderTargetLoadMask, readMask, mClearMask);
  277. renderPassBeginInfo.renderArea.offset.x = 0;
  278. renderPassBeginInfo.renderArea.offset.y = 0;
  279. renderPassBeginInfo.renderArea.extent.width = mRenderTargetWidth;
  280. renderPassBeginInfo.renderArea.extent.height = mRenderTargetHeight;
  281. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  282. renderPassBeginInfo.pClearValues = mClearValues.data();
  283. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  284. mClearMask = CLEAR_NONE;
  285. mState = State::RecordingRenderPass;
  286. }
  287. void VulkanCmdBuffer::endRenderPass()
  288. {
  289. assert(mState == State::RecordingRenderPass);
  290. vkCmdEndRenderPass(mCmdBuffer);
  291. // Execute any queued events
  292. for(auto& entry : mQueuedEvents)
  293. vkCmdSetEvent(mCmdBuffer, entry->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  294. mQueuedEvents.clear();
  295. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  296. // and reset read-only state.
  297. for(auto& entry : mPassTouchedSubresourceInfos)
  298. {
  299. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  300. subresourceInfo.isShaderInput = false;
  301. subresourceInfo.isReadOnly = true;
  302. subresourceInfo.needsBarrier = false;
  303. }
  304. for (auto& entry : mBuffers)
  305. entry.second.needsBarrier = false;
  306. mPassTouchedSubresourceInfos.clear();
  307. updateFinalLayouts();
  308. mState = State::Recording;
  309. // In case the same GPU params from last pass get used, this makes sure the states we reset above, get re-applied
  310. mBoundParamsDirty = true;
  311. }
  312. void VulkanCmdBuffer::allocateSemaphores(VkSemaphore* semaphores)
  313. {
  314. if (mIntraQueueSemaphore != nullptr)
  315. mIntraQueueSemaphore->destroy();
  316. mIntraQueueSemaphore = mDevice.getResourceManager().create<VulkanSemaphore>();
  317. semaphores[0] = mIntraQueueSemaphore->getHandle();
  318. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  319. {
  320. if (mInterQueueSemaphores[i] != nullptr)
  321. mInterQueueSemaphores[i]->destroy();
  322. mInterQueueSemaphores[i] = mDevice.getResourceManager().create<VulkanSemaphore>();
  323. semaphores[i + 1] = mInterQueueSemaphores[i]->getHandle();
  324. }
  325. mNumUsedInterQueueSemaphores = 0;
  326. }
  327. VulkanSemaphore* VulkanCmdBuffer::requestInterQueueSemaphore() const
  328. {
  329. if (mNumUsedInterQueueSemaphores >= BS_MAX_VULKAN_CB_DEPENDENCIES)
  330. return nullptr;
  331. return mInterQueueSemaphores[mNumUsedInterQueueSemaphores++];
  332. }
  333. void VulkanCmdBuffer::submit(VulkanQueue* queue, UINT32 queueIdx, UINT32 syncMask)
  334. {
  335. assert(isReadyForSubmit());
  336. // Make sure to reset the CB fence before we submit it
  337. VkResult result = vkResetFences(mDevice.getLogical(), 1, &mFence);
  338. assert(result == VK_SUCCESS);
  339. // If there are any query resets needed, execute those first
  340. VulkanDevice& device = queue->getDevice();
  341. if(!mQueuedQueryResets.empty())
  342. {
  343. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  344. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  345. for (auto& entry : mQueuedQueryResets)
  346. entry->reset(vkCmdBuffer);
  347. cmdBuffer->end();
  348. queue->queueSubmit(cmdBuffer, nullptr, 0);
  349. mQueuedQueryResets.clear();
  350. }
  351. // Issue pipeline barriers for queue transitions (need to happen on original queue first, then on new queue)
  352. for (auto& entry : mBuffers)
  353. {
  354. VulkanBuffer* resource = static_cast<VulkanBuffer*>(entry.first);
  355. if (!resource->isExclusive())
  356. continue;
  357. UINT32 currentQueueFamily = resource->getQueueFamily();
  358. if (currentQueueFamily != (UINT32)-1 && currentQueueFamily != mQueueFamily)
  359. {
  360. Vector<VkBufferMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].bufferBarriers;
  361. barriers.push_back(VkBufferMemoryBarrier());
  362. VkBufferMemoryBarrier& barrier = barriers.back();
  363. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  364. barrier.pNext = nullptr;
  365. barrier.srcAccessMask = entry.second.accessFlags;
  366. barrier.dstAccessMask = entry.second.accessFlags;
  367. barrier.srcQueueFamilyIndex = currentQueueFamily;
  368. barrier.dstQueueFamilyIndex = mQueueFamily;
  369. barrier.buffer = resource->getHandle();
  370. barrier.offset = 0;
  371. barrier.size = VK_WHOLE_SIZE;
  372. }
  373. }
  374. // For images issue queue transitions, as above. Also issue layout transitions to their inital layouts.
  375. Vector<VkImageMemoryBarrier>& localBarriers = mTransitionInfoTemp[mQueueFamily].imageBarriers;
  376. for (auto& entry : mImages)
  377. {
  378. VulkanImage* resource = static_cast<VulkanImage*>(entry.first);
  379. ImageInfo& imageInfo = mImageInfos[entry.second];
  380. UINT32 currentQueueFamily = resource->getQueueFamily();
  381. bool queueMismatch = resource->isExclusive() && currentQueueFamily != (UINT32)-1
  382. && currentQueueFamily != mQueueFamily;
  383. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  384. if (queueMismatch)
  385. {
  386. Vector<VkImageMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].imageBarriers;
  387. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  388. {
  389. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  390. UINT32 startIdx = (UINT32)barriers.size();
  391. resource->getBarriers(subresourceInfo.range, barriers);
  392. for(UINT32 j = startIdx; j < (UINT32)barriers.size(); j++)
  393. {
  394. VkImageMemoryBarrier& barrier = barriers[j];
  395. barrier.dstAccessMask = resource->getAccessFlags(barrier.oldLayout);
  396. barrier.newLayout = barrier.oldLayout;
  397. barrier.srcQueueFamilyIndex = currentQueueFamily;
  398. barrier.dstQueueFamilyIndex = mQueueFamily;
  399. }
  400. }
  401. }
  402. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  403. {
  404. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  405. VkImageLayout initialLayout = subresourceInfo.initialLayout;
  406. if (initialLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  407. continue;
  408. const VkImageSubresourceRange& range = subresourceInfo.range;
  409. UINT32 mipEnd = range.baseMipLevel + range.levelCount;
  410. UINT32 faceEnd = range.baseArrayLayer + range.layerCount;
  411. bool layoutMismatch = false;
  412. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  413. {
  414. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  415. {
  416. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  417. if(subresource->getLayout() != initialLayout)
  418. {
  419. layoutMismatch = true;
  420. break;
  421. }
  422. }
  423. if (layoutMismatch)
  424. break;
  425. }
  426. if(layoutMismatch)
  427. {
  428. UINT32 startIdx = (UINT32)localBarriers.size();
  429. resource->getBarriers(subresourceInfo.range, localBarriers);
  430. for (UINT32 j = startIdx; j < (UINT32)localBarriers.size(); j++)
  431. {
  432. VkImageMemoryBarrier& barrier = localBarriers[j];
  433. barrier.dstAccessMask = resource->getAccessFlags(initialLayout, subresourceInfo.isInitialReadOnly);
  434. barrier.newLayout = initialLayout;
  435. }
  436. }
  437. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  438. {
  439. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  440. {
  441. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  442. subresource->setLayout(subresourceInfo.finalLayout);
  443. }
  444. }
  445. }
  446. }
  447. for (auto& entry : mTransitionInfoTemp)
  448. {
  449. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  450. if (empty)
  451. continue;
  452. UINT32 entryQueueFamily = entry.first;
  453. // No queue transition needed for entries on this queue (this entry is most likely an image layout transition)
  454. if (entryQueueFamily == (UINT32)-1 || entryQueueFamily == mQueueFamily)
  455. continue;
  456. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(entryQueueFamily, false);
  457. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  458. TransitionInfo& barriers = entry.second;
  459. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  460. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  461. VkPipelineStageFlags srcStage = 0;
  462. VkPipelineStageFlags dstStage = 0;
  463. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  464. vkCmdPipelineBarrier(vkCmdBuffer,
  465. srcStage, dstStage,
  466. 0, 0, nullptr,
  467. numBufferBarriers, barriers.bufferBarriers.data(),
  468. numImgBarriers, barriers.imageBarriers.data());
  469. // Find an appropriate queue to execute on
  470. UINT32 otherQueueIdx = 0;
  471. VulkanQueue* otherQueue = nullptr;
  472. GpuQueueType otherQueueType = GQT_GRAPHICS;
  473. for (UINT32 i = 0; i < GQT_COUNT; i++)
  474. {
  475. otherQueueType = (GpuQueueType)i;
  476. if (device.getQueueFamily(otherQueueType) != entryQueueFamily)
  477. continue;
  478. UINT32 numQueues = device.getNumQueues(otherQueueType);
  479. for (UINT32 j = 0; j < numQueues; j++)
  480. {
  481. // Try to find a queue not currently executing
  482. VulkanQueue* curQueue = device.getQueue(otherQueueType, j);
  483. if (!curQueue->isExecuting())
  484. {
  485. otherQueue = curQueue;
  486. otherQueueIdx = j;
  487. }
  488. }
  489. // Can't find empty one, use the first one then
  490. if (otherQueue == nullptr)
  491. {
  492. otherQueue = device.getQueue(otherQueueType, 0);
  493. otherQueueIdx = 0;
  494. }
  495. break;
  496. }
  497. syncMask |= CommandSyncMask::getGlobalQueueMask(otherQueueType, otherQueueIdx);
  498. cmdBuffer->end();
  499. // Note: If I switch back to doing layout transitions here, I need to wait on present semaphore
  500. otherQueue->submit(cmdBuffer, nullptr, 0);
  501. }
  502. UINT32 deviceIdx = device.getIndex();
  503. VulkanCommandBufferManager& cbm = static_cast<VulkanCommandBufferManager&>(CommandBufferManager::instance());
  504. UINT32 numSemaphores;
  505. cbm.getSyncSemaphores(deviceIdx, syncMask, mSemaphoresTemp.data(), numSemaphores);
  506. // Wait on present (i.e. until the back buffer becomes available) for any swap chains
  507. for(auto& entry : mSwapChains)
  508. {
  509. const SwapChainSurface& surface = entry->getBackBuffer();
  510. if (surface.needsWait)
  511. {
  512. VulkanSemaphore* semaphore = entry->getBackBuffer().sync;
  513. if (numSemaphores >= (UINT32)mSemaphoresTemp.size())
  514. mSemaphoresTemp.push_back(semaphore);
  515. else
  516. mSemaphoresTemp[numSemaphores] = semaphore;
  517. numSemaphores++;
  518. entry->notifyBackBufferWaitIssued();
  519. }
  520. }
  521. // Issue second part of transition pipeline barriers (on this queue)
  522. for (auto& entry : mTransitionInfoTemp)
  523. {
  524. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  525. if (empty)
  526. continue;
  527. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  528. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  529. TransitionInfo& barriers = entry.second;
  530. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  531. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  532. VkPipelineStageFlags srcStage = 0;
  533. VkPipelineStageFlags dstStage = 0;
  534. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  535. vkCmdPipelineBarrier(vkCmdBuffer,
  536. srcStage, dstStage,
  537. 0, 0, nullptr,
  538. numBufferBarriers, barriers.bufferBarriers.data(),
  539. numImgBarriers, barriers.imageBarriers.data());
  540. cmdBuffer->end();
  541. queue->queueSubmit(cmdBuffer, mSemaphoresTemp.data(), numSemaphores);
  542. numSemaphores = 0; // Semaphores are only needed the first time, since we're adding the buffers on the same queue
  543. }
  544. queue->queueSubmit(this, mSemaphoresTemp.data(), numSemaphores);
  545. queue->submitQueued();
  546. mGlobalQueueIdx = CommandSyncMask::getGlobalQueueIdx(queue->getType(), queueIdx);
  547. for (auto& entry : mResources)
  548. {
  549. ResourceUseHandle& useHandle = entry.second;
  550. assert(!useHandle.used);
  551. useHandle.used = true;
  552. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  553. }
  554. for (auto& entry : mImages)
  555. {
  556. UINT32 imageInfoIdx = entry.second;
  557. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  558. ResourceUseHandle& useHandle = imageInfo.useHandle;
  559. assert(!useHandle.used);
  560. useHandle.used = true;
  561. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  562. }
  563. for (auto& entry : mBuffers)
  564. {
  565. ResourceUseHandle& useHandle = entry.second.useHandle;
  566. assert(!useHandle.used);
  567. useHandle.used = true;
  568. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  569. }
  570. // Note: Uncomment for debugging only, prevents any device concurrency issues.
  571. // vkQueueWaitIdle(queue->getHandle());
  572. // Clear vectors but don't clear the actual map, as we want to re-use the memory since we expect queue family
  573. // indices to be the same
  574. for (auto& entry : mTransitionInfoTemp)
  575. {
  576. entry.second.imageBarriers.clear();
  577. entry.second.bufferBarriers.clear();
  578. }
  579. mGraphicsPipeline = nullptr;
  580. mComputePipeline = nullptr;
  581. mGfxPipelineRequiresBind = true;
  582. mCmpPipelineRequiresBind = true;
  583. mFramebuffer = nullptr;
  584. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  585. mQueuedLayoutTransitions.clear();
  586. mBoundParams = nullptr;
  587. mSwapChains.clear();
  588. }
  589. bool VulkanCmdBuffer::checkFenceStatus(bool block) const
  590. {
  591. VkResult result = vkWaitForFences(mDevice.getLogical(), 1, &mFence, true, block ? 1'000'000'000 : 0);
  592. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  593. return result == VK_SUCCESS;
  594. }
  595. void VulkanCmdBuffer::reset()
  596. {
  597. bool wasSubmitted = mState == State::Submitted;
  598. mState = State::Ready;
  599. vkResetCommandBuffer(mCmdBuffer, VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT); // Note: Maybe better not to release resources?
  600. if (wasSubmitted)
  601. {
  602. for (auto& entry : mResources)
  603. {
  604. ResourceUseHandle& useHandle = entry.second;
  605. assert(useHandle.used);
  606. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  607. }
  608. for (auto& entry : mImages)
  609. {
  610. UINT32 imageInfoIdx = entry.second;
  611. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  612. ResourceUseHandle& useHandle = imageInfo.useHandle;
  613. assert(useHandle.used);
  614. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  615. }
  616. for (auto& entry : mBuffers)
  617. {
  618. ResourceUseHandle& useHandle = entry.second.useHandle;
  619. assert(useHandle.used);
  620. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  621. }
  622. }
  623. else
  624. {
  625. for (auto& entry : mResources)
  626. entry.first->notifyUnbound();
  627. for (auto& entry : mImages)
  628. entry.first->notifyUnbound();
  629. for (auto& entry : mBuffers)
  630. entry.first->notifyUnbound();
  631. }
  632. mResources.clear();
  633. mImages.clear();
  634. mBuffers.clear();
  635. mOcclusionQueries.clear();
  636. mTimerQueries.clear();
  637. mImageInfos.clear();
  638. mSubresourceInfoStorage.clear();
  639. mPassTouchedSubresourceInfos.clear();
  640. }
  641. void VulkanCmdBuffer::setRenderTarget(const SPtr<RenderTarget>& rt, UINT32 readOnlyFlags,
  642. RenderSurfaceMask loadMask)
  643. {
  644. assert(mState != State::Submitted);
  645. VulkanFramebuffer* newFB;
  646. if(rt != nullptr)
  647. {
  648. if (rt->getProperties().isWindow)
  649. {
  650. #if BS_PLATFORM == BS_PLATFORM_WIN32
  651. Win32RenderWindow* window = static_cast<Win32RenderWindow*>(rt.get());
  652. #elif BS_PLATFORM == BS_PLATFORM_LINUX
  653. LinuxRenderWindow* window = static_cast<LinuxRenderWindow*>(rt.get());
  654. #endif
  655. window->acquireBackBuffer();
  656. VulkanSwapChain* swapChain;
  657. rt->getCustomAttribute("SC", &swapChain);
  658. mSwapChains.insert(swapChain);
  659. }
  660. rt->getCustomAttribute("FB", &newFB);
  661. }
  662. else
  663. {
  664. newFB = nullptr;
  665. }
  666. // Warn if invalid load mask
  667. if (loadMask.isSet(RT_DEPTH) && !loadMask.isSet(RT_STENCIL))
  668. {
  669. LOGWRN("setRenderTarget() invalid load mask, depth enabled but stencil disabled. This is not supported. Both \
  670. will be loaded.");
  671. loadMask.set(RT_STENCIL);
  672. }
  673. if (!loadMask.isSet(RT_DEPTH) && loadMask.isSet(RT_STENCIL))
  674. {
  675. LOGWRN("setRenderTarget() invalid load mask, stencil enabled but depth disabled. This is not supported. Both \
  676. will be loaded.");
  677. loadMask.set(RT_DEPTH);
  678. }
  679. if (mFramebuffer == newFB && mRenderTargetReadOnlyFlags == readOnlyFlags && mRenderTargetLoadMask == loadMask)
  680. return;
  681. if (isInRenderPass())
  682. endRenderPass();
  683. else
  684. {
  685. // If a clear is queued for previous FB, execute the render pass with no additional instructions
  686. if (mClearMask)
  687. executeClearPass();
  688. }
  689. // Reset isFBAttachment flags for subresources from the old framebuffer
  690. if(mFramebuffer != nullptr)
  691. {
  692. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  693. for(UINT32 i = 0; i < numColorAttachments; i++)
  694. {
  695. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  696. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  697. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  698. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  699. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  700. {
  701. ImageSubresourceInfo& entry = subresourceInfos[j];
  702. entry.isFBAttachment = false;
  703. }
  704. }
  705. if(mFramebuffer->hasDepthAttachment())
  706. {
  707. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  708. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  709. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  710. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  711. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  712. {
  713. ImageSubresourceInfo& entry = subresourceInfos[j];
  714. entry.isFBAttachment = false;
  715. }
  716. }
  717. }
  718. if(newFB == nullptr)
  719. {
  720. mFramebuffer = nullptr;
  721. mRenderTargetWidth = 0;
  722. mRenderTargetHeight = 0;
  723. mRenderTargetReadOnlyFlags = 0;
  724. mRenderTargetLoadMask = RT_NONE;
  725. }
  726. else
  727. {
  728. mFramebuffer = newFB;
  729. mRenderTargetWidth = rt->getProperties().width;
  730. mRenderTargetHeight = rt->getProperties().height;
  731. mRenderTargetReadOnlyFlags = readOnlyFlags;
  732. mRenderTargetLoadMask = loadMask;
  733. }
  734. // Re-set the params as they will need to be re-bound
  735. setGpuParams(mBoundParams);
  736. if (mFramebuffer != nullptr)
  737. registerResource(mFramebuffer, loadMask, readOnlyFlags);
  738. mGfxPipelineRequiresBind = true;
  739. }
  740. void VulkanCmdBuffer::clearViewport(const Rect2I& area, UINT32 buffers, const Color& color, float depth, UINT16 stencil,
  741. UINT8 targetMask)
  742. {
  743. if (buffers == 0 || mFramebuffer == nullptr)
  744. return;
  745. // Add clear command if currently in render pass
  746. if (isInRenderPass())
  747. {
  748. VkClearAttachment attachments[BS_MAX_MULTIPLE_RENDER_TARGETS + 1];
  749. UINT32 baseLayer = 0;
  750. UINT32 attachmentIdx = 0;
  751. if ((buffers & FBT_COLOR) != 0)
  752. {
  753. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  754. for (UINT32 i = 0; i < numColorAttachments; i++)
  755. {
  756. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  757. if (((1 << attachment.index) & targetMask) == 0)
  758. continue;
  759. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  760. attachments[attachmentIdx].colorAttachment = i;
  761. VkClearColorValue& colorValue = attachments[attachmentIdx].clearValue.color;
  762. colorValue.float32[0] = color.r;
  763. colorValue.float32[1] = color.g;
  764. colorValue.float32[2] = color.b;
  765. colorValue.float32[3] = color.a;
  766. UINT32 curBaseLayer = attachment.baseLayer;
  767. if (attachmentIdx == 0)
  768. baseLayer = curBaseLayer;
  769. else
  770. {
  771. if (baseLayer != curBaseLayer)
  772. {
  773. // Note: This could be supported relatively easily: we would need to issue multiple separate
  774. // clear commands for such framebuffers.
  775. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  776. "starting layers. This is currently not supported.");
  777. }
  778. }
  779. attachmentIdx++;
  780. }
  781. }
  782. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  783. {
  784. if (mFramebuffer->hasDepthAttachment())
  785. {
  786. attachments[attachmentIdx].aspectMask = 0;
  787. if ((buffers & FBT_DEPTH) != 0)
  788. {
  789. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_DEPTH_BIT;
  790. attachments[attachmentIdx].clearValue.depthStencil.depth = depth;
  791. }
  792. if ((buffers & FBT_STENCIL) != 0)
  793. {
  794. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_STENCIL_BIT;
  795. attachments[attachmentIdx].clearValue.depthStencil.stencil = stencil;
  796. }
  797. attachments[attachmentIdx].colorAttachment = 0;
  798. UINT32 curBaseLayer = mFramebuffer->getDepthStencilAttachment().baseLayer;
  799. if (attachmentIdx == 0)
  800. baseLayer = curBaseLayer;
  801. else
  802. {
  803. if (baseLayer != curBaseLayer)
  804. {
  805. // Note: This could be supported relatively easily: we would need to issue multiple separate
  806. // clear commands for such framebuffers.
  807. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  808. "starting layers. This is currently not supported.");
  809. }
  810. }
  811. attachmentIdx++;
  812. }
  813. }
  814. UINT32 numAttachments = attachmentIdx;
  815. if (numAttachments == 0)
  816. return;
  817. VkClearRect clearRect;
  818. clearRect.baseArrayLayer = baseLayer;
  819. clearRect.layerCount = mFramebuffer->getNumLayers();
  820. clearRect.rect.offset.x = area.x;
  821. clearRect.rect.offset.y = area.y;
  822. clearRect.rect.extent.width = area.width;
  823. clearRect.rect.extent.height = area.height;
  824. vkCmdClearAttachments(mCmdBuffer, numAttachments, attachments, 1, &clearRect);
  825. }
  826. // Otherwise we use a render pass that performs a clear on begin
  827. else
  828. {
  829. ClearMask clearMask;
  830. std::array<VkClearValue, BS_MAX_MULTIPLE_RENDER_TARGETS + 1> clearValues = mClearValues;
  831. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  832. if ((buffers & FBT_COLOR) != 0)
  833. {
  834. for (UINT32 i = 0; i < numColorAttachments; i++)
  835. {
  836. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  837. if (((1 << attachment.index) & targetMask) == 0)
  838. continue;
  839. clearMask |= (ClearMaskBits)(1 << attachment.index);
  840. VkClearColorValue& colorValue = clearValues[i].color;
  841. colorValue.float32[0] = color.r;
  842. colorValue.float32[1] = color.g;
  843. colorValue.float32[2] = color.b;
  844. colorValue.float32[3] = color.a;
  845. }
  846. }
  847. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  848. {
  849. if (mFramebuffer->hasDepthAttachment())
  850. {
  851. UINT32 depthAttachmentIdx = numColorAttachments;
  852. if ((buffers & FBT_DEPTH) != 0)
  853. {
  854. clearValues[depthAttachmentIdx].depthStencil.depth = depth;
  855. clearMask |= CLEAR_DEPTH;
  856. }
  857. if ((buffers & FBT_STENCIL) != 0)
  858. {
  859. clearValues[depthAttachmentIdx].depthStencil.stencil = stencil;
  860. clearMask |= CLEAR_STENCIL;
  861. }
  862. }
  863. }
  864. if (!clearMask)
  865. return;
  866. // Some previous clear operation is already queued, execute it first
  867. bool previousClearNeedsToFinish = (mClearMask & clearMask) != CLEAR_NONE;
  868. if(previousClearNeedsToFinish)
  869. executeClearPass();
  870. mClearMask |= clearMask;
  871. mClearValues = clearValues;
  872. mClearArea = area;
  873. }
  874. }
  875. void VulkanCmdBuffer::clearRenderTarget(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  876. {
  877. Rect2I area(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  878. clearViewport(area, buffers, color, depth, stencil, targetMask);
  879. }
  880. void VulkanCmdBuffer::clearViewport(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  881. {
  882. Rect2I area;
  883. area.x = (UINT32)(mViewport.x * mRenderTargetWidth);
  884. area.y = (UINT32)(mViewport.y * mRenderTargetHeight);
  885. area.width = (UINT32)(mViewport.width * mRenderTargetWidth);
  886. area.height = (UINT32)(mViewport.height * mRenderTargetHeight);
  887. clearViewport(area, buffers, color, depth, stencil, targetMask);
  888. }
  889. void VulkanCmdBuffer::setPipelineState(const SPtr<GraphicsPipelineState>& state)
  890. {
  891. if (mGraphicsPipeline == state)
  892. return;
  893. mGraphicsPipeline = std::static_pointer_cast<VulkanGraphicsPipelineState>(state);
  894. mGfxPipelineRequiresBind = true;
  895. }
  896. void VulkanCmdBuffer::setPipelineState(const SPtr<ComputePipelineState>& state)
  897. {
  898. if (mComputePipeline == state)
  899. return;
  900. mComputePipeline = std::static_pointer_cast<VulkanComputePipelineState>(state);
  901. mCmpPipelineRequiresBind = true;
  902. }
  903. void VulkanCmdBuffer::setGpuParams(const SPtr<GpuParams>& gpuParams)
  904. {
  905. // Note: We keep an internal reference to GPU params even though we shouldn't keep a reference to a core thread
  906. // object. But it should be fine since we expect the resource to be externally synchronized so it should never
  907. // be allowed to go out of scope on a non-core thread anyway.
  908. mBoundParams = std::static_pointer_cast<VulkanGpuParams>(gpuParams);
  909. if (mBoundParams != nullptr)
  910. mBoundParamsDirty = true;
  911. else
  912. {
  913. mNumBoundDescriptorSets = 0;
  914. mBoundParamsDirty = false;
  915. }
  916. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  917. }
  918. void VulkanCmdBuffer::setViewport(const Rect2& area)
  919. {
  920. if (mViewport == area)
  921. return;
  922. mViewport = area;
  923. mViewportRequiresBind = true;
  924. }
  925. void VulkanCmdBuffer::setScissorRect(const Rect2I& value)
  926. {
  927. if (mScissor == value)
  928. return;
  929. mScissor = value;
  930. mScissorRequiresBind = true;
  931. }
  932. void VulkanCmdBuffer::setStencilRef(UINT32 value)
  933. {
  934. if (mStencilRef == value)
  935. return;
  936. mStencilRef = value;
  937. mStencilRefRequiresBind = true;
  938. }
  939. void VulkanCmdBuffer::setDrawOp(DrawOperationType drawOp)
  940. {
  941. if (mDrawOp == drawOp)
  942. return;
  943. mDrawOp = drawOp;
  944. mGfxPipelineRequiresBind = true;
  945. }
  946. void VulkanCmdBuffer::setVertexBuffers(UINT32 index, SPtr<VertexBuffer>* buffers, UINT32 numBuffers)
  947. {
  948. if (numBuffers == 0)
  949. return;
  950. for(UINT32 i = 0; i < numBuffers; i++)
  951. {
  952. VulkanVertexBuffer* vertexBuffer = static_cast<VulkanVertexBuffer*>(buffers[i].get());
  953. if (vertexBuffer != nullptr)
  954. {
  955. VulkanBuffer* resource = vertexBuffer->getResource(mDevice.getIndex());
  956. if (resource != nullptr)
  957. {
  958. mVertexBuffersTemp[i] = resource->getHandle();
  959. registerResource(resource, VulkanUseFlag::Read);
  960. }
  961. else
  962. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  963. }
  964. else
  965. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  966. }
  967. vkCmdBindVertexBuffers(mCmdBuffer, index, numBuffers, mVertexBuffersTemp, mVertexBufferOffsetsTemp);
  968. }
  969. void VulkanCmdBuffer::setIndexBuffer(const SPtr<IndexBuffer>& buffer)
  970. {
  971. VulkanIndexBuffer* indexBuffer = static_cast<VulkanIndexBuffer*>(buffer.get());
  972. VkBuffer vkBuffer = VK_NULL_HANDLE;
  973. VkIndexType indexType = VK_INDEX_TYPE_UINT32;
  974. if (indexBuffer != nullptr)
  975. {
  976. VulkanBuffer* resource = indexBuffer->getResource(mDevice.getIndex());
  977. if (resource != nullptr)
  978. {
  979. vkBuffer = resource->getHandle();
  980. indexType = VulkanUtility::getIndexType(buffer->getProperties().getType());
  981. registerResource(resource, VulkanUseFlag::Read);
  982. }
  983. }
  984. vkCmdBindIndexBuffer(mCmdBuffer, vkBuffer, 0, indexType);
  985. }
  986. void VulkanCmdBuffer::setVertexDeclaration(const SPtr<VertexDeclaration>& decl)
  987. {
  988. if (mVertexDecl == decl)
  989. return;
  990. mVertexDecl = decl;
  991. mGfxPipelineRequiresBind = true;
  992. }
  993. bool VulkanCmdBuffer::isReadyForRender()
  994. {
  995. if (mGraphicsPipeline == nullptr)
  996. return false;
  997. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  998. if (inputDecl == nullptr)
  999. return false;
  1000. return mFramebuffer != nullptr && mVertexDecl != nullptr;
  1001. }
  1002. bool VulkanCmdBuffer::bindGraphicsPipeline()
  1003. {
  1004. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  1005. SPtr<VulkanVertexInput> vertexInput = VulkanVertexInputManager::instance().getVertexInfo(mVertexDecl, inputDecl);
  1006. VulkanPipeline* pipeline = mGraphicsPipeline->getPipeline(mDevice.getIndex(), mFramebuffer,
  1007. mRenderTargetReadOnlyFlags, mDrawOp, vertexInput);
  1008. if (pipeline == nullptr)
  1009. return false;
  1010. // Check that pipeline matches the read-only state of any framebuffer attachments
  1011. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1012. for (UINT32 i = 0; i < numColorAttachments; i++)
  1013. {
  1014. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1015. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1016. fbAttachment.surface.mipLevel);
  1017. if (subresourceInfo.isShaderInput && !pipeline->isColorReadOnly(i))
  1018. {
  1019. LOGWRN("Framebuffer attachment also used as a shader input, but color writes aren't disabled. This will"
  1020. " result in undefined behavior.");
  1021. }
  1022. }
  1023. if (mFramebuffer->hasDepthAttachment())
  1024. {
  1025. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1026. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1027. fbAttachment.surface.mipLevel);
  1028. if (subresourceInfo.isShaderInput && !pipeline->isDepthReadOnly())
  1029. {
  1030. LOGWRN("Framebuffer attachment also used as a shader input, but depth/stencil writes aren't disabled. "
  1031. "This will result in undefined behavior.");
  1032. }
  1033. }
  1034. mGraphicsPipeline->registerPipelineResources(this);
  1035. registerResource(pipeline, VulkanUseFlag::Read);
  1036. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline->getHandle());
  1037. bindDynamicStates(true);
  1038. mGfxPipelineRequiresBind = false;
  1039. return true;
  1040. }
  1041. void VulkanCmdBuffer::bindDynamicStates(bool forceAll)
  1042. {
  1043. if (mViewportRequiresBind || forceAll)
  1044. {
  1045. VkViewport viewport;
  1046. viewport.x = mViewport.x * mRenderTargetWidth;
  1047. viewport.y = mViewport.y * mRenderTargetHeight;
  1048. viewport.width = mViewport.width * mRenderTargetWidth;
  1049. viewport.height = mViewport.height * mRenderTargetHeight;
  1050. viewport.minDepth = 0.0f;
  1051. viewport.maxDepth = 1.0f;
  1052. vkCmdSetViewport(mCmdBuffer, 0, 1, &viewport);
  1053. mViewportRequiresBind = false;
  1054. }
  1055. if(mStencilRefRequiresBind || forceAll)
  1056. {
  1057. vkCmdSetStencilReference(mCmdBuffer, VK_STENCIL_FRONT_AND_BACK, mStencilRef);
  1058. mStencilRefRequiresBind = false;
  1059. }
  1060. if(mScissorRequiresBind || forceAll)
  1061. {
  1062. VkRect2D scissorRect;
  1063. if(mGraphicsPipeline->isScissorEnabled())
  1064. {
  1065. scissorRect.offset.x = mScissor.x;
  1066. scissorRect.offset.y = mScissor.y;
  1067. scissorRect.extent.width = mScissor.width;
  1068. scissorRect.extent.height = mScissor.height;
  1069. }
  1070. else
  1071. {
  1072. scissorRect.offset.x = 0;
  1073. scissorRect.offset.y = 0;
  1074. scissorRect.extent.width = mRenderTargetWidth;
  1075. scissorRect.extent.height = mRenderTargetHeight;
  1076. }
  1077. vkCmdSetScissor(mCmdBuffer, 0, 1, &scissorRect);
  1078. mScissorRequiresBind = false;
  1079. }
  1080. }
  1081. void VulkanCmdBuffer::bindGpuParams()
  1082. {
  1083. if (mBoundParamsDirty)
  1084. {
  1085. if (mBoundParams != nullptr)
  1086. {
  1087. mNumBoundDescriptorSets = mBoundParams->getNumSets();
  1088. mBoundParams->prepareForBind(*this, mDescriptorSetsTemp);
  1089. }
  1090. else
  1091. mNumBoundDescriptorSets = 0;
  1092. mBoundParamsDirty = false;
  1093. }
  1094. else
  1095. {
  1096. mNumBoundDescriptorSets = 0;
  1097. }
  1098. }
  1099. void VulkanCmdBuffer::executeLayoutTransitions()
  1100. {
  1101. auto createLayoutTransitionBarrier = [&](VulkanImage* image, ImageInfo& imageInfo)
  1102. {
  1103. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1104. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1105. {
  1106. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  1107. if (!subresourceInfo.hasTransitioned || subresourceInfo.currentLayout == subresourceInfo.requiredLayout)
  1108. continue;
  1109. mLayoutTransitionBarriersTemp.push_back(VkImageMemoryBarrier());
  1110. VkImageMemoryBarrier& barrier = mLayoutTransitionBarriersTemp.back();
  1111. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1112. barrier.pNext = nullptr;
  1113. barrier.srcAccessMask = image->getAccessFlags(subresourceInfo.currentLayout);
  1114. barrier.dstAccessMask = image->getAccessFlags(subresourceInfo.requiredLayout, subresourceInfo.isReadOnly);
  1115. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1116. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1117. barrier.oldLayout = subresourceInfo.currentLayout;
  1118. barrier.newLayout = subresourceInfo.requiredLayout;
  1119. barrier.image = image->getHandle();
  1120. barrier.subresourceRange = subresourceInfo.range;
  1121. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1122. subresourceInfo.isReadOnly = true;
  1123. subresourceInfo.hasTransitioned = true;
  1124. }
  1125. };
  1126. // Note: These layout transitions will contain transitions for offscreen framebuffer attachments (while they
  1127. // transition to shader read-only layout). This can be avoided, since they're immediately used by the render pass
  1128. // as color attachments, making the layout change redundant.
  1129. for (auto& entry : mQueuedLayoutTransitions)
  1130. {
  1131. UINT32 imageInfoIdx = entry.second;
  1132. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1133. createLayoutTransitionBarrier(entry.first, imageInfo);
  1134. }
  1135. VkPipelineStageFlags srcStage = 0;
  1136. VkPipelineStageFlags dstStage = 0;
  1137. getPipelineStageFlags(mLayoutTransitionBarriersTemp, srcStage, dstStage);
  1138. vkCmdPipelineBarrier(mCmdBuffer,
  1139. srcStage, dstStage,
  1140. 0, 0, nullptr,
  1141. 0, nullptr,
  1142. (UINT32)mLayoutTransitionBarriersTemp.size(), mLayoutTransitionBarriersTemp.data());
  1143. mQueuedLayoutTransitions.clear();
  1144. mLayoutTransitionBarriersTemp.clear();
  1145. }
  1146. void VulkanCmdBuffer::updateFinalLayouts()
  1147. {
  1148. if (mFramebuffer == nullptr)
  1149. return;
  1150. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1151. for (UINT32 i = 0; i < numColorAttachments; i++)
  1152. {
  1153. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1154. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1155. fbAttachment.surface.mipLevel);
  1156. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1157. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1158. subresourceInfo.hasTransitioned = true;
  1159. }
  1160. if (mFramebuffer->hasDepthAttachment())
  1161. {
  1162. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1163. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1164. fbAttachment.surface.mipLevel);
  1165. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1166. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1167. subresourceInfo.hasTransitioned = true;
  1168. }
  1169. }
  1170. void VulkanCmdBuffer::executeClearPass()
  1171. {
  1172. assert(mState == State::Recording);
  1173. executeLayoutTransitions();
  1174. VkRenderPassBeginInfo renderPassBeginInfo;
  1175. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  1176. renderPassBeginInfo.pNext = nullptr;
  1177. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(RT_NONE, RT_NONE, mClearMask);
  1178. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(RT_NONE, RT_NONE, mClearMask);
  1179. renderPassBeginInfo.renderArea.offset.x = mClearArea.x;
  1180. renderPassBeginInfo.renderArea.offset.y = mClearArea.y;
  1181. renderPassBeginInfo.renderArea.extent.width = mClearArea.width;
  1182. renderPassBeginInfo.renderArea.extent.height = mClearArea.height;
  1183. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  1184. renderPassBeginInfo.pClearValues = mClearValues.data();
  1185. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  1186. vkCmdEndRenderPass(mCmdBuffer);
  1187. updateFinalLayouts();
  1188. mClearMask = CLEAR_NONE;
  1189. }
  1190. void VulkanCmdBuffer::draw(UINT32 vertexOffset, UINT32 vertexCount, UINT32 instanceCount)
  1191. {
  1192. if (!isReadyForRender())
  1193. return;
  1194. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1195. bindGpuParams();
  1196. if (!isInRenderPass())
  1197. beginRenderPass();
  1198. if (mGfxPipelineRequiresBind)
  1199. {
  1200. if (!bindGraphicsPipeline())
  1201. return;
  1202. }
  1203. else
  1204. bindDynamicStates(false);
  1205. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1206. {
  1207. if (mNumBoundDescriptorSets > 0)
  1208. {
  1209. UINT32 deviceIdx = mDevice.getIndex();
  1210. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1211. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1212. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1213. }
  1214. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1215. }
  1216. if (instanceCount <= 0)
  1217. instanceCount = 1;
  1218. vkCmdDraw(mCmdBuffer, vertexCount, instanceCount, vertexOffset, 0);
  1219. }
  1220. void VulkanCmdBuffer::drawIndexed(UINT32 startIndex, UINT32 indexCount, UINT32 vertexOffset, UINT32 instanceCount)
  1221. {
  1222. if (!isReadyForRender())
  1223. return;
  1224. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1225. bindGpuParams();
  1226. if (!isInRenderPass())
  1227. beginRenderPass();
  1228. if (mGfxPipelineRequiresBind)
  1229. {
  1230. if (!bindGraphicsPipeline())
  1231. return;
  1232. }
  1233. else
  1234. bindDynamicStates(false);
  1235. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1236. {
  1237. if (mNumBoundDescriptorSets > 0)
  1238. {
  1239. UINT32 deviceIdx = mDevice.getIndex();
  1240. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1241. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1242. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1243. }
  1244. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1245. }
  1246. if (instanceCount <= 0)
  1247. instanceCount = 1;
  1248. vkCmdDrawIndexed(mCmdBuffer, indexCount, instanceCount, startIndex, vertexOffset, 0);
  1249. }
  1250. void VulkanCmdBuffer::dispatch(UINT32 numGroupsX, UINT32 numGroupsY, UINT32 numGroupsZ)
  1251. {
  1252. if (mComputePipeline == nullptr)
  1253. return;
  1254. if (isInRenderPass())
  1255. endRenderPass();
  1256. // Note: Should I restore the render target after? Note that this is only being done is framebuffer subresources
  1257. // have their "isFBAttachment" flag reset, potentially I can just clear/restore those
  1258. setRenderTarget(nullptr, 0, RT_ALL);
  1259. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1260. bindGpuParams();
  1261. executeLayoutTransitions();
  1262. UINT32 deviceIdx = mDevice.getIndex();
  1263. if(mCmpPipelineRequiresBind)
  1264. {
  1265. VulkanPipeline* pipeline = mComputePipeline->getPipeline(deviceIdx);
  1266. if (pipeline == nullptr)
  1267. return;
  1268. registerResource(pipeline, VulkanUseFlag::Read);
  1269. mComputePipeline->registerPipelineResources(this);
  1270. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline->getHandle());
  1271. mCmpPipelineRequiresBind = false;
  1272. }
  1273. if(mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Compute))
  1274. {
  1275. if (mNumBoundDescriptorSets > 0)
  1276. {
  1277. VkPipelineLayout pipelineLayout = mComputePipeline->getPipelineLayout(deviceIdx);
  1278. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipelineLayout, 0,
  1279. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1280. }
  1281. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Compute);
  1282. }
  1283. vkCmdDispatch(mCmdBuffer, numGroupsX, numGroupsY, numGroupsZ);
  1284. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  1285. // and reset read-only state.
  1286. for(auto& entry : mPassTouchedSubresourceInfos)
  1287. {
  1288. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  1289. subresourceInfo.isShaderInput = false;
  1290. subresourceInfo.isReadOnly = true;
  1291. subresourceInfo.needsBarrier = false;
  1292. }
  1293. for (auto& entry : mBuffers)
  1294. entry.second.needsBarrier = false;
  1295. mPassTouchedSubresourceInfos.clear();
  1296. }
  1297. void VulkanCmdBuffer::setEvent(VulkanEvent* event)
  1298. {
  1299. if(isInRenderPass())
  1300. mQueuedEvents.push_back(event);
  1301. else
  1302. vkCmdSetEvent(mCmdBuffer, event->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  1303. }
  1304. void VulkanCmdBuffer::resetQuery(VulkanQuery* query)
  1305. {
  1306. if (isInRenderPass())
  1307. mQueuedQueryResets.push_back(query);
  1308. else
  1309. query->reset(mCmdBuffer);
  1310. }
  1311. void VulkanCmdBuffer::memoryBarrier(VkBuffer buffer, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1312. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage)
  1313. {
  1314. VkBufferMemoryBarrier barrier;
  1315. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  1316. barrier.pNext = nullptr;
  1317. barrier.srcAccessMask = srcAccessFlags;
  1318. barrier.dstAccessMask = dstAccessFlags;
  1319. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1320. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1321. barrier.buffer = buffer;
  1322. barrier.offset = 0;
  1323. barrier.size = VK_WHOLE_SIZE;
  1324. vkCmdPipelineBarrier(getHandle(),
  1325. srcStage,
  1326. dstStage,
  1327. 0, 0, nullptr,
  1328. 1, &barrier,
  1329. 0, nullptr);
  1330. }
  1331. void VulkanCmdBuffer::memoryBarrier(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1332. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage, VkImageLayout layout,
  1333. const VkImageSubresourceRange& range)
  1334. {
  1335. VkImageMemoryBarrier barrier;
  1336. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1337. barrier.pNext = nullptr;
  1338. barrier.srcAccessMask = srcAccessFlags;
  1339. barrier.dstAccessMask = dstAccessFlags;
  1340. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1341. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1342. barrier.image = image;
  1343. barrier.subresourceRange = range;
  1344. barrier.oldLayout = layout;
  1345. barrier.newLayout = layout;
  1346. vkCmdPipelineBarrier(getHandle(),
  1347. srcStage,
  1348. dstStage,
  1349. 0, 0, nullptr,
  1350. 0, nullptr,
  1351. 1, &barrier);
  1352. }
  1353. void VulkanCmdBuffer::setLayout(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1354. VkImageLayout oldLayout, VkImageLayout newLayout, const VkImageSubresourceRange& range)
  1355. {
  1356. VkImageMemoryBarrier barrier;
  1357. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1358. barrier.pNext = nullptr;
  1359. barrier.srcAccessMask = srcAccessFlags;
  1360. barrier.dstAccessMask = dstAccessFlags;
  1361. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1362. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1363. barrier.oldLayout = oldLayout;
  1364. barrier.newLayout = newLayout;
  1365. barrier.image = image;
  1366. barrier.subresourceRange = range;
  1367. VkPipelineStageFlags srcStage = getPipelineStageFlags(srcAccessFlags);
  1368. VkPipelineStageFlags dstStage = getPipelineStageFlags(dstAccessFlags);
  1369. vkCmdPipelineBarrier(getHandle(),
  1370. srcStage,
  1371. dstStage,
  1372. 0, 0, nullptr,
  1373. 0, nullptr,
  1374. 1, &barrier);
  1375. }
  1376. VkImageLayout VulkanCmdBuffer::getCurrentLayout(VulkanImage* image, const VkImageSubresourceRange& range,
  1377. bool inRenderPass)
  1378. {
  1379. UINT32 face = range.baseArrayLayer;
  1380. UINT32 mip = range.baseMipLevel;
  1381. // The assumption is that all the subresources in the range will have the same layout, as this should be handled
  1382. // by registerResource(), or by external code (in the case of transfers). So we only check the first subresource.
  1383. VulkanImageSubresource* subresource = image->getSubresource(face, mip);
  1384. auto iterFind = mImages.find(image);
  1385. if (iterFind == mImages.end())
  1386. return subresource->getLayout();
  1387. UINT32 imageInfoIdx = iterFind->second;
  1388. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1389. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1390. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1391. {
  1392. ImageSubresourceInfo& entry = subresourceInfos[i];
  1393. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1394. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1395. {
  1396. // If it's a FB attachment, retrieve its layout after the render pass begins
  1397. if(entry.isFBAttachment && inRenderPass && mFramebuffer)
  1398. {
  1399. RenderSurfaceMask readMask = getFBReadMask();
  1400. // Is it a depth-stencil attachment?
  1401. if(mFramebuffer->hasDepthAttachment() && mFramebuffer->getDepthStencilAttachment().image == image)
  1402. {
  1403. if (readMask.isSet(RT_DEPTH))
  1404. {
  1405. if (readMask.isSet(RT_STENCIL))
  1406. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL;
  1407. else // Depth readable but stencil isn't
  1408. return VK_IMAGE_LAYOUT_GENERAL;
  1409. }
  1410. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1411. }
  1412. else // It is a color attachment
  1413. {
  1414. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1415. for (UINT32 j = 0; j < numColorAttachments; j++)
  1416. {
  1417. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(j);
  1418. if (attachment.image == image)
  1419. {
  1420. if (readMask.isSet((RenderSurfaceMaskBits)(1 << attachment.index)))
  1421. return VK_IMAGE_LAYOUT_GENERAL;
  1422. else
  1423. return VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1424. }
  1425. }
  1426. }
  1427. }
  1428. return entry.requiredLayout;
  1429. }
  1430. }
  1431. return image->getOptimalLayout();
  1432. }
  1433. void VulkanCmdBuffer::registerResource(VulkanResource* res, VulkanUseFlags flags)
  1434. {
  1435. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1436. if(insertResult.second) // New element
  1437. {
  1438. ResourceUseHandle& useHandle = insertResult.first->second;
  1439. useHandle.used = false;
  1440. useHandle.flags = flags;
  1441. res->notifyBound();
  1442. }
  1443. else // Existing element
  1444. {
  1445. ResourceUseHandle& useHandle = insertResult.first->second;
  1446. assert(!useHandle.used);
  1447. useHandle.flags |= flags;
  1448. }
  1449. }
  1450. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range, VulkanUseFlags flags,
  1451. ResourceUsage usage)
  1452. {
  1453. VkImageLayout layout = res->getOptimalLayout();
  1454. registerResource(res, range, layout, layout, flags, usage);
  1455. }
  1456. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range,
  1457. VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags, ResourceUsage usage)
  1458. {
  1459. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1460. bool isTransfer = usage == ResourceUsage::Transfer;
  1461. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1462. // If binding it for write in a shader (not as color attachment or transfer op), we will need to issue a memory
  1463. // barrier if the image gets used again during this render pass, so remember this information.
  1464. bool needsBarrier = isShaderBind && flags.isSet(VulkanUseFlag::Write);
  1465. UINT32 nextImageInfoIdx = (UINT32)mImageInfos.size();
  1466. auto registerSubresourceInfo = [&](const VkImageSubresourceRange& subresourceRange)
  1467. {
  1468. mSubresourceInfoStorage.push_back(ImageSubresourceInfo());
  1469. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage.back();
  1470. subresourceInfo.currentLayout = newLayout;
  1471. subresourceInfo.initialLayout = newLayout;
  1472. subresourceInfo.requiredLayout = newLayout;
  1473. subresourceInfo.finalLayout = finalLayout;
  1474. subresourceInfo.range = subresourceRange;
  1475. subresourceInfo.isFBAttachment = isFBAttachment;
  1476. subresourceInfo.isShaderInput = isShaderBind;
  1477. if (!isTransfer)
  1478. {
  1479. subresourceInfo.hasTransitioned = false;
  1480. subresourceInfo.isReadOnly = !flags.isSet(VulkanUseFlag::Write);
  1481. }
  1482. else
  1483. {
  1484. subresourceInfo.hasTransitioned = true; // Transfers handle layout transitions externally (at this point they are assumed to already be done)
  1485. subresourceInfo.isReadOnly = true; // Doesn't matter for transfers
  1486. }
  1487. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1488. subresourceInfo.needsBarrier = needsBarrier;
  1489. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1490. };
  1491. auto insertResult = mImages.insert(std::make_pair(res, nextImageInfoIdx));
  1492. if (insertResult.second) // New element
  1493. {
  1494. UINT32 imageInfoIdx = insertResult.first->second;
  1495. mImageInfos.push_back(ImageInfo());
  1496. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1497. imageInfo.subresourceInfoIdx = (UINT32)mSubresourceInfoStorage.size();
  1498. imageInfo.numSubresourceInfos = 1;
  1499. imageInfo.useHandle.used = false;
  1500. imageInfo.useHandle.flags = flags;
  1501. registerSubresourceInfo(range);
  1502. res->notifyBound();
  1503. }
  1504. else // Existing element
  1505. {
  1506. UINT32 imageInfoIdx = insertResult.first->second;
  1507. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1508. assert(!imageInfo.useHandle.used);
  1509. imageInfo.useHandle.flags |= flags;
  1510. // See if there is an overlap between existing ranges and the new range. And if so break them up accordingly.
  1511. //// First test for the simplest and most common case (same range or no overlap) to avoid more complex
  1512. //// computations.
  1513. ImageSubresourceInfo* subresources = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1514. bool foundRange = false;
  1515. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1516. {
  1517. if(VulkanUtility::rangeOverlaps(subresources[i].range, range))
  1518. {
  1519. if (subresources[i].range.layerCount == range.layerCount &&
  1520. subresources[i].range.levelCount == range.levelCount &&
  1521. subresources[i].range.baseArrayLayer == range.baseArrayLayer &&
  1522. subresources[i].range.baseMipLevel == range.baseMipLevel)
  1523. {
  1524. // Just update existing range
  1525. updateSubresourceInfo(res, imageInfoIdx, subresources[i], newLayout, finalLayout, flags, usage);
  1526. mPassTouchedSubresourceInfos.insert(imageInfo.subresourceInfoIdx + i);
  1527. foundRange = true;
  1528. break;
  1529. }
  1530. break;
  1531. }
  1532. }
  1533. //// We'll need to update subresource ranges or add new ones. The hope is that this code is trigger VERY rarely
  1534. //// (for just a few specific textures per frame).
  1535. if (!foundRange)
  1536. {
  1537. std::array<VkImageSubresourceRange, 5> tempCutRanges;
  1538. bs_frame_mark();
  1539. {
  1540. // We orphan previously allocated memory (we reset it after submit() anyway)
  1541. UINT32 newSubresourceIdx = (UINT32)mSubresourceInfoStorage.size();
  1542. FrameVector<UINT32> cutOverlappingRanges;
  1543. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1544. {
  1545. UINT32 subresourceIdx = imageInfo.subresourceInfoIdx + i;
  1546. ImageSubresourceInfo& subresource = mSubresourceInfoStorage[subresourceIdx];
  1547. if (!VulkanUtility::rangeOverlaps(subresource.range, range))
  1548. {
  1549. // Just copy as is
  1550. mSubresourceInfoStorage.push_back(subresource);
  1551. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1552. }
  1553. else // Need to cut
  1554. {
  1555. UINT32 numCutRanges;
  1556. VulkanUtility::cutRange(subresource.range, range, tempCutRanges, numCutRanges);
  1557. for(UINT32 j = 0; j < numCutRanges; j++)
  1558. {
  1559. // Create a copy of the original subresource with the new range
  1560. ImageSubresourceInfo newInfo = mSubresourceInfoStorage[subresourceIdx];
  1561. newInfo.range = tempCutRanges[j];
  1562. if(VulkanUtility::rangeOverlaps(tempCutRanges[j], range))
  1563. {
  1564. // Update overlapping sub-resource range with new data from this range
  1565. updateSubresourceInfo(res, imageInfoIdx, newInfo, newLayout, finalLayout, flags, usage);
  1566. // Keep track of the overlapping ranges for later
  1567. cutOverlappingRanges.push_back((UINT32)mSubresourceInfoStorage.size());
  1568. }
  1569. mSubresourceInfoStorage.push_back(newInfo);
  1570. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1571. }
  1572. }
  1573. }
  1574. // Our range doesn't overlap with any existing ranges, so just add it
  1575. if(cutOverlappingRanges.size() == 0)
  1576. {
  1577. registerSubresourceInfo(range);
  1578. }
  1579. else // Search if overlapping ranges fully cover the requested range, and insert non-covered regions
  1580. {
  1581. FrameQueue<VkImageSubresourceRange> sourceRanges;
  1582. sourceRanges.push(range);
  1583. for(auto& entry : cutOverlappingRanges)
  1584. {
  1585. VkImageSubresourceRange& overlappingRange = mSubresourceInfoStorage[entry].range;
  1586. UINT32 numSourceRanges = (UINT32)sourceRanges.size();
  1587. for(UINT32 i = 0; i < numSourceRanges; i++)
  1588. {
  1589. VkImageSubresourceRange sourceRange = sourceRanges.front();
  1590. sourceRanges.pop();
  1591. UINT32 numCutRanges;
  1592. VulkanUtility::cutRange(sourceRange, overlappingRange, tempCutRanges, numCutRanges);
  1593. for(UINT32 j = 0; j < numCutRanges; j++)
  1594. {
  1595. // We only care about ranges outside of the ones we already covered
  1596. if(!VulkanUtility::rangeOverlaps(tempCutRanges[j], overlappingRange))
  1597. sourceRanges.push(tempCutRanges[j]);
  1598. }
  1599. }
  1600. }
  1601. // Any remaining range hasn't been covered yet
  1602. while(!sourceRanges.empty())
  1603. {
  1604. registerSubresourceInfo(sourceRanges.front());
  1605. sourceRanges.pop();
  1606. }
  1607. }
  1608. imageInfo.subresourceInfoIdx = newSubresourceIdx;
  1609. imageInfo.numSubresourceInfos = (UINT32)mSubresourceInfoStorage.size() - newSubresourceIdx;
  1610. }
  1611. bs_frame_clear();
  1612. }
  1613. }
  1614. // Register any sub-resources
  1615. for(UINT32 i = 0; i < range.layerCount; i++)
  1616. {
  1617. for(UINT32 j = 0; j < range.levelCount; j++)
  1618. {
  1619. UINT32 layer = range.baseArrayLayer + i;
  1620. UINT32 mipLevel = range.baseMipLevel + j;
  1621. registerResource(res->getSubresource(layer, mipLevel), flags);
  1622. }
  1623. }
  1624. }
  1625. void VulkanCmdBuffer::registerResource(VulkanBuffer* res, VkAccessFlags accessFlags, VulkanUseFlags flags)
  1626. {
  1627. bool isShaderWrite = (accessFlags & VK_ACCESS_SHADER_WRITE_BIT) != 0;
  1628. auto insertResult = mBuffers.insert(std::make_pair(res, BufferInfo()));
  1629. if (insertResult.second) // New element
  1630. {
  1631. BufferInfo& bufferInfo = insertResult.first->second;
  1632. bufferInfo.accessFlags = accessFlags;
  1633. bufferInfo.useHandle.used = false;
  1634. bufferInfo.useHandle.flags = flags;
  1635. // Any writes done on storage buffers will need explicit memory barriers (if read during the same pass) so
  1636. // we remember this, in case this buffers gets used later in this pass.
  1637. bufferInfo.needsBarrier = isShaderWrite;
  1638. res->notifyBound();
  1639. }
  1640. else // Existing element
  1641. {
  1642. BufferInfo& bufferInfo = insertResult.first->second;
  1643. assert(!bufferInfo.useHandle.used);
  1644. bufferInfo.useHandle.flags |= flags;
  1645. bufferInfo.accessFlags |= accessFlags;
  1646. // If the buffer was written to previously in this pass, and is now being used by a shader we need to issue
  1647. // a barrier to make those writes visible.
  1648. bool isShaderRead = (accessFlags & VK_ACCESS_SHADER_READ_BIT) != 0;
  1649. if(bufferInfo.needsBarrier && (isShaderRead || isShaderWrite))
  1650. {
  1651. // Need to end render pass in order to execute the barrier. Hopefully this won't trigger much since most
  1652. // shader writes are done during compute
  1653. if (isInRenderPass())
  1654. endRenderPass();
  1655. VkPipelineStageFlags stages =
  1656. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1657. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1658. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1659. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1660. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1661. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1662. VkBuffer buffer = res->getHandle();
  1663. memoryBarrier(buffer, VK_ACCESS_SHADER_WRITE_BIT, accessFlags, stages, stages);
  1664. bufferInfo.needsBarrier = isShaderWrite;
  1665. }
  1666. }
  1667. }
  1668. void VulkanCmdBuffer::registerResource(VulkanFramebuffer* res, RenderSurfaceMask loadMask, UINT32 readMask)
  1669. {
  1670. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1671. if (insertResult.second) // New element
  1672. {
  1673. ResourceUseHandle& useHandle = insertResult.first->second;
  1674. useHandle.used = false;
  1675. useHandle.flags = VulkanUseFlag::Write;
  1676. res->notifyBound();
  1677. }
  1678. else // Existing element
  1679. {
  1680. ResourceUseHandle& useHandle = insertResult.first->second;
  1681. assert(!useHandle.used);
  1682. useHandle.flags |= VulkanUseFlag::Write;
  1683. }
  1684. // Register any sub-resources
  1685. UINT32 numColorAttachments = res->getNumColorAttachments();
  1686. for (UINT32 i = 0; i < numColorAttachments; i++)
  1687. {
  1688. const VulkanFramebufferAttachment& attachment = res->getColorAttachment(i);
  1689. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1690. VkImageLayout layout;
  1691. if (loadMask.isSet((RenderSurfaceMaskBits)(1 << i)))
  1692. {
  1693. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1694. // transition this to a different layout when it begins, but we handle that elsewhere
  1695. layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1696. }
  1697. else
  1698. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1699. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1700. registerResource(attachment.image, range, layout, attachment.finalLayout, VulkanUseFlag::Write,
  1701. ResourceUsage::Framebuffer);
  1702. }
  1703. if(res->hasDepthAttachment())
  1704. {
  1705. const VulkanFramebufferAttachment& attachment = res->getDepthStencilAttachment();
  1706. VulkanUseFlag useFlag = VulkanUseFlag::Write;
  1707. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1708. VkImageLayout layout;
  1709. if (loadMask.isSet(RT_DEPTH) || loadMask.isSet(RT_STENCIL)) // Can't load one without the other
  1710. {
  1711. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1712. // transition this to a different layout when it begins, but we handle that elsewhere
  1713. layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1714. }
  1715. else
  1716. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1717. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1718. registerResource(attachment.image, range, layout, attachment.finalLayout, useFlag, ResourceUsage::Framebuffer);
  1719. }
  1720. }
  1721. bool VulkanCmdBuffer::updateSubresourceInfo(VulkanImage* image, UINT32 imageInfoIdx,
  1722. ImageSubresourceInfo& subresourceInfo, VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags,
  1723. ResourceUsage usage)
  1724. {
  1725. bool isTransfer = usage == ResourceUsage::Transfer;
  1726. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1727. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1728. // Transfers are always considered read only because this only matters for destination access mask, and transfers
  1729. // handle that externally
  1730. if(!isTransfer)
  1731. subresourceInfo.isReadOnly &= !flags.isSet(VulkanUseFlag::Write);
  1732. // For transfers, just assign new layout, external code does all the work
  1733. if (isTransfer)
  1734. subresourceInfo.requiredLayout = newLayout;
  1735. else
  1736. {
  1737. // New layout is valid, check for transitions (UNDEFINED signifies the caller doesn't want a layout transition)
  1738. if (newLayout != VK_IMAGE_LAYOUT_UNDEFINED)
  1739. {
  1740. // If layout transition was requested by framebuffer bind, respect it because render-pass will only accept a
  1741. // specific layout (in certain cases), and we have no choice.
  1742. // In the case when a FB attachment is also bound for shader reads, this will override the layout required for
  1743. // shader read (GENERAL or DEPTH_READ_ONLY), but that is fine because those transitions are handled
  1744. // automatically by render-pass layout transitions.
  1745. // Any other texture (non FB attachment) will only even be bound in a single layout and we can keep the one it
  1746. // was originally registered with.
  1747. if (isFBAttachment)
  1748. subresourceInfo.requiredLayout = newLayout;
  1749. else if (!subresourceInfo.isFBAttachment) // Layout transition is not being done on a FB image
  1750. {
  1751. // Check if the image had a layout previously assigned, and if so check if multiple different layouts
  1752. // were requested. In that case we wish to transfer the image to GENERAL layout.
  1753. bool firstUseInRenderPass = !subresourceInfo.isShaderInput && !subresourceInfo.isFBAttachment;
  1754. if (firstUseInRenderPass || subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  1755. subresourceInfo.requiredLayout = newLayout;
  1756. else if (subresourceInfo.requiredLayout != newLayout)
  1757. subresourceInfo.requiredLayout = VK_IMAGE_LAYOUT_GENERAL;
  1758. }
  1759. }
  1760. }
  1761. // If attached to FB, then the final layout is set by the FB (provided as layout param here), otherwise its
  1762. // the same as required layout
  1763. if (!isFBAttachment && !subresourceInfo.isFBAttachment)
  1764. subresourceInfo.finalLayout = subresourceInfo.requiredLayout;
  1765. else
  1766. {
  1767. if (isFBAttachment)
  1768. subresourceInfo.finalLayout = finalLayout;
  1769. }
  1770. // Queue layout transition, if not transfer. Transfer handle layout transitions externally.
  1771. if (!isTransfer)
  1772. {
  1773. // If we haven't done a layout transition yet, we can just overwrite the previously written values, and the
  1774. // transition will be handled as the first thing in submit(), otherwise we queue a non-initial transition
  1775. // below.
  1776. if (!subresourceInfo.hasTransitioned)
  1777. {
  1778. subresourceInfo.initialLayout = subresourceInfo.requiredLayout;
  1779. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1780. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1781. }
  1782. else
  1783. {
  1784. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1785. mQueuedLayoutTransitions[image] = imageInfoIdx;
  1786. }
  1787. }
  1788. else
  1789. {
  1790. // Resource has already transitioned its layout externally since this is a transfer. We cannot allow a
  1791. // transition in submit().
  1792. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1793. subresourceInfo.hasTransitioned = true;
  1794. }
  1795. // If a FB attachment was just bound as a shader input, we might need to restart the render pass with a FB
  1796. // attachment that supports read-only attachments using the GENERAL or DEPTH_READ_ONLY layout
  1797. bool resetRenderPass = false;
  1798. if (isFBAttachment)
  1799. {
  1800. if (!subresourceInfo.isFBAttachment)
  1801. {
  1802. subresourceInfo.isFBAttachment = true;
  1803. resetRenderPass = subresourceInfo.isShaderInput;
  1804. }
  1805. }
  1806. else
  1807. {
  1808. if (!subresourceInfo.isShaderInput)
  1809. {
  1810. subresourceInfo.isShaderInput = true;
  1811. if (subresourceInfo.isFBAttachment)
  1812. {
  1813. // Special case for depth: If user has set up proper read-only flags, then the render pass will have
  1814. // taken care of setting the valid state anyway, so no need to end the render pass
  1815. if (subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL)
  1816. {
  1817. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0 || (mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1818. resetRenderPass = false;
  1819. else
  1820. resetRenderPass = true;
  1821. }
  1822. else
  1823. resetRenderPass = true;
  1824. }
  1825. else
  1826. {
  1827. // Subresource has been bound as shader input, and it wasn't bound as a FB attachment this render pass.
  1828. // However is could have been bound in a previous pass, so check the layouts and force a layout
  1829. // transition if required.
  1830. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1831. resetRenderPass = true;
  1832. else
  1833. resetRenderPass = false;
  1834. }
  1835. }
  1836. }
  1837. // If we need to switch frame-buffers, end current render pass
  1838. if (resetRenderPass && isInRenderPass())
  1839. endRenderPass();
  1840. else
  1841. {
  1842. // Since we won't be ending the render pass, check if this same sub-resource was written earlier in the pass,
  1843. // in which case we need to issue a memory barrier so those writes are visible.
  1844. // Memory barrier only matters if image is bound for shader use (no need for color attachments or transfers)
  1845. if(subresourceInfo.needsBarrier && isShaderBind)
  1846. {
  1847. bool isWrite = flags.isSet(VulkanUseFlag::Write);
  1848. VkPipelineStageFlags stages =
  1849. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1850. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1851. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1852. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1853. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1854. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1855. memoryBarrier(image->getHandle(), VK_ACCESS_SHADER_WRITE_BIT,
  1856. image->getAccessFlags(subresourceInfo.requiredLayout, !isWrite),
  1857. stages, stages, subresourceInfo.requiredLayout, subresourceInfo.range);
  1858. subresourceInfo.needsBarrier = isWrite;
  1859. }
  1860. }
  1861. return resetRenderPass;
  1862. }
  1863. VulkanCmdBuffer::ImageSubresourceInfo& VulkanCmdBuffer::findSubresourceInfo(VulkanImage* image, UINT32 face, UINT32 mip)
  1864. {
  1865. UINT32 imageInfoIdx = mImages[image];
  1866. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1867. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1868. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1869. {
  1870. ImageSubresourceInfo& entry = subresourceInfos[i];
  1871. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1872. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1873. {
  1874. return entry;
  1875. }
  1876. }
  1877. assert(false); // Caller should ensure the subresource actually exists, so this shouldn't happen
  1878. return subresourceInfos[0];
  1879. }
  1880. void VulkanCmdBuffer::getInProgressQueries(Vector<VulkanTimerQuery*>& timer, Vector<VulkanOcclusionQuery*>& occlusion) const
  1881. {
  1882. for(auto& query : mTimerQueries)
  1883. {
  1884. if (query->_isInProgress())
  1885. timer.push_back(query);
  1886. }
  1887. for (auto& query : mOcclusionQueries)
  1888. {
  1889. if (query->_isInProgress())
  1890. occlusion.push_back(query);
  1891. }
  1892. }
  1893. VulkanCommandBuffer::VulkanCommandBuffer(VulkanDevice& device, GpuQueueType type, UINT32 deviceIdx,
  1894. UINT32 queueIdx, bool secondary)
  1895. : CommandBuffer(type, deviceIdx, queueIdx, secondary), mBuffer(nullptr)
  1896. , mDevice(device), mQueue(nullptr), mIdMask(0)
  1897. {
  1898. UINT32 numQueues = device.getNumQueues(mType);
  1899. if (numQueues == 0) // Fall back to graphics queue
  1900. {
  1901. mType = GQT_GRAPHICS;
  1902. numQueues = device.getNumQueues(GQT_GRAPHICS);
  1903. }
  1904. mQueue = device.getQueue(mType, mQueueIdx % numQueues);
  1905. mIdMask = device.getQueueMask(mType, mQueueIdx);
  1906. acquireNewBuffer();
  1907. }
  1908. RenderSurfaceMask VulkanCmdBuffer::getFBReadMask()
  1909. {
  1910. // Check if any frame-buffer attachments are also used as shader inputs, in which case we make them read-only
  1911. RenderSurfaceMask readMask = RT_NONE;
  1912. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1913. for(UINT32 i = 0; i < numColorAttachments; i++)
  1914. {
  1915. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1916. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1917. fbAttachment.surface.mipLevel);
  1918. bool readOnly = subresourceInfo.isShaderInput;
  1919. if(readOnly)
  1920. readMask.set((RenderSurfaceMaskBits)(1 << i));
  1921. }
  1922. if (mFramebuffer->hasDepthAttachment())
  1923. {
  1924. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1925. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1926. fbAttachment.surface.mipLevel);
  1927. bool readOnly = subresourceInfo.isShaderInput;
  1928. if (readOnly)
  1929. readMask.set(RT_DEPTH);
  1930. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0)
  1931. readMask.set(RT_DEPTH);
  1932. if ((mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1933. readMask.set(RT_STENCIL);
  1934. }
  1935. return readMask;
  1936. }
  1937. VulkanCommandBuffer::~VulkanCommandBuffer()
  1938. {
  1939. mBuffer->reset();
  1940. }
  1941. void VulkanCommandBuffer::acquireNewBuffer()
  1942. {
  1943. VulkanCmdBufferPool& pool = mDevice.getCmdBufferPool();
  1944. if (mBuffer != nullptr)
  1945. assert(mBuffer->isSubmitted());
  1946. UINT32 queueFamily = mDevice.getQueueFamily(mType);
  1947. mBuffer = pool.getBuffer(queueFamily, mIsSecondary);
  1948. }
  1949. void VulkanCommandBuffer::submit(UINT32 syncMask)
  1950. {
  1951. // Ignore myself
  1952. syncMask &= ~mIdMask;
  1953. if (mBuffer->isInRenderPass())
  1954. mBuffer->endRenderPass();
  1955. // Execute any queued layout transitions that weren't already handled by the render pass
  1956. mBuffer->executeLayoutTransitions();
  1957. // Interrupt any in-progress queries (no in-progress queries allowed during command buffer submit)
  1958. Vector<VulkanTimerQuery*> timerQueries;
  1959. Vector<VulkanOcclusionQuery*> occlusionQueries;
  1960. mBuffer->getInProgressQueries(timerQueries, occlusionQueries);
  1961. for (auto& query : timerQueries)
  1962. query->_interrupt(*mBuffer);
  1963. for (auto& query : occlusionQueries)
  1964. query->_interrupt(*mBuffer);
  1965. if (mBuffer->isRecording())
  1966. mBuffer->end();
  1967. if (mBuffer->isReadyForSubmit()) // Possibly nothing was recorded in the buffer
  1968. {
  1969. mBuffer->submit(mQueue, mQueueIdx, syncMask);
  1970. acquireNewBuffer();
  1971. gVulkanCBManager().refreshStates(mDeviceIdx);
  1972. }
  1973. // Resume interrupted queries on the new command buffer
  1974. for (auto& query : timerQueries)
  1975. query->_resume(*mBuffer);
  1976. for (auto& query : occlusionQueries)
  1977. query->_resume(*mBuffer);
  1978. }
  1979. }}