BsVulkanCommandBuffer.cpp 71 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanCommandBuffer.h"
  4. #include "Managers/BsVulkanCommandBufferManager.h"
  5. #include "BsVulkanUtility.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParams.h"
  8. #include "BsVulkanQueue.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanIndexBuffer.h"
  11. #include "BsVulkanVertexBuffer.h"
  12. #include "BsVulkanHardwareBuffer.h"
  13. #include "BsVulkanFramebuffer.h"
  14. #include "Managers/BsVulkanVertexInputManager.h"
  15. #include "BsVulkanEventQuery.h"
  16. #include "Managers/BsVulkanQueryManager.h"
  17. #include "BsVulkanSwapChain.h"
  18. #include "BsVulkanTimerQuery.h"
  19. #include "BsVulkanOcclusionQuery.h"
  20. #if BS_PLATFORM == BS_PLATFORM_WIN32
  21. #include "Win32/BsWin32RenderWindow.h"
  22. #else
  23. static_assert(false, "Other platforms go here");
  24. #endif
  25. namespace bs { namespace ct
  26. {
  27. VulkanSemaphore::VulkanSemaphore(VulkanResourceManager* owner)
  28. :VulkanResource(owner, true)
  29. {
  30. VkSemaphoreCreateInfo semaphoreCI;
  31. semaphoreCI.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  32. semaphoreCI.pNext = nullptr;
  33. semaphoreCI.flags = 0;
  34. VkResult result = vkCreateSemaphore(owner->getDevice().getLogical(), &semaphoreCI, gVulkanAllocator, &mSemaphore);
  35. assert(result == VK_SUCCESS);
  36. }
  37. VulkanSemaphore::~VulkanSemaphore()
  38. {
  39. vkDestroySemaphore(mOwner->getDevice().getLogical(), mSemaphore, gVulkanAllocator);
  40. }
  41. VulkanCmdBufferPool::VulkanCmdBufferPool(VulkanDevice& device)
  42. :mDevice(device), mNextId(1)
  43. {
  44. for (UINT32 i = 0; i < GQT_COUNT; i++)
  45. {
  46. UINT32 familyIdx = device.getQueueFamily((GpuQueueType)i);
  47. if (familyIdx == (UINT32)-1)
  48. continue;
  49. VkCommandPoolCreateInfo poolCI;
  50. poolCI.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  51. poolCI.pNext = nullptr;
  52. poolCI.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
  53. poolCI.queueFamilyIndex = familyIdx;
  54. PoolInfo& poolInfo = mPools[familyIdx];
  55. poolInfo.queueFamily = familyIdx;
  56. memset(poolInfo.buffers, 0, sizeof(poolInfo.buffers));
  57. vkCreateCommandPool(device.getLogical(), &poolCI, gVulkanAllocator, &poolInfo.pool);
  58. }
  59. }
  60. VulkanCmdBufferPool::~VulkanCmdBufferPool()
  61. {
  62. // Note: Shutdown should be the only place command buffers are destroyed at, as the system relies on the fact that
  63. // they won't be destroyed during normal operation.
  64. for(auto& entry : mPools)
  65. {
  66. PoolInfo& poolInfo = entry.second;
  67. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  68. {
  69. VulkanCmdBuffer* buffer = poolInfo.buffers[i];
  70. if (buffer == nullptr)
  71. break;
  72. bs_delete(buffer);
  73. }
  74. vkDestroyCommandPool(mDevice.getLogical(), poolInfo.pool, gVulkanAllocator);
  75. }
  76. }
  77. VulkanCmdBuffer* VulkanCmdBufferPool::getBuffer(UINT32 queueFamily, bool secondary)
  78. {
  79. auto iterFind = mPools.find(queueFamily);
  80. if (iterFind == mPools.end())
  81. return nullptr;
  82. VulkanCmdBuffer** buffers = iterFind->second.buffers;
  83. UINT32 i = 0;
  84. for(; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  85. {
  86. if (buffers[i] == nullptr)
  87. break;
  88. if(buffers[i]->mState == VulkanCmdBuffer::State::Ready)
  89. {
  90. buffers[i]->begin();
  91. return buffers[i];
  92. }
  93. }
  94. assert(i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY &&
  95. "Too many command buffers allocated. Increment BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY to a higher value. ");
  96. buffers[i] = createBuffer(queueFamily, secondary);
  97. buffers[i]->begin();
  98. return buffers[i];
  99. }
  100. VulkanCmdBuffer* VulkanCmdBufferPool::createBuffer(UINT32 queueFamily, bool secondary)
  101. {
  102. auto iterFind = mPools.find(queueFamily);
  103. if (iterFind == mPools.end())
  104. return nullptr;
  105. const PoolInfo& poolInfo = iterFind->second;
  106. return bs_new<VulkanCmdBuffer>(mDevice, mNextId++, poolInfo.pool, poolInfo.queueFamily, secondary);
  107. }
  108. VulkanCmdBuffer::VulkanCmdBuffer(VulkanDevice& device, UINT32 id, VkCommandPool pool, UINT32 queueFamily, bool secondary)
  109. : mId(id), mQueueFamily(queueFamily), mState(State::Ready), mDevice(device), mPool(pool)
  110. , mIntraQueueSemaphore(nullptr), mInterQueueSemaphores(), mNumUsedInterQueueSemaphores(0)
  111. , mFramebuffer(nullptr), mRenderTargetWidth(0)
  112. , mRenderTargetHeight(0), mRenderTargetReadOnlyFlags(0), mRenderTargetLoadMask(RT_NONE), mGlobalQueueIdx(-1)
  113. , mViewport(0.0f, 0.0f, 1.0f, 1.0f), mScissor(0, 0, 0, 0), mStencilRef(0), mDrawOp(DOT_TRIANGLE_LIST)
  114. , mNumBoundDescriptorSets(0), mGfxPipelineRequiresBind(true), mCmpPipelineRequiresBind(true)
  115. , mViewportRequiresBind(true), mStencilRefRequiresBind(true), mScissorRequiresBind(true), mBoundParamsDirty(false)
  116. , mClearValues(), mClearMask(), mSemaphoresTemp(BS_MAX_UNIQUE_QUEUES), mVertexBuffersTemp()
  117. , mVertexBufferOffsetsTemp()
  118. {
  119. UINT32 maxBoundDescriptorSets = device.getDeviceProperties().limits.maxBoundDescriptorSets;
  120. mDescriptorSetsTemp = (VkDescriptorSet*)bs_alloc(sizeof(VkDescriptorSet) * maxBoundDescriptorSets);
  121. VkCommandBufferAllocateInfo cmdBufferAllocInfo;
  122. cmdBufferAllocInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  123. cmdBufferAllocInfo.pNext = nullptr;
  124. cmdBufferAllocInfo.commandPool = pool;
  125. cmdBufferAllocInfo.level = secondary ? VK_COMMAND_BUFFER_LEVEL_SECONDARY : VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  126. cmdBufferAllocInfo.commandBufferCount = 1;
  127. VkResult result = vkAllocateCommandBuffers(mDevice.getLogical(), &cmdBufferAllocInfo, &mCmdBuffer);
  128. assert(result == VK_SUCCESS);
  129. VkFenceCreateInfo fenceCI;
  130. fenceCI.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  131. fenceCI.pNext = nullptr;
  132. fenceCI.flags = 0;
  133. result = vkCreateFence(mDevice.getLogical(), &fenceCI, gVulkanAllocator, &mFence);
  134. assert(result == VK_SUCCESS);
  135. }
  136. VulkanCmdBuffer::~VulkanCmdBuffer()
  137. {
  138. VkDevice device = mDevice.getLogical();
  139. if(mState == State::Submitted)
  140. {
  141. // Wait 1s
  142. UINT64 waitTime = 1000 * 1000 * 1000;
  143. VkResult result = vkWaitForFences(device, 1, &mFence, true, waitTime);
  144. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  145. if (result == VK_TIMEOUT)
  146. LOGWRN("Freeing a command buffer before done executing because fence wait expired!");
  147. // Resources have been marked as used, make sure to notify them we're done with them
  148. reset();
  149. }
  150. else if(mState != State::Ready)
  151. {
  152. // Notify any resources that they are no longer bound
  153. for (auto& entry : mResources)
  154. {
  155. ResourceUseHandle& useHandle = entry.second;
  156. assert(!useHandle.used);
  157. entry.first->notifyUnbound();
  158. }
  159. for (auto& entry : mImages)
  160. {
  161. UINT32 imageInfoIdx = entry.second;
  162. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  163. ResourceUseHandle& useHandle = imageInfo.useHandle;
  164. assert(!useHandle.used);
  165. entry.first->notifyUnbound();
  166. }
  167. for (auto& entry : mBuffers)
  168. {
  169. ResourceUseHandle& useHandle = entry.second.useHandle;
  170. assert(!useHandle.used);
  171. entry.first->notifyUnbound();
  172. }
  173. }
  174. if (mIntraQueueSemaphore != nullptr)
  175. mIntraQueueSemaphore->destroy();
  176. for(UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  177. {
  178. if (mInterQueueSemaphores[i] != nullptr)
  179. mInterQueueSemaphores[i]->destroy();
  180. }
  181. vkDestroyFence(device, mFence, gVulkanAllocator);
  182. vkFreeCommandBuffers(device, mPool, 1, &mCmdBuffer);
  183. bs_free(mDescriptorSetsTemp);
  184. }
  185. UINT32 VulkanCmdBuffer::getDeviceIdx() const
  186. {
  187. return mDevice.getIndex();
  188. }
  189. void VulkanCmdBuffer::begin()
  190. {
  191. assert(mState == State::Ready);
  192. VkCommandBufferBeginInfo beginInfo;
  193. beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  194. beginInfo.pNext = nullptr;
  195. beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  196. beginInfo.pInheritanceInfo = nullptr;
  197. VkResult result = vkBeginCommandBuffer(mCmdBuffer, &beginInfo);
  198. assert(result == VK_SUCCESS);
  199. mState = State::Recording;
  200. }
  201. void VulkanCmdBuffer::end()
  202. {
  203. assert(mState == State::Recording);
  204. // If a clear is queued, execute the render pass with no additional instructions
  205. if (mClearMask)
  206. executeClearPass();
  207. VkResult result = vkEndCommandBuffer(mCmdBuffer);
  208. assert(result == VK_SUCCESS);
  209. mState = State::RecordingDone;
  210. }
  211. void VulkanCmdBuffer::beginRenderPass()
  212. {
  213. assert(mState == State::Recording);
  214. if (mFramebuffer == nullptr)
  215. {
  216. LOGWRN("Attempting to begin a render pass but no render target is bound to the command buffer.");
  217. return;
  218. }
  219. if(mClearMask != CLEAR_NONE)
  220. {
  221. // If a previous clear is queued, but it doesn't match the rendered area, need to execute a separate pass
  222. // just for it
  223. Rect2I rtArea(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  224. if (mClearArea != rtArea)
  225. executeClearPass();
  226. }
  227. executeLayoutTransitions();
  228. // Check if any frame-buffer attachments are also used as shader inputs, in which case we make them read-only
  229. RenderSurfaceMask readMask = RT_NONE;
  230. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  231. for(UINT32 i = 0; i < numColorAttachments; i++)
  232. {
  233. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  234. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  235. fbAttachment.surface.mipLevel);
  236. bool readOnly = subresourceInfo.isShaderInput;
  237. if(readOnly)
  238. readMask.set((RenderSurfaceMaskBits)(1 << i));
  239. }
  240. if (mFramebuffer->hasDepthAttachment())
  241. {
  242. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  243. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  244. fbAttachment.surface.mipLevel);
  245. bool readOnly = subresourceInfo.isShaderInput;
  246. if (readOnly)
  247. readMask.set(RT_DEPTH);
  248. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0)
  249. readMask.set(RT_DEPTH);
  250. if ((mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  251. readMask.set(RT_STENCIL);
  252. }
  253. for(auto& entry : mPassTouchedSubresourceInfos)
  254. {
  255. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  256. subresourceInfo.needsBarrier = false;
  257. }
  258. for (auto& entry : mBuffers)
  259. entry.second.needsBarrier = false;
  260. VkRenderPassBeginInfo renderPassBeginInfo;
  261. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  262. renderPassBeginInfo.pNext = nullptr;
  263. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(mRenderTargetLoadMask, readMask, mClearMask);
  264. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(mRenderTargetLoadMask, readMask, mClearMask);
  265. renderPassBeginInfo.renderArea.offset.x = 0;
  266. renderPassBeginInfo.renderArea.offset.y = 0;
  267. renderPassBeginInfo.renderArea.extent.width = mRenderTargetWidth;
  268. renderPassBeginInfo.renderArea.extent.height = mRenderTargetHeight;
  269. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  270. renderPassBeginInfo.pClearValues = mClearValues.data();
  271. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  272. mClearMask = CLEAR_NONE;
  273. mState = State::RecordingRenderPass;
  274. }
  275. void VulkanCmdBuffer::endRenderPass()
  276. {
  277. assert(mState == State::RecordingRenderPass);
  278. vkCmdEndRenderPass(mCmdBuffer);
  279. // Execute any queued events
  280. for(auto& entry : mQueuedEvents)
  281. vkCmdSetEvent(mCmdBuffer, entry->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  282. mQueuedEvents.clear();
  283. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  284. // and reset read-only state.
  285. for(auto& entry : mPassTouchedSubresourceInfos)
  286. {
  287. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  288. subresourceInfo.isShaderInput = false;
  289. subresourceInfo.isReadOnly = true;
  290. }
  291. mPassTouchedSubresourceInfos.clear();
  292. updateFinalLayouts();
  293. mState = State::Recording;
  294. // In case the same GPU params from last pass get used, this makes sure the states we reset above, get re-applied
  295. mBoundParamsDirty = true;
  296. }
  297. void VulkanCmdBuffer::allocateSemaphores(VkSemaphore* semaphores)
  298. {
  299. if (mIntraQueueSemaphore != nullptr)
  300. mIntraQueueSemaphore->destroy();
  301. mIntraQueueSemaphore = mDevice.getResourceManager().create<VulkanSemaphore>();
  302. semaphores[0] = mIntraQueueSemaphore->getHandle();
  303. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  304. {
  305. if (mInterQueueSemaphores[i] != nullptr)
  306. mInterQueueSemaphores[i]->destroy();
  307. mInterQueueSemaphores[i] = mDevice.getResourceManager().create<VulkanSemaphore>();
  308. semaphores[i + 1] = mInterQueueSemaphores[i]->getHandle();
  309. }
  310. mNumUsedInterQueueSemaphores = 0;
  311. }
  312. VulkanSemaphore* VulkanCmdBuffer::requestInterQueueSemaphore() const
  313. {
  314. if (mNumUsedInterQueueSemaphores >= BS_MAX_VULKAN_CB_DEPENDENCIES)
  315. return nullptr;
  316. return mInterQueueSemaphores[mNumUsedInterQueueSemaphores++];
  317. }
  318. void VulkanCmdBuffer::submit(VulkanQueue* queue, UINT32 queueIdx, UINT32 syncMask)
  319. {
  320. assert(isReadyForSubmit());
  321. // Make sure to reset the CB fence before we submit it
  322. VkResult result = vkResetFences(mDevice.getLogical(), 1, &mFence);
  323. assert(result == VK_SUCCESS);
  324. // If there are any query resets needed, execute those first
  325. VulkanDevice& device = queue->getDevice();
  326. if(!mQueuedQueryResets.empty())
  327. {
  328. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  329. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  330. for (auto& entry : mQueuedQueryResets)
  331. entry->reset(vkCmdBuffer);
  332. cmdBuffer->end();
  333. queue->queueSubmit(cmdBuffer, nullptr, 0);
  334. mQueuedQueryResets.clear();
  335. }
  336. // Issue pipeline barriers for queue transitions (need to happen on original queue first, then on new queue)
  337. for (auto& entry : mBuffers)
  338. {
  339. VulkanBuffer* resource = static_cast<VulkanBuffer*>(entry.first);
  340. if (!resource->isExclusive())
  341. continue;
  342. UINT32 currentQueueFamily = resource->getQueueFamily();
  343. if (currentQueueFamily != -1 && currentQueueFamily != mQueueFamily)
  344. {
  345. Vector<VkBufferMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].bufferBarriers;
  346. barriers.push_back(VkBufferMemoryBarrier());
  347. VkBufferMemoryBarrier& barrier = barriers.back();
  348. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  349. barrier.pNext = nullptr;
  350. barrier.srcAccessMask = entry.second.accessFlags;
  351. barrier.dstAccessMask = entry.second.accessFlags;
  352. barrier.srcQueueFamilyIndex = currentQueueFamily;
  353. barrier.dstQueueFamilyIndex = mQueueFamily;
  354. barrier.buffer = resource->getHandle();
  355. barrier.offset = 0;
  356. barrier.size = VK_WHOLE_SIZE;
  357. }
  358. }
  359. // For images issue queue transitions, as above. Also issue layout transitions to their inital layouts.
  360. Vector<VkImageMemoryBarrier>& localBarriers = mTransitionInfoTemp[mQueueFamily].imageBarriers;
  361. for (auto& entry : mImages)
  362. {
  363. VulkanImage* resource = static_cast<VulkanImage*>(entry.first);
  364. ImageInfo& imageInfo = mImageInfos[entry.second];
  365. UINT32 currentQueueFamily = resource->getQueueFamily();
  366. bool queueMismatch = resource->isExclusive() && currentQueueFamily != -1 && currentQueueFamily != mQueueFamily;
  367. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  368. if (queueMismatch)
  369. {
  370. Vector<VkImageMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].imageBarriers;
  371. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  372. {
  373. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  374. UINT32 startIdx = (UINT32)barriers.size();
  375. resource->getBarriers(subresourceInfo.range, barriers);
  376. for(UINT32 j = startIdx; j < (UINT32)barriers.size(); j++)
  377. {
  378. VkImageMemoryBarrier& barrier = barriers[j];
  379. barrier.dstAccessMask = resource->getAccessFlags(barrier.oldLayout);
  380. barrier.newLayout = barrier.oldLayout;
  381. barrier.srcQueueFamilyIndex = currentQueueFamily;
  382. barrier.dstQueueFamilyIndex = mQueueFamily;
  383. }
  384. }
  385. }
  386. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  387. {
  388. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  389. VkImageLayout initialLayout = subresourceInfo.initialLayout;
  390. if (initialLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  391. continue;
  392. const VkImageSubresourceRange& range = subresourceInfo.range;
  393. UINT32 mipEnd = range.baseMipLevel + range.levelCount;
  394. UINT32 faceEnd = range.baseArrayLayer + range.layerCount;
  395. bool layoutMismatch = false;
  396. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  397. {
  398. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  399. {
  400. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  401. if(subresource->getLayout() != initialLayout)
  402. {
  403. layoutMismatch = true;
  404. break;
  405. }
  406. }
  407. if (layoutMismatch)
  408. break;
  409. }
  410. if(layoutMismatch)
  411. {
  412. UINT32 startIdx = (UINT32)localBarriers.size();
  413. resource->getBarriers(subresourceInfo.range, localBarriers);
  414. for (UINT32 j = startIdx; j < (UINT32)localBarriers.size(); j++)
  415. {
  416. VkImageMemoryBarrier& barrier = localBarriers[j];
  417. barrier.dstAccessMask = resource->getAccessFlags(initialLayout, subresourceInfo.isInitialReadOnly);
  418. barrier.newLayout = initialLayout;
  419. }
  420. }
  421. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  422. {
  423. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  424. {
  425. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  426. subresource->setLayout(subresourceInfo.finalLayout);
  427. }
  428. }
  429. }
  430. }
  431. for (auto& entry : mTransitionInfoTemp)
  432. {
  433. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  434. if (empty)
  435. continue;
  436. UINT32 entryQueueFamily = entry.first;
  437. // No queue transition needed for entries on this queue (this entry is most likely an image layout transition)
  438. if (entryQueueFamily == -1 || entryQueueFamily == mQueueFamily)
  439. continue;
  440. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(entryQueueFamily, false);
  441. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  442. TransitionInfo& barriers = entry.second;
  443. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  444. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  445. vkCmdPipelineBarrier(vkCmdBuffer,
  446. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // Note: VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT might be more correct here, according to the spec
  447. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // The main idea is that the barrier executes before the semaphore triggers, no actual stage dependencies are needed.
  448. 0, 0, nullptr,
  449. numBufferBarriers, barriers.bufferBarriers.data(),
  450. numImgBarriers, barriers.imageBarriers.data());
  451. // Find an appropriate queue to execute on
  452. UINT32 otherQueueIdx = 0;
  453. VulkanQueue* otherQueue = nullptr;
  454. GpuQueueType otherQueueType = GQT_GRAPHICS;
  455. for (UINT32 i = 0; i < GQT_COUNT; i++)
  456. {
  457. otherQueueType = (GpuQueueType)i;
  458. if (device.getQueueFamily(otherQueueType) != entryQueueFamily)
  459. continue;
  460. UINT32 numQueues = device.getNumQueues(otherQueueType);
  461. for (UINT32 j = 0; j < numQueues; j++)
  462. {
  463. // Try to find a queue not currently executing
  464. VulkanQueue* curQueue = device.getQueue(otherQueueType, j);
  465. if (!curQueue->isExecuting())
  466. {
  467. otherQueue = curQueue;
  468. otherQueueIdx = j;
  469. }
  470. }
  471. // Can't find empty one, use the first one then
  472. if (otherQueue == nullptr)
  473. {
  474. otherQueue = device.getQueue(otherQueueType, 0);
  475. otherQueueIdx = 0;
  476. }
  477. break;
  478. }
  479. syncMask |= CommandSyncMask::getGlobalQueueMask(otherQueueType, otherQueueIdx);
  480. cmdBuffer->end();
  481. // Note: If I switch back to doing layout transitions here, I need to wait on present semaphore
  482. otherQueue->submit(cmdBuffer, nullptr, 0);
  483. }
  484. UINT32 deviceIdx = device.getIndex();
  485. VulkanCommandBufferManager& cbm = static_cast<VulkanCommandBufferManager&>(CommandBufferManager::instance());
  486. UINT32 numSemaphores;
  487. cbm.getSyncSemaphores(deviceIdx, syncMask, mSemaphoresTemp.data(), numSemaphores);
  488. // Wait on present (i.e. until the back buffer becomes available) for any swap chains
  489. for(auto& entry : mSwapChains)
  490. {
  491. const SwapChainSurface& surface = entry->getBackBuffer();
  492. if (surface.needsWait)
  493. {
  494. VulkanSemaphore* semaphore = entry->getBackBuffer().sync;
  495. if (numSemaphores >= (UINT32)mSemaphoresTemp.size())
  496. mSemaphoresTemp.push_back(semaphore);
  497. else
  498. mSemaphoresTemp[numSemaphores] = semaphore;
  499. numSemaphores++;
  500. entry->notifyBackBufferWaitIssued();
  501. }
  502. }
  503. // Issue second part of transition pipeline barriers (on this queue)
  504. for (auto& entry : mTransitionInfoTemp)
  505. {
  506. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  507. if (empty)
  508. continue;
  509. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  510. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  511. TransitionInfo& barriers = entry.second;
  512. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  513. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  514. vkCmdPipelineBarrier(vkCmdBuffer,
  515. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // Note: VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT might be more correct here, according to the spec
  516. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT,
  517. 0, 0, nullptr,
  518. numBufferBarriers, barriers.bufferBarriers.data(),
  519. numImgBarriers, barriers.imageBarriers.data());
  520. cmdBuffer->end();
  521. queue->queueSubmit(cmdBuffer, mSemaphoresTemp.data(), numSemaphores);
  522. numSemaphores = 0; // Semaphores are only needed the first time, since we're adding the buffers on the same queue
  523. }
  524. queue->queueSubmit(this, mSemaphoresTemp.data(), numSemaphores);
  525. queue->submitQueued();
  526. mGlobalQueueIdx = CommandSyncMask::getGlobalQueueIdx(queue->getType(), queueIdx);
  527. for (auto& entry : mResources)
  528. {
  529. ResourceUseHandle& useHandle = entry.second;
  530. assert(!useHandle.used);
  531. useHandle.used = true;
  532. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  533. }
  534. for (auto& entry : mImages)
  535. {
  536. UINT32 imageInfoIdx = entry.second;
  537. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  538. ResourceUseHandle& useHandle = imageInfo.useHandle;
  539. assert(!useHandle.used);
  540. useHandle.used = true;
  541. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  542. }
  543. for (auto& entry : mBuffers)
  544. {
  545. ResourceUseHandle& useHandle = entry.second.useHandle;
  546. assert(!useHandle.used);
  547. useHandle.used = true;
  548. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  549. }
  550. // Note: Uncomment for debugging only, prevents any device concurrency issues.
  551. // vkQueueWaitIdle(queue->getHandle());
  552. // Clear vectors but don't clear the actual map, as we want to re-use the memory since we expect queue family
  553. // indices to be the same
  554. for (auto& entry : mTransitionInfoTemp)
  555. {
  556. entry.second.imageBarriers.clear();
  557. entry.second.bufferBarriers.clear();
  558. }
  559. mGraphicsPipeline = nullptr;
  560. mComputePipeline = nullptr;
  561. mGfxPipelineRequiresBind = true;
  562. mCmpPipelineRequiresBind = true;
  563. mFramebuffer = nullptr;
  564. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  565. mQueuedLayoutTransitions.clear();
  566. mBoundParams = nullptr;
  567. mBoundParams = false;
  568. mSwapChains.clear();
  569. }
  570. bool VulkanCmdBuffer::checkFenceStatus(bool block) const
  571. {
  572. VkResult result = vkWaitForFences(mDevice.getLogical(), 1, &mFence, true, block ? 1'000'000'000 : 0);
  573. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  574. return result == VK_SUCCESS;
  575. }
  576. void VulkanCmdBuffer::reset()
  577. {
  578. bool wasSubmitted = mState == State::Submitted;
  579. mState = State::Ready;
  580. vkResetCommandBuffer(mCmdBuffer, VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT); // Note: Maybe better not to release resources?
  581. if (wasSubmitted)
  582. {
  583. for (auto& entry : mResources)
  584. {
  585. ResourceUseHandle& useHandle = entry.second;
  586. assert(useHandle.used);
  587. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  588. }
  589. for (auto& entry : mImages)
  590. {
  591. UINT32 imageInfoIdx = entry.second;
  592. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  593. ResourceUseHandle& useHandle = imageInfo.useHandle;
  594. assert(useHandle.used);
  595. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  596. }
  597. for (auto& entry : mBuffers)
  598. {
  599. ResourceUseHandle& useHandle = entry.second.useHandle;
  600. assert(useHandle.used);
  601. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  602. }
  603. }
  604. else
  605. {
  606. for (auto& entry : mResources)
  607. entry.first->notifyUnbound();
  608. for (auto& entry : mImages)
  609. entry.first->notifyUnbound();
  610. for (auto& entry : mBuffers)
  611. entry.first->notifyUnbound();
  612. }
  613. mResources.clear();
  614. mImages.clear();
  615. mBuffers.clear();
  616. mOcclusionQueries.clear();
  617. mTimerQueries.clear();
  618. mImageInfos.clear();
  619. mSubresourceInfoStorage.clear();
  620. mPassTouchedSubresourceInfos.clear();
  621. }
  622. void VulkanCmdBuffer::setRenderTarget(const SPtr<RenderTarget>& rt, UINT32 readOnlyFlags,
  623. RenderSurfaceMask loadMask)
  624. {
  625. assert(mState != State::Submitted);
  626. VulkanFramebuffer* newFB;
  627. if(rt != nullptr)
  628. {
  629. if (rt->getProperties().isWindow())
  630. {
  631. Win32RenderWindow* window = static_cast<Win32RenderWindow*>(rt.get());
  632. window->acquireBackBuffer();
  633. VulkanSwapChain* swapChain;
  634. rt->getCustomAttribute("SC", &swapChain);
  635. mSwapChains.insert(swapChain);
  636. }
  637. rt->getCustomAttribute("FB", &newFB);
  638. }
  639. else
  640. {
  641. newFB = nullptr;
  642. }
  643. // Warn if invalid load mask
  644. if (loadMask.isSet(RT_DEPTH) && !loadMask.isSet(RT_STENCIL))
  645. {
  646. LOGWRN("setRenderTarget() invalid load mask, depth enabled but stencil disabled. This is not supported. Both \
  647. will be loaded.");
  648. loadMask.set(RT_STENCIL);
  649. }
  650. if (!loadMask.isSet(RT_DEPTH) && loadMask.isSet(RT_STENCIL))
  651. {
  652. LOGWRN("setRenderTarget() invalid load mask, stencil enabled but depth disabled. This is not supported. Both \
  653. will be loaded.");
  654. loadMask.set(RT_DEPTH);
  655. }
  656. if (mFramebuffer == newFB && mRenderTargetReadOnlyFlags == readOnlyFlags && mRenderTargetLoadMask == loadMask)
  657. return;
  658. if (isInRenderPass())
  659. endRenderPass();
  660. else
  661. {
  662. // If a clear is queued for previous FB, execute the render pass with no additional instructions
  663. if (mClearMask)
  664. executeClearPass();
  665. }
  666. // Reset isFBAttachment flags for subresources from the old framebuffer
  667. if(mFramebuffer != nullptr)
  668. {
  669. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  670. for(UINT32 i = 0; i < numColorAttachments; i++)
  671. {
  672. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  673. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  674. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  675. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  676. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  677. {
  678. ImageSubresourceInfo& entry = subresourceInfos[j];
  679. entry.isFBAttachment = false;
  680. }
  681. }
  682. if(mFramebuffer->hasDepthAttachment())
  683. {
  684. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  685. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  686. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  687. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  688. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  689. {
  690. ImageSubresourceInfo& entry = subresourceInfos[j];
  691. entry.isFBAttachment = false;
  692. }
  693. }
  694. }
  695. if(newFB == nullptr)
  696. {
  697. mFramebuffer = nullptr;
  698. mRenderTargetWidth = 0;
  699. mRenderTargetHeight = 0;
  700. mRenderTargetReadOnlyFlags = 0;
  701. mRenderTargetLoadMask = RT_NONE;
  702. }
  703. else
  704. {
  705. mFramebuffer = newFB;
  706. mRenderTargetWidth = rt->getProperties().getWidth();
  707. mRenderTargetHeight = rt->getProperties().getHeight();
  708. mRenderTargetReadOnlyFlags = readOnlyFlags;
  709. mRenderTargetLoadMask = loadMask;
  710. }
  711. setGpuParams(nullptr);
  712. if (mFramebuffer != nullptr)
  713. registerResource(mFramebuffer, loadMask, readOnlyFlags);
  714. mGfxPipelineRequiresBind = true;
  715. }
  716. void VulkanCmdBuffer::clearViewport(const Rect2I& area, UINT32 buffers, const Color& color, float depth, UINT16 stencil,
  717. UINT8 targetMask)
  718. {
  719. if (buffers == 0 || mFramebuffer == nullptr)
  720. return;
  721. // Add clear command if currently in render pass
  722. if (isInRenderPass())
  723. {
  724. VkClearAttachment attachments[BS_MAX_MULTIPLE_RENDER_TARGETS + 1];
  725. UINT32 baseLayer = 0;
  726. UINT32 attachmentIdx = 0;
  727. if ((buffers & FBT_COLOR) != 0)
  728. {
  729. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  730. for (UINT32 i = 0; i < numColorAttachments; i++)
  731. {
  732. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  733. if (((1 << attachment.index) & targetMask) == 0)
  734. continue;
  735. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  736. attachments[attachmentIdx].colorAttachment = i;
  737. VkClearColorValue& colorValue = attachments[attachmentIdx].clearValue.color;
  738. colorValue.float32[0] = color.r;
  739. colorValue.float32[1] = color.g;
  740. colorValue.float32[2] = color.b;
  741. colorValue.float32[3] = color.a;
  742. UINT32 curBaseLayer = attachment.baseLayer;
  743. if (attachmentIdx == 0)
  744. baseLayer = curBaseLayer;
  745. else
  746. {
  747. if (baseLayer != curBaseLayer)
  748. {
  749. // Note: This could be supported relatively easily: we would need to issue multiple separate
  750. // clear commands for such framebuffers.
  751. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  752. "starting layers. This is currently not supported.");
  753. }
  754. }
  755. attachmentIdx++;
  756. }
  757. }
  758. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  759. {
  760. if (mFramebuffer->hasDepthAttachment())
  761. {
  762. attachments[attachmentIdx].aspectMask = 0;
  763. if ((buffers & FBT_DEPTH) != 0)
  764. {
  765. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_DEPTH_BIT;
  766. attachments[attachmentIdx].clearValue.depthStencil.depth = depth;
  767. }
  768. if ((buffers & FBT_STENCIL) != 0)
  769. {
  770. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_STENCIL_BIT;
  771. attachments[attachmentIdx].clearValue.depthStencil.stencil = stencil;
  772. }
  773. attachments[attachmentIdx].colorAttachment = 0;
  774. UINT32 curBaseLayer = mFramebuffer->getDepthStencilAttachment().baseLayer;
  775. if (attachmentIdx == 0)
  776. baseLayer = curBaseLayer;
  777. else
  778. {
  779. if (baseLayer != curBaseLayer)
  780. {
  781. // Note: This could be supported relatively easily: we would need to issue multiple separate
  782. // clear commands for such framebuffers.
  783. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  784. "starting layers. This is currently not supported.");
  785. }
  786. }
  787. attachmentIdx++;
  788. }
  789. }
  790. UINT32 numAttachments = attachmentIdx;
  791. if (numAttachments == 0)
  792. return;
  793. VkClearRect clearRect;
  794. clearRect.baseArrayLayer = baseLayer;
  795. clearRect.layerCount = mFramebuffer->getNumLayers();
  796. clearRect.rect.offset.x = area.x;
  797. clearRect.rect.offset.y = area.y;
  798. clearRect.rect.extent.width = area.width;
  799. clearRect.rect.extent.height = area.height;
  800. vkCmdClearAttachments(mCmdBuffer, numAttachments, attachments, 1, &clearRect);
  801. }
  802. // Otherwise we use a render pass that performs a clear on begin
  803. else
  804. {
  805. ClearMask clearMask;
  806. std::array<VkClearValue, BS_MAX_MULTIPLE_RENDER_TARGETS + 1> clearValues = mClearValues;
  807. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  808. if ((buffers & FBT_COLOR) != 0)
  809. {
  810. for (UINT32 i = 0; i < numColorAttachments; i++)
  811. {
  812. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  813. if (((1 << attachment.index) & targetMask) == 0)
  814. continue;
  815. clearMask |= (ClearMaskBits)(1 << attachment.index);
  816. VkClearColorValue& colorValue = clearValues[i].color;
  817. colorValue.float32[0] = color.r;
  818. colorValue.float32[1] = color.g;
  819. colorValue.float32[2] = color.b;
  820. colorValue.float32[3] = color.a;
  821. }
  822. }
  823. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  824. {
  825. if (mFramebuffer->hasDepthAttachment())
  826. {
  827. UINT32 depthAttachmentIdx = numColorAttachments;
  828. if ((buffers & FBT_DEPTH) != 0)
  829. {
  830. clearValues[depthAttachmentIdx].depthStencil.depth = depth;
  831. clearMask |= CLEAR_DEPTH;
  832. }
  833. if ((buffers & FBT_STENCIL) != 0)
  834. {
  835. clearValues[depthAttachmentIdx].depthStencil.stencil = stencil;
  836. clearMask |= CLEAR_STENCIL;
  837. }
  838. }
  839. }
  840. if (!clearMask)
  841. return;
  842. // Some previous clear operation is already queued, execute it first
  843. bool previousClearNeedsToFinish = (mClearMask & clearMask) != CLEAR_NONE;
  844. if(previousClearNeedsToFinish)
  845. executeClearPass();
  846. mClearMask |= clearMask;
  847. mClearValues = clearValues;
  848. mClearArea = area;
  849. }
  850. }
  851. void VulkanCmdBuffer::clearRenderTarget(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  852. {
  853. Rect2I area(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  854. clearViewport(area, buffers, color, depth, stencil, targetMask);
  855. }
  856. void VulkanCmdBuffer::clearViewport(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  857. {
  858. Rect2I area;
  859. area.x = (UINT32)(mViewport.x * mRenderTargetWidth);
  860. area.y = (UINT32)(mViewport.y * mRenderTargetHeight);
  861. area.width = (UINT32)(mViewport.width * mRenderTargetWidth);
  862. area.height = (UINT32)(mViewport.height * mRenderTargetHeight);
  863. clearViewport(area, buffers, color, depth, stencil, targetMask);
  864. }
  865. void VulkanCmdBuffer::setPipelineState(const SPtr<GraphicsPipelineState>& state)
  866. {
  867. if (mGraphicsPipeline == state)
  868. return;
  869. mGraphicsPipeline = std::static_pointer_cast<VulkanGraphicsPipelineState>(state);
  870. mGfxPipelineRequiresBind = true;
  871. }
  872. void VulkanCmdBuffer::setPipelineState(const SPtr<ComputePipelineState>& state)
  873. {
  874. if (mComputePipeline == state)
  875. return;
  876. mComputePipeline = std::static_pointer_cast<VulkanComputePipelineState>(state);
  877. mCmpPipelineRequiresBind = true;
  878. }
  879. void VulkanCmdBuffer::setGpuParams(const SPtr<GpuParams>& gpuParams)
  880. {
  881. // Note: We keep an internal reference to GPU params even though we shouldn't keep a reference to a core thread
  882. // object. But it should be fine since we expect the resource to be externally synchronized so it should never
  883. // be allowed to go out of scope on a non-core thread anyway.
  884. mBoundParams = std::static_pointer_cast<VulkanGpuParams>(gpuParams);
  885. if (mBoundParams != nullptr)
  886. mBoundParamsDirty = true;
  887. else
  888. {
  889. mNumBoundDescriptorSets = 0;
  890. mBoundParamsDirty = false;
  891. }
  892. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  893. }
  894. void VulkanCmdBuffer::setViewport(const Rect2& area)
  895. {
  896. if (mViewport == area)
  897. return;
  898. mViewport = area;
  899. mViewportRequiresBind = true;
  900. }
  901. void VulkanCmdBuffer::setScissorRect(const Rect2I& value)
  902. {
  903. if (mScissor == value)
  904. return;
  905. mScissor = value;
  906. mScissorRequiresBind = true;
  907. }
  908. void VulkanCmdBuffer::setStencilRef(UINT32 value)
  909. {
  910. if (mStencilRef == value)
  911. return;
  912. mStencilRef = value;
  913. mStencilRefRequiresBind = true;
  914. }
  915. void VulkanCmdBuffer::setDrawOp(DrawOperationType drawOp)
  916. {
  917. if (mDrawOp == drawOp)
  918. return;
  919. mDrawOp = drawOp;
  920. mGfxPipelineRequiresBind = true;
  921. }
  922. void VulkanCmdBuffer::setVertexBuffers(UINT32 index, SPtr<VertexBuffer>* buffers, UINT32 numBuffers)
  923. {
  924. if (numBuffers == 0)
  925. return;
  926. for(UINT32 i = 0; i < numBuffers; i++)
  927. {
  928. VulkanVertexBuffer* vertexBuffer = static_cast<VulkanVertexBuffer*>(buffers[i].get());
  929. if (vertexBuffer != nullptr)
  930. {
  931. VulkanBuffer* resource = vertexBuffer->getResource(mDevice.getIndex());
  932. if (resource != nullptr)
  933. {
  934. mVertexBuffersTemp[i] = resource->getHandle();
  935. registerResource(resource, VulkanUseFlag::Read);
  936. }
  937. else
  938. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  939. }
  940. else
  941. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  942. }
  943. vkCmdBindVertexBuffers(mCmdBuffer, index, numBuffers, mVertexBuffersTemp, mVertexBufferOffsetsTemp);
  944. }
  945. void VulkanCmdBuffer::setIndexBuffer(const SPtr<IndexBuffer>& buffer)
  946. {
  947. VulkanIndexBuffer* indexBuffer = static_cast<VulkanIndexBuffer*>(buffer.get());
  948. VkBuffer vkBuffer = VK_NULL_HANDLE;
  949. VkIndexType indexType = VK_INDEX_TYPE_UINT32;
  950. if (indexBuffer != nullptr)
  951. {
  952. VulkanBuffer* resource = indexBuffer->getResource(mDevice.getIndex());
  953. if (resource != nullptr)
  954. {
  955. vkBuffer = resource->getHandle();
  956. indexType = VulkanUtility::getIndexType(buffer->getProperties().getType());
  957. registerResource(resource, VulkanUseFlag::Read);
  958. }
  959. }
  960. vkCmdBindIndexBuffer(mCmdBuffer, vkBuffer, 0, indexType);
  961. }
  962. void VulkanCmdBuffer::setVertexDeclaration(const SPtr<VertexDeclaration>& decl)
  963. {
  964. if (mVertexDecl == decl)
  965. return;
  966. mVertexDecl = decl;
  967. mGfxPipelineRequiresBind = true;
  968. }
  969. bool VulkanCmdBuffer::isReadyForRender()
  970. {
  971. if (mGraphicsPipeline == nullptr)
  972. return false;
  973. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  974. if (inputDecl == nullptr)
  975. return false;
  976. return mFramebuffer != nullptr && mVertexDecl != nullptr;
  977. }
  978. bool VulkanCmdBuffer::bindGraphicsPipeline()
  979. {
  980. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  981. SPtr<VulkanVertexInput> vertexInput = VulkanVertexInputManager::instance().getVertexInfo(mVertexDecl, inputDecl);
  982. VulkanPipeline* pipeline = mGraphicsPipeline->getPipeline(mDevice.getIndex(), mFramebuffer,
  983. mRenderTargetReadOnlyFlags, mDrawOp, vertexInput);
  984. if (pipeline == nullptr)
  985. return false;
  986. // Check that pipeline matches the read-only state of any framebuffer attachments
  987. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  988. for (UINT32 i = 0; i < numColorAttachments; i++)
  989. {
  990. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  991. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  992. fbAttachment.surface.mipLevel);
  993. if (subresourceInfo.isShaderInput && !pipeline->isColorReadOnly(i))
  994. {
  995. LOGWRN("Framebuffer attachment also used as a shader input, but color writes aren't disabled. This will"
  996. " result in undefined behavior.");
  997. }
  998. }
  999. if (mFramebuffer->hasDepthAttachment())
  1000. {
  1001. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1002. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  1003. fbAttachment.surface.mipLevel);
  1004. if (subresourceInfo.isShaderInput && !pipeline->isDepthReadOnly())
  1005. {
  1006. LOGWRN("Framebuffer attachment also used as a shader input, but depth/stencil writes aren't disabled. "
  1007. "This will result in undefined behavior.");
  1008. }
  1009. }
  1010. mGraphicsPipeline->registerPipelineResources(this);
  1011. registerResource(pipeline, VulkanUseFlag::Read);
  1012. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline->getHandle());
  1013. bindDynamicStates(true);
  1014. mGfxPipelineRequiresBind = false;
  1015. return true;
  1016. }
  1017. void VulkanCmdBuffer::bindDynamicStates(bool forceAll)
  1018. {
  1019. if (mViewportRequiresBind || forceAll)
  1020. {
  1021. VkViewport viewport;
  1022. viewport.x = mViewport.x * mRenderTargetWidth;
  1023. viewport.y = mViewport.y * mRenderTargetHeight;
  1024. viewport.width = mViewport.width * mRenderTargetWidth;
  1025. viewport.height = mViewport.height * mRenderTargetHeight;
  1026. viewport.minDepth = 0.0f;
  1027. viewport.maxDepth = 1.0f;
  1028. vkCmdSetViewport(mCmdBuffer, 0, 1, &viewport);
  1029. mViewportRequiresBind = false;
  1030. }
  1031. if(mStencilRefRequiresBind || forceAll)
  1032. {
  1033. vkCmdSetStencilReference(mCmdBuffer, VK_STENCIL_FRONT_AND_BACK, mStencilRef);
  1034. mStencilRefRequiresBind = false;
  1035. }
  1036. if(mScissorRequiresBind || forceAll)
  1037. {
  1038. VkRect2D scissorRect;
  1039. if(mGraphicsPipeline->isScissorEnabled())
  1040. {
  1041. scissorRect.offset.x = mScissor.x;
  1042. scissorRect.offset.y = mScissor.y;
  1043. scissorRect.extent.width = mScissor.width;
  1044. scissorRect.extent.height = mScissor.height;
  1045. }
  1046. else
  1047. {
  1048. scissorRect.offset.x = 0;
  1049. scissorRect.offset.y = 0;
  1050. scissorRect.extent.width = mRenderTargetWidth;
  1051. scissorRect.extent.height = mRenderTargetHeight;
  1052. }
  1053. vkCmdSetScissor(mCmdBuffer, 0, 1, &scissorRect);
  1054. mScissorRequiresBind = false;
  1055. }
  1056. }
  1057. void VulkanCmdBuffer::bindGpuParams()
  1058. {
  1059. if (mBoundParamsDirty)
  1060. {
  1061. if (mBoundParams != nullptr)
  1062. {
  1063. mNumBoundDescriptorSets = mBoundParams->getNumSets();
  1064. mBoundParams->prepareForBind(*this, mDescriptorSetsTemp);
  1065. }
  1066. else
  1067. mNumBoundDescriptorSets = 0;
  1068. mBoundParamsDirty = false;
  1069. }
  1070. else
  1071. {
  1072. mNumBoundDescriptorSets = 0;
  1073. }
  1074. }
  1075. void VulkanCmdBuffer::executeLayoutTransitions()
  1076. {
  1077. auto createLayoutTransitionBarrier = [&](VulkanImage* image, ImageInfo& imageInfo)
  1078. {
  1079. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1080. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1081. {
  1082. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  1083. if (!subresourceInfo.hasTransitioned || subresourceInfo.currentLayout == subresourceInfo.requiredLayout)
  1084. continue;
  1085. mLayoutTransitionBarriersTemp.push_back(VkImageMemoryBarrier());
  1086. VkImageMemoryBarrier& barrier = mLayoutTransitionBarriersTemp.back();
  1087. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1088. barrier.pNext = nullptr;
  1089. barrier.srcAccessMask = image->getAccessFlags(subresourceInfo.currentLayout);
  1090. barrier.dstAccessMask = image->getAccessFlags(subresourceInfo.requiredLayout, subresourceInfo.isReadOnly);
  1091. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1092. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1093. barrier.oldLayout = subresourceInfo.currentLayout;
  1094. barrier.newLayout = subresourceInfo.requiredLayout;
  1095. barrier.image = image->getHandle();
  1096. barrier.subresourceRange = subresourceInfo.range;
  1097. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1098. subresourceInfo.isReadOnly = true;
  1099. subresourceInfo.hasTransitioned = true;
  1100. }
  1101. };
  1102. // Note: These layout transitions will contain transitions for offscreen framebuffer attachments (while they
  1103. // transition to shader read-only layout). This can be avoided, since they're immediately used by the render pass
  1104. // as color attachments, making the layout change redundant.
  1105. for (auto& entry : mQueuedLayoutTransitions)
  1106. {
  1107. UINT32 imageInfoIdx = entry.second;
  1108. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1109. createLayoutTransitionBarrier(entry.first, imageInfo);
  1110. }
  1111. vkCmdPipelineBarrier(mCmdBuffer,
  1112. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, // Note: VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT might be more correct here, according to the spec
  1113. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT,
  1114. 0, 0, nullptr,
  1115. 0, nullptr,
  1116. (UINT32)mLayoutTransitionBarriersTemp.size(), mLayoutTransitionBarriersTemp.data());
  1117. mQueuedLayoutTransitions.clear();
  1118. mLayoutTransitionBarriersTemp.clear();
  1119. }
  1120. void VulkanCmdBuffer::updateFinalLayouts()
  1121. {
  1122. if (mFramebuffer == nullptr)
  1123. return;
  1124. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1125. for (UINT32 i = 0; i < numColorAttachments; i++)
  1126. {
  1127. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1128. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  1129. fbAttachment.surface.mipLevel);
  1130. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1131. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1132. subresourceInfo.hasTransitioned = true;
  1133. }
  1134. if (mFramebuffer->hasDepthAttachment())
  1135. {
  1136. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1137. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.arraySlice,
  1138. fbAttachment.surface.mipLevel);
  1139. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1140. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1141. subresourceInfo.hasTransitioned = true;
  1142. }
  1143. }
  1144. void VulkanCmdBuffer::executeClearPass()
  1145. {
  1146. assert(mState == State::Recording);
  1147. executeLayoutTransitions();
  1148. VkRenderPassBeginInfo renderPassBeginInfo;
  1149. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  1150. renderPassBeginInfo.pNext = nullptr;
  1151. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(RT_NONE, RT_NONE, mClearMask);
  1152. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(RT_NONE, RT_NONE, mClearMask);
  1153. renderPassBeginInfo.renderArea.offset.x = mClearArea.x;
  1154. renderPassBeginInfo.renderArea.offset.y = mClearArea.y;
  1155. renderPassBeginInfo.renderArea.extent.width = mClearArea.width;
  1156. renderPassBeginInfo.renderArea.extent.height = mClearArea.height;
  1157. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  1158. renderPassBeginInfo.pClearValues = mClearValues.data();
  1159. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  1160. vkCmdEndRenderPass(mCmdBuffer);
  1161. updateFinalLayouts();
  1162. mClearMask = CLEAR_NONE;
  1163. }
  1164. void VulkanCmdBuffer::draw(UINT32 vertexOffset, UINT32 vertexCount, UINT32 instanceCount)
  1165. {
  1166. if (!isReadyForRender())
  1167. return;
  1168. // Note: Must begin render pass before binding GPU params as some GPU param related data gets cleared on begin, and
  1169. // we don't want to clear the currently used one
  1170. if (!isInRenderPass())
  1171. beginRenderPass();
  1172. bindGpuParams();
  1173. if (mGfxPipelineRequiresBind)
  1174. {
  1175. if (!bindGraphicsPipeline())
  1176. return;
  1177. }
  1178. else
  1179. bindDynamicStates(false);
  1180. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1181. {
  1182. if (mNumBoundDescriptorSets > 0)
  1183. {
  1184. UINT32 deviceIdx = mDevice.getIndex();
  1185. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1186. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1187. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1188. }
  1189. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1190. }
  1191. if (instanceCount <= 0)
  1192. instanceCount = 1;
  1193. vkCmdDraw(mCmdBuffer, vertexCount, instanceCount, vertexOffset, 0);
  1194. }
  1195. void VulkanCmdBuffer::drawIndexed(UINT32 startIndex, UINT32 indexCount, UINT32 vertexOffset, UINT32 instanceCount)
  1196. {
  1197. if (!isReadyForRender())
  1198. return;
  1199. // Note: Must begin render pass before binding GPU params as some GPU param related data gets cleared on begin, and
  1200. // we don't want to clear the currently used one
  1201. if (!isInRenderPass())
  1202. beginRenderPass();
  1203. bindGpuParams();
  1204. if (mGfxPipelineRequiresBind)
  1205. {
  1206. if (!bindGraphicsPipeline())
  1207. return;
  1208. }
  1209. else
  1210. bindDynamicStates(false);
  1211. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1212. {
  1213. if (mNumBoundDescriptorSets > 0)
  1214. {
  1215. UINT32 deviceIdx = mDevice.getIndex();
  1216. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1217. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1218. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1219. }
  1220. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1221. }
  1222. if (instanceCount <= 0)
  1223. instanceCount = 1;
  1224. vkCmdDrawIndexed(mCmdBuffer, indexCount, instanceCount, startIndex, vertexOffset, 0);
  1225. }
  1226. void VulkanCmdBuffer::dispatch(UINT32 numGroupsX, UINT32 numGroupsY, UINT32 numGroupsZ)
  1227. {
  1228. if (mComputePipeline == nullptr)
  1229. return;
  1230. if (isInRenderPass())
  1231. endRenderPass();
  1232. bindGpuParams();
  1233. executeLayoutTransitions();
  1234. UINT32 deviceIdx = mDevice.getIndex();
  1235. if(mCmpPipelineRequiresBind)
  1236. {
  1237. VulkanPipeline* pipeline = mComputePipeline->getPipeline(deviceIdx);
  1238. if (pipeline == nullptr)
  1239. return;
  1240. registerResource(pipeline, VulkanUseFlag::Read);
  1241. mComputePipeline->registerPipelineResources(this);
  1242. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline->getHandle());
  1243. mCmpPipelineRequiresBind = false;
  1244. }
  1245. if(mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Compute))
  1246. {
  1247. if (mNumBoundDescriptorSets > 0)
  1248. {
  1249. VkPipelineLayout pipelineLayout = mComputePipeline->getPipelineLayout(deviceIdx);
  1250. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipelineLayout, 0,
  1251. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1252. }
  1253. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Compute);
  1254. }
  1255. vkCmdDispatch(mCmdBuffer, numGroupsX, numGroupsY, numGroupsZ);
  1256. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  1257. // and reset read-only state.
  1258. for(auto& entry : mPassTouchedSubresourceInfos)
  1259. {
  1260. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  1261. subresourceInfo.isShaderInput = false;
  1262. subresourceInfo.isReadOnly = true;
  1263. }
  1264. mPassTouchedSubresourceInfos.clear();
  1265. }
  1266. void VulkanCmdBuffer::setEvent(VulkanEvent* event)
  1267. {
  1268. if(isInRenderPass())
  1269. mQueuedEvents.push_back(event);
  1270. else
  1271. vkCmdSetEvent(mCmdBuffer, event->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  1272. }
  1273. void VulkanCmdBuffer::resetQuery(VulkanQuery* query)
  1274. {
  1275. if (isInRenderPass())
  1276. mQueuedQueryResets.push_back(query);
  1277. else
  1278. query->reset(mCmdBuffer);
  1279. }
  1280. void VulkanCmdBuffer::memoryBarrier(VkBuffer buffer, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1281. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage)
  1282. {
  1283. VkBufferMemoryBarrier barrier;
  1284. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  1285. barrier.pNext = nullptr;
  1286. barrier.srcAccessMask = srcAccessFlags;
  1287. barrier.dstAccessMask = dstAccessFlags;
  1288. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1289. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1290. barrier.buffer = buffer;
  1291. barrier.offset = 0;
  1292. barrier.size = VK_WHOLE_SIZE;
  1293. vkCmdPipelineBarrier(getHandle(),
  1294. srcStage,
  1295. dstStage,
  1296. 0, 0, nullptr,
  1297. 1, &barrier,
  1298. 0, nullptr);
  1299. }
  1300. void VulkanCmdBuffer::memoryBarrier(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1301. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage, VkImageLayout layout,
  1302. const VkImageSubresourceRange& range)
  1303. {
  1304. VkImageMemoryBarrier barrier;
  1305. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1306. barrier.pNext = nullptr;
  1307. barrier.srcAccessMask = srcAccessFlags;
  1308. barrier.dstAccessMask = dstAccessFlags;
  1309. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1310. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1311. barrier.image = image;
  1312. barrier.subresourceRange = range;
  1313. barrier.oldLayout = layout;
  1314. barrier.newLayout = layout;
  1315. vkCmdPipelineBarrier(getHandle(),
  1316. srcStage,
  1317. dstStage,
  1318. 0, 0, nullptr,
  1319. 0, nullptr,
  1320. 1, &barrier);
  1321. }
  1322. void VulkanCmdBuffer::setLayout(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1323. VkImageLayout oldLayout, VkImageLayout newLayout, const VkImageSubresourceRange& range)
  1324. {
  1325. VkImageMemoryBarrier barrier;
  1326. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1327. barrier.pNext = nullptr;
  1328. barrier.srcAccessMask = srcAccessFlags;
  1329. barrier.dstAccessMask = dstAccessFlags;
  1330. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1331. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1332. barrier.oldLayout = oldLayout;
  1333. barrier.newLayout = newLayout;
  1334. barrier.image = image;
  1335. barrier.subresourceRange = range;
  1336. vkCmdPipelineBarrier(getHandle(),
  1337. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT,
  1338. VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT,
  1339. 0, 0, nullptr,
  1340. 0, nullptr,
  1341. 1, &barrier);
  1342. }
  1343. void VulkanCmdBuffer::registerResource(VulkanResource* res, VulkanUseFlags flags)
  1344. {
  1345. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1346. if(insertResult.second) // New element
  1347. {
  1348. ResourceUseHandle& useHandle = insertResult.first->second;
  1349. useHandle.used = false;
  1350. useHandle.flags = flags;
  1351. res->notifyBound();
  1352. }
  1353. else // Existing element
  1354. {
  1355. ResourceUseHandle& useHandle = insertResult.first->second;
  1356. assert(!useHandle.used);
  1357. useHandle.flags |= flags;
  1358. }
  1359. }
  1360. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range, VulkanUseFlags flags)
  1361. {
  1362. VkImageLayout layout = res->getOptimalLayout();
  1363. registerResource(res, range, VK_IMAGE_LAYOUT_UNDEFINED, layout, flags, false);
  1364. }
  1365. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range, VkImageLayout newLayout,
  1366. VkImageLayout finalLayout, VulkanUseFlags flags, bool isFBAttachment)
  1367. {
  1368. // Check if we're binding for shader use, or as a color attachment, or just for transfer purposes
  1369. bool isShaderBind = !isFBAttachment && newLayout != VK_IMAGE_LAYOUT_UNDEFINED;
  1370. // If binding it for write in a shader (not as color attachment or transfer op), we will need to issue a memory
  1371. // barrier if the image gets used again during this render pass, so remember this information.
  1372. bool needsBarrier = isShaderBind && flags.isSet(VulkanUseFlag::Write);
  1373. UINT32 nextImageInfoIdx = (UINT32)mImageInfos.size();
  1374. auto registerSubresourceInfo = [&](const VkImageSubresourceRange& subresourceRange)
  1375. {
  1376. mSubresourceInfoStorage.push_back(ImageSubresourceInfo());
  1377. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage.back();
  1378. subresourceInfo.currentLayout = newLayout;
  1379. subresourceInfo.initialLayout = newLayout;
  1380. subresourceInfo.requiredLayout = newLayout;
  1381. subresourceInfo.finalLayout = finalLayout;
  1382. subresourceInfo.range = subresourceRange;
  1383. subresourceInfo.isFBAttachment = isFBAttachment;
  1384. subresourceInfo.isShaderInput = !isFBAttachment;
  1385. subresourceInfo.hasTransitioned = false;
  1386. subresourceInfo.isReadOnly = !flags.isSet(VulkanUseFlag::Write);
  1387. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1388. subresourceInfo.needsBarrier = needsBarrier;
  1389. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1390. };
  1391. auto insertResult = mImages.insert(std::make_pair(res, nextImageInfoIdx));
  1392. if (insertResult.second) // New element
  1393. {
  1394. UINT32 imageInfoIdx = insertResult.first->second;
  1395. mImageInfos.push_back(ImageInfo());
  1396. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1397. imageInfo.subresourceInfoIdx = (UINT32)mSubresourceInfoStorage.size();
  1398. imageInfo.numSubresourceInfos = 1;
  1399. imageInfo.useHandle.used = false;
  1400. imageInfo.useHandle.flags = flags;
  1401. registerSubresourceInfo(range);
  1402. res->notifyBound();
  1403. }
  1404. else // Existing element
  1405. {
  1406. UINT32 imageInfoIdx = insertResult.first->second;
  1407. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1408. assert(!imageInfo.useHandle.used);
  1409. imageInfo.useHandle.flags |= flags;
  1410. // See if there is an overlap between existing ranges and the new range. And if so break them up accordingly.
  1411. //// First test for the simplest and most common case (same range or no overlap) to avoid more complex
  1412. //// computations.
  1413. ImageSubresourceInfo* subresources = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1414. bool foundRange = false;
  1415. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1416. {
  1417. if(VulkanUtility::rangeOverlaps(subresources[i].range, range))
  1418. {
  1419. if (subresources[i].range.layerCount == range.layerCount &&
  1420. subresources[i].range.levelCount == range.levelCount &&
  1421. subresources[i].range.baseArrayLayer == range.baseArrayLayer &&
  1422. subresources[i].range.baseMipLevel == range.baseMipLevel)
  1423. {
  1424. // Just update existing range
  1425. updateSubresourceInfo(res, imageInfoIdx, subresources[i], newLayout, finalLayout, flags,
  1426. isFBAttachment);
  1427. mPassTouchedSubresourceInfos.insert(imageInfo.subresourceInfoIdx + i);
  1428. foundRange = true;
  1429. break;
  1430. }
  1431. break;
  1432. }
  1433. }
  1434. //// We'll need to update subresource ranges or add new ones. The hope is that this code is trigger VERY rarely
  1435. //// (for just a few specific textures per frame).
  1436. if (!foundRange)
  1437. {
  1438. std::array<VkImageSubresourceRange, 5> tempCutRanges;
  1439. bs_frame_mark();
  1440. {
  1441. // We orphan previously allocated memory (we reset it after submit() anyway)
  1442. UINT32 newSubresourceIdx = (UINT32)mSubresourceInfoStorage.size();
  1443. FrameVector<UINT32> cutOverlappingRanges;
  1444. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1445. {
  1446. UINT32 subresourceIdx = imageInfo.subresourceInfoIdx + i;
  1447. ImageSubresourceInfo& subresource = mSubresourceInfoStorage[subresourceIdx];
  1448. if (!VulkanUtility::rangeOverlaps(subresource.range, range))
  1449. {
  1450. // Just copy as is
  1451. mSubresourceInfoStorage.push_back(subresource);
  1452. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1453. }
  1454. else // Need to cut
  1455. {
  1456. UINT32 numCutRanges;
  1457. VulkanUtility::cutRange(subresource.range, range, tempCutRanges, numCutRanges);
  1458. for(UINT32 j = 0; j < numCutRanges; j++)
  1459. {
  1460. // Create a copy of the original subresource with the new range
  1461. ImageSubresourceInfo newInfo = mSubresourceInfoStorage[subresourceIdx];
  1462. newInfo.range = tempCutRanges[j];
  1463. if(VulkanUtility::rangeOverlaps(tempCutRanges[j], range))
  1464. {
  1465. // Update overlapping sub-resource range with new data from this range
  1466. updateSubresourceInfo(res, imageInfoIdx, newInfo, newLayout, finalLayout, flags,
  1467. isFBAttachment);
  1468. // Keep track of the overlapping ranges for later
  1469. cutOverlappingRanges.push_back((UINT32)mSubresourceInfoStorage.size());
  1470. }
  1471. mSubresourceInfoStorage.push_back(newInfo);
  1472. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1473. }
  1474. }
  1475. }
  1476. // Our range doesn't overlap with any existing ranges, so just add it
  1477. if(cutOverlappingRanges.size() == 0)
  1478. {
  1479. registerSubresourceInfo(range);
  1480. }
  1481. else // Search if overlapping ranges fully cover the requested range, and insert non-covered regions
  1482. {
  1483. FrameQueue<VkImageSubresourceRange> sourceRanges;
  1484. sourceRanges.push(range);
  1485. for(auto& entry : cutOverlappingRanges)
  1486. {
  1487. VkImageSubresourceRange& overlappingRange = mSubresourceInfoStorage[entry].range;
  1488. UINT32 numSourceRanges = (UINT32)sourceRanges.size();
  1489. for(UINT32 i = 0; i < numSourceRanges; i++)
  1490. {
  1491. VkImageSubresourceRange sourceRange = sourceRanges.front();
  1492. sourceRanges.pop();
  1493. UINT32 numCutRanges;
  1494. VulkanUtility::cutRange(sourceRange, overlappingRange, tempCutRanges, numCutRanges);
  1495. for(UINT32 j = 0; j < numCutRanges; j++)
  1496. {
  1497. // We only care about ranges outside of the ones we already covered
  1498. if(!VulkanUtility::rangeOverlaps(tempCutRanges[j], overlappingRange))
  1499. sourceRanges.push(tempCutRanges[j]);
  1500. }
  1501. }
  1502. }
  1503. // Any remaining range hasn't been covered yet
  1504. while(!sourceRanges.empty())
  1505. {
  1506. registerSubresourceInfo(sourceRanges.front());
  1507. sourceRanges.pop();
  1508. }
  1509. }
  1510. imageInfo.subresourceInfoIdx = newSubresourceIdx;
  1511. imageInfo.numSubresourceInfos = (UINT32)mSubresourceInfoStorage.size() - newSubresourceIdx;
  1512. }
  1513. bs_frame_clear();
  1514. }
  1515. }
  1516. // Register any sub-resources
  1517. for(UINT32 i = 0; i < range.layerCount; i++)
  1518. {
  1519. for(UINT32 j = 0; j < range.levelCount; j++)
  1520. {
  1521. UINT32 layer = range.baseArrayLayer + i;
  1522. UINT32 mipLevel = range.baseMipLevel + j;
  1523. registerResource(res->getSubresource(layer, mipLevel), flags);
  1524. }
  1525. }
  1526. }
  1527. void VulkanCmdBuffer::registerResource(VulkanBuffer* res, VkAccessFlags accessFlags, VulkanUseFlags flags)
  1528. {
  1529. bool isShaderWrite = (accessFlags & VK_ACCESS_SHADER_WRITE_BIT) != 0;
  1530. auto insertResult = mBuffers.insert(std::make_pair(res, BufferInfo()));
  1531. if (insertResult.second) // New element
  1532. {
  1533. BufferInfo& bufferInfo = insertResult.first->second;
  1534. bufferInfo.accessFlags = accessFlags;
  1535. bufferInfo.useHandle.used = false;
  1536. bufferInfo.useHandle.flags = flags;
  1537. // Any writes done on storage buffers will need explicit memory barriers (if read during the same pass) so
  1538. // we remember this, in case this buffers gets used later in this pass.
  1539. bufferInfo.needsBarrier = isShaderWrite;
  1540. res->notifyBound();
  1541. }
  1542. else // Existing element
  1543. {
  1544. BufferInfo& bufferInfo = insertResult.first->second;
  1545. assert(!bufferInfo.useHandle.used);
  1546. bufferInfo.useHandle.flags |= flags;
  1547. bufferInfo.accessFlags |= accessFlags;
  1548. // If the buffer was written to previously in this pass, and is now being used by a shader we need to issue
  1549. // a barrier to make those writes visible.
  1550. bool isShaderRead = (accessFlags & VK_ACCESS_SHADER_READ_BIT) != 0;
  1551. if(bufferInfo.needsBarrier && (isShaderRead || isShaderWrite))
  1552. {
  1553. VkPipelineStageFlags stages =
  1554. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1555. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1556. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1557. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1558. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1559. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1560. VkBuffer buffer = res->getHandle();
  1561. memoryBarrier(buffer, VK_ACCESS_SHADER_WRITE_BIT, accessFlags, stages, stages);
  1562. bufferInfo.needsBarrier = isShaderWrite;
  1563. }
  1564. }
  1565. }
  1566. void VulkanCmdBuffer::registerResource(VulkanFramebuffer* res, RenderSurfaceMask loadMask, UINT32 readMask)
  1567. {
  1568. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1569. if (insertResult.second) // New element
  1570. {
  1571. ResourceUseHandle& useHandle = insertResult.first->second;
  1572. useHandle.used = false;
  1573. useHandle.flags = VulkanUseFlag::Write;
  1574. res->notifyBound();
  1575. }
  1576. else // Existing element
  1577. {
  1578. ResourceUseHandle& useHandle = insertResult.first->second;
  1579. assert(!useHandle.used);
  1580. useHandle.flags |= VulkanUseFlag::Write;
  1581. }
  1582. // Register any sub-resources
  1583. UINT32 numColorAttachments = res->getNumColorAttachments();
  1584. for (UINT32 i = 0; i < numColorAttachments; i++)
  1585. {
  1586. const VulkanFramebufferAttachment& attachment = res->getColorAttachment(i);
  1587. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1588. VkImageLayout layout;
  1589. if (loadMask.isSet((RenderSurfaceMaskBits)(1 << i)))
  1590. {
  1591. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1592. // transition this to a different layout when it begins, but we handle that elsewhere
  1593. layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1594. }
  1595. else
  1596. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1597. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1598. registerResource(attachment.image, range, layout, attachment.finalLayout, VulkanUseFlag::Write, true);
  1599. }
  1600. if(res->hasDepthAttachment())
  1601. {
  1602. const VulkanFramebufferAttachment& attachment = res->getDepthStencilAttachment();
  1603. VulkanUseFlag useFlag = VulkanUseFlag::Write;
  1604. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1605. VkImageLayout layout;
  1606. if (loadMask.isSet(RT_DEPTH) || loadMask.isSet(RT_STENCIL)) // Can't load one without the other
  1607. {
  1608. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1609. // transition this to a different layout when it begins, but we handle that elsewhere
  1610. layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1611. }
  1612. else
  1613. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1614. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1615. registerResource(attachment.image, range, layout, attachment.finalLayout, useFlag, true);
  1616. }
  1617. }
  1618. bool VulkanCmdBuffer::updateSubresourceInfo(VulkanImage* image, UINT32 imageInfoIdx,
  1619. ImageSubresourceInfo& subresourceInfo, VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags,
  1620. bool isFBAttachment)
  1621. {
  1622. subresourceInfo.isReadOnly &= !flags.isSet(VulkanUseFlag::Write);
  1623. // New layout is valid, check for transitions (UNDEFINED signifies the caller doesn't want a layout transition)
  1624. if (newLayout != VK_IMAGE_LAYOUT_UNDEFINED)
  1625. {
  1626. // If layout transition was requested by framebuffer bind, respect it because render-pass will only accept a
  1627. // specific layout (in certain cases), and we have no choice.
  1628. // In the case when a FB attachment is also bound for shader reads, this will override the layout required for
  1629. // shader read (GENERAL or DEPTH_READ_ONLY), but that is fine because those transitions are handled
  1630. // automatically by render-pass layout transitions.
  1631. // Any other texture (non FB attachment) will only even be bound in a single layout and we can keep the one it
  1632. // was originally registered with.
  1633. if (isFBAttachment)
  1634. subresourceInfo.requiredLayout = newLayout;
  1635. else if (!subresourceInfo.isFBAttachment) // Layout transition is not being done on a FB image
  1636. {
  1637. // Check if the image had a layout previously assigned, and if so check if multiple different layouts
  1638. // were requested. In that case we wish to transfer the image to GENERAL layout.
  1639. bool firstUseInRenderPass = !subresourceInfo.isShaderInput && !subresourceInfo.isFBAttachment;
  1640. if (firstUseInRenderPass || subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  1641. subresourceInfo.requiredLayout = newLayout;
  1642. else if (subresourceInfo.requiredLayout != newLayout)
  1643. subresourceInfo.requiredLayout = VK_IMAGE_LAYOUT_GENERAL;
  1644. }
  1645. }
  1646. // If attached to FB, then the final layout is set by the FB (provided as layout param here), otherwise its
  1647. // the same as required layout
  1648. if (!isFBAttachment && !subresourceInfo.isFBAttachment)
  1649. subresourceInfo.finalLayout = subresourceInfo.requiredLayout;
  1650. else
  1651. {
  1652. if (isFBAttachment)
  1653. subresourceInfo.finalLayout = finalLayout;
  1654. }
  1655. // If we haven't done a layout transition yet, we can just overwrite the previously written values, and the
  1656. // transition will be handled as the first thing in submit(), otherwise we queue a non-initial transition
  1657. // below.
  1658. if (!subresourceInfo.hasTransitioned)
  1659. {
  1660. subresourceInfo.initialLayout = subresourceInfo.requiredLayout;
  1661. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1662. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1663. }
  1664. else
  1665. {
  1666. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1667. mQueuedLayoutTransitions[image] = imageInfoIdx;
  1668. }
  1669. // If a FB attachment was just bound as a shader input, we might need to restart the render pass with a FB
  1670. // attachment that supports read-only attachments using the GENERAL or DEPTH_READ_ONLY layout
  1671. bool resetRenderPass = false;
  1672. if (isFBAttachment)
  1673. {
  1674. if (!subresourceInfo.isFBAttachment)
  1675. {
  1676. subresourceInfo.isFBAttachment = true;
  1677. resetRenderPass = subresourceInfo.isShaderInput;
  1678. }
  1679. }
  1680. else
  1681. {
  1682. if (!subresourceInfo.isShaderInput)
  1683. {
  1684. subresourceInfo.isShaderInput = true;
  1685. if (subresourceInfo.isFBAttachment)
  1686. {
  1687. // Special case for depth: If user has set up proper read-only flags, then the render pass will have
  1688. // taken care of setting the valid state anyway, so no need to end the render pass
  1689. if (subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL)
  1690. {
  1691. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0 || (mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1692. resetRenderPass = false;
  1693. else
  1694. resetRenderPass = true;
  1695. }
  1696. else
  1697. resetRenderPass = true;
  1698. }
  1699. else
  1700. resetRenderPass = false;
  1701. }
  1702. }
  1703. // If we need to switch frame-buffers, end current render pass
  1704. if (resetRenderPass && isInRenderPass())
  1705. endRenderPass();
  1706. else
  1707. {
  1708. // Since we won't be ending the render pass, check if this same sub-resource was written earlier in the pass,
  1709. // in which case we need to issue a memory barrier so those writes are visible.
  1710. // Memory barrier only matters if image is bound for shader use (no need for color attachments or transfers)
  1711. bool isShaderBind = !isFBAttachment && newLayout != VK_IMAGE_LAYOUT_UNDEFINED;
  1712. if(subresourceInfo.needsBarrier && isShaderBind)
  1713. {
  1714. bool isWrite = flags.isSet(VulkanUseFlag::Write);
  1715. VkPipelineStageFlags stages =
  1716. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1717. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1718. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1719. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1720. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1721. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1722. memoryBarrier(image->getHandle(), VK_ACCESS_SHADER_WRITE_BIT,
  1723. image->getAccessFlags(subresourceInfo.requiredLayout, !isWrite),
  1724. stages, stages, subresourceInfo.requiredLayout, subresourceInfo.range);
  1725. subresourceInfo.needsBarrier = isWrite;
  1726. }
  1727. }
  1728. return resetRenderPass;
  1729. }
  1730. VulkanCmdBuffer::ImageSubresourceInfo& VulkanCmdBuffer::findSubresourceInfo(VulkanImage* image, UINT32 face, UINT32 mip)
  1731. {
  1732. UINT32 imageInfoIdx = mImages[image];
  1733. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1734. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1735. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1736. {
  1737. ImageSubresourceInfo& entry = subresourceInfos[i];
  1738. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1739. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1740. {
  1741. return entry;
  1742. }
  1743. }
  1744. assert(false); // Caller should ensure the subresource actually exists, so this shouldn't happen
  1745. return subresourceInfos[0];
  1746. }
  1747. void VulkanCmdBuffer::getInProgressQueries(Vector<VulkanTimerQuery*>& timer, Vector<VulkanOcclusionQuery*>& occlusion) const
  1748. {
  1749. for(auto& query : mTimerQueries)
  1750. {
  1751. if (query->_isInProgress())
  1752. timer.push_back(query);
  1753. }
  1754. for (auto& query : mOcclusionQueries)
  1755. {
  1756. if (query->_isInProgress())
  1757. occlusion.push_back(query);
  1758. }
  1759. }
  1760. VulkanCommandBuffer::VulkanCommandBuffer(VulkanDevice& device, GpuQueueType type, UINT32 deviceIdx,
  1761. UINT32 queueIdx, bool secondary)
  1762. : CommandBuffer(type, deviceIdx, queueIdx, secondary), mBuffer(nullptr)
  1763. , mDevice(device), mQueue(nullptr), mIdMask(0)
  1764. {
  1765. UINT32 numQueues = device.getNumQueues(mType);
  1766. if (numQueues == 0) // Fall back to graphics queue
  1767. {
  1768. mType = GQT_GRAPHICS;
  1769. numQueues = device.getNumQueues(GQT_GRAPHICS);
  1770. }
  1771. mQueue = device.getQueue(mType, mQueueIdx % numQueues);
  1772. mIdMask = device.getQueueMask(mType, mQueueIdx);
  1773. acquireNewBuffer();
  1774. }
  1775. VulkanCommandBuffer::~VulkanCommandBuffer()
  1776. {
  1777. mBuffer->reset();
  1778. }
  1779. void VulkanCommandBuffer::acquireNewBuffer()
  1780. {
  1781. VulkanCmdBufferPool& pool = mDevice.getCmdBufferPool();
  1782. if (mBuffer != nullptr)
  1783. assert(mBuffer->isSubmitted());
  1784. UINT32 queueFamily = mDevice.getQueueFamily(mType);
  1785. mBuffer = pool.getBuffer(queueFamily, mIsSecondary);
  1786. }
  1787. void VulkanCommandBuffer::submit(UINT32 syncMask)
  1788. {
  1789. // Ignore myself
  1790. syncMask &= ~mIdMask;
  1791. if (mBuffer->isInRenderPass())
  1792. mBuffer->endRenderPass();
  1793. // Execute any queued layout transitions that weren't already handled by the render pass
  1794. mBuffer->executeLayoutTransitions();
  1795. // Interrupt any in-progress queries (no in-progress queries allowed during command buffer submit)
  1796. Vector<VulkanTimerQuery*> timerQueries;
  1797. Vector<VulkanOcclusionQuery*> occlusionQueries;
  1798. mBuffer->getInProgressQueries(timerQueries, occlusionQueries);
  1799. for (auto& query : timerQueries)
  1800. query->_interrupt(*mBuffer);
  1801. for (auto& query : occlusionQueries)
  1802. query->_interrupt(*mBuffer);
  1803. if (mBuffer->isRecording())
  1804. mBuffer->end();
  1805. if (mBuffer->isReadyForSubmit()) // Possibly nothing was recorded in the buffer
  1806. {
  1807. mBuffer->submit(mQueue, mQueueIdx, syncMask);
  1808. acquireNewBuffer();
  1809. gVulkanCBManager().refreshStates(mDeviceIdx);
  1810. }
  1811. // Resume interrupted queries on the new command buffer
  1812. for (auto& query : timerQueries)
  1813. query->_resume(*mBuffer);
  1814. for (auto& query : occlusionQueries)
  1815. query->_resume(*mBuffer);
  1816. }
  1817. }}