BsVulkanCommandBuffer.cpp 78 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanCommandBuffer.h"
  4. #include "Managers/BsVulkanCommandBufferManager.h"
  5. #include "BsVulkanUtility.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParams.h"
  8. #include "BsVulkanQueue.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanIndexBuffer.h"
  11. #include "BsVulkanVertexBuffer.h"
  12. #include "BsVulkanHardwareBuffer.h"
  13. #include "BsVulkanFramebuffer.h"
  14. #include "Managers/BsVulkanVertexInputManager.h"
  15. #include "BsVulkanEventQuery.h"
  16. #include "Managers/BsVulkanQueryManager.h"
  17. #include "BsVulkanSwapChain.h"
  18. #include "BsVulkanTimerQuery.h"
  19. #include "BsVulkanOcclusionQuery.h"
  20. #if BS_PLATFORM == BS_PLATFORM_WIN32
  21. #include "Win32/BsWin32RenderWindow.h"
  22. #else
  23. static_assert(false, "Other platforms go here");
  24. #endif
  25. namespace bs { namespace ct
  26. {
  27. VulkanSemaphore::VulkanSemaphore(VulkanResourceManager* owner)
  28. :VulkanResource(owner, true)
  29. {
  30. VkSemaphoreCreateInfo semaphoreCI;
  31. semaphoreCI.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  32. semaphoreCI.pNext = nullptr;
  33. semaphoreCI.flags = 0;
  34. VkResult result = vkCreateSemaphore(owner->getDevice().getLogical(), &semaphoreCI, gVulkanAllocator, &mSemaphore);
  35. assert(result == VK_SUCCESS);
  36. }
  37. VulkanSemaphore::~VulkanSemaphore()
  38. {
  39. vkDestroySemaphore(mOwner->getDevice().getLogical(), mSemaphore, gVulkanAllocator);
  40. }
  41. VulkanCmdBufferPool::VulkanCmdBufferPool(VulkanDevice& device)
  42. :mDevice(device), mNextId(1)
  43. {
  44. for (UINT32 i = 0; i < GQT_COUNT; i++)
  45. {
  46. UINT32 familyIdx = device.getQueueFamily((GpuQueueType)i);
  47. if (familyIdx == (UINT32)-1)
  48. continue;
  49. VkCommandPoolCreateInfo poolCI;
  50. poolCI.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  51. poolCI.pNext = nullptr;
  52. poolCI.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
  53. poolCI.queueFamilyIndex = familyIdx;
  54. PoolInfo& poolInfo = mPools[familyIdx];
  55. poolInfo.queueFamily = familyIdx;
  56. memset(poolInfo.buffers, 0, sizeof(poolInfo.buffers));
  57. vkCreateCommandPool(device.getLogical(), &poolCI, gVulkanAllocator, &poolInfo.pool);
  58. }
  59. }
  60. VulkanCmdBufferPool::~VulkanCmdBufferPool()
  61. {
  62. // Note: Shutdown should be the only place command buffers are destroyed at, as the system relies on the fact that
  63. // they won't be destroyed during normal operation.
  64. for(auto& entry : mPools)
  65. {
  66. PoolInfo& poolInfo = entry.second;
  67. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  68. {
  69. VulkanCmdBuffer* buffer = poolInfo.buffers[i];
  70. if (buffer == nullptr)
  71. break;
  72. bs_delete(buffer);
  73. }
  74. vkDestroyCommandPool(mDevice.getLogical(), poolInfo.pool, gVulkanAllocator);
  75. }
  76. }
  77. VulkanCmdBuffer* VulkanCmdBufferPool::getBuffer(UINT32 queueFamily, bool secondary)
  78. {
  79. auto iterFind = mPools.find(queueFamily);
  80. if (iterFind == mPools.end())
  81. return nullptr;
  82. VulkanCmdBuffer** buffers = iterFind->second.buffers;
  83. UINT32 i = 0;
  84. for(; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  85. {
  86. if (buffers[i] == nullptr)
  87. break;
  88. if(buffers[i]->mState == VulkanCmdBuffer::State::Ready)
  89. {
  90. buffers[i]->begin();
  91. return buffers[i];
  92. }
  93. }
  94. assert(i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY &&
  95. "Too many command buffers allocated. Increment BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY to a higher value. ");
  96. buffers[i] = createBuffer(queueFamily, secondary);
  97. buffers[i]->begin();
  98. return buffers[i];
  99. }
  100. VulkanCmdBuffer* VulkanCmdBufferPool::createBuffer(UINT32 queueFamily, bool secondary)
  101. {
  102. auto iterFind = mPools.find(queueFamily);
  103. if (iterFind == mPools.end())
  104. return nullptr;
  105. const PoolInfo& poolInfo = iterFind->second;
  106. return bs_new<VulkanCmdBuffer>(mDevice, mNextId++, poolInfo.pool, poolInfo.queueFamily, secondary);
  107. }
  108. /** Returns a set of pipeline stages that can are allowed to be used for the specified set of access flags. */
  109. VkPipelineStageFlags getPipelineStageFlags(VkAccessFlags accessFlags)
  110. {
  111. VkPipelineStageFlags flags = 0;
  112. if ((accessFlags & VK_ACCESS_INDIRECT_COMMAND_READ_BIT) != 0)
  113. flags |= VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT;
  114. if ((accessFlags & (VK_ACCESS_INDEX_READ_BIT | VK_ACCESS_VERTEX_ATTRIBUTE_READ_BIT)) != 0)
  115. flags |= VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
  116. if ((accessFlags & (VK_ACCESS_UNIFORM_READ_BIT | VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT)) != 0)
  117. {
  118. flags |= VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT;
  119. flags |= VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT | VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT;
  120. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  121. }
  122. if ((accessFlags & VK_ACCESS_INPUT_ATTACHMENT_READ_BIT) != 0)
  123. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
  124. if ((accessFlags & (VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT)) != 0)
  125. flags |= VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  126. if ((accessFlags & (VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT)) != 0)
  127. flags |= VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT;
  128. if ((accessFlags & (VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT)) != 0)
  129. flags |= VK_PIPELINE_STAGE_TRANSFER_BIT;
  130. if ((accessFlags & (VK_ACCESS_HOST_READ_BIT | VK_ACCESS_HOST_WRITE_BIT)) != 0)
  131. flags |= VK_PIPELINE_STAGE_HOST_BIT;
  132. if (flags == 0)
  133. flags = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  134. return flags;
  135. }
  136. template<class T>
  137. void getPipelineStageFlags(const Vector<T>& barriers, VkPipelineStageFlags& src, VkPipelineStageFlags& dst)
  138. {
  139. for(auto& entry : barriers)
  140. {
  141. src |= getPipelineStageFlags(entry.srcAccessMask);
  142. dst |= getPipelineStageFlags(entry.dstAccessMask);
  143. }
  144. if (src == 0)
  145. src = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  146. if (dst == 0)
  147. dst = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  148. }
  149. VulkanCmdBuffer::VulkanCmdBuffer(VulkanDevice& device, UINT32 id, VkCommandPool pool, UINT32 queueFamily, bool secondary)
  150. : mId(id), mQueueFamily(queueFamily), mState(State::Ready), mDevice(device), mPool(pool)
  151. , mIntraQueueSemaphore(nullptr), mInterQueueSemaphores(), mNumUsedInterQueueSemaphores(0)
  152. , mFramebuffer(nullptr), mRenderTargetWidth(0)
  153. , mRenderTargetHeight(0), mRenderTargetReadOnlyFlags(0), mRenderTargetLoadMask(RT_NONE), mGlobalQueueIdx(-1)
  154. , mViewport(0.0f, 0.0f, 1.0f, 1.0f), mScissor(0, 0, 0, 0), mStencilRef(0), mDrawOp(DOT_TRIANGLE_LIST)
  155. , mNumBoundDescriptorSets(0), mGfxPipelineRequiresBind(true), mCmpPipelineRequiresBind(true)
  156. , mViewportRequiresBind(true), mStencilRefRequiresBind(true), mScissorRequiresBind(true), mBoundParamsDirty(false)
  157. , mClearValues(), mClearMask(), mSemaphoresTemp(BS_MAX_UNIQUE_QUEUES), mVertexBuffersTemp()
  158. , mVertexBufferOffsetsTemp()
  159. {
  160. UINT32 maxBoundDescriptorSets = device.getDeviceProperties().limits.maxBoundDescriptorSets;
  161. mDescriptorSetsTemp = (VkDescriptorSet*)bs_alloc(sizeof(VkDescriptorSet) * maxBoundDescriptorSets);
  162. VkCommandBufferAllocateInfo cmdBufferAllocInfo;
  163. cmdBufferAllocInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  164. cmdBufferAllocInfo.pNext = nullptr;
  165. cmdBufferAllocInfo.commandPool = pool;
  166. cmdBufferAllocInfo.level = secondary ? VK_COMMAND_BUFFER_LEVEL_SECONDARY : VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  167. cmdBufferAllocInfo.commandBufferCount = 1;
  168. VkResult result = vkAllocateCommandBuffers(mDevice.getLogical(), &cmdBufferAllocInfo, &mCmdBuffer);
  169. assert(result == VK_SUCCESS);
  170. VkFenceCreateInfo fenceCI;
  171. fenceCI.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  172. fenceCI.pNext = nullptr;
  173. fenceCI.flags = 0;
  174. result = vkCreateFence(mDevice.getLogical(), &fenceCI, gVulkanAllocator, &mFence);
  175. assert(result == VK_SUCCESS);
  176. }
  177. VulkanCmdBuffer::~VulkanCmdBuffer()
  178. {
  179. VkDevice device = mDevice.getLogical();
  180. if(mState == State::Submitted)
  181. {
  182. // Wait 1s
  183. UINT64 waitTime = 1000 * 1000 * 1000;
  184. VkResult result = vkWaitForFences(device, 1, &mFence, true, waitTime);
  185. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  186. if (result == VK_TIMEOUT)
  187. LOGWRN("Freeing a command buffer before done executing because fence wait expired!");
  188. // Resources have been marked as used, make sure to notify them we're done with them
  189. reset();
  190. }
  191. else if(mState != State::Ready)
  192. {
  193. // Notify any resources that they are no longer bound
  194. for (auto& entry : mResources)
  195. {
  196. ResourceUseHandle& useHandle = entry.second;
  197. assert(!useHandle.used);
  198. entry.first->notifyUnbound();
  199. }
  200. for (auto& entry : mImages)
  201. {
  202. UINT32 imageInfoIdx = entry.second;
  203. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  204. ResourceUseHandle& useHandle = imageInfo.useHandle;
  205. assert(!useHandle.used);
  206. entry.first->notifyUnbound();
  207. }
  208. for (auto& entry : mBuffers)
  209. {
  210. ResourceUseHandle& useHandle = entry.second.useHandle;
  211. assert(!useHandle.used);
  212. entry.first->notifyUnbound();
  213. }
  214. }
  215. if (mIntraQueueSemaphore != nullptr)
  216. mIntraQueueSemaphore->destroy();
  217. for(UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  218. {
  219. if (mInterQueueSemaphores[i] != nullptr)
  220. mInterQueueSemaphores[i]->destroy();
  221. }
  222. vkDestroyFence(device, mFence, gVulkanAllocator);
  223. vkFreeCommandBuffers(device, mPool, 1, &mCmdBuffer);
  224. bs_free(mDescriptorSetsTemp);
  225. }
  226. UINT32 VulkanCmdBuffer::getDeviceIdx() const
  227. {
  228. return mDevice.getIndex();
  229. }
  230. void VulkanCmdBuffer::begin()
  231. {
  232. assert(mState == State::Ready);
  233. VkCommandBufferBeginInfo beginInfo;
  234. beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  235. beginInfo.pNext = nullptr;
  236. beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  237. beginInfo.pInheritanceInfo = nullptr;
  238. VkResult result = vkBeginCommandBuffer(mCmdBuffer, &beginInfo);
  239. assert(result == VK_SUCCESS);
  240. mState = State::Recording;
  241. }
  242. void VulkanCmdBuffer::end()
  243. {
  244. assert(mState == State::Recording);
  245. // If a clear is queued, execute the render pass with no additional instructions
  246. if (mClearMask)
  247. executeClearPass();
  248. VkResult result = vkEndCommandBuffer(mCmdBuffer);
  249. assert(result == VK_SUCCESS);
  250. mState = State::RecordingDone;
  251. }
  252. void VulkanCmdBuffer::beginRenderPass()
  253. {
  254. assert(mState == State::Recording);
  255. if (mFramebuffer == nullptr)
  256. {
  257. LOGWRN("Attempting to begin a render pass but no render target is bound to the command buffer.");
  258. return;
  259. }
  260. if(mClearMask != CLEAR_NONE)
  261. {
  262. // If a previous clear is queued, but it doesn't match the rendered area, need to execute a separate pass
  263. // just for it
  264. Rect2I rtArea(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  265. if (mClearArea != rtArea)
  266. executeClearPass();
  267. }
  268. executeLayoutTransitions();
  269. RenderSurfaceMask readMask = getFBReadMask();
  270. VkRenderPassBeginInfo renderPassBeginInfo;
  271. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  272. renderPassBeginInfo.pNext = nullptr;
  273. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(mRenderTargetLoadMask, readMask, mClearMask);
  274. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(mRenderTargetLoadMask, readMask, mClearMask);
  275. renderPassBeginInfo.renderArea.offset.x = 0;
  276. renderPassBeginInfo.renderArea.offset.y = 0;
  277. renderPassBeginInfo.renderArea.extent.width = mRenderTargetWidth;
  278. renderPassBeginInfo.renderArea.extent.height = mRenderTargetHeight;
  279. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  280. renderPassBeginInfo.pClearValues = mClearValues.data();
  281. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  282. mClearMask = CLEAR_NONE;
  283. mState = State::RecordingRenderPass;
  284. }
  285. void VulkanCmdBuffer::endRenderPass()
  286. {
  287. assert(mState == State::RecordingRenderPass);
  288. vkCmdEndRenderPass(mCmdBuffer);
  289. // Execute any queued events
  290. for(auto& entry : mQueuedEvents)
  291. vkCmdSetEvent(mCmdBuffer, entry->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  292. mQueuedEvents.clear();
  293. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  294. // and reset read-only state.
  295. for(auto& entry : mPassTouchedSubresourceInfos)
  296. {
  297. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  298. subresourceInfo.isShaderInput = false;
  299. subresourceInfo.isReadOnly = true;
  300. subresourceInfo.needsBarrier = false;
  301. }
  302. for (auto& entry : mBuffers)
  303. entry.second.needsBarrier = false;
  304. mPassTouchedSubresourceInfos.clear();
  305. updateFinalLayouts();
  306. mState = State::Recording;
  307. // In case the same GPU params from last pass get used, this makes sure the states we reset above, get re-applied
  308. mBoundParamsDirty = true;
  309. }
  310. void VulkanCmdBuffer::allocateSemaphores(VkSemaphore* semaphores)
  311. {
  312. if (mIntraQueueSemaphore != nullptr)
  313. mIntraQueueSemaphore->destroy();
  314. mIntraQueueSemaphore = mDevice.getResourceManager().create<VulkanSemaphore>();
  315. semaphores[0] = mIntraQueueSemaphore->getHandle();
  316. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  317. {
  318. if (mInterQueueSemaphores[i] != nullptr)
  319. mInterQueueSemaphores[i]->destroy();
  320. mInterQueueSemaphores[i] = mDevice.getResourceManager().create<VulkanSemaphore>();
  321. semaphores[i + 1] = mInterQueueSemaphores[i]->getHandle();
  322. }
  323. mNumUsedInterQueueSemaphores = 0;
  324. }
  325. VulkanSemaphore* VulkanCmdBuffer::requestInterQueueSemaphore() const
  326. {
  327. if (mNumUsedInterQueueSemaphores >= BS_MAX_VULKAN_CB_DEPENDENCIES)
  328. return nullptr;
  329. return mInterQueueSemaphores[mNumUsedInterQueueSemaphores++];
  330. }
  331. void VulkanCmdBuffer::submit(VulkanQueue* queue, UINT32 queueIdx, UINT32 syncMask)
  332. {
  333. assert(isReadyForSubmit());
  334. // Make sure to reset the CB fence before we submit it
  335. VkResult result = vkResetFences(mDevice.getLogical(), 1, &mFence);
  336. assert(result == VK_SUCCESS);
  337. // If there are any query resets needed, execute those first
  338. VulkanDevice& device = queue->getDevice();
  339. if(!mQueuedQueryResets.empty())
  340. {
  341. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  342. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  343. for (auto& entry : mQueuedQueryResets)
  344. entry->reset(vkCmdBuffer);
  345. cmdBuffer->end();
  346. queue->queueSubmit(cmdBuffer, nullptr, 0);
  347. mQueuedQueryResets.clear();
  348. }
  349. // Issue pipeline barriers for queue transitions (need to happen on original queue first, then on new queue)
  350. for (auto& entry : mBuffers)
  351. {
  352. VulkanBuffer* resource = static_cast<VulkanBuffer*>(entry.first);
  353. if (!resource->isExclusive())
  354. continue;
  355. UINT32 currentQueueFamily = resource->getQueueFamily();
  356. if (currentQueueFamily != -1 && currentQueueFamily != mQueueFamily)
  357. {
  358. Vector<VkBufferMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].bufferBarriers;
  359. barriers.push_back(VkBufferMemoryBarrier());
  360. VkBufferMemoryBarrier& barrier = barriers.back();
  361. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  362. barrier.pNext = nullptr;
  363. barrier.srcAccessMask = entry.second.accessFlags;
  364. barrier.dstAccessMask = entry.second.accessFlags;
  365. barrier.srcQueueFamilyIndex = currentQueueFamily;
  366. barrier.dstQueueFamilyIndex = mQueueFamily;
  367. barrier.buffer = resource->getHandle();
  368. barrier.offset = 0;
  369. barrier.size = VK_WHOLE_SIZE;
  370. }
  371. }
  372. // For images issue queue transitions, as above. Also issue layout transitions to their inital layouts.
  373. Vector<VkImageMemoryBarrier>& localBarriers = mTransitionInfoTemp[mQueueFamily].imageBarriers;
  374. for (auto& entry : mImages)
  375. {
  376. VulkanImage* resource = static_cast<VulkanImage*>(entry.first);
  377. ImageInfo& imageInfo = mImageInfos[entry.second];
  378. UINT32 currentQueueFamily = resource->getQueueFamily();
  379. bool queueMismatch = resource->isExclusive() && currentQueueFamily != -1 && currentQueueFamily != mQueueFamily;
  380. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  381. if (queueMismatch)
  382. {
  383. Vector<VkImageMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].imageBarriers;
  384. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  385. {
  386. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  387. UINT32 startIdx = (UINT32)barriers.size();
  388. resource->getBarriers(subresourceInfo.range, barriers);
  389. for(UINT32 j = startIdx; j < (UINT32)barriers.size(); j++)
  390. {
  391. VkImageMemoryBarrier& barrier = barriers[j];
  392. barrier.dstAccessMask = resource->getAccessFlags(barrier.oldLayout);
  393. barrier.newLayout = barrier.oldLayout;
  394. barrier.srcQueueFamilyIndex = currentQueueFamily;
  395. barrier.dstQueueFamilyIndex = mQueueFamily;
  396. }
  397. }
  398. }
  399. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  400. {
  401. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  402. VkImageLayout initialLayout = subresourceInfo.initialLayout;
  403. if (initialLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  404. continue;
  405. const VkImageSubresourceRange& range = subresourceInfo.range;
  406. UINT32 mipEnd = range.baseMipLevel + range.levelCount;
  407. UINT32 faceEnd = range.baseArrayLayer + range.layerCount;
  408. bool layoutMismatch = false;
  409. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  410. {
  411. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  412. {
  413. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  414. if(subresource->getLayout() != initialLayout)
  415. {
  416. layoutMismatch = true;
  417. break;
  418. }
  419. }
  420. if (layoutMismatch)
  421. break;
  422. }
  423. if(layoutMismatch)
  424. {
  425. UINT32 startIdx = (UINT32)localBarriers.size();
  426. resource->getBarriers(subresourceInfo.range, localBarriers);
  427. for (UINT32 j = startIdx; j < (UINT32)localBarriers.size(); j++)
  428. {
  429. VkImageMemoryBarrier& barrier = localBarriers[j];
  430. barrier.dstAccessMask = resource->getAccessFlags(initialLayout, subresourceInfo.isInitialReadOnly);
  431. barrier.newLayout = initialLayout;
  432. }
  433. }
  434. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  435. {
  436. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  437. {
  438. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  439. subresource->setLayout(subresourceInfo.finalLayout);
  440. }
  441. }
  442. }
  443. }
  444. for (auto& entry : mTransitionInfoTemp)
  445. {
  446. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  447. if (empty)
  448. continue;
  449. UINT32 entryQueueFamily = entry.first;
  450. // No queue transition needed for entries on this queue (this entry is most likely an image layout transition)
  451. if (entryQueueFamily == -1 || entryQueueFamily == mQueueFamily)
  452. continue;
  453. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(entryQueueFamily, false);
  454. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  455. TransitionInfo& barriers = entry.second;
  456. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  457. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  458. VkPipelineStageFlags srcStage = 0;
  459. VkPipelineStageFlags dstStage = 0;
  460. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  461. vkCmdPipelineBarrier(vkCmdBuffer,
  462. srcStage, dstStage,
  463. 0, 0, nullptr,
  464. numBufferBarriers, barriers.bufferBarriers.data(),
  465. numImgBarriers, barriers.imageBarriers.data());
  466. // Find an appropriate queue to execute on
  467. UINT32 otherQueueIdx = 0;
  468. VulkanQueue* otherQueue = nullptr;
  469. GpuQueueType otherQueueType = GQT_GRAPHICS;
  470. for (UINT32 i = 0; i < GQT_COUNT; i++)
  471. {
  472. otherQueueType = (GpuQueueType)i;
  473. if (device.getQueueFamily(otherQueueType) != entryQueueFamily)
  474. continue;
  475. UINT32 numQueues = device.getNumQueues(otherQueueType);
  476. for (UINT32 j = 0; j < numQueues; j++)
  477. {
  478. // Try to find a queue not currently executing
  479. VulkanQueue* curQueue = device.getQueue(otherQueueType, j);
  480. if (!curQueue->isExecuting())
  481. {
  482. otherQueue = curQueue;
  483. otherQueueIdx = j;
  484. }
  485. }
  486. // Can't find empty one, use the first one then
  487. if (otherQueue == nullptr)
  488. {
  489. otherQueue = device.getQueue(otherQueueType, 0);
  490. otherQueueIdx = 0;
  491. }
  492. break;
  493. }
  494. syncMask |= CommandSyncMask::getGlobalQueueMask(otherQueueType, otherQueueIdx);
  495. cmdBuffer->end();
  496. // Note: If I switch back to doing layout transitions here, I need to wait on present semaphore
  497. otherQueue->submit(cmdBuffer, nullptr, 0);
  498. }
  499. UINT32 deviceIdx = device.getIndex();
  500. VulkanCommandBufferManager& cbm = static_cast<VulkanCommandBufferManager&>(CommandBufferManager::instance());
  501. UINT32 numSemaphores;
  502. cbm.getSyncSemaphores(deviceIdx, syncMask, mSemaphoresTemp.data(), numSemaphores);
  503. // Wait on present (i.e. until the back buffer becomes available) for any swap chains
  504. for(auto& entry : mSwapChains)
  505. {
  506. const SwapChainSurface& surface = entry->getBackBuffer();
  507. if (surface.needsWait)
  508. {
  509. VulkanSemaphore* semaphore = entry->getBackBuffer().sync;
  510. if (numSemaphores >= (UINT32)mSemaphoresTemp.size())
  511. mSemaphoresTemp.push_back(semaphore);
  512. else
  513. mSemaphoresTemp[numSemaphores] = semaphore;
  514. numSemaphores++;
  515. entry->notifyBackBufferWaitIssued();
  516. }
  517. }
  518. // Issue second part of transition pipeline barriers (on this queue)
  519. for (auto& entry : mTransitionInfoTemp)
  520. {
  521. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  522. if (empty)
  523. continue;
  524. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  525. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  526. TransitionInfo& barriers = entry.second;
  527. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  528. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  529. VkPipelineStageFlags srcStage = 0;
  530. VkPipelineStageFlags dstStage = 0;
  531. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  532. vkCmdPipelineBarrier(vkCmdBuffer,
  533. srcStage, dstStage,
  534. 0, 0, nullptr,
  535. numBufferBarriers, barriers.bufferBarriers.data(),
  536. numImgBarriers, barriers.imageBarriers.data());
  537. cmdBuffer->end();
  538. queue->queueSubmit(cmdBuffer, mSemaphoresTemp.data(), numSemaphores);
  539. numSemaphores = 0; // Semaphores are only needed the first time, since we're adding the buffers on the same queue
  540. }
  541. queue->queueSubmit(this, mSemaphoresTemp.data(), numSemaphores);
  542. queue->submitQueued();
  543. mGlobalQueueIdx = CommandSyncMask::getGlobalQueueIdx(queue->getType(), queueIdx);
  544. for (auto& entry : mResources)
  545. {
  546. ResourceUseHandle& useHandle = entry.second;
  547. assert(!useHandle.used);
  548. useHandle.used = true;
  549. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  550. }
  551. for (auto& entry : mImages)
  552. {
  553. UINT32 imageInfoIdx = entry.second;
  554. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  555. ResourceUseHandle& useHandle = imageInfo.useHandle;
  556. assert(!useHandle.used);
  557. useHandle.used = true;
  558. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  559. }
  560. for (auto& entry : mBuffers)
  561. {
  562. ResourceUseHandle& useHandle = entry.second.useHandle;
  563. assert(!useHandle.used);
  564. useHandle.used = true;
  565. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  566. }
  567. // Note: Uncomment for debugging only, prevents any device concurrency issues.
  568. // vkQueueWaitIdle(queue->getHandle());
  569. // Clear vectors but don't clear the actual map, as we want to re-use the memory since we expect queue family
  570. // indices to be the same
  571. for (auto& entry : mTransitionInfoTemp)
  572. {
  573. entry.second.imageBarriers.clear();
  574. entry.second.bufferBarriers.clear();
  575. }
  576. mGraphicsPipeline = nullptr;
  577. mComputePipeline = nullptr;
  578. mGfxPipelineRequiresBind = true;
  579. mCmpPipelineRequiresBind = true;
  580. mFramebuffer = nullptr;
  581. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  582. mQueuedLayoutTransitions.clear();
  583. mBoundParams = nullptr;
  584. mSwapChains.clear();
  585. }
  586. bool VulkanCmdBuffer::checkFenceStatus(bool block) const
  587. {
  588. VkResult result = vkWaitForFences(mDevice.getLogical(), 1, &mFence, true, block ? 1'000'000'000 : 0);
  589. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  590. return result == VK_SUCCESS;
  591. }
  592. void VulkanCmdBuffer::reset()
  593. {
  594. bool wasSubmitted = mState == State::Submitted;
  595. mState = State::Ready;
  596. vkResetCommandBuffer(mCmdBuffer, VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT); // Note: Maybe better not to release resources?
  597. if (wasSubmitted)
  598. {
  599. for (auto& entry : mResources)
  600. {
  601. ResourceUseHandle& useHandle = entry.second;
  602. assert(useHandle.used);
  603. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  604. }
  605. for (auto& entry : mImages)
  606. {
  607. UINT32 imageInfoIdx = entry.second;
  608. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  609. ResourceUseHandle& useHandle = imageInfo.useHandle;
  610. assert(useHandle.used);
  611. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  612. }
  613. for (auto& entry : mBuffers)
  614. {
  615. ResourceUseHandle& useHandle = entry.second.useHandle;
  616. assert(useHandle.used);
  617. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  618. }
  619. }
  620. else
  621. {
  622. for (auto& entry : mResources)
  623. entry.first->notifyUnbound();
  624. for (auto& entry : mImages)
  625. entry.first->notifyUnbound();
  626. for (auto& entry : mBuffers)
  627. entry.first->notifyUnbound();
  628. }
  629. mResources.clear();
  630. mImages.clear();
  631. mBuffers.clear();
  632. mOcclusionQueries.clear();
  633. mTimerQueries.clear();
  634. mImageInfos.clear();
  635. mSubresourceInfoStorage.clear();
  636. mPassTouchedSubresourceInfos.clear();
  637. }
  638. void VulkanCmdBuffer::setRenderTarget(const SPtr<RenderTarget>& rt, UINT32 readOnlyFlags,
  639. RenderSurfaceMask loadMask)
  640. {
  641. assert(mState != State::Submitted);
  642. VulkanFramebuffer* newFB;
  643. if(rt != nullptr)
  644. {
  645. if (rt->getProperties().isWindow)
  646. {
  647. Win32RenderWindow* window = static_cast<Win32RenderWindow*>(rt.get());
  648. window->acquireBackBuffer();
  649. VulkanSwapChain* swapChain;
  650. rt->getCustomAttribute("SC", &swapChain);
  651. mSwapChains.insert(swapChain);
  652. }
  653. rt->getCustomAttribute("FB", &newFB);
  654. }
  655. else
  656. {
  657. newFB = nullptr;
  658. }
  659. // Warn if invalid load mask
  660. if (loadMask.isSet(RT_DEPTH) && !loadMask.isSet(RT_STENCIL))
  661. {
  662. LOGWRN("setRenderTarget() invalid load mask, depth enabled but stencil disabled. This is not supported. Both \
  663. will be loaded.");
  664. loadMask.set(RT_STENCIL);
  665. }
  666. if (!loadMask.isSet(RT_DEPTH) && loadMask.isSet(RT_STENCIL))
  667. {
  668. LOGWRN("setRenderTarget() invalid load mask, stencil enabled but depth disabled. This is not supported. Both \
  669. will be loaded.");
  670. loadMask.set(RT_DEPTH);
  671. }
  672. if (mFramebuffer == newFB && mRenderTargetReadOnlyFlags == readOnlyFlags && mRenderTargetLoadMask == loadMask)
  673. return;
  674. if (isInRenderPass())
  675. endRenderPass();
  676. else
  677. {
  678. // If a clear is queued for previous FB, execute the render pass with no additional instructions
  679. if (mClearMask)
  680. executeClearPass();
  681. }
  682. // Reset isFBAttachment flags for subresources from the old framebuffer
  683. if(mFramebuffer != nullptr)
  684. {
  685. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  686. for(UINT32 i = 0; i < numColorAttachments; i++)
  687. {
  688. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  689. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  690. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  691. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  692. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  693. {
  694. ImageSubresourceInfo& entry = subresourceInfos[j];
  695. entry.isFBAttachment = false;
  696. }
  697. }
  698. if(mFramebuffer->hasDepthAttachment())
  699. {
  700. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  701. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  702. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  703. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  704. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  705. {
  706. ImageSubresourceInfo& entry = subresourceInfos[j];
  707. entry.isFBAttachment = false;
  708. }
  709. }
  710. }
  711. if(newFB == nullptr)
  712. {
  713. mFramebuffer = nullptr;
  714. mRenderTargetWidth = 0;
  715. mRenderTargetHeight = 0;
  716. mRenderTargetReadOnlyFlags = 0;
  717. mRenderTargetLoadMask = RT_NONE;
  718. }
  719. else
  720. {
  721. mFramebuffer = newFB;
  722. mRenderTargetWidth = rt->getProperties().width;
  723. mRenderTargetHeight = rt->getProperties().height;
  724. mRenderTargetReadOnlyFlags = readOnlyFlags;
  725. mRenderTargetLoadMask = loadMask;
  726. }
  727. // Re-set the params as they will need to be re-bound
  728. setGpuParams(mBoundParams);
  729. if (mFramebuffer != nullptr)
  730. registerResource(mFramebuffer, loadMask, readOnlyFlags);
  731. mGfxPipelineRequiresBind = true;
  732. }
  733. void VulkanCmdBuffer::clearViewport(const Rect2I& area, UINT32 buffers, const Color& color, float depth, UINT16 stencil,
  734. UINT8 targetMask)
  735. {
  736. if (buffers == 0 || mFramebuffer == nullptr)
  737. return;
  738. // Add clear command if currently in render pass
  739. if (isInRenderPass())
  740. {
  741. VkClearAttachment attachments[BS_MAX_MULTIPLE_RENDER_TARGETS + 1];
  742. UINT32 baseLayer = 0;
  743. UINT32 attachmentIdx = 0;
  744. if ((buffers & FBT_COLOR) != 0)
  745. {
  746. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  747. for (UINT32 i = 0; i < numColorAttachments; i++)
  748. {
  749. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  750. if (((1 << attachment.index) & targetMask) == 0)
  751. continue;
  752. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  753. attachments[attachmentIdx].colorAttachment = i;
  754. VkClearColorValue& colorValue = attachments[attachmentIdx].clearValue.color;
  755. colorValue.float32[0] = color.r;
  756. colorValue.float32[1] = color.g;
  757. colorValue.float32[2] = color.b;
  758. colorValue.float32[3] = color.a;
  759. UINT32 curBaseLayer = attachment.baseLayer;
  760. if (attachmentIdx == 0)
  761. baseLayer = curBaseLayer;
  762. else
  763. {
  764. if (baseLayer != curBaseLayer)
  765. {
  766. // Note: This could be supported relatively easily: we would need to issue multiple separate
  767. // clear commands for such framebuffers.
  768. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  769. "starting layers. This is currently not supported.");
  770. }
  771. }
  772. attachmentIdx++;
  773. }
  774. }
  775. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  776. {
  777. if (mFramebuffer->hasDepthAttachment())
  778. {
  779. attachments[attachmentIdx].aspectMask = 0;
  780. if ((buffers & FBT_DEPTH) != 0)
  781. {
  782. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_DEPTH_BIT;
  783. attachments[attachmentIdx].clearValue.depthStencil.depth = depth;
  784. }
  785. if ((buffers & FBT_STENCIL) != 0)
  786. {
  787. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_STENCIL_BIT;
  788. attachments[attachmentIdx].clearValue.depthStencil.stencil = stencil;
  789. }
  790. attachments[attachmentIdx].colorAttachment = 0;
  791. UINT32 curBaseLayer = mFramebuffer->getDepthStencilAttachment().baseLayer;
  792. if (attachmentIdx == 0)
  793. baseLayer = curBaseLayer;
  794. else
  795. {
  796. if (baseLayer != curBaseLayer)
  797. {
  798. // Note: This could be supported relatively easily: we would need to issue multiple separate
  799. // clear commands for such framebuffers.
  800. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  801. "starting layers. This is currently not supported.");
  802. }
  803. }
  804. attachmentIdx++;
  805. }
  806. }
  807. UINT32 numAttachments = attachmentIdx;
  808. if (numAttachments == 0)
  809. return;
  810. VkClearRect clearRect;
  811. clearRect.baseArrayLayer = baseLayer;
  812. clearRect.layerCount = mFramebuffer->getNumLayers();
  813. clearRect.rect.offset.x = area.x;
  814. clearRect.rect.offset.y = area.y;
  815. clearRect.rect.extent.width = area.width;
  816. clearRect.rect.extent.height = area.height;
  817. vkCmdClearAttachments(mCmdBuffer, numAttachments, attachments, 1, &clearRect);
  818. }
  819. // Otherwise we use a render pass that performs a clear on begin
  820. else
  821. {
  822. ClearMask clearMask;
  823. std::array<VkClearValue, BS_MAX_MULTIPLE_RENDER_TARGETS + 1> clearValues = mClearValues;
  824. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  825. if ((buffers & FBT_COLOR) != 0)
  826. {
  827. for (UINT32 i = 0; i < numColorAttachments; i++)
  828. {
  829. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  830. if (((1 << attachment.index) & targetMask) == 0)
  831. continue;
  832. clearMask |= (ClearMaskBits)(1 << attachment.index);
  833. VkClearColorValue& colorValue = clearValues[i].color;
  834. colorValue.float32[0] = color.r;
  835. colorValue.float32[1] = color.g;
  836. colorValue.float32[2] = color.b;
  837. colorValue.float32[3] = color.a;
  838. }
  839. }
  840. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  841. {
  842. if (mFramebuffer->hasDepthAttachment())
  843. {
  844. UINT32 depthAttachmentIdx = numColorAttachments;
  845. if ((buffers & FBT_DEPTH) != 0)
  846. {
  847. clearValues[depthAttachmentIdx].depthStencil.depth = depth;
  848. clearMask |= CLEAR_DEPTH;
  849. }
  850. if ((buffers & FBT_STENCIL) != 0)
  851. {
  852. clearValues[depthAttachmentIdx].depthStencil.stencil = stencil;
  853. clearMask |= CLEAR_STENCIL;
  854. }
  855. }
  856. }
  857. if (!clearMask)
  858. return;
  859. // Some previous clear operation is already queued, execute it first
  860. bool previousClearNeedsToFinish = (mClearMask & clearMask) != CLEAR_NONE;
  861. if(previousClearNeedsToFinish)
  862. executeClearPass();
  863. mClearMask |= clearMask;
  864. mClearValues = clearValues;
  865. mClearArea = area;
  866. }
  867. }
  868. void VulkanCmdBuffer::clearRenderTarget(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  869. {
  870. Rect2I area(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  871. clearViewport(area, buffers, color, depth, stencil, targetMask);
  872. }
  873. void VulkanCmdBuffer::clearViewport(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  874. {
  875. Rect2I area;
  876. area.x = (UINT32)(mViewport.x * mRenderTargetWidth);
  877. area.y = (UINT32)(mViewport.y * mRenderTargetHeight);
  878. area.width = (UINT32)(mViewport.width * mRenderTargetWidth);
  879. area.height = (UINT32)(mViewport.height * mRenderTargetHeight);
  880. clearViewport(area, buffers, color, depth, stencil, targetMask);
  881. }
  882. void VulkanCmdBuffer::setPipelineState(const SPtr<GraphicsPipelineState>& state)
  883. {
  884. if (mGraphicsPipeline == state)
  885. return;
  886. mGraphicsPipeline = std::static_pointer_cast<VulkanGraphicsPipelineState>(state);
  887. mGfxPipelineRequiresBind = true;
  888. }
  889. void VulkanCmdBuffer::setPipelineState(const SPtr<ComputePipelineState>& state)
  890. {
  891. if (mComputePipeline == state)
  892. return;
  893. mComputePipeline = std::static_pointer_cast<VulkanComputePipelineState>(state);
  894. mCmpPipelineRequiresBind = true;
  895. }
  896. void VulkanCmdBuffer::setGpuParams(const SPtr<GpuParams>& gpuParams)
  897. {
  898. // Note: We keep an internal reference to GPU params even though we shouldn't keep a reference to a core thread
  899. // object. But it should be fine since we expect the resource to be externally synchronized so it should never
  900. // be allowed to go out of scope on a non-core thread anyway.
  901. mBoundParams = std::static_pointer_cast<VulkanGpuParams>(gpuParams);
  902. if (mBoundParams != nullptr)
  903. mBoundParamsDirty = true;
  904. else
  905. {
  906. mNumBoundDescriptorSets = 0;
  907. mBoundParamsDirty = false;
  908. }
  909. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  910. }
  911. void VulkanCmdBuffer::setViewport(const Rect2& area)
  912. {
  913. if (mViewport == area)
  914. return;
  915. mViewport = area;
  916. mViewportRequiresBind = true;
  917. }
  918. void VulkanCmdBuffer::setScissorRect(const Rect2I& value)
  919. {
  920. if (mScissor == value)
  921. return;
  922. mScissor = value;
  923. mScissorRequiresBind = true;
  924. }
  925. void VulkanCmdBuffer::setStencilRef(UINT32 value)
  926. {
  927. if (mStencilRef == value)
  928. return;
  929. mStencilRef = value;
  930. mStencilRefRequiresBind = true;
  931. }
  932. void VulkanCmdBuffer::setDrawOp(DrawOperationType drawOp)
  933. {
  934. if (mDrawOp == drawOp)
  935. return;
  936. mDrawOp = drawOp;
  937. mGfxPipelineRequiresBind = true;
  938. }
  939. void VulkanCmdBuffer::setVertexBuffers(UINT32 index, SPtr<VertexBuffer>* buffers, UINT32 numBuffers)
  940. {
  941. if (numBuffers == 0)
  942. return;
  943. for(UINT32 i = 0; i < numBuffers; i++)
  944. {
  945. VulkanVertexBuffer* vertexBuffer = static_cast<VulkanVertexBuffer*>(buffers[i].get());
  946. if (vertexBuffer != nullptr)
  947. {
  948. VulkanBuffer* resource = vertexBuffer->getResource(mDevice.getIndex());
  949. if (resource != nullptr)
  950. {
  951. mVertexBuffersTemp[i] = resource->getHandle();
  952. registerResource(resource, VulkanUseFlag::Read);
  953. }
  954. else
  955. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  956. }
  957. else
  958. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  959. }
  960. vkCmdBindVertexBuffers(mCmdBuffer, index, numBuffers, mVertexBuffersTemp, mVertexBufferOffsetsTemp);
  961. }
  962. void VulkanCmdBuffer::setIndexBuffer(const SPtr<IndexBuffer>& buffer)
  963. {
  964. VulkanIndexBuffer* indexBuffer = static_cast<VulkanIndexBuffer*>(buffer.get());
  965. VkBuffer vkBuffer = VK_NULL_HANDLE;
  966. VkIndexType indexType = VK_INDEX_TYPE_UINT32;
  967. if (indexBuffer != nullptr)
  968. {
  969. VulkanBuffer* resource = indexBuffer->getResource(mDevice.getIndex());
  970. if (resource != nullptr)
  971. {
  972. vkBuffer = resource->getHandle();
  973. indexType = VulkanUtility::getIndexType(buffer->getProperties().getType());
  974. registerResource(resource, VulkanUseFlag::Read);
  975. }
  976. }
  977. vkCmdBindIndexBuffer(mCmdBuffer, vkBuffer, 0, indexType);
  978. }
  979. void VulkanCmdBuffer::setVertexDeclaration(const SPtr<VertexDeclaration>& decl)
  980. {
  981. if (mVertexDecl == decl)
  982. return;
  983. mVertexDecl = decl;
  984. mGfxPipelineRequiresBind = true;
  985. }
  986. bool VulkanCmdBuffer::isReadyForRender()
  987. {
  988. if (mGraphicsPipeline == nullptr)
  989. return false;
  990. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  991. if (inputDecl == nullptr)
  992. return false;
  993. return mFramebuffer != nullptr && mVertexDecl != nullptr;
  994. }
  995. bool VulkanCmdBuffer::bindGraphicsPipeline()
  996. {
  997. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  998. SPtr<VulkanVertexInput> vertexInput = VulkanVertexInputManager::instance().getVertexInfo(mVertexDecl, inputDecl);
  999. VulkanPipeline* pipeline = mGraphicsPipeline->getPipeline(mDevice.getIndex(), mFramebuffer,
  1000. mRenderTargetReadOnlyFlags, mDrawOp, vertexInput);
  1001. if (pipeline == nullptr)
  1002. return false;
  1003. // Check that pipeline matches the read-only state of any framebuffer attachments
  1004. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1005. for (UINT32 i = 0; i < numColorAttachments; i++)
  1006. {
  1007. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1008. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1009. fbAttachment.surface.mipLevel);
  1010. if (subresourceInfo.isShaderInput && !pipeline->isColorReadOnly(i))
  1011. {
  1012. LOGWRN("Framebuffer attachment also used as a shader input, but color writes aren't disabled. This will"
  1013. " result in undefined behavior.");
  1014. }
  1015. }
  1016. if (mFramebuffer->hasDepthAttachment())
  1017. {
  1018. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1019. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1020. fbAttachment.surface.mipLevel);
  1021. if (subresourceInfo.isShaderInput && !pipeline->isDepthReadOnly())
  1022. {
  1023. LOGWRN("Framebuffer attachment also used as a shader input, but depth/stencil writes aren't disabled. "
  1024. "This will result in undefined behavior.");
  1025. }
  1026. }
  1027. mGraphicsPipeline->registerPipelineResources(this);
  1028. registerResource(pipeline, VulkanUseFlag::Read);
  1029. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline->getHandle());
  1030. bindDynamicStates(true);
  1031. mGfxPipelineRequiresBind = false;
  1032. return true;
  1033. }
  1034. void VulkanCmdBuffer::bindDynamicStates(bool forceAll)
  1035. {
  1036. if (mViewportRequiresBind || forceAll)
  1037. {
  1038. VkViewport viewport;
  1039. viewport.x = mViewport.x * mRenderTargetWidth;
  1040. viewport.y = mViewport.y * mRenderTargetHeight;
  1041. viewport.width = mViewport.width * mRenderTargetWidth;
  1042. viewport.height = mViewport.height * mRenderTargetHeight;
  1043. viewport.minDepth = 0.0f;
  1044. viewport.maxDepth = 1.0f;
  1045. vkCmdSetViewport(mCmdBuffer, 0, 1, &viewport);
  1046. mViewportRequiresBind = false;
  1047. }
  1048. if(mStencilRefRequiresBind || forceAll)
  1049. {
  1050. vkCmdSetStencilReference(mCmdBuffer, VK_STENCIL_FRONT_AND_BACK, mStencilRef);
  1051. mStencilRefRequiresBind = false;
  1052. }
  1053. if(mScissorRequiresBind || forceAll)
  1054. {
  1055. VkRect2D scissorRect;
  1056. if(mGraphicsPipeline->isScissorEnabled())
  1057. {
  1058. scissorRect.offset.x = mScissor.x;
  1059. scissorRect.offset.y = mScissor.y;
  1060. scissorRect.extent.width = mScissor.width;
  1061. scissorRect.extent.height = mScissor.height;
  1062. }
  1063. else
  1064. {
  1065. scissorRect.offset.x = 0;
  1066. scissorRect.offset.y = 0;
  1067. scissorRect.extent.width = mRenderTargetWidth;
  1068. scissorRect.extent.height = mRenderTargetHeight;
  1069. }
  1070. vkCmdSetScissor(mCmdBuffer, 0, 1, &scissorRect);
  1071. mScissorRequiresBind = false;
  1072. }
  1073. }
  1074. void VulkanCmdBuffer::bindGpuParams()
  1075. {
  1076. if (mBoundParamsDirty)
  1077. {
  1078. if (mBoundParams != nullptr)
  1079. {
  1080. mNumBoundDescriptorSets = mBoundParams->getNumSets();
  1081. mBoundParams->prepareForBind(*this, mDescriptorSetsTemp);
  1082. }
  1083. else
  1084. mNumBoundDescriptorSets = 0;
  1085. mBoundParamsDirty = false;
  1086. }
  1087. else
  1088. {
  1089. mNumBoundDescriptorSets = 0;
  1090. }
  1091. }
  1092. void VulkanCmdBuffer::executeLayoutTransitions()
  1093. {
  1094. auto createLayoutTransitionBarrier = [&](VulkanImage* image, ImageInfo& imageInfo)
  1095. {
  1096. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1097. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1098. {
  1099. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  1100. if (!subresourceInfo.hasTransitioned || subresourceInfo.currentLayout == subresourceInfo.requiredLayout)
  1101. continue;
  1102. mLayoutTransitionBarriersTemp.push_back(VkImageMemoryBarrier());
  1103. VkImageMemoryBarrier& barrier = mLayoutTransitionBarriersTemp.back();
  1104. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1105. barrier.pNext = nullptr;
  1106. barrier.srcAccessMask = image->getAccessFlags(subresourceInfo.currentLayout);
  1107. barrier.dstAccessMask = image->getAccessFlags(subresourceInfo.requiredLayout, subresourceInfo.isReadOnly);
  1108. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1109. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1110. barrier.oldLayout = subresourceInfo.currentLayout;
  1111. barrier.newLayout = subresourceInfo.requiredLayout;
  1112. barrier.image = image->getHandle();
  1113. barrier.subresourceRange = subresourceInfo.range;
  1114. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1115. subresourceInfo.isReadOnly = true;
  1116. subresourceInfo.hasTransitioned = true;
  1117. }
  1118. };
  1119. // Note: These layout transitions will contain transitions for offscreen framebuffer attachments (while they
  1120. // transition to shader read-only layout). This can be avoided, since they're immediately used by the render pass
  1121. // as color attachments, making the layout change redundant.
  1122. for (auto& entry : mQueuedLayoutTransitions)
  1123. {
  1124. UINT32 imageInfoIdx = entry.second;
  1125. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1126. createLayoutTransitionBarrier(entry.first, imageInfo);
  1127. }
  1128. VkPipelineStageFlags srcStage = 0;
  1129. VkPipelineStageFlags dstStage = 0;
  1130. getPipelineStageFlags(mLayoutTransitionBarriersTemp, srcStage, dstStage);
  1131. vkCmdPipelineBarrier(mCmdBuffer,
  1132. srcStage, dstStage,
  1133. 0, 0, nullptr,
  1134. 0, nullptr,
  1135. (UINT32)mLayoutTransitionBarriersTemp.size(), mLayoutTransitionBarriersTemp.data());
  1136. mQueuedLayoutTransitions.clear();
  1137. mLayoutTransitionBarriersTemp.clear();
  1138. }
  1139. void VulkanCmdBuffer::updateFinalLayouts()
  1140. {
  1141. if (mFramebuffer == nullptr)
  1142. return;
  1143. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1144. for (UINT32 i = 0; i < numColorAttachments; i++)
  1145. {
  1146. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1147. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1148. fbAttachment.surface.mipLevel);
  1149. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1150. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1151. subresourceInfo.hasTransitioned = true;
  1152. }
  1153. if (mFramebuffer->hasDepthAttachment())
  1154. {
  1155. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1156. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1157. fbAttachment.surface.mipLevel);
  1158. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1159. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1160. subresourceInfo.hasTransitioned = true;
  1161. }
  1162. }
  1163. void VulkanCmdBuffer::executeClearPass()
  1164. {
  1165. assert(mState == State::Recording);
  1166. executeLayoutTransitions();
  1167. VkRenderPassBeginInfo renderPassBeginInfo;
  1168. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  1169. renderPassBeginInfo.pNext = nullptr;
  1170. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(RT_NONE, RT_NONE, mClearMask);
  1171. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(RT_NONE, RT_NONE, mClearMask);
  1172. renderPassBeginInfo.renderArea.offset.x = mClearArea.x;
  1173. renderPassBeginInfo.renderArea.offset.y = mClearArea.y;
  1174. renderPassBeginInfo.renderArea.extent.width = mClearArea.width;
  1175. renderPassBeginInfo.renderArea.extent.height = mClearArea.height;
  1176. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  1177. renderPassBeginInfo.pClearValues = mClearValues.data();
  1178. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  1179. vkCmdEndRenderPass(mCmdBuffer);
  1180. updateFinalLayouts();
  1181. mClearMask = CLEAR_NONE;
  1182. }
  1183. void VulkanCmdBuffer::draw(UINT32 vertexOffset, UINT32 vertexCount, UINT32 instanceCount)
  1184. {
  1185. if (!isReadyForRender())
  1186. return;
  1187. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1188. bindGpuParams();
  1189. if (!isInRenderPass())
  1190. beginRenderPass();
  1191. if (mGfxPipelineRequiresBind)
  1192. {
  1193. if (!bindGraphicsPipeline())
  1194. return;
  1195. }
  1196. else
  1197. bindDynamicStates(false);
  1198. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1199. {
  1200. if (mNumBoundDescriptorSets > 0)
  1201. {
  1202. UINT32 deviceIdx = mDevice.getIndex();
  1203. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1204. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1205. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1206. }
  1207. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1208. }
  1209. if (instanceCount <= 0)
  1210. instanceCount = 1;
  1211. vkCmdDraw(mCmdBuffer, vertexCount, instanceCount, vertexOffset, 0);
  1212. }
  1213. void VulkanCmdBuffer::drawIndexed(UINT32 startIndex, UINT32 indexCount, UINT32 vertexOffset, UINT32 instanceCount)
  1214. {
  1215. if (!isReadyForRender())
  1216. return;
  1217. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1218. bindGpuParams();
  1219. if (!isInRenderPass())
  1220. beginRenderPass();
  1221. if (mGfxPipelineRequiresBind)
  1222. {
  1223. if (!bindGraphicsPipeline())
  1224. return;
  1225. }
  1226. else
  1227. bindDynamicStates(false);
  1228. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1229. {
  1230. if (mNumBoundDescriptorSets > 0)
  1231. {
  1232. UINT32 deviceIdx = mDevice.getIndex();
  1233. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1234. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1235. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1236. }
  1237. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1238. }
  1239. if (instanceCount <= 0)
  1240. instanceCount = 1;
  1241. vkCmdDrawIndexed(mCmdBuffer, indexCount, instanceCount, startIndex, vertexOffset, 0);
  1242. }
  1243. void VulkanCmdBuffer::dispatch(UINT32 numGroupsX, UINT32 numGroupsY, UINT32 numGroupsZ)
  1244. {
  1245. if (mComputePipeline == nullptr)
  1246. return;
  1247. if (isInRenderPass())
  1248. endRenderPass();
  1249. // Note: Should I restore the render target after? Note that this is only being done is framebuffer subresources
  1250. // have their "isFBAttachment" flag reset, potentially I can just clear/restore those
  1251. setRenderTarget(nullptr, 0, RT_ALL);
  1252. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1253. bindGpuParams();
  1254. executeLayoutTransitions();
  1255. UINT32 deviceIdx = mDevice.getIndex();
  1256. if(mCmpPipelineRequiresBind)
  1257. {
  1258. VulkanPipeline* pipeline = mComputePipeline->getPipeline(deviceIdx);
  1259. if (pipeline == nullptr)
  1260. return;
  1261. registerResource(pipeline, VulkanUseFlag::Read);
  1262. mComputePipeline->registerPipelineResources(this);
  1263. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline->getHandle());
  1264. mCmpPipelineRequiresBind = false;
  1265. }
  1266. if(mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Compute))
  1267. {
  1268. if (mNumBoundDescriptorSets > 0)
  1269. {
  1270. VkPipelineLayout pipelineLayout = mComputePipeline->getPipelineLayout(deviceIdx);
  1271. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipelineLayout, 0,
  1272. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1273. }
  1274. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Compute);
  1275. }
  1276. vkCmdDispatch(mCmdBuffer, numGroupsX, numGroupsY, numGroupsZ);
  1277. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  1278. // and reset read-only state.
  1279. for(auto& entry : mPassTouchedSubresourceInfos)
  1280. {
  1281. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  1282. subresourceInfo.isShaderInput = false;
  1283. subresourceInfo.isReadOnly = true;
  1284. subresourceInfo.needsBarrier = false;
  1285. }
  1286. for (auto& entry : mBuffers)
  1287. entry.second.needsBarrier = false;
  1288. mPassTouchedSubresourceInfos.clear();
  1289. }
  1290. void VulkanCmdBuffer::setEvent(VulkanEvent* event)
  1291. {
  1292. if(isInRenderPass())
  1293. mQueuedEvents.push_back(event);
  1294. else
  1295. vkCmdSetEvent(mCmdBuffer, event->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  1296. }
  1297. void VulkanCmdBuffer::resetQuery(VulkanQuery* query)
  1298. {
  1299. if (isInRenderPass())
  1300. mQueuedQueryResets.push_back(query);
  1301. else
  1302. query->reset(mCmdBuffer);
  1303. }
  1304. void VulkanCmdBuffer::memoryBarrier(VkBuffer buffer, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1305. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage)
  1306. {
  1307. VkBufferMemoryBarrier barrier;
  1308. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  1309. barrier.pNext = nullptr;
  1310. barrier.srcAccessMask = srcAccessFlags;
  1311. barrier.dstAccessMask = dstAccessFlags;
  1312. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1313. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1314. barrier.buffer = buffer;
  1315. barrier.offset = 0;
  1316. barrier.size = VK_WHOLE_SIZE;
  1317. vkCmdPipelineBarrier(getHandle(),
  1318. srcStage,
  1319. dstStage,
  1320. 0, 0, nullptr,
  1321. 1, &barrier,
  1322. 0, nullptr);
  1323. }
  1324. void VulkanCmdBuffer::memoryBarrier(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1325. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage, VkImageLayout layout,
  1326. const VkImageSubresourceRange& range)
  1327. {
  1328. VkImageMemoryBarrier barrier;
  1329. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1330. barrier.pNext = nullptr;
  1331. barrier.srcAccessMask = srcAccessFlags;
  1332. barrier.dstAccessMask = dstAccessFlags;
  1333. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1334. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1335. barrier.image = image;
  1336. barrier.subresourceRange = range;
  1337. barrier.oldLayout = layout;
  1338. barrier.newLayout = layout;
  1339. vkCmdPipelineBarrier(getHandle(),
  1340. srcStage,
  1341. dstStage,
  1342. 0, 0, nullptr,
  1343. 0, nullptr,
  1344. 1, &barrier);
  1345. }
  1346. void VulkanCmdBuffer::setLayout(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1347. VkImageLayout oldLayout, VkImageLayout newLayout, const VkImageSubresourceRange& range)
  1348. {
  1349. VkImageMemoryBarrier barrier;
  1350. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1351. barrier.pNext = nullptr;
  1352. barrier.srcAccessMask = srcAccessFlags;
  1353. barrier.dstAccessMask = dstAccessFlags;
  1354. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1355. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1356. barrier.oldLayout = oldLayout;
  1357. barrier.newLayout = newLayout;
  1358. barrier.image = image;
  1359. barrier.subresourceRange = range;
  1360. VkPipelineStageFlags srcStage = getPipelineStageFlags(srcAccessFlags);
  1361. VkPipelineStageFlags dstStage = getPipelineStageFlags(dstAccessFlags);
  1362. vkCmdPipelineBarrier(getHandle(),
  1363. srcStage,
  1364. dstStage,
  1365. 0, 0, nullptr,
  1366. 0, nullptr,
  1367. 1, &barrier);
  1368. }
  1369. VkImageLayout VulkanCmdBuffer::getCurrentLayout(VulkanImage* image, const VkImageSubresourceRange& range,
  1370. bool inRenderPass)
  1371. {
  1372. UINT32 face = range.baseArrayLayer;
  1373. UINT32 mip = range.baseMipLevel;
  1374. // The assumption is that all the subresources in the range will have the same layout, as this should be handled
  1375. // by registerResource(), or by external code (in the case of transfers). So we only check the first subresource.
  1376. VulkanImageSubresource* subresource = image->getSubresource(face, mip);
  1377. auto iterFind = mImages.find(image);
  1378. if (iterFind == mImages.end())
  1379. return subresource->getLayout();
  1380. UINT32 imageInfoIdx = iterFind->second;
  1381. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1382. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1383. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1384. {
  1385. ImageSubresourceInfo& entry = subresourceInfos[i];
  1386. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1387. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1388. {
  1389. // If it's a FB attachment, retrieve its layout after the render pass begins
  1390. if(entry.isFBAttachment && inRenderPass && mFramebuffer)
  1391. {
  1392. RenderSurfaceMask readMask = getFBReadMask();
  1393. // Is it a depth-stencil attachment?
  1394. if(mFramebuffer->hasDepthAttachment() && mFramebuffer->getDepthStencilAttachment().image == image)
  1395. {
  1396. if (readMask.isSet(RT_DEPTH))
  1397. {
  1398. if (readMask.isSet(RT_STENCIL))
  1399. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL;
  1400. else // Depth readable but stencil isn't
  1401. return VK_IMAGE_LAYOUT_GENERAL;
  1402. }
  1403. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1404. }
  1405. else // It is a color attachment
  1406. {
  1407. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1408. for (UINT32 j = 0; j < numColorAttachments; j++)
  1409. {
  1410. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(j);
  1411. if (attachment.image == image)
  1412. {
  1413. if (readMask.isSet((RenderSurfaceMaskBits)(1 << attachment.index)))
  1414. return VK_IMAGE_LAYOUT_GENERAL;
  1415. else
  1416. return VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1417. }
  1418. }
  1419. }
  1420. }
  1421. return entry.requiredLayout;
  1422. }
  1423. }
  1424. return image->getOptimalLayout();
  1425. }
  1426. void VulkanCmdBuffer::registerResource(VulkanResource* res, VulkanUseFlags flags)
  1427. {
  1428. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1429. if(insertResult.second) // New element
  1430. {
  1431. ResourceUseHandle& useHandle = insertResult.first->second;
  1432. useHandle.used = false;
  1433. useHandle.flags = flags;
  1434. res->notifyBound();
  1435. }
  1436. else // Existing element
  1437. {
  1438. ResourceUseHandle& useHandle = insertResult.first->second;
  1439. assert(!useHandle.used);
  1440. useHandle.flags |= flags;
  1441. }
  1442. }
  1443. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range, VulkanUseFlags flags,
  1444. ResourceUsage usage)
  1445. {
  1446. VkImageLayout layout = res->getOptimalLayout();
  1447. registerResource(res, range, layout, layout, flags, usage);
  1448. }
  1449. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range,
  1450. VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags, ResourceUsage usage)
  1451. {
  1452. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1453. bool isTransfer = usage == ResourceUsage::Transfer;
  1454. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1455. // If binding it for write in a shader (not as color attachment or transfer op), we will need to issue a memory
  1456. // barrier if the image gets used again during this render pass, so remember this information.
  1457. bool needsBarrier = isShaderBind && flags.isSet(VulkanUseFlag::Write);
  1458. UINT32 nextImageInfoIdx = (UINT32)mImageInfos.size();
  1459. auto registerSubresourceInfo = [&](const VkImageSubresourceRange& subresourceRange)
  1460. {
  1461. mSubresourceInfoStorage.push_back(ImageSubresourceInfo());
  1462. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage.back();
  1463. subresourceInfo.currentLayout = newLayout;
  1464. subresourceInfo.initialLayout = newLayout;
  1465. subresourceInfo.requiredLayout = newLayout;
  1466. subresourceInfo.finalLayout = finalLayout;
  1467. subresourceInfo.range = subresourceRange;
  1468. subresourceInfo.isFBAttachment = isFBAttachment;
  1469. subresourceInfo.isShaderInput = isShaderBind;
  1470. if (!isTransfer)
  1471. {
  1472. subresourceInfo.hasTransitioned = false;
  1473. subresourceInfo.isReadOnly = !flags.isSet(VulkanUseFlag::Write);
  1474. }
  1475. else
  1476. {
  1477. subresourceInfo.hasTransitioned = true; // Transfers handle layout transitions externally (at this point they are assumed to already be done)
  1478. subresourceInfo.isReadOnly = true; // Doesn't matter for transfers
  1479. }
  1480. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1481. subresourceInfo.needsBarrier = needsBarrier;
  1482. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1483. };
  1484. auto insertResult = mImages.insert(std::make_pair(res, nextImageInfoIdx));
  1485. if (insertResult.second) // New element
  1486. {
  1487. UINT32 imageInfoIdx = insertResult.first->second;
  1488. mImageInfos.push_back(ImageInfo());
  1489. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1490. imageInfo.subresourceInfoIdx = (UINT32)mSubresourceInfoStorage.size();
  1491. imageInfo.numSubresourceInfos = 1;
  1492. imageInfo.useHandle.used = false;
  1493. imageInfo.useHandle.flags = flags;
  1494. registerSubresourceInfo(range);
  1495. res->notifyBound();
  1496. }
  1497. else // Existing element
  1498. {
  1499. UINT32 imageInfoIdx = insertResult.first->second;
  1500. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1501. assert(!imageInfo.useHandle.used);
  1502. imageInfo.useHandle.flags |= flags;
  1503. // See if there is an overlap between existing ranges and the new range. And if so break them up accordingly.
  1504. //// First test for the simplest and most common case (same range or no overlap) to avoid more complex
  1505. //// computations.
  1506. ImageSubresourceInfo* subresources = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1507. bool foundRange = false;
  1508. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1509. {
  1510. if(VulkanUtility::rangeOverlaps(subresources[i].range, range))
  1511. {
  1512. if (subresources[i].range.layerCount == range.layerCount &&
  1513. subresources[i].range.levelCount == range.levelCount &&
  1514. subresources[i].range.baseArrayLayer == range.baseArrayLayer &&
  1515. subresources[i].range.baseMipLevel == range.baseMipLevel)
  1516. {
  1517. // Just update existing range
  1518. updateSubresourceInfo(res, imageInfoIdx, subresources[i], newLayout, finalLayout, flags, usage);
  1519. mPassTouchedSubresourceInfos.insert(imageInfo.subresourceInfoIdx + i);
  1520. foundRange = true;
  1521. break;
  1522. }
  1523. break;
  1524. }
  1525. }
  1526. //// We'll need to update subresource ranges or add new ones. The hope is that this code is trigger VERY rarely
  1527. //// (for just a few specific textures per frame).
  1528. if (!foundRange)
  1529. {
  1530. std::array<VkImageSubresourceRange, 5> tempCutRanges;
  1531. bs_frame_mark();
  1532. {
  1533. // We orphan previously allocated memory (we reset it after submit() anyway)
  1534. UINT32 newSubresourceIdx = (UINT32)mSubresourceInfoStorage.size();
  1535. FrameVector<UINT32> cutOverlappingRanges;
  1536. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1537. {
  1538. UINT32 subresourceIdx = imageInfo.subresourceInfoIdx + i;
  1539. ImageSubresourceInfo& subresource = mSubresourceInfoStorage[subresourceIdx];
  1540. if (!VulkanUtility::rangeOverlaps(subresource.range, range))
  1541. {
  1542. // Just copy as is
  1543. mSubresourceInfoStorage.push_back(subresource);
  1544. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1545. }
  1546. else // Need to cut
  1547. {
  1548. UINT32 numCutRanges;
  1549. VulkanUtility::cutRange(subresource.range, range, tempCutRanges, numCutRanges);
  1550. for(UINT32 j = 0; j < numCutRanges; j++)
  1551. {
  1552. // Create a copy of the original subresource with the new range
  1553. ImageSubresourceInfo newInfo = mSubresourceInfoStorage[subresourceIdx];
  1554. newInfo.range = tempCutRanges[j];
  1555. if(VulkanUtility::rangeOverlaps(tempCutRanges[j], range))
  1556. {
  1557. // Update overlapping sub-resource range with new data from this range
  1558. updateSubresourceInfo(res, imageInfoIdx, newInfo, newLayout, finalLayout, flags, usage);
  1559. // Keep track of the overlapping ranges for later
  1560. cutOverlappingRanges.push_back((UINT32)mSubresourceInfoStorage.size());
  1561. }
  1562. mSubresourceInfoStorage.push_back(newInfo);
  1563. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1564. }
  1565. }
  1566. }
  1567. // Our range doesn't overlap with any existing ranges, so just add it
  1568. if(cutOverlappingRanges.size() == 0)
  1569. {
  1570. registerSubresourceInfo(range);
  1571. }
  1572. else // Search if overlapping ranges fully cover the requested range, and insert non-covered regions
  1573. {
  1574. FrameQueue<VkImageSubresourceRange> sourceRanges;
  1575. sourceRanges.push(range);
  1576. for(auto& entry : cutOverlappingRanges)
  1577. {
  1578. VkImageSubresourceRange& overlappingRange = mSubresourceInfoStorage[entry].range;
  1579. UINT32 numSourceRanges = (UINT32)sourceRanges.size();
  1580. for(UINT32 i = 0; i < numSourceRanges; i++)
  1581. {
  1582. VkImageSubresourceRange sourceRange = sourceRanges.front();
  1583. sourceRanges.pop();
  1584. UINT32 numCutRanges;
  1585. VulkanUtility::cutRange(sourceRange, overlappingRange, tempCutRanges, numCutRanges);
  1586. for(UINT32 j = 0; j < numCutRanges; j++)
  1587. {
  1588. // We only care about ranges outside of the ones we already covered
  1589. if(!VulkanUtility::rangeOverlaps(tempCutRanges[j], overlappingRange))
  1590. sourceRanges.push(tempCutRanges[j]);
  1591. }
  1592. }
  1593. }
  1594. // Any remaining range hasn't been covered yet
  1595. while(!sourceRanges.empty())
  1596. {
  1597. registerSubresourceInfo(sourceRanges.front());
  1598. sourceRanges.pop();
  1599. }
  1600. }
  1601. imageInfo.subresourceInfoIdx = newSubresourceIdx;
  1602. imageInfo.numSubresourceInfos = (UINT32)mSubresourceInfoStorage.size() - newSubresourceIdx;
  1603. }
  1604. bs_frame_clear();
  1605. }
  1606. }
  1607. // Register any sub-resources
  1608. for(UINT32 i = 0; i < range.layerCount; i++)
  1609. {
  1610. for(UINT32 j = 0; j < range.levelCount; j++)
  1611. {
  1612. UINT32 layer = range.baseArrayLayer + i;
  1613. UINT32 mipLevel = range.baseMipLevel + j;
  1614. registerResource(res->getSubresource(layer, mipLevel), flags);
  1615. }
  1616. }
  1617. }
  1618. void VulkanCmdBuffer::registerResource(VulkanBuffer* res, VkAccessFlags accessFlags, VulkanUseFlags flags)
  1619. {
  1620. bool isShaderWrite = (accessFlags & VK_ACCESS_SHADER_WRITE_BIT) != 0;
  1621. auto insertResult = mBuffers.insert(std::make_pair(res, BufferInfo()));
  1622. if (insertResult.second) // New element
  1623. {
  1624. BufferInfo& bufferInfo = insertResult.first->second;
  1625. bufferInfo.accessFlags = accessFlags;
  1626. bufferInfo.useHandle.used = false;
  1627. bufferInfo.useHandle.flags = flags;
  1628. // Any writes done on storage buffers will need explicit memory barriers (if read during the same pass) so
  1629. // we remember this, in case this buffers gets used later in this pass.
  1630. bufferInfo.needsBarrier = isShaderWrite;
  1631. res->notifyBound();
  1632. }
  1633. else // Existing element
  1634. {
  1635. BufferInfo& bufferInfo = insertResult.first->second;
  1636. assert(!bufferInfo.useHandle.used);
  1637. bufferInfo.useHandle.flags |= flags;
  1638. bufferInfo.accessFlags |= accessFlags;
  1639. // If the buffer was written to previously in this pass, and is now being used by a shader we need to issue
  1640. // a barrier to make those writes visible.
  1641. bool isShaderRead = (accessFlags & VK_ACCESS_SHADER_READ_BIT) != 0;
  1642. if(bufferInfo.needsBarrier && (isShaderRead || isShaderWrite))
  1643. {
  1644. // Need to end render pass in order to execute the barrier. Hopefully this won't trigger much since most
  1645. // shader writes are done during compute
  1646. if (isInRenderPass())
  1647. endRenderPass();
  1648. VkPipelineStageFlags stages =
  1649. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1650. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1651. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1652. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1653. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1654. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1655. VkBuffer buffer = res->getHandle();
  1656. memoryBarrier(buffer, VK_ACCESS_SHADER_WRITE_BIT, accessFlags, stages, stages);
  1657. bufferInfo.needsBarrier = isShaderWrite;
  1658. }
  1659. }
  1660. }
  1661. void VulkanCmdBuffer::registerResource(VulkanFramebuffer* res, RenderSurfaceMask loadMask, UINT32 readMask)
  1662. {
  1663. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1664. if (insertResult.second) // New element
  1665. {
  1666. ResourceUseHandle& useHandle = insertResult.first->second;
  1667. useHandle.used = false;
  1668. useHandle.flags = VulkanUseFlag::Write;
  1669. res->notifyBound();
  1670. }
  1671. else // Existing element
  1672. {
  1673. ResourceUseHandle& useHandle = insertResult.first->second;
  1674. assert(!useHandle.used);
  1675. useHandle.flags |= VulkanUseFlag::Write;
  1676. }
  1677. // Register any sub-resources
  1678. UINT32 numColorAttachments = res->getNumColorAttachments();
  1679. for (UINT32 i = 0; i < numColorAttachments; i++)
  1680. {
  1681. const VulkanFramebufferAttachment& attachment = res->getColorAttachment(i);
  1682. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1683. VkImageLayout layout;
  1684. if (loadMask.isSet((RenderSurfaceMaskBits)(1 << i)))
  1685. {
  1686. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1687. // transition this to a different layout when it begins, but we handle that elsewhere
  1688. layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1689. }
  1690. else
  1691. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1692. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1693. registerResource(attachment.image, range, layout, attachment.finalLayout, VulkanUseFlag::Write,
  1694. ResourceUsage::Framebuffer);
  1695. }
  1696. if(res->hasDepthAttachment())
  1697. {
  1698. const VulkanFramebufferAttachment& attachment = res->getDepthStencilAttachment();
  1699. VulkanUseFlag useFlag = VulkanUseFlag::Write;
  1700. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1701. VkImageLayout layout;
  1702. if (loadMask.isSet(RT_DEPTH) || loadMask.isSet(RT_STENCIL)) // Can't load one without the other
  1703. {
  1704. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1705. // transition this to a different layout when it begins, but we handle that elsewhere
  1706. layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1707. }
  1708. else
  1709. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1710. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1711. registerResource(attachment.image, range, layout, attachment.finalLayout, useFlag, ResourceUsage::Framebuffer);
  1712. }
  1713. }
  1714. bool VulkanCmdBuffer::updateSubresourceInfo(VulkanImage* image, UINT32 imageInfoIdx,
  1715. ImageSubresourceInfo& subresourceInfo, VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags,
  1716. ResourceUsage usage)
  1717. {
  1718. bool isTransfer = usage == ResourceUsage::Transfer;
  1719. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1720. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1721. // Transfers are always considered read only because this only matters for destination access mask, and transfers
  1722. // handle that externally
  1723. if(!isTransfer)
  1724. subresourceInfo.isReadOnly &= !flags.isSet(VulkanUseFlag::Write);
  1725. // For transfers, just assign new layout, external code does all the work
  1726. if (isTransfer)
  1727. subresourceInfo.requiredLayout = newLayout;
  1728. else
  1729. {
  1730. // New layout is valid, check for transitions (UNDEFINED signifies the caller doesn't want a layout transition)
  1731. if (newLayout != VK_IMAGE_LAYOUT_UNDEFINED)
  1732. {
  1733. // If layout transition was requested by framebuffer bind, respect it because render-pass will only accept a
  1734. // specific layout (in certain cases), and we have no choice.
  1735. // In the case when a FB attachment is also bound for shader reads, this will override the layout required for
  1736. // shader read (GENERAL or DEPTH_READ_ONLY), but that is fine because those transitions are handled
  1737. // automatically by render-pass layout transitions.
  1738. // Any other texture (non FB attachment) will only even be bound in a single layout and we can keep the one it
  1739. // was originally registered with.
  1740. if (isFBAttachment)
  1741. subresourceInfo.requiredLayout = newLayout;
  1742. else if (!subresourceInfo.isFBAttachment) // Layout transition is not being done on a FB image
  1743. {
  1744. // Check if the image had a layout previously assigned, and if so check if multiple different layouts
  1745. // were requested. In that case we wish to transfer the image to GENERAL layout.
  1746. bool firstUseInRenderPass = !subresourceInfo.isShaderInput && !subresourceInfo.isFBAttachment;
  1747. if (firstUseInRenderPass || subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  1748. subresourceInfo.requiredLayout = newLayout;
  1749. else if (subresourceInfo.requiredLayout != newLayout)
  1750. subresourceInfo.requiredLayout = VK_IMAGE_LAYOUT_GENERAL;
  1751. }
  1752. }
  1753. }
  1754. // If attached to FB, then the final layout is set by the FB (provided as layout param here), otherwise its
  1755. // the same as required layout
  1756. if (!isFBAttachment && !subresourceInfo.isFBAttachment)
  1757. subresourceInfo.finalLayout = subresourceInfo.requiredLayout;
  1758. else
  1759. {
  1760. if (isFBAttachment)
  1761. subresourceInfo.finalLayout = finalLayout;
  1762. }
  1763. // Queue layout transition, if not transfer. Transfer handle layout transitions externally.
  1764. if (!isTransfer)
  1765. {
  1766. // If we haven't done a layout transition yet, we can just overwrite the previously written values, and the
  1767. // transition will be handled as the first thing in submit(), otherwise we queue a non-initial transition
  1768. // below.
  1769. if (!subresourceInfo.hasTransitioned)
  1770. {
  1771. subresourceInfo.initialLayout = subresourceInfo.requiredLayout;
  1772. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1773. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1774. }
  1775. else
  1776. {
  1777. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1778. mQueuedLayoutTransitions[image] = imageInfoIdx;
  1779. }
  1780. }
  1781. else
  1782. {
  1783. // Resource has already transitioned its layout externally since this is a transfer. We cannot allow a
  1784. // transition in submit().
  1785. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1786. subresourceInfo.hasTransitioned = true;
  1787. }
  1788. // If a FB attachment was just bound as a shader input, we might need to restart the render pass with a FB
  1789. // attachment that supports read-only attachments using the GENERAL or DEPTH_READ_ONLY layout
  1790. bool resetRenderPass = false;
  1791. if (isFBAttachment)
  1792. {
  1793. if (!subresourceInfo.isFBAttachment)
  1794. {
  1795. subresourceInfo.isFBAttachment = true;
  1796. resetRenderPass = subresourceInfo.isShaderInput;
  1797. }
  1798. }
  1799. else
  1800. {
  1801. if (!subresourceInfo.isShaderInput)
  1802. {
  1803. subresourceInfo.isShaderInput = true;
  1804. if (subresourceInfo.isFBAttachment)
  1805. {
  1806. // Special case for depth: If user has set up proper read-only flags, then the render pass will have
  1807. // taken care of setting the valid state anyway, so no need to end the render pass
  1808. if (subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL)
  1809. {
  1810. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0 || (mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1811. resetRenderPass = false;
  1812. else
  1813. resetRenderPass = true;
  1814. }
  1815. else
  1816. resetRenderPass = true;
  1817. }
  1818. else
  1819. {
  1820. // Subresource has been bound as shader input, and it wasn't bound as a FB attachment this render pass.
  1821. // However is could have been bound in a previous pass, so check the layouts and force a layout
  1822. // transition if required.
  1823. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1824. resetRenderPass = true;
  1825. else
  1826. resetRenderPass = false;
  1827. }
  1828. }
  1829. }
  1830. // If we need to switch frame-buffers, end current render pass
  1831. if (resetRenderPass && isInRenderPass())
  1832. endRenderPass();
  1833. else
  1834. {
  1835. // Since we won't be ending the render pass, check if this same sub-resource was written earlier in the pass,
  1836. // in which case we need to issue a memory barrier so those writes are visible.
  1837. // Memory barrier only matters if image is bound for shader use (no need for color attachments or transfers)
  1838. if(subresourceInfo.needsBarrier && isShaderBind)
  1839. {
  1840. bool isWrite = flags.isSet(VulkanUseFlag::Write);
  1841. VkPipelineStageFlags stages =
  1842. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1843. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1844. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1845. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1846. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1847. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1848. memoryBarrier(image->getHandle(), VK_ACCESS_SHADER_WRITE_BIT,
  1849. image->getAccessFlags(subresourceInfo.requiredLayout, !isWrite),
  1850. stages, stages, subresourceInfo.requiredLayout, subresourceInfo.range);
  1851. subresourceInfo.needsBarrier = isWrite;
  1852. }
  1853. }
  1854. return resetRenderPass;
  1855. }
  1856. VulkanCmdBuffer::ImageSubresourceInfo& VulkanCmdBuffer::findSubresourceInfo(VulkanImage* image, UINT32 face, UINT32 mip)
  1857. {
  1858. UINT32 imageInfoIdx = mImages[image];
  1859. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1860. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1861. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1862. {
  1863. ImageSubresourceInfo& entry = subresourceInfos[i];
  1864. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1865. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1866. {
  1867. return entry;
  1868. }
  1869. }
  1870. assert(false); // Caller should ensure the subresource actually exists, so this shouldn't happen
  1871. return subresourceInfos[0];
  1872. }
  1873. void VulkanCmdBuffer::getInProgressQueries(Vector<VulkanTimerQuery*>& timer, Vector<VulkanOcclusionQuery*>& occlusion) const
  1874. {
  1875. for(auto& query : mTimerQueries)
  1876. {
  1877. if (query->_isInProgress())
  1878. timer.push_back(query);
  1879. }
  1880. for (auto& query : mOcclusionQueries)
  1881. {
  1882. if (query->_isInProgress())
  1883. occlusion.push_back(query);
  1884. }
  1885. }
  1886. VulkanCommandBuffer::VulkanCommandBuffer(VulkanDevice& device, GpuQueueType type, UINT32 deviceIdx,
  1887. UINT32 queueIdx, bool secondary)
  1888. : CommandBuffer(type, deviceIdx, queueIdx, secondary), mBuffer(nullptr)
  1889. , mDevice(device), mQueue(nullptr), mIdMask(0)
  1890. {
  1891. UINT32 numQueues = device.getNumQueues(mType);
  1892. if (numQueues == 0) // Fall back to graphics queue
  1893. {
  1894. mType = GQT_GRAPHICS;
  1895. numQueues = device.getNumQueues(GQT_GRAPHICS);
  1896. }
  1897. mQueue = device.getQueue(mType, mQueueIdx % numQueues);
  1898. mIdMask = device.getQueueMask(mType, mQueueIdx);
  1899. acquireNewBuffer();
  1900. }
  1901. RenderSurfaceMask VulkanCmdBuffer::getFBReadMask()
  1902. {
  1903. // Check if any frame-buffer attachments are also used as shader inputs, in which case we make them read-only
  1904. RenderSurfaceMask readMask = RT_NONE;
  1905. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1906. for(UINT32 i = 0; i < numColorAttachments; i++)
  1907. {
  1908. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1909. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1910. fbAttachment.surface.mipLevel);
  1911. bool readOnly = subresourceInfo.isShaderInput;
  1912. if(readOnly)
  1913. readMask.set((RenderSurfaceMaskBits)(1 << i));
  1914. }
  1915. if (mFramebuffer->hasDepthAttachment())
  1916. {
  1917. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1918. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1919. fbAttachment.surface.mipLevel);
  1920. bool readOnly = subresourceInfo.isShaderInput;
  1921. if (readOnly)
  1922. readMask.set(RT_DEPTH);
  1923. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0)
  1924. readMask.set(RT_DEPTH);
  1925. if ((mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1926. readMask.set(RT_STENCIL);
  1927. }
  1928. return readMask;
  1929. }
  1930. VulkanCommandBuffer::~VulkanCommandBuffer()
  1931. {
  1932. mBuffer->reset();
  1933. }
  1934. void VulkanCommandBuffer::acquireNewBuffer()
  1935. {
  1936. VulkanCmdBufferPool& pool = mDevice.getCmdBufferPool();
  1937. if (mBuffer != nullptr)
  1938. assert(mBuffer->isSubmitted());
  1939. UINT32 queueFamily = mDevice.getQueueFamily(mType);
  1940. mBuffer = pool.getBuffer(queueFamily, mIsSecondary);
  1941. }
  1942. void VulkanCommandBuffer::submit(UINT32 syncMask)
  1943. {
  1944. // Ignore myself
  1945. syncMask &= ~mIdMask;
  1946. if (mBuffer->isInRenderPass())
  1947. mBuffer->endRenderPass();
  1948. // Execute any queued layout transitions that weren't already handled by the render pass
  1949. mBuffer->executeLayoutTransitions();
  1950. // Interrupt any in-progress queries (no in-progress queries allowed during command buffer submit)
  1951. Vector<VulkanTimerQuery*> timerQueries;
  1952. Vector<VulkanOcclusionQuery*> occlusionQueries;
  1953. mBuffer->getInProgressQueries(timerQueries, occlusionQueries);
  1954. for (auto& query : timerQueries)
  1955. query->_interrupt(*mBuffer);
  1956. for (auto& query : occlusionQueries)
  1957. query->_interrupt(*mBuffer);
  1958. if (mBuffer->isRecording())
  1959. mBuffer->end();
  1960. if (mBuffer->isReadyForSubmit()) // Possibly nothing was recorded in the buffer
  1961. {
  1962. mBuffer->submit(mQueue, mQueueIdx, syncMask);
  1963. acquireNewBuffer();
  1964. gVulkanCBManager().refreshStates(mDeviceIdx);
  1965. }
  1966. // Resume interrupted queries on the new command buffer
  1967. for (auto& query : timerQueries)
  1968. query->_resume(*mBuffer);
  1969. for (auto& query : occlusionQueries)
  1970. query->_resume(*mBuffer);
  1971. }
  1972. }}