BsVulkanCommandBuffer.cpp 78 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425
  1. //********************************** Banshee Engine (www.banshee3d.com) **************************************************//
  2. //**************** Copyright (c) 2016 Marko Pintera ([email protected]). All rights reserved. **********************//
  3. #include "BsVulkanCommandBuffer.h"
  4. #include "Managers/BsVulkanCommandBufferManager.h"
  5. #include "BsVulkanUtility.h"
  6. #include "BsVulkanDevice.h"
  7. #include "BsVulkanGpuParams.h"
  8. #include "BsVulkanQueue.h"
  9. #include "BsVulkanTexture.h"
  10. #include "BsVulkanIndexBuffer.h"
  11. #include "BsVulkanVertexBuffer.h"
  12. #include "BsVulkanHardwareBuffer.h"
  13. #include "BsVulkanFramebuffer.h"
  14. #include "Managers/BsVulkanVertexInputManager.h"
  15. #include "BsVulkanEventQuery.h"
  16. #include "Managers/BsVulkanQueryManager.h"
  17. #include "BsVulkanSwapChain.h"
  18. #include "BsVulkanTimerQuery.h"
  19. #include "BsVulkanOcclusionQuery.h"
  20. #if BS_PLATFORM == BS_PLATFORM_WIN32
  21. #include "Win32/BsWin32RenderWindow.h"
  22. #elif BS_PLATFORM == BS_PLATFORM_LINUX
  23. #include "Linux/BsLinuxRenderWindow.h"
  24. #else
  25. static_assert(false, "Other platforms go here");
  26. #endif
  27. namespace bs { namespace ct
  28. {
  29. VulkanSemaphore::VulkanSemaphore(VulkanResourceManager* owner)
  30. :VulkanResource(owner, true)
  31. {
  32. VkSemaphoreCreateInfo semaphoreCI;
  33. semaphoreCI.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  34. semaphoreCI.pNext = nullptr;
  35. semaphoreCI.flags = 0;
  36. VkResult result = vkCreateSemaphore(owner->getDevice().getLogical(), &semaphoreCI, gVulkanAllocator, &mSemaphore);
  37. assert(result == VK_SUCCESS);
  38. }
  39. VulkanSemaphore::~VulkanSemaphore()
  40. {
  41. vkDestroySemaphore(mOwner->getDevice().getLogical(), mSemaphore, gVulkanAllocator);
  42. }
  43. VulkanCmdBufferPool::VulkanCmdBufferPool(VulkanDevice& device)
  44. :mDevice(device), mNextId(1)
  45. {
  46. for (UINT32 i = 0; i < GQT_COUNT; i++)
  47. {
  48. UINT32 familyIdx = device.getQueueFamily((GpuQueueType)i);
  49. if (familyIdx == (UINT32)-1)
  50. continue;
  51. VkCommandPoolCreateInfo poolCI;
  52. poolCI.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  53. poolCI.pNext = nullptr;
  54. poolCI.flags = VK_COMMAND_POOL_CREATE_RESET_COMMAND_BUFFER_BIT;
  55. poolCI.queueFamilyIndex = familyIdx;
  56. PoolInfo& poolInfo = mPools[familyIdx];
  57. poolInfo.queueFamily = familyIdx;
  58. memset(poolInfo.buffers, 0, sizeof(poolInfo.buffers));
  59. vkCreateCommandPool(device.getLogical(), &poolCI, gVulkanAllocator, &poolInfo.pool);
  60. }
  61. }
  62. VulkanCmdBufferPool::~VulkanCmdBufferPool()
  63. {
  64. // Note: Shutdown should be the only place command buffers are destroyed at, as the system relies on the fact that
  65. // they won't be destroyed during normal operation.
  66. for(auto& entry : mPools)
  67. {
  68. PoolInfo& poolInfo = entry.second;
  69. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  70. {
  71. VulkanCmdBuffer* buffer = poolInfo.buffers[i];
  72. if (buffer == nullptr)
  73. break;
  74. bs_delete(buffer);
  75. }
  76. vkDestroyCommandPool(mDevice.getLogical(), poolInfo.pool, gVulkanAllocator);
  77. }
  78. }
  79. VulkanCmdBuffer* VulkanCmdBufferPool::getBuffer(UINT32 queueFamily, bool secondary)
  80. {
  81. auto iterFind = mPools.find(queueFamily);
  82. if (iterFind == mPools.end())
  83. return nullptr;
  84. VulkanCmdBuffer** buffers = iterFind->second.buffers;
  85. UINT32 i = 0;
  86. for(; i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY; i++)
  87. {
  88. if (buffers[i] == nullptr)
  89. break;
  90. if(buffers[i]->mState == VulkanCmdBuffer::State::Ready)
  91. {
  92. buffers[i]->begin();
  93. return buffers[i];
  94. }
  95. }
  96. assert(i < BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY &&
  97. "Too many command buffers allocated. Increment BS_MAX_VULKAN_CB_PER_QUEUE_FAMILY to a higher value. ");
  98. buffers[i] = createBuffer(queueFamily, secondary);
  99. buffers[i]->begin();
  100. return buffers[i];
  101. }
  102. VulkanCmdBuffer* VulkanCmdBufferPool::createBuffer(UINT32 queueFamily, bool secondary)
  103. {
  104. auto iterFind = mPools.find(queueFamily);
  105. if (iterFind == mPools.end())
  106. return nullptr;
  107. const PoolInfo& poolInfo = iterFind->second;
  108. return bs_new<VulkanCmdBuffer>(mDevice, mNextId++, poolInfo.pool, poolInfo.queueFamily, secondary);
  109. }
  110. /** Returns a set of pipeline stages that can are allowed to be used for the specified set of access flags. */
  111. VkPipelineStageFlags getPipelineStageFlags(VkAccessFlags accessFlags)
  112. {
  113. VkPipelineStageFlags flags = 0;
  114. if ((accessFlags & VK_ACCESS_INDIRECT_COMMAND_READ_BIT) != 0)
  115. flags |= VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT;
  116. if ((accessFlags & (VK_ACCESS_INDEX_READ_BIT | VK_ACCESS_VERTEX_ATTRIBUTE_READ_BIT)) != 0)
  117. flags |= VK_PIPELINE_STAGE_VERTEX_INPUT_BIT;
  118. if ((accessFlags & (VK_ACCESS_UNIFORM_READ_BIT | VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_SHADER_WRITE_BIT)) != 0)
  119. {
  120. flags |= VK_PIPELINE_STAGE_VERTEX_SHADER_BIT | VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT;
  121. flags |= VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT | VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT;
  122. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  123. }
  124. if ((accessFlags & VK_ACCESS_INPUT_ATTACHMENT_READ_BIT) != 0)
  125. flags |= VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT;
  126. if ((accessFlags & (VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT)) != 0)
  127. flags |= VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  128. if ((accessFlags & (VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT)) != 0)
  129. flags |= VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT;
  130. if ((accessFlags & (VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT)) != 0)
  131. flags |= VK_PIPELINE_STAGE_TRANSFER_BIT;
  132. if ((accessFlags & (VK_ACCESS_HOST_READ_BIT | VK_ACCESS_HOST_WRITE_BIT)) != 0)
  133. flags |= VK_PIPELINE_STAGE_HOST_BIT;
  134. if (flags == 0)
  135. flags = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  136. return flags;
  137. }
  138. template<class T>
  139. void getPipelineStageFlags(const Vector<T>& barriers, VkPipelineStageFlags& src, VkPipelineStageFlags& dst)
  140. {
  141. for(auto& entry : barriers)
  142. {
  143. src |= getPipelineStageFlags(entry.srcAccessMask);
  144. dst |= getPipelineStageFlags(entry.dstAccessMask);
  145. }
  146. if (src == 0)
  147. src = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  148. if (dst == 0)
  149. dst = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  150. }
  151. VulkanCmdBuffer::VulkanCmdBuffer(VulkanDevice& device, UINT32 id, VkCommandPool pool, UINT32 queueFamily, bool secondary)
  152. : mId(id), mQueueFamily(queueFamily), mState(State::Ready), mDevice(device), mPool(pool)
  153. , mIntraQueueSemaphore(nullptr), mInterQueueSemaphores(), mNumUsedInterQueueSemaphores(0)
  154. , mFramebuffer(nullptr), mRenderTargetWidth(0)
  155. , mRenderTargetHeight(0), mRenderTargetReadOnlyFlags(0), mRenderTargetLoadMask(RT_NONE), mGlobalQueueIdx(-1)
  156. , mViewport(0.0f, 0.0f, 1.0f, 1.0f), mScissor(0, 0, 0, 0), mStencilRef(0), mDrawOp(DOT_TRIANGLE_LIST)
  157. , mNumBoundDescriptorSets(0), mGfxPipelineRequiresBind(true), mCmpPipelineRequiresBind(true)
  158. , mViewportRequiresBind(true), mStencilRefRequiresBind(true), mScissorRequiresBind(true), mBoundParamsDirty(false)
  159. , mClearValues(), mClearMask(), mSemaphoresTemp(BS_MAX_UNIQUE_QUEUES), mVertexBuffersTemp()
  160. , mVertexBufferOffsetsTemp()
  161. {
  162. UINT32 maxBoundDescriptorSets = device.getDeviceProperties().limits.maxBoundDescriptorSets;
  163. mDescriptorSetsTemp = (VkDescriptorSet*)bs_alloc(sizeof(VkDescriptorSet) * maxBoundDescriptorSets);
  164. VkCommandBufferAllocateInfo cmdBufferAllocInfo;
  165. cmdBufferAllocInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  166. cmdBufferAllocInfo.pNext = nullptr;
  167. cmdBufferAllocInfo.commandPool = pool;
  168. cmdBufferAllocInfo.level = secondary ? VK_COMMAND_BUFFER_LEVEL_SECONDARY : VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  169. cmdBufferAllocInfo.commandBufferCount = 1;
  170. VkResult result = vkAllocateCommandBuffers(mDevice.getLogical(), &cmdBufferAllocInfo, &mCmdBuffer);
  171. assert(result == VK_SUCCESS);
  172. VkFenceCreateInfo fenceCI;
  173. fenceCI.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  174. fenceCI.pNext = nullptr;
  175. fenceCI.flags = 0;
  176. result = vkCreateFence(mDevice.getLogical(), &fenceCI, gVulkanAllocator, &mFence);
  177. assert(result == VK_SUCCESS);
  178. }
  179. VulkanCmdBuffer::~VulkanCmdBuffer()
  180. {
  181. VkDevice device = mDevice.getLogical();
  182. if(mState == State::Submitted)
  183. {
  184. // Wait 1s
  185. UINT64 waitTime = 1000 * 1000 * 1000;
  186. VkResult result = vkWaitForFences(device, 1, &mFence, true, waitTime);
  187. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  188. if (result == VK_TIMEOUT)
  189. LOGWRN("Freeing a command buffer before done executing because fence wait expired!");
  190. // Resources have been marked as used, make sure to notify them we're done with them
  191. reset();
  192. }
  193. else if(mState != State::Ready)
  194. {
  195. // Notify any resources that they are no longer bound
  196. for (auto& entry : mResources)
  197. {
  198. ResourceUseHandle& useHandle = entry.second;
  199. assert(!useHandle.used);
  200. entry.first->notifyUnbound();
  201. }
  202. for (auto& entry : mImages)
  203. {
  204. UINT32 imageInfoIdx = entry.second;
  205. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  206. ResourceUseHandle& useHandle = imageInfo.useHandle;
  207. assert(!useHandle.used);
  208. entry.first->notifyUnbound();
  209. }
  210. for (auto& entry : mBuffers)
  211. {
  212. ResourceUseHandle& useHandle = entry.second.useHandle;
  213. assert(!useHandle.used);
  214. entry.first->notifyUnbound();
  215. }
  216. }
  217. if (mIntraQueueSemaphore != nullptr)
  218. mIntraQueueSemaphore->destroy();
  219. for(UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  220. {
  221. if (mInterQueueSemaphores[i] != nullptr)
  222. mInterQueueSemaphores[i]->destroy();
  223. }
  224. vkDestroyFence(device, mFence, gVulkanAllocator);
  225. vkFreeCommandBuffers(device, mPool, 1, &mCmdBuffer);
  226. bs_free(mDescriptorSetsTemp);
  227. }
  228. UINT32 VulkanCmdBuffer::getDeviceIdx() const
  229. {
  230. return mDevice.getIndex();
  231. }
  232. void VulkanCmdBuffer::begin()
  233. {
  234. assert(mState == State::Ready);
  235. VkCommandBufferBeginInfo beginInfo;
  236. beginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  237. beginInfo.pNext = nullptr;
  238. beginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  239. beginInfo.pInheritanceInfo = nullptr;
  240. VkResult result = vkBeginCommandBuffer(mCmdBuffer, &beginInfo);
  241. assert(result == VK_SUCCESS);
  242. mState = State::Recording;
  243. }
  244. void VulkanCmdBuffer::end()
  245. {
  246. assert(mState == State::Recording);
  247. // If a clear is queued, execute the render pass with no additional instructions
  248. if (mClearMask)
  249. executeClearPass();
  250. VkResult result = vkEndCommandBuffer(mCmdBuffer);
  251. assert(result == VK_SUCCESS);
  252. mState = State::RecordingDone;
  253. }
  254. void VulkanCmdBuffer::beginRenderPass()
  255. {
  256. assert(mState == State::Recording);
  257. if (mFramebuffer == nullptr)
  258. {
  259. LOGWRN("Attempting to begin a render pass but no render target is bound to the command buffer.");
  260. return;
  261. }
  262. if(mClearMask != CLEAR_NONE)
  263. {
  264. // If a previous clear is queued, but it doesn't match the rendered area, need to execute a separate pass
  265. // just for it
  266. Rect2I rtArea(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  267. if (mClearArea != rtArea)
  268. executeClearPass();
  269. }
  270. executeLayoutTransitions();
  271. RenderSurfaceMask readMask = getFBReadMask();
  272. VkRenderPassBeginInfo renderPassBeginInfo;
  273. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  274. renderPassBeginInfo.pNext = nullptr;
  275. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(mRenderTargetLoadMask, readMask, mClearMask);
  276. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(mRenderTargetLoadMask, readMask, mClearMask);
  277. renderPassBeginInfo.renderArea.offset.x = 0;
  278. renderPassBeginInfo.renderArea.offset.y = 0;
  279. renderPassBeginInfo.renderArea.extent.width = mRenderTargetWidth;
  280. renderPassBeginInfo.renderArea.extent.height = mRenderTargetHeight;
  281. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  282. renderPassBeginInfo.pClearValues = mClearValues.data();
  283. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  284. mClearMask = CLEAR_NONE;
  285. mState = State::RecordingRenderPass;
  286. }
  287. void VulkanCmdBuffer::endRenderPass()
  288. {
  289. assert(mState == State::RecordingRenderPass);
  290. vkCmdEndRenderPass(mCmdBuffer);
  291. // Execute any queued events
  292. for(auto& entry : mQueuedEvents)
  293. vkCmdSetEvent(mCmdBuffer, entry->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  294. mQueuedEvents.clear();
  295. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  296. // and reset read-only state.
  297. for(auto& entry : mPassTouchedSubresourceInfos)
  298. {
  299. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  300. subresourceInfo.isShaderInput = false;
  301. subresourceInfo.isReadOnly = true;
  302. subresourceInfo.needsBarrier = false;
  303. }
  304. for (auto& entry : mBuffers)
  305. entry.second.needsBarrier = false;
  306. mPassTouchedSubresourceInfos.clear();
  307. updateFinalLayouts();
  308. mState = State::Recording;
  309. // In case the same GPU params from last pass get used, this makes sure the states we reset above, get re-applied
  310. mBoundParamsDirty = true;
  311. }
  312. void VulkanCmdBuffer::allocateSemaphores(VkSemaphore* semaphores)
  313. {
  314. if (mIntraQueueSemaphore != nullptr)
  315. mIntraQueueSemaphore->destroy();
  316. mIntraQueueSemaphore = mDevice.getResourceManager().create<VulkanSemaphore>();
  317. semaphores[0] = mIntraQueueSemaphore->getHandle();
  318. for (UINT32 i = 0; i < BS_MAX_VULKAN_CB_DEPENDENCIES; i++)
  319. {
  320. if (mInterQueueSemaphores[i] != nullptr)
  321. mInterQueueSemaphores[i]->destroy();
  322. mInterQueueSemaphores[i] = mDevice.getResourceManager().create<VulkanSemaphore>();
  323. semaphores[i + 1] = mInterQueueSemaphores[i]->getHandle();
  324. }
  325. mNumUsedInterQueueSemaphores = 0;
  326. }
  327. VulkanSemaphore* VulkanCmdBuffer::requestInterQueueSemaphore() const
  328. {
  329. if (mNumUsedInterQueueSemaphores >= BS_MAX_VULKAN_CB_DEPENDENCIES)
  330. return nullptr;
  331. return mInterQueueSemaphores[mNumUsedInterQueueSemaphores++];
  332. }
  333. void VulkanCmdBuffer::submit(VulkanQueue* queue, UINT32 queueIdx, UINT32 syncMask)
  334. {
  335. assert(isReadyForSubmit());
  336. // Make sure to reset the CB fence before we submit it
  337. VkResult result = vkResetFences(mDevice.getLogical(), 1, &mFence);
  338. assert(result == VK_SUCCESS);
  339. // If there are any query resets needed, execute those first
  340. VulkanDevice& device = queue->getDevice();
  341. if(!mQueuedQueryResets.empty())
  342. {
  343. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  344. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  345. for (auto& entry : mQueuedQueryResets)
  346. entry->reset(vkCmdBuffer);
  347. cmdBuffer->end();
  348. queue->queueSubmit(cmdBuffer, nullptr, 0);
  349. mQueuedQueryResets.clear();
  350. }
  351. // Issue pipeline barriers for queue transitions (need to happen on original queue first, then on new queue)
  352. for (auto& entry : mBuffers)
  353. {
  354. VulkanBuffer* resource = static_cast<VulkanBuffer*>(entry.first);
  355. if (!resource->isExclusive())
  356. continue;
  357. UINT32 currentQueueFamily = resource->getQueueFamily();
  358. if (currentQueueFamily != (UINT32)-1 && currentQueueFamily != mQueueFamily)
  359. {
  360. Vector<VkBufferMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].bufferBarriers;
  361. barriers.push_back(VkBufferMemoryBarrier());
  362. VkBufferMemoryBarrier& barrier = barriers.back();
  363. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  364. barrier.pNext = nullptr;
  365. barrier.srcAccessMask = entry.second.accessFlags;
  366. barrier.dstAccessMask = entry.second.accessFlags;
  367. barrier.srcQueueFamilyIndex = currentQueueFamily;
  368. barrier.dstQueueFamilyIndex = mQueueFamily;
  369. barrier.buffer = resource->getHandle();
  370. barrier.offset = 0;
  371. barrier.size = VK_WHOLE_SIZE;
  372. }
  373. }
  374. // For images issue queue transitions, as above. Also issue layout transitions to their inital layouts.
  375. Vector<VkImageMemoryBarrier>& localBarriers = mTransitionInfoTemp[mQueueFamily].imageBarriers;
  376. for (auto& entry : mImages)
  377. {
  378. VulkanImage* resource = static_cast<VulkanImage*>(entry.first);
  379. ImageInfo& imageInfo = mImageInfos[entry.second];
  380. UINT32 currentQueueFamily = resource->getQueueFamily();
  381. bool queueMismatch = resource->isExclusive() && currentQueueFamily != (UINT32)-1
  382. && currentQueueFamily != mQueueFamily;
  383. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  384. if (queueMismatch)
  385. {
  386. Vector<VkImageMemoryBarrier>& barriers = mTransitionInfoTemp[currentQueueFamily].imageBarriers;
  387. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  388. {
  389. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  390. UINT32 startIdx = (UINT32)barriers.size();
  391. resource->getBarriers(subresourceInfo.range, barriers);
  392. for(UINT32 j = startIdx; j < (UINT32)barriers.size(); j++)
  393. {
  394. VkImageMemoryBarrier& barrier = barriers[j];
  395. barrier.dstAccessMask = resource->getAccessFlags(barrier.oldLayout);
  396. barrier.newLayout = barrier.oldLayout;
  397. barrier.srcQueueFamilyIndex = currentQueueFamily;
  398. barrier.dstQueueFamilyIndex = mQueueFamily;
  399. }
  400. }
  401. }
  402. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  403. {
  404. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  405. VkImageLayout initialLayout = subresourceInfo.initialLayout;
  406. if (initialLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  407. continue;
  408. const VkImageSubresourceRange& range = subresourceInfo.range;
  409. UINT32 mipEnd = range.baseMipLevel + range.levelCount;
  410. UINT32 faceEnd = range.baseArrayLayer + range.layerCount;
  411. if(!subresourceInfo.hasExternalTransition)
  412. {
  413. bool layoutMismatch = false;
  414. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  415. {
  416. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  417. {
  418. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  419. if (subresource->getLayout() != initialLayout)
  420. {
  421. layoutMismatch = true;
  422. break;
  423. }
  424. }
  425. if (layoutMismatch)
  426. break;
  427. }
  428. if (layoutMismatch)
  429. {
  430. UINT32 startIdx = (UINT32)localBarriers.size();
  431. resource->getBarriers(subresourceInfo.range, localBarriers);
  432. for (UINT32 j = startIdx; j < (UINT32)localBarriers.size(); j++)
  433. {
  434. VkImageMemoryBarrier& barrier = localBarriers[j];
  435. barrier.dstAccessMask = resource->getAccessFlags(initialLayout, subresourceInfo.isInitialReadOnly);
  436. barrier.newLayout = initialLayout;
  437. }
  438. }
  439. }
  440. for (UINT32 mip = range.baseMipLevel; mip < mipEnd; mip++)
  441. {
  442. for (UINT32 face = range.baseArrayLayer; face < faceEnd; face++)
  443. {
  444. VulkanImageSubresource* subresource = resource->getSubresource(face, mip);
  445. subresource->setLayout(subresourceInfo.finalLayout);
  446. }
  447. }
  448. }
  449. }
  450. for (auto& entry : mTransitionInfoTemp)
  451. {
  452. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  453. if (empty)
  454. continue;
  455. UINT32 entryQueueFamily = entry.first;
  456. // No queue transition needed for entries on this queue (this entry is most likely an image layout transition)
  457. if (entryQueueFamily == (UINT32)-1 || entryQueueFamily == mQueueFamily)
  458. continue;
  459. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(entryQueueFamily, false);
  460. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  461. TransitionInfo& barriers = entry.second;
  462. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  463. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  464. VkPipelineStageFlags srcStage = 0;
  465. VkPipelineStageFlags dstStage = 0;
  466. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  467. vkCmdPipelineBarrier(vkCmdBuffer,
  468. srcStage, dstStage,
  469. 0, 0, nullptr,
  470. numBufferBarriers, barriers.bufferBarriers.data(),
  471. numImgBarriers, barriers.imageBarriers.data());
  472. // Find an appropriate queue to execute on
  473. UINT32 otherQueueIdx = 0;
  474. VulkanQueue* otherQueue = nullptr;
  475. GpuQueueType otherQueueType = GQT_GRAPHICS;
  476. for (UINT32 i = 0; i < GQT_COUNT; i++)
  477. {
  478. otherQueueType = (GpuQueueType)i;
  479. if (device.getQueueFamily(otherQueueType) != entryQueueFamily)
  480. continue;
  481. UINT32 numQueues = device.getNumQueues(otherQueueType);
  482. for (UINT32 j = 0; j < numQueues; j++)
  483. {
  484. // Try to find a queue not currently executing
  485. VulkanQueue* curQueue = device.getQueue(otherQueueType, j);
  486. if (!curQueue->isExecuting())
  487. {
  488. otherQueue = curQueue;
  489. otherQueueIdx = j;
  490. }
  491. }
  492. // Can't find empty one, use the first one then
  493. if (otherQueue == nullptr)
  494. {
  495. otherQueue = device.getQueue(otherQueueType, 0);
  496. otherQueueIdx = 0;
  497. }
  498. break;
  499. }
  500. syncMask |= CommandSyncMask::getGlobalQueueMask(otherQueueType, otherQueueIdx);
  501. cmdBuffer->end();
  502. // Note: If I switch back to doing layout transitions here, I need to wait on present semaphore
  503. otherQueue->submit(cmdBuffer, nullptr, 0);
  504. }
  505. UINT32 deviceIdx = device.getIndex();
  506. VulkanCommandBufferManager& cbm = static_cast<VulkanCommandBufferManager&>(CommandBufferManager::instance());
  507. UINT32 numSemaphores;
  508. cbm.getSyncSemaphores(deviceIdx, syncMask, mSemaphoresTemp.data(), numSemaphores);
  509. // Wait on present (i.e. until the back buffer becomes available) for any swap chains
  510. for(auto& entry : mSwapChains)
  511. {
  512. const SwapChainSurface& surface = entry->getBackBuffer();
  513. if (surface.needsWait)
  514. {
  515. VulkanSemaphore* semaphore = entry->getBackBuffer().sync;
  516. if (numSemaphores >= (UINT32)mSemaphoresTemp.size())
  517. mSemaphoresTemp.push_back(semaphore);
  518. else
  519. mSemaphoresTemp[numSemaphores] = semaphore;
  520. numSemaphores++;
  521. entry->notifyBackBufferWaitIssued();
  522. }
  523. }
  524. // Issue second part of transition pipeline barriers (on this queue)
  525. for (auto& entry : mTransitionInfoTemp)
  526. {
  527. bool empty = entry.second.imageBarriers.size() == 0 && entry.second.bufferBarriers.size() == 0;
  528. if (empty)
  529. continue;
  530. VulkanCmdBuffer* cmdBuffer = device.getCmdBufferPool().getBuffer(mQueueFamily, false);
  531. VkCommandBuffer vkCmdBuffer = cmdBuffer->getHandle();
  532. TransitionInfo& barriers = entry.second;
  533. UINT32 numImgBarriers = (UINT32)barriers.imageBarriers.size();
  534. UINT32 numBufferBarriers = (UINT32)barriers.bufferBarriers.size();
  535. VkPipelineStageFlags srcStage = 0;
  536. VkPipelineStageFlags dstStage = 0;
  537. getPipelineStageFlags(barriers.imageBarriers, srcStage, dstStage);
  538. vkCmdPipelineBarrier(vkCmdBuffer,
  539. srcStage, dstStage,
  540. 0, 0, nullptr,
  541. numBufferBarriers, barriers.bufferBarriers.data(),
  542. numImgBarriers, barriers.imageBarriers.data());
  543. cmdBuffer->end();
  544. queue->queueSubmit(cmdBuffer, mSemaphoresTemp.data(), numSemaphores);
  545. numSemaphores = 0; // Semaphores are only needed the first time, since we're adding the buffers on the same queue
  546. }
  547. queue->queueSubmit(this, mSemaphoresTemp.data(), numSemaphores);
  548. queue->submitQueued();
  549. mGlobalQueueIdx = CommandSyncMask::getGlobalQueueIdx(queue->getType(), queueIdx);
  550. for (auto& entry : mResources)
  551. {
  552. ResourceUseHandle& useHandle = entry.second;
  553. assert(!useHandle.used);
  554. useHandle.used = true;
  555. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  556. }
  557. for (auto& entry : mImages)
  558. {
  559. UINT32 imageInfoIdx = entry.second;
  560. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  561. ResourceUseHandle& useHandle = imageInfo.useHandle;
  562. assert(!useHandle.used);
  563. useHandle.used = true;
  564. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  565. }
  566. for (auto& entry : mBuffers)
  567. {
  568. ResourceUseHandle& useHandle = entry.second.useHandle;
  569. assert(!useHandle.used);
  570. useHandle.used = true;
  571. entry.first->notifyUsed(mGlobalQueueIdx, mQueueFamily, useHandle.flags);
  572. }
  573. // Note: Uncomment for debugging only, prevents any device concurrency issues.
  574. // vkQueueWaitIdle(queue->getHandle());
  575. // Clear vectors but don't clear the actual map, as we want to re-use the memory since we expect queue family
  576. // indices to be the same
  577. for (auto& entry : mTransitionInfoTemp)
  578. {
  579. entry.second.imageBarriers.clear();
  580. entry.second.bufferBarriers.clear();
  581. }
  582. mGraphicsPipeline = nullptr;
  583. mComputePipeline = nullptr;
  584. mGfxPipelineRequiresBind = true;
  585. mCmpPipelineRequiresBind = true;
  586. mFramebuffer = nullptr;
  587. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  588. mQueuedLayoutTransitions.clear();
  589. mBoundParams = nullptr;
  590. mSwapChains.clear();
  591. }
  592. bool VulkanCmdBuffer::checkFenceStatus(bool block) const
  593. {
  594. VkResult result = vkWaitForFences(mDevice.getLogical(), 1, &mFence, true, block ? 1'000'000'000 : 0);
  595. assert(result == VK_SUCCESS || result == VK_TIMEOUT);
  596. return result == VK_SUCCESS;
  597. }
  598. void VulkanCmdBuffer::reset()
  599. {
  600. bool wasSubmitted = mState == State::Submitted;
  601. mState = State::Ready;
  602. vkResetCommandBuffer(mCmdBuffer, VK_COMMAND_BUFFER_RESET_RELEASE_RESOURCES_BIT); // Note: Maybe better not to release resources?
  603. if (wasSubmitted)
  604. {
  605. for (auto& entry : mResources)
  606. {
  607. ResourceUseHandle& useHandle = entry.second;
  608. assert(useHandle.used);
  609. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  610. }
  611. for (auto& entry : mImages)
  612. {
  613. UINT32 imageInfoIdx = entry.second;
  614. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  615. ResourceUseHandle& useHandle = imageInfo.useHandle;
  616. assert(useHandle.used);
  617. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  618. }
  619. for (auto& entry : mBuffers)
  620. {
  621. ResourceUseHandle& useHandle = entry.second.useHandle;
  622. assert(useHandle.used);
  623. entry.first->notifyDone(mGlobalQueueIdx, useHandle.flags);
  624. }
  625. }
  626. else
  627. {
  628. for (auto& entry : mResources)
  629. entry.first->notifyUnbound();
  630. for (auto& entry : mImages)
  631. entry.first->notifyUnbound();
  632. for (auto& entry : mBuffers)
  633. entry.first->notifyUnbound();
  634. }
  635. mResources.clear();
  636. mImages.clear();
  637. mBuffers.clear();
  638. mOcclusionQueries.clear();
  639. mTimerQueries.clear();
  640. mImageInfos.clear();
  641. mSubresourceInfoStorage.clear();
  642. mPassTouchedSubresourceInfos.clear();
  643. }
  644. void VulkanCmdBuffer::setRenderTarget(const SPtr<RenderTarget>& rt, UINT32 readOnlyFlags,
  645. RenderSurfaceMask loadMask)
  646. {
  647. assert(mState != State::Submitted);
  648. VulkanFramebuffer* newFB;
  649. if(rt != nullptr)
  650. {
  651. if (rt->getProperties().isWindow)
  652. {
  653. #if BS_PLATFORM == BS_PLATFORM_WIN32
  654. Win32RenderWindow* window = static_cast<Win32RenderWindow*>(rt.get());
  655. #elif BS_PLATFORM == BS_PLATFORM_LINUX
  656. LinuxRenderWindow* window = static_cast<LinuxRenderWindow*>(rt.get());
  657. #endif
  658. window->acquireBackBuffer();
  659. VulkanSwapChain* swapChain;
  660. rt->getCustomAttribute("SC", &swapChain);
  661. mSwapChains.insert(swapChain);
  662. }
  663. rt->getCustomAttribute("FB", &newFB);
  664. }
  665. else
  666. {
  667. newFB = nullptr;
  668. }
  669. // Warn if invalid load mask
  670. if (loadMask.isSet(RT_DEPTH) && !loadMask.isSet(RT_STENCIL))
  671. {
  672. LOGWRN("setRenderTarget() invalid load mask, depth enabled but stencil disabled. This is not supported. Both \
  673. will be loaded.");
  674. loadMask.set(RT_STENCIL);
  675. }
  676. if (!loadMask.isSet(RT_DEPTH) && loadMask.isSet(RT_STENCIL))
  677. {
  678. LOGWRN("setRenderTarget() invalid load mask, stencil enabled but depth disabled. This is not supported. Both \
  679. will be loaded.");
  680. loadMask.set(RT_DEPTH);
  681. }
  682. if (mFramebuffer == newFB && mRenderTargetReadOnlyFlags == readOnlyFlags && mRenderTargetLoadMask == loadMask)
  683. return;
  684. if (isInRenderPass())
  685. endRenderPass();
  686. else
  687. {
  688. // If a clear is queued for previous FB, execute the render pass with no additional instructions
  689. if (mClearMask)
  690. executeClearPass();
  691. }
  692. // Reset isFBAttachment flags for subresources from the old framebuffer
  693. if(mFramebuffer != nullptr)
  694. {
  695. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  696. for(UINT32 i = 0; i < numColorAttachments; i++)
  697. {
  698. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  699. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  700. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  701. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  702. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  703. {
  704. ImageSubresourceInfo& entry = subresourceInfos[j];
  705. entry.isFBAttachment = false;
  706. }
  707. }
  708. if(mFramebuffer->hasDepthAttachment())
  709. {
  710. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  711. UINT32 imageInfoIdx = mImages[fbAttachment.image];
  712. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  713. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  714. for(UINT32 j = 0; j < imageInfo.numSubresourceInfos; j++)
  715. {
  716. ImageSubresourceInfo& entry = subresourceInfos[j];
  717. entry.isFBAttachment = false;
  718. }
  719. }
  720. }
  721. if(newFB == nullptr)
  722. {
  723. mFramebuffer = nullptr;
  724. mRenderTargetWidth = 0;
  725. mRenderTargetHeight = 0;
  726. mRenderTargetReadOnlyFlags = 0;
  727. mRenderTargetLoadMask = RT_NONE;
  728. }
  729. else
  730. {
  731. mFramebuffer = newFB;
  732. mRenderTargetWidth = rt->getProperties().width;
  733. mRenderTargetHeight = rt->getProperties().height;
  734. mRenderTargetReadOnlyFlags = readOnlyFlags;
  735. mRenderTargetLoadMask = loadMask;
  736. }
  737. // Re-set the params as they will need to be re-bound
  738. setGpuParams(mBoundParams);
  739. if (mFramebuffer != nullptr)
  740. registerResource(mFramebuffer, loadMask, readOnlyFlags);
  741. mGfxPipelineRequiresBind = true;
  742. }
  743. void VulkanCmdBuffer::clearViewport(const Rect2I& area, UINT32 buffers, const Color& color, float depth, UINT16 stencil,
  744. UINT8 targetMask)
  745. {
  746. if (buffers == 0 || mFramebuffer == nullptr)
  747. return;
  748. // Add clear command if currently in render pass
  749. if (isInRenderPass())
  750. {
  751. VkClearAttachment attachments[BS_MAX_MULTIPLE_RENDER_TARGETS + 1];
  752. UINT32 baseLayer = 0;
  753. UINT32 attachmentIdx = 0;
  754. if ((buffers & FBT_COLOR) != 0)
  755. {
  756. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  757. for (UINT32 i = 0; i < numColorAttachments; i++)
  758. {
  759. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  760. if (((1 << attachment.index) & targetMask) == 0)
  761. continue;
  762. attachments[attachmentIdx].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  763. attachments[attachmentIdx].colorAttachment = i;
  764. VkClearColorValue& colorValue = attachments[attachmentIdx].clearValue.color;
  765. colorValue.float32[0] = color.r;
  766. colorValue.float32[1] = color.g;
  767. colorValue.float32[2] = color.b;
  768. colorValue.float32[3] = color.a;
  769. UINT32 curBaseLayer = attachment.baseLayer;
  770. if (attachmentIdx == 0)
  771. baseLayer = curBaseLayer;
  772. else
  773. {
  774. if (baseLayer != curBaseLayer)
  775. {
  776. // Note: This could be supported relatively easily: we would need to issue multiple separate
  777. // clear commands for such framebuffers.
  778. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  779. "starting layers. This is currently not supported.");
  780. }
  781. }
  782. attachmentIdx++;
  783. }
  784. }
  785. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  786. {
  787. if (mFramebuffer->hasDepthAttachment())
  788. {
  789. attachments[attachmentIdx].aspectMask = 0;
  790. if ((buffers & FBT_DEPTH) != 0)
  791. {
  792. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_DEPTH_BIT;
  793. attachments[attachmentIdx].clearValue.depthStencil.depth = depth;
  794. }
  795. if ((buffers & FBT_STENCIL) != 0)
  796. {
  797. attachments[attachmentIdx].aspectMask |= VK_IMAGE_ASPECT_STENCIL_BIT;
  798. attachments[attachmentIdx].clearValue.depthStencil.stencil = stencil;
  799. }
  800. attachments[attachmentIdx].colorAttachment = 0;
  801. UINT32 curBaseLayer = mFramebuffer->getDepthStencilAttachment().baseLayer;
  802. if (attachmentIdx == 0)
  803. baseLayer = curBaseLayer;
  804. else
  805. {
  806. if (baseLayer != curBaseLayer)
  807. {
  808. // Note: This could be supported relatively easily: we would need to issue multiple separate
  809. // clear commands for such framebuffers.
  810. LOGERR("Attempting to clear a texture that has multiple multi-layer surfaces with mismatching "
  811. "starting layers. This is currently not supported.");
  812. }
  813. }
  814. attachmentIdx++;
  815. }
  816. }
  817. UINT32 numAttachments = attachmentIdx;
  818. if (numAttachments == 0)
  819. return;
  820. VkClearRect clearRect;
  821. clearRect.baseArrayLayer = baseLayer;
  822. clearRect.layerCount = mFramebuffer->getNumLayers();
  823. clearRect.rect.offset.x = area.x;
  824. clearRect.rect.offset.y = area.y;
  825. clearRect.rect.extent.width = area.width;
  826. clearRect.rect.extent.height = area.height;
  827. vkCmdClearAttachments(mCmdBuffer, numAttachments, attachments, 1, &clearRect);
  828. }
  829. // Otherwise we use a render pass that performs a clear on begin
  830. else
  831. {
  832. ClearMask clearMask;
  833. std::array<VkClearValue, BS_MAX_MULTIPLE_RENDER_TARGETS + 1> clearValues = mClearValues;
  834. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  835. if ((buffers & FBT_COLOR) != 0)
  836. {
  837. for (UINT32 i = 0; i < numColorAttachments; i++)
  838. {
  839. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(i);
  840. if (((1 << attachment.index) & targetMask) == 0)
  841. continue;
  842. clearMask |= (ClearMaskBits)(1 << attachment.index);
  843. VkClearColorValue& colorValue = clearValues[i].color;
  844. colorValue.float32[0] = color.r;
  845. colorValue.float32[1] = color.g;
  846. colorValue.float32[2] = color.b;
  847. colorValue.float32[3] = color.a;
  848. }
  849. }
  850. if ((buffers & FBT_DEPTH) != 0 || (buffers & FBT_STENCIL) != 0)
  851. {
  852. if (mFramebuffer->hasDepthAttachment())
  853. {
  854. UINT32 depthAttachmentIdx = numColorAttachments;
  855. if ((buffers & FBT_DEPTH) != 0)
  856. {
  857. clearValues[depthAttachmentIdx].depthStencil.depth = depth;
  858. clearMask |= CLEAR_DEPTH;
  859. }
  860. if ((buffers & FBT_STENCIL) != 0)
  861. {
  862. clearValues[depthAttachmentIdx].depthStencil.stencil = stencil;
  863. clearMask |= CLEAR_STENCIL;
  864. }
  865. }
  866. }
  867. if (!clearMask)
  868. return;
  869. // Some previous clear operation is already queued, execute it first
  870. bool previousClearNeedsToFinish = (mClearMask & clearMask) != CLEAR_NONE;
  871. if(previousClearNeedsToFinish)
  872. executeClearPass();
  873. mClearMask |= clearMask;
  874. mClearValues = clearValues;
  875. mClearArea = area;
  876. }
  877. }
  878. void VulkanCmdBuffer::clearRenderTarget(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  879. {
  880. Rect2I area(0, 0, mRenderTargetWidth, mRenderTargetHeight);
  881. clearViewport(area, buffers, color, depth, stencil, targetMask);
  882. }
  883. void VulkanCmdBuffer::clearViewport(UINT32 buffers, const Color& color, float depth, UINT16 stencil, UINT8 targetMask)
  884. {
  885. Rect2I area;
  886. area.x = (UINT32)(mViewport.x * mRenderTargetWidth);
  887. area.y = (UINT32)(mViewport.y * mRenderTargetHeight);
  888. area.width = (UINT32)(mViewport.width * mRenderTargetWidth);
  889. area.height = (UINT32)(mViewport.height * mRenderTargetHeight);
  890. clearViewport(area, buffers, color, depth, stencil, targetMask);
  891. }
  892. void VulkanCmdBuffer::setPipelineState(const SPtr<GraphicsPipelineState>& state)
  893. {
  894. if (mGraphicsPipeline == state)
  895. return;
  896. mGraphicsPipeline = std::static_pointer_cast<VulkanGraphicsPipelineState>(state);
  897. mGfxPipelineRequiresBind = true;
  898. }
  899. void VulkanCmdBuffer::setPipelineState(const SPtr<ComputePipelineState>& state)
  900. {
  901. if (mComputePipeline == state)
  902. return;
  903. mComputePipeline = std::static_pointer_cast<VulkanComputePipelineState>(state);
  904. mCmpPipelineRequiresBind = true;
  905. }
  906. void VulkanCmdBuffer::setGpuParams(const SPtr<GpuParams>& gpuParams)
  907. {
  908. // Note: We keep an internal reference to GPU params even though we shouldn't keep a reference to a core thread
  909. // object. But it should be fine since we expect the resource to be externally synchronized so it should never
  910. // be allowed to go out of scope on a non-core thread anyway.
  911. mBoundParams = std::static_pointer_cast<VulkanGpuParams>(gpuParams);
  912. if (mBoundParams != nullptr)
  913. mBoundParamsDirty = true;
  914. else
  915. {
  916. mNumBoundDescriptorSets = 0;
  917. mBoundParamsDirty = false;
  918. }
  919. mDescriptorSetsBindState = DescriptorSetBindFlag::Graphics | DescriptorSetBindFlag::Compute;
  920. }
  921. void VulkanCmdBuffer::setViewport(const Rect2& area)
  922. {
  923. if (mViewport == area)
  924. return;
  925. mViewport = area;
  926. mViewportRequiresBind = true;
  927. }
  928. void VulkanCmdBuffer::setScissorRect(const Rect2I& value)
  929. {
  930. if (mScissor == value)
  931. return;
  932. mScissor = value;
  933. mScissorRequiresBind = true;
  934. }
  935. void VulkanCmdBuffer::setStencilRef(UINT32 value)
  936. {
  937. if (mStencilRef == value)
  938. return;
  939. mStencilRef = value;
  940. mStencilRefRequiresBind = true;
  941. }
  942. void VulkanCmdBuffer::setDrawOp(DrawOperationType drawOp)
  943. {
  944. if (mDrawOp == drawOp)
  945. return;
  946. mDrawOp = drawOp;
  947. mGfxPipelineRequiresBind = true;
  948. }
  949. void VulkanCmdBuffer::setVertexBuffers(UINT32 index, SPtr<VertexBuffer>* buffers, UINT32 numBuffers)
  950. {
  951. if (numBuffers == 0)
  952. return;
  953. for(UINT32 i = 0; i < numBuffers; i++)
  954. {
  955. VulkanVertexBuffer* vertexBuffer = static_cast<VulkanVertexBuffer*>(buffers[i].get());
  956. if (vertexBuffer != nullptr)
  957. {
  958. VulkanBuffer* resource = vertexBuffer->getResource(mDevice.getIndex());
  959. if (resource != nullptr)
  960. {
  961. mVertexBuffersTemp[i] = resource->getHandle();
  962. registerResource(resource, VulkanUseFlag::Read);
  963. }
  964. else
  965. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  966. }
  967. else
  968. mVertexBuffersTemp[i] = VK_NULL_HANDLE;
  969. }
  970. vkCmdBindVertexBuffers(mCmdBuffer, index, numBuffers, mVertexBuffersTemp, mVertexBufferOffsetsTemp);
  971. }
  972. void VulkanCmdBuffer::setIndexBuffer(const SPtr<IndexBuffer>& buffer)
  973. {
  974. VulkanIndexBuffer* indexBuffer = static_cast<VulkanIndexBuffer*>(buffer.get());
  975. VkBuffer vkBuffer = VK_NULL_HANDLE;
  976. VkIndexType indexType = VK_INDEX_TYPE_UINT32;
  977. if (indexBuffer != nullptr)
  978. {
  979. VulkanBuffer* resource = indexBuffer->getResource(mDevice.getIndex());
  980. if (resource != nullptr)
  981. {
  982. vkBuffer = resource->getHandle();
  983. indexType = VulkanUtility::getIndexType(buffer->getProperties().getType());
  984. registerResource(resource, VulkanUseFlag::Read);
  985. }
  986. }
  987. vkCmdBindIndexBuffer(mCmdBuffer, vkBuffer, 0, indexType);
  988. }
  989. void VulkanCmdBuffer::setVertexDeclaration(const SPtr<VertexDeclaration>& decl)
  990. {
  991. if (mVertexDecl == decl)
  992. return;
  993. mVertexDecl = decl;
  994. mGfxPipelineRequiresBind = true;
  995. }
  996. bool VulkanCmdBuffer::isReadyForRender()
  997. {
  998. if (mGraphicsPipeline == nullptr)
  999. return false;
  1000. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  1001. if (inputDecl == nullptr)
  1002. return false;
  1003. return mFramebuffer != nullptr && mVertexDecl != nullptr;
  1004. }
  1005. bool VulkanCmdBuffer::bindGraphicsPipeline()
  1006. {
  1007. SPtr<VertexDeclaration> inputDecl = mGraphicsPipeline->getInputDeclaration();
  1008. SPtr<VulkanVertexInput> vertexInput = VulkanVertexInputManager::instance().getVertexInfo(mVertexDecl, inputDecl);
  1009. VulkanPipeline* pipeline = mGraphicsPipeline->getPipeline(mDevice.getIndex(), mFramebuffer,
  1010. mRenderTargetReadOnlyFlags, mDrawOp, vertexInput);
  1011. if (pipeline == nullptr)
  1012. return false;
  1013. // Check that pipeline matches the read-only state of any framebuffer attachments
  1014. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1015. for (UINT32 i = 0; i < numColorAttachments; i++)
  1016. {
  1017. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1018. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1019. fbAttachment.surface.mipLevel);
  1020. if (subresourceInfo.isShaderInput && !pipeline->isColorReadOnly(i))
  1021. {
  1022. LOGWRN("Framebuffer attachment also used as a shader input, but color writes aren't disabled. This will"
  1023. " result in undefined behavior.");
  1024. }
  1025. }
  1026. if (mFramebuffer->hasDepthAttachment())
  1027. {
  1028. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1029. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1030. fbAttachment.surface.mipLevel);
  1031. if (subresourceInfo.isShaderInput && !pipeline->isDepthReadOnly())
  1032. {
  1033. LOGWRN("Framebuffer attachment also used as a shader input, but depth/stencil writes aren't disabled. "
  1034. "This will result in undefined behavior.");
  1035. }
  1036. }
  1037. mGraphicsPipeline->registerPipelineResources(this);
  1038. registerResource(pipeline, VulkanUseFlag::Read);
  1039. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline->getHandle());
  1040. bindDynamicStates(true);
  1041. mGfxPipelineRequiresBind = false;
  1042. return true;
  1043. }
  1044. void VulkanCmdBuffer::bindDynamicStates(bool forceAll)
  1045. {
  1046. if (mViewportRequiresBind || forceAll)
  1047. {
  1048. VkViewport viewport;
  1049. viewport.x = mViewport.x * mRenderTargetWidth;
  1050. viewport.y = mViewport.y * mRenderTargetHeight;
  1051. viewport.width = mViewport.width * mRenderTargetWidth;
  1052. viewport.height = mViewport.height * mRenderTargetHeight;
  1053. viewport.minDepth = 0.0f;
  1054. viewport.maxDepth = 1.0f;
  1055. vkCmdSetViewport(mCmdBuffer, 0, 1, &viewport);
  1056. mViewportRequiresBind = false;
  1057. }
  1058. if(mStencilRefRequiresBind || forceAll)
  1059. {
  1060. vkCmdSetStencilReference(mCmdBuffer, VK_STENCIL_FRONT_AND_BACK, mStencilRef);
  1061. mStencilRefRequiresBind = false;
  1062. }
  1063. if(mScissorRequiresBind || forceAll)
  1064. {
  1065. VkRect2D scissorRect;
  1066. if(mGraphicsPipeline->isScissorEnabled())
  1067. {
  1068. scissorRect.offset.x = mScissor.x;
  1069. scissorRect.offset.y = mScissor.y;
  1070. scissorRect.extent.width = mScissor.width;
  1071. scissorRect.extent.height = mScissor.height;
  1072. }
  1073. else
  1074. {
  1075. scissorRect.offset.x = 0;
  1076. scissorRect.offset.y = 0;
  1077. scissorRect.extent.width = mRenderTargetWidth;
  1078. scissorRect.extent.height = mRenderTargetHeight;
  1079. }
  1080. vkCmdSetScissor(mCmdBuffer, 0, 1, &scissorRect);
  1081. mScissorRequiresBind = false;
  1082. }
  1083. }
  1084. void VulkanCmdBuffer::bindGpuParams()
  1085. {
  1086. if (mBoundParamsDirty)
  1087. {
  1088. if (mBoundParams != nullptr)
  1089. {
  1090. mNumBoundDescriptorSets = mBoundParams->getNumSets();
  1091. mBoundParams->prepareForBind(*this, mDescriptorSetsTemp);
  1092. }
  1093. else
  1094. mNumBoundDescriptorSets = 0;
  1095. mBoundParamsDirty = false;
  1096. }
  1097. else
  1098. {
  1099. mNumBoundDescriptorSets = 0;
  1100. }
  1101. }
  1102. void VulkanCmdBuffer::executeLayoutTransitions()
  1103. {
  1104. auto createLayoutTransitionBarrier = [&](VulkanImage* image, ImageInfo& imageInfo)
  1105. {
  1106. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1107. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1108. {
  1109. ImageSubresourceInfo& subresourceInfo = subresourceInfos[i];
  1110. if (!subresourceInfo.hasTransitioned || subresourceInfo.currentLayout == subresourceInfo.requiredLayout)
  1111. continue;
  1112. mLayoutTransitionBarriersTemp.push_back(VkImageMemoryBarrier());
  1113. VkImageMemoryBarrier& barrier = mLayoutTransitionBarriersTemp.back();
  1114. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1115. barrier.pNext = nullptr;
  1116. barrier.srcAccessMask = image->getAccessFlags(subresourceInfo.currentLayout);
  1117. barrier.dstAccessMask = image->getAccessFlags(subresourceInfo.requiredLayout, subresourceInfo.isReadOnly);
  1118. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1119. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1120. barrier.oldLayout = subresourceInfo.currentLayout;
  1121. barrier.newLayout = subresourceInfo.requiredLayout;
  1122. barrier.image = image->getHandle();
  1123. barrier.subresourceRange = subresourceInfo.range;
  1124. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1125. subresourceInfo.isReadOnly = true;
  1126. subresourceInfo.hasTransitioned = true;
  1127. }
  1128. };
  1129. // Note: These layout transitions will contain transitions for offscreen framebuffer attachments (while they
  1130. // transition to shader read-only layout). This can be avoided, since they're immediately used by the render pass
  1131. // as color attachments, making the layout change redundant.
  1132. for (auto& entry : mQueuedLayoutTransitions)
  1133. {
  1134. UINT32 imageInfoIdx = entry.second;
  1135. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1136. createLayoutTransitionBarrier(entry.first, imageInfo);
  1137. }
  1138. VkPipelineStageFlags srcStage = 0;
  1139. VkPipelineStageFlags dstStage = 0;
  1140. getPipelineStageFlags(mLayoutTransitionBarriersTemp, srcStage, dstStage);
  1141. vkCmdPipelineBarrier(mCmdBuffer,
  1142. srcStage, dstStage,
  1143. 0, 0, nullptr,
  1144. 0, nullptr,
  1145. (UINT32)mLayoutTransitionBarriersTemp.size(), mLayoutTransitionBarriersTemp.data());
  1146. mQueuedLayoutTransitions.clear();
  1147. mLayoutTransitionBarriersTemp.clear();
  1148. }
  1149. void VulkanCmdBuffer::updateFinalLayouts()
  1150. {
  1151. if (mFramebuffer == nullptr)
  1152. return;
  1153. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1154. for (UINT32 i = 0; i < numColorAttachments; i++)
  1155. {
  1156. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1157. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1158. fbAttachment.surface.mipLevel);
  1159. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1160. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1161. subresourceInfo.hasTransitioned = true;
  1162. }
  1163. if (mFramebuffer->hasDepthAttachment())
  1164. {
  1165. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1166. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1167. fbAttachment.surface.mipLevel);
  1168. subresourceInfo.currentLayout = subresourceInfo.finalLayout;
  1169. subresourceInfo.requiredLayout = subresourceInfo.finalLayout;
  1170. subresourceInfo.hasTransitioned = true;
  1171. }
  1172. }
  1173. void VulkanCmdBuffer::executeClearPass()
  1174. {
  1175. assert(mState == State::Recording);
  1176. executeLayoutTransitions();
  1177. VkRenderPassBeginInfo renderPassBeginInfo;
  1178. renderPassBeginInfo.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  1179. renderPassBeginInfo.pNext = nullptr;
  1180. renderPassBeginInfo.framebuffer = mFramebuffer->getFramebuffer(RT_NONE, RT_NONE, mClearMask);
  1181. renderPassBeginInfo.renderPass = mFramebuffer->getRenderPass(RT_NONE, RT_NONE, mClearMask);
  1182. renderPassBeginInfo.renderArea.offset.x = mClearArea.x;
  1183. renderPassBeginInfo.renderArea.offset.y = mClearArea.y;
  1184. renderPassBeginInfo.renderArea.extent.width = mClearArea.width;
  1185. renderPassBeginInfo.renderArea.extent.height = mClearArea.height;
  1186. renderPassBeginInfo.clearValueCount = mFramebuffer->getNumClearEntries(mClearMask);
  1187. renderPassBeginInfo.pClearValues = mClearValues.data();
  1188. vkCmdBeginRenderPass(mCmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
  1189. vkCmdEndRenderPass(mCmdBuffer);
  1190. updateFinalLayouts();
  1191. mClearMask = CLEAR_NONE;
  1192. }
  1193. void VulkanCmdBuffer::draw(UINT32 vertexOffset, UINT32 vertexCount, UINT32 instanceCount)
  1194. {
  1195. if (!isReadyForRender())
  1196. return;
  1197. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1198. bindGpuParams();
  1199. if (!isInRenderPass())
  1200. beginRenderPass();
  1201. if (mGfxPipelineRequiresBind)
  1202. {
  1203. if (!bindGraphicsPipeline())
  1204. return;
  1205. }
  1206. else
  1207. bindDynamicStates(false);
  1208. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1209. {
  1210. if (mNumBoundDescriptorSets > 0)
  1211. {
  1212. UINT32 deviceIdx = mDevice.getIndex();
  1213. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1214. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1215. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1216. }
  1217. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1218. }
  1219. if (instanceCount <= 0)
  1220. instanceCount = 1;
  1221. vkCmdDraw(mCmdBuffer, vertexCount, instanceCount, vertexOffset, 0);
  1222. }
  1223. void VulkanCmdBuffer::drawIndexed(UINT32 startIndex, UINT32 indexCount, UINT32 vertexOffset, UINT32 instanceCount)
  1224. {
  1225. if (!isReadyForRender())
  1226. return;
  1227. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1228. bindGpuParams();
  1229. if (!isInRenderPass())
  1230. beginRenderPass();
  1231. if (mGfxPipelineRequiresBind)
  1232. {
  1233. if (!bindGraphicsPipeline())
  1234. return;
  1235. }
  1236. else
  1237. bindDynamicStates(false);
  1238. if (mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Graphics))
  1239. {
  1240. if (mNumBoundDescriptorSets > 0)
  1241. {
  1242. UINT32 deviceIdx = mDevice.getIndex();
  1243. VkPipelineLayout pipelineLayout = mGraphicsPipeline->getPipelineLayout(deviceIdx);
  1244. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipelineLayout, 0,
  1245. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1246. }
  1247. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Graphics);
  1248. }
  1249. if (instanceCount <= 0)
  1250. instanceCount = 1;
  1251. vkCmdDrawIndexed(mCmdBuffer, indexCount, instanceCount, startIndex, vertexOffset, 0);
  1252. }
  1253. void VulkanCmdBuffer::dispatch(UINT32 numGroupsX, UINT32 numGroupsY, UINT32 numGroupsZ)
  1254. {
  1255. if (mComputePipeline == nullptr)
  1256. return;
  1257. if (isInRenderPass())
  1258. endRenderPass();
  1259. // Note: Should I restore the render target after? Note that this is only being done is framebuffer subresources
  1260. // have their "isFBAttachment" flag reset, potentially I can just clear/restore those
  1261. setRenderTarget(nullptr, 0, RT_ALL);
  1262. // Need to bind gpu params before starting render pass, in order to make sure any layout transitions execute
  1263. bindGpuParams();
  1264. executeLayoutTransitions();
  1265. UINT32 deviceIdx = mDevice.getIndex();
  1266. if(mCmpPipelineRequiresBind)
  1267. {
  1268. VulkanPipeline* pipeline = mComputePipeline->getPipeline(deviceIdx);
  1269. if (pipeline == nullptr)
  1270. return;
  1271. registerResource(pipeline, VulkanUseFlag::Read);
  1272. mComputePipeline->registerPipelineResources(this);
  1273. vkCmdBindPipeline(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline->getHandle());
  1274. mCmpPipelineRequiresBind = false;
  1275. }
  1276. if(mDescriptorSetsBindState.isSet(DescriptorSetBindFlag::Compute))
  1277. {
  1278. if (mNumBoundDescriptorSets > 0)
  1279. {
  1280. VkPipelineLayout pipelineLayout = mComputePipeline->getPipelineLayout(deviceIdx);
  1281. vkCmdBindDescriptorSets(mCmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipelineLayout, 0,
  1282. mNumBoundDescriptorSets, mDescriptorSetsTemp, 0, nullptr);
  1283. }
  1284. mDescriptorSetsBindState.unset(DescriptorSetBindFlag::Compute);
  1285. }
  1286. vkCmdDispatch(mCmdBuffer, numGroupsX, numGroupsY, numGroupsZ);
  1287. // Update any layout transitions that were performed by subpass dependencies, reset flags that signal image usage
  1288. // and reset read-only state.
  1289. for(auto& entry : mPassTouchedSubresourceInfos)
  1290. {
  1291. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage[entry];
  1292. subresourceInfo.isShaderInput = false;
  1293. subresourceInfo.isReadOnly = true;
  1294. subresourceInfo.needsBarrier = false;
  1295. }
  1296. for (auto& entry : mBuffers)
  1297. entry.second.needsBarrier = false;
  1298. mPassTouchedSubresourceInfos.clear();
  1299. }
  1300. void VulkanCmdBuffer::setEvent(VulkanEvent* event)
  1301. {
  1302. if(isInRenderPass())
  1303. mQueuedEvents.push_back(event);
  1304. else
  1305. vkCmdSetEvent(mCmdBuffer, event->getHandle(), VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  1306. }
  1307. void VulkanCmdBuffer::resetQuery(VulkanQuery* query)
  1308. {
  1309. if (isInRenderPass())
  1310. mQueuedQueryResets.push_back(query);
  1311. else
  1312. query->reset(mCmdBuffer);
  1313. }
  1314. void VulkanCmdBuffer::memoryBarrier(VkBuffer buffer, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1315. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage)
  1316. {
  1317. VkBufferMemoryBarrier barrier;
  1318. barrier.sType = VK_STRUCTURE_TYPE_BUFFER_MEMORY_BARRIER;
  1319. barrier.pNext = nullptr;
  1320. barrier.srcAccessMask = srcAccessFlags;
  1321. barrier.dstAccessMask = dstAccessFlags;
  1322. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1323. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1324. barrier.buffer = buffer;
  1325. barrier.offset = 0;
  1326. barrier.size = VK_WHOLE_SIZE;
  1327. vkCmdPipelineBarrier(getHandle(),
  1328. srcStage,
  1329. dstStage,
  1330. 0, 0, nullptr,
  1331. 1, &barrier,
  1332. 0, nullptr);
  1333. }
  1334. void VulkanCmdBuffer::memoryBarrier(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1335. VkPipelineStageFlags srcStage, VkPipelineStageFlags dstStage, VkImageLayout layout,
  1336. const VkImageSubresourceRange& range)
  1337. {
  1338. VkImageMemoryBarrier barrier;
  1339. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1340. barrier.pNext = nullptr;
  1341. barrier.srcAccessMask = srcAccessFlags;
  1342. barrier.dstAccessMask = dstAccessFlags;
  1343. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1344. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1345. barrier.image = image;
  1346. barrier.subresourceRange = range;
  1347. barrier.oldLayout = layout;
  1348. barrier.newLayout = layout;
  1349. vkCmdPipelineBarrier(getHandle(),
  1350. srcStage,
  1351. dstStage,
  1352. 0, 0, nullptr,
  1353. 0, nullptr,
  1354. 1, &barrier);
  1355. }
  1356. void VulkanCmdBuffer::setLayout(VkImage image, VkAccessFlags srcAccessFlags, VkAccessFlags dstAccessFlags,
  1357. VkImageLayout oldLayout, VkImageLayout newLayout, const VkImageSubresourceRange& range)
  1358. {
  1359. VkImageMemoryBarrier barrier;
  1360. barrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  1361. barrier.pNext = nullptr;
  1362. barrier.srcAccessMask = srcAccessFlags;
  1363. barrier.dstAccessMask = dstAccessFlags;
  1364. barrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1365. barrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  1366. barrier.oldLayout = oldLayout;
  1367. barrier.newLayout = newLayout;
  1368. barrier.image = image;
  1369. barrier.subresourceRange = range;
  1370. VkPipelineStageFlags srcStage = getPipelineStageFlags(srcAccessFlags);
  1371. VkPipelineStageFlags dstStage = getPipelineStageFlags(dstAccessFlags);
  1372. vkCmdPipelineBarrier(getHandle(),
  1373. srcStage,
  1374. dstStage,
  1375. 0, 0, nullptr,
  1376. 0, nullptr,
  1377. 1, &barrier);
  1378. }
  1379. VkImageLayout VulkanCmdBuffer::getCurrentLayout(VulkanImage* image, const VkImageSubresourceRange& range,
  1380. bool inRenderPass)
  1381. {
  1382. UINT32 face = range.baseArrayLayer;
  1383. UINT32 mip = range.baseMipLevel;
  1384. // The assumption is that all the subresources in the range will have the same layout, as this should be handled
  1385. // by registerResource(), or by external code (in the case of transfers). So we only check the first subresource.
  1386. VulkanImageSubresource* subresource = image->getSubresource(face, mip);
  1387. auto iterFind = mImages.find(image);
  1388. if (iterFind == mImages.end())
  1389. return subresource->getLayout();
  1390. UINT32 imageInfoIdx = iterFind->second;
  1391. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1392. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1393. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1394. {
  1395. ImageSubresourceInfo& entry = subresourceInfos[i];
  1396. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1397. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1398. {
  1399. // If it's a FB attachment, retrieve its layout after the render pass begins
  1400. if(entry.isFBAttachment && inRenderPass && mFramebuffer)
  1401. {
  1402. RenderSurfaceMask readMask = getFBReadMask();
  1403. // Is it a depth-stencil attachment?
  1404. if(mFramebuffer->hasDepthAttachment() && mFramebuffer->getDepthStencilAttachment().image == image)
  1405. {
  1406. if (readMask.isSet(RT_DEPTH))
  1407. {
  1408. if (readMask.isSet(RT_STENCIL))
  1409. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL;
  1410. else // Depth readable but stencil isn't
  1411. return VK_IMAGE_LAYOUT_GENERAL;
  1412. }
  1413. return VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1414. }
  1415. else // It is a color attachment
  1416. {
  1417. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1418. for (UINT32 j = 0; j < numColorAttachments; j++)
  1419. {
  1420. const VulkanFramebufferAttachment& attachment = mFramebuffer->getColorAttachment(j);
  1421. if (attachment.image == image)
  1422. {
  1423. if (readMask.isSet((RenderSurfaceMaskBits)(1 << attachment.index)))
  1424. return VK_IMAGE_LAYOUT_GENERAL;
  1425. else
  1426. return VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1427. }
  1428. }
  1429. }
  1430. }
  1431. return entry.requiredLayout;
  1432. }
  1433. }
  1434. return subresource->getLayout();
  1435. }
  1436. void VulkanCmdBuffer::registerResource(VulkanResource* res, VulkanUseFlags flags)
  1437. {
  1438. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1439. if(insertResult.second) // New element
  1440. {
  1441. ResourceUseHandle& useHandle = insertResult.first->second;
  1442. useHandle.used = false;
  1443. useHandle.flags = flags;
  1444. res->notifyBound();
  1445. }
  1446. else // Existing element
  1447. {
  1448. ResourceUseHandle& useHandle = insertResult.first->second;
  1449. assert(!useHandle.used);
  1450. useHandle.flags |= flags;
  1451. }
  1452. }
  1453. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range, VulkanUseFlags flags,
  1454. ResourceUsage usage)
  1455. {
  1456. VkImageLayout layout = res->getOptimalLayout();
  1457. registerResource(res, range, layout, layout, flags, usage);
  1458. }
  1459. void VulkanCmdBuffer::registerResource(VulkanImage* res, const VkImageSubresourceRange& range,
  1460. VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags, ResourceUsage usage)
  1461. {
  1462. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1463. bool isTransfer = usage == ResourceUsage::Transfer;
  1464. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1465. // If binding it for write in a shader (not as color attachment or transfer op), we will need to issue a memory
  1466. // barrier if the image gets used again during this render pass, so remember this information.
  1467. bool needsBarrier = isShaderBind && flags.isSet(VulkanUseFlag::Write);
  1468. UINT32 nextImageInfoIdx = (UINT32)mImageInfos.size();
  1469. auto registerSubresourceInfo = [&](const VkImageSubresourceRange& subresourceRange)
  1470. {
  1471. mSubresourceInfoStorage.push_back(ImageSubresourceInfo());
  1472. ImageSubresourceInfo& subresourceInfo = mSubresourceInfoStorage.back();
  1473. subresourceInfo.currentLayout = newLayout;
  1474. subresourceInfo.initialLayout = newLayout;
  1475. subresourceInfo.requiredLayout = newLayout;
  1476. subresourceInfo.finalLayout = finalLayout;
  1477. subresourceInfo.range = subresourceRange;
  1478. subresourceInfo.isFBAttachment = isFBAttachment;
  1479. subresourceInfo.isShaderInput = isShaderBind;
  1480. if (!isTransfer)
  1481. {
  1482. subresourceInfo.hasExternalTransition = false;
  1483. subresourceInfo.hasTransitioned = false;
  1484. subresourceInfo.isReadOnly = !flags.isSet(VulkanUseFlag::Write);
  1485. }
  1486. else
  1487. {
  1488. // Transfers handle layout transitions externally (at this point they are assumed to already be done)
  1489. subresourceInfo.hasExternalTransition = true;
  1490. subresourceInfo.hasTransitioned = true;
  1491. subresourceInfo.isReadOnly = true; // Doesn't matter for transfers
  1492. }
  1493. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1494. subresourceInfo.needsBarrier = needsBarrier;
  1495. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1496. };
  1497. auto insertResult = mImages.insert(std::make_pair(res, nextImageInfoIdx));
  1498. if (insertResult.second) // New element
  1499. {
  1500. UINT32 imageInfoIdx = insertResult.first->second;
  1501. mImageInfos.push_back(ImageInfo());
  1502. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1503. imageInfo.subresourceInfoIdx = (UINT32)mSubresourceInfoStorage.size();
  1504. imageInfo.numSubresourceInfos = 1;
  1505. imageInfo.useHandle.used = false;
  1506. imageInfo.useHandle.flags = flags;
  1507. registerSubresourceInfo(range);
  1508. res->notifyBound();
  1509. }
  1510. else // Existing element
  1511. {
  1512. UINT32 imageInfoIdx = insertResult.first->second;
  1513. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1514. assert(!imageInfo.useHandle.used);
  1515. imageInfo.useHandle.flags |= flags;
  1516. // See if there is an overlap between existing ranges and the new range. And if so break them up accordingly.
  1517. //// First test for the simplest and most common case (same range or no overlap) to avoid more complex
  1518. //// computations.
  1519. ImageSubresourceInfo* subresources = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1520. bool foundRange = false;
  1521. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1522. {
  1523. if(VulkanUtility::rangeOverlaps(subresources[i].range, range))
  1524. {
  1525. if (subresources[i].range.layerCount == range.layerCount &&
  1526. subresources[i].range.levelCount == range.levelCount &&
  1527. subresources[i].range.baseArrayLayer == range.baseArrayLayer &&
  1528. subresources[i].range.baseMipLevel == range.baseMipLevel)
  1529. {
  1530. // Just update existing range
  1531. updateSubresourceInfo(res, imageInfoIdx, subresources[i], newLayout, finalLayout, flags, usage);
  1532. mPassTouchedSubresourceInfos.insert(imageInfo.subresourceInfoIdx + i);
  1533. foundRange = true;
  1534. break;
  1535. }
  1536. break;
  1537. }
  1538. }
  1539. //// We'll need to update subresource ranges or add new ones. The hope is that this code is trigger VERY rarely
  1540. //// (for just a few specific textures per frame).
  1541. if (!foundRange)
  1542. {
  1543. std::array<VkImageSubresourceRange, 5> tempCutRanges;
  1544. bs_frame_mark();
  1545. {
  1546. // We orphan previously allocated memory (we reset it after submit() anyway)
  1547. UINT32 newSubresourceIdx = (UINT32)mSubresourceInfoStorage.size();
  1548. FrameVector<UINT32> cutOverlappingRanges;
  1549. for (UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1550. {
  1551. UINT32 subresourceIdx = imageInfo.subresourceInfoIdx + i;
  1552. ImageSubresourceInfo& subresource = mSubresourceInfoStorage[subresourceIdx];
  1553. if (!VulkanUtility::rangeOverlaps(subresource.range, range))
  1554. {
  1555. // Just copy as is
  1556. mSubresourceInfoStorage.push_back(subresource);
  1557. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1558. }
  1559. else // Need to cut
  1560. {
  1561. UINT32 numCutRanges;
  1562. VulkanUtility::cutRange(subresource.range, range, tempCutRanges, numCutRanges);
  1563. for(UINT32 j = 0; j < numCutRanges; j++)
  1564. {
  1565. // Create a copy of the original subresource with the new range
  1566. ImageSubresourceInfo newInfo = mSubresourceInfoStorage[subresourceIdx];
  1567. newInfo.range = tempCutRanges[j];
  1568. if(VulkanUtility::rangeOverlaps(tempCutRanges[j], range))
  1569. {
  1570. // Update overlapping sub-resource range with new data from this range
  1571. updateSubresourceInfo(res, imageInfoIdx, newInfo, newLayout, finalLayout, flags, usage);
  1572. // Keep track of the overlapping ranges for later
  1573. cutOverlappingRanges.push_back((UINT32)mSubresourceInfoStorage.size());
  1574. }
  1575. mSubresourceInfoStorage.push_back(newInfo);
  1576. mPassTouchedSubresourceInfos.insert((UINT32)mSubresourceInfoStorage.size() - 1);
  1577. }
  1578. }
  1579. }
  1580. // Our range doesn't overlap with any existing ranges, so just add it
  1581. if(cutOverlappingRanges.size() == 0)
  1582. {
  1583. registerSubresourceInfo(range);
  1584. }
  1585. else // Search if overlapping ranges fully cover the requested range, and insert non-covered regions
  1586. {
  1587. FrameQueue<VkImageSubresourceRange> sourceRanges;
  1588. sourceRanges.push(range);
  1589. for(auto& entry : cutOverlappingRanges)
  1590. {
  1591. VkImageSubresourceRange& overlappingRange = mSubresourceInfoStorage[entry].range;
  1592. UINT32 numSourceRanges = (UINT32)sourceRanges.size();
  1593. for(UINT32 i = 0; i < numSourceRanges; i++)
  1594. {
  1595. VkImageSubresourceRange sourceRange = sourceRanges.front();
  1596. sourceRanges.pop();
  1597. UINT32 numCutRanges;
  1598. VulkanUtility::cutRange(sourceRange, overlappingRange, tempCutRanges, numCutRanges);
  1599. for(UINT32 j = 0; j < numCutRanges; j++)
  1600. {
  1601. // We only care about ranges outside of the ones we already covered
  1602. if(!VulkanUtility::rangeOverlaps(tempCutRanges[j], overlappingRange))
  1603. sourceRanges.push(tempCutRanges[j]);
  1604. }
  1605. }
  1606. }
  1607. // Any remaining range hasn't been covered yet
  1608. while(!sourceRanges.empty())
  1609. {
  1610. registerSubresourceInfo(sourceRanges.front());
  1611. sourceRanges.pop();
  1612. }
  1613. }
  1614. imageInfo.subresourceInfoIdx = newSubresourceIdx;
  1615. imageInfo.numSubresourceInfos = (UINT32)mSubresourceInfoStorage.size() - newSubresourceIdx;
  1616. }
  1617. bs_frame_clear();
  1618. }
  1619. }
  1620. // Register any sub-resources
  1621. for(UINT32 i = 0; i < range.layerCount; i++)
  1622. {
  1623. for(UINT32 j = 0; j < range.levelCount; j++)
  1624. {
  1625. UINT32 layer = range.baseArrayLayer + i;
  1626. UINT32 mipLevel = range.baseMipLevel + j;
  1627. registerResource(res->getSubresource(layer, mipLevel), flags);
  1628. }
  1629. }
  1630. }
  1631. void VulkanCmdBuffer::registerResource(VulkanBuffer* res, VkAccessFlags accessFlags, VulkanUseFlags flags)
  1632. {
  1633. bool isShaderWrite = (accessFlags & VK_ACCESS_SHADER_WRITE_BIT) != 0;
  1634. auto insertResult = mBuffers.insert(std::make_pair(res, BufferInfo()));
  1635. if (insertResult.second) // New element
  1636. {
  1637. BufferInfo& bufferInfo = insertResult.first->second;
  1638. bufferInfo.accessFlags = accessFlags;
  1639. bufferInfo.useHandle.used = false;
  1640. bufferInfo.useHandle.flags = flags;
  1641. // Any writes done on storage buffers will need explicit memory barriers (if read during the same pass) so
  1642. // we remember this, in case this buffers gets used later in this pass.
  1643. bufferInfo.needsBarrier = isShaderWrite;
  1644. res->notifyBound();
  1645. }
  1646. else // Existing element
  1647. {
  1648. BufferInfo& bufferInfo = insertResult.first->second;
  1649. assert(!bufferInfo.useHandle.used);
  1650. bufferInfo.useHandle.flags |= flags;
  1651. bufferInfo.accessFlags |= accessFlags;
  1652. // If the buffer was written to previously in this pass, and is now being used by a shader we need to issue
  1653. // a barrier to make those writes visible.
  1654. bool isShaderRead = (accessFlags & VK_ACCESS_SHADER_READ_BIT) != 0;
  1655. if(bufferInfo.needsBarrier && (isShaderRead || isShaderWrite))
  1656. {
  1657. // Need to end render pass in order to execute the barrier. Hopefully this won't trigger much since most
  1658. // shader writes are done during compute
  1659. if (isInRenderPass())
  1660. endRenderPass();
  1661. VkPipelineStageFlags stages =
  1662. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1663. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1664. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1665. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1666. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1667. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1668. VkBuffer buffer = res->getHandle();
  1669. memoryBarrier(buffer, VK_ACCESS_SHADER_WRITE_BIT, accessFlags, stages, stages);
  1670. bufferInfo.needsBarrier = isShaderWrite;
  1671. }
  1672. }
  1673. }
  1674. void VulkanCmdBuffer::registerResource(VulkanFramebuffer* res, RenderSurfaceMask loadMask, UINT32 readMask)
  1675. {
  1676. auto insertResult = mResources.insert(std::make_pair(res, ResourceUseHandle()));
  1677. if (insertResult.second) // New element
  1678. {
  1679. ResourceUseHandle& useHandle = insertResult.first->second;
  1680. useHandle.used = false;
  1681. useHandle.flags = VulkanUseFlag::Write;
  1682. res->notifyBound();
  1683. }
  1684. else // Existing element
  1685. {
  1686. ResourceUseHandle& useHandle = insertResult.first->second;
  1687. assert(!useHandle.used);
  1688. useHandle.flags |= VulkanUseFlag::Write;
  1689. }
  1690. // Register any sub-resources
  1691. UINT32 numColorAttachments = res->getNumColorAttachments();
  1692. for (UINT32 i = 0; i < numColorAttachments; i++)
  1693. {
  1694. const VulkanFramebufferAttachment& attachment = res->getColorAttachment(i);
  1695. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1696. VkImageLayout layout;
  1697. if (loadMask.isSet((RenderSurfaceMaskBits)(1 << i)))
  1698. {
  1699. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1700. // transition this to a different layout when it begins, but we handle that elsewhere
  1701. layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  1702. }
  1703. else
  1704. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1705. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1706. registerResource(attachment.image, range, layout, attachment.finalLayout, VulkanUseFlag::Write,
  1707. ResourceUsage::Framebuffer);
  1708. }
  1709. if(res->hasDepthAttachment())
  1710. {
  1711. const VulkanFramebufferAttachment& attachment = res->getDepthStencilAttachment();
  1712. VulkanUseFlag useFlag = VulkanUseFlag::Write;
  1713. // If image is being loaded, we need to transfer it to correct layout, otherwise it doesn't matter
  1714. VkImageLayout layout;
  1715. if (loadMask.isSet(RT_DEPTH) || loadMask.isSet(RT_STENCIL)) // Can't load one without the other
  1716. {
  1717. // Note that this might not be the actual layout used during the render pass, as the render pass can
  1718. // transition this to a different layout when it begins, but we handle that elsewhere
  1719. layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  1720. }
  1721. else
  1722. layout = VK_IMAGE_LAYOUT_UNDEFINED;
  1723. VkImageSubresourceRange range = attachment.image->getRange(attachment.surface);
  1724. registerResource(attachment.image, range, layout, attachment.finalLayout, useFlag, ResourceUsage::Framebuffer);
  1725. }
  1726. }
  1727. bool VulkanCmdBuffer::updateSubresourceInfo(VulkanImage* image, UINT32 imageInfoIdx,
  1728. ImageSubresourceInfo& subresourceInfo, VkImageLayout newLayout, VkImageLayout finalLayout, VulkanUseFlags flags,
  1729. ResourceUsage usage)
  1730. {
  1731. bool isTransfer = usage == ResourceUsage::Transfer;
  1732. bool isFBAttachment = usage == ResourceUsage::Framebuffer;
  1733. bool isShaderBind = usage == ResourceUsage::ShaderBind;
  1734. // Transfers are always considered read only because this only matters for destination access mask, and transfers
  1735. // handle that externally
  1736. if(!isTransfer)
  1737. subresourceInfo.isReadOnly &= !flags.isSet(VulkanUseFlag::Write);
  1738. // For transfers, just assign new layout, external code does all the work
  1739. if (isTransfer)
  1740. subresourceInfo.requiredLayout = newLayout;
  1741. else
  1742. {
  1743. // New layout is valid, check for transitions (UNDEFINED signifies the caller doesn't want a layout transition)
  1744. if (newLayout != VK_IMAGE_LAYOUT_UNDEFINED)
  1745. {
  1746. // If layout transition was requested by framebuffer bind, respect it because render-pass will only accept a
  1747. // specific layout (in certain cases), and we have no choice.
  1748. // In the case when a FB attachment is also bound for shader reads, this will override the layout required for
  1749. // shader read (GENERAL or DEPTH_READ_ONLY), but that is fine because those transitions are handled
  1750. // automatically by render-pass layout transitions.
  1751. // Any other texture (non FB attachment) will only even be bound in a single layout and we can keep the one it
  1752. // was originally registered with.
  1753. if (isFBAttachment)
  1754. subresourceInfo.requiredLayout = newLayout;
  1755. else if (!subresourceInfo.isFBAttachment) // Layout transition is not being done on a FB image
  1756. {
  1757. // Check if the image had a layout previously assigned, and if so check if multiple different layouts
  1758. // were requested. In that case we wish to transfer the image to GENERAL layout.
  1759. bool firstUseInRenderPass = !subresourceInfo.isShaderInput && !subresourceInfo.isFBAttachment;
  1760. if (firstUseInRenderPass || subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_UNDEFINED)
  1761. subresourceInfo.requiredLayout = newLayout;
  1762. else if (subresourceInfo.requiredLayout != newLayout)
  1763. subresourceInfo.requiredLayout = VK_IMAGE_LAYOUT_GENERAL;
  1764. }
  1765. }
  1766. }
  1767. // If attached to FB, then the final layout is set by the FB (provided as layout param here), otherwise its
  1768. // the same as required layout
  1769. if (!isFBAttachment && !subresourceInfo.isFBAttachment)
  1770. subresourceInfo.finalLayout = subresourceInfo.requiredLayout;
  1771. else
  1772. {
  1773. if (isFBAttachment)
  1774. subresourceInfo.finalLayout = finalLayout;
  1775. }
  1776. // Queue layout transition, if not transfer. Transfer handle layout transitions externally.
  1777. if (!isTransfer)
  1778. {
  1779. // If we haven't done a layout transition yet, we can just overwrite the previously written values, and the
  1780. // transition will be handled as the first thing in submit(), otherwise we queue a non-initial transition
  1781. // below.
  1782. if (!subresourceInfo.hasTransitioned)
  1783. {
  1784. subresourceInfo.initialLayout = subresourceInfo.requiredLayout;
  1785. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1786. subresourceInfo.isInitialReadOnly = subresourceInfo.isReadOnly;
  1787. }
  1788. else
  1789. {
  1790. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1791. mQueuedLayoutTransitions[image] = imageInfoIdx;
  1792. }
  1793. }
  1794. else
  1795. {
  1796. // Resource has already transitioned its layout externally since this is a transfer. We cannot allow a
  1797. // transition in submit().
  1798. subresourceInfo.currentLayout = subresourceInfo.requiredLayout;
  1799. subresourceInfo.hasTransitioned = true;
  1800. }
  1801. // If a FB attachment was just bound as a shader input, we might need to restart the render pass with a FB
  1802. // attachment that supports read-only attachments using the GENERAL or DEPTH_READ_ONLY layout
  1803. bool resetRenderPass = false;
  1804. if (isFBAttachment)
  1805. {
  1806. if (!subresourceInfo.isFBAttachment)
  1807. {
  1808. subresourceInfo.isFBAttachment = true;
  1809. resetRenderPass = subresourceInfo.isShaderInput;
  1810. }
  1811. }
  1812. else
  1813. {
  1814. if (!subresourceInfo.isShaderInput)
  1815. {
  1816. subresourceInfo.isShaderInput = true;
  1817. if (subresourceInfo.isFBAttachment)
  1818. {
  1819. // Special case for depth: If user has set up proper read-only flags, then the render pass will have
  1820. // taken care of setting the valid state anyway, so no need to end the render pass
  1821. if (subresourceInfo.requiredLayout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL)
  1822. {
  1823. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0 || (mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1824. resetRenderPass = false;
  1825. else
  1826. resetRenderPass = true;
  1827. }
  1828. else
  1829. resetRenderPass = true;
  1830. }
  1831. else
  1832. {
  1833. // Subresource has been bound as shader input, and it wasn't bound as a FB attachment this render pass.
  1834. // However is could have been bound in a previous pass, so check the layouts and force a layout
  1835. // transition if required.
  1836. if (subresourceInfo.currentLayout != subresourceInfo.requiredLayout)
  1837. resetRenderPass = true;
  1838. else
  1839. resetRenderPass = false;
  1840. }
  1841. }
  1842. }
  1843. // If we need to switch frame-buffers, end current render pass
  1844. if (resetRenderPass && isInRenderPass())
  1845. endRenderPass();
  1846. else
  1847. {
  1848. // Since we won't be ending the render pass, check if this same sub-resource was written earlier in the pass,
  1849. // in which case we need to issue a memory barrier so those writes are visible.
  1850. // Memory barrier only matters if image is bound for shader use (no need for color attachments or transfers)
  1851. if(subresourceInfo.needsBarrier && isShaderBind)
  1852. {
  1853. bool isWrite = flags.isSet(VulkanUseFlag::Write);
  1854. VkPipelineStageFlags stages =
  1855. VK_PIPELINE_STAGE_VERTEX_SHADER_BIT |
  1856. VK_PIPELINE_STAGE_TESSELLATION_CONTROL_SHADER_BIT |
  1857. VK_PIPELINE_STAGE_TESSELLATION_EVALUATION_SHADER_BIT |
  1858. VK_PIPELINE_STAGE_GEOMETRY_SHADER_BIT |
  1859. VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT |
  1860. VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT;
  1861. memoryBarrier(image->getHandle(), VK_ACCESS_SHADER_WRITE_BIT,
  1862. image->getAccessFlags(subresourceInfo.requiredLayout, !isWrite),
  1863. stages, stages, subresourceInfo.requiredLayout, subresourceInfo.range);
  1864. subresourceInfo.needsBarrier = isWrite;
  1865. }
  1866. }
  1867. return resetRenderPass;
  1868. }
  1869. VulkanCmdBuffer::ImageSubresourceInfo& VulkanCmdBuffer::findSubresourceInfo(VulkanImage* image, UINT32 face, UINT32 mip)
  1870. {
  1871. UINT32 imageInfoIdx = mImages[image];
  1872. ImageInfo& imageInfo = mImageInfos[imageInfoIdx];
  1873. ImageSubresourceInfo* subresourceInfos = &mSubresourceInfoStorage[imageInfo.subresourceInfoIdx];
  1874. for(UINT32 i = 0; i < imageInfo.numSubresourceInfos; i++)
  1875. {
  1876. ImageSubresourceInfo& entry = subresourceInfos[i];
  1877. if(face >= entry.range.baseArrayLayer && face < (entry.range.baseArrayLayer + entry.range.layerCount) &&
  1878. mip >= entry.range.baseMipLevel && mip < (entry.range.baseMipLevel + entry.range.levelCount))
  1879. {
  1880. return entry;
  1881. }
  1882. }
  1883. assert(false); // Caller should ensure the subresource actually exists, so this shouldn't happen
  1884. return subresourceInfos[0];
  1885. }
  1886. void VulkanCmdBuffer::getInProgressQueries(Vector<VulkanTimerQuery*>& timer, Vector<VulkanOcclusionQuery*>& occlusion) const
  1887. {
  1888. for(auto& query : mTimerQueries)
  1889. {
  1890. if (query->_isInProgress())
  1891. timer.push_back(query);
  1892. }
  1893. for (auto& query : mOcclusionQueries)
  1894. {
  1895. if (query->_isInProgress())
  1896. occlusion.push_back(query);
  1897. }
  1898. }
  1899. VulkanCommandBuffer::VulkanCommandBuffer(VulkanDevice& device, GpuQueueType type, UINT32 deviceIdx,
  1900. UINT32 queueIdx, bool secondary)
  1901. : CommandBuffer(type, deviceIdx, queueIdx, secondary), mBuffer(nullptr)
  1902. , mDevice(device), mQueue(nullptr), mIdMask(0)
  1903. {
  1904. UINT32 numQueues = device.getNumQueues(mType);
  1905. if (numQueues == 0) // Fall back to graphics queue
  1906. {
  1907. mType = GQT_GRAPHICS;
  1908. numQueues = device.getNumQueues(GQT_GRAPHICS);
  1909. }
  1910. mQueue = device.getQueue(mType, mQueueIdx % numQueues);
  1911. mIdMask = device.getQueueMask(mType, mQueueIdx);
  1912. acquireNewBuffer();
  1913. }
  1914. RenderSurfaceMask VulkanCmdBuffer::getFBReadMask()
  1915. {
  1916. // Check if any frame-buffer attachments are also used as shader inputs, in which case we make them read-only
  1917. RenderSurfaceMask readMask = RT_NONE;
  1918. UINT32 numColorAttachments = mFramebuffer->getNumColorAttachments();
  1919. for(UINT32 i = 0; i < numColorAttachments; i++)
  1920. {
  1921. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getColorAttachment(i);
  1922. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1923. fbAttachment.surface.mipLevel);
  1924. bool readOnly = subresourceInfo.isShaderInput;
  1925. if(readOnly)
  1926. readMask.set((RenderSurfaceMaskBits)(1 << i));
  1927. }
  1928. if (mFramebuffer->hasDepthAttachment())
  1929. {
  1930. const VulkanFramebufferAttachment& fbAttachment = mFramebuffer->getDepthStencilAttachment();
  1931. ImageSubresourceInfo& subresourceInfo = findSubresourceInfo(fbAttachment.image, fbAttachment.surface.face,
  1932. fbAttachment.surface.mipLevel);
  1933. bool readOnly = subresourceInfo.isShaderInput;
  1934. if (readOnly)
  1935. readMask.set(RT_DEPTH);
  1936. if ((mRenderTargetReadOnlyFlags & FBT_DEPTH) != 0)
  1937. readMask.set(RT_DEPTH);
  1938. if ((mRenderTargetReadOnlyFlags & FBT_STENCIL) != 0)
  1939. readMask.set(RT_STENCIL);
  1940. }
  1941. return readMask;
  1942. }
  1943. VulkanCommandBuffer::~VulkanCommandBuffer()
  1944. {
  1945. mBuffer->reset();
  1946. }
  1947. void VulkanCommandBuffer::acquireNewBuffer()
  1948. {
  1949. VulkanCmdBufferPool& pool = mDevice.getCmdBufferPool();
  1950. if (mBuffer != nullptr)
  1951. assert(mBuffer->isSubmitted());
  1952. UINT32 queueFamily = mDevice.getQueueFamily(mType);
  1953. mBuffer = pool.getBuffer(queueFamily, mIsSecondary);
  1954. }
  1955. void VulkanCommandBuffer::submit(UINT32 syncMask)
  1956. {
  1957. // Ignore myself
  1958. syncMask &= ~mIdMask;
  1959. if (mBuffer->isInRenderPass())
  1960. mBuffer->endRenderPass();
  1961. // Execute any queued layout transitions that weren't already handled by the render pass
  1962. mBuffer->executeLayoutTransitions();
  1963. // Interrupt any in-progress queries (no in-progress queries allowed during command buffer submit)
  1964. Vector<VulkanTimerQuery*> timerQueries;
  1965. Vector<VulkanOcclusionQuery*> occlusionQueries;
  1966. mBuffer->getInProgressQueries(timerQueries, occlusionQueries);
  1967. for (auto& query : timerQueries)
  1968. query->_interrupt(*mBuffer);
  1969. for (auto& query : occlusionQueries)
  1970. query->_interrupt(*mBuffer);
  1971. if (mBuffer->isRecording())
  1972. mBuffer->end();
  1973. if (mBuffer->isReadyForSubmit()) // Possibly nothing was recorded in the buffer
  1974. {
  1975. mBuffer->submit(mQueue, mQueueIdx, syncMask);
  1976. acquireNewBuffer();
  1977. gVulkanCBManager().refreshStates(mDeviceIdx);
  1978. }
  1979. // Resume interrupted queries on the new command buffer
  1980. for (auto& query : timerQueries)
  1981. query->_resume(*mBuffer);
  1982. for (auto& query : occlusionQueries)
  1983. query->_resume(*mBuffer);
  1984. }
  1985. }}