VkGrManager.cpp 56 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723
  1. // Copyright (C) 2009-present, Panagiotis Christopoulos Charitos and contributors.
  2. // All rights reserved.
  3. // Code licensed under the BSD License.
  4. // http://www.anki3d.org/LICENSE
  5. #include <AnKi/Gr/Vulkan/VkGrManager.h>
  6. #include <AnKi/Util/StringList.h>
  7. #include <AnKi/Core/App.h>
  8. #include <AnKi/Gr/Vulkan/VkBuffer.h>
  9. #include <AnKi/Gr/Vulkan/VkTexture.h>
  10. #include <AnKi/Gr/Vulkan/VkSampler.h>
  11. #include <AnKi/Gr/Vulkan/VkShader.h>
  12. #include <AnKi/Gr/Vulkan/VkShaderProgram.h>
  13. #include <AnKi/Gr/Vulkan/VkCommandBuffer.h>
  14. #include <AnKi/Gr/Vulkan/VkOcclusionQuery.h>
  15. #include <AnKi/Gr/Vulkan/VkTimestampQuery.h>
  16. #include <AnKi/Gr/Vulkan/VkPipelineQuery.h>
  17. #include <AnKi/Gr/RenderGraph.h>
  18. #include <AnKi/Gr/Vulkan/VkAccelerationStructure.h>
  19. #include <AnKi/Gr/Vulkan/VkGrUpscaler.h>
  20. #include <AnKi/Gr/Vulkan/VkFence.h>
  21. #include <AnKi/Gr/Vulkan/VkGpuMemoryManager.h>
  22. #include <AnKi/Gr/Vulkan/VkDescriptor.h>
  23. #include <AnKi/Window/NativeWindow.h>
  24. #if ANKI_WINDOWING_SYSTEM_SDL
  25. # include <AnKi/Window/NativeWindowSdl.h>
  26. # include <SDL3/SDL_vulkan.h>
  27. #elif ANKI_WINDOWING_SYSTEM_ANDROID
  28. # include <AnKi/Window/NativeWindowAndroid.h>
  29. #elif ANKI_WINDOWING_SYSTEM_HEADLESS
  30. // Nothing extra
  31. #else
  32. # error "Unsupported"
  33. #endif
  34. namespace anki {
  35. // DLSS related
  36. #define ANKI_VK_NVX_BINARY_IMPORT "VK_NVX_binary_import"
  37. template<>
  38. template<>
  39. GrManager& MakeSingletonPtr<GrManager>::allocateSingleton<>()
  40. {
  41. ANKI_ASSERT(m_global == nullptr);
  42. m_global = new GrManagerImpl;
  43. #if ANKI_ASSERTIONS_ENABLED
  44. ++g_singletonsAllocated;
  45. #endif
  46. return *m_global;
  47. }
  48. template<>
  49. void MakeSingletonPtr<GrManager>::freeSingleton()
  50. {
  51. if(m_global)
  52. {
  53. delete static_cast<GrManagerImpl*>(m_global);
  54. m_global = nullptr;
  55. #if ANKI_ASSERTIONS_ENABLED
  56. --g_singletonsAllocated;
  57. #endif
  58. }
  59. }
  60. GrManager::GrManager()
  61. {
  62. }
  63. GrManager::~GrManager()
  64. {
  65. }
  66. Error GrManager::init(GrManagerInitInfo& inf)
  67. {
  68. ANKI_VK_SELF(GrManagerImpl);
  69. return self.init(inf);
  70. }
  71. void GrManager::beginFrame()
  72. {
  73. ANKI_VK_SELF(GrManagerImpl);
  74. self.beginFrameInternal();
  75. }
  76. TexturePtr GrManager::acquireNextPresentableTexture()
  77. {
  78. ANKI_VK_SELF(GrManagerImpl);
  79. return self.acquireNextPresentableTexture();
  80. }
  81. void GrManager::endFrame()
  82. {
  83. ANKI_VK_SELF(GrManagerImpl);
  84. self.endFrameInternal();
  85. }
  86. void GrManager::finish()
  87. {
  88. ANKI_VK_SELF(GrManagerImpl);
  89. self.finishInternal();
  90. }
  91. #define ANKI_NEW_GR_OBJECT(type) \
  92. type##Ptr GrManager::new##type(const type##InitInfo& init) \
  93. { \
  94. type##Ptr ptr(type::newInstance(init)); \
  95. if(!ptr.isCreated()) [[unlikely]] \
  96. { \
  97. ANKI_VK_LOGF("Failed to create a " ANKI_STRINGIZE(type) " object"); \
  98. } \
  99. return ptr; \
  100. }
  101. #define ANKI_NEW_GR_OBJECT_NO_INIT_INFO(type) \
  102. type##Ptr GrManager::new##type() \
  103. { \
  104. type##Ptr ptr(type::newInstance()); \
  105. if(!ptr.isCreated()) [[unlikely]] \
  106. { \
  107. ANKI_VK_LOGF("Failed to create a " ANKI_STRINGIZE(type) " object"); \
  108. } \
  109. return ptr; \
  110. }
  111. ANKI_NEW_GR_OBJECT(Buffer)
  112. ANKI_NEW_GR_OBJECT(Texture)
  113. ANKI_NEW_GR_OBJECT(Sampler)
  114. ANKI_NEW_GR_OBJECT(Shader)
  115. ANKI_NEW_GR_OBJECT(ShaderProgram)
  116. ANKI_NEW_GR_OBJECT(CommandBuffer)
  117. ANKI_NEW_GR_OBJECT_NO_INIT_INFO(OcclusionQuery)
  118. ANKI_NEW_GR_OBJECT_NO_INIT_INFO(TimestampQuery)
  119. ANKI_NEW_GR_OBJECT(PipelineQuery)
  120. ANKI_NEW_GR_OBJECT_NO_INIT_INFO(RenderGraph)
  121. ANKI_NEW_GR_OBJECT(AccelerationStructure)
  122. ANKI_NEW_GR_OBJECT(GrUpscaler)
  123. #undef ANKI_NEW_GR_OBJECT
  124. #undef ANKI_NEW_GR_OBJECT_NO_INIT_INFO
  125. void GrManager::submit(WeakArray<CommandBuffer*> cmdbs, WeakArray<Fence*> waitFences, FencePtr* signalFence)
  126. {
  127. ANKI_VK_SELF(GrManagerImpl);
  128. self.submitInternal(cmdbs, waitFences, signalFence);
  129. }
  130. PtrSize GrManager::getAccelerationStructureMemoryRequirement(const AccelerationStructureInitInfo& init) const
  131. {
  132. ANKI_VK_SELF_CONST(GrManagerImpl);
  133. PtrSize asSize, unused;
  134. AccelerationStructureImpl::getMemoryRequirement(init, asSize, unused);
  135. return asSize + self.m_caps.m_asBufferAlignment;
  136. }
  137. GrManagerImpl::~GrManagerImpl()
  138. {
  139. ANKI_VK_LOGI("Destroying Vulkan backend");
  140. // 1st THING: wait for the GPU
  141. finishInternal();
  142. // 2nd THING: Destroy everything that has a reference to GrObjects.
  143. m_crntSwapchain.reset(nullptr);
  144. SwapchainFactory::freeSingleton();
  145. for(U32 frame = 0; frame < m_perFrame.getSize(); ++frame)
  146. {
  147. m_frame = frame;
  148. deleteObjectsMarkedForDeletion();
  149. }
  150. // 3rd THING: Continue with the rest
  151. CommandBufferFactory::freeSingleton();
  152. OcclusionQueryFactory::freeSingleton();
  153. TimestampQueryFactory::freeSingleton();
  154. PrimitivesPassedClippingFactory::freeSingleton();
  155. SemaphoreFactory::freeSingleton();
  156. SamplerFactory::freeSingleton();
  157. GpuMemoryManager::freeSingleton();
  158. PipelineLayoutFactory2::freeSingleton();
  159. BindlessDescriptorSet::freeSingleton();
  160. PipelineCache::freeSingleton();
  161. FenceFactory::freeSingleton();
  162. if(m_device)
  163. {
  164. vkDestroyDevice(m_device, nullptr);
  165. }
  166. if(m_surface)
  167. {
  168. vkDestroySurfaceKHR(m_instance, m_surface, nullptr);
  169. }
  170. if(m_debugUtilsMessager)
  171. {
  172. vkDestroyDebugUtilsMessengerEXT(m_instance, m_debugUtilsMessager, nullptr);
  173. }
  174. if(m_instance)
  175. {
  176. #if ANKI_GR_MANAGER_DEBUG_MEMMORY
  177. VkAllocationCallbacks* pallocCbs = &m_debugAllocCbs;
  178. #else
  179. VkAllocationCallbacks* pallocCbs = nullptr;
  180. #endif
  181. vkDestroyInstance(m_instance, pallocCbs);
  182. }
  183. m_cacheDir.destroy();
  184. GrMemoryPool::freeSingleton();
  185. }
  186. Error GrManagerImpl::init(const GrManagerInitInfo& init)
  187. {
  188. const Error err = initInternal(init);
  189. if(err)
  190. {
  191. ANKI_VK_LOGE("Vulkan initialization failed");
  192. return Error::kFunctionFailed;
  193. }
  194. return Error::kNone;
  195. }
  196. Error GrManagerImpl::initInternal(const GrManagerInitInfo& init)
  197. {
  198. ANKI_VK_LOGI("Initializing Vulkan backend");
  199. GrMemoryPool::allocateSingleton(init.m_allocCallback, init.m_allocCallbackUserData);
  200. m_cacheDir = init.m_cacheDirectory;
  201. ANKI_CHECK(initInstance());
  202. ANKI_CHECK(initSurface());
  203. ANKI_CHECK(initDevice());
  204. PipelineCache::allocateSingleton();
  205. ANKI_CHECK(PipelineCache::getSingleton().init(init.m_cacheDirectory));
  206. ANKI_CHECK(initMemory());
  207. CommandBufferFactory::allocateSingleton();
  208. FenceFactory::allocateSingleton();
  209. SemaphoreFactory::allocateSingleton();
  210. OcclusionQueryFactory::allocateSingleton();
  211. TimestampQueryFactory::allocateSingleton();
  212. PrimitivesPassedClippingFactory::allocateSingleton();
  213. SamplerFactory::allocateSingleton();
  214. SwapchainFactory::allocateSingleton(U32(g_cvarGrVsync));
  215. m_crntSwapchain = SwapchainFactory::getSingleton().newInstance();
  216. // See if unaligned formats are supported
  217. {
  218. m_capabilities.m_unalignedBbpTextureFormats = true;
  219. VkImageFormatProperties props = {};
  220. VkResult res = vkGetPhysicalDeviceImageFormatProperties(m_physicalDevice, VK_FORMAT_R8G8B8_UNORM, VK_IMAGE_TYPE_2D, VK_IMAGE_TILING_OPTIMAL,
  221. VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT, 0, &props);
  222. if(res == VK_ERROR_FORMAT_NOT_SUPPORTED)
  223. {
  224. m_capabilities.m_unalignedBbpTextureFormats = false;
  225. }
  226. res = vkGetPhysicalDeviceImageFormatProperties(m_physicalDevice, VK_FORMAT_R32G32B32_SFLOAT, VK_IMAGE_TYPE_2D, VK_IMAGE_TILING_OPTIMAL,
  227. VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT, 0, &props);
  228. if(res == VK_ERROR_FORMAT_NOT_SUPPORTED)
  229. {
  230. m_capabilities.m_unalignedBbpTextureFormats = false;
  231. }
  232. if(!m_capabilities.m_unalignedBbpTextureFormats)
  233. {
  234. ANKI_VK_LOGV("R8G8B8, R32G32B32 image formats are not supported");
  235. }
  236. }
  237. BindlessDescriptorSet::allocateSingleton();
  238. ANKI_CHECK(BindlessDescriptorSet::getSingleton().init());
  239. PipelineLayoutFactory2::allocateSingleton();
  240. return Error::kNone;
  241. }
  242. Error GrManagerImpl::initInstance()
  243. {
  244. // Init VOLK
  245. //
  246. ANKI_VK_CHECK(volkInitialize());
  247. // Create the instance
  248. //
  249. const U8 vulkanMinor = 1;
  250. const U8 vulkanMajor = 3;
  251. VkApplicationInfo app = {};
  252. app.sType = VK_STRUCTURE_TYPE_APPLICATION_INFO;
  253. app.pApplicationName = "unamed";
  254. app.applicationVersion = 1;
  255. app.pEngineName = "AnKi 3D Engine";
  256. app.engineVersion = (ANKI_VERSION_MAJOR << 16) | ANKI_VERSION_MINOR;
  257. app.apiVersion = VK_MAKE_VERSION(vulkanMajor, vulkanMinor, 0);
  258. VkInstanceCreateInfo ci = {};
  259. ci.sType = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO;
  260. ci.pApplicationInfo = &app;
  261. // Instance layers
  262. GrDynamicArray<const char*> layersToEnable;
  263. GrList<GrString> layersToEnableStrings;
  264. {
  265. U32 layerCount;
  266. vkEnumerateInstanceLayerProperties(&layerCount, nullptr);
  267. if(layerCount)
  268. {
  269. GrDynamicArray<VkLayerProperties> layerProps;
  270. layerProps.resize(layerCount);
  271. vkEnumerateInstanceLayerProperties(&layerCount, &layerProps[0]);
  272. ANKI_VK_LOGV("Found the following instance layers:");
  273. for(const VkLayerProperties& layer : layerProps)
  274. {
  275. ANKI_VK_LOGV("\t%s", layer.layerName);
  276. CString layerName = layer.layerName;
  277. Bool enableLayer = (g_cvarGrValidation || g_cvarGrDebugPrintf) && layerName == "VK_LAYER_KHRONOS_validation";
  278. enableLayer = enableLayer || (!CString(g_cvarGrVkLayers).isEmpty() && CString(g_cvarGrVkLayers).find(layerName) != CString::kNpos);
  279. if(enableLayer)
  280. {
  281. layersToEnableStrings.emplaceBack(layer.layerName);
  282. layersToEnable.emplaceBack(layersToEnableStrings.getBack().cstr());
  283. }
  284. }
  285. }
  286. if(layersToEnable.getSize())
  287. {
  288. ANKI_VK_LOGI("Will enable the following instance layers:");
  289. for(const char* name : layersToEnable)
  290. {
  291. ANKI_VK_LOGI("\t%s", name);
  292. }
  293. ci.enabledLayerCount = layersToEnable.getSize();
  294. ci.ppEnabledLayerNames = &layersToEnable[0];
  295. }
  296. }
  297. // Validation features
  298. GrDynamicArray<VkValidationFeatureEnableEXT> enabledValidationFeatures;
  299. GrDynamicArray<VkValidationFeatureDisableEXT> disabledValidationFeatures;
  300. if(g_cvarGrDebugPrintf)
  301. {
  302. enabledValidationFeatures.emplaceBack(VK_VALIDATION_FEATURE_ENABLE_DEBUG_PRINTF_EXT);
  303. }
  304. if(g_cvarGrDebugPrintf && !g_cvarGrValidation)
  305. {
  306. disabledValidationFeatures.emplaceBack(VK_VALIDATION_FEATURE_DISABLE_ALL_EXT);
  307. }
  308. if(g_cvarGrValidation && g_cvarGrGpuValidation)
  309. {
  310. enabledValidationFeatures.emplaceBack(VK_VALIDATION_FEATURE_ENABLE_GPU_ASSISTED_EXT);
  311. }
  312. VkValidationFeaturesEXT validationFeatures = {};
  313. if(enabledValidationFeatures.getSize() || disabledValidationFeatures.getSize())
  314. {
  315. validationFeatures.sType = VK_STRUCTURE_TYPE_VALIDATION_FEATURES_EXT;
  316. validationFeatures.disabledValidationFeatureCount = disabledValidationFeatures.getSize();
  317. validationFeatures.enabledValidationFeatureCount = enabledValidationFeatures.getSize();
  318. validationFeatures.pDisabledValidationFeatures = disabledValidationFeatures.getBegin();
  319. validationFeatures.pEnabledValidationFeatures = enabledValidationFeatures.getBegin();
  320. validationFeatures.pNext = ci.pNext;
  321. ci.pNext = &validationFeatures;
  322. }
  323. // Extensions
  324. GrDynamicArray<const char*> instExtensions;
  325. GrDynamicArray<VkExtensionProperties> instExtensionInf;
  326. U32 extCount = 0;
  327. vkEnumerateInstanceExtensionProperties(nullptr, &extCount, nullptr);
  328. if(extCount)
  329. {
  330. instExtensions.resize(extCount);
  331. instExtensionInf.resize(extCount);
  332. vkEnumerateInstanceExtensionProperties(nullptr, &extCount, &instExtensionInf[0]);
  333. ANKI_VK_LOGV("Found the following instance extensions:");
  334. for(U32 i = 0; i < extCount; ++i)
  335. {
  336. ANKI_VK_LOGV("\t%s", instExtensionInf[i].extensionName);
  337. }
  338. U32 instExtensionCount = 0;
  339. for(U32 i = 0; i < extCount; ++i)
  340. {
  341. const CString extensionName = instExtensionInf[i].extensionName;
  342. #if ANKI_WINDOWING_SYSTEM_HEADLESS
  343. if(extensionName == VK_EXT_HEADLESS_SURFACE_EXTENSION_NAME)
  344. {
  345. m_extensions |= VulkanExtensions::kEXT_headless_surface;
  346. instExtensions[instExtensionCount++] = VK_EXT_HEADLESS_SURFACE_EXTENSION_NAME;
  347. }
  348. #elif ANKI_OS_LINUX
  349. if(extensionName == VK_KHR_WAYLAND_SURFACE_EXTENSION_NAME)
  350. {
  351. m_extensions |= VulkanExtensions::kKHR_wayland_surface;
  352. instExtensions[instExtensionCount++] = VK_KHR_WAYLAND_SURFACE_EXTENSION_NAME;
  353. }
  354. else if(extensionName == VK_KHR_XCB_SURFACE_EXTENSION_NAME)
  355. {
  356. m_extensions |= VulkanExtensions::kKHR_xcb_surface;
  357. instExtensions[instExtensionCount++] = VK_KHR_XCB_SURFACE_EXTENSION_NAME;
  358. }
  359. else if(extensionName == VK_KHR_XLIB_SURFACE_EXTENSION_NAME)
  360. {
  361. m_extensions |= VulkanExtensions::kKHR_xlib_surface;
  362. instExtensions[instExtensionCount++] = VK_KHR_XLIB_SURFACE_EXTENSION_NAME;
  363. }
  364. #elif ANKI_OS_WINDOWS
  365. if(extensionName == VK_KHR_WIN32_SURFACE_EXTENSION_NAME)
  366. {
  367. m_extensions |= VulkanExtensions::kKHR_win32_surface;
  368. instExtensions[instExtensionCount++] = VK_KHR_WIN32_SURFACE_EXTENSION_NAME;
  369. }
  370. #elif ANKI_OS_ANDROID
  371. if(extensionName == VK_KHR_ANDROID_SURFACE_EXTENSION_NAME)
  372. {
  373. m_extensions |= VulkanExtensions::kKHR_android_surface;
  374. instExtensions[instExtensionCount++] = VK_KHR_ANDROID_SURFACE_EXTENSION_NAME;
  375. }
  376. #else
  377. # error Not implemented
  378. #endif
  379. else if(extensionName == VK_KHR_SURFACE_EXTENSION_NAME)
  380. {
  381. m_extensions |= VulkanExtensions::kKHR_surface;
  382. instExtensions[instExtensionCount++] = VK_KHR_SURFACE_EXTENSION_NAME;
  383. }
  384. else if(extensionName == VK_EXT_DEBUG_UTILS_EXTENSION_NAME && (g_cvarGrDebugMarkers || g_cvarGrValidation || g_cvarGrDebugPrintf))
  385. {
  386. m_extensions |= VulkanExtensions::kEXT_debug_utils;
  387. instExtensions[instExtensionCount++] = VK_EXT_DEBUG_UTILS_EXTENSION_NAME;
  388. }
  389. }
  390. if(!(m_extensions
  391. & (VulkanExtensions::kEXT_headless_surface | VulkanExtensions::kKHR_wayland_surface | VulkanExtensions::kKHR_xcb_surface
  392. | VulkanExtensions::kKHR_xlib_surface | VulkanExtensions::kKHR_win32_surface | VulkanExtensions::kKHR_android_surface)))
  393. {
  394. ANKI_VK_LOGE("Couldn't find suitable surface extension");
  395. return Error::kFunctionFailed;
  396. }
  397. if(instExtensionCount)
  398. {
  399. ANKI_VK_LOGI("Will enable the following instance extensions:");
  400. for(U32 i = 0; i < instExtensionCount; ++i)
  401. {
  402. ANKI_VK_LOGI("\t%s", instExtensions[i]);
  403. }
  404. ci.enabledExtensionCount = instExtensionCount;
  405. ci.ppEnabledExtensionNames = &instExtensions[0];
  406. }
  407. }
  408. #if ANKI_GR_MANAGER_DEBUG_MEMMORY
  409. m_debugAllocCbs = {};
  410. m_debugAllocCbs.pUserData = this;
  411. m_debugAllocCbs.pfnAllocation = allocateCallback;
  412. m_debugAllocCbs.pfnReallocation = reallocateCallback;
  413. m_debugAllocCbs.pfnFree = freeCallback;
  414. VkAllocationCallbacks* pallocCbs = &m_debugAllocCbs;
  415. #else
  416. VkAllocationCallbacks* pallocCbs = nullptr;
  417. #endif
  418. ANKI_VK_CHECK(vkCreateInstance(&ci, pallocCbs, &m_instance));
  419. // Get symbolx
  420. //
  421. volkLoadInstance(m_instance);
  422. // Set debug callbacks
  423. if(!!(m_extensions & VulkanExtensions::kEXT_debug_utils))
  424. {
  425. VkDebugUtilsMessengerCreateInfoEXT info = {};
  426. info.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_MESSENGER_CREATE_INFO_EXT;
  427. info.messageSeverity = VK_DEBUG_UTILS_MESSAGE_SEVERITY_INFO_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_SEVERITY_WARNING_BIT_EXT
  428. | VK_DEBUG_UTILS_MESSAGE_SEVERITY_ERROR_BIT_EXT;
  429. info.messageType = VK_DEBUG_UTILS_MESSAGE_TYPE_VALIDATION_BIT_EXT | VK_DEBUG_UTILS_MESSAGE_TYPE_PERFORMANCE_BIT_EXT
  430. | VK_DEBUG_UTILS_MESSAGE_TYPE_GENERAL_BIT_EXT;
  431. info.pfnUserCallback = debugReportCallbackEXT;
  432. info.pUserData = this;
  433. vkCreateDebugUtilsMessengerEXT(m_instance, &info, nullptr, &m_debugUtilsMessager);
  434. }
  435. // Create the physical device
  436. //
  437. {
  438. uint32_t count = 0;
  439. ANKI_VK_CHECK(vkEnumeratePhysicalDevices(m_instance, &count, nullptr));
  440. if(count < 1)
  441. {
  442. ANKI_VK_LOGE("Wrong number of physical devices");
  443. return Error::kFunctionFailed;
  444. }
  445. GrDynamicArray<VkPhysicalDevice> physicalDevices;
  446. physicalDevices.resize(count);
  447. ANKI_VK_CHECK(vkEnumeratePhysicalDevices(m_instance, &count, &physicalDevices[0]));
  448. class Dev
  449. {
  450. public:
  451. VkPhysicalDevice m_pdev;
  452. VkPhysicalDeviceProperties2 m_vkProps;
  453. };
  454. GrDynamicArray<Dev> devs;
  455. devs.resize(count);
  456. for(U32 devIdx = 0; devIdx < count; ++devIdx)
  457. {
  458. devs[devIdx].m_pdev = physicalDevices[devIdx];
  459. devs[devIdx].m_vkProps.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2;
  460. vkGetPhysicalDeviceProperties2(physicalDevices[devIdx], &devs[devIdx].m_vkProps);
  461. }
  462. // Sort the devices with the most powerful first
  463. std::sort(devs.getBegin(), devs.getEnd(), [](const Dev& a, const Dev& b) {
  464. if(a.m_vkProps.properties.deviceType != b.m_vkProps.properties.deviceType)
  465. {
  466. auto findDeviceTypeWeight = [](VkPhysicalDeviceType type) {
  467. switch(type)
  468. {
  469. case VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU:
  470. return 1.0;
  471. case VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU:
  472. return 2.0;
  473. default:
  474. return 0.0;
  475. }
  476. };
  477. // Put descrete GPUs first
  478. return findDeviceTypeWeight(a.m_vkProps.properties.deviceType) > findDeviceTypeWeight(b.m_vkProps.properties.deviceType);
  479. }
  480. else
  481. {
  482. return a.m_vkProps.properties.apiVersion >= b.m_vkProps.properties.apiVersion;
  483. }
  484. });
  485. const U32 chosenPhysDevIdx = min<U32>(g_cvarGrDevice, devs.getSize() - 1);
  486. ANKI_VK_LOGI("Physical devices:");
  487. for(U32 devIdx = 0; devIdx < count; ++devIdx)
  488. {
  489. ANKI_VK_LOGI((devIdx == chosenPhysDevIdx) ? "\t(Selected) %s" : "\t%s", devs[devIdx].m_vkProps.properties.deviceName);
  490. }
  491. m_capabilities.m_discreteGpu = devs[chosenPhysDevIdx].m_vkProps.properties.deviceType == VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU;
  492. m_physicalDevice = devs[chosenPhysDevIdx].m_pdev;
  493. }
  494. VkPhysicalDeviceProperties2 props2 = {};
  495. props2.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PROPERTIES_2;
  496. vkGetPhysicalDeviceProperties2(m_physicalDevice, &props2);
  497. VkPhysicalDeviceVulkan12Properties props12 = {};
  498. props12.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_VULKAN_1_2_PROPERTIES;
  499. getPhysicalDeviceProperties2(props12);
  500. VkPhysicalDeviceVulkan13Properties props13 = {};
  501. props13.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_VULKAN_1_3_PROPERTIES;
  502. getPhysicalDeviceProperties2(props13);
  503. m_capabilities.m_minWaveSize = props13.minSubgroupSize;
  504. m_capabilities.m_maxWaveSize = props13.maxSubgroupSize;
  505. if(props2.properties.limits.maxComputeWorkGroupInvocations < 1024)
  506. {
  507. ANKI_VK_LOGE("GPU doesn't support at least 1024 workgroup invocations");
  508. return Error::kFunctionFailed;
  509. }
  510. // Find vendor
  511. switch(props2.properties.vendorID)
  512. {
  513. case 0x13B5:
  514. m_capabilities.m_gpuVendor = GpuVendor::kArm;
  515. break;
  516. case 0x10DE:
  517. m_capabilities.m_gpuVendor = GpuVendor::kNvidia;
  518. break;
  519. case 0x1002:
  520. case 0x1022:
  521. m_capabilities.m_gpuVendor = GpuVendor::kAMD;
  522. break;
  523. case 0x8086:
  524. m_capabilities.m_gpuVendor = GpuVendor::kIntel;
  525. break;
  526. case 0x5143:
  527. m_capabilities.m_gpuVendor = GpuVendor::kQualcomm;
  528. break;
  529. default:
  530. m_capabilities.m_gpuVendor = GpuVendor::kUnknown;
  531. }
  532. ANKI_VK_LOGI("GPU is %s. Vendor identified as %s, Driver %s", props2.properties.deviceName, &kGPUVendorStrings[m_capabilities.m_gpuVendor][0],
  533. props12.driverInfo);
  534. // Set limits
  535. m_capabilities.m_constantBufferBindOffsetAlignment =
  536. computeCompoundAlignment<U32>(ANKI_SAFE_ALIGNMENT, U32(props2.properties.limits.minUniformBufferOffsetAlignment));
  537. m_capabilities.m_structuredBufferBindOffsetAlignment =
  538. computeCompoundAlignment<U32>(ANKI_SAFE_ALIGNMENT, U32(props2.properties.limits.minStorageBufferOffsetAlignment));
  539. m_capabilities.m_structuredBufferNaturalAlignment = false;
  540. m_capabilities.m_texelBufferBindOffsetAlignment = max<U32>(ANKI_SAFE_ALIGNMENT, U32(props2.properties.limits.minTexelBufferOffsetAlignment));
  541. m_capabilities.m_computeSharedMemorySize = props2.properties.limits.maxComputeSharedMemorySize;
  542. m_capabilities.m_maxDrawIndirectCount = props2.properties.limits.maxDrawIndirectCount;
  543. m_capabilities.m_majorApiVersion = vulkanMajor;
  544. m_capabilities.m_minorApiVersion = vulkanMinor;
  545. m_caps.m_nonCoherentAtomSize = props2.properties.limits.nonCoherentAtomSize;
  546. m_caps.m_maxTexelBufferElements = props2.properties.limits.maxTexelBufferElements;
  547. m_caps.m_timestampPeriod = props2.properties.limits.timestampPeriod;
  548. // DLSS checks
  549. m_capabilities.m_dlss = ANKI_DLSS && m_capabilities.m_gpuVendor == GpuVendor::kNvidia;
  550. return Error::kNone;
  551. }
  552. Error GrManagerImpl::initDevice()
  553. {
  554. uint32_t count = 0;
  555. vkGetPhysicalDeviceQueueFamilyProperties(m_physicalDevice, &count, nullptr);
  556. ANKI_VK_LOGI("Number of queue families: %u", count);
  557. GrDynamicArray<VkQueueFamilyProperties> queueInfos;
  558. queueInfos.resize(count);
  559. vkGetPhysicalDeviceQueueFamilyProperties(m_physicalDevice, &count, &queueInfos[0]);
  560. Bool generalQueueFamilySupportsMultipleQueues = false;
  561. const VkQueueFlags generalQueueFlags = VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT;
  562. for(U32 i = 0; i < count; ++i)
  563. {
  564. VkBool32 supportsPresent = false;
  565. ANKI_VK_CHECK(vkGetPhysicalDeviceSurfaceSupportKHR(m_physicalDevice, i, m_surface, &supportsPresent));
  566. if(!supportsPresent)
  567. {
  568. continue;
  569. }
  570. if((queueInfos[i].queueFlags & generalQueueFlags) == generalQueueFlags)
  571. {
  572. m_queueFamilyIndices[GpuQueueType::kGeneral] = i;
  573. if(queueInfos[i].queueCount > 1)
  574. {
  575. generalQueueFamilySupportsMultipleQueues = true;
  576. }
  577. }
  578. else if((queueInfos[i].queueFlags & VK_QUEUE_COMPUTE_BIT) && !(queueInfos[i].queueFlags & VK_QUEUE_GRAPHICS_BIT))
  579. {
  580. // This must be the async compute
  581. m_queueFamilyIndices[GpuQueueType::kCompute] = i;
  582. }
  583. }
  584. if(m_queueFamilyIndices[GpuQueueType::kGeneral] == kMaxU32)
  585. {
  586. ANKI_VK_LOGE("Couldn't find a queue family with graphics+compute+transfer+present. Something is wrong");
  587. return Error::kFunctionFailed;
  588. }
  589. const Bool pureAsyncCompute = m_queueFamilyIndices[GpuQueueType::kCompute] != kMaxU32 && g_cvarGrAsyncCompute == 0;
  590. const Bool lowPriorityQueueAsyncCompute = !pureAsyncCompute && generalQueueFamilySupportsMultipleQueues && g_cvarGrAsyncCompute <= 1;
  591. Array<F32, U32(GpuQueueType::kCount)> priorities = {1.0f, 0.5f};
  592. Array<VkDeviceQueueCreateInfo, U32(GpuQueueType::kCount)> q = {};
  593. VkDeviceQueueCreateInfo queueCreateInfo = {};
  594. queueCreateInfo.sType = VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO;
  595. q.fill(queueCreateInfo);
  596. VkDeviceCreateInfo ci = {};
  597. ci.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO;
  598. ci.pQueueCreateInfos = &q[0];
  599. CString asyncComputeMsg;
  600. if(pureAsyncCompute)
  601. {
  602. asyncComputeMsg = "Using pure async compute queue";
  603. q[GpuQueueType::kGeneral].queueFamilyIndex = m_queueFamilyIndices[GpuQueueType::kGeneral];
  604. q[GpuQueueType::kGeneral].queueCount = 1;
  605. q[GpuQueueType::kGeneral].pQueuePriorities = &priorities[0];
  606. q[GpuQueueType::kCompute].queueFamilyIndex = m_queueFamilyIndices[GpuQueueType::kCompute];
  607. q[GpuQueueType::kCompute].queueCount = 1;
  608. q[GpuQueueType::kCompute].pQueuePriorities = &priorities[0];
  609. ci.queueCreateInfoCount = 2;
  610. }
  611. else if(lowPriorityQueueAsyncCompute)
  612. {
  613. asyncComputeMsg = "Using low priority queue in same family as general queue (fallback #1)";
  614. m_queueFamilyIndices[GpuQueueType::kCompute] = m_queueFamilyIndices[GpuQueueType::kGeneral];
  615. q[0].queueFamilyIndex = m_queueFamilyIndices[GpuQueueType::kGeneral];
  616. q[0].queueCount = 2;
  617. q[0].pQueuePriorities = &priorities[0];
  618. ci.queueCreateInfoCount = 1;
  619. }
  620. else
  621. {
  622. asyncComputeMsg = "Can't do much, using general queue (fallback #2)";
  623. m_queueFamilyIndices[GpuQueueType::kCompute] = m_queueFamilyIndices[GpuQueueType::kGeneral];
  624. q[0].queueFamilyIndex = m_queueFamilyIndices[GpuQueueType::kGeneral];
  625. q[0].queueCount = 1;
  626. q[0].pQueuePriorities = &priorities[0];
  627. ci.queueCreateInfoCount = 1;
  628. }
  629. ANKI_VK_LOGI("Async compute: %s", asyncComputeMsg.cstr());
  630. // Extensions
  631. U32 extCount = 0;
  632. vkEnumerateDeviceExtensionProperties(m_physicalDevice, nullptr, &extCount, nullptr);
  633. GrDynamicArray<VkExtensionProperties> extensionInfos; // Keep it alive in the stack
  634. GrDynamicArray<const char*> extensionsToEnable;
  635. if(extCount)
  636. {
  637. extensionInfos.resize(extCount);
  638. extensionsToEnable.resize(extCount);
  639. U32 extensionsToEnableCount = 0;
  640. vkEnumerateDeviceExtensionProperties(m_physicalDevice, nullptr, &extCount, &extensionInfos[0]);
  641. ANKI_VK_LOGV("Found the following device extensions:");
  642. for(U32 i = 0; i < extCount; ++i)
  643. {
  644. ANKI_VK_LOGV("\t%s", extensionInfos[i].extensionName);
  645. }
  646. while(extCount-- != 0)
  647. {
  648. const CString extensionName(&extensionInfos[extCount].extensionName[0]);
  649. if(extensionName == VK_KHR_SWAPCHAIN_EXTENSION_NAME)
  650. {
  651. m_extensions |= VulkanExtensions::kKHR_swapchain;
  652. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  653. }
  654. else if(extensionName == VK_KHR_RAY_TRACING_PIPELINE_EXTENSION_NAME && g_cvarGrRayTracing)
  655. {
  656. m_extensions |= VulkanExtensions::kKHR_ray_tracing;
  657. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  658. m_capabilities.m_rayTracingEnabled = true;
  659. }
  660. else if(extensionName == VK_KHR_RAY_QUERY_EXTENSION_NAME && g_cvarGrRayTracing)
  661. {
  662. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  663. }
  664. else if(extensionName == VK_KHR_ACCELERATION_STRUCTURE_EXTENSION_NAME && g_cvarGrRayTracing)
  665. {
  666. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  667. }
  668. else if(extensionName == VK_KHR_DEFERRED_HOST_OPERATIONS_EXTENSION_NAME && g_cvarGrRayTracing)
  669. {
  670. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  671. }
  672. else if(extensionName == VK_KHR_PIPELINE_LIBRARY_EXTENSION_NAME && g_cvarGrRayTracing)
  673. {
  674. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  675. }
  676. else if(extensionName == VK_KHR_PIPELINE_EXECUTABLE_PROPERTIES_EXTENSION_NAME && g_cvarCoreDisplayStats > 1)
  677. {
  678. m_extensions |= VulkanExtensions::kKHR_pipeline_executable_properties;
  679. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  680. }
  681. else if(extensionName == VK_KHR_SHADER_NON_SEMANTIC_INFO_EXTENSION_NAME && g_cvarGrDebugPrintf)
  682. {
  683. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  684. }
  685. else if(extensionName == VK_KHR_FRAGMENT_SHADING_RATE_EXTENSION_NAME && g_cvarGrVrs)
  686. {
  687. m_extensions |= VulkanExtensions::kKHR_fragment_shading_rate;
  688. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  689. }
  690. else if(extensionName == VK_EXT_ASTC_DECODE_MODE_EXTENSION_NAME)
  691. {
  692. m_extensions |= VulkanExtensions::kEXT_astc_decode_mode;
  693. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  694. }
  695. else if(extensionName == VK_EXT_TEXTURE_COMPRESSION_ASTC_HDR_EXTENSION_NAME)
  696. {
  697. m_extensions |= VulkanExtensions::kEXT_texture_compression_astc_hdr;
  698. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  699. }
  700. else if(m_capabilities.m_dlss && extensionName == VK_KHR_PUSH_DESCRIPTOR_EXTENSION_NAME)
  701. {
  702. m_extensions |= VulkanExtensions::kKHR_push_descriptor;
  703. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  704. }
  705. else if(m_capabilities.m_dlss && extensionName == ANKI_VK_NVX_BINARY_IMPORT)
  706. {
  707. m_extensions |= VulkanExtensions::kNVX_binary_import;
  708. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  709. }
  710. else if(m_capabilities.m_dlss && extensionName == VK_NVX_IMAGE_VIEW_HANDLE_EXTENSION_NAME)
  711. {
  712. m_extensions |= VulkanExtensions::kNVX_image_view_handle;
  713. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  714. }
  715. else if(extensionName == VK_EXT_MESH_SHADER_EXTENSION_NAME && g_cvarGrMeshShaders)
  716. {
  717. m_extensions |= VulkanExtensions::kEXT_mesh_shader;
  718. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  719. }
  720. else if(extensionName == VK_KHR_FRAGMENT_SHADER_BARYCENTRIC_EXTENSION_NAME)
  721. {
  722. m_extensions |= VulkanExtensions::kKHR_fragment_shader_barycentric;
  723. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  724. }
  725. else if(extensionName == VK_KHR_RAY_TRACING_POSITION_FETCH_EXTENSION_NAME && g_cvarGrRayTracing)
  726. {
  727. m_extensions |= VulkanExtensions::kKHR_ray_tracing_position_fetch;
  728. extensionsToEnable[extensionsToEnableCount++] = extensionName.cstr();
  729. }
  730. }
  731. ANKI_VK_LOGI("Will enable the following device extensions:");
  732. for(U32 i = 0; i < extensionsToEnableCount; ++i)
  733. {
  734. ANKI_VK_LOGI("\t%s", extensionsToEnable[i]);
  735. }
  736. ci.enabledExtensionCount = extensionsToEnableCount;
  737. ci.ppEnabledExtensionNames = &extensionsToEnable[0];
  738. }
  739. // Enable/disable generic features
  740. VkPhysicalDeviceFeatures2 devFeatures = {};
  741. {
  742. devFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FEATURES_2;
  743. vkGetPhysicalDeviceFeatures2(m_physicalDevice, &devFeatures);
  744. devFeatures.features.robustBufferAccess = (g_cvarGrValidation && devFeatures.features.robustBufferAccess) ? true : false;
  745. ANKI_VK_LOGI("Robust buffer access is %s", (devFeatures.features.robustBufferAccess) ? "enabled" : "disabled");
  746. if(devFeatures.features.pipelineStatisticsQuery)
  747. {
  748. m_capabilities.m_pipelineQuery = true;
  749. ANKI_VK_LOGV("GPU supports pipeline statistics queries");
  750. }
  751. appendPNextList(ci, &devFeatures);
  752. }
  753. // 1.1 features
  754. VkPhysicalDeviceVulkan11Features features11 = {};
  755. {
  756. features11.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_VULKAN_1_1_FEATURES;
  757. getPhysicalDevicaFeatures2(features11);
  758. appendPNextList(ci, &features11);
  759. }
  760. #define ANKI_ASSERT_SUPPORTED(features, feature) \
  761. if(!features.feature) \
  762. { \
  763. ANKI_VK_LOGE(#feature " not supported"); \
  764. return Error::kFunctionFailed; \
  765. }
  766. // 1.2 features
  767. VkPhysicalDeviceVulkan12Features features12 = {};
  768. {
  769. features12.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_VULKAN_1_2_FEATURES;
  770. getPhysicalDevicaFeatures2(features12);
  771. ANKI_ASSERT_SUPPORTED(features12, descriptorIndexing)
  772. ANKI_ASSERT_SUPPORTED(features12, shaderSampledImageArrayNonUniformIndexing)
  773. ANKI_ASSERT_SUPPORTED(features12, shaderStorageImageArrayNonUniformIndexing)
  774. ANKI_ASSERT_SUPPORTED(features12, descriptorBindingSampledImageUpdateAfterBind)
  775. ANKI_ASSERT_SUPPORTED(features12, descriptorBindingStorageImageUpdateAfterBind)
  776. ANKI_ASSERT_SUPPORTED(features12, descriptorBindingUpdateUnusedWhilePending)
  777. ANKI_ASSERT_SUPPORTED(features12, samplerFilterMinmax)
  778. ANKI_ASSERT_SUPPORTED(features12, hostQueryReset)
  779. ANKI_ASSERT_SUPPORTED(features12, timelineSemaphore)
  780. ANKI_ASSERT_SUPPORTED(features12, drawIndirectCount)
  781. ANKI_ASSERT_SUPPORTED(features12, bufferDeviceAddress)
  782. features12.bufferDeviceAddressCaptureReplay = !!features12.bufferDeviceAddressCaptureReplay && g_cvarGrDebugMarkers;
  783. features12.bufferDeviceAddressMultiDevice = false;
  784. ANKI_ASSERT_SUPPORTED(features12, shaderFloat16)
  785. ANKI_ASSERT_SUPPORTED(features12, scalarBlockLayout)
  786. ANKI_ASSERT_SUPPORTED(features12, shaderBufferInt64Atomics)
  787. appendPNextList(ci, &features12);
  788. }
  789. // 1.3 features
  790. VkPhysicalDeviceVulkan13Features features13 = {};
  791. {
  792. features13.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_VULKAN_1_3_FEATURES;
  793. getPhysicalDevicaFeatures2(features13);
  794. ANKI_ASSERT_SUPPORTED(features13, dynamicRendering)
  795. ANKI_ASSERT_SUPPORTED(features13, maintenance4)
  796. #if ANKI_PLATFORM_MOBILE
  797. ANKI_ASSERT_SUPPORTED(features13, textureCompressionASTC_HDR)
  798. #endif
  799. appendPNextList(ci, &features13);
  800. }
  801. // Set RT features
  802. VkPhysicalDeviceRayTracingPipelineFeaturesKHR rtPipelineFeatures = {};
  803. VkPhysicalDeviceRayQueryFeaturesKHR rayQueryFeatures = {};
  804. VkPhysicalDeviceAccelerationStructureFeaturesKHR accelerationStructureFeatures = {};
  805. VkPhysicalDeviceRayTracingPositionFetchFeaturesKHR positionFetchFeatures = {};
  806. if(!!(m_extensions & VulkanExtensions::kKHR_ray_tracing))
  807. {
  808. if(!(m_extensions & VulkanExtensions::kKHR_ray_tracing_position_fetch))
  809. {
  810. ANKI_VK_LOGE(VK_KHR_RAY_TRACING_POSITION_FETCH_EXTENSION_NAME " is required");
  811. return Error::kFunctionFailed;
  812. }
  813. rtPipelineFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_RAY_TRACING_PIPELINE_FEATURES_KHR;
  814. rayQueryFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_RAY_QUERY_FEATURES_KHR;
  815. accelerationStructureFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_ACCELERATION_STRUCTURE_FEATURES_KHR;
  816. positionFetchFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_RAY_TRACING_POSITION_FETCH_FEATURES_KHR;
  817. VkPhysicalDeviceFeatures2 features = {};
  818. features.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FEATURES_2;
  819. features.pNext = &rtPipelineFeatures;
  820. rtPipelineFeatures.pNext = &rayQueryFeatures;
  821. rayQueryFeatures.pNext = &accelerationStructureFeatures;
  822. accelerationStructureFeatures.pNext = &positionFetchFeatures;
  823. vkGetPhysicalDeviceFeatures2(m_physicalDevice, &features);
  824. if(!rtPipelineFeatures.rayTracingPipeline || !rayQueryFeatures.rayQuery || !accelerationStructureFeatures.accelerationStructure)
  825. {
  826. ANKI_VK_LOGE("Ray tracing and ray query are both required");
  827. return Error::kFunctionFailed;
  828. }
  829. if(!positionFetchFeatures.rayTracingPositionFetch)
  830. {
  831. ANKI_VK_LOGE(VK_KHR_RAY_TRACING_POSITION_FETCH_EXTENSION_NAME " should be really really supported");
  832. return Error::kFunctionFailed;
  833. }
  834. // Only enable what's necessary
  835. rtPipelineFeatures.rayTracingPipelineShaderGroupHandleCaptureReplay = false;
  836. rtPipelineFeatures.rayTracingPipelineShaderGroupHandleCaptureReplayMixed = false;
  837. rtPipelineFeatures.rayTraversalPrimitiveCulling = false;
  838. accelerationStructureFeatures.accelerationStructureCaptureReplay = false;
  839. accelerationStructureFeatures.accelerationStructureHostCommands = false;
  840. accelerationStructureFeatures.descriptorBindingAccelerationStructureUpdateAfterBind = false;
  841. appendPNextList(ci, &accelerationStructureFeatures);
  842. appendPNextList(ci, &rayQueryFeatures);
  843. appendPNextList(ci, &rtPipelineFeatures);
  844. appendPNextList(ci, &positionFetchFeatures);
  845. // Get some more stuff
  846. VkPhysicalDeviceAccelerationStructurePropertiesKHR props = {};
  847. props.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_ACCELERATION_STRUCTURE_PROPERTIES_KHR;
  848. getPhysicalDeviceProperties2(props);
  849. m_caps.m_asBuildScratchAlignment = props.minAccelerationStructureScratchOffsetAlignment;
  850. VkPhysicalDeviceRayTracingPipelinePropertiesKHR rtprops = {};
  851. rtprops.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_RAY_TRACING_PIPELINE_PROPERTIES_KHR;
  852. getPhysicalDeviceProperties2(rtprops);
  853. m_capabilities.m_shaderGroupHandleSize = rtprops.shaderGroupHandleSize;
  854. m_capabilities.m_sbtRecordAlignment = rtprops.shaderGroupBaseAlignment;
  855. }
  856. // Pipeline features
  857. VkPhysicalDevicePipelineExecutablePropertiesFeaturesKHR pplineExecutablePropertiesFeatures = {};
  858. if(!!(m_extensions & VulkanExtensions::kKHR_pipeline_executable_properties))
  859. {
  860. pplineExecutablePropertiesFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_PIPELINE_EXECUTABLE_PROPERTIES_FEATURES_KHR;
  861. pplineExecutablePropertiesFeatures.pipelineExecutableInfo = true;
  862. appendPNextList(ci, &pplineExecutablePropertiesFeatures);
  863. }
  864. // VRS
  865. VkPhysicalDeviceFragmentShadingRateFeaturesKHR fragmentShadingRateFeatures = {};
  866. if(!(m_extensions & VulkanExtensions::kKHR_fragment_shading_rate))
  867. {
  868. ANKI_VK_LOGI(VK_KHR_FRAGMENT_SHADING_RATE_EXTENSION_NAME " is not supported or disabled");
  869. m_capabilities.m_vrs = false;
  870. }
  871. else
  872. {
  873. m_capabilities.m_vrs = true;
  874. fragmentShadingRateFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FRAGMENT_SHADING_RATE_FEATURES_KHR;
  875. getPhysicalDevicaFeatures2(fragmentShadingRateFeatures);
  876. // Some checks
  877. if(!fragmentShadingRateFeatures.attachmentFragmentShadingRate || !fragmentShadingRateFeatures.pipelineFragmentShadingRate)
  878. {
  879. ANKI_VK_LOGW(VK_KHR_FRAGMENT_SHADING_RATE_EXTENSION_NAME " doesn't support attachment and/or pipeline rates. Will disable VRS");
  880. m_capabilities.m_vrs = false;
  881. }
  882. else
  883. {
  884. // Disable some things
  885. fragmentShadingRateFeatures.primitiveFragmentShadingRate = false;
  886. }
  887. if(m_capabilities.m_vrs)
  888. {
  889. VkPhysicalDeviceFragmentShadingRatePropertiesKHR fragmentShadingRateProperties = {};
  890. fragmentShadingRateProperties.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FRAGMENT_SHADING_RATE_PROPERTIES_KHR;
  891. getPhysicalDeviceProperties2(fragmentShadingRateProperties);
  892. if(fragmentShadingRateProperties.minFragmentShadingRateAttachmentTexelSize.width > 16
  893. || fragmentShadingRateProperties.minFragmentShadingRateAttachmentTexelSize.height > 16
  894. || fragmentShadingRateProperties.maxFragmentShadingRateAttachmentTexelSize.width < 8
  895. || fragmentShadingRateProperties.maxFragmentShadingRateAttachmentTexelSize.height < 8)
  896. {
  897. ANKI_VK_LOGW(VK_KHR_FRAGMENT_SHADING_RATE_EXTENSION_NAME
  898. " doesn't support 8x8 or 16x16 shading rate attachment texel size. Will disable VRS");
  899. m_capabilities.m_vrs = false;
  900. }
  901. else
  902. {
  903. m_capabilities.m_minShadingRateImageTexelSize = max(fragmentShadingRateProperties.minFragmentShadingRateAttachmentTexelSize.width,
  904. fragmentShadingRateProperties.minFragmentShadingRateAttachmentTexelSize.height);
  905. }
  906. }
  907. if(m_capabilities.m_vrs)
  908. {
  909. appendPNextList(ci, &fragmentShadingRateFeatures);
  910. }
  911. }
  912. // Mesh shaders
  913. VkPhysicalDeviceMeshShaderFeaturesEXT meshShadersFeatures = {};
  914. if(!!(m_extensions & VulkanExtensions::kEXT_mesh_shader))
  915. {
  916. m_capabilities.m_meshShaders = true;
  917. meshShadersFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MESH_SHADER_FEATURES_EXT;
  918. getPhysicalDevicaFeatures2(meshShadersFeatures);
  919. if(meshShadersFeatures.taskShader == false)
  920. {
  921. ANKI_LOGE(VK_EXT_MESH_SHADER_EXTENSION_NAME " doesn't support task shaders");
  922. return Error::kFunctionFailed;
  923. }
  924. meshShadersFeatures.multiviewMeshShader = false;
  925. meshShadersFeatures.primitiveFragmentShadingRateMeshShader = false;
  926. appendPNextList(ci, &meshShadersFeatures);
  927. ANKI_VK_LOGI(VK_EXT_MESH_SHADER_EXTENSION_NAME " is supported and enabled");
  928. }
  929. else
  930. {
  931. ANKI_VK_LOGI(VK_EXT_MESH_SHADER_EXTENSION_NAME " is not supported or disabled ");
  932. }
  933. // Barycentrics
  934. VkPhysicalDeviceFragmentShaderBarycentricFeaturesKHR baryFeatures = {};
  935. if(!!(m_extensions & VulkanExtensions::kKHR_fragment_shader_barycentric))
  936. {
  937. baryFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FRAGMENT_SHADER_BARYCENTRIC_FEATURES_KHR;
  938. getPhysicalDevicaFeatures2(baryFeatures);
  939. if(baryFeatures.fragmentShaderBarycentric == false)
  940. {
  941. ANKI_VK_LOGE("VkPhysicalDeviceFragmentShaderBarycentricFeaturesKHR::fragmentShaderBarycentric is false");
  942. return Error::kFunctionFailed;
  943. }
  944. appendPNextList(ci, &baryFeatures);
  945. m_capabilities.m_barycentrics = true;
  946. }
  947. ANKI_VK_CHECK(vkCreateDevice(m_physicalDevice, &ci, nullptr, &m_device));
  948. // Get the queues
  949. vkGetDeviceQueue(m_device, m_queueFamilyIndices[GpuQueueType::kGeneral], 0, &m_queues[GpuQueueType::kGeneral]);
  950. trySetVulkanHandleName("General", VK_OBJECT_TYPE_QUEUE, m_queues[GpuQueueType::kGeneral]);
  951. if(pureAsyncCompute)
  952. {
  953. vkGetDeviceQueue(m_device, m_queueFamilyIndices[GpuQueueType::kCompute], 0, &m_queues[GpuQueueType::kCompute]);
  954. trySetVulkanHandleName("AsyncCompute", VK_OBJECT_TYPE_QUEUE, m_queues[GpuQueueType::kGeneral]);
  955. }
  956. else if(lowPriorityQueueAsyncCompute)
  957. {
  958. vkGetDeviceQueue(m_device, m_queueFamilyIndices[GpuQueueType::kGeneral], 1, &m_queues[GpuQueueType::kCompute]);
  959. trySetVulkanHandleName("GeneralLowPriority", VK_OBJECT_TYPE_QUEUE, m_queues[GpuQueueType::kCompute]);
  960. }
  961. else
  962. {
  963. m_queues[GpuQueueType::kCompute] = nullptr;
  964. }
  965. return Error::kNone;
  966. }
  967. Error GrManagerImpl::initMemory()
  968. {
  969. vkGetPhysicalDeviceMemoryProperties(m_physicalDevice, &m_memoryProperties);
  970. // Print some info
  971. ANKI_VK_LOGV("Vulkan memory info:");
  972. for(U32 i = 0; i < m_memoryProperties.memoryHeapCount; ++i)
  973. {
  974. ANKI_VK_LOGV("\tHeap %u size %zu", i, m_memoryProperties.memoryHeaps[i].size);
  975. }
  976. for(U32 i = 0; i < m_memoryProperties.memoryTypeCount; ++i)
  977. {
  978. ANKI_VK_LOGV("\tMem type %u points to heap %u, flags %" ANKI_PRIb32, i, m_memoryProperties.memoryTypes[i].heapIndex,
  979. ANKI_FORMAT_U32(m_memoryProperties.memoryTypes[i].propertyFlags));
  980. }
  981. GpuMemoryManager::allocateSingleton();
  982. return Error::kNone;
  983. }
  984. #if ANKI_GR_MANAGER_DEBUG_MEMMORY
  985. void* GrManagerImpl::allocateCallback(void* userData, size_t size, size_t alignment, VkSystemAllocationScope allocationScope)
  986. {
  987. if(size == 0) [[unlikely]]
  988. {
  989. return nullptr;
  990. }
  991. ANKI_ASSERT(userData);
  992. ANKI_ASSERT(size);
  993. ANKI_ASSERT(isPowerOfTwo(alignment));
  994. ANKI_ASSERT(alignment <= MAX_ALLOC_ALIGNMENT);
  995. auto alloc = static_cast<GrManagerImpl*>(userData)->getAllocator();
  996. PtrSize newSize = size + sizeof(AllocHeader);
  997. AllocHeader* header = static_cast<AllocHeader*>(alloc.getMemoryPool().allocate(newSize, MAX_ALLOC_ALIGNMENT));
  998. header->m_sig = ALLOC_SIG;
  999. header->m_size = size;
  1000. ++header;
  1001. return static_cast<AllocHeader*>(header);
  1002. }
  1003. void* GrManagerImpl::reallocateCallback(void* userData, void* original, size_t size, size_t alignment, VkSystemAllocationScope allocationScope)
  1004. {
  1005. if(original && size == 0)
  1006. {
  1007. freeCallback(userData, original);
  1008. return nullptr;
  1009. }
  1010. void* mem = allocateCallback(userData, size, alignment, allocationScope);
  1011. if(original)
  1012. {
  1013. // Move the data
  1014. AllocHeader* header = static_cast<AllocHeader*>(original);
  1015. --header;
  1016. ANKI_ASSERT(header->m_sig == ALLOC_SIG);
  1017. memcpy(mem, original, header->m_size);
  1018. }
  1019. return mem;
  1020. }
  1021. void GrManagerImpl::freeCallback(void* userData, void* ptr)
  1022. {
  1023. if(ptr)
  1024. {
  1025. ANKI_ASSERT(userData);
  1026. auto alloc = static_cast<GrManagerImpl*>(userData)->getAllocator();
  1027. AllocHeader* header = static_cast<AllocHeader*>(ptr);
  1028. --header;
  1029. ANKI_ASSERT(header->m_sig == ALLOC_SIG);
  1030. alloc.getMemoryPool().free(header);
  1031. }
  1032. }
  1033. #endif
  1034. void GrManagerImpl::beginFrameInternal()
  1035. {
  1036. ANKI_TRACE_FUNCTION();
  1037. LockGuard<Mutex> lock(m_globalMtx);
  1038. // Do that at begin frame, ALWAYS
  1039. ++m_frame;
  1040. PerFrame& frame = m_perFrame[m_frame % m_perFrame.getSize()];
  1041. ANKI_ASSERT(m_frameState == kFrameEnded);
  1042. m_frameState = kFrameStarted;
  1043. // Wait for the oldest frame because we don't want to start the new one too early
  1044. {
  1045. ANKI_TRACE_SCOPED_EVENT(WaitFences);
  1046. for(MicroFencePtr& fence : frame.m_fences)
  1047. {
  1048. if(fence)
  1049. {
  1050. const Bool signaled = fence->clientWait(kMaxSecond);
  1051. if(!signaled)
  1052. {
  1053. ANKI_VK_LOGF("Timeout detected");
  1054. }
  1055. }
  1056. }
  1057. }
  1058. frame.m_fences.destroy();
  1059. frame.m_queueWroteToSwapchainImage = GpuQueueType::kCount;
  1060. // Clear garbage
  1061. deleteObjectsMarkedForDeletion();
  1062. }
  1063. TexturePtr GrManagerImpl::acquireNextPresentableTexture()
  1064. {
  1065. ANKI_TRACE_FUNCTION();
  1066. // Create some objets outside the lock
  1067. Array<Char, 16> name;
  1068. snprintf(name.getBegin(), name.getSize(), "Acquire %" PRIu64, m_frame);
  1069. MicroFencePtr fence = FenceFactory::getSingleton().newInstance(name.getBegin());
  1070. LockGuard<Mutex> lock(m_globalMtx);
  1071. ANKI_ASSERT(m_frameState == kFrameStarted);
  1072. m_frameState = kPresentableAcquired;
  1073. PerFrame& frame = m_perFrame[m_frame % m_perFrame.getSize()];
  1074. frame.m_fences.emplaceBack(fence);
  1075. // Get new image
  1076. const MicroSemaphore& acquireSemaphore = *m_crntSwapchain->m_acquireSemaphores[m_frame % m_crntSwapchain->m_acquireSemaphores.getSize()];
  1077. uint32_t imageIdx;
  1078. const VkResult res = vkAcquireNextImageKHR(m_device, m_crntSwapchain->m_swapchain, UINT64_MAX, acquireSemaphore.getHandle(),
  1079. fence->getImplementation().m_handle, &imageIdx);
  1080. if(res == VK_ERROR_OUT_OF_DATE_KHR)
  1081. {
  1082. ANKI_VK_LOGW("Swapchain is out of date. Will wait for the queue and create a new one");
  1083. for(VkQueue queue : m_queues)
  1084. {
  1085. if(queue)
  1086. {
  1087. vkQueueWaitIdle(queue);
  1088. }
  1089. }
  1090. m_crntSwapchain.reset(nullptr);
  1091. m_crntSwapchain = SwapchainFactory::getSingleton().newInstance();
  1092. const MicroSemaphore& acquireSemaphore = *m_crntSwapchain->m_acquireSemaphores[m_frame % m_crntSwapchain->m_acquireSemaphores.getSize()];
  1093. // Can't fail a second time
  1094. ANKI_VK_CHECKF(vkAcquireNextImageKHR(m_device, m_crntSwapchain->m_swapchain, UINT64_MAX, acquireSemaphore.getHandle(),
  1095. fence->getImplementation().m_handle, &imageIdx));
  1096. }
  1097. else
  1098. {
  1099. ANKI_VK_CHECKF(res);
  1100. }
  1101. ANKI_ASSERT(m_acquiredImageIdx == kMaxU8 && "Tried to acquire multiple times in a frame?");
  1102. m_acquiredImageIdx = U8(imageIdx);
  1103. return TexturePtr(m_crntSwapchain->m_textures[imageIdx].get());
  1104. }
  1105. void GrManagerImpl::endFrameInternal()
  1106. {
  1107. ANKI_TRACE_FUNCTION();
  1108. LockGuard<Mutex> lock(m_globalMtx);
  1109. PerFrame& frame = m_perFrame[m_frame % m_perFrame.getSize()];
  1110. const Bool imageAcquired = m_acquiredImageIdx < kMaxU8;
  1111. // Present
  1112. if(imageAcquired)
  1113. {
  1114. ANKI_ASSERT(m_frameState == kPresentableDrawn && "Acquired an image but didn't draw to it?");
  1115. ANKI_ASSERT(m_acquiredImageIdx < kMaxU8);
  1116. VkResult res;
  1117. VkPresentInfoKHR present = {};
  1118. present.sType = VK_STRUCTURE_TYPE_PRESENT_INFO_KHR;
  1119. present.waitSemaphoreCount = 1;
  1120. present.pWaitSemaphores = &m_crntSwapchain->m_renderSemaphores[m_acquiredImageIdx]->getHandle();
  1121. present.swapchainCount = 1;
  1122. present.pSwapchains = &m_crntSwapchain->m_swapchain;
  1123. const U32 idx = m_acquiredImageIdx;
  1124. present.pImageIndices = &idx;
  1125. present.pResults = &res;
  1126. const VkResult res1 = vkQueuePresentKHR(m_queues[frame.m_queueWroteToSwapchainImage], &present);
  1127. if(res1 == VK_ERROR_OUT_OF_DATE_KHR)
  1128. {
  1129. ANKI_VK_LOGW("Swapchain is out of date. Will wait for the queues and create a new one");
  1130. for(VkQueue queue : m_queues)
  1131. {
  1132. if(queue)
  1133. {
  1134. vkQueueWaitIdle(queue);
  1135. }
  1136. }
  1137. vkDeviceWaitIdle(m_device);
  1138. m_crntSwapchain.reset(nullptr);
  1139. m_crntSwapchain = SwapchainFactory::getSingleton().newInstance();
  1140. }
  1141. else
  1142. {
  1143. ANKI_VK_CHECKF(res1);
  1144. ANKI_VK_CHECKF(res);
  1145. }
  1146. m_acquiredImageIdx = kMaxU8;
  1147. }
  1148. else
  1149. {
  1150. ANKI_ASSERT(m_frameState == kFrameStarted);
  1151. }
  1152. m_frameState = kFrameEnded;
  1153. GpuMemoryManager::getSingleton().updateStats();
  1154. }
  1155. void GrManagerImpl::submitInternal(WeakArray<CommandBuffer*> cmdbs, WeakArray<Fence*> waitFences, FencePtr* signalFence)
  1156. {
  1157. // First thing, create a fence
  1158. MicroFencePtr fence = FenceFactory::getSingleton().newInstance("Submit");
  1159. // Gather command buffers
  1160. GrDynamicArray<VkCommandBuffer> vkCmdbs;
  1161. vkCmdbs.resizeStorage(cmdbs.getSize());
  1162. Bool renderedToDefaultFb = false;
  1163. GpuQueueType queueType = GpuQueueType::kCount;
  1164. for(U32 i = 0; i < cmdbs.getSize(); ++i)
  1165. {
  1166. CommandBufferImpl& cmdb = *static_cast<CommandBufferImpl*>(cmdbs[i]);
  1167. ANKI_ASSERT(cmdb.isFinalized());
  1168. renderedToDefaultFb = renderedToDefaultFb || cmdb.renderedToDefaultFramebuffer();
  1169. #if ANKI_ASSERTIONS_ENABLED
  1170. cmdb.setSubmitted();
  1171. #endif
  1172. const MicroCommandBuffer& mcmdb = *cmdb.getMicroCommandBuffer();
  1173. if(i == 0)
  1174. {
  1175. queueType = mcmdb.getVulkanQueueType();
  1176. }
  1177. else
  1178. {
  1179. ANKI_ASSERT(queueType == mcmdb.getVulkanQueueType() && "All cmdbs should be for the same queue");
  1180. }
  1181. vkCmdbs.emplaceBack(cmdb.getHandle());
  1182. }
  1183. // Gather wait semaphores
  1184. GrDynamicArray<VkSemaphore> waitSemaphores;
  1185. GrDynamicArray<VkPipelineStageFlags> waitStages;
  1186. GrDynamicArray<U64> waitTimelineValues;
  1187. waitSemaphores.resizeStorage(waitFences.getSize());
  1188. waitStages.resizeStorage(waitFences.getSize());
  1189. waitTimelineValues.resizeStorage(waitFences.getSize());
  1190. for(U32 i = 0; i < waitFences.getSize(); ++i)
  1191. {
  1192. FenceImpl& impl = static_cast<FenceImpl&>(*waitFences[i]);
  1193. MicroSemaphore& msem = *impl.m_semaphore;
  1194. ANKI_ASSERT(msem.isTimeline());
  1195. waitSemaphores.emplaceBack(msem.getHandle());
  1196. waitStages.emplaceBack(VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  1197. waitTimelineValues.emplaceBack(msem.getSemaphoreValue());
  1198. }
  1199. // Signal semaphore
  1200. GrDynamicArray<VkSemaphore> signalSemaphores;
  1201. GrDynamicArray<U64> signalTimelineValues;
  1202. if(signalFence)
  1203. {
  1204. FenceImpl* fenceImpl = anki::newInstance<FenceImpl>(GrMemoryPool::getSingleton(), "SignalFence");
  1205. fenceImpl->m_semaphore = SemaphoreFactory::getSingleton().newInstance(true, "SubmitSignal");
  1206. signalFence->reset(fenceImpl);
  1207. signalSemaphores.emplaceBack(fenceImpl->m_semaphore->getHandle());
  1208. signalTimelineValues.emplaceBack(fenceImpl->m_semaphore->getNextSemaphoreValue());
  1209. }
  1210. // Submit
  1211. {
  1212. // Protect the class, the queue and other stuff
  1213. LockGuard<Mutex> lock(m_globalMtx);
  1214. // Do some special stuff for the last command buffer
  1215. GrManagerImpl::PerFrame& frame = m_perFrame[m_frame % m_perFrame.getSize()];
  1216. if(renderedToDefaultFb)
  1217. {
  1218. ANKI_ASSERT(m_frameState == kPresentableAcquired);
  1219. m_frameState = kPresentableDrawn;
  1220. const MicroSemaphore& acquireSemaphore = *m_crntSwapchain->m_acquireSemaphores[m_frame % m_crntSwapchain->m_acquireSemaphores.getSize()];
  1221. // Wait semaphore
  1222. waitSemaphores.emplaceBack(acquireSemaphore.getHandle());
  1223. // That depends on how we use the swapchain img. Be a bit conservative
  1224. waitStages.emplaceBack(VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT);
  1225. // Set something
  1226. waitTimelineValues.emplaceBack(0);
  1227. // Get the semaphore to signal and then wait on present
  1228. const MicroSemaphore& renderSemaphore = *m_crntSwapchain->m_renderSemaphores[m_acquiredImageIdx];
  1229. signalSemaphores.emplaceBack(renderSemaphore.getHandle());
  1230. // Increment the timeline values as well because the spec wants a dummy value even for non-timeline semaphores
  1231. signalTimelineValues.emplaceBack(0);
  1232. frame.m_queueWroteToSwapchainImage = queueType;
  1233. }
  1234. frame.m_fences.emplaceBack(fence);
  1235. // Submit
  1236. VkSubmitInfo submit = {};
  1237. submit.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO;
  1238. submit.waitSemaphoreCount = waitSemaphores.getSize();
  1239. submit.pWaitSemaphores = (waitSemaphores.getSize()) ? waitSemaphores.getBegin() : nullptr;
  1240. submit.signalSemaphoreCount = signalSemaphores.getSize();
  1241. submit.pSignalSemaphores = (signalSemaphores.getSize()) ? signalSemaphores.getBegin() : nullptr;
  1242. submit.pWaitDstStageMask = (waitStages.getSize()) ? waitStages.getBegin() : nullptr;
  1243. submit.commandBufferCount = vkCmdbs.getSize();
  1244. submit.pCommandBuffers = (vkCmdbs.getSize()) ? vkCmdbs.getBegin() : nullptr;
  1245. VkTimelineSemaphoreSubmitInfo timelineInfo = {};
  1246. timelineInfo.sType = VK_STRUCTURE_TYPE_TIMELINE_SEMAPHORE_SUBMIT_INFO;
  1247. timelineInfo.waitSemaphoreValueCount = waitSemaphores.getSize();
  1248. timelineInfo.pWaitSemaphoreValues = (waitSemaphores.getSize()) ? waitTimelineValues.getBegin() : nullptr;
  1249. timelineInfo.signalSemaphoreValueCount = signalTimelineValues.getSize();
  1250. timelineInfo.pSignalSemaphoreValues = (signalTimelineValues.getSize()) ? signalTimelineValues.getBegin() : nullptr;
  1251. appendPNextList(submit, &timelineInfo);
  1252. ANKI_TRACE_SCOPED_EVENT(VkQueueSubmit);
  1253. ANKI_VK_CHECKF(vkQueueSubmit(m_queues[queueType], 1, &submit, fence->getImplementation().m_handle));
  1254. }
  1255. }
  1256. void GrManagerImpl::finishInternal()
  1257. {
  1258. LockGuard<Mutex> lock(m_globalMtx);
  1259. for(VkQueue queue : m_queues)
  1260. {
  1261. if(queue)
  1262. {
  1263. vkQueueWaitIdle(queue);
  1264. }
  1265. }
  1266. for(PerFrame& frame : m_perFrame)
  1267. {
  1268. for(MicroFencePtr& fence : frame.m_fences)
  1269. {
  1270. const Bool signaled = fence->clientWait(kMaxSecond);
  1271. if(!signaled)
  1272. {
  1273. ANKI_VK_LOGF("Timeout detected");
  1274. }
  1275. }
  1276. frame.m_fences.destroy();
  1277. }
  1278. // Since we waited for the GPU do a cleanup as well
  1279. const U64 oldFrame = m_frame;
  1280. for(U32 frame = 0; frame < m_perFrame.getSize(); ++frame)
  1281. {
  1282. m_frame = frame;
  1283. deleteObjectsMarkedForDeletion();
  1284. }
  1285. m_frame = oldFrame;
  1286. }
  1287. void GrManagerImpl::trySetVulkanHandleName(CString name, VkObjectType type, U64 handle) const
  1288. {
  1289. if(name && name.getLength() && !!(m_extensions & VulkanExtensions::kEXT_debug_utils))
  1290. {
  1291. VkDebugUtilsObjectNameInfoEXT info = {};
  1292. info.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT;
  1293. info.objectHandle = handle;
  1294. info.objectType = type;
  1295. info.pObjectName = name.cstr();
  1296. vkSetDebugUtilsObjectNameEXT(m_device, &info);
  1297. }
  1298. }
  1299. void GrManagerImpl::printPipelineShaderInfo(VkPipeline ppline, CString name, U64 hash) const
  1300. {
  1301. if(printPipelineShaderInfoInternal(ppline, name, hash))
  1302. {
  1303. ANKI_VK_LOGE("Ignoring previous errors");
  1304. }
  1305. }
  1306. Error GrManagerImpl::printPipelineShaderInfoInternal(VkPipeline ppline, CString name, U64 hash) const
  1307. {
  1308. if(!!(m_extensions & VulkanExtensions::kKHR_pipeline_executable_properties) && Logger::getSingleton().verbosityEnabled())
  1309. {
  1310. VkPipelineInfoKHR pplineInf = {};
  1311. pplineInf.sType = VK_STRUCTURE_TYPE_PIPELINE_INFO_KHR;
  1312. pplineInf.pipeline = ppline;
  1313. U32 executableCount = 0;
  1314. ANKI_VK_CHECK(vkGetPipelineExecutablePropertiesKHR(m_device, &pplineInf, &executableCount, nullptr));
  1315. GrDynamicArray<VkPipelineExecutablePropertiesKHR> executableProps;
  1316. LockGuard lock(m_shaderStatsMtx); // Lock so that all messages appear together
  1317. ANKI_VK_LOGV("Pipeline info \"%s\" (0x%016" PRIx64 "):", name.cstr(), hash);
  1318. if(executableCount > 0)
  1319. {
  1320. executableProps.resize(executableCount);
  1321. for(VkPipelineExecutablePropertiesKHR& prop : executableProps)
  1322. {
  1323. prop = {};
  1324. prop.sType = VK_STRUCTURE_TYPE_PIPELINE_EXECUTABLE_PROPERTIES_KHR;
  1325. }
  1326. ANKI_VK_CHECK(vkGetPipelineExecutablePropertiesKHR(m_device, &pplineInf, &executableCount, &executableProps[0]));
  1327. }
  1328. else
  1329. {
  1330. ANKI_VK_LOGV("\tNo executable count!!!");
  1331. }
  1332. for(U32 i = 0; i < executableCount; ++i)
  1333. {
  1334. const VkPipelineExecutablePropertiesKHR& p = executableProps[i];
  1335. ANKI_VK_LOGV("\tDescription: %s, stages: 0x%X:", p.description, p.stages);
  1336. // Get stats
  1337. VkPipelineExecutableInfoKHR exeInf = {};
  1338. exeInf.sType = VK_STRUCTURE_TYPE_PIPELINE_EXECUTABLE_INFO_KHR;
  1339. exeInf.executableIndex = i;
  1340. exeInf.pipeline = ppline;
  1341. U32 statCount = 0;
  1342. vkGetPipelineExecutableStatisticsKHR(m_device, &exeInf, &statCount, nullptr);
  1343. GrDynamicArray<VkPipelineExecutableStatisticKHR> stats;
  1344. stats.resize(statCount);
  1345. for(VkPipelineExecutableStatisticKHR& s : stats)
  1346. {
  1347. s = {};
  1348. s.sType = VK_STRUCTURE_TYPE_PIPELINE_EXECUTABLE_STATISTIC_KHR;
  1349. }
  1350. vkGetPipelineExecutableStatisticsKHR(m_device, &exeInf, &statCount, &stats[0]);
  1351. for(U32 s = 0; s < statCount; ++s)
  1352. {
  1353. const VkPipelineExecutableStatisticKHR& ss = stats[s];
  1354. switch(ss.format)
  1355. {
  1356. case VK_PIPELINE_EXECUTABLE_STATISTIC_FORMAT_BOOL32_KHR:
  1357. ANKI_VK_LOGV("\t\t%s: %u", ss.name, ss.value.b32);
  1358. break;
  1359. case VK_PIPELINE_EXECUTABLE_STATISTIC_FORMAT_INT64_KHR:
  1360. ANKI_VK_LOGV("\t\t%s: %" PRId64, ss.name, ss.value.i64);
  1361. break;
  1362. case VK_PIPELINE_EXECUTABLE_STATISTIC_FORMAT_UINT64_KHR:
  1363. ANKI_VK_LOGV("\t\t%s: %" PRIu64, ss.name, ss.value.u64);
  1364. break;
  1365. case VK_PIPELINE_EXECUTABLE_STATISTIC_FORMAT_FLOAT64_KHR:
  1366. ANKI_VK_LOGV("\t\t%s: %f", ss.name, ss.value.f64);
  1367. break;
  1368. default:
  1369. ANKI_ASSERT(0);
  1370. }
  1371. }
  1372. ANKI_VK_LOGV("\t\tSubgroup size: %u", p.subgroupSize);
  1373. }
  1374. }
  1375. return Error::kNone;
  1376. }
  1377. Error GrManagerImpl::initSurface()
  1378. {
  1379. #if ANKI_WINDOWING_SYSTEM_SDL
  1380. if(!SDL_Vulkan_CreateSurface(static_cast<NativeWindowSdl&>(NativeWindow::getSingleton()).m_sdlWindow, m_instance, nullptr, &m_surface))
  1381. {
  1382. ANKI_VK_LOGE("SDL_Vulkan_CreateSurface() failed: %s", SDL_GetError());
  1383. return Error::kFunctionFailed;
  1384. }
  1385. #elif ANKI_WINDOWING_SYSTEM_ANDROID
  1386. VkAndroidSurfaceCreateInfoKHR createInfo = {};
  1387. createInfo.sType = VK_STRUCTURE_TYPE_ANDROID_SURFACE_CREATE_INFO_KHR;
  1388. createInfo.window = static_cast<NativeWindowAndroid&>(NativeWindow::getSingleton()).m_nativeWindowAndroid;
  1389. ANKI_VK_CHECK(vkCreateAndroidSurfaceKHR(m_instance, &createInfo, nullptr, &m_surface));
  1390. #elif ANKI_WINDOWING_SYSTEM_HEADLESS
  1391. VkHeadlessSurfaceCreateInfoEXT createInfo = {};
  1392. createInfo.sType = VK_STRUCTURE_TYPE_HEADLESS_SURFACE_CREATE_INFO_EXT;
  1393. ANKI_VK_CHECK(vkCreateHeadlessSurfaceEXT(m_instance, &createInfo, nullptr, &m_surface));
  1394. #else
  1395. # error Unsupported
  1396. #endif
  1397. m_nativeWindowWidth = NativeWindow::getSingleton().getWidth();
  1398. m_nativeWindowHeight = NativeWindow::getSingleton().getHeight();
  1399. return Error::kNone;
  1400. }
  1401. VkBool32 GrManagerImpl::debugReportCallbackEXT(VkDebugUtilsMessageSeverityFlagBitsEXT messageSeverity,
  1402. [[maybe_unused]] VkDebugUtilsMessageTypeFlagsEXT messageTypes,
  1403. const VkDebugUtilsMessengerCallbackDataEXT* pCallbackData, [[maybe_unused]] void* pUserData)
  1404. {
  1405. #if ANKI_PLATFORM_MOBILE
  1406. if(pCallbackData->messageIdNumber == 101294395)
  1407. {
  1408. // Interface mismatch error. Eg vert shader is writing to varying that is not consumed by frag. Ignore this
  1409. // stupid error because I'm not going to create more shader variants to fix it. Especially when mobile drivers
  1410. // do linking anyway. On desktop just enable the maintenance4 extension
  1411. return false;
  1412. }
  1413. #endif
  1414. if(pCallbackData->messageIdNumber == 20145586 || pCallbackData->messageIdNumber == 979140054)
  1415. {
  1416. // Mismatch of the format of the storage image (or the storage texel buffer) in SPIR-V and the actual VkImage. Ignore it because DXC puts
  1417. // almost random shit as formats
  1418. return false;
  1419. }
  1420. if(pCallbackData->messageIdNumber == 1944932341 || pCallbackData->messageIdNumber == 1303270965)
  1421. {
  1422. // Not sure why I'm getting that
  1423. return false;
  1424. }
  1425. // Get all names of affected objects
  1426. GrString objectNames;
  1427. if(pCallbackData->objectCount)
  1428. {
  1429. for(U32 i = 0; i < pCallbackData->objectCount; ++i)
  1430. {
  1431. const Char* name = pCallbackData->pObjects[i].pObjectName;
  1432. objectNames += (name) ? name : "?";
  1433. if(i < pCallbackData->objectCount - 1)
  1434. {
  1435. objectNames += ", ";
  1436. }
  1437. }
  1438. }
  1439. else
  1440. {
  1441. objectNames = "N/A";
  1442. }
  1443. if(messageSeverity & VK_DEBUG_UTILS_MESSAGE_SEVERITY_ERROR_BIT_EXT)
  1444. {
  1445. ANKI_VK_LOGE("VK debug report: %s. Affected objects: %s", pCallbackData->pMessage, objectNames.cstr());
  1446. }
  1447. else if(messageSeverity & VK_DEBUG_UTILS_MESSAGE_SEVERITY_WARNING_BIT_EXT)
  1448. {
  1449. ANKI_VK_LOGW("VK debug report: %s. Affected objects: %s", pCallbackData->pMessage, objectNames.cstr());
  1450. }
  1451. else
  1452. {
  1453. ANKI_VK_LOGI("VK debug report: %s. Affected objects: %s", pCallbackData->pMessage, objectNames.cstr());
  1454. }
  1455. return false;
  1456. }
  1457. } // end namespace anki