| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635 |
- // Copyright (C) 2009-2023, Panagiotis Christopoulos Charitos and contributors.
- // All rights reserved.
- // Code licensed under the BSD License.
- // http://www.anki3d.org/LICENSE
- #include <AnKi/Gr/RenderGraph.h>
- #include <AnKi/Gr/GrManager.h>
- #include <AnKi/Gr/Texture.h>
- #include <AnKi/Gr/Sampler.h>
- #include <AnKi/Gr/CommandBuffer.h>
- #include <AnKi/Util/Tracer.h>
- #include <AnKi/Util/BitSet.h>
- #include <AnKi/Util/File.h>
- #include <AnKi/Util/StringList.h>
- #include <AnKi/Util/HighRezTimer.h>
- #include <AnKi/Core/Common.h>
- namespace anki {
- #define ANKI_DBG_RENDER_GRAPH 0
- static inline U32 getTextureSurfOrVolCount(const TexturePtr& tex)
- {
- return tex->getMipmapCount() * tex->getLayerCount() * (textureTypeIsCube(tex->getTextureType()) ? 6 : 1);
- }
- /// Contains some extra things for render targets.
- class RenderGraph::RT
- {
- public:
- DynamicArray<TextureUsageBit, MemoryPoolPtrWrapper<StackMemoryPool>> m_surfOrVolUsages;
- DynamicArray<U16, MemoryPoolPtrWrapper<StackMemoryPool>> m_lastBatchThatTransitionedIt;
- TexturePtr m_texture; ///< Hold a reference.
- Bool m_imported;
- RT(StackMemoryPool* pool)
- : m_surfOrVolUsages(pool)
- , m_lastBatchThatTransitionedIt(pool)
- {
- }
- };
- /// Same as RT but for buffers.
- class RenderGraph::BufferRange
- {
- public:
- BufferUsageBit m_usage;
- BufferPtr m_buffer; ///< Hold a reference.
- PtrSize m_offset;
- PtrSize m_range;
- };
- class RenderGraph::AS
- {
- public:
- AccelerationStructureUsageBit m_usage;
- AccelerationStructurePtr m_as; ///< Hold a reference.
- };
- /// Pipeline barrier.
- class RenderGraph::TextureBarrier
- {
- public:
- U32 m_idx;
- TextureUsageBit m_usageBefore;
- TextureUsageBit m_usageAfter;
- TextureSurfaceDescriptor m_surface;
- DepthStencilAspectBit m_dsAspect;
- TextureBarrier(U32 rtIdx, TextureUsageBit usageBefore, TextureUsageBit usageAfter, const TextureSurfaceDescriptor& surf, DepthStencilAspectBit dsAspect)
- : m_idx(rtIdx)
- , m_usageBefore(usageBefore)
- , m_usageAfter(usageAfter)
- , m_surface(surf)
- , m_dsAspect(dsAspect)
- {
- }
- };
- /// Pipeline barrier.
- class RenderGraph::BufferBarrier
- {
- public:
- U32 m_idx;
- BufferUsageBit m_usageBefore;
- BufferUsageBit m_usageAfter;
- BufferBarrier(U32 buffIdx, BufferUsageBit usageBefore, BufferUsageBit usageAfter)
- : m_idx(buffIdx)
- , m_usageBefore(usageBefore)
- , m_usageAfter(usageAfter)
- {
- }
- };
- /// Pipeline barrier.
- class RenderGraph::ASBarrier
- {
- public:
- U32 m_idx;
- AccelerationStructureUsageBit m_usageBefore;
- AccelerationStructureUsageBit m_usageAfter;
- ASBarrier(U32 asIdx, AccelerationStructureUsageBit usageBefore, AccelerationStructureUsageBit usageAfter)
- : m_idx(asIdx)
- , m_usageBefore(usageBefore)
- , m_usageAfter(usageAfter)
- {
- }
- };
- /// Contains some extra things the RenderPassBase cannot hold.
- class RenderGraph::Pass
- {
- public:
- // WARNING!!!!!: Whatever you put here needs manual destruction in RenderGraph::reset()
- DynamicArray<U32, MemoryPoolPtrWrapper<StackMemoryPool>> m_dependsOn;
- DynamicArray<RenderPassDependency::TextureInfo, MemoryPoolPtrWrapper<StackMemoryPool>> m_consumedTextures;
- Function<void(RenderPassWorkContext&), MemoryPoolPtrWrapper<StackMemoryPool>> m_callback;
- class
- {
- public:
- Array<RenderTarget, kMaxColorRenderTargets> m_colorRts;
- RenderTarget m_dsRt;
- TextureView* m_vrsRt = nullptr;
- Array<U32, 4> m_renderArea = {};
- U8 m_colorRtCount = 0;
- U8 m_vrsTexelSizeX = 0;
- U8 m_vrsTexelSizeY = 0;
- Array<TextureViewPtr, kMaxColorRenderTargets + 2> m_refs;
- Bool hasRenderpass() const
- {
- return m_renderArea[3] != 0;
- }
- } m_beginRenderpassInfo;
- BaseString<MemoryPoolPtrWrapper<StackMemoryPool>> m_name;
- U32 m_batchIdx ANKI_DEBUG_CODE(= kMaxU32);
- Bool m_drawsToPresentable = false;
- Pass(StackMemoryPool* pool)
- : m_dependsOn(pool)
- , m_consumedTextures(pool)
- , m_name(pool)
- {
- }
- };
- /// A batch of render passes. These passes can run in parallel.
- /// @warning It's POD. Destructor won't be called.
- class RenderGraph::Batch
- {
- public:
- DynamicArray<U32, MemoryPoolPtrWrapper<StackMemoryPool>> m_passIndices;
- DynamicArray<TextureBarrier, MemoryPoolPtrWrapper<StackMemoryPool>> m_textureBarriersBefore;
- DynamicArray<BufferBarrier, MemoryPoolPtrWrapper<StackMemoryPool>> m_bufferBarriersBefore;
- DynamicArray<ASBarrier, MemoryPoolPtrWrapper<StackMemoryPool>> m_asBarriersBefore;
- Batch(StackMemoryPool* pool)
- : m_passIndices(pool)
- , m_textureBarriersBefore(pool)
- , m_bufferBarriersBefore(pool)
- , m_asBarriersBefore(pool)
- {
- }
- Batch(Batch&& b)
- {
- *this = std::move(b);
- }
- Batch& operator=(Batch&& b)
- {
- m_passIndices = std::move(b.m_passIndices);
- m_textureBarriersBefore = std::move(b.m_textureBarriersBefore);
- m_bufferBarriersBefore = std::move(b.m_bufferBarriersBefore);
- m_asBarriersBefore = std::move(b.m_asBarriersBefore);
- return *this;
- }
- };
- /// The RenderGraph build context.
- class RenderGraph::BakeContext
- {
- public:
- DynamicArray<Pass, MemoryPoolPtrWrapper<StackMemoryPool>> m_passes;
- BitSet<kMaxRenderGraphPasses, U64> m_passIsInBatch{false};
- DynamicArray<Batch, MemoryPoolPtrWrapper<StackMemoryPool>> m_batches;
- DynamicArray<RT, MemoryPoolPtrWrapper<StackMemoryPool>> m_rts;
- DynamicArray<BufferRange, MemoryPoolPtrWrapper<StackMemoryPool>> m_buffers;
- DynamicArray<AS, MemoryPoolPtrWrapper<StackMemoryPool>> m_as;
- Bool m_gatherStatistics = false;
- BakeContext(StackMemoryPool* pool)
- : m_passes(pool)
- , m_batches(pool)
- , m_rts(pool)
- , m_buffers(pool)
- , m_as(pool)
- {
- }
- };
- RenderGraph::RenderGraph(CString name)
- : GrObject(kClassType, name)
- {
- }
- RenderGraph::~RenderGraph()
- {
- ANKI_ASSERT(m_ctx == nullptr);
- }
- RenderGraph* RenderGraph::newInstance()
- {
- return anki::newInstance<RenderGraph>(GrMemoryPool::getSingleton(), "N/A");
- }
- void RenderGraph::reset()
- {
- ANKI_TRACE_SCOPED_EVENT(GrRenderGraphReset);
- if(!m_ctx)
- {
- return;
- }
- if((m_version % kPeriodicCleanupEvery) == 0)
- {
- // Do cleanup
- periodicCleanup();
- }
- // Extract the final usage of the imported RTs and clean all RTs
- for(RT& rt : m_ctx->m_rts)
- {
- if(rt.m_imported)
- {
- const U32 surfOrVolumeCount = getTextureSurfOrVolCount(rt.m_texture);
- // Create a new hash because our hash map dislikes concurent keys.
- const U64 uuid = rt.m_texture->getUuid();
- const U64 hash = computeHash(&uuid, sizeof(uuid));
- auto it = m_importedRenderTargets.find(hash);
- if(it != m_importedRenderTargets.getEnd())
- {
- // Found
- ANKI_ASSERT(it->m_surfOrVolLastUsages.getSize() == surfOrVolumeCount);
- ANKI_ASSERT(rt.m_surfOrVolUsages.getSize() == surfOrVolumeCount);
- }
- else
- {
- // Not found, create
- it = m_importedRenderTargets.emplace(hash);
- it->m_surfOrVolLastUsages.resize(surfOrVolumeCount);
- }
- // Update the usage
- for(U32 surfOrVolIdx = 0; surfOrVolIdx < surfOrVolumeCount; ++surfOrVolIdx)
- {
- it->m_surfOrVolLastUsages[surfOrVolIdx] = rt.m_surfOrVolUsages[surfOrVolIdx];
- }
- }
- rt.m_texture.reset(nullptr);
- }
- for(BufferRange& buff : m_ctx->m_buffers)
- {
- buff.m_buffer.reset(nullptr);
- }
- for(AS& as : m_ctx->m_as)
- {
- as.m_as.reset(nullptr);
- }
- for(auto& it : m_renderTargetCache)
- {
- it.m_texturesInUse = 0;
- }
- for(Pass& p : m_ctx->m_passes)
- {
- p.m_beginRenderpassInfo.m_refs.fill(TextureViewPtr(nullptr));
- p.m_callback.destroy();
- p.m_name.destroy();
- }
- m_ctx = nullptr;
- ++m_version;
- }
- TexturePtr RenderGraph::getOrCreateRenderTarget(const TextureInitInfo& initInf, U64 hash)
- {
- ANKI_ASSERT(hash);
- // Find a cache entry
- RenderTargetCacheEntry* entry = nullptr;
- auto it = m_renderTargetCache.find(hash);
- if(it == m_renderTargetCache.getEnd()) [[unlikely]]
- {
- // Didn't found the entry, create a new one
- auto it2 = m_renderTargetCache.emplace(hash);
- entry = &(*it2);
- }
- else
- {
- entry = &(*it);
- }
- ANKI_ASSERT(entry);
- // Create or pop one tex from the cache
- TexturePtr tex;
- const Bool createNewTex = entry->m_textures.getSize() == entry->m_texturesInUse;
- if(!createNewTex)
- {
- // Pop
- tex = entry->m_textures[entry->m_texturesInUse++];
- }
- else
- {
- // Create it
- tex = GrManager::getSingleton().newTexture(initInf);
- ANKI_ASSERT(entry->m_texturesInUse == entry->m_textures.getSize());
- entry->m_textures.resize(entry->m_textures.getSize() + 1);
- entry->m_textures[entry->m_textures.getSize() - 1] = tex;
- ++entry->m_texturesInUse;
- }
- return tex;
- }
- Bool RenderGraph::overlappingTextureSubresource(const TextureSubresourceInfo& suba, const TextureSubresourceInfo& subb)
- {
- #define ANKI_OVERLAPPING(first, count) ((suba.first < subb.first + subb.count) && (subb.first < suba.first + suba.count))
- const Bool overlappingFaces = ANKI_OVERLAPPING(m_firstFace, m_faceCount);
- const Bool overlappingMips = ANKI_OVERLAPPING(m_firstMipmap, m_mipmapCount);
- const Bool overlappingLayers = ANKI_OVERLAPPING(m_firstLayer, m_layerCount);
- #undef ANKI_OVERLAPPING
- return overlappingFaces && overlappingLayers && overlappingMips;
- }
- Bool RenderGraph::passADependsOnB(const RenderPassDescriptionBase& a, const RenderPassDescriptionBase& b)
- {
- // Render targets
- {
- // Compute the 3 types of dependencies
- const BitSet<kMaxRenderGraphRenderTargets, U64> aReadBWrite = a.m_readRtMask & b.m_writeRtMask;
- const BitSet<kMaxRenderGraphRenderTargets, U64> aWriteBRead = a.m_writeRtMask & b.m_readRtMask;
- const BitSet<kMaxRenderGraphRenderTargets, U64> aWriteBWrite = a.m_writeRtMask & b.m_writeRtMask;
- const BitSet<kMaxRenderGraphRenderTargets, U64> fullDep = aReadBWrite | aWriteBRead | aWriteBWrite;
- if(fullDep.getAnySet())
- {
- // There might be an overlap
- for(const RenderPassDependency& aDep : a.m_rtDeps)
- {
- if(!fullDep.get(aDep.m_texture.m_handle.m_idx))
- {
- continue;
- }
- for(const RenderPassDependency& bDep : b.m_rtDeps)
- {
- if(aDep.m_texture.m_handle != bDep.m_texture.m_handle)
- {
- continue;
- }
- if(!((aDep.m_texture.m_usage | bDep.m_texture.m_usage) & TextureUsageBit::kAllWrite))
- {
- // Don't care about read to read deps
- continue;
- }
- if(overlappingTextureSubresource(aDep.m_texture.m_subresource, bDep.m_texture.m_subresource))
- {
- return true;
- }
- }
- }
- }
- }
- // Buffers
- if(a.m_readBuffMask || a.m_writeBuffMask)
- {
- const BitSet<kMaxRenderGraphBuffers, U64> aReadBWrite = a.m_readBuffMask & b.m_writeBuffMask;
- const BitSet<kMaxRenderGraphBuffers, U64> aWriteBRead = a.m_writeBuffMask & b.m_readBuffMask;
- const BitSet<kMaxRenderGraphBuffers, U64> aWriteBWrite = a.m_writeBuffMask & b.m_writeBuffMask;
- const BitSet<kMaxRenderGraphBuffers, U64> fullDep = aReadBWrite | aWriteBRead | aWriteBWrite;
- if(fullDep.getAnySet())
- {
- // There might be an overlap
- for(const RenderPassDependency& aDep : a.m_buffDeps)
- {
- if(!fullDep.get(aDep.m_buffer.m_handle.m_idx))
- {
- continue;
- }
- for(const RenderPassDependency& bDep : b.m_buffDeps)
- {
- if(aDep.m_buffer.m_handle != bDep.m_buffer.m_handle)
- {
- continue;
- }
- if(!((aDep.m_buffer.m_usage | bDep.m_buffer.m_usage) & BufferUsageBit::kAllWrite))
- {
- // Don't care about read to read deps
- continue;
- }
- // TODO: Take into account the ranges
- return true;
- }
- }
- }
- }
- // AS
- if(a.m_readAsMask || a.m_writeAsMask)
- {
- const BitSet<kMaxRenderGraphAccelerationStructures, U32> aReadBWrite = a.m_readAsMask & b.m_writeAsMask;
- const BitSet<kMaxRenderGraphAccelerationStructures, U32> aWriteBRead = a.m_writeAsMask & b.m_readAsMask;
- const BitSet<kMaxRenderGraphAccelerationStructures, U32> aWriteBWrite = a.m_writeAsMask & b.m_writeAsMask;
- const BitSet<kMaxRenderGraphAccelerationStructures, U32> fullDep = aReadBWrite | aWriteBRead | aWriteBWrite;
- if(fullDep)
- {
- for(const RenderPassDependency& aDep : a.m_asDeps)
- {
- if(!fullDep.get(aDep.m_as.m_handle.m_idx))
- {
- continue;
- }
- for(const RenderPassDependency& bDep : b.m_asDeps)
- {
- if(aDep.m_as.m_handle != bDep.m_as.m_handle)
- {
- continue;
- }
- if(!((aDep.m_as.m_usage | bDep.m_as.m_usage) & AccelerationStructureUsageBit::kAllWrite))
- {
- // Don't care about read to read deps
- continue;
- }
- return true;
- }
- }
- }
- }
- return false;
- }
- Bool RenderGraph::passHasUnmetDependencies(const BakeContext& ctx, U32 passIdx)
- {
- Bool depends = false;
- if(ctx.m_batches.getSize() > 0)
- {
- // Check if the deps of passIdx are all in a batch
- for(const U32 depPassIdx : ctx.m_passes[passIdx].m_dependsOn)
- {
- if(!ctx.m_passIsInBatch.get(depPassIdx))
- {
- // Dependency pass is not in a batch
- depends = true;
- break;
- }
- }
- }
- else
- {
- // First batch, check if passIdx depends on any pass
- depends = ctx.m_passes[passIdx].m_dependsOn.getSize() != 0;
- }
- return depends;
- }
- RenderGraph::BakeContext* RenderGraph::newContext(const RenderGraphDescription& descr, StackMemoryPool& pool)
- {
- // Allocate
- BakeContext* ctx = anki::newInstance<BakeContext>(pool, &pool);
- // Init the resources
- ctx->m_rts.resizeStorage(descr.m_renderTargets.getSize());
- for(U32 rtIdx = 0; rtIdx < descr.m_renderTargets.getSize(); ++rtIdx)
- {
- RT& outRt = *ctx->m_rts.emplaceBack(&pool);
- const RenderGraphDescription::RT& inRt = descr.m_renderTargets[rtIdx];
- const Bool imported = inRt.m_importedTex.isCreated();
- if(imported)
- {
- // It's imported
- outRt.m_texture = inRt.m_importedTex;
- }
- else
- {
- // Need to create new
- // Create a new TextureInitInfo with the derived usage
- TextureInitInfo initInf = inRt.m_initInfo;
- initInf.m_usage = inRt.m_usageDerivedByDeps;
- ANKI_ASSERT(initInf.m_usage != TextureUsageBit::kNone && "Probably not referenced by any pass");
- // Create the new hash
- const U64 hash = appendHash(&initInf.m_usage, sizeof(initInf.m_usage), inRt.m_hash);
- // Get or create the texture
- outRt.m_texture = getOrCreateRenderTarget(initInf, hash);
- }
- // Init the usage
- const U32 surfOrVolumeCount = getTextureSurfOrVolCount(outRt.m_texture);
- outRt.m_surfOrVolUsages.resize(surfOrVolumeCount, TextureUsageBit::kNone);
- if(imported && inRt.m_importedAndUndefinedUsage)
- {
- // Get the usage from previous frames
- // Create a new hash because our hash map dislikes concurent keys.
- const U64 uuid = outRt.m_texture->getUuid();
- const U64 hash = computeHash(&uuid, sizeof(uuid));
- auto it = m_importedRenderTargets.find(hash);
- ANKI_ASSERT(it != m_importedRenderTargets.getEnd() && "Can't find the imported RT");
- ANKI_ASSERT(it->m_surfOrVolLastUsages.getSize() == surfOrVolumeCount);
- for(U32 surfOrVolIdx = 0; surfOrVolIdx < surfOrVolumeCount; ++surfOrVolIdx)
- {
- outRt.m_surfOrVolUsages[surfOrVolIdx] = it->m_surfOrVolLastUsages[surfOrVolIdx];
- }
- }
- else if(imported)
- {
- // Set the usage that was given by the user
- for(U32 surfOrVolIdx = 0; surfOrVolIdx < surfOrVolumeCount; ++surfOrVolIdx)
- {
- outRt.m_surfOrVolUsages[surfOrVolIdx] = inRt.m_importedLastKnownUsage;
- }
- }
- outRt.m_lastBatchThatTransitionedIt.resize(surfOrVolumeCount, kMaxU16);
- outRt.m_imported = imported;
- }
- // Buffers
- ctx->m_buffers.resize(descr.m_buffers.getSize());
- for(U32 buffIdx = 0; buffIdx < ctx->m_buffers.getSize(); ++buffIdx)
- {
- ctx->m_buffers[buffIdx].m_usage = descr.m_buffers[buffIdx].m_usage;
- ANKI_ASSERT(descr.m_buffers[buffIdx].m_importedBuff.isCreated());
- ctx->m_buffers[buffIdx].m_buffer = descr.m_buffers[buffIdx].m_importedBuff;
- ctx->m_buffers[buffIdx].m_offset = descr.m_buffers[buffIdx].m_offset;
- ctx->m_buffers[buffIdx].m_range = descr.m_buffers[buffIdx].m_range;
- }
- // AS
- ctx->m_as.resize(descr.m_as.getSize());
- for(U32 i = 0; i < descr.m_as.getSize(); ++i)
- {
- ctx->m_as[i].m_usage = descr.m_as[i].m_usage;
- ctx->m_as[i].m_as = descr.m_as[i].m_importedAs;
- ANKI_ASSERT(ctx->m_as[i].m_as.isCreated());
- }
- ctx->m_gatherStatistics = descr.m_gatherStatistics;
- return ctx;
- }
- void RenderGraph::initRenderPassesAndSetDeps(const RenderGraphDescription& descr)
- {
- BakeContext& ctx = *m_ctx;
- const U32 passCount = descr.m_passes.getSize();
- ANKI_ASSERT(passCount > 0);
- ctx.m_passes.resizeStorage(passCount);
- for(U32 passIdx = 0; passIdx < passCount; ++passIdx)
- {
- const RenderPassDescriptionBase& inPass = *descr.m_passes[passIdx];
- Pass& outPass = *ctx.m_passes.emplaceBack(ctx.m_as.getMemoryPool().m_pool);
- outPass.m_callback = inPass.m_callback;
- outPass.m_name = inPass.m_name;
- // Create consumer info
- outPass.m_consumedTextures.resize(inPass.m_rtDeps.getSize());
- for(U32 depIdx = 0; depIdx < inPass.m_rtDeps.getSize(); ++depIdx)
- {
- const RenderPassDependency& inDep = inPass.m_rtDeps[depIdx];
- ANKI_ASSERT(inDep.m_type == RenderPassDependency::Type::kTexture);
- RenderPassDependency::TextureInfo& inf = outPass.m_consumedTextures[depIdx];
- ANKI_ASSERT(sizeof(inf) == sizeof(inDep.m_texture));
- memcpy(&inf, &inDep.m_texture, sizeof(inf));
- }
- // Set dependencies by checking all previous subpasses.
- U32 prevPassIdx = passIdx;
- while(prevPassIdx--)
- {
- const RenderPassDescriptionBase& prevPass = *descr.m_passes[prevPassIdx];
- if(passADependsOnB(inPass, prevPass))
- {
- outPass.m_dependsOn.emplaceBack(prevPassIdx);
- }
- }
- }
- }
- void RenderGraph::initBatches()
- {
- ANKI_ASSERT(m_ctx);
- U passesAssignedToBatchCount = 0;
- const U passCount = m_ctx->m_passes.getSize();
- ANKI_ASSERT(passCount > 0);
- while(passesAssignedToBatchCount < passCount)
- {
- Batch batch(m_ctx->m_as.getMemoryPool().m_pool);
- for(U32 i = 0; i < passCount; ++i)
- {
- if(!m_ctx->m_passIsInBatch.get(i) && !passHasUnmetDependencies(*m_ctx, i))
- {
- // Add to the batch
- ++passesAssignedToBatchCount;
- batch.m_passIndices.emplaceBack(i);
- }
- }
- // Mark batch's passes done
- for(U32 passIdx : batch.m_passIndices)
- {
- m_ctx->m_passIsInBatch.set(passIdx);
- m_ctx->m_passes[passIdx].m_batchIdx = m_ctx->m_batches.getSize();
- }
- m_ctx->m_batches.emplaceBack(std::move(batch));
- }
- }
- void RenderGraph::initGraphicsPasses(const RenderGraphDescription& descr)
- {
- BakeContext& ctx = *m_ctx;
- const U32 passCount = descr.m_passes.getSize();
- ANKI_ASSERT(passCount > 0);
- for(U32 passIdx = 0; passIdx < passCount; ++passIdx)
- {
- const RenderPassDescriptionBase& baseInPass = *descr.m_passes[passIdx];
- Pass& outPass = ctx.m_passes[passIdx];
- // Create command buffers and framebuffer
- if(baseInPass.m_type == RenderPassDescriptionBase::Type::kGraphics)
- {
- const GraphicsRenderPassDescription& inPass = static_cast<const GraphicsRenderPassDescription&>(baseInPass);
- if(inPass.hasRenderpass())
- {
- outPass.m_beginRenderpassInfo.m_renderArea = inPass.m_rpassRenderArea;
- outPass.m_beginRenderpassInfo.m_colorRtCount = inPass.m_colorRtCount;
- // Init the usage bits
- for(U32 i = 0; i < inPass.m_colorRtCount; ++i)
- {
- const RenderTargetInfo& inAttachment = inPass.m_rts[i];
- RenderTarget& outAttachment = outPass.m_beginRenderpassInfo.m_colorRts[i];
- getCrntUsage(inAttachment.m_handle, outPass.m_batchIdx, TextureSubresourceInfo(inAttachment.m_surface), outAttachment.m_usage);
- const TextureViewInitInfo viewInit(m_ctx->m_rts[inAttachment.m_handle.m_idx].m_texture.get(),
- TextureSubresourceInfo(inAttachment.m_surface), "RenderGraph");
- TextureViewPtr view = GrManager::getSingleton().newTextureView(viewInit);
- outAttachment.m_view = view.get();
- outPass.m_beginRenderpassInfo.m_refs[i] = view;
- outAttachment.m_loadOperation = inAttachment.m_loadOperation;
- outAttachment.m_storeOperation = inAttachment.m_storeOperation;
- outAttachment.m_clearValue = inAttachment.m_clearValue;
- }
- if(!!inPass.m_rts[kMaxColorRenderTargets].m_aspect)
- {
- const RenderTargetInfo& inAttachment = inPass.m_rts[kMaxColorRenderTargets];
- RenderTarget& outAttachment = outPass.m_beginRenderpassInfo.m_dsRt;
- const TextureSubresourceInfo subresource = TextureSubresourceInfo(inAttachment.m_surface, inAttachment.m_aspect);
- getCrntUsage(inAttachment.m_handle, outPass.m_batchIdx, subresource, outAttachment.m_usage);
- const TextureViewInitInfo viewInit(m_ctx->m_rts[inAttachment.m_handle.m_idx].m_texture.get(),
- TextureSubresourceInfo(inAttachment.m_surface, inAttachment.m_aspect), "RenderGraph");
- TextureViewPtr view = GrManager::getSingleton().newTextureView(viewInit);
- outAttachment.m_view = view.get();
- outPass.m_beginRenderpassInfo.m_refs[kMaxColorRenderTargets] = view;
- outAttachment.m_loadOperation = inAttachment.m_loadOperation;
- outAttachment.m_storeOperation = inAttachment.m_storeOperation;
- outAttachment.m_stencilLoadOperation = inAttachment.m_stencilLoadOperation;
- outAttachment.m_stencilStoreOperation = inAttachment.m_stencilStoreOperation;
- outAttachment.m_clearValue = inAttachment.m_clearValue;
- outAttachment.m_aspect = inAttachment.m_aspect;
- }
- if(inPass.m_vrsRtTexelSizeX > 0)
- {
- const RenderTargetInfo& inAttachment = inPass.m_rts[kMaxColorRenderTargets + 1];
- const TextureViewInitInfo viewInit(m_ctx->m_rts[inAttachment.m_handle.m_idx].m_texture.get(), inAttachment.m_surface,
- "RenderGraph SRI");
- TextureViewPtr view = GrManager::getSingleton().newTextureView(viewInit);
- outPass.m_beginRenderpassInfo.m_vrsRt = view.get();
- outPass.m_beginRenderpassInfo.m_refs[kMaxColorRenderTargets + 1] = view;
- outPass.m_beginRenderpassInfo.m_vrsTexelSizeX = inPass.m_vrsRtTexelSizeX;
- outPass.m_beginRenderpassInfo.m_vrsTexelSizeY = inPass.m_vrsRtTexelSizeY;
- }
- }
- }
- }
- }
- template<typename TFunc>
- void RenderGraph::iterateSurfsOrVolumes(const Texture& tex, const TextureSubresourceInfo& subresource, TFunc func)
- {
- for(U32 mip = subresource.m_firstMipmap; mip < subresource.m_firstMipmap + subresource.m_mipmapCount; ++mip)
- {
- for(U32 layer = subresource.m_firstLayer; layer < subresource.m_firstLayer + subresource.m_layerCount; ++layer)
- {
- for(U32 face = subresource.m_firstFace; face < U32(subresource.m_firstFace + subresource.m_faceCount); ++face)
- {
- // Compute surf or vol idx
- const U32 faceCount = textureTypeIsCube(tex.getTextureType()) ? 6 : 1;
- const U32 idx = (faceCount * tex.getLayerCount()) * mip + faceCount * layer + face;
- const TextureSurfaceDescriptor surf(mip, face, layer);
- if(!func(idx, surf))
- {
- return;
- }
- }
- }
- }
- }
- void RenderGraph::setTextureBarrier(Batch& batch, const RenderPassDependency& dep)
- {
- ANKI_ASSERT(dep.m_type == RenderPassDependency::Type::kTexture);
- BakeContext& ctx = *m_ctx;
- const U32 batchIdx = U32(&batch - &ctx.m_batches[0]);
- const U32 rtIdx = dep.m_texture.m_handle.m_idx;
- const TextureUsageBit depUsage = dep.m_texture.m_usage;
- RT& rt = ctx.m_rts[rtIdx];
- iterateSurfsOrVolumes(*rt.m_texture, dep.m_texture.m_subresource, [&](U32 surfOrVolIdx, const TextureSurfaceDescriptor& surf) {
- TextureUsageBit& crntUsage = rt.m_surfOrVolUsages[surfOrVolIdx];
- const Bool skipBarrier = crntUsage == depUsage && !(crntUsage & TextureUsageBit::kAllWrite);
- if(!skipBarrier)
- {
- // Check if we can merge barriers
- if(rt.m_lastBatchThatTransitionedIt[surfOrVolIdx] == batchIdx)
- {
- // Will merge the barriers
- crntUsage |= depUsage;
- [[maybe_unused]] Bool found = false;
- for(TextureBarrier& b : batch.m_textureBarriersBefore)
- {
- if(b.m_idx == rtIdx && b.m_surface == surf)
- {
- b.m_usageAfter |= depUsage;
- found = true;
- break;
- }
- }
- ANKI_ASSERT(found);
- }
- else
- {
- // Create a new barrier for this surface
- batch.m_textureBarriersBefore.emplaceBack(rtIdx, crntUsage, depUsage, surf, dep.m_texture.m_subresource.m_depthStencilAspect);
- crntUsage = depUsage;
- rt.m_lastBatchThatTransitionedIt[surfOrVolIdx] = U16(batchIdx);
- }
- }
- return true;
- });
- }
- void RenderGraph::setBatchBarriers(const RenderGraphDescription& descr)
- {
- BakeContext& ctx = *m_ctx;
- // For all batches
- for(Batch& batch : ctx.m_batches)
- {
- BitSet<kMaxRenderGraphBuffers, U64> buffHasBarrierMask(false);
- BitSet<kMaxRenderGraphAccelerationStructures, U32> asHasBarrierMask(false);
- // For all passes of that batch
- for(U32 passIdx : batch.m_passIndices)
- {
- const RenderPassDescriptionBase& pass = *descr.m_passes[passIdx];
- // Do textures
- for(const RenderPassDependency& dep : pass.m_rtDeps)
- {
- setTextureBarrier(batch, dep);
- }
- // Do buffers
- for(const RenderPassDependency& dep : pass.m_buffDeps)
- {
- const U32 buffIdx = dep.m_buffer.m_handle.m_idx;
- const BufferUsageBit depUsage = dep.m_buffer.m_usage;
- BufferUsageBit& crntUsage = ctx.m_buffers[buffIdx].m_usage;
- const Bool skipBarrier = crntUsage == depUsage && !(crntUsage & BufferUsageBit::kAllWrite);
- if(skipBarrier)
- {
- continue;
- }
- const Bool buffHasBarrier = buffHasBarrierMask.get(buffIdx);
- if(!buffHasBarrier)
- {
- // Buff hasn't had a barrier in this batch, add a new barrier
- batch.m_bufferBarriersBefore.emplaceBack(buffIdx, crntUsage, depUsage);
- crntUsage = depUsage;
- buffHasBarrierMask.set(buffIdx);
- }
- else
- {
- // Buff already in a barrier, merge the 2 barriers
- BufferBarrier* barrierToMergeTo = nullptr;
- for(BufferBarrier& b : batch.m_bufferBarriersBefore)
- {
- if(b.m_idx == buffIdx)
- {
- barrierToMergeTo = &b;
- break;
- }
- }
- ANKI_ASSERT(barrierToMergeTo);
- ANKI_ASSERT(!!barrierToMergeTo->m_usageAfter);
- barrierToMergeTo->m_usageAfter |= depUsage;
- crntUsage = barrierToMergeTo->m_usageAfter;
- }
- }
- // Do AS
- for(const RenderPassDependency& dep : pass.m_asDeps)
- {
- const U32 asIdx = dep.m_as.m_handle.m_idx;
- const AccelerationStructureUsageBit depUsage = dep.m_as.m_usage;
- AccelerationStructureUsageBit& crntUsage = ctx.m_as[asIdx].m_usage;
- const Bool skipBarrier = crntUsage == depUsage && !(crntUsage & AccelerationStructureUsageBit::kAllWrite);
- if(skipBarrier)
- {
- continue;
- }
- const Bool asHasBarrierInThisBatch = asHasBarrierMask.get(asIdx);
- if(!asHasBarrierInThisBatch)
- {
- // AS doesn't have a barrier in this batch, create a new one
- batch.m_asBarriersBefore.emplaceBack(asIdx, crntUsage, depUsage);
- crntUsage = depUsage;
- asHasBarrierMask.set(asIdx);
- }
- else
- {
- // AS already has a barrier, merge the 2 barriers
- ASBarrier* barrierToMergeTo = nullptr;
- for(ASBarrier& other : batch.m_asBarriersBefore)
- {
- if(other.m_idx == asIdx)
- {
- barrierToMergeTo = &other;
- break;
- }
- }
- ANKI_ASSERT(barrierToMergeTo);
- ANKI_ASSERT(!!barrierToMergeTo->m_usageAfter);
- barrierToMergeTo->m_usageAfter |= depUsage;
- crntUsage = barrierToMergeTo->m_usageAfter;
- }
- }
- } // For all passes
- ANKI_ASSERT(batch.m_bufferBarriersBefore.getSize() || batch.m_textureBarriersBefore.getSize() || batch.m_asBarriersBefore.getSize());
- #if ANKI_DBG_RENDER_GRAPH
- // Sort the barriers to ease the dumped graph
- std::sort(batch.m_textureBarriersBefore.getBegin(), batch.m_textureBarriersBefore.getEnd(),
- [&](const TextureBarrier& a, const TextureBarrier& b) {
- const U aidx = a.m_idx;
- const U bidx = b.m_idx;
- if(aidx == bidx)
- {
- if(a.m_surface.m_level != b.m_surface.m_level)
- {
- return a.m_surface.m_level < b.m_surface.m_level;
- }
- else if(a.m_surface.m_face != b.m_surface.m_face)
- {
- return a.m_surface.m_face < b.m_surface.m_face;
- }
- else if(a.m_surface.m_layer != b.m_surface.m_layer)
- {
- return a.m_surface.m_layer < b.m_surface.m_layer;
- }
- else
- {
- return false;
- }
- }
- else
- {
- return aidx < bidx;
- }
- });
- std::sort(batch.m_bufferBarriersBefore.getBegin(), batch.m_bufferBarriersBefore.getEnd(),
- [&](const BufferBarrier& a, const BufferBarrier& b) {
- return a.m_idx < b.m_idx;
- });
- std::sort(batch.m_asBarriersBefore.getBegin(), batch.m_asBarriersBefore.getEnd(), [&](const ASBarrier& a, const ASBarrier& b) {
- return a.m_idx < b.m_idx;
- });
- #endif
- } // For all batches
- }
- void RenderGraph::minimizeSubchannelSwitches()
- {
- BakeContext& ctx = *m_ctx;
- Bool computeFirst = true;
- for(Batch& batch : ctx.m_batches)
- {
- U32 graphicsPasses = 0;
- U32 computePasses = 0;
- std::sort(batch.m_passIndices.getBegin(), batch.m_passIndices.getEnd(), [&](U32 a, U32 b) {
- const Bool aIsCompute = !ctx.m_passes[a].m_beginRenderpassInfo.hasRenderpass();
- const Bool bIsCompute = !ctx.m_passes[b].m_beginRenderpassInfo.hasRenderpass();
- graphicsPasses += !aIsCompute + !bIsCompute;
- computePasses += aIsCompute + bIsCompute;
- if(computeFirst)
- {
- return !aIsCompute < !bIsCompute;
- }
- else
- {
- return aIsCompute < bIsCompute;
- }
- });
- if(graphicsPasses && !computePasses)
- {
- // Only graphics passes in this batch, start next batch from graphics
- computeFirst = false;
- }
- else if(computePasses && !graphicsPasses)
- {
- // Only compute passes in this batch, start next batch from compute
- computeFirst = true;
- }
- else
- {
- // This batch ends in compute start next batch in compute and if it ends with graphics start next in graphics
- computeFirst = !computeFirst;
- }
- }
- }
- void RenderGraph::sortBatchPasses()
- {
- BakeContext& ctx = *m_ctx;
- for(Batch& batch : ctx.m_batches)
- {
- std::sort(batch.m_passIndices.getBegin(), batch.m_passIndices.getEnd(), [&](U32 a, U32 b) {
- const Bool aIsCompute = !ctx.m_passes[a].m_beginRenderpassInfo.hasRenderpass();
- const Bool bIsCompute = !ctx.m_passes[b].m_beginRenderpassInfo.hasRenderpass();
- return aIsCompute < bIsCompute;
- });
- }
- }
- void RenderGraph::compileNewGraph(const RenderGraphDescription& descr, StackMemoryPool& pool)
- {
- ANKI_TRACE_SCOPED_EVENT(GrRenderGraphCompile);
- // Init the context
- BakeContext& ctx = *newContext(descr, pool);
- m_ctx = &ctx;
- // Init the passes and find the dependencies between passes
- initRenderPassesAndSetDeps(descr);
- // Walk the graph and create pass batches
- initBatches();
- // Now that we know the batches every pass belongs init the graphics passes
- initGraphicsPasses(descr);
- // Create barriers between batches
- setBatchBarriers(descr);
- // Sort passes in batches
- if(GrManager::getSingleton().getDeviceCapabilities().m_gpuVendor == GpuVendor::kNvidia)
- {
- minimizeSubchannelSwitches();
- }
- else
- {
- sortBatchPasses();
- }
- #if ANKI_DBG_RENDER_GRAPH
- if(dumpDependencyDotFile(descr, ctx, "./"))
- {
- ANKI_LOGF("Won't recover on debug code");
- }
- #endif
- }
- Texture& RenderGraph::getTexture(RenderTargetHandle handle) const
- {
- ANKI_ASSERT(m_ctx->m_rts[handle.m_idx].m_texture.isCreated());
- return *m_ctx->m_rts[handle.m_idx].m_texture;
- }
- void RenderGraph::getCachedBuffer(BufferHandle handle, Buffer*& buff, PtrSize& offset, PtrSize& range) const
- {
- const BufferRange& record = m_ctx->m_buffers[handle.m_idx];
- buff = record.m_buffer.get();
- offset = record.m_offset;
- range = record.m_range;
- }
- AccelerationStructure* RenderGraph::getAs(AccelerationStructureHandle handle) const
- {
- ANKI_ASSERT(m_ctx->m_as[handle.m_idx].m_as.isCreated());
- return m_ctx->m_as[handle.m_idx].m_as.get();
- }
- void RenderGraph::recordAndSubmitCommandBuffers(FencePtr* optionalFence)
- {
- ANKI_TRACE_SCOPED_EVENT(GrRenderGraphRecordAndSubmit);
- ANKI_ASSERT(m_ctx);
- const U32 batchGroupCount = min(CoreThreadJobManager::getSingleton().getThreadCount(), m_ctx->m_batches.getSize());
- StackMemoryPool* pool = m_ctx->m_rts.getMemoryPool().m_pool;
- DynamicArray<CommandBufferPtr, MemoryPoolPtrWrapper<StackMemoryPool>> cmdbs(pool);
- cmdbs.resize(batchGroupCount);
- SpinLock cmdbsMtx;
- for(U32 group = 0; group < batchGroupCount; ++group)
- {
- U32 start, end;
- splitThreadedProblem(group, batchGroupCount, m_ctx->m_batches.getSize(), start, end);
- if(start == end)
- {
- continue;
- }
- CoreThreadJobManager::getSingleton().dispatchTask(
- [this, start, end, pool, &cmdbs, &cmdbsMtx, group, batchGroupCount]([[maybe_unused]] U32 tid) {
- ANKI_TRACE_SCOPED_EVENT(GrRenderGraphTask);
- CommandBufferInitInfo cmdbInit("RenderGraph cmdb");
- cmdbInit.m_flags = CommandBufferFlag::kGeneralWork;
- CommandBufferPtr cmdb = GrManager::getSingleton().newCommandBuffer(cmdbInit);
- // Write timestamp
- const Bool setPreQuery = m_ctx->m_gatherStatistics && group == 0;
- const Bool setPostQuery = m_ctx->m_gatherStatistics && group == batchGroupCount - 1;
- TimestampQueryPtr preQuery, postQuery;
- if(setPreQuery)
- {
- preQuery = GrManager::getSingleton().newTimestampQuery();
- cmdb->writeTimestamp(preQuery.get());
- }
- if(setPostQuery)
- {
- postQuery = GrManager::getSingleton().newTimestampQuery();
- }
- // Bookkeeping
- {
- LockGuard lock(cmdbsMtx);
- cmdbs[group] = cmdb;
- if(preQuery.isCreated())
- {
- m_statistics.m_nextTimestamp = (m_statistics.m_nextTimestamp + 1) % kMaxBufferedTimestamps;
- m_statistics.m_timestamps[m_statistics.m_nextTimestamp * 2] = preQuery;
- }
- if(postQuery.isCreated())
- {
- m_statistics.m_timestamps[m_statistics.m_nextTimestamp * 2 + 1] = postQuery;
- m_statistics.m_cpuStartTimes[m_statistics.m_nextTimestamp] = HighRezTimer::getCurrentTime();
- }
- }
- RenderPassWorkContext ctx;
- ctx.m_rgraph = this;
- for(U32 i = start; i < end; ++i)
- {
- const Batch& batch = m_ctx->m_batches[i];
- // Set the barriers
- DynamicArray<TextureBarrierInfo, MemoryPoolPtrWrapper<StackMemoryPool>> texBarriers(pool);
- texBarriers.resizeStorage(batch.m_textureBarriersBefore.getSize());
- for(const TextureBarrier& barrier : batch.m_textureBarriersBefore)
- {
- TextureBarrierInfo& inf = *texBarriers.emplaceBack();
- inf.m_previousUsage = barrier.m_usageBefore;
- inf.m_nextUsage = barrier.m_usageAfter;
- inf.m_subresource = barrier.m_surface;
- inf.m_subresource.m_depthStencilAspect = barrier.m_dsAspect;
- inf.m_texture = m_ctx->m_rts[barrier.m_idx].m_texture.get();
- }
- DynamicArray<BufferBarrierInfo, MemoryPoolPtrWrapper<StackMemoryPool>> buffBarriers(pool);
- buffBarriers.resizeStorage(batch.m_bufferBarriersBefore.getSize());
- for(const BufferBarrier& barrier : batch.m_bufferBarriersBefore)
- {
- BufferBarrierInfo& inf = *buffBarriers.emplaceBack();
- inf.m_previousUsage = barrier.m_usageBefore;
- inf.m_nextUsage = barrier.m_usageAfter;
- inf.m_bufferView = BufferView(m_ctx->m_buffers[barrier.m_idx].m_buffer.get(), m_ctx->m_buffers[barrier.m_idx].m_offset,
- m_ctx->m_buffers[barrier.m_idx].m_range);
- }
- DynamicArray<AccelerationStructureBarrierInfo, MemoryPoolPtrWrapper<StackMemoryPool>> asBarriers(pool);
- for(const ASBarrier& barrier : batch.m_asBarriersBefore)
- {
- AccelerationStructureBarrierInfo& inf = *asBarriers.emplaceBack();
- inf.m_previousUsage = barrier.m_usageBefore;
- inf.m_nextUsage = barrier.m_usageAfter;
- inf.m_as = m_ctx->m_as[barrier.m_idx].m_as.get();
- }
- cmdb->pushDebugMarker("Barrier", Vec3(1.0f, 0.0f, 0.0f));
- cmdb->setPipelineBarrier(texBarriers, buffBarriers, asBarriers);
- cmdb->popDebugMarker();
- ctx.m_commandBuffer = cmdb.get();
- ctx.m_batchIdx = i;
- // Call the passes
- for(U32 passIdx : batch.m_passIndices)
- {
- Pass& pass = m_ctx->m_passes[passIdx];
- const Vec3 passColor = (pass.m_beginRenderpassInfo.hasRenderpass()) ? Vec3(0.0f, 1.0f, 0.0f) : Vec3(1.0f, 1.0f, 0.0f);
- cmdb->pushDebugMarker(pass.m_name, passColor);
- if(pass.m_beginRenderpassInfo.hasRenderpass())
- {
- cmdb->beginRenderPass({pass.m_beginRenderpassInfo.m_colorRts.getBegin(), U32(pass.m_beginRenderpassInfo.m_colorRtCount)},
- pass.m_beginRenderpassInfo.m_dsRt.m_view ? &pass.m_beginRenderpassInfo.m_dsRt : nullptr,
- pass.m_beginRenderpassInfo.m_renderArea[0], pass.m_beginRenderpassInfo.m_renderArea[1],
- pass.m_beginRenderpassInfo.m_renderArea[2], pass.m_beginRenderpassInfo.m_renderArea[3],
- pass.m_beginRenderpassInfo.m_vrsRt, pass.m_beginRenderpassInfo.m_vrsTexelSizeX,
- pass.m_beginRenderpassInfo.m_vrsTexelSizeY);
- }
- {
- ANKI_TRACE_SCOPED_EVENT(GrRenderGraphCallback);
- ctx.m_passIdx = passIdx;
- pass.m_callback(ctx);
- }
- if(pass.m_beginRenderpassInfo.hasRenderpass())
- {
- cmdb->endRenderPass();
- }
- cmdb->popDebugMarker();
- }
- } // end for batches
- if(setPostQuery)
- {
- // Write a timestamp before the last flush
- cmdb->writeTimestamp(postQuery.get());
- }
- cmdb->endRecording();
- });
- }
- CoreThreadJobManager::getSingleton().waitForAllTasksToFinish();
- // Submit
- if(cmdbs.getSize() == 1) [[unlikely]]
- {
- GrManager::getSingleton().submit(cmdbs[0].get(), {}, optionalFence);
- }
- else
- {
- // 2 submits. The 1st contains all the batches minus the last. Then the last batch is alone given that it most likely it writes to the
- // swapchain
- DynamicArray<CommandBuffer*, MemoryPoolPtrWrapper<StackMemoryPool>> pCmdbs(pool);
- pCmdbs.resize(cmdbs.getSize() - 1);
- for(U32 i = 0; i < cmdbs.getSize() - 1; ++i)
- {
- pCmdbs[i] = cmdbs[i].get();
- }
- GrManager::getSingleton().submit(WeakArray(pCmdbs), {}, nullptr);
- GrManager::getSingleton().submit(cmdbs.getBack().get(), {}, optionalFence);
- }
- }
- void RenderGraph::getCrntUsage(RenderTargetHandle handle, U32 batchIdx, const TextureSubresourceInfo& subresource, TextureUsageBit& usage) const
- {
- usage = TextureUsageBit::kNone;
- const Batch& batch = m_ctx->m_batches[batchIdx];
- for(U32 passIdx : batch.m_passIndices)
- {
- for(const RenderPassDependency::TextureInfo& consumer : m_ctx->m_passes[passIdx].m_consumedTextures)
- {
- if(consumer.m_handle == handle && overlappingTextureSubresource(subresource, consumer.m_subresource))
- {
- usage |= consumer.m_usage;
- break;
- }
- }
- }
- }
- void RenderGraph::periodicCleanup()
- {
- U32 rtsCleanedCount = 0;
- for(RenderTargetCacheEntry& entry : m_renderTargetCache)
- {
- if(entry.m_texturesInUse < entry.m_textures.getSize())
- {
- // Should cleanup
- rtsCleanedCount += entry.m_textures.getSize() - entry.m_texturesInUse;
- // New array
- GrDynamicArray<TexturePtr> newArray;
- if(entry.m_texturesInUse > 0)
- {
- newArray.resize(entry.m_texturesInUse);
- }
- // Populate the new array
- for(U32 i = 0; i < newArray.getSize(); ++i)
- {
- newArray[i] = std::move(entry.m_textures[i]);
- }
- // Destroy the old array and the rest of the textures
- entry.m_textures.destroy();
- // Move new array
- entry.m_textures = std::move(newArray);
- }
- }
- if(rtsCleanedCount > 0)
- {
- ANKI_GR_LOGI("Cleaned %u render targets", rtsCleanedCount);
- }
- }
- void RenderGraph::getStatistics(RenderGraphStatistics& statistics) const
- {
- const U32 oldFrame = (m_statistics.m_nextTimestamp + 1) % kMaxBufferedTimestamps;
- if(m_statistics.m_timestamps[oldFrame * 2] && m_statistics.m_timestamps[oldFrame * 2 + 1])
- {
- Second start, end;
- [[maybe_unused]] TimestampQueryResult res = m_statistics.m_timestamps[oldFrame * 2]->getResult(start);
- ANKI_ASSERT(res == TimestampQueryResult::kAvailable);
- res = m_statistics.m_timestamps[oldFrame * 2 + 1]->getResult(end);
- ANKI_ASSERT(res == TimestampQueryResult::kAvailable);
- const Second diff = end - start;
- statistics.m_gpuTime = diff;
- statistics.m_cpuStartTime = m_statistics.m_cpuStartTimes[oldFrame];
- }
- else
- {
- statistics.m_gpuTime = -1.0;
- statistics.m_cpuStartTime = -1.0;
- }
- }
- #if ANKI_DBG_RENDER_GRAPH
- StringRaii RenderGraph::textureUsageToStr(StackMemoryPool& pool, TextureUsageBit usage)
- {
- if(!usage)
- {
- return StringRaii(&pool, "None");
- }
- StringListRaii slist(&pool);
- # define ANKI_TEX_USAGE(u) \
- if(!!(usage & TextureUsageBit::u)) \
- { \
- slist.pushBackSprintf("%s", #u); \
- }
- ANKI_TEX_USAGE(kSampledGeometry);
- ANKI_TEX_USAGE(kSampledFragment);
- ANKI_TEX_USAGE(kSampledCompute);
- ANKI_TEX_USAGE(kSampledTraceRays);
- ANKI_TEX_USAGE(kUavGeometryRead);
- ANKI_TEX_USAGE(kUavGeometryWrite);
- ANKI_TEX_USAGE(kUavFragmentRead);
- ANKI_TEX_USAGE(kUavFragmentWrite);
- ANKI_TEX_USAGE(kUavComputeRead);
- ANKI_TEX_USAGE(kUavComputeWrite);
- ANKI_TEX_USAGE(kUavTraceRaysRead);
- ANKI_TEX_USAGE(kUavTraceRaysWrite);
- ANKI_TEX_USAGE(kFramebufferRead);
- ANKI_TEX_USAGE(kFramebufferWrite);
- ANKI_TEX_USAGE(kTransferDestination);
- ANKI_TEX_USAGE(kGenerateMipmaps);
- ANKI_TEX_USAGE(kPresent);
- ANKI_TEX_USAGE(kFramebufferShadingRate);
- if(!usage)
- {
- slist.pushBackSprintf("?");
- }
- # undef ANKI_TEX_USAGE
- ANKI_ASSERT(!slist.isEmpty());
- StringRaii str(&pool);
- slist.join(" | ", str);
- return str;
- }
- StringRaii RenderGraph::bufferUsageToStr(StackMemoryPool& pool, BufferUsageBit usage)
- {
- StringListRaii slist(&pool);
- # define ANKI_BUFF_USAGE(u) \
- if(!!(usage & BufferUsageBit::u)) \
- { \
- slist.pushBackSprintf("%s", #u); \
- }
- ANKI_BUFF_USAGE(kConstantGeometry);
- ANKI_BUFF_USAGE(kConstantFragment);
- ANKI_BUFF_USAGE(kConstantCompute);
- ANKI_BUFF_USAGE(kConstantTraceRays);
- ANKI_BUFF_USAGE(kStorageGeometryRead);
- ANKI_BUFF_USAGE(kStorageGeometryWrite);
- ANKI_BUFF_USAGE(kStorageFragmentRead);
- ANKI_BUFF_USAGE(kStorageFragmentWrite);
- ANKI_BUFF_USAGE(kStorageComputeRead);
- ANKI_BUFF_USAGE(kStorageComputeWrite);
- ANKI_BUFF_USAGE(kStorageTraceRaysRead);
- ANKI_BUFF_USAGE(kStorageTraceRaysWrite);
- ANKI_BUFF_USAGE(kTextureGeometryRead);
- ANKI_BUFF_USAGE(kTextureGeometryWrite);
- ANKI_BUFF_USAGE(kTextureFragmentRead);
- ANKI_BUFF_USAGE(kTextureFragmentWrite);
- ANKI_BUFF_USAGE(kTextureComputeRead);
- ANKI_BUFF_USAGE(kTextureComputeWrite);
- ANKI_BUFF_USAGE(kTextureTraceRaysRead);
- ANKI_BUFF_USAGE(kTextureTraceRaysWrite);
- ANKI_BUFF_USAGE(kIndex);
- ANKI_BUFF_USAGE(kVertex);
- ANKI_BUFF_USAGE(kIndirectCompute);
- ANKI_BUFF_USAGE(kIndirectDraw);
- ANKI_BUFF_USAGE(kIndirectTraceRays);
- ANKI_BUFF_USAGE(kTransferSource);
- ANKI_BUFF_USAGE(kTransferDestination);
- ANKI_BUFF_USAGE(kAccelerationStructureBuild);
- if(!usage)
- {
- slist.pushBackSprintf("NONE");
- }
- # undef ANKI_BUFF_USAGE
- ANKI_ASSERT(!slist.isEmpty());
- StringRaii str(&pool);
- slist.join(" | ", str);
- return str;
- }
- StringRaii RenderGraph::asUsageToStr(StackMemoryPool& pool, AccelerationStructureUsageBit usage)
- {
- StringListRaii slist(&pool);
- # define ANKI_AS_USAGE(u) \
- if(!!(usage & AccelerationStructureUsageBit::u)) \
- { \
- slist.pushBackSprintf("%s", #u); \
- }
- ANKI_AS_USAGE(kBuild);
- ANKI_AS_USAGE(kAttach);
- ANKI_AS_USAGE(kGeometryRead);
- ANKI_AS_USAGE(kFragmentRead);
- ANKI_AS_USAGE(kComputeRead);
- ANKI_AS_USAGE(kTraceRaysRead);
- if(!usage)
- {
- slist.pushBackSprintf("NONE");
- }
- # undef ANKI_AS_USAGE
- ANKI_ASSERT(!slist.isEmpty());
- StringRaii str(&pool);
- slist.join(" | ", str);
- return str;
- }
- Error RenderGraph::dumpDependencyDotFile(const RenderGraphDescription& descr, const BakeContext& ctx, CString path) const
- {
- ANKI_GR_LOGW("Running with debug code");
- static constexpr Array<const char*, 5> COLORS = {"red", "green", "blue", "magenta", "cyan"};
- StackMemoryPool& pool = *ctx.m_pool;
- StringListRaii slist(&pool);
- slist.pushBackSprintf("digraph {\n");
- slist.pushBackSprintf("\t//splines = ortho;\nconcentrate = true;\n");
- for(U32 batchIdx = 0; batchIdx < ctx.m_batches.getSize(); ++batchIdx)
- {
- // Set same rank
- slist.pushBackSprintf("\t{rank=\"same\";");
- for(U32 passIdx : ctx.m_batches[batchIdx].m_passIndices)
- {
- slist.pushBackSprintf("\"%s\";", descr.m_passes[passIdx]->m_name.cstr());
- }
- slist.pushBackSprintf("}\n");
- // Print passes
- for(U32 passIdx : ctx.m_batches[batchIdx].m_passIndices)
- {
- CString passName = descr.m_passes[passIdx]->m_name.toCString();
- slist.pushBackSprintf("\t\"%s\"[color=%s,style=%s,shape=box];\n", passName.cstr(), COLORS[batchIdx % COLORS.getSize()],
- (descr.m_passes[passIdx]->m_type == RenderPassDescriptionBase::Type::kGraphics) ? "bold" : "dashed");
- for(U32 depIdx : ctx.m_passes[passIdx].m_dependsOn)
- {
- slist.pushBackSprintf("\t\"%s\"->\"%s\";\n", descr.m_passes[depIdx]->m_name.cstr(), passName.cstr());
- }
- if(ctx.m_passes[passIdx].m_dependsOn.getSize() == 0)
- {
- slist.pushBackSprintf("\tNONE->\"%s\";\n", descr.m_passes[passIdx]->m_name.cstr());
- }
- }
- }
- # if 0
- // Color the resources
- slist.pushBackSprintf("subgraph cluster_0 {\n");
- for(U rtIdx = 0; rtIdx < descr.m_renderTargets.getSize(); ++rtIdx)
- {
- slist.pushBackSprintf(
- "\t\"%s\"[color=%s];\n", &descr.m_renderTargets[rtIdx].m_name[0], COLORS[rtIdx % COLORS.getSize()]);
- }
- slist.pushBackSprintf("}\n");
- # endif
- // Barriers
- // slist.pushBackSprintf("subgraph cluster_1 {\n");
- StringRaii prevBubble(&pool);
- prevBubble.create("START");
- for(U32 batchIdx = 0; batchIdx < ctx.m_batches.getSize(); ++batchIdx)
- {
- const Batch& batch = ctx.m_batches[batchIdx];
- StringRaii batchName(&pool);
- batchName.sprintf("batch%u", batchIdx);
- for(U32 barrierIdx = 0; barrierIdx < batch.m_textureBarriersBefore.getSize(); ++barrierIdx)
- {
- const TextureBarrier& barrier = batch.m_textureBarriersBefore[barrierIdx];
- StringRaii barrierLabel(&pool);
- barrierLabel.sprintf("<b>%s</b> (mip,dp,f,l)=(%u,%u,%u,%u)<br/>%s <b>to</b> %s", &descr.m_renderTargets[barrier.m_idx].m_name[0],
- barrier.m_surface.m_level, barrier.m_surface.m_depth, barrier.m_surface.m_face, barrier.m_surface.m_layer,
- textureUsageToStr(pool, barrier.m_usageBefore).cstr(), textureUsageToStr(pool, barrier.m_usageAfter).cstr());
- StringRaii barrierName(&pool);
- barrierName.sprintf("%s tex barrier%u", batchName.cstr(), barrierIdx);
- slist.pushBackSprintf("\t\"%s\"[color=%s,style=bold,shape=box,label=< %s >];\n", barrierName.cstr(), COLORS[batchIdx % COLORS.getSize()],
- barrierLabel.cstr());
- slist.pushBackSprintf("\t\"%s\"->\"%s\";\n", prevBubble.cstr(), barrierName.cstr());
- prevBubble = barrierName;
- }
- for(U32 barrierIdx = 0; barrierIdx < batch.m_bufferBarriersBefore.getSize(); ++barrierIdx)
- {
- const BufferBarrier& barrier = batch.m_bufferBarriersBefore[barrierIdx];
- StringRaii barrierLabel(&pool);
- barrierLabel.sprintf("<b>%s</b><br/>%s <b>to</b> %s", &descr.m_buffers[barrier.m_idx].m_name[0],
- bufferUsageToStr(pool, barrier.m_usageBefore).cstr(), bufferUsageToStr(pool, barrier.m_usageAfter).cstr());
- StringRaii barrierName(&pool);
- barrierName.sprintf("%s buff barrier%u", batchName.cstr(), barrierIdx);
- slist.pushBackSprintf("\t\"%s\"[color=%s,style=bold,shape=box,label=< %s >];\n", barrierName.cstr(), COLORS[batchIdx % COLORS.getSize()],
- barrierLabel.cstr());
- slist.pushBackSprintf("\t\"%s\"->\"%s\";\n", prevBubble.cstr(), barrierName.cstr());
- prevBubble = barrierName;
- }
- for(U32 barrierIdx = 0; barrierIdx < batch.m_asBarriersBefore.getSize(); ++barrierIdx)
- {
- const ASBarrier& barrier = batch.m_asBarriersBefore[barrierIdx];
- StringRaii barrierLabel(&pool);
- barrierLabel.sprintf("<b>%s</b><br/>%s <b>to</b> %s", descr.m_as[barrier.m_idx].m_name.getBegin(),
- asUsageToStr(pool, barrier.m_usageBefore).cstr(), asUsageToStr(pool, barrier.m_usageAfter).cstr());
- StringRaii barrierName(&pool);
- barrierName.sprintf("%s AS barrier%u", batchName.cstr(), barrierIdx);
- slist.pushBackSprintf("\t\"%s\"[color=%s,style=bold,shape=box,label=< %s >];\n", barrierName.cstr(), COLORS[batchIdx % COLORS.getSize()],
- barrierLabel.cstr());
- slist.pushBackSprintf("\t\"%s\"->\"%s\";\n", prevBubble.cstr(), barrierName.cstr());
- prevBubble = barrierName;
- }
- for(U32 passIdx : batch.m_passIndices)
- {
- const RenderPassDescriptionBase& pass = *descr.m_passes[passIdx];
- StringRaii passName(&pool);
- passName.sprintf("%s pass", pass.m_name.cstr());
- slist.pushBackSprintf("\t\"%s\"[color=%s,style=bold];\n", passName.cstr(), COLORS[batchIdx % COLORS.getSize()]);
- slist.pushBackSprintf("\t\"%s\"->\"%s\";\n", prevBubble.cstr(), passName.cstr());
- prevBubble = passName;
- }
- }
- // slist.pushBackSprintf("}\n");
- slist.pushBackSprintf("}");
- File file;
- ANKI_CHECK(file.open(StringRaii(&pool).sprintf("%s/rgraph_%05u.dot", &path[0], m_version).toCString(), FileOpenFlag::kWrite));
- for(const String& s : slist)
- {
- ANKI_CHECK(file.writeTextf("%s", &s[0]));
- }
- return Error::kNone;
- }
- #endif
- } // end namespace anki
|