renderer_vk.cpp 287 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223222422252226222722282229223022312232223322342235223622372238223922402241224222432244224522462247224822492250225122522253225422552256225722582259226022612262226322642265226622672268226922702271227222732274227522762277227822792280228122822283228422852286228722882289229022912292229322942295229622972298229923002301230223032304230523062307230823092310231123122313231423152316231723182319232023212322232323242325232623272328232923302331233223332334233523362337233823392340234123422343234423452346234723482349235023512352235323542355235623572358235923602361236223632364236523662367236823692370237123722373237423752376237723782379238023812382238323842385238623872388238923902391239223932394239523962397239823992400240124022403240424052406240724082409241024112412241324142415241624172418241924202421242224232424242524262427242824292430243124322433243424352436243724382439244024412442244324442445244624472448244924502451245224532454245524562457245824592460246124622463246424652466246724682469247024712472247324742475247624772478247924802481248224832484248524862487248824892490249124922493249424952496249724982499250025012502250325042505250625072508250925102511251225132514251525162517251825192520252125222523252425252526252725282529253025312532253325342535253625372538253925402541254225432544254525462547254825492550255125522553255425552556255725582559256025612562256325642565256625672568256925702571257225732574257525762577257825792580258125822583258425852586258725882589259025912592259325942595259625972598259926002601260226032604260526062607260826092610261126122613261426152616261726182619262026212622262326242625262626272628262926302631263226332634263526362637263826392640264126422643264426452646264726482649265026512652265326542655265626572658265926602661266226632664266526662667266826692670267126722673267426752676267726782679268026812682268326842685268626872688268926902691269226932694269526962697269826992700270127022703270427052706270727082709271027112712271327142715271627172718271927202721272227232724272527262727272827292730273127322733273427352736273727382739274027412742274327442745274627472748274927502751275227532754275527562757275827592760276127622763276427652766276727682769277027712772277327742775277627772778277927802781278227832784278527862787278827892790279127922793279427952796279727982799280028012802280328042805280628072808280928102811281228132814281528162817281828192820282128222823282428252826282728282829283028312832283328342835283628372838283928402841284228432844284528462847284828492850285128522853285428552856285728582859286028612862286328642865286628672868286928702871287228732874287528762877287828792880288128822883288428852886288728882889289028912892289328942895289628972898289929002901290229032904290529062907290829092910291129122913291429152916291729182919292029212922292329242925292629272928292929302931293229332934293529362937293829392940294129422943294429452946294729482949295029512952295329542955295629572958295929602961296229632964296529662967296829692970297129722973297429752976297729782979298029812982298329842985298629872988298929902991299229932994299529962997299829993000300130023003300430053006300730083009301030113012301330143015301630173018301930203021302230233024302530263027302830293030303130323033303430353036303730383039304030413042304330443045304630473048304930503051305230533054305530563057305830593060306130623063306430653066306730683069307030713072307330743075307630773078307930803081308230833084308530863087308830893090309130923093309430953096309730983099310031013102310331043105310631073108310931103111311231133114311531163117311831193120312131223123312431253126312731283129313031313132313331343135313631373138313931403141314231433144314531463147314831493150315131523153315431553156315731583159316031613162316331643165316631673168316931703171317231733174317531763177317831793180318131823183318431853186318731883189319031913192319331943195319631973198319932003201320232033204320532063207320832093210321132123213321432153216321732183219322032213222322332243225322632273228322932303231323232333234323532363237323832393240324132423243324432453246324732483249325032513252325332543255325632573258325932603261326232633264326532663267326832693270327132723273327432753276327732783279328032813282328332843285328632873288328932903291329232933294329532963297329832993300330133023303330433053306330733083309331033113312331333143315331633173318331933203321332233233324332533263327332833293330333133323333333433353336333733383339334033413342334333443345334633473348334933503351335233533354335533563357335833593360336133623363336433653366336733683369337033713372337333743375337633773378337933803381338233833384338533863387338833893390339133923393339433953396339733983399340034013402340334043405340634073408340934103411341234133414341534163417341834193420342134223423342434253426342734283429343034313432343334343435343634373438343934403441344234433444344534463447344834493450345134523453345434553456345734583459346034613462346334643465346634673468346934703471347234733474347534763477347834793480348134823483348434853486348734883489349034913492349334943495349634973498349935003501350235033504350535063507350835093510351135123513351435153516351735183519352035213522352335243525352635273528352935303531353235333534353535363537353835393540354135423543354435453546354735483549355035513552355335543555355635573558355935603561356235633564356535663567356835693570357135723573357435753576357735783579358035813582358335843585358635873588358935903591359235933594359535963597359835993600360136023603360436053606360736083609361036113612361336143615361636173618361936203621362236233624362536263627362836293630363136323633363436353636363736383639364036413642364336443645364636473648364936503651365236533654365536563657365836593660366136623663366436653666366736683669367036713672367336743675367636773678367936803681368236833684368536863687368836893690369136923693369436953696369736983699370037013702370337043705370637073708370937103711371237133714371537163717371837193720372137223723372437253726372737283729373037313732373337343735373637373738373937403741374237433744374537463747374837493750375137523753375437553756375737583759376037613762376337643765376637673768376937703771377237733774377537763777377837793780378137823783378437853786378737883789379037913792379337943795379637973798379938003801380238033804380538063807380838093810381138123813381438153816381738183819382038213822382338243825382638273828382938303831383238333834383538363837383838393840384138423843384438453846384738483849385038513852385338543855385638573858385938603861386238633864386538663867386838693870387138723873387438753876387738783879388038813882388338843885388638873888388938903891389238933894389538963897389838993900390139023903390439053906390739083909391039113912391339143915391639173918391939203921392239233924392539263927392839293930393139323933393439353936393739383939394039413942394339443945394639473948394939503951395239533954395539563957395839593960396139623963396439653966396739683969397039713972397339743975397639773978397939803981398239833984398539863987398839893990399139923993399439953996399739983999400040014002400340044005400640074008400940104011401240134014401540164017401840194020402140224023402440254026402740284029403040314032403340344035403640374038403940404041404240434044404540464047404840494050405140524053405440554056405740584059406040614062406340644065406640674068406940704071407240734074407540764077407840794080408140824083408440854086408740884089409040914092409340944095409640974098409941004101410241034104410541064107410841094110411141124113411441154116411741184119412041214122412341244125412641274128412941304131413241334134413541364137413841394140414141424143414441454146414741484149415041514152415341544155415641574158415941604161416241634164416541664167416841694170417141724173417441754176417741784179418041814182418341844185418641874188418941904191419241934194419541964197419841994200420142024203420442054206420742084209421042114212421342144215421642174218421942204221422242234224422542264227422842294230423142324233423442354236423742384239424042414242424342444245424642474248424942504251425242534254425542564257425842594260426142624263426442654266426742684269427042714272427342744275427642774278427942804281428242834284428542864287428842894290429142924293429442954296429742984299430043014302430343044305430643074308430943104311431243134314431543164317431843194320432143224323432443254326432743284329433043314332433343344335433643374338433943404341434243434344434543464347434843494350435143524353435443554356435743584359436043614362436343644365436643674368436943704371437243734374437543764377437843794380438143824383438443854386438743884389439043914392439343944395439643974398439944004401440244034404440544064407440844094410441144124413441444154416441744184419442044214422442344244425442644274428442944304431443244334434443544364437443844394440444144424443444444454446444744484449445044514452445344544455445644574458445944604461446244634464446544664467446844694470447144724473447444754476447744784479448044814482448344844485448644874488448944904491449244934494449544964497449844994500450145024503450445054506450745084509451045114512451345144515451645174518451945204521452245234524452545264527452845294530453145324533453445354536453745384539454045414542454345444545454645474548454945504551455245534554455545564557455845594560456145624563456445654566456745684569457045714572457345744575457645774578457945804581458245834584458545864587458845894590459145924593459445954596459745984599460046014602460346044605460646074608460946104611461246134614461546164617461846194620462146224623462446254626462746284629463046314632463346344635463646374638463946404641464246434644464546464647464846494650465146524653465446554656465746584659466046614662466346644665466646674668466946704671467246734674467546764677467846794680468146824683468446854686468746884689469046914692469346944695469646974698469947004701470247034704470547064707470847094710471147124713471447154716471747184719472047214722472347244725472647274728472947304731473247334734473547364737473847394740474147424743474447454746474747484749475047514752475347544755475647574758475947604761476247634764476547664767476847694770477147724773477447754776477747784779478047814782478347844785478647874788478947904791479247934794479547964797479847994800480148024803480448054806480748084809481048114812481348144815481648174818481948204821482248234824482548264827482848294830483148324833483448354836483748384839484048414842484348444845484648474848484948504851485248534854485548564857485848594860486148624863486448654866486748684869487048714872487348744875487648774878487948804881488248834884488548864887488848894890489148924893489448954896489748984899490049014902490349044905490649074908490949104911491249134914491549164917491849194920492149224923492449254926492749284929493049314932493349344935493649374938493949404941494249434944494549464947494849494950495149524953495449554956495749584959496049614962496349644965496649674968496949704971497249734974497549764977497849794980498149824983498449854986498749884989499049914992499349944995499649974998499950005001500250035004500550065007500850095010501150125013501450155016501750185019502050215022502350245025502650275028502950305031503250335034503550365037503850395040504150425043504450455046504750485049505050515052505350545055505650575058505950605061506250635064506550665067506850695070507150725073507450755076507750785079508050815082508350845085508650875088508950905091509250935094509550965097509850995100510151025103510451055106510751085109511051115112511351145115511651175118511951205121512251235124512551265127512851295130513151325133513451355136513751385139514051415142514351445145514651475148514951505151515251535154515551565157515851595160516151625163516451655166516751685169517051715172517351745175517651775178517951805181518251835184518551865187518851895190519151925193519451955196519751985199520052015202520352045205520652075208520952105211521252135214521552165217521852195220522152225223522452255226522752285229523052315232523352345235523652375238523952405241524252435244524552465247524852495250525152525253525452555256525752585259526052615262526352645265526652675268526952705271527252735274527552765277527852795280528152825283528452855286528752885289529052915292529352945295529652975298529953005301530253035304530553065307530853095310531153125313531453155316531753185319532053215322532353245325532653275328532953305331533253335334533553365337533853395340534153425343534453455346534753485349535053515352535353545355535653575358535953605361536253635364536553665367536853695370537153725373537453755376537753785379538053815382538353845385538653875388538953905391539253935394539553965397539853995400540154025403540454055406540754085409541054115412541354145415541654175418541954205421542254235424542554265427542854295430543154325433543454355436543754385439544054415442544354445445544654475448544954505451545254535454545554565457545854595460546154625463546454655466546754685469547054715472547354745475547654775478547954805481548254835484548554865487548854895490549154925493549454955496549754985499550055015502550355045505550655075508550955105511551255135514551555165517551855195520552155225523552455255526552755285529553055315532553355345535553655375538553955405541554255435544554555465547554855495550555155525553555455555556555755585559556055615562556355645565556655675568556955705571557255735574557555765577557855795580558155825583558455855586558755885589559055915592559355945595559655975598559956005601560256035604560556065607560856095610561156125613561456155616561756185619562056215622562356245625562656275628562956305631563256335634563556365637563856395640564156425643564456455646564756485649565056515652565356545655565656575658565956605661566256635664566556665667566856695670567156725673567456755676567756785679568056815682568356845685568656875688568956905691569256935694569556965697569856995700570157025703570457055706570757085709571057115712571357145715571657175718571957205721572257235724572557265727572857295730573157325733573457355736573757385739574057415742574357445745574657475748574957505751575257535754575557565757575857595760576157625763576457655766576757685769577057715772577357745775577657775778577957805781578257835784578557865787578857895790579157925793579457955796579757985799580058015802580358045805580658075808580958105811581258135814581558165817581858195820582158225823582458255826582758285829583058315832583358345835583658375838583958405841584258435844584558465847584858495850585158525853585458555856585758585859586058615862586358645865586658675868586958705871587258735874587558765877587858795880588158825883588458855886588758885889589058915892589358945895589658975898589959005901590259035904590559065907590859095910591159125913591459155916591759185919592059215922592359245925592659275928592959305931593259335934593559365937593859395940594159425943594459455946594759485949595059515952595359545955595659575958595959605961596259635964596559665967596859695970597159725973597459755976597759785979598059815982598359845985598659875988598959905991599259935994599559965997599859996000600160026003600460056006600760086009601060116012601360146015601660176018601960206021602260236024602560266027602860296030603160326033603460356036603760386039604060416042604360446045604660476048604960506051605260536054605560566057605860596060606160626063606460656066606760686069607060716072607360746075607660776078607960806081608260836084608560866087608860896090609160926093609460956096609760986099610061016102610361046105610661076108610961106111611261136114611561166117611861196120612161226123612461256126612761286129613061316132613361346135613661376138613961406141614261436144614561466147614861496150615161526153615461556156615761586159616061616162616361646165616661676168616961706171617261736174617561766177617861796180618161826183618461856186618761886189619061916192619361946195619661976198619962006201620262036204620562066207620862096210621162126213621462156216621762186219622062216222622362246225622662276228622962306231623262336234623562366237623862396240624162426243624462456246624762486249625062516252625362546255625662576258625962606261626262636264626562666267626862696270627162726273627462756276627762786279628062816282628362846285628662876288628962906291629262936294629562966297629862996300630163026303630463056306630763086309631063116312631363146315631663176318631963206321632263236324632563266327632863296330633163326333633463356336633763386339634063416342634363446345634663476348634963506351635263536354635563566357635863596360636163626363636463656366636763686369637063716372637363746375637663776378637963806381638263836384638563866387638863896390639163926393639463956396639763986399640064016402640364046405640664076408640964106411641264136414641564166417641864196420642164226423642464256426642764286429643064316432643364346435643664376438643964406441644264436444644564466447644864496450645164526453645464556456645764586459646064616462646364646465646664676468646964706471647264736474647564766477647864796480648164826483648464856486648764886489649064916492649364946495649664976498649965006501650265036504650565066507650865096510651165126513651465156516651765186519652065216522652365246525652665276528652965306531653265336534653565366537653865396540654165426543654465456546654765486549655065516552655365546555655665576558655965606561656265636564656565666567656865696570657165726573657465756576657765786579658065816582658365846585658665876588658965906591659265936594659565966597659865996600660166026603660466056606660766086609661066116612661366146615661666176618661966206621662266236624662566266627662866296630663166326633663466356636663766386639664066416642664366446645664666476648664966506651665266536654665566566657665866596660666166626663666466656666666766686669667066716672667366746675667666776678667966806681668266836684668566866687668866896690669166926693669466956696669766986699670067016702670367046705670667076708670967106711671267136714671567166717671867196720672167226723672467256726672767286729673067316732673367346735673667376738673967406741674267436744674567466747674867496750675167526753675467556756675767586759676067616762676367646765676667676768676967706771677267736774677567766777677867796780678167826783678467856786678767886789679067916792679367946795679667976798679968006801680268036804680568066807680868096810681168126813681468156816681768186819682068216822682368246825682668276828682968306831683268336834683568366837683868396840684168426843684468456846684768486849685068516852685368546855685668576858685968606861686268636864686568666867686868696870687168726873687468756876687768786879688068816882688368846885688668876888688968906891689268936894689568966897689868996900690169026903690469056906690769086909691069116912691369146915691669176918691969206921692269236924692569266927692869296930693169326933693469356936693769386939694069416942694369446945694669476948694969506951695269536954695569566957695869596960696169626963696469656966696769686969697069716972697369746975697669776978697969806981698269836984698569866987698869896990699169926993699469956996699769986999700070017002700370047005700670077008700970107011701270137014701570167017701870197020702170227023702470257026702770287029703070317032703370347035703670377038703970407041704270437044704570467047704870497050705170527053705470557056705770587059706070617062706370647065706670677068706970707071707270737074707570767077707870797080708170827083708470857086708770887089709070917092709370947095709670977098709971007101710271037104710571067107710871097110711171127113711471157116711771187119712071217122712371247125712671277128712971307131713271337134713571367137713871397140714171427143714471457146714771487149715071517152715371547155715671577158715971607161716271637164716571667167716871697170717171727173717471757176717771787179718071817182718371847185718671877188718971907191719271937194719571967197719871997200720172027203720472057206720772087209721072117212721372147215721672177218721972207221722272237224722572267227722872297230723172327233723472357236723772387239724072417242724372447245724672477248724972507251725272537254725572567257725872597260726172627263726472657266726772687269727072717272727372747275727672777278727972807281728272837284728572867287728872897290729172927293729472957296729772987299730073017302730373047305730673077308730973107311731273137314731573167317731873197320732173227323732473257326732773287329733073317332733373347335733673377338733973407341734273437344734573467347734873497350735173527353735473557356735773587359736073617362736373647365736673677368736973707371737273737374737573767377737873797380738173827383738473857386738773887389739073917392739373947395739673977398739974007401740274037404740574067407740874097410741174127413741474157416741774187419742074217422742374247425742674277428742974307431743274337434743574367437743874397440744174427443744474457446744774487449745074517452745374547455745674577458745974607461746274637464746574667467746874697470747174727473747474757476747774787479748074817482748374847485748674877488748974907491749274937494749574967497749874997500750175027503750475057506750775087509751075117512751375147515751675177518751975207521752275237524752575267527752875297530753175327533753475357536753775387539754075417542754375447545754675477548754975507551755275537554755575567557755875597560756175627563756475657566756775687569757075717572757375747575757675777578757975807581758275837584758575867587758875897590759175927593759475957596759775987599760076017602760376047605760676077608760976107611761276137614761576167617761876197620762176227623762476257626762776287629763076317632763376347635763676377638763976407641764276437644764576467647764876497650765176527653765476557656765776587659766076617662766376647665766676677668766976707671767276737674767576767677767876797680768176827683768476857686768776887689769076917692769376947695769676977698769977007701770277037704770577067707770877097710771177127713771477157716771777187719772077217722772377247725772677277728772977307731773277337734773577367737773877397740774177427743774477457746774777487749775077517752775377547755775677577758775977607761776277637764776577667767776877697770777177727773777477757776777777787779778077817782778377847785778677877788778977907791779277937794779577967797779877997800780178027803780478057806780778087809781078117812781378147815781678177818781978207821782278237824782578267827782878297830783178327833783478357836783778387839784078417842784378447845784678477848784978507851785278537854785578567857785878597860786178627863786478657866786778687869787078717872787378747875787678777878787978807881788278837884788578867887788878897890789178927893789478957896789778987899790079017902790379047905790679077908790979107911791279137914791579167917791879197920792179227923792479257926792779287929793079317932793379347935793679377938793979407941794279437944794579467947794879497950795179527953795479557956795779587959796079617962796379647965796679677968796979707971797279737974797579767977797879797980798179827983798479857986798779887989799079917992799379947995799679977998799980008001800280038004800580068007800880098010801180128013801480158016801780188019802080218022802380248025802680278028802980308031803280338034803580368037803880398040804180428043804480458046804780488049805080518052805380548055805680578058805980608061806280638064806580668067806880698070807180728073807480758076807780788079808080818082808380848085808680878088808980908091809280938094809580968097809880998100810181028103810481058106810781088109811081118112811381148115811681178118811981208121812281238124812581268127812881298130813181328133813481358136813781388139814081418142814381448145814681478148814981508151815281538154815581568157815881598160816181628163816481658166816781688169817081718172817381748175817681778178817981808181818281838184818581868187818881898190819181928193819481958196819781988199820082018202820382048205820682078208820982108211821282138214821582168217821882198220822182228223822482258226822782288229823082318232823382348235823682378238823982408241824282438244824582468247824882498250825182528253825482558256825782588259826082618262826382648265826682678268826982708271827282738274827582768277827882798280828182828283828482858286828782888289829082918292829382948295829682978298829983008301830283038304830583068307830883098310831183128313831483158316831783188319832083218322832383248325832683278328832983308331833283338334833583368337833883398340834183428343834483458346834783488349835083518352835383548355835683578358835983608361836283638364836583668367836883698370837183728373837483758376837783788379838083818382838383848385838683878388838983908391839283938394839583968397839883998400840184028403840484058406840784088409841084118412841384148415841684178418841984208421842284238424842584268427842884298430843184328433843484358436843784388439844084418442844384448445844684478448844984508451845284538454845584568457845884598460846184628463846484658466846784688469847084718472847384748475847684778478847984808481848284838484848584868487848884898490849184928493849484958496849784988499850085018502850385048505850685078508850985108511851285138514851585168517851885198520852185228523852485258526852785288529853085318532853385348535853685378538853985408541854285438544854585468547854885498550855185528553855485558556855785588559856085618562856385648565856685678568856985708571857285738574857585768577857885798580858185828583858485858586858785888589859085918592859385948595859685978598859986008601860286038604860586068607860886098610861186128613861486158616861786188619862086218622862386248625862686278628862986308631863286338634863586368637863886398640864186428643864486458646864786488649865086518652865386548655865686578658865986608661866286638664866586668667866886698670867186728673867486758676867786788679868086818682868386848685868686878688868986908691869286938694869586968697869886998700870187028703870487058706870787088709871087118712871387148715871687178718871987208721872287238724872587268727872887298730873187328733873487358736873787388739874087418742874387448745874687478748874987508751875287538754875587568757875887598760876187628763876487658766876787688769877087718772877387748775877687778778877987808781878287838784878587868787878887898790879187928793879487958796879787988799880088018802880388048805880688078808880988108811881288138814881588168817881888198820882188228823882488258826882788288829883088318832883388348835883688378838883988408841884288438844884588468847884888498850885188528853885488558856885788588859886088618862886388648865886688678868886988708871887288738874887588768877887888798880888188828883888488858886888788888889889088918892889388948895889688978898889989008901890289038904890589068907890889098910891189128913891489158916891789188919892089218922892389248925892689278928892989308931893289338934893589368937893889398940894189428943894489458946894789488949895089518952895389548955895689578958895989608961896289638964896589668967896889698970897189728973897489758976897789788979898089818982898389848985898689878988898989908991899289938994899589968997899889999000900190029003900490059006900790089009901090119012901390149015901690179018901990209021902290239024902590269027902890299030903190329033903490359036903790389039904090419042904390449045904690479048904990509051905290539054905590569057905890599060906190629063906490659066906790689069907090719072907390749075907690779078907990809081908290839084908590869087908890899090909190929093909490959096909790989099910091019102910391049105910691079108910991109111911291139114911591169117911891199120912191229123912491259126912791289129913091319132913391349135913691379138913991409141914291439144914591469147914891499150915191529153915491559156915791589159916091619162916391649165916691679168916991709171917291739174917591769177917891799180918191829183918491859186918791889189919091919192919391949195919691979198919992009201920292039204920592069207920892099210921192129213921492159216921792189219922092219222922392249225922692279228922992309231923292339234923592369237923892399240
  1. /*
  2. * Copyright 2011-2024 Branimir Karadzic. All rights reserved.
  3. * License: https://github.com/bkaradzic/bgfx/blob/master/LICENSE
  4. */
  5. #include "bgfx_p.h"
  6. #if BGFX_CONFIG_RENDERER_VULKAN
  7. # include <bx/pixelformat.h>
  8. # include "renderer_vk.h"
  9. # include "shader_spirv.h"
  10. #if BX_PLATFORM_OSX
  11. # import <Cocoa/Cocoa.h>
  12. # import <Foundation/Foundation.h>
  13. # import <QuartzCore/QuartzCore.h>
  14. # import <Metal/Metal.h>
  15. #endif // BX_PLATFORM_OSX
  16. #if defined(WL_EGL_PLATFORM)
  17. # include <wayland-egl-backend.h>
  18. #endif // defined(WL_EGL_PLATFORM)
  19. namespace bgfx { namespace vk
  20. {
  21. static char s_viewName[BGFX_CONFIG_MAX_VIEWS][BGFX_CONFIG_MAX_VIEW_NAME];
  22. inline void setViewType(ViewId _view, const bx::StringView _str)
  23. {
  24. if (BX_ENABLED(BGFX_CONFIG_DEBUG_ANNOTATION || BGFX_CONFIG_PROFILER) )
  25. {
  26. bx::memCopy(&s_viewName[_view][3], _str.getPtr(), _str.getLength() );
  27. }
  28. }
  29. struct PrimInfo
  30. {
  31. VkPrimitiveTopology m_topology;
  32. uint32_t m_min;
  33. uint32_t m_div;
  34. uint32_t m_sub;
  35. };
  36. static const PrimInfo s_primInfo[] =
  37. {
  38. { VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST, 3, 3, 0 },
  39. { VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP, 3, 1, 2 },
  40. { VK_PRIMITIVE_TOPOLOGY_LINE_LIST, 2, 2, 0 },
  41. { VK_PRIMITIVE_TOPOLOGY_LINE_STRIP, 2, 1, 1 },
  42. { VK_PRIMITIVE_TOPOLOGY_POINT_LIST, 1, 1, 0 },
  43. { VK_PRIMITIVE_TOPOLOGY_MAX_ENUM, 0, 0, 0 },
  44. };
  45. BX_STATIC_ASSERT(Topology::Count == BX_COUNTOF(s_primInfo)-1);
  46. static MsaaSamplerVK s_msaa[] =
  47. {
  48. { 1, VK_SAMPLE_COUNT_1_BIT },
  49. { 2, VK_SAMPLE_COUNT_2_BIT },
  50. { 4, VK_SAMPLE_COUNT_4_BIT },
  51. { 8, VK_SAMPLE_COUNT_8_BIT },
  52. { 16, VK_SAMPLE_COUNT_16_BIT },
  53. };
  54. static const VkBlendFactor s_blendFactor[][2] =
  55. {
  56. { VkBlendFactor(0), VkBlendFactor(0) }, // ignored
  57. { VK_BLEND_FACTOR_ZERO, VK_BLEND_FACTOR_ZERO }, // ZERO
  58. { VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_ONE }, // ONE
  59. { VK_BLEND_FACTOR_SRC_COLOR, VK_BLEND_FACTOR_SRC_ALPHA }, // SRC_COLOR
  60. { VK_BLEND_FACTOR_ONE_MINUS_SRC_COLOR, VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA }, // INV_SRC_COLOR
  61. { VK_BLEND_FACTOR_SRC_ALPHA, VK_BLEND_FACTOR_SRC_ALPHA }, // SRC_ALPHA
  62. { VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA, VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA }, // INV_SRC_ALPHA
  63. { VK_BLEND_FACTOR_DST_ALPHA, VK_BLEND_FACTOR_DST_ALPHA }, // DST_ALPHA
  64. { VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA, VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA }, // INV_DST_ALPHA
  65. { VK_BLEND_FACTOR_DST_COLOR, VK_BLEND_FACTOR_DST_ALPHA }, // DST_COLOR
  66. { VK_BLEND_FACTOR_ONE_MINUS_DST_COLOR, VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA }, // INV_DST_COLOR
  67. { VK_BLEND_FACTOR_SRC_ALPHA, VK_BLEND_FACTOR_ONE }, // SRC_ALPHA_SAT
  68. { VK_BLEND_FACTOR_CONSTANT_COLOR, VK_BLEND_FACTOR_CONSTANT_COLOR }, // FACTOR
  69. { VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR, VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR }, // INV_FACTOR
  70. };
  71. static const VkBlendOp s_blendEquation[] =
  72. {
  73. VK_BLEND_OP_ADD,
  74. VK_BLEND_OP_SUBTRACT,
  75. VK_BLEND_OP_REVERSE_SUBTRACT,
  76. VK_BLEND_OP_MIN,
  77. VK_BLEND_OP_MAX,
  78. };
  79. static const VkCompareOp s_cmpFunc[] =
  80. {
  81. VkCompareOp(0), // ignored
  82. VK_COMPARE_OP_LESS,
  83. VK_COMPARE_OP_LESS_OR_EQUAL,
  84. VK_COMPARE_OP_EQUAL,
  85. VK_COMPARE_OP_GREATER_OR_EQUAL,
  86. VK_COMPARE_OP_GREATER,
  87. VK_COMPARE_OP_NOT_EQUAL,
  88. VK_COMPARE_OP_NEVER,
  89. VK_COMPARE_OP_ALWAYS,
  90. };
  91. static const VkStencilOp s_stencilOp[] =
  92. {
  93. VK_STENCIL_OP_ZERO,
  94. VK_STENCIL_OP_KEEP,
  95. VK_STENCIL_OP_REPLACE,
  96. VK_STENCIL_OP_INCREMENT_AND_WRAP,
  97. VK_STENCIL_OP_INCREMENT_AND_CLAMP,
  98. VK_STENCIL_OP_DECREMENT_AND_WRAP,
  99. VK_STENCIL_OP_DECREMENT_AND_CLAMP,
  100. VK_STENCIL_OP_INVERT,
  101. };
  102. static const VkCullModeFlagBits s_cullMode[] =
  103. {
  104. VK_CULL_MODE_NONE,
  105. VK_CULL_MODE_FRONT_BIT,
  106. VK_CULL_MODE_BACK_BIT,
  107. };
  108. static const VkSamplerAddressMode s_textureAddress[] =
  109. {
  110. VK_SAMPLER_ADDRESS_MODE_REPEAT,
  111. VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT,
  112. VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE,
  113. VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER,
  114. };
  115. struct PresentMode
  116. {
  117. VkPresentModeKHR mode;
  118. bool vsync;
  119. const char* name;
  120. };
  121. static const PresentMode s_presentMode[] =
  122. {
  123. { VK_PRESENT_MODE_FIFO_KHR, true, "VK_PRESENT_MODE_FIFO_KHR" },
  124. { VK_PRESENT_MODE_FIFO_RELAXED_KHR, true, "VK_PRESENT_MODE_FIFO_RELAXED_KHR" },
  125. { VK_PRESENT_MODE_MAILBOX_KHR, true, "VK_PRESENT_MODE_MAILBOX_KHR" },
  126. { VK_PRESENT_MODE_IMMEDIATE_KHR, false, "VK_PRESENT_MODE_IMMEDIATE_KHR" },
  127. };
  128. #define VK_IMPORT_FUNC(_optional, _func) PFN_##_func _func
  129. #define VK_IMPORT_INSTANCE_FUNC VK_IMPORT_FUNC
  130. #define VK_IMPORT_DEVICE_FUNC VK_IMPORT_FUNC
  131. VK_IMPORT
  132. VK_IMPORT_INSTANCE
  133. VK_IMPORT_DEVICE
  134. #undef VK_IMPORT_DEVICE_FUNC
  135. #undef VK_IMPORT_INSTANCE_FUNC
  136. #undef VK_IMPORT_FUNC
  137. struct TextureFormatInfo
  138. {
  139. VkFormat m_fmt;
  140. VkFormat m_fmtSrv;
  141. VkFormat m_fmtDsv;
  142. VkFormat m_fmtSrgb;
  143. VkComponentMapping m_mapping;
  144. };
  145. static const TextureFormatInfo s_textureFormat[] =
  146. {
  147. #define $_ VK_COMPONENT_SWIZZLE_IDENTITY
  148. #define $0 VK_COMPONENT_SWIZZLE_ZERO
  149. #define $1 VK_COMPONENT_SWIZZLE_ONE
  150. #define $R VK_COMPONENT_SWIZZLE_R
  151. #define $G VK_COMPONENT_SWIZZLE_G
  152. #define $B VK_COMPONENT_SWIZZLE_B
  153. #define $A VK_COMPONENT_SWIZZLE_A
  154. { VK_FORMAT_BC1_RGB_UNORM_BLOCK, VK_FORMAT_BC1_RGB_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC1_RGB_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC1
  155. { VK_FORMAT_BC2_UNORM_BLOCK, VK_FORMAT_BC2_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC2_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC2
  156. { VK_FORMAT_BC3_UNORM_BLOCK, VK_FORMAT_BC3_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC3_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC3
  157. { VK_FORMAT_BC4_UNORM_BLOCK, VK_FORMAT_BC4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC4
  158. { VK_FORMAT_BC5_UNORM_BLOCK, VK_FORMAT_BC5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC5
  159. { VK_FORMAT_BC6H_SFLOAT_BLOCK, VK_FORMAT_BC6H_SFLOAT_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC6H
  160. { VK_FORMAT_BC7_UNORM_BLOCK, VK_FORMAT_BC7_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC7_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC7
  161. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ETC1
  162. { VK_FORMAT_ETC2_R8G8B8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2
  163. { VK_FORMAT_ETC2_R8G8B8A8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8A8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2A
  164. { VK_FORMAT_ETC2_R8G8B8A1_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8A1_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2A1
  165. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC12
  166. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC14
  167. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC12A
  168. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC14A
  169. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC22
  170. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC24
  171. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATC
  172. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATCE
  173. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATCI
  174. { VK_FORMAT_ASTC_4x4_UNORM_BLOCK, VK_FORMAT_ASTC_4x4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_4x4_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC4x4
  175. { VK_FORMAT_ASTC_5x4_UNORM_BLOCK, VK_FORMAT_ASTC_5x4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_5x4_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC5x4
  176. { VK_FORMAT_ASTC_5x5_UNORM_BLOCK, VK_FORMAT_ASTC_5x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_5x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC5x5
  177. { VK_FORMAT_ASTC_6x5_UNORM_BLOCK, VK_FORMAT_ASTC_6x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_6x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC6x5
  178. { VK_FORMAT_ASTC_6x6_UNORM_BLOCK, VK_FORMAT_ASTC_6x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_6x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC6x6
  179. { VK_FORMAT_ASTC_8x5_UNORM_BLOCK, VK_FORMAT_ASTC_8x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x5
  180. { VK_FORMAT_ASTC_8x6_UNORM_BLOCK, VK_FORMAT_ASTC_8x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x6
  181. { VK_FORMAT_ASTC_8x8_UNORM_BLOCK, VK_FORMAT_ASTC_8x8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x8
  182. { VK_FORMAT_ASTC_10x5_UNORM_BLOCK, VK_FORMAT_ASTC_10x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x5
  183. { VK_FORMAT_ASTC_10x6_UNORM_BLOCK, VK_FORMAT_ASTC_10x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x6
  184. { VK_FORMAT_ASTC_10x8_UNORM_BLOCK, VK_FORMAT_ASTC_10x8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x8
  185. { VK_FORMAT_ASTC_10x10_UNORM_BLOCK, VK_FORMAT_ASTC_10x10_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x10_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x10
  186. { VK_FORMAT_ASTC_12x10_UNORM_BLOCK, VK_FORMAT_ASTC_12x10_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_12x10_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC12x10
  187. { VK_FORMAT_ASTC_12x12_UNORM_BLOCK, VK_FORMAT_ASTC_12x12_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_12x12_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC12x12
  188. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // Unknown
  189. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R1
  190. { VK_FORMAT_R8_UNORM, VK_FORMAT_R8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $0, $0, $0, $R } }, // A8
  191. { VK_FORMAT_R8_UNORM, VK_FORMAT_R8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8_SRGB, { $_, $_, $_, $_ } }, // R8
  192. { VK_FORMAT_R8_SINT, VK_FORMAT_R8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8I
  193. { VK_FORMAT_R8_UINT, VK_FORMAT_R8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8U
  194. { VK_FORMAT_R8_SNORM, VK_FORMAT_R8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8S
  195. { VK_FORMAT_R16_UNORM, VK_FORMAT_R16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16
  196. { VK_FORMAT_R16_SINT, VK_FORMAT_R16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16I
  197. { VK_FORMAT_R16_UINT, VK_FORMAT_R16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16U
  198. { VK_FORMAT_R16_SFLOAT, VK_FORMAT_R16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16F
  199. { VK_FORMAT_R16_SNORM, VK_FORMAT_R16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16S
  200. { VK_FORMAT_R32_SINT, VK_FORMAT_R32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32I
  201. { VK_FORMAT_R32_UINT, VK_FORMAT_R32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32U
  202. { VK_FORMAT_R32_SFLOAT, VK_FORMAT_R32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32F
  203. { VK_FORMAT_R8G8_UNORM, VK_FORMAT_R8G8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8_SRGB, { $_, $_, $_, $_ } }, // RG8
  204. { VK_FORMAT_R8G8_SINT, VK_FORMAT_R8G8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8I
  205. { VK_FORMAT_R8G8_UINT, VK_FORMAT_R8G8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8U
  206. { VK_FORMAT_R8G8_SNORM, VK_FORMAT_R8G8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8S
  207. { VK_FORMAT_R16G16_UNORM, VK_FORMAT_R16G16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16
  208. { VK_FORMAT_R16G16_SINT, VK_FORMAT_R16G16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16I
  209. { VK_FORMAT_R16G16_UINT, VK_FORMAT_R16G16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16U
  210. { VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16F
  211. { VK_FORMAT_R16G16_SNORM, VK_FORMAT_R16G16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16S
  212. { VK_FORMAT_R32G32_SINT, VK_FORMAT_R32G32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32I
  213. { VK_FORMAT_R32G32_UINT, VK_FORMAT_R32G32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32U
  214. { VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32F
  215. { VK_FORMAT_R8G8B8_UNORM, VK_FORMAT_R8G8B8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8
  216. { VK_FORMAT_R8G8B8_SINT, VK_FORMAT_R8G8B8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8I
  217. { VK_FORMAT_R8G8B8_UINT, VK_FORMAT_R8G8B8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8U
  218. { VK_FORMAT_R8G8B8_SNORM, VK_FORMAT_R8G8B8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGB8S
  219. { VK_FORMAT_E5B9G9R9_UFLOAT_PACK32, VK_FORMAT_E5B9G9R9_UFLOAT_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGB9E5F
  220. { VK_FORMAT_B8G8R8A8_UNORM, VK_FORMAT_B8G8R8A8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_B8G8R8A8_SRGB, { $_, $_, $_, $_ } }, // BGRA8
  221. { VK_FORMAT_R8G8B8A8_UNORM, VK_FORMAT_R8G8B8A8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8
  222. { VK_FORMAT_R8G8B8A8_SINT, VK_FORMAT_R8G8B8A8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8I
  223. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8U
  224. { VK_FORMAT_R8G8B8A8_SNORM, VK_FORMAT_R8G8B8A8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA8S
  225. { VK_FORMAT_R16G16B16A16_UNORM, VK_FORMAT_R16G16B16A16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16
  226. { VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16I
  227. { VK_FORMAT_R16G16B16A16_UINT, VK_FORMAT_R16G16B16A16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16U
  228. { VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16F
  229. { VK_FORMAT_R16G16B16A16_SNORM, VK_FORMAT_R16G16B16A16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16S
  230. { VK_FORMAT_R32G32B32A32_SINT, VK_FORMAT_R32G32B32A32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32I
  231. { VK_FORMAT_R32G32B32A32_UINT, VK_FORMAT_R32G32B32A32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32U
  232. { VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32F
  233. { VK_FORMAT_R5G6B5_UNORM_PACK16, VK_FORMAT_R5G6B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // B5G6R5
  234. { VK_FORMAT_B5G6R5_UNORM_PACK16, VK_FORMAT_B5G6R5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R5G6B5
  235. { VK_FORMAT_B4G4R4A4_UNORM_PACK16, VK_FORMAT_B4G4R4A4_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $G, $R, $A, $B } }, // BGRA4
  236. { VK_FORMAT_R4G4B4A4_UNORM_PACK16, VK_FORMAT_R4G4B4A4_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $A, $B, $G, $R } }, // RGBA4
  237. { VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BGR5A1
  238. { VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $B, $G, $R, $A } }, // RGB5A1
  239. { VK_FORMAT_A2R10G10B10_UNORM_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $B, $G, $R, $A } }, // RGB10A2
  240. { VK_FORMAT_B10G11R11_UFLOAT_PACK32, VK_FORMAT_B10G11R11_UFLOAT_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG11B10F
  241. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // UnknownDepth
  242. { VK_FORMAT_UNDEFINED, VK_FORMAT_R16_UNORM, VK_FORMAT_D16_UNORM, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D16
  243. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24
  244. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24S8
  245. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D32
  246. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D16F
  247. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24F
  248. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D32F
  249. { VK_FORMAT_UNDEFINED, VK_FORMAT_R8_UINT, VK_FORMAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D0S8
  250. #undef $_
  251. #undef $0
  252. #undef $1
  253. #undef $R
  254. #undef $G
  255. #undef $B
  256. #undef $A
  257. };
  258. BX_STATIC_ASSERT(TextureFormat::Count == BX_COUNTOF(s_textureFormat) );
  259. struct ImageTest
  260. {
  261. VkImageType type;
  262. VkImageUsageFlags usage;
  263. VkImageCreateFlags flags;
  264. uint32_t formatCaps[2];
  265. };
  266. static const ImageTest s_imageTest[] =
  267. {
  268. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_SAMPLED_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_2D, BGFX_CAPS_FORMAT_TEXTURE_2D_SRGB } },
  269. { VK_IMAGE_TYPE_3D, VK_IMAGE_USAGE_SAMPLED_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_3D, BGFX_CAPS_FORMAT_TEXTURE_3D_SRGB } },
  270. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_SAMPLED_BIT, VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT, { BGFX_CAPS_FORMAT_TEXTURE_CUBE, BGFX_CAPS_FORMAT_TEXTURE_CUBE_SRGB } },
  271. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER, 0 } },
  272. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER, 0 } },
  273. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_STORAGE_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_IMAGE_READ, 0 } },
  274. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_STORAGE_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_IMAGE_WRITE, 0 } },
  275. };
  276. struct LayerInfo
  277. {
  278. bool m_supported;
  279. bool m_initialize;
  280. };
  281. struct Layer
  282. {
  283. enum Enum
  284. {
  285. VK_LAYER_KHRONOS_validation,
  286. VK_LAYER_LUNARG_standard_validation,
  287. Count
  288. };
  289. const char* m_name;
  290. uint32_t m_minVersion;
  291. LayerInfo m_instance;
  292. LayerInfo m_device;
  293. };
  294. // Layer registry
  295. //
  296. static Layer s_layer[] =
  297. {
  298. { "VK_LAYER_KHRONOS_validation", 1, { false, false }, { false, false } },
  299. { "VK_LAYER_LUNARG_standard_validation", 1, { false, false }, { false, false } },
  300. { "", 0, { false, false }, { false, false } },
  301. };
  302. BX_STATIC_ASSERT(Layer::Count == BX_COUNTOF(s_layer)-1);
  303. void updateLayer(const char* _name, uint32_t _version, bool _instanceLayer)
  304. {
  305. bx::StringView layerName(_name);
  306. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  307. {
  308. Layer& layer = s_layer[ii];
  309. LayerInfo& layerInfo = _instanceLayer
  310. ? layer.m_instance
  311. : layer.m_device
  312. ;
  313. if (!layerInfo.m_supported && layerInfo.m_initialize)
  314. {
  315. if ( 0 == bx::strCmp(layerName, layer.m_name)
  316. && _version >= layer.m_minVersion)
  317. {
  318. layerInfo.m_supported = true;
  319. break;
  320. }
  321. }
  322. }
  323. }
  324. struct Extension
  325. {
  326. enum Enum
  327. {
  328. EXT_conservative_rasterization,
  329. EXT_custom_border_color,
  330. EXT_debug_report,
  331. EXT_debug_utils,
  332. EXT_line_rasterization,
  333. EXT_memory_budget,
  334. EXT_shader_viewport_index_layer,
  335. KHR_draw_indirect_count,
  336. KHR_get_physical_device_properties2,
  337. Count
  338. };
  339. const char* m_name;
  340. uint32_t m_minVersion;
  341. bool m_instanceExt;
  342. bool m_supported;
  343. bool m_initialize;
  344. Layer::Enum m_layer;
  345. };
  346. // Extension registry
  347. //
  348. static Extension s_extension[] =
  349. {
  350. { "VK_EXT_conservative_rasterization", 1, false, false, true, Layer::Count },
  351. { "VK_EXT_custom_border_color", 1, false, false, true, Layer::Count },
  352. { "VK_EXT_debug_report", 1, false, false, false, Layer::Count },
  353. { "VK_EXT_debug_utils", 1, false, false, BGFX_CONFIG_DEBUG_OBJECT_NAME || BGFX_CONFIG_DEBUG_ANNOTATION, Layer::Count },
  354. { "VK_EXT_line_rasterization", 1, false, false, true, Layer::Count },
  355. { "VK_EXT_memory_budget", 1, false, false, true, Layer::Count },
  356. { "VK_EXT_shader_viewport_index_layer", 1, false, false, true, Layer::Count },
  357. { "VK_KHR_draw_indirect_count", 1, false, false, true, Layer::Count },
  358. { "VK_KHR_get_physical_device_properties2", 1, false, false, true, Layer::Count },
  359. };
  360. BX_STATIC_ASSERT(Extension::Count == BX_COUNTOF(s_extension) );
  361. bool updateExtension(const char* _name, uint32_t _version, bool _instanceExt, Extension _extensions[Extension::Count])
  362. {
  363. bool supported = false;
  364. if (BX_ENABLED(BGFX_CONFIG_RENDERER_USE_EXTENSIONS) )
  365. {
  366. const bx::StringView ext(_name);
  367. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  368. {
  369. Extension& extension = _extensions[ii];
  370. const LayerInfo& layerInfo = _instanceExt
  371. ? s_layer[extension.m_layer].m_instance
  372. : s_layer[extension.m_layer].m_device
  373. ;
  374. if (!extension.m_supported
  375. && extension.m_initialize
  376. && (extension.m_layer == Layer::Count || layerInfo.m_supported) )
  377. {
  378. if ( 0 == bx::strCmp(ext, extension.m_name)
  379. && _version >= extension.m_minVersion)
  380. {
  381. extension.m_supported = true;
  382. extension.m_instanceExt = _instanceExt;
  383. supported = true;
  384. break;
  385. }
  386. }
  387. }
  388. }
  389. return supported;
  390. }
  391. static const VkFormat s_attribType[][4][2] =
  392. {
  393. { // Uint8
  394. { VK_FORMAT_R8_UINT, VK_FORMAT_R8_UNORM },
  395. { VK_FORMAT_R8G8_UINT, VK_FORMAT_R8G8_UNORM },
  396. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UNORM },
  397. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UNORM },
  398. },
  399. { // Uint10
  400. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  401. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  402. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  403. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  404. },
  405. { // Int16
  406. { VK_FORMAT_R16_SINT, VK_FORMAT_R16_SNORM },
  407. { VK_FORMAT_R16G16_SINT, VK_FORMAT_R16G16_SNORM },
  408. { VK_FORMAT_R16G16B16_SINT, VK_FORMAT_R16G16B16_SNORM },
  409. { VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_R16G16B16A16_SNORM },
  410. },
  411. { // Half
  412. { VK_FORMAT_R16_SFLOAT, VK_FORMAT_R16_SFLOAT },
  413. { VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_R16G16_SFLOAT },
  414. { VK_FORMAT_R16G16B16_SFLOAT, VK_FORMAT_R16G16B16_SFLOAT },
  415. { VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_R16G16B16A16_SFLOAT },
  416. },
  417. { // Float
  418. { VK_FORMAT_R32_SFLOAT, VK_FORMAT_R32_SFLOAT },
  419. { VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_R32G32_SFLOAT },
  420. { VK_FORMAT_R32G32B32_SFLOAT, VK_FORMAT_R32G32B32_SFLOAT },
  421. { VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_R32G32B32A32_SFLOAT },
  422. },
  423. };
  424. BX_STATIC_ASSERT(AttribType::Count == BX_COUNTOF(s_attribType) );
  425. void fillVertexLayout(const ShaderVK* _vsh, VkPipelineVertexInputStateCreateInfo& _vertexInputState, const VertexLayout& _layout)
  426. {
  427. uint32_t numBindings = _vertexInputState.vertexBindingDescriptionCount;
  428. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  429. VkVertexInputBindingDescription* inputBinding = const_cast<VkVertexInputBindingDescription*>(_vertexInputState.pVertexBindingDescriptions + numBindings);
  430. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  431. inputBinding->binding = numBindings;
  432. inputBinding->stride = _layout.m_stride;
  433. inputBinding->inputRate = VK_VERTEX_INPUT_RATE_VERTEX;
  434. for (uint32_t attr = 0; attr < Attrib::Count; ++attr)
  435. {
  436. if (UINT16_MAX != _layout.m_attributes[attr])
  437. {
  438. inputAttrib->location = _vsh->m_attrRemap[attr];
  439. inputAttrib->binding = numBindings;
  440. uint8_t num;
  441. AttribType::Enum type;
  442. bool normalized;
  443. bool asInt;
  444. _layout.decode(Attrib::Enum(attr), num, type, normalized, asInt);
  445. inputAttrib->format = s_attribType[type][num-1][normalized];
  446. inputAttrib->offset = _layout.m_offset[attr];
  447. ++inputAttrib;
  448. ++numAttribs;
  449. }
  450. }
  451. _vertexInputState.vertexBindingDescriptionCount = numBindings + 1;
  452. _vertexInputState.vertexAttributeDescriptionCount = numAttribs;
  453. }
  454. void fillInstanceBinding(const ShaderVK* _vsh, VkPipelineVertexInputStateCreateInfo& _vertexInputState, uint32_t _numInstanceData)
  455. {
  456. BX_UNUSED(_vsh);
  457. uint32_t numBindings = _vertexInputState.vertexBindingDescriptionCount;
  458. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  459. VkVertexInputBindingDescription* inputBinding = const_cast<VkVertexInputBindingDescription*>(_vertexInputState.pVertexBindingDescriptions + numBindings);
  460. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  461. inputBinding->binding = numBindings;
  462. inputBinding->stride = _numInstanceData * 16;
  463. inputBinding->inputRate = VK_VERTEX_INPUT_RATE_INSTANCE;
  464. for (uint32_t inst = 0; inst < _numInstanceData; ++inst)
  465. {
  466. inputAttrib->location = numAttribs;
  467. inputAttrib->binding = numBindings;
  468. inputAttrib->format = VK_FORMAT_R32G32B32A32_SFLOAT;
  469. inputAttrib->offset = inst * 16;
  470. ++numAttribs;
  471. ++inputAttrib;
  472. }
  473. _vertexInputState.vertexBindingDescriptionCount = numBindings + 1;
  474. _vertexInputState.vertexAttributeDescriptionCount = numAttribs;
  475. }
  476. static const char* s_deviceTypeName[] =
  477. {
  478. "Other",
  479. "Integrated GPU",
  480. "Discrete GPU",
  481. "Virtual GPU",
  482. "CPU",
  483. "Unknown?!"
  484. };
  485. const char* getName(VkPhysicalDeviceType _type)
  486. {
  487. return s_deviceTypeName[bx::min<int32_t>(_type, BX_COUNTOF(s_deviceTypeName)-1 )];
  488. }
  489. static const char* s_allocScopeName[] =
  490. {
  491. "vkCommand",
  492. "vkObject",
  493. "vkCache",
  494. "vkDevice",
  495. "vkInstance",
  496. };
  497. BX_STATIC_ASSERT(VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE == BX_COUNTOF(s_allocScopeName)-1);
  498. constexpr size_t kMinAlignment = 16;
  499. static void* VKAPI_PTR allocationFunction(void* _userData, size_t _size, size_t _alignment, VkSystemAllocationScope _allocationScope)
  500. {
  501. bx::AllocatorI* allocator = (bx::AllocatorI*)_userData;
  502. return bx::alignedAlloc(allocator, _size, bx::max(kMinAlignment, _alignment), bx::Location(s_allocScopeName[_allocationScope], 0) );
  503. }
  504. static void* VKAPI_PTR reallocationFunction(void* _userData, void* _ptr, size_t _size, size_t _alignment, VkSystemAllocationScope _allocationScope)
  505. {
  506. bx::AllocatorI* allocator = (bx::AllocatorI*)_userData;
  507. BX_UNUSED(_userData);
  508. if (0 == _size)
  509. {
  510. bx::alignedFree(allocator, _ptr, 0);
  511. return NULL;
  512. }
  513. return bx::alignedRealloc(allocator, _ptr, _size, bx::max(kMinAlignment, _alignment), bx::Location(s_allocScopeName[_allocationScope], 0) );
  514. }
  515. static void VKAPI_PTR freeFunction(void* _userData, void* _ptr)
  516. {
  517. if (NULL == _ptr)
  518. {
  519. return;
  520. }
  521. bx::AllocatorI* allocator = (bx::AllocatorI*)_userData;
  522. bx::alignedFree(allocator, _ptr, kMinAlignment);
  523. }
  524. static void VKAPI_PTR internalAllocationNotification(void* _userData, size_t _size, VkInternalAllocationType _allocationType, VkSystemAllocationScope _allocationScope)
  525. {
  526. BX_UNUSED(_userData, _size, _allocationType, _allocationScope);
  527. }
  528. static void VKAPI_PTR internalFreeNotification(void* _userData, size_t _size, VkInternalAllocationType _allocationType, VkSystemAllocationScope _allocationScope)
  529. {
  530. BX_UNUSED(_userData, _size, _allocationType, _allocationScope);
  531. }
  532. static VkAllocationCallbacks s_allocationCb =
  533. {
  534. NULL,
  535. allocationFunction,
  536. reallocationFunction,
  537. freeFunction,
  538. internalAllocationNotification,
  539. internalFreeNotification,
  540. };
  541. VkResult VKAPI_PTR stubSetDebugUtilsObjectNameEXT(VkDevice _device, const VkDebugUtilsObjectNameInfoEXT* _nameInfo)
  542. {
  543. BX_UNUSED(_device, _nameInfo);
  544. return VK_SUCCESS;
  545. }
  546. void VKAPI_PTR stubCmdInsertDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer, const VkDebugUtilsLabelEXT* _labelInfo)
  547. {
  548. BX_UNUSED(_commandBuffer, _labelInfo);
  549. }
  550. void VKAPI_PTR stubCmdBeginDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer, const VkDebugUtilsLabelEXT* _labelInfo)
  551. {
  552. BX_UNUSED(_commandBuffer, _labelInfo);
  553. }
  554. void VKAPI_PTR stubCmdEndDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer)
  555. {
  556. BX_UNUSED(_commandBuffer);
  557. }
  558. static const char* s_debugReportObjectType[] =
  559. {
  560. "Unknown",
  561. "Instance",
  562. "PhysicalDevice",
  563. "Device",
  564. "Queue",
  565. "Semaphore",
  566. "CommandBuffer",
  567. "Fence",
  568. "DeviceMemory",
  569. "Buffer",
  570. "Image",
  571. "Event",
  572. "QueryPool",
  573. "BufferView",
  574. "ImageView",
  575. "ShaderModule",
  576. "PipelineCache",
  577. "PipelineLayout",
  578. "RenderPass",
  579. "Pipeline",
  580. "DescriptorSetLayout",
  581. "Sampler",
  582. "DescriptorPool",
  583. "DescriptorSet",
  584. "Framebuffer",
  585. "CommandPool",
  586. "SurfaceKHR",
  587. "SwapchainKHR",
  588. "DebugReport",
  589. };
  590. VkBool32 VKAPI_PTR debugReportCb(
  591. VkDebugReportFlagsEXT _flags
  592. , VkDebugReportObjectTypeEXT _objectType
  593. , uint64_t _object
  594. , size_t _location
  595. , int32_t _messageCode
  596. , const char* _layerPrefix
  597. , const char* _message
  598. , void* _userData
  599. )
  600. {
  601. BX_UNUSED(_flags, _objectType, _object, _location, _messageCode, _layerPrefix, _message, _userData, s_debugReportObjectType);
  602. // For more info about 'VUID-VkSwapchainCreateInfoKHR-imageExtent-01274'
  603. // check https://github.com/KhronosGroup/Vulkan-Docs/issues/1144
  604. if (!bx::strFind(_message, "PointSizeMissing").isEmpty()
  605. || !bx::strFind(_message, "SwapchainTooManyImages").isEmpty()
  606. || !bx::strFind(_message, "SwapchainImageNotAcquired").isEmpty()
  607. || !bx::strFind(_message, "VUID-VkSwapchainCreateInfoKHR-imageExtent-01274").isEmpty() )
  608. {
  609. return VK_FALSE;
  610. }
  611. BX_TRACE("%c%c%c%c%c %19s, %s, %d: %s"
  612. , 0 != (_flags & VK_DEBUG_REPORT_INFORMATION_BIT_EXT ) ? 'I' : '-'
  613. , 0 != (_flags & VK_DEBUG_REPORT_WARNING_BIT_EXT ) ? 'W' : '-'
  614. , 0 != (_flags & VK_DEBUG_REPORT_PERFORMANCE_WARNING_BIT_EXT) ? 'P' : '-'
  615. , 0 != (_flags & VK_DEBUG_REPORT_ERROR_BIT_EXT ) ? 'E' : '-'
  616. , 0 != (_flags & VK_DEBUG_REPORT_DEBUG_BIT_EXT ) ? 'D' : '-'
  617. , s_debugReportObjectType[_objectType]
  618. , _layerPrefix
  619. , _messageCode
  620. , _message
  621. );
  622. return VK_FALSE;
  623. }
  624. VkResult enumerateLayerProperties(VkPhysicalDevice _physicalDevice, uint32_t* _propertyCount, VkLayerProperties* _properties)
  625. {
  626. return (VK_NULL_HANDLE == _physicalDevice)
  627. ? vkEnumerateInstanceLayerProperties(_propertyCount, _properties)
  628. : vkEnumerateDeviceLayerProperties(_physicalDevice, _propertyCount, _properties)
  629. ;
  630. }
  631. VkResult enumerateExtensionProperties(VkPhysicalDevice _physicalDevice, const char* _layerName, uint32_t* _propertyCount, VkExtensionProperties* _properties)
  632. {
  633. return (VK_NULL_HANDLE == _physicalDevice)
  634. ? vkEnumerateInstanceExtensionProperties(_layerName, _propertyCount, _properties)
  635. : vkEnumerateDeviceExtensionProperties(_physicalDevice, _layerName, _propertyCount, _properties)
  636. ;
  637. }
  638. void dumpExtensions(VkPhysicalDevice _physicalDevice, Extension _extensions[Extension::Count])
  639. {
  640. { // Global extensions.
  641. uint32_t numExtensionProperties;
  642. VkResult result = enumerateExtensionProperties(_physicalDevice
  643. , NULL
  644. , &numExtensionProperties
  645. , NULL
  646. );
  647. if (VK_SUCCESS == result
  648. && 0 < numExtensionProperties)
  649. {
  650. VkExtensionProperties* extensionProperties = (VkExtensionProperties*)bx::alloc(g_allocator, numExtensionProperties * sizeof(VkExtensionProperties) );
  651. result = enumerateExtensionProperties(_physicalDevice
  652. , NULL
  653. , &numExtensionProperties
  654. , extensionProperties
  655. );
  656. BX_TRACE("Global extensions (%d):"
  657. , numExtensionProperties
  658. );
  659. for (uint32_t extension = 0; extension < numExtensionProperties; ++extension)
  660. {
  661. bool supported = updateExtension(
  662. extensionProperties[extension].extensionName
  663. , extensionProperties[extension].specVersion
  664. , VK_NULL_HANDLE == _physicalDevice
  665. , _extensions
  666. );
  667. BX_TRACE("\tv%-3d %s%s"
  668. , extensionProperties[extension].specVersion
  669. , extensionProperties[extension].extensionName
  670. , supported ? " (supported)" : "", extensionProperties[extension].extensionName
  671. );
  672. BX_UNUSED(supported);
  673. }
  674. bx::free(g_allocator, extensionProperties);
  675. }
  676. }
  677. // Layer extensions.
  678. uint32_t numLayerProperties;
  679. VkResult result = enumerateLayerProperties(_physicalDevice, &numLayerProperties, NULL);
  680. if (VK_SUCCESS == result
  681. && 0 < numLayerProperties)
  682. {
  683. VkLayerProperties* layerProperties = (VkLayerProperties*)bx::alloc(g_allocator, numLayerProperties * sizeof(VkLayerProperties) );
  684. result = enumerateLayerProperties(_physicalDevice, &numLayerProperties, layerProperties);
  685. char indent = VK_NULL_HANDLE == _physicalDevice ? '\0' : '\t';
  686. BX_UNUSED(indent);
  687. BX_TRACE("%cLayer extensions (%d):"
  688. , indent
  689. , numLayerProperties
  690. );
  691. for (uint32_t layer = 0; layer < numLayerProperties; ++layer)
  692. {
  693. updateLayer(
  694. layerProperties[layer].layerName
  695. , layerProperties[layer].implementationVersion
  696. , VK_NULL_HANDLE == _physicalDevice
  697. );
  698. BX_TRACE("%c\t%s (s: 0x%08x, i: 0x%08x), %s"
  699. , indent
  700. , layerProperties[layer].layerName
  701. , layerProperties[layer].specVersion
  702. , layerProperties[layer].implementationVersion
  703. , layerProperties[layer].description
  704. );
  705. uint32_t numExtensionProperties;
  706. result = enumerateExtensionProperties(_physicalDevice
  707. , layerProperties[layer].layerName
  708. , &numExtensionProperties
  709. , NULL
  710. );
  711. if (VK_SUCCESS == result
  712. && 0 < numExtensionProperties)
  713. {
  714. VkExtensionProperties* extensionProperties = (VkExtensionProperties*)bx::alloc(g_allocator, numExtensionProperties * sizeof(VkExtensionProperties) );
  715. result = enumerateExtensionProperties(_physicalDevice
  716. , layerProperties[layer].layerName
  717. , &numExtensionProperties
  718. , extensionProperties
  719. );
  720. for (uint32_t extension = 0; extension < numExtensionProperties; ++extension)
  721. {
  722. const bool supported = updateExtension(
  723. extensionProperties[extension].extensionName
  724. , extensionProperties[extension].specVersion
  725. , VK_NULL_HANDLE == _physicalDevice
  726. , _extensions
  727. );
  728. BX_TRACE("%c\t\t%s (s: 0x%08x)"
  729. , indent
  730. , extensionProperties[extension].extensionName
  731. , extensionProperties[extension].specVersion
  732. , supported ? " (supported)" : "", extensionProperties[extension].extensionName
  733. );
  734. BX_UNUSED(supported);
  735. }
  736. bx::free(g_allocator, extensionProperties);
  737. }
  738. }
  739. bx::free(g_allocator, layerProperties);
  740. }
  741. }
  742. const char* getName(VkResult _result)
  743. {
  744. switch (_result)
  745. {
  746. #define VKENUM(_ty) case _ty: return #_ty
  747. VKENUM(VK_SUCCESS);
  748. VKENUM(VK_NOT_READY);
  749. VKENUM(VK_TIMEOUT);
  750. VKENUM(VK_EVENT_SET);
  751. VKENUM(VK_EVENT_RESET);
  752. VKENUM(VK_INCOMPLETE);
  753. VKENUM(VK_ERROR_OUT_OF_HOST_MEMORY);
  754. VKENUM(VK_ERROR_OUT_OF_DEVICE_MEMORY);
  755. VKENUM(VK_ERROR_OUT_OF_POOL_MEMORY);
  756. VKENUM(VK_ERROR_FRAGMENTED_POOL);
  757. VKENUM(VK_ERROR_INITIALIZATION_FAILED);
  758. VKENUM(VK_ERROR_DEVICE_LOST);
  759. VKENUM(VK_ERROR_MEMORY_MAP_FAILED);
  760. VKENUM(VK_ERROR_LAYER_NOT_PRESENT);
  761. VKENUM(VK_ERROR_EXTENSION_NOT_PRESENT);
  762. VKENUM(VK_ERROR_FEATURE_NOT_PRESENT);
  763. VKENUM(VK_ERROR_INCOMPATIBLE_DRIVER);
  764. VKENUM(VK_ERROR_TOO_MANY_OBJECTS);
  765. VKENUM(VK_ERROR_FORMAT_NOT_SUPPORTED);
  766. VKENUM(VK_ERROR_SURFACE_LOST_KHR);
  767. VKENUM(VK_ERROR_NATIVE_WINDOW_IN_USE_KHR);
  768. VKENUM(VK_SUBOPTIMAL_KHR);
  769. VKENUM(VK_ERROR_OUT_OF_DATE_KHR);
  770. VKENUM(VK_ERROR_INCOMPATIBLE_DISPLAY_KHR);
  771. VKENUM(VK_ERROR_VALIDATION_FAILED_EXT);
  772. #undef VKENUM
  773. default: break;
  774. }
  775. BX_WARN(false, "Unknown VkResult? %x", _result);
  776. return "<VkResult?>";
  777. }
  778. template<typename Ty>
  779. constexpr VkObjectType getType();
  780. template<> VkObjectType getType<VkBuffer >() { return VK_OBJECT_TYPE_BUFFER; }
  781. template<> VkObjectType getType<VkCommandPool >() { return VK_OBJECT_TYPE_COMMAND_POOL; }
  782. template<> VkObjectType getType<VkDescriptorPool >() { return VK_OBJECT_TYPE_DESCRIPTOR_POOL; }
  783. template<> VkObjectType getType<VkDescriptorSet >() { return VK_OBJECT_TYPE_DESCRIPTOR_SET; }
  784. template<> VkObjectType getType<VkDescriptorSetLayout>() { return VK_OBJECT_TYPE_DESCRIPTOR_SET_LAYOUT; }
  785. template<> VkObjectType getType<VkDeviceMemory >() { return VK_OBJECT_TYPE_DEVICE_MEMORY; }
  786. template<> VkObjectType getType<VkFence >() { return VK_OBJECT_TYPE_FENCE; }
  787. template<> VkObjectType getType<VkFramebuffer >() { return VK_OBJECT_TYPE_FRAMEBUFFER; }
  788. template<> VkObjectType getType<VkImage >() { return VK_OBJECT_TYPE_IMAGE; }
  789. template<> VkObjectType getType<VkImageView >() { return VK_OBJECT_TYPE_IMAGE_VIEW; }
  790. template<> VkObjectType getType<VkPipeline >() { return VK_OBJECT_TYPE_PIPELINE; }
  791. template<> VkObjectType getType<VkPipelineCache >() { return VK_OBJECT_TYPE_PIPELINE_CACHE; }
  792. template<> VkObjectType getType<VkPipelineLayout >() { return VK_OBJECT_TYPE_PIPELINE_LAYOUT; }
  793. template<> VkObjectType getType<VkQueryPool >() { return VK_OBJECT_TYPE_QUERY_POOL; }
  794. template<> VkObjectType getType<VkRenderPass >() { return VK_OBJECT_TYPE_RENDER_PASS; }
  795. template<> VkObjectType getType<VkSampler >() { return VK_OBJECT_TYPE_SAMPLER; }
  796. template<> VkObjectType getType<VkSemaphore >() { return VK_OBJECT_TYPE_SEMAPHORE; }
  797. template<> VkObjectType getType<VkShaderModule >() { return VK_OBJECT_TYPE_SHADER_MODULE; }
  798. template<> VkObjectType getType<VkSurfaceKHR >() { return VK_OBJECT_TYPE_SURFACE_KHR; }
  799. template<> VkObjectType getType<VkSwapchainKHR >() { return VK_OBJECT_TYPE_SWAPCHAIN_KHR; }
  800. template<typename Ty>
  801. static BX_NO_INLINE void setDebugObjectName(VkDevice _device, Ty _object, const char* _format, ...)
  802. {
  803. if (BX_ENABLED(BGFX_CONFIG_DEBUG_OBJECT_NAME)
  804. && s_extension[Extension::EXT_debug_utils].m_supported)
  805. {
  806. char temp[2048];
  807. va_list argList;
  808. va_start(argList, _format);
  809. int32_t size = bx::min<int32_t>(sizeof(temp)-1, bx::vsnprintf(temp, sizeof(temp), _format, argList) );
  810. va_end(argList);
  811. temp[size] = '\0';
  812. VkDebugUtilsObjectNameInfoEXT ni;
  813. ni.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT;
  814. ni.pNext = NULL;
  815. ni.objectType = getType<Ty>();
  816. ni.objectHandle = uint64_t(_object.vk);
  817. ni.pObjectName = temp;
  818. VK_CHECK(vkSetDebugUtilsObjectNameEXT(_device, &ni) );
  819. }
  820. }
  821. void setMemoryBarrier(
  822. VkCommandBuffer _commandBuffer
  823. , VkPipelineStageFlags _srcStages
  824. , VkPipelineStageFlags _dstStages
  825. )
  826. {
  827. VkMemoryBarrier mb;
  828. mb.sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER;
  829. mb.pNext = NULL;
  830. mb.srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  831. mb.dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  832. vkCmdPipelineBarrier(
  833. _commandBuffer
  834. , _srcStages
  835. , _dstStages
  836. , 0
  837. , 1
  838. , &mb
  839. , 0
  840. , NULL
  841. , 0
  842. , NULL
  843. );
  844. }
  845. void setImageMemoryBarrier(
  846. VkCommandBuffer _commandBuffer
  847. , VkImage _image
  848. , VkImageAspectFlags _aspectMask
  849. , VkImageLayout _oldLayout
  850. , VkImageLayout _newLayout
  851. , uint32_t _baseMipLevel = 0
  852. , uint32_t _levelCount = VK_REMAINING_MIP_LEVELS
  853. , uint32_t _baseArrayLayer = 0
  854. , uint32_t _layerCount = VK_REMAINING_ARRAY_LAYERS
  855. )
  856. {
  857. BX_ASSERT(true
  858. && _newLayout != VK_IMAGE_LAYOUT_UNDEFINED
  859. && _newLayout != VK_IMAGE_LAYOUT_PREINITIALIZED
  860. , "_newLayout cannot use VK_IMAGE_LAYOUT_UNDEFINED or VK_IMAGE_LAYOUT_PREINITIALIZED."
  861. );
  862. constexpr VkPipelineStageFlags depthStageMask = 0
  863. | VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT
  864. | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT
  865. ;
  866. constexpr VkPipelineStageFlags sampledStageMask = 0
  867. | VK_PIPELINE_STAGE_VERTEX_SHADER_BIT
  868. | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT
  869. | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  870. ;
  871. VkPipelineStageFlags srcStageMask = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  872. VkPipelineStageFlags dstStageMask = VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT;
  873. VkAccessFlags srcAccessMask = 0;
  874. VkAccessFlags dstAccessMask = 0;
  875. switch (_oldLayout)
  876. {
  877. case VK_IMAGE_LAYOUT_UNDEFINED:
  878. break;
  879. case VK_IMAGE_LAYOUT_GENERAL:
  880. srcStageMask = VK_PIPELINE_STAGE_ALL_COMMANDS_BIT;
  881. srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  882. break;
  883. case VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL:
  884. srcStageMask = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  885. srcAccessMask = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
  886. break;
  887. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL:
  888. srcStageMask = depthStageMask;
  889. srcAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
  890. break;
  891. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL:
  892. srcStageMask = depthStageMask | sampledStageMask;
  893. break;
  894. case VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL:
  895. srcStageMask = sampledStageMask;
  896. break;
  897. case VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL:
  898. srcStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  899. break;
  900. case VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL:
  901. srcStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  902. srcAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
  903. break;
  904. case VK_IMAGE_LAYOUT_PREINITIALIZED:
  905. srcStageMask = VK_PIPELINE_STAGE_HOST_BIT;
  906. srcAccessMask = VK_ACCESS_HOST_WRITE_BIT;
  907. break;
  908. case VK_IMAGE_LAYOUT_PRESENT_SRC_KHR:
  909. break;
  910. default:
  911. BX_ASSERT(false, "Unknown image layout.");
  912. break;
  913. }
  914. switch (_newLayout)
  915. {
  916. case VK_IMAGE_LAYOUT_GENERAL:
  917. dstStageMask = VK_PIPELINE_STAGE_ALL_COMMANDS_BIT;
  918. dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  919. break;
  920. case VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL:
  921. dstStageMask = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  922. dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
  923. break;
  924. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL:
  925. dstStageMask = depthStageMask;
  926. dstAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
  927. break;
  928. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL:
  929. dstStageMask = depthStageMask | sampledStageMask;
  930. dstAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_INPUT_ATTACHMENT_READ_BIT;
  931. break;
  932. case VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL:
  933. dstStageMask = sampledStageMask;
  934. dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_INPUT_ATTACHMENT_READ_BIT;
  935. break;
  936. case VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL:
  937. dstStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  938. dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
  939. break;
  940. case VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL:
  941. dstStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  942. dstAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
  943. break;
  944. case VK_IMAGE_LAYOUT_PRESENT_SRC_KHR:
  945. // vkQueuePresentKHR performs automatic visibility operations
  946. break;
  947. default:
  948. BX_ASSERT(false, "Unknown image layout.");
  949. break;
  950. }
  951. VkImageMemoryBarrier imb;
  952. imb.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  953. imb.pNext = NULL;
  954. imb.srcAccessMask = srcAccessMask;
  955. imb.dstAccessMask = dstAccessMask;
  956. imb.oldLayout = _oldLayout;
  957. imb.newLayout = _newLayout;
  958. imb.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  959. imb.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  960. imb.image = _image;
  961. imb.subresourceRange.aspectMask = _aspectMask;
  962. imb.subresourceRange.baseMipLevel = _baseMipLevel;
  963. imb.subresourceRange.levelCount = _levelCount;
  964. imb.subresourceRange.baseArrayLayer = _baseArrayLayer;
  965. imb.subresourceRange.layerCount = _layerCount;
  966. vkCmdPipelineBarrier(
  967. _commandBuffer
  968. , srcStageMask
  969. , dstStageMask
  970. , 0
  971. , 0
  972. , NULL
  973. , 0
  974. , NULL
  975. , 1
  976. , &imb
  977. );
  978. }
  979. #define MAX_DESCRIPTOR_SETS (1024 * BGFX_CONFIG_MAX_FRAME_LATENCY)
  980. struct RendererContextVK : public RendererContextI
  981. {
  982. RendererContextVK()
  983. : m_allocatorCb(NULL)
  984. , m_renderDocDll(NULL)
  985. , m_vulkan1Dll(NULL)
  986. , m_maxAnisotropy(1.0f)
  987. , m_depthClamp(false)
  988. , m_wireframe(false)
  989. , m_captureBuffer(VK_NULL_HANDLE)
  990. , m_captureMemory(VK_NULL_HANDLE)
  991. , m_captureSize(0)
  992. {
  993. }
  994. ~RendererContextVK()
  995. {
  996. }
  997. bool init(const Init& _init)
  998. {
  999. struct ErrorState
  1000. {
  1001. enum Enum
  1002. {
  1003. Default,
  1004. LoadedVulkan1,
  1005. InstanceCreated,
  1006. DeviceCreated,
  1007. CommandQueueCreated,
  1008. SwapChainCreated,
  1009. DescriptorCreated,
  1010. TimerQueryCreated,
  1011. };
  1012. };
  1013. ErrorState::Enum errorState = ErrorState::Default;
  1014. const bool headless = NULL == g_platformData.nwh;
  1015. const void* nextFeatures = NULL;
  1016. VkPhysicalDeviceLineRasterizationFeaturesEXT lineRasterizationFeatures;
  1017. VkPhysicalDeviceCustomBorderColorFeaturesEXT customBorderColorFeatures;
  1018. bx::memSet(&lineRasterizationFeatures, 0, sizeof(lineRasterizationFeatures) );
  1019. bx::memSet(&customBorderColorFeatures, 0, sizeof(customBorderColorFeatures) );
  1020. m_fbh.idx = kInvalidHandle;
  1021. bx::memSet(m_uniforms, 0, sizeof(m_uniforms) );
  1022. bx::memSet(&m_resolution, 0, sizeof(m_resolution) );
  1023. bool imported = true;
  1024. VkResult result;
  1025. m_globalQueueFamily = UINT32_MAX;
  1026. if (_init.debug
  1027. || _init.profile)
  1028. {
  1029. m_renderDocDll = loadRenderDoc();
  1030. }
  1031. setGraphicsDebuggerPresent(NULL != m_renderDocDll);
  1032. m_vulkan1Dll = bx::dlopen(
  1033. #if BX_PLATFORM_WINDOWS
  1034. "vulkan-1.dll"
  1035. #elif BX_PLATFORM_ANDROID
  1036. "libvulkan.so"
  1037. #elif BX_PLATFORM_OSX
  1038. "libMoltenVK.dylib"
  1039. #else
  1040. "libvulkan.so.1"
  1041. #endif // BX_PLATFORM_*
  1042. );
  1043. if (NULL == m_vulkan1Dll)
  1044. {
  1045. BX_TRACE("Init error: Failed to load vulkan dynamic library.");
  1046. goto error;
  1047. }
  1048. errorState = ErrorState::LoadedVulkan1;
  1049. BX_TRACE("Shared library functions:");
  1050. #define VK_IMPORT_FUNC(_optional, _func) \
  1051. _func = (PFN_##_func)bx::dlsym(m_vulkan1Dll, #_func); \
  1052. BX_TRACE("\t%p " #_func, _func); \
  1053. imported &= _optional || NULL != _func
  1054. VK_IMPORT
  1055. #undef VK_IMPORT_FUNC
  1056. if (!imported)
  1057. {
  1058. BX_TRACE("Init error: Failed to load shared library functions.");
  1059. goto error;
  1060. }
  1061. {
  1062. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_device.m_initialize = _init.debug;
  1063. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_instance.m_initialize = _init.debug;
  1064. s_layer[Layer::VK_LAYER_KHRONOS_validation ].m_device.m_initialize = _init.debug;
  1065. s_layer[Layer::VK_LAYER_KHRONOS_validation ].m_instance.m_initialize = _init.debug;
  1066. s_extension[Extension::EXT_debug_report].m_initialize = _init.debug;
  1067. s_extension[Extension::EXT_shader_viewport_index_layer].m_initialize = !!(_init.capabilities & BGFX_CAPS_VIEWPORT_LAYER_ARRAY);
  1068. s_extension[Extension::EXT_conservative_rasterization ].m_initialize = !!(_init.capabilities & BGFX_CAPS_CONSERVATIVE_RASTER );
  1069. s_extension[Extension::KHR_draw_indirect_count ].m_initialize = !!(_init.capabilities & BGFX_CAPS_DRAW_INDIRECT_COUNT );
  1070. dumpExtensions(VK_NULL_HANDLE, s_extension);
  1071. if (s_layer[Layer::VK_LAYER_KHRONOS_validation].m_device.m_supported
  1072. || s_layer[Layer::VK_LAYER_KHRONOS_validation].m_instance.m_supported)
  1073. {
  1074. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_device.m_supported = false;
  1075. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_instance.m_supported = false;
  1076. }
  1077. uint32_t numEnabledLayers = 0;
  1078. const char* enabledLayer[Layer::Count];
  1079. BX_TRACE("Enabled instance layers:");
  1080. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  1081. {
  1082. const Layer& layer = s_layer[ii];
  1083. if (layer.m_instance.m_supported
  1084. && layer.m_instance.m_initialize)
  1085. {
  1086. enabledLayer[numEnabledLayers++] = layer.m_name;
  1087. BX_TRACE("\t%s", layer.m_name);
  1088. }
  1089. }
  1090. uint32_t numEnabledExtensions = 0;
  1091. const char* enabledExtension[Extension::Count + 2];
  1092. if (!headless)
  1093. {
  1094. enabledExtension[numEnabledExtensions++] = VK_KHR_SURFACE_EXTENSION_NAME;
  1095. enabledExtension[numEnabledExtensions++] = KHR_SURFACE_EXTENSION_NAME;
  1096. }
  1097. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  1098. {
  1099. const Extension& extension = s_extension[ii];
  1100. const LayerInfo& layerInfo = s_layer[extension.m_layer].m_instance;
  1101. const bool layerEnabled = false
  1102. || extension.m_layer == Layer::Count || (layerInfo.m_supported && layerInfo.m_initialize)
  1103. ;
  1104. if (extension.m_supported
  1105. && extension.m_initialize
  1106. && extension.m_instanceExt
  1107. && layerEnabled)
  1108. {
  1109. enabledExtension[numEnabledExtensions++] = extension.m_name;
  1110. }
  1111. }
  1112. BX_TRACE("Enabled instance extensions:");
  1113. for (uint32_t ii = 0; ii < numEnabledExtensions; ++ii)
  1114. {
  1115. BX_TRACE("\t%s", enabledExtension[ii]);
  1116. }
  1117. uint32_t vulkanApiVersionSelector;
  1118. if (NULL != vkEnumerateInstanceVersion)
  1119. {
  1120. result = vkEnumerateInstanceVersion(&vulkanApiVersionSelector);
  1121. if (VK_SUCCESS != result)
  1122. {
  1123. BX_TRACE(
  1124. "Init error: vkEnumerateInstanceVersion failed %d: %s."
  1125. , result
  1126. , getName(result)
  1127. );
  1128. goto error;
  1129. }
  1130. }
  1131. else
  1132. {
  1133. vulkanApiVersionSelector = VK_API_VERSION_1_0;
  1134. }
  1135. VkApplicationInfo appInfo;
  1136. appInfo.sType = VK_STRUCTURE_TYPE_APPLICATION_INFO;
  1137. appInfo.pNext = NULL;
  1138. appInfo.pApplicationName = "bgfx";
  1139. appInfo.applicationVersion = BGFX_API_VERSION;
  1140. appInfo.pEngineName = "bgfx";
  1141. appInfo.engineVersion = BGFX_API_VERSION;
  1142. appInfo.apiVersion = vulkanApiVersionSelector;
  1143. VkInstanceCreateInfo ici;
  1144. ici.sType = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO;
  1145. ici.pNext = NULL;
  1146. ici.flags = 0
  1147. | (BX_ENABLED(BX_PLATFORM_OSX) ? VK_INSTANCE_CREATE_ENUMERATE_PORTABILITY_BIT_KHR : 0)
  1148. ;
  1149. ici.pApplicationInfo = &appInfo;
  1150. ici.enabledLayerCount = numEnabledLayers;
  1151. ici.ppEnabledLayerNames = enabledLayer;
  1152. ici.enabledExtensionCount = numEnabledExtensions;
  1153. ici.ppEnabledExtensionNames = enabledExtension;
  1154. if (BX_ENABLED(BGFX_CONFIG_DEBUG) )
  1155. {
  1156. s_allocationCb.pUserData = g_allocator;
  1157. m_allocatorCb = &s_allocationCb;
  1158. BX_UNUSED(s_allocationCb);
  1159. }
  1160. result = vkCreateInstance(
  1161. &ici
  1162. , m_allocatorCb
  1163. , &m_instance
  1164. );
  1165. if (VK_SUCCESS != result)
  1166. {
  1167. BX_TRACE("Init error: vkCreateInstance failed %d: %s.", result, getName(result) );
  1168. goto error;
  1169. }
  1170. m_instanceApiVersion = vulkanApiVersionSelector;
  1171. BX_TRACE("Instance API version: %d.%d.%d"
  1172. , VK_API_VERSION_MAJOR(m_instanceApiVersion)
  1173. , VK_API_VERSION_MINOR(m_instanceApiVersion)
  1174. , VK_API_VERSION_PATCH(m_instanceApiVersion)
  1175. );
  1176. BX_TRACE("Instance variant: %d", VK_API_VERSION_VARIANT(m_instanceApiVersion) );
  1177. }
  1178. errorState = ErrorState::InstanceCreated;
  1179. BX_TRACE("Instance functions:");
  1180. #define VK_IMPORT_INSTANCE_FUNC(_optional, _func) \
  1181. _func = (PFN_##_func)vkGetInstanceProcAddr(m_instance, #_func); \
  1182. BX_TRACE("\t%p " #_func, _func); \
  1183. imported &= _optional || NULL != _func
  1184. VK_IMPORT_INSTANCE
  1185. #undef VK_IMPORT_INSTANCE_FUNC
  1186. if (!imported)
  1187. {
  1188. BX_TRACE("Init error: Failed to load instance functions.");
  1189. goto error;
  1190. }
  1191. m_debugReportCallback = VK_NULL_HANDLE;
  1192. if (s_extension[Extension::EXT_debug_report].m_supported)
  1193. {
  1194. VkDebugReportCallbackCreateInfoEXT drcb;
  1195. drcb.sType = VK_STRUCTURE_TYPE_DEBUG_REPORT_CREATE_INFO_EXT;
  1196. drcb.pNext = NULL;
  1197. drcb.pfnCallback = debugReportCb;
  1198. drcb.pUserData = NULL;
  1199. drcb.flags = 0
  1200. | VK_DEBUG_REPORT_ERROR_BIT_EXT
  1201. | VK_DEBUG_REPORT_WARNING_BIT_EXT
  1202. ;
  1203. result = vkCreateDebugReportCallbackEXT(m_instance
  1204. , &drcb
  1205. , m_allocatorCb
  1206. , &m_debugReportCallback
  1207. );
  1208. BX_WARN(VK_SUCCESS == result, "vkCreateDebugReportCallbackEXT failed %d: %s.", result, getName(result) );
  1209. }
  1210. {
  1211. BX_TRACE("---");
  1212. uint32_t numPhysicalDevices;
  1213. result = vkEnumeratePhysicalDevices(m_instance
  1214. , &numPhysicalDevices
  1215. , NULL
  1216. );
  1217. if (VK_SUCCESS != result)
  1218. {
  1219. BX_TRACE("Init error: vkEnumeratePhysicalDevices failed %d: %s.", result, getName(result) );
  1220. goto error;
  1221. }
  1222. VkPhysicalDevice physicalDevices[4];
  1223. numPhysicalDevices = bx::min<uint32_t>(numPhysicalDevices, BX_COUNTOF(physicalDevices) );
  1224. result = vkEnumeratePhysicalDevices(m_instance
  1225. , &numPhysicalDevices
  1226. , physicalDevices
  1227. );
  1228. if (VK_SUCCESS != result)
  1229. {
  1230. BX_TRACE("Init error: vkEnumeratePhysicalDevices failed %d: %s.", result, getName(result) );
  1231. goto error;
  1232. }
  1233. Extension physicalDeviceExtensions[4][Extension::Count];
  1234. uint32_t physicalDeviceIdx = UINT32_MAX;
  1235. uint32_t fallbackPhysicalDeviceIdx = UINT32_MAX;
  1236. for (uint32_t ii = 0; ii < numPhysicalDevices; ++ii)
  1237. {
  1238. VkPhysicalDeviceProperties pdp;
  1239. vkGetPhysicalDeviceProperties(physicalDevices[ii], &pdp);
  1240. BX_TRACE("Physical device %d:", ii);
  1241. BX_TRACE("\t Name: %s", pdp.deviceName);
  1242. BX_TRACE("\t API version: %d.%d.%d"
  1243. , VK_API_VERSION_MAJOR(pdp.apiVersion)
  1244. , VK_API_VERSION_MINOR(pdp.apiVersion)
  1245. , VK_API_VERSION_PATCH(pdp.apiVersion)
  1246. );
  1247. BX_TRACE("\t API variant: %d", VK_API_VERSION_VARIANT(pdp.apiVersion) );
  1248. BX_TRACE("\tDriver version: %x", pdp.driverVersion);
  1249. BX_TRACE("\t VendorId: %x", pdp.vendorID);
  1250. BX_TRACE("\t DeviceId: %x", pdp.deviceID);
  1251. BX_TRACE("\t Type: %d", pdp.deviceType);
  1252. if (VK_PHYSICAL_DEVICE_TYPE_CPU == pdp.deviceType)
  1253. {
  1254. pdp.vendorID = BGFX_PCI_ID_SOFTWARE_RASTERIZER;
  1255. }
  1256. g_caps.gpu[ii].vendorId = uint16_t(pdp.vendorID);
  1257. g_caps.gpu[ii].deviceId = uint16_t(pdp.deviceID);
  1258. ++g_caps.numGPUs;
  1259. if ( (BGFX_PCI_ID_NONE != g_caps.vendorId || 0 != g_caps.deviceId)
  1260. && (BGFX_PCI_ID_NONE == g_caps.vendorId || pdp.vendorID == g_caps.vendorId)
  1261. && ( 0 == g_caps.deviceId || pdp.deviceID == g_caps.deviceId) )
  1262. {
  1263. if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU
  1264. || pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU)
  1265. {
  1266. fallbackPhysicalDeviceIdx = ii;
  1267. }
  1268. physicalDeviceIdx = ii;
  1269. }
  1270. else if (UINT32_MAX == physicalDeviceIdx)
  1271. {
  1272. if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU)
  1273. {
  1274. fallbackPhysicalDeviceIdx = ii;
  1275. }
  1276. else if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU)
  1277. {
  1278. physicalDeviceIdx = ii;
  1279. }
  1280. }
  1281. VkPhysicalDeviceMemoryProperties pdmp;
  1282. vkGetPhysicalDeviceMemoryProperties(physicalDevices[ii], &pdmp);
  1283. BX_TRACE("\tMemory type count: %d", pdmp.memoryTypeCount);
  1284. for (uint32_t jj = 0; jj < pdmp.memoryTypeCount; ++jj)
  1285. {
  1286. BX_TRACE("\t%3d: flags 0x%08x, index %d"
  1287. , jj
  1288. , pdmp.memoryTypes[jj].propertyFlags
  1289. , pdmp.memoryTypes[jj].heapIndex
  1290. );
  1291. }
  1292. BX_TRACE("\tMemory heap count: %d", pdmp.memoryHeapCount);
  1293. for (uint32_t jj = 0; jj < pdmp.memoryHeapCount; ++jj)
  1294. {
  1295. char size[16];
  1296. bx::prettify(size, BX_COUNTOF(size), pdmp.memoryHeaps[jj].size);
  1297. BX_TRACE("\t%3d: flags 0x%08x, size %10s"
  1298. , jj
  1299. , pdmp.memoryHeaps[jj].flags
  1300. , size
  1301. );
  1302. }
  1303. bx::memCopy(&physicalDeviceExtensions[ii][0], &s_extension[0], sizeof(s_extension) );
  1304. dumpExtensions(physicalDevices[ii], physicalDeviceExtensions[ii]);
  1305. }
  1306. if (UINT32_MAX == physicalDeviceIdx)
  1307. {
  1308. physicalDeviceIdx = UINT32_MAX == fallbackPhysicalDeviceIdx
  1309. ? 0
  1310. : fallbackPhysicalDeviceIdx
  1311. ;
  1312. }
  1313. m_physicalDevice = physicalDevices[physicalDeviceIdx];
  1314. bx::memCopy(&s_extension[0], &physicalDeviceExtensions[physicalDeviceIdx][0], sizeof(s_extension) );
  1315. vkGetPhysicalDeviceProperties(m_physicalDevice, &m_deviceProperties);
  1316. g_caps.vendorId = uint16_t(m_deviceProperties.vendorID);
  1317. g_caps.deviceId = uint16_t(m_deviceProperties.deviceID);
  1318. BX_TRACE("Using physical device %d: %s", physicalDeviceIdx, m_deviceProperties.deviceName);
  1319. VkPhysicalDeviceFeatures supportedFeatures;
  1320. if (s_extension[Extension::KHR_get_physical_device_properties2].m_supported)
  1321. {
  1322. VkPhysicalDeviceFeatures2KHR deviceFeatures2;
  1323. deviceFeatures2.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FEATURES_2_KHR;
  1324. deviceFeatures2.pNext = NULL;
  1325. VkBaseOutStructure* next = (VkBaseOutStructure*)&deviceFeatures2;
  1326. if (s_extension[Extension::EXT_line_rasterization].m_supported)
  1327. {
  1328. next->pNext = (VkBaseOutStructure*)&lineRasterizationFeatures;
  1329. next = (VkBaseOutStructure*)&lineRasterizationFeatures;
  1330. lineRasterizationFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_LINE_RASTERIZATION_FEATURES_EXT;
  1331. lineRasterizationFeatures.pNext = NULL;
  1332. }
  1333. if (s_extension[Extension::EXT_custom_border_color].m_supported)
  1334. {
  1335. next->pNext = (VkBaseOutStructure*)&customBorderColorFeatures;
  1336. next = (VkBaseOutStructure*)&customBorderColorFeatures;
  1337. customBorderColorFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_CUSTOM_BORDER_COLOR_FEATURES_EXT;
  1338. customBorderColorFeatures.pNext = NULL;
  1339. }
  1340. nextFeatures = deviceFeatures2.pNext;
  1341. vkGetPhysicalDeviceFeatures2KHR(m_physicalDevice, &deviceFeatures2);
  1342. supportedFeatures = deviceFeatures2.features;
  1343. }
  1344. else
  1345. {
  1346. vkGetPhysicalDeviceFeatures(m_physicalDevice, &supportedFeatures);
  1347. }
  1348. bx::memSet(&m_deviceFeatures, 0, sizeof(m_deviceFeatures) );
  1349. m_deviceFeatures.fullDrawIndexUint32 = supportedFeatures.fullDrawIndexUint32;
  1350. m_deviceFeatures.imageCubeArray = supportedFeatures.imageCubeArray && (_init.capabilities & BGFX_CAPS_TEXTURE_CUBE_ARRAY);
  1351. m_deviceFeatures.independentBlend = supportedFeatures.independentBlend && (_init.capabilities & BGFX_CAPS_BLEND_INDEPENDENT);
  1352. m_deviceFeatures.multiDrawIndirect = supportedFeatures.multiDrawIndirect && (_init.capabilities & BGFX_CAPS_DRAW_INDIRECT);
  1353. m_deviceFeatures.drawIndirectFirstInstance = supportedFeatures.drawIndirectFirstInstance && (_init.capabilities & BGFX_CAPS_DRAW_INDIRECT);
  1354. m_deviceFeatures.depthClamp = supportedFeatures.depthClamp;
  1355. m_deviceFeatures.fillModeNonSolid = supportedFeatures.fillModeNonSolid;
  1356. m_deviceFeatures.largePoints = supportedFeatures.largePoints;
  1357. m_deviceFeatures.samplerAnisotropy = supportedFeatures.samplerAnisotropy;
  1358. m_deviceFeatures.textureCompressionETC2 = supportedFeatures.textureCompressionETC2;
  1359. m_deviceFeatures.textureCompressionBC = supportedFeatures.textureCompressionBC;
  1360. m_deviceFeatures.vertexPipelineStoresAndAtomics = supportedFeatures.vertexPipelineStoresAndAtomics;
  1361. m_deviceFeatures.fragmentStoresAndAtomics = supportedFeatures.fragmentStoresAndAtomics;
  1362. m_deviceFeatures.shaderImageGatherExtended = supportedFeatures.shaderImageGatherExtended;
  1363. m_deviceFeatures.shaderStorageImageExtendedFormats = supportedFeatures.shaderStorageImageExtendedFormats;
  1364. m_deviceFeatures.shaderClipDistance = supportedFeatures.shaderClipDistance;
  1365. m_deviceFeatures.shaderCullDistance = supportedFeatures.shaderCullDistance;
  1366. m_deviceFeatures.shaderResourceMinLod = supportedFeatures.shaderResourceMinLod;
  1367. m_deviceFeatures.geometryShader = supportedFeatures.geometryShader;
  1368. m_lineAASupport = true
  1369. && s_extension[Extension::EXT_line_rasterization].m_supported
  1370. && lineRasterizationFeatures.smoothLines
  1371. ;
  1372. m_borderColorSupport = true
  1373. && s_extension[Extension::EXT_custom_border_color].m_supported
  1374. && customBorderColorFeatures.customBorderColors
  1375. ;
  1376. m_timerQuerySupport = m_deviceProperties.limits.timestampComputeAndGraphics;
  1377. const bool indirectDrawSupport = true
  1378. && m_deviceFeatures.multiDrawIndirect
  1379. && m_deviceFeatures.drawIndirectFirstInstance
  1380. ;
  1381. g_caps.supported |= ( 0
  1382. | BGFX_CAPS_ALPHA_TO_COVERAGE
  1383. | (m_deviceFeatures.independentBlend ? BGFX_CAPS_BLEND_INDEPENDENT : 0)
  1384. | BGFX_CAPS_COMPUTE
  1385. | (indirectDrawSupport ? BGFX_CAPS_DRAW_INDIRECT : 0)
  1386. | BGFX_CAPS_FRAGMENT_DEPTH
  1387. | BGFX_CAPS_IMAGE_RW
  1388. | (m_deviceFeatures.fullDrawIndexUint32 ? BGFX_CAPS_INDEX32 : 0)
  1389. | BGFX_CAPS_INSTANCING
  1390. | BGFX_CAPS_OCCLUSION_QUERY
  1391. | (!headless ? BGFX_CAPS_SWAP_CHAIN : 0)
  1392. | BGFX_CAPS_TEXTURE_2D_ARRAY
  1393. | BGFX_CAPS_TEXTURE_3D
  1394. | BGFX_CAPS_TEXTURE_BLIT
  1395. | BGFX_CAPS_TEXTURE_COMPARE_ALL
  1396. | (m_deviceFeatures.imageCubeArray ? BGFX_CAPS_TEXTURE_CUBE_ARRAY : 0)
  1397. | BGFX_CAPS_TEXTURE_READ_BACK
  1398. | BGFX_CAPS_VERTEX_ATTRIB_HALF
  1399. | BGFX_CAPS_VERTEX_ATTRIB_UINT10
  1400. | BGFX_CAPS_VERTEX_ID
  1401. | (m_deviceFeatures.geometryShader ? BGFX_CAPS_PRIMITIVE_ID : 0)
  1402. );
  1403. g_caps.supported |= 0
  1404. | (s_extension[Extension::EXT_conservative_rasterization ].m_supported ? BGFX_CAPS_CONSERVATIVE_RASTER : 0)
  1405. | (s_extension[Extension::EXT_shader_viewport_index_layer].m_supported ? BGFX_CAPS_VIEWPORT_LAYER_ARRAY : 0)
  1406. | (s_extension[Extension::KHR_draw_indirect_count ].m_supported && indirectDrawSupport ? BGFX_CAPS_DRAW_INDIRECT_COUNT : 0)
  1407. ;
  1408. const uint32_t maxAttachments = bx::min<uint32_t>(m_deviceProperties.limits.maxFragmentOutputAttachments, m_deviceProperties.limits.maxColorAttachments);
  1409. g_caps.limits.maxTextureSize = m_deviceProperties.limits.maxImageDimension2D;
  1410. g_caps.limits.maxTextureLayers = m_deviceProperties.limits.maxImageArrayLayers;
  1411. g_caps.limits.maxFBAttachments = bx::min<uint32_t>(maxAttachments, BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS);
  1412. g_caps.limits.maxTextureSamplers = bx::min<uint32_t>(m_deviceProperties.limits.maxPerStageResources, BGFX_CONFIG_MAX_TEXTURE_SAMPLERS);
  1413. g_caps.limits.maxComputeBindings = bx::min<uint32_t>(m_deviceProperties.limits.maxPerStageResources, BGFX_MAX_COMPUTE_BINDINGS);
  1414. g_caps.limits.maxVertexStreams = bx::min<uint32_t>(m_deviceProperties.limits.maxVertexInputBindings, BGFX_CONFIG_MAX_VERTEX_STREAMS);
  1415. {
  1416. const VkSampleCountFlags sampleMask = ~0
  1417. & m_deviceProperties.limits.framebufferColorSampleCounts
  1418. & m_deviceProperties.limits.framebufferDepthSampleCounts
  1419. ;
  1420. for (uint16_t ii = 0, last = 0; ii < BX_COUNTOF(s_msaa); ii++)
  1421. {
  1422. const VkSampleCountFlags sampleBit = s_msaa[ii].Sample;
  1423. if (sampleBit & sampleMask)
  1424. {
  1425. last = ii;
  1426. }
  1427. else
  1428. {
  1429. s_msaa[ii] = s_msaa[last];
  1430. }
  1431. }
  1432. }
  1433. for (uint32_t ii = 0; ii < TextureFormat::Count; ++ii)
  1434. {
  1435. uint16_t support = BGFX_CAPS_FORMAT_TEXTURE_NONE;
  1436. const bool depth = bimg::isDepth(bimg::TextureFormat::Enum(ii) );
  1437. VkFormat fmt = depth
  1438. ? s_textureFormat[ii].m_fmtDsv
  1439. : s_textureFormat[ii].m_fmt
  1440. ;
  1441. for (uint32_t jj = 0, num = depth ? 1 : 2; jj < num; ++jj)
  1442. {
  1443. if (VK_FORMAT_UNDEFINED != fmt)
  1444. {
  1445. for (uint32_t test = 0; test < BX_COUNTOF(s_imageTest); ++test)
  1446. {
  1447. const ImageTest& it = s_imageTest[test];
  1448. VkImageFormatProperties ifp;
  1449. result = vkGetPhysicalDeviceImageFormatProperties(
  1450. m_physicalDevice
  1451. , fmt
  1452. , it.type
  1453. , VK_IMAGE_TILING_OPTIMAL
  1454. , it.usage
  1455. , it.flags
  1456. , &ifp
  1457. );
  1458. if (VK_SUCCESS == result)
  1459. {
  1460. support |= it.formatCaps[jj];
  1461. const bool multisample = VK_SAMPLE_COUNT_1_BIT < ifp.sampleCounts;
  1462. if (it.usage & VK_IMAGE_USAGE_SAMPLED_BIT)
  1463. {
  1464. support |= 0
  1465. | BGFX_CAPS_FORMAT_TEXTURE_VERTEX
  1466. | (multisample ? BGFX_CAPS_FORMAT_TEXTURE_MSAA : 0)
  1467. ;
  1468. }
  1469. if (it.usage & (VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT) )
  1470. {
  1471. support |= 0
  1472. | BGFX_CAPS_FORMAT_TEXTURE_MIP_AUTOGEN
  1473. | (multisample ? BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER_MSAA : 0)
  1474. ;
  1475. }
  1476. }
  1477. }
  1478. }
  1479. fmt = s_textureFormat[ii].m_fmtSrgb;
  1480. }
  1481. g_caps.formats[ii] = support;
  1482. }
  1483. vkGetPhysicalDeviceMemoryProperties(m_physicalDevice, &m_memoryProperties);
  1484. }
  1485. {
  1486. BX_TRACE("---");
  1487. uint32_t queueFamilyPropertyCount = 0;
  1488. vkGetPhysicalDeviceQueueFamilyProperties(
  1489. m_physicalDevice
  1490. , &queueFamilyPropertyCount
  1491. , NULL
  1492. );
  1493. VkQueueFamilyProperties* queueFamilyPropertices = (VkQueueFamilyProperties*)bx::alloc(g_allocator, queueFamilyPropertyCount * sizeof(VkQueueFamilyProperties) );
  1494. vkGetPhysicalDeviceQueueFamilyProperties(
  1495. m_physicalDevice
  1496. , &queueFamilyPropertyCount
  1497. , queueFamilyPropertices
  1498. );
  1499. for (uint32_t ii = 0; ii < queueFamilyPropertyCount; ++ii)
  1500. {
  1501. const VkQueueFamilyProperties& qfp = queueFamilyPropertices[ii];
  1502. BX_TRACE("Queue family property %d:", ii);
  1503. BX_TRACE("\t Queue flags: 0x%08x", qfp.queueFlags);
  1504. BX_TRACE("\t Queue count: %d", qfp.queueCount);
  1505. BX_TRACE("\tTS valid bits: 0x%08x", qfp.timestampValidBits);
  1506. BX_TRACE("\t Min image: %d x %d x %d"
  1507. , qfp.minImageTransferGranularity.width
  1508. , qfp.minImageTransferGranularity.height
  1509. , qfp.minImageTransferGranularity.depth
  1510. );
  1511. constexpr VkQueueFlags requiredFlags = VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT;
  1512. if (UINT32_MAX == m_globalQueueFamily
  1513. && requiredFlags == (requiredFlags & qfp.queueFlags) )
  1514. {
  1515. m_globalQueueFamily = ii;
  1516. }
  1517. }
  1518. bx::free(g_allocator, queueFamilyPropertices);
  1519. if (UINT32_MAX == m_globalQueueFamily)
  1520. {
  1521. BX_TRACE("Init error: Unable to find combined graphics and compute queue.");
  1522. goto error;
  1523. }
  1524. }
  1525. {
  1526. uint32_t numEnabledLayers = 0;
  1527. const char* enabledLayer[Layer::Count];
  1528. BX_TRACE("Enabled device layers:");
  1529. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  1530. {
  1531. const Layer& layer = s_layer[ii];
  1532. if (layer.m_device.m_supported
  1533. && layer.m_device.m_initialize)
  1534. {
  1535. enabledLayer[numEnabledLayers++] = layer.m_name;
  1536. BX_TRACE("\t%s", layer.m_name);
  1537. }
  1538. }
  1539. uint32_t numEnabledExtensions = 0;
  1540. const char* enabledExtension[Extension::Count + 3];
  1541. enabledExtension[numEnabledExtensions++] = VK_KHR_MAINTENANCE1_EXTENSION_NAME;
  1542. if (!headless)
  1543. {
  1544. enabledExtension[numEnabledExtensions++] = VK_KHR_SWAPCHAIN_EXTENSION_NAME;
  1545. }
  1546. if (BX_ENABLED(BX_PLATFORM_OSX) )
  1547. {
  1548. enabledExtension[numEnabledExtensions++] = VK_KHR_PORTABILITY_SUBSET_EXTENSION_NAME;
  1549. }
  1550. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  1551. {
  1552. const Extension& extension = s_extension[ii];
  1553. bool layerEnabled = extension.m_layer == Layer::Count
  1554. || (s_layer[extension.m_layer].m_device.m_supported && s_layer[extension.m_layer].m_device.m_initialize)
  1555. ;
  1556. if (extension.m_supported
  1557. && extension.m_initialize
  1558. && !extension.m_instanceExt
  1559. && layerEnabled)
  1560. {
  1561. enabledExtension[numEnabledExtensions++] = extension.m_name;
  1562. }
  1563. }
  1564. BX_TRACE("Enabled device extensions:");
  1565. for (uint32_t ii = 0; ii < numEnabledExtensions; ++ii)
  1566. {
  1567. BX_TRACE("\t%s", enabledExtension[ii]);
  1568. }
  1569. float queuePriorities[1] = { 0.0f };
  1570. VkDeviceQueueCreateInfo dcqi;
  1571. dcqi.sType = VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO;
  1572. dcqi.pNext = NULL;
  1573. dcqi.flags = 0;
  1574. dcqi.queueFamilyIndex = m_globalQueueFamily;
  1575. dcqi.queueCount = 1;
  1576. dcqi.pQueuePriorities = queuePriorities;
  1577. VkDeviceCreateInfo dci;
  1578. dci.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO;
  1579. dci.pNext = nextFeatures;
  1580. dci.flags = 0;
  1581. dci.queueCreateInfoCount = 1;
  1582. dci.pQueueCreateInfos = &dcqi;
  1583. dci.enabledLayerCount = numEnabledLayers;
  1584. dci.ppEnabledLayerNames = enabledLayer;
  1585. dci.enabledExtensionCount = numEnabledExtensions;
  1586. dci.ppEnabledExtensionNames = enabledExtension;
  1587. dci.pEnabledFeatures = &m_deviceFeatures;
  1588. result = vkCreateDevice(
  1589. m_physicalDevice
  1590. , &dci
  1591. , m_allocatorCb
  1592. , &m_device
  1593. );
  1594. if (VK_SUCCESS != result)
  1595. {
  1596. BX_TRACE("Init error: vkCreateDevice failed %d: %s.", result, getName(result) );
  1597. goto error;
  1598. }
  1599. }
  1600. errorState = ErrorState::DeviceCreated;
  1601. BX_TRACE("Device functions:");
  1602. #define VK_IMPORT_DEVICE_FUNC(_optional, _func) \
  1603. _func = (PFN_##_func)vkGetDeviceProcAddr(m_device, #_func); \
  1604. BX_TRACE("\t%p " #_func, _func); \
  1605. imported &= _optional || NULL != _func
  1606. VK_IMPORT_DEVICE
  1607. #undef VK_IMPORT_DEVICE_FUNC
  1608. if (!imported)
  1609. {
  1610. BX_TRACE("Init error: Failed to load device functions.");
  1611. goto error;
  1612. }
  1613. vkGetDeviceQueue(m_device, m_globalQueueFamily, 0, &m_globalQueue);
  1614. {
  1615. m_numFramesInFlight = _init.resolution.maxFrameLatency == 0
  1616. ? BGFX_CONFIG_MAX_FRAME_LATENCY
  1617. : _init.resolution.maxFrameLatency
  1618. ;
  1619. result = m_cmd.init(m_globalQueueFamily, m_globalQueue, m_numFramesInFlight);
  1620. if (VK_SUCCESS != result)
  1621. {
  1622. BX_TRACE("Init error: creating command queue failed %d: %s.", result, getName(result) );
  1623. goto error;
  1624. }
  1625. result = m_cmd.alloc(&m_commandBuffer);
  1626. if (VK_SUCCESS != result)
  1627. {
  1628. BX_TRACE("Init error: allocating command buffer failed %d: %s.", result, getName(result) );
  1629. goto error;
  1630. }
  1631. }
  1632. errorState = ErrorState::CommandQueueCreated;
  1633. m_presentElapsed = 0;
  1634. {
  1635. m_resolution = _init.resolution;
  1636. m_resolution.reset &= ~BGFX_RESET_INTERNAL_FORCE;
  1637. m_numWindows = 0;
  1638. if (!headless)
  1639. {
  1640. m_textVideoMem.resize(false, _init.resolution.width, _init.resolution.height);
  1641. m_textVideoMem.clear();
  1642. for (uint8_t ii = 0; ii < BX_COUNTOF(m_swapchainFormats); ++ii)
  1643. {
  1644. m_swapchainFormats[ii] = TextureFormat::Enum(ii);
  1645. }
  1646. result = m_backBuffer.create(UINT16_MAX, g_platformData.nwh, m_resolution.width, m_resolution.height, m_resolution.format);
  1647. if (VK_SUCCESS != result)
  1648. {
  1649. BX_TRACE("Init error: creating swap chain failed %d: %s.", result, getName(result) );
  1650. goto error;
  1651. }
  1652. m_windows[0] = BGFX_INVALID_HANDLE;
  1653. m_numWindows++;
  1654. postReset();
  1655. }
  1656. }
  1657. errorState = ErrorState::SwapChainCreated;
  1658. {
  1659. VkDescriptorPoolSize dps[] =
  1660. {
  1661. { VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1662. { VK_DESCRIPTOR_TYPE_SAMPLER, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1663. { VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC, MAX_DESCRIPTOR_SETS * 2 },
  1664. { VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1665. { VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1666. };
  1667. VkDescriptorPoolCreateInfo dpci;
  1668. dpci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO;
  1669. dpci.pNext = NULL;
  1670. dpci.flags = VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT;
  1671. dpci.maxSets = MAX_DESCRIPTOR_SETS;
  1672. dpci.poolSizeCount = BX_COUNTOF(dps);
  1673. dpci.pPoolSizes = dps;
  1674. result = vkCreateDescriptorPool(m_device, &dpci, m_allocatorCb, &m_descriptorPool);
  1675. if (VK_SUCCESS != result)
  1676. {
  1677. BX_TRACE("Init error: vkCreateDescriptorPool failed %d: %s.", result, getName(result) );
  1678. goto error;
  1679. }
  1680. VkPipelineCacheCreateInfo pcci;
  1681. pcci.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO;
  1682. pcci.pNext = NULL;
  1683. pcci.flags = 0;
  1684. pcci.initialDataSize = 0;
  1685. pcci.pInitialData = NULL;
  1686. result = vkCreatePipelineCache(m_device, &pcci, m_allocatorCb, &m_pipelineCache);
  1687. if (VK_SUCCESS != result)
  1688. {
  1689. BX_TRACE("Init error: vkCreatePipelineCache failed %d: %s.", result, getName(result) );
  1690. goto error;
  1691. }
  1692. }
  1693. {
  1694. const uint32_t size = 128;
  1695. const uint32_t count = BGFX_CONFIG_MAX_DRAW_CALLS;
  1696. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1697. {
  1698. BX_TRACE("Create scratch buffer %d", ii);
  1699. m_scratchBuffer[ii].create(size, count);
  1700. }
  1701. }
  1702. errorState = ErrorState::DescriptorCreated;
  1703. if (NULL == vkSetDebugUtilsObjectNameEXT)
  1704. {
  1705. vkSetDebugUtilsObjectNameEXT = stubSetDebugUtilsObjectNameEXT;
  1706. }
  1707. if (NULL == vkCmdBeginDebugUtilsLabelEXT
  1708. || NULL == vkCmdEndDebugUtilsLabelEXT)
  1709. {
  1710. vkCmdBeginDebugUtilsLabelEXT = stubCmdBeginDebugUtilsLabelEXT;
  1711. vkCmdEndDebugUtilsLabelEXT = stubCmdEndDebugUtilsLabelEXT;
  1712. }
  1713. if (NULL == vkCmdInsertDebugUtilsLabelEXT)
  1714. {
  1715. vkCmdInsertDebugUtilsLabelEXT = stubCmdInsertDebugUtilsLabelEXT;
  1716. }
  1717. // Init reserved part of view name.
  1718. for (uint32_t ii = 0; ii < BGFX_CONFIG_MAX_VIEWS; ++ii)
  1719. {
  1720. bx::snprintf(s_viewName[ii], BGFX_CONFIG_MAX_VIEW_NAME_RESERVED+1, "%3d ", ii);
  1721. }
  1722. if (m_timerQuerySupport)
  1723. {
  1724. result = m_gpuTimer.init();
  1725. if (VK_SUCCESS != result)
  1726. {
  1727. BX_TRACE("Init error: creating GPU timer failed %d: %s.", result, getName(result) );
  1728. goto error;
  1729. }
  1730. }
  1731. errorState = ErrorState::TimerQueryCreated;
  1732. result = m_occlusionQuery.init();
  1733. if (VK_SUCCESS != result)
  1734. {
  1735. BX_TRACE("Init error: creating occlusion query failed %d: %s.", result, getName(result) );
  1736. goto error;
  1737. }
  1738. g_internalData.context = m_device;
  1739. return true;
  1740. error:
  1741. BX_TRACE("errorState %d", errorState);
  1742. switch (errorState)
  1743. {
  1744. case ErrorState::TimerQueryCreated:
  1745. if (m_timerQuerySupport)
  1746. {
  1747. m_gpuTimer.shutdown();
  1748. }
  1749. [[fallthrough]];
  1750. case ErrorState::DescriptorCreated:
  1751. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1752. {
  1753. m_scratchBuffer[ii].destroy();
  1754. }
  1755. vkDestroy(m_pipelineCache);
  1756. vkDestroy(m_descriptorPool);
  1757. [[fallthrough]];
  1758. case ErrorState::SwapChainCreated:
  1759. m_backBuffer.destroy();
  1760. [[fallthrough]];
  1761. case ErrorState::CommandQueueCreated:
  1762. m_cmd.shutdown();
  1763. [[fallthrough]];
  1764. case ErrorState::DeviceCreated:
  1765. vkDestroyDevice(m_device, m_allocatorCb);
  1766. [[fallthrough]];
  1767. case ErrorState::InstanceCreated:
  1768. if (VK_NULL_HANDLE != m_debugReportCallback)
  1769. {
  1770. vkDestroyDebugReportCallbackEXT(m_instance, m_debugReportCallback, m_allocatorCb);
  1771. }
  1772. vkDestroyInstance(m_instance, m_allocatorCb);
  1773. [[fallthrough]];
  1774. case ErrorState::LoadedVulkan1:
  1775. bx::dlclose(m_vulkan1Dll);
  1776. m_vulkan1Dll = NULL;
  1777. m_allocatorCb = NULL;
  1778. unloadRenderDoc(m_renderDocDll);
  1779. [[fallthrough]];
  1780. case ErrorState::Default:
  1781. break;
  1782. };
  1783. return false;
  1784. }
  1785. void shutdown()
  1786. {
  1787. VK_CHECK(vkDeviceWaitIdle(m_device) );
  1788. if (m_timerQuerySupport)
  1789. {
  1790. m_gpuTimer.shutdown();
  1791. }
  1792. m_occlusionQuery.shutdown();
  1793. preReset();
  1794. m_pipelineStateCache.invalidate();
  1795. m_descriptorSetLayoutCache.invalidate();
  1796. m_renderPassCache.invalidate();
  1797. m_samplerCache.invalidate();
  1798. m_samplerBorderColorCache.invalidate();
  1799. m_imageViewCache.invalidate();
  1800. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1801. {
  1802. m_scratchBuffer[ii].destroy();
  1803. }
  1804. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  1805. {
  1806. m_frameBuffers[ii].destroy();
  1807. }
  1808. for (uint32_t ii = 0; ii < BX_COUNTOF(m_indexBuffers); ++ii)
  1809. {
  1810. m_indexBuffers[ii].destroy();
  1811. }
  1812. for (uint32_t ii = 0; ii < BX_COUNTOF(m_vertexBuffers); ++ii)
  1813. {
  1814. m_vertexBuffers[ii].destroy();
  1815. }
  1816. for (uint32_t ii = 0; ii < BX_COUNTOF(m_shaders); ++ii)
  1817. {
  1818. m_shaders[ii].destroy();
  1819. }
  1820. for (uint32_t ii = 0; ii < BX_COUNTOF(m_textures); ++ii)
  1821. {
  1822. m_textures[ii].destroy();
  1823. }
  1824. m_backBuffer.destroy();
  1825. m_cmd.shutdown();
  1826. vkDestroy(m_pipelineCache);
  1827. vkDestroy(m_descriptorPool);
  1828. vkDestroyDevice(m_device, m_allocatorCb);
  1829. if (VK_NULL_HANDLE != m_debugReportCallback)
  1830. {
  1831. vkDestroyDebugReportCallbackEXT(m_instance, m_debugReportCallback, m_allocatorCb);
  1832. }
  1833. vkDestroyInstance(m_instance, m_allocatorCb);
  1834. bx::dlclose(m_vulkan1Dll);
  1835. m_vulkan1Dll = NULL;
  1836. m_allocatorCb = NULL;
  1837. unloadRenderDoc(m_renderDocDll);
  1838. }
  1839. RendererType::Enum getRendererType() const override
  1840. {
  1841. return RendererType::Vulkan;
  1842. }
  1843. const char* getRendererName() const override
  1844. {
  1845. return BGFX_RENDERER_VULKAN_NAME;
  1846. }
  1847. bool isDeviceRemoved() override
  1848. {
  1849. return false;
  1850. }
  1851. void flip() override
  1852. {
  1853. int64_t start = bx::getHPCounter();
  1854. for (uint16_t ii = 0; ii < m_numWindows; ++ii)
  1855. {
  1856. FrameBufferVK& fb = isValid(m_windows[ii])
  1857. ? m_frameBuffers[m_windows[ii].idx]
  1858. : m_backBuffer
  1859. ;
  1860. fb.present();
  1861. }
  1862. int64_t now = bx::getHPCounter();
  1863. m_presentElapsed += now - start;
  1864. }
  1865. void createIndexBuffer(IndexBufferHandle _handle, const Memory* _mem, uint16_t _flags) override
  1866. {
  1867. m_indexBuffers[_handle.idx].create(m_commandBuffer, _mem->size, _mem->data, _flags, false);
  1868. }
  1869. void destroyIndexBuffer(IndexBufferHandle _handle) override
  1870. {
  1871. m_indexBuffers[_handle.idx].destroy();
  1872. }
  1873. void createVertexLayout(VertexLayoutHandle _handle, const VertexLayout& _layout) override
  1874. {
  1875. VertexLayout& layout = m_vertexLayouts[_handle.idx];
  1876. bx::memCopy(&layout, &_layout, sizeof(VertexLayout) );
  1877. dump(layout);
  1878. }
  1879. void destroyVertexLayout(VertexLayoutHandle /*_handle*/) override
  1880. {
  1881. }
  1882. void createVertexBuffer(VertexBufferHandle _handle, const Memory* _mem, VertexLayoutHandle _layoutHandle, uint16_t _flags) override
  1883. {
  1884. m_vertexBuffers[_handle.idx].create(m_commandBuffer, _mem->size, _mem->data, _layoutHandle, _flags);
  1885. }
  1886. void destroyVertexBuffer(VertexBufferHandle _handle) override
  1887. {
  1888. m_vertexBuffers[_handle.idx].destroy();
  1889. }
  1890. void createDynamicIndexBuffer(IndexBufferHandle _handle, uint32_t _size, uint16_t _flags) override
  1891. {
  1892. m_indexBuffers[_handle.idx].create(m_commandBuffer, _size, NULL, _flags, false);
  1893. }
  1894. void updateDynamicIndexBuffer(IndexBufferHandle _handle, uint32_t _offset, uint32_t _size, const Memory* _mem) override
  1895. {
  1896. m_indexBuffers[_handle.idx].update(m_commandBuffer, _offset, bx::min<uint32_t>(_size, _mem->size), _mem->data);
  1897. }
  1898. void destroyDynamicIndexBuffer(IndexBufferHandle _handle) override
  1899. {
  1900. m_indexBuffers[_handle.idx].destroy();
  1901. }
  1902. void createDynamicVertexBuffer(VertexBufferHandle _handle, uint32_t _size, uint16_t _flags) override
  1903. {
  1904. VertexLayoutHandle layoutHandle = BGFX_INVALID_HANDLE;
  1905. m_vertexBuffers[_handle.idx].create(m_commandBuffer, _size, NULL, layoutHandle, _flags);
  1906. }
  1907. void updateDynamicVertexBuffer(VertexBufferHandle _handle, uint32_t _offset, uint32_t _size, const Memory* _mem) override
  1908. {
  1909. m_vertexBuffers[_handle.idx].update(m_commandBuffer, _offset, bx::min<uint32_t>(_size, _mem->size), _mem->data);
  1910. }
  1911. void destroyDynamicVertexBuffer(VertexBufferHandle _handle) override
  1912. {
  1913. m_vertexBuffers[_handle.idx].destroy();
  1914. }
  1915. void createShader(ShaderHandle _handle, const Memory* _mem) override
  1916. {
  1917. m_shaders[_handle.idx].create(_mem);
  1918. }
  1919. void destroyShader(ShaderHandle _handle) override
  1920. {
  1921. m_shaders[_handle.idx].destroy();
  1922. }
  1923. void createProgram(ProgramHandle _handle, ShaderHandle _vsh, ShaderHandle _fsh) override
  1924. {
  1925. m_program[_handle.idx].create(&m_shaders[_vsh.idx], isValid(_fsh) ? &m_shaders[_fsh.idx] : NULL);
  1926. }
  1927. void destroyProgram(ProgramHandle _handle) override
  1928. {
  1929. m_program[_handle.idx].destroy();
  1930. }
  1931. void* createTexture(TextureHandle _handle, const Memory* _mem, uint64_t _flags, uint8_t _skip) override
  1932. {
  1933. return m_textures[_handle.idx].create(m_commandBuffer, _mem, _flags, _skip);
  1934. }
  1935. void updateTextureBegin(TextureHandle /*_handle*/, uint8_t /*_side*/, uint8_t /*_mip*/) override
  1936. {
  1937. }
  1938. void updateTexture(TextureHandle _handle, uint8_t _side, uint8_t _mip, const Rect& _rect, uint16_t _z, uint16_t _depth, uint16_t _pitch, const Memory* _mem) override
  1939. {
  1940. m_textures[_handle.idx].update(m_commandBuffer, _side, _mip, _rect, _z, _depth, _pitch, _mem);
  1941. }
  1942. void updateTextureEnd() override
  1943. {
  1944. }
  1945. void readTexture(TextureHandle _handle, void* _data, uint8_t _mip) override
  1946. {
  1947. TextureVK& texture = m_textures[_handle.idx];
  1948. uint32_t height = bx::uint32_max(1, texture.m_height >> _mip);
  1949. uint32_t pitch = texture.m_readback.pitch(_mip);
  1950. uint32_t size = height * pitch;
  1951. VkDeviceMemory stagingMemory;
  1952. VkBuffer stagingBuffer;
  1953. VK_CHECK(createReadbackBuffer(size, &stagingBuffer, &stagingMemory) );
  1954. texture.m_readback.copyImageToBuffer(
  1955. m_commandBuffer
  1956. , stagingBuffer
  1957. , texture.m_currentImageLayout
  1958. , texture.m_aspectMask
  1959. , _mip
  1960. );
  1961. kick(true);
  1962. texture.m_readback.readback(stagingMemory, 0, _data, _mip);
  1963. vkDestroy(stagingBuffer);
  1964. vkDestroy(stagingMemory);
  1965. }
  1966. void resizeTexture(TextureHandle _handle, uint16_t _width, uint16_t _height, uint8_t _numMips, uint16_t _numLayers) override
  1967. {
  1968. const TextureVK& texture = m_textures[_handle.idx];
  1969. const TextureFormat::Enum format = TextureFormat::Enum(texture.m_requestedFormat);
  1970. const uint64_t flags = texture.m_flags;
  1971. const uint32_t size = sizeof(uint32_t) + sizeof(TextureCreate);
  1972. const Memory* mem = alloc(size);
  1973. bx::StaticMemoryBlockWriter writer(mem->data, mem->size);
  1974. uint32_t magic = BGFX_CHUNK_MAGIC_TEX;
  1975. bx::write(&writer, magic, bx::ErrorAssert{});
  1976. TextureCreate tc;
  1977. tc.m_width = _width;
  1978. tc.m_height = _height;
  1979. tc.m_depth = 0;
  1980. tc.m_numLayers = _numLayers;
  1981. tc.m_numMips = _numMips;
  1982. tc.m_format = format;
  1983. tc.m_cubeMap = false;
  1984. tc.m_mem = NULL;
  1985. bx::write(&writer, tc, bx::ErrorAssert{});
  1986. destroyTexture(_handle);
  1987. createTexture(_handle, mem, flags, 0);
  1988. bgfx::release(mem);
  1989. }
  1990. void overrideInternal(TextureHandle /*_handle*/, uintptr_t /*_ptr*/) override
  1991. {
  1992. }
  1993. uintptr_t getInternal(TextureHandle /*_handle*/) override
  1994. {
  1995. return 0;
  1996. }
  1997. void destroyTexture(TextureHandle _handle) override
  1998. {
  1999. m_imageViewCache.invalidateWithParent(_handle.idx);
  2000. m_textures[_handle.idx].destroy();
  2001. }
  2002. void createFrameBuffer(FrameBufferHandle _handle, uint8_t _num, const Attachment* _attachment) override
  2003. {
  2004. m_frameBuffers[_handle.idx].create(_num, _attachment);
  2005. }
  2006. void createFrameBuffer(FrameBufferHandle _handle, void* _nwh, uint32_t _width, uint32_t _height, TextureFormat::Enum _format, TextureFormat::Enum _depthFormat) override
  2007. {
  2008. for (uint32_t ii = 0, num = m_numWindows; ii < num; ++ii)
  2009. {
  2010. FrameBufferHandle handle = m_windows[ii];
  2011. if (isValid(handle)
  2012. && m_frameBuffers[handle.idx].m_nwh == _nwh)
  2013. {
  2014. destroyFrameBuffer(handle);
  2015. }
  2016. }
  2017. uint16_t denseIdx = m_numWindows++;
  2018. m_windows[denseIdx] = _handle;
  2019. VK_CHECK(m_frameBuffers[_handle.idx].create(denseIdx, _nwh, _width, _height, _format, _depthFormat) );
  2020. }
  2021. void destroyFrameBuffer(FrameBufferHandle _handle) override
  2022. {
  2023. FrameBufferVK& frameBuffer = m_frameBuffers[_handle.idx];
  2024. if (_handle.idx == m_fbh.idx)
  2025. {
  2026. setFrameBuffer(BGFX_INVALID_HANDLE, false);
  2027. }
  2028. uint16_t denseIdx = frameBuffer.destroy();
  2029. if (UINT16_MAX != denseIdx)
  2030. {
  2031. --m_numWindows;
  2032. if (m_numWindows > 1)
  2033. {
  2034. FrameBufferHandle handle = m_windows[m_numWindows];
  2035. m_windows[m_numWindows] = {kInvalidHandle};
  2036. if (m_numWindows != denseIdx)
  2037. {
  2038. m_windows[denseIdx] = handle;
  2039. m_frameBuffers[handle.idx].m_denseIdx = denseIdx;
  2040. }
  2041. }
  2042. }
  2043. }
  2044. void createUniform(UniformHandle _handle, UniformType::Enum _type, uint16_t _num, const char* _name) override
  2045. {
  2046. if (NULL != m_uniforms[_handle.idx])
  2047. {
  2048. bx::free(g_allocator, m_uniforms[_handle.idx]);
  2049. }
  2050. const uint32_t size = bx::alignUp(g_uniformTypeSize[_type] * _num, 16);
  2051. void* data = bx::alloc(g_allocator, size);
  2052. bx::memSet(data, 0, size);
  2053. m_uniforms[_handle.idx] = data;
  2054. m_uniformReg.add(_handle, _name);
  2055. }
  2056. void destroyUniform(UniformHandle _handle) override
  2057. {
  2058. bx::free(g_allocator, m_uniforms[_handle.idx]);
  2059. m_uniforms[_handle.idx] = NULL;
  2060. }
  2061. void requestScreenShot(FrameBufferHandle _fbh, const char* _filePath) override
  2062. {
  2063. const FrameBufferVK& frameBuffer = isValid(_fbh)
  2064. ? m_frameBuffers[_fbh.idx]
  2065. : m_backBuffer
  2066. ;
  2067. const SwapChainVK& swapChain = frameBuffer.m_swapChain;
  2068. if (!isSwapChainReadable(swapChain) )
  2069. {
  2070. BX_TRACE("Unable to capture screenshot %s.", _filePath);
  2071. return;
  2072. }
  2073. auto callback = [](void* _src, uint32_t _width, uint32_t _height, uint32_t _pitch, const void* _userData)
  2074. {
  2075. const char* filePath = (const char*)_userData;
  2076. g_callback->screenShot(
  2077. filePath
  2078. , _width
  2079. , _height
  2080. , _pitch
  2081. , _src
  2082. , _height * _pitch
  2083. , false
  2084. );
  2085. };
  2086. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(swapChain.m_colorFormat) );
  2087. const uint32_t size = frameBuffer.m_width * frameBuffer.m_height * bpp / 8;
  2088. VkDeviceMemory stagingMemory;
  2089. VkBuffer stagingBuffer;
  2090. VK_CHECK(createReadbackBuffer(size, &stagingBuffer, &stagingMemory) );
  2091. readSwapChain(swapChain, stagingBuffer, stagingMemory, callback, _filePath);
  2092. vkDestroy(stagingBuffer);
  2093. vkDestroy(stagingMemory);
  2094. }
  2095. void updateViewName(ViewId _id, const char* _name) override
  2096. {
  2097. bx::strCopy(&s_viewName[_id][BGFX_CONFIG_MAX_VIEW_NAME_RESERVED]
  2098. , BX_COUNTOF(s_viewName[0]) - BGFX_CONFIG_MAX_VIEW_NAME_RESERVED
  2099. , _name
  2100. );
  2101. }
  2102. void updateUniform(uint16_t _loc, const void* _data, uint32_t _size) override
  2103. {
  2104. bx::memCopy(m_uniforms[_loc], _data, _size);
  2105. }
  2106. void invalidateOcclusionQuery(OcclusionQueryHandle _handle) override
  2107. {
  2108. m_occlusionQuery.invalidate(_handle);
  2109. }
  2110. void setMarker(const char* _marker, uint16_t _len) override
  2111. {
  2112. if (BX_ENABLED(BGFX_CONFIG_DEBUG_ANNOTATION) )
  2113. {
  2114. BX_UNUSED(_len);
  2115. const uint32_t abgr = kColorMarker;
  2116. VkDebugUtilsLabelEXT dul;
  2117. dul.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT;
  2118. dul.pNext = NULL;
  2119. dul.pLabelName = _marker;
  2120. dul.color[0] = ((abgr >> 24) & 0xff) / 255.0f;
  2121. dul.color[1] = ((abgr >> 16) & 0xff) / 255.0f;
  2122. dul.color[2] = ((abgr >> 8) & 0xff) / 255.0f;
  2123. dul.color[3] = ((abgr >> 0) & 0xff) / 255.0f;
  2124. vkCmdInsertDebugUtilsLabelEXT(m_commandBuffer, &dul);
  2125. }
  2126. }
  2127. virtual void setName(Handle _handle, const char* _name, uint16_t _len) override
  2128. {
  2129. switch (_handle.type)
  2130. {
  2131. case Handle::IndexBuffer:
  2132. setDebugObjectName(m_device, m_indexBuffers[_handle.idx].m_buffer, "%.*s", _len, _name);
  2133. break;
  2134. case Handle::Shader:
  2135. setDebugObjectName(m_device, m_shaders[_handle.idx].m_module, "%.*s", _len, _name);
  2136. break;
  2137. case Handle::Texture:
  2138. setDebugObjectName(m_device, m_textures[_handle.idx].m_textureImage, "%.*s", _len, _name);
  2139. if (VK_NULL_HANDLE != m_textures[_handle.idx].m_singleMsaaImage)
  2140. {
  2141. setDebugObjectName(m_device, m_textures[_handle.idx].m_singleMsaaImage, "%.*s", _len, _name);
  2142. }
  2143. break;
  2144. case Handle::VertexBuffer:
  2145. setDebugObjectName(m_device, m_vertexBuffers[_handle.idx].m_buffer, "%.*s", _len, _name);
  2146. break;
  2147. default:
  2148. BX_ASSERT(false, "Invalid handle type?! %d", _handle.type);
  2149. break;
  2150. }
  2151. }
  2152. template<typename Ty>
  2153. void release(Ty& _object)
  2154. {
  2155. if (VK_NULL_HANDLE != _object)
  2156. {
  2157. m_cmd.release(uint64_t(_object.vk), getType<Ty>() );
  2158. _object = VK_NULL_HANDLE;
  2159. }
  2160. }
  2161. void submitBlit(BlitState& _bs, uint16_t _view);
  2162. void submit(Frame* _render, ClearQuad& _clearQuad, TextVideoMemBlitter& _textVideoMemBlitter) override;
  2163. void blitSetup(TextVideoMemBlitter& _blitter) override
  2164. {
  2165. const uint32_t width = m_backBuffer.m_width;
  2166. const uint32_t height = m_backBuffer.m_height;
  2167. setFrameBuffer(BGFX_INVALID_HANDLE);
  2168. VkViewport vp;
  2169. vp.x = 0.0f;
  2170. vp.y = float(height);
  2171. vp.width = float(width);
  2172. vp.height = -float(height);
  2173. vp.minDepth = 0.0f;
  2174. vp.maxDepth = 1.0f;
  2175. vkCmdSetViewport(m_commandBuffer, 0, 1, &vp);
  2176. VkRect2D rc;
  2177. rc.offset.x = 0;
  2178. rc.offset.y = 0;
  2179. rc.extent.width = width;
  2180. rc.extent.height = height;
  2181. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  2182. const uint64_t state = 0
  2183. | BGFX_STATE_WRITE_RGB
  2184. | BGFX_STATE_WRITE_A
  2185. | BGFX_STATE_DEPTH_TEST_ALWAYS
  2186. | BGFX_STATE_MSAA
  2187. ;
  2188. const VertexLayout* layout = &m_vertexLayouts[_blitter.m_vb->layoutHandle.idx];
  2189. VkPipeline pso = getPipeline(state
  2190. , packStencil(BGFX_STENCIL_DEFAULT, BGFX_STENCIL_DEFAULT)
  2191. , 1
  2192. , &layout
  2193. , _blitter.m_program
  2194. , 0
  2195. );
  2196. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pso);
  2197. ProgramVK& program = m_program[_blitter.m_program.idx];
  2198. float proj[16];
  2199. bx::mtxOrtho(proj, 0.0f, (float)width, (float)height, 0.0f, 0.0f, 1000.0f, 0.0f, false);
  2200. PredefinedUniform& predefined = m_program[_blitter.m_program.idx].m_predefined[0];
  2201. uint8_t flags = predefined.m_type;
  2202. setShaderUniform(flags, predefined.m_loc, proj, 4);
  2203. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  2204. if (NULL != vcb)
  2205. {
  2206. commit(*vcb);
  2207. }
  2208. ScratchBufferVK& scratchBuffer = m_scratchBuffer[m_cmd.m_currentFrameInFlight];
  2209. const uint32_t bufferOffset = scratchBuffer.write(m_vsScratch, program.m_vsh->m_size);
  2210. const TextureVK& texture = m_textures[_blitter.m_texture.idx];
  2211. RenderBind bind;
  2212. bind.clear();
  2213. bind.m_bind[0].m_type = Binding::Texture;
  2214. bind.m_bind[0].m_idx = _blitter.m_texture.idx;
  2215. bind.m_bind[0].m_samplerFlags = (uint32_t)(texture.m_flags & BGFX_SAMPLER_BITS_MASK);
  2216. const VkDescriptorSet descriptorSet = getDescriptorSet(program, bind, scratchBuffer, NULL);
  2217. vkCmdBindDescriptorSets(
  2218. m_commandBuffer
  2219. , VK_PIPELINE_BIND_POINT_GRAPHICS
  2220. , program.m_pipelineLayout
  2221. , 0
  2222. , 1
  2223. , &descriptorSet
  2224. , 1
  2225. , &bufferOffset
  2226. );
  2227. const VertexBufferVK& vb = m_vertexBuffers[_blitter.m_vb->handle.idx];
  2228. const VkDeviceSize offset = 0;
  2229. vkCmdBindVertexBuffers(m_commandBuffer, 0, 1, &vb.m_buffer, &offset);
  2230. const BufferVK& ib = m_indexBuffers[_blitter.m_ib->handle.idx];
  2231. vkCmdBindIndexBuffer(
  2232. m_commandBuffer
  2233. , ib.m_buffer
  2234. , 0
  2235. , VK_INDEX_TYPE_UINT16
  2236. );
  2237. }
  2238. void blitRender(TextVideoMemBlitter& _blitter, uint32_t _numIndices) override
  2239. {
  2240. const uint32_t numVertices = _numIndices*4/6;
  2241. if (0 < numVertices && m_backBuffer.isRenderable() )
  2242. {
  2243. m_indexBuffers[_blitter.m_ib->handle.idx].update(m_commandBuffer, 0, _numIndices*2, _blitter.m_ib->data);
  2244. m_vertexBuffers[_blitter.m_vb->handle.idx].update(m_commandBuffer, 0, numVertices*_blitter.m_layout.m_stride, _blitter.m_vb->data, true);
  2245. VkRenderPassBeginInfo rpbi;
  2246. rpbi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  2247. rpbi.pNext = NULL;
  2248. rpbi.renderPass = m_backBuffer.m_renderPass;
  2249. rpbi.framebuffer = m_backBuffer.m_currentFramebuffer;
  2250. rpbi.renderArea.offset.x = 0;
  2251. rpbi.renderArea.offset.y = 0;
  2252. rpbi.renderArea.extent.width = m_backBuffer.m_width;
  2253. rpbi.renderArea.extent.height = m_backBuffer.m_height;
  2254. rpbi.clearValueCount = 0;
  2255. rpbi.pClearValues = NULL;
  2256. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  2257. vkCmdDrawIndexed(m_commandBuffer, _numIndices, 1, 0, 0, 0);
  2258. vkCmdEndRenderPass(m_commandBuffer);
  2259. }
  2260. }
  2261. void preReset()
  2262. {
  2263. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  2264. {
  2265. m_frameBuffers[ii].preReset();
  2266. }
  2267. if (m_captureSize > 0)
  2268. {
  2269. g_callback->captureEnd();
  2270. release(m_captureBuffer);
  2271. release(m_captureMemory);
  2272. m_captureSize = 0;
  2273. }
  2274. }
  2275. void postReset()
  2276. {
  2277. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  2278. {
  2279. m_frameBuffers[ii].postReset();
  2280. }
  2281. if (m_resolution.reset & BGFX_RESET_CAPTURE)
  2282. {
  2283. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_backBuffer.m_swapChain.m_colorFormat) );
  2284. const uint32_t captureSize = m_backBuffer.m_width * m_backBuffer.m_height * bpp / 8;
  2285. const uint8_t dstBpp = bimg::getBitsPerPixel(bimg::TextureFormat::BGRA8);
  2286. const uint32_t dstPitch = m_backBuffer.m_width * dstBpp / 8;
  2287. if (captureSize > m_captureSize)
  2288. {
  2289. release(m_captureBuffer);
  2290. release(m_captureMemory);
  2291. m_captureSize = captureSize;
  2292. VK_CHECK(createReadbackBuffer(m_captureSize, &m_captureBuffer, &m_captureMemory) );
  2293. }
  2294. g_callback->captureBegin(m_resolution.width, m_resolution.height, dstPitch, TextureFormat::BGRA8, false);
  2295. }
  2296. }
  2297. bool updateResolution(const Resolution& _resolution)
  2298. {
  2299. const bool suspended = !!(_resolution.reset & BGFX_RESET_SUSPEND);
  2300. float maxAnisotropy = 1.0f;
  2301. if (!!(_resolution.reset & BGFX_RESET_MAXANISOTROPY) )
  2302. {
  2303. maxAnisotropy = m_deviceProperties.limits.maxSamplerAnisotropy;
  2304. }
  2305. if (m_maxAnisotropy != maxAnisotropy)
  2306. {
  2307. m_maxAnisotropy = maxAnisotropy;
  2308. m_samplerCache.invalidate();
  2309. m_samplerBorderColorCache.invalidate();
  2310. }
  2311. bool depthClamp = m_deviceFeatures.depthClamp && !!(_resolution.reset & BGFX_RESET_DEPTH_CLAMP);
  2312. if (m_depthClamp != depthClamp)
  2313. {
  2314. m_depthClamp = depthClamp;
  2315. m_pipelineStateCache.invalidate();
  2316. }
  2317. if (NULL == m_backBuffer.m_nwh)
  2318. {
  2319. return suspended;
  2320. }
  2321. uint32_t flags = _resolution.reset & ~(0
  2322. | BGFX_RESET_SUSPEND
  2323. | BGFX_RESET_MAXANISOTROPY
  2324. | BGFX_RESET_DEPTH_CLAMP
  2325. );
  2326. // Note: m_needToRefreshSwapchain is deliberately ignored when deciding whether to recreate the swapchain
  2327. // because it can happen several frames before submit is called with the new resolution.
  2328. // Instead, vkAcquireNextImageKHR and all draws to the backbuffer are skipped until the window size is updated.
  2329. // That also fixes a related issue where VK_ERROR_OUT_OF_DATE_KHR is returned from
  2330. // vkQueuePresentKHR when the window doesn't exist anymore, and vkGetPhysicalDeviceSurfaceCapabilitiesKHR
  2331. // fails with VK_ERROR_SURFACE_LOST_KHR.
  2332. if (false
  2333. || m_resolution.format != _resolution.format
  2334. || m_resolution.width != _resolution.width
  2335. || m_resolution.height != _resolution.height
  2336. || m_resolution.reset != flags
  2337. || m_backBuffer.m_swapChain.m_needToRecreateSurface)
  2338. {
  2339. flags &= ~BGFX_RESET_INTERNAL_FORCE;
  2340. if (m_backBuffer.m_nwh != g_platformData.nwh)
  2341. {
  2342. m_backBuffer.m_nwh = g_platformData.nwh;
  2343. }
  2344. m_resolution = _resolution;
  2345. m_resolution.reset = flags;
  2346. m_textVideoMem.resize(false, _resolution.width, _resolution.height);
  2347. m_textVideoMem.clear();
  2348. preReset();
  2349. m_backBuffer.update(m_commandBuffer, m_resolution);
  2350. postReset();
  2351. }
  2352. return suspended;
  2353. }
  2354. void setShaderUniform(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2355. {
  2356. if (_flags & kUniformFragmentBit)
  2357. {
  2358. bx::memCopy(&m_fsScratch[_regIndex], _val, _numRegs*16);
  2359. }
  2360. else
  2361. {
  2362. bx::memCopy(&m_vsScratch[_regIndex], _val, _numRegs*16);
  2363. }
  2364. }
  2365. void setShaderUniform4f(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2366. {
  2367. setShaderUniform(_flags, _regIndex, _val, _numRegs);
  2368. }
  2369. void setShaderUniform4x4f(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2370. {
  2371. setShaderUniform(_flags, _regIndex, _val, _numRegs);
  2372. }
  2373. void setFrameBuffer(FrameBufferHandle _fbh, bool _acquire = true)
  2374. {
  2375. BGFX_PROFILER_SCOPE("Vk::setFrameBuffer()", kColorFrame);
  2376. BX_ASSERT(false
  2377. || isValid(_fbh)
  2378. || NULL != m_backBuffer.m_nwh
  2379. , "Rendering to backbuffer in headless mode."
  2380. );
  2381. FrameBufferVK& newFrameBuffer = isValid(_fbh)
  2382. ? m_frameBuffers[_fbh.idx]
  2383. : m_backBuffer
  2384. ;
  2385. FrameBufferVK& oldFrameBuffer = isValid(m_fbh)
  2386. ? m_frameBuffers[m_fbh.idx]
  2387. : m_backBuffer
  2388. ;
  2389. if (NULL == oldFrameBuffer.m_nwh
  2390. && m_fbh.idx != _fbh.idx)
  2391. {
  2392. oldFrameBuffer.resolve();
  2393. for (uint8_t ii = 0, num = oldFrameBuffer.m_num; ii < num; ++ii)
  2394. {
  2395. TextureVK& texture = m_textures[oldFrameBuffer.m_texture[ii].idx];
  2396. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout);
  2397. if (VK_NULL_HANDLE != texture.m_singleMsaaImage)
  2398. {
  2399. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout, true);
  2400. }
  2401. }
  2402. if (isValid(oldFrameBuffer.m_depth) )
  2403. {
  2404. TextureVK& texture = m_textures[oldFrameBuffer.m_depth.idx];
  2405. const bool writeOnly = 0 != (texture.m_flags&BGFX_TEXTURE_RT_WRITE_ONLY);
  2406. if (!writeOnly)
  2407. {
  2408. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout);
  2409. }
  2410. }
  2411. }
  2412. if (NULL == newFrameBuffer.m_nwh)
  2413. {
  2414. for (uint8_t ii = 0, num = newFrameBuffer.m_num; ii < num; ++ii)
  2415. {
  2416. TextureVK& texture = m_textures[newFrameBuffer.m_texture[ii].idx];
  2417. texture.setImageMemoryBarrier(
  2418. m_commandBuffer
  2419. , VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  2420. );
  2421. }
  2422. if (isValid(newFrameBuffer.m_depth) )
  2423. {
  2424. TextureVK& texture = m_textures[newFrameBuffer.m_depth.idx];
  2425. texture.setImageMemoryBarrier(
  2426. m_commandBuffer
  2427. , VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL
  2428. );
  2429. }
  2430. newFrameBuffer.acquire(m_commandBuffer);
  2431. }
  2432. if (_acquire)
  2433. {
  2434. int64_t start = bx::getHPCounter();
  2435. newFrameBuffer.acquire(m_commandBuffer);
  2436. int64_t now = bx::getHPCounter();
  2437. if (NULL != newFrameBuffer.m_nwh)
  2438. {
  2439. m_presentElapsed += now - start;
  2440. }
  2441. }
  2442. m_fbh = _fbh;
  2443. }
  2444. void setDebugWireframe(bool _wireframe)
  2445. {
  2446. const bool wireframe = m_deviceFeatures.fillModeNonSolid && _wireframe;
  2447. if (m_wireframe != wireframe)
  2448. {
  2449. m_wireframe = wireframe;
  2450. m_pipelineStateCache.invalidate();
  2451. }
  2452. }
  2453. void setBlendState(VkPipelineColorBlendStateCreateInfo& _desc, uint64_t _state, uint32_t _rgba = 0)
  2454. {
  2455. VkPipelineColorBlendAttachmentState* bas = const_cast<VkPipelineColorBlendAttachmentState*>(_desc.pAttachments);
  2456. uint8_t writeMask = 0;
  2457. writeMask |= (_state & BGFX_STATE_WRITE_R) ? VK_COLOR_COMPONENT_R_BIT : 0;
  2458. writeMask |= (_state & BGFX_STATE_WRITE_G) ? VK_COLOR_COMPONENT_G_BIT : 0;
  2459. writeMask |= (_state & BGFX_STATE_WRITE_B) ? VK_COLOR_COMPONENT_B_BIT : 0;
  2460. writeMask |= (_state & BGFX_STATE_WRITE_A) ? VK_COLOR_COMPONENT_A_BIT : 0;
  2461. bas->blendEnable = !!(BGFX_STATE_BLEND_MASK & _state);
  2462. {
  2463. const uint32_t blend = uint32_t( (_state & BGFX_STATE_BLEND_MASK ) >> BGFX_STATE_BLEND_SHIFT);
  2464. const uint32_t equation = uint32_t( (_state & BGFX_STATE_BLEND_EQUATION_MASK) >> BGFX_STATE_BLEND_EQUATION_SHIFT);
  2465. const uint32_t srcRGB = (blend ) & 0xf;
  2466. const uint32_t dstRGB = (blend >> 4) & 0xf;
  2467. const uint32_t srcA = (blend >> 8) & 0xf;
  2468. const uint32_t dstA = (blend >> 12) & 0xf;
  2469. const uint32_t equRGB = (equation ) & 0x7;
  2470. const uint32_t equA = (equation >> 3) & 0x7;
  2471. bas->srcColorBlendFactor = s_blendFactor[srcRGB][0];
  2472. bas->dstColorBlendFactor = s_blendFactor[dstRGB][0];
  2473. bas->colorBlendOp = s_blendEquation[equRGB];
  2474. bas->srcAlphaBlendFactor = s_blendFactor[srcA][1];
  2475. bas->dstAlphaBlendFactor = s_blendFactor[dstA][1];
  2476. bas->alphaBlendOp = s_blendEquation[equA];
  2477. bas->colorWriteMask = writeMask;
  2478. }
  2479. const FrameBufferVK& frameBuffer = isValid(m_fbh)
  2480. ? m_frameBuffers[m_fbh.idx]
  2481. : m_backBuffer
  2482. ;
  2483. const uint32_t numAttachments = NULL == frameBuffer.m_nwh
  2484. ? frameBuffer.m_num
  2485. : 1
  2486. ;
  2487. if (!!(BGFX_STATE_BLEND_INDEPENDENT & _state)
  2488. && m_deviceFeatures.independentBlend )
  2489. {
  2490. for (uint32_t ii = 1, rgba = _rgba; ii < numAttachments; ++ii, rgba >>= 11)
  2491. {
  2492. ++bas;
  2493. bas->blendEnable = 0 != (rgba & 0x7ff);
  2494. const uint32_t src = (rgba ) & 0xf;
  2495. const uint32_t dst = (rgba >> 4) & 0xf;
  2496. const uint32_t equation = (rgba >> 8) & 0x7;
  2497. bas->srcColorBlendFactor = s_blendFactor[src][0];
  2498. bas->dstColorBlendFactor = s_blendFactor[dst][0];
  2499. bas->colorBlendOp = s_blendEquation[equation];
  2500. bas->srcAlphaBlendFactor = s_blendFactor[src][1];
  2501. bas->dstAlphaBlendFactor = s_blendFactor[dst][1];
  2502. bas->alphaBlendOp = s_blendEquation[equation];
  2503. bas->colorWriteMask = writeMask;
  2504. }
  2505. }
  2506. else
  2507. {
  2508. for (uint32_t ii = 1; ii < numAttachments; ++ii)
  2509. {
  2510. bx::memCopy(&bas[ii], bas, sizeof(VkPipelineColorBlendAttachmentState) );
  2511. }
  2512. }
  2513. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_COLOR_BLEND_STATE_CREATE_INFO;
  2514. _desc.pNext = NULL;
  2515. _desc.flags = 0;
  2516. _desc.logicOpEnable = VK_FALSE;
  2517. _desc.logicOp = VK_LOGIC_OP_CLEAR;
  2518. _desc.attachmentCount = numAttachments;
  2519. _desc.blendConstants[0] = 0.0f;
  2520. _desc.blendConstants[1] = 0.0f;
  2521. _desc.blendConstants[2] = 0.0f;
  2522. _desc.blendConstants[3] = 0.0f;
  2523. }
  2524. void setRasterizerState(VkPipelineRasterizationStateCreateInfo& _desc, uint64_t _state, bool _wireframe = false)
  2525. {
  2526. const uint32_t cull = (_state&BGFX_STATE_CULL_MASK) >> BGFX_STATE_CULL_SHIFT;
  2527. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO;
  2528. _desc.pNext = NULL;
  2529. _desc.flags = 0;
  2530. _desc.depthClampEnable = m_deviceFeatures.depthClamp && m_depthClamp;
  2531. _desc.rasterizerDiscardEnable = VK_FALSE;
  2532. _desc.polygonMode = m_deviceFeatures.fillModeNonSolid && _wireframe
  2533. ? VK_POLYGON_MODE_LINE
  2534. : VK_POLYGON_MODE_FILL
  2535. ;
  2536. _desc.cullMode = s_cullMode[cull];
  2537. _desc.frontFace = (_state&BGFX_STATE_FRONT_CCW) ? VK_FRONT_FACE_COUNTER_CLOCKWISE : VK_FRONT_FACE_CLOCKWISE;
  2538. _desc.depthBiasEnable = VK_FALSE;
  2539. _desc.depthBiasConstantFactor = 0.0f;
  2540. _desc.depthBiasClamp = 0.0f;
  2541. _desc.depthBiasSlopeFactor = 0.0f;
  2542. _desc.lineWidth = 1.0f;
  2543. }
  2544. void setConservativeRasterizerState(VkPipelineRasterizationConservativeStateCreateInfoEXT& _desc, uint64_t _state)
  2545. {
  2546. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_CONSERVATIVE_STATE_CREATE_INFO_EXT;
  2547. _desc.pNext = NULL;
  2548. _desc.flags = 0;
  2549. _desc.conservativeRasterizationMode = (_state&BGFX_STATE_CONSERVATIVE_RASTER)
  2550. ? VK_CONSERVATIVE_RASTERIZATION_MODE_OVERESTIMATE_EXT
  2551. : VK_CONSERVATIVE_RASTERIZATION_MODE_DISABLED_EXT
  2552. ;
  2553. _desc.extraPrimitiveOverestimationSize = 0.0f;
  2554. }
  2555. void setLineRasterizerState(VkPipelineRasterizationLineStateCreateInfoEXT& _desc, uint64_t _state)
  2556. {
  2557. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_LINE_STATE_CREATE_INFO_EXT;
  2558. _desc.pNext = NULL;
  2559. _desc.lineRasterizationMode = (_state & BGFX_STATE_LINEAA)
  2560. ? VK_LINE_RASTERIZATION_MODE_RECTANGULAR_SMOOTH_EXT
  2561. : VK_LINE_RASTERIZATION_MODE_DEFAULT_EXT
  2562. ;
  2563. _desc.stippledLineEnable = VK_FALSE;
  2564. _desc.lineStippleFactor = 0;
  2565. _desc.lineStipplePattern = 0;
  2566. }
  2567. void setDepthStencilState(VkPipelineDepthStencilStateCreateInfo& _desc, uint64_t _state, uint64_t _stencil = 0)
  2568. {
  2569. const uint32_t fstencil = unpackStencil(0, _stencil);
  2570. uint32_t func = (_state&BGFX_STATE_DEPTH_TEST_MASK)>>BGFX_STATE_DEPTH_TEST_SHIFT;
  2571. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_DEPTH_STENCIL_STATE_CREATE_INFO;
  2572. _desc.pNext = NULL;
  2573. _desc.flags = 0;
  2574. _desc.depthTestEnable = 0 != func;
  2575. _desc.depthWriteEnable = !!(BGFX_STATE_WRITE_Z & _state);
  2576. _desc.depthCompareOp = s_cmpFunc[func];
  2577. _desc.depthBoundsTestEnable = VK_FALSE;
  2578. _desc.stencilTestEnable = 0 != _stencil;
  2579. uint32_t bstencil = unpackStencil(1, _stencil);
  2580. uint32_t frontAndBack = bstencil != BGFX_STENCIL_NONE && bstencil != fstencil;
  2581. bstencil = frontAndBack ? bstencil : fstencil;
  2582. _desc.front.failOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_FAIL_S_MASK) >> BGFX_STENCIL_OP_FAIL_S_SHIFT];
  2583. _desc.front.passOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_PASS_Z_MASK) >> BGFX_STENCIL_OP_PASS_Z_SHIFT];
  2584. _desc.front.depthFailOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_FAIL_Z_MASK) >> BGFX_STENCIL_OP_FAIL_Z_SHIFT];
  2585. _desc.front.compareOp = s_cmpFunc[(fstencil & BGFX_STENCIL_TEST_MASK) >> BGFX_STENCIL_TEST_SHIFT];
  2586. _desc.front.compareMask = UINT32_MAX;
  2587. _desc.front.writeMask = UINT32_MAX;
  2588. _desc.front.reference = 0;
  2589. _desc.back.failOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_FAIL_S_MASK) >> BGFX_STENCIL_OP_FAIL_S_SHIFT];
  2590. _desc.back.passOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_PASS_Z_MASK) >> BGFX_STENCIL_OP_PASS_Z_SHIFT];
  2591. _desc.back.depthFailOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_FAIL_Z_MASK) >> BGFX_STENCIL_OP_FAIL_Z_SHIFT];
  2592. _desc.back.compareOp = s_cmpFunc[(bstencil&BGFX_STENCIL_TEST_MASK) >> BGFX_STENCIL_TEST_SHIFT];
  2593. _desc.back.compareMask = UINT32_MAX;
  2594. _desc.back.writeMask = UINT32_MAX;
  2595. _desc.back.reference = 0;
  2596. _desc.minDepthBounds = 0.0f;
  2597. _desc.maxDepthBounds = 1.0f;
  2598. }
  2599. void setInputLayout(VkPipelineVertexInputStateCreateInfo& _vertexInputState, uint8_t _numStream, const VertexLayout** _layout, const ProgramVK& _program, uint8_t _numInstanceData)
  2600. {
  2601. _vertexInputState.sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO;
  2602. _vertexInputState.pNext = NULL;
  2603. _vertexInputState.flags = 0;
  2604. _vertexInputState.vertexBindingDescriptionCount = 0;
  2605. _vertexInputState.vertexAttributeDescriptionCount = 0;
  2606. uint16_t unsettedAttr[Attrib::Count];
  2607. bx::memCopy(unsettedAttr, _program.m_vsh->m_attrMask, sizeof(uint16_t) * Attrib::Count);
  2608. for (uint8_t stream = 0; stream < _numStream; ++stream)
  2609. {
  2610. VertexLayout layout;
  2611. bx::memCopy(&layout, _layout[stream], sizeof(VertexLayout) );
  2612. const uint16_t* attrMask = _program.m_vsh->m_attrMask;
  2613. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  2614. {
  2615. uint16_t mask = attrMask[ii];
  2616. uint16_t attr = (layout.m_attributes[ii] & mask);
  2617. layout.m_attributes[ii] = attr == 0 || attr == UINT16_MAX ? UINT16_MAX : attr;
  2618. if (unsettedAttr[ii] && attr != UINT16_MAX)
  2619. {
  2620. unsettedAttr[ii] = 0;
  2621. }
  2622. }
  2623. fillVertexLayout(_program.m_vsh, _vertexInputState, layout);
  2624. }
  2625. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  2626. {
  2627. if (0 < unsettedAttr[ii])
  2628. {
  2629. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  2630. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  2631. inputAttrib->location = _program.m_vsh->m_attrRemap[ii];
  2632. inputAttrib->binding = 0;
  2633. inputAttrib->format = VK_FORMAT_R32G32B32_SFLOAT;
  2634. inputAttrib->offset = 0;
  2635. _vertexInputState.vertexAttributeDescriptionCount++;
  2636. }
  2637. }
  2638. if (0 < _numInstanceData)
  2639. {
  2640. fillInstanceBinding(_program.m_vsh, _vertexInputState, _numInstanceData);
  2641. }
  2642. }
  2643. VkResult getRenderPass(uint8_t _num, const VkFormat* _formats, const VkImageAspectFlags* _aspects, const bool* _resolve, VkSampleCountFlagBits _samples, ::VkRenderPass* _renderPass)
  2644. {
  2645. VkResult result = VK_SUCCESS;
  2646. if (VK_SAMPLE_COUNT_1_BIT == _samples)
  2647. {
  2648. _resolve = NULL;
  2649. }
  2650. bx::HashMurmur2A hash;
  2651. hash.begin();
  2652. hash.add(_samples);
  2653. hash.add(_formats, sizeof(VkFormat) * _num);
  2654. if (NULL != _resolve)
  2655. {
  2656. hash.add(_resolve, sizeof(bool) * _num);
  2657. }
  2658. uint32_t hashKey = hash.end();
  2659. VkRenderPass renderPass = m_renderPassCache.find(hashKey);
  2660. if (VK_NULL_HANDLE != renderPass)
  2661. {
  2662. *_renderPass = renderPass;
  2663. return result;
  2664. }
  2665. VkAttachmentDescription ad[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS * 2];
  2666. for (uint8_t ii = 0; ii < (_num * 2); ++ii)
  2667. {
  2668. ad[ii].flags = 0;
  2669. ad[ii].format = VK_FORMAT_UNDEFINED;
  2670. ad[ii].samples = _samples;
  2671. ad[ii].loadOp = VK_ATTACHMENT_LOAD_OP_LOAD;
  2672. ad[ii].storeOp = VK_ATTACHMENT_STORE_OP_STORE;
  2673. ad[ii].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
  2674. ad[ii].stencilStoreOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
  2675. ad[ii].initialLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2676. ad[ii].finalLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2677. }
  2678. VkAttachmentReference colorAr[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2679. VkAttachmentReference resolveAr[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2680. VkAttachmentReference depthAr;
  2681. uint32_t numColorAr = 0;
  2682. uint32_t numResolveAr = 0;
  2683. colorAr[0].attachment = VK_ATTACHMENT_UNUSED;
  2684. colorAr[0].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2685. resolveAr[0].attachment = VK_ATTACHMENT_UNUSED;
  2686. resolveAr[0].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2687. depthAr.attachment = VK_ATTACHMENT_UNUSED;
  2688. depthAr.layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2689. for (uint8_t ii = 0; ii < _num; ++ii)
  2690. {
  2691. ad[ii].format = _formats[ii];
  2692. if (_aspects[ii] & VK_IMAGE_ASPECT_COLOR_BIT)
  2693. {
  2694. colorAr[numColorAr].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2695. colorAr[numColorAr].attachment = ii;
  2696. resolveAr[numColorAr].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2697. resolveAr[numColorAr].attachment = VK_ATTACHMENT_UNUSED;
  2698. if (NULL != _resolve
  2699. && _resolve[ii])
  2700. {
  2701. const uint32_t resolve = _num + numResolveAr;
  2702. ad[resolve].format = _formats[ii];
  2703. ad[resolve].samples = VK_SAMPLE_COUNT_1_BIT;
  2704. ad[resolve].loadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
  2705. resolveAr[numColorAr].attachment = resolve;
  2706. numResolveAr++;
  2707. }
  2708. numColorAr++;
  2709. }
  2710. else if (_aspects[ii] & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  2711. {
  2712. ad[ii].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_LOAD;
  2713. ad[ii].stencilStoreOp = VK_ATTACHMENT_STORE_OP_STORE;
  2714. ad[ii].initialLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2715. ad[ii].finalLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2716. depthAr.layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2717. depthAr.attachment = ii;
  2718. }
  2719. }
  2720. VkSubpassDescription sd[1];
  2721. sd[0].flags = 0;
  2722. sd[0].pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS;
  2723. sd[0].inputAttachmentCount = 0;
  2724. sd[0].pInputAttachments = NULL;
  2725. sd[0].colorAttachmentCount = bx::max<uint32_t>(numColorAr, 1);
  2726. sd[0].pColorAttachments = colorAr;
  2727. sd[0].pResolveAttachments = resolveAr;
  2728. sd[0].pDepthStencilAttachment = &depthAr;
  2729. sd[0].preserveAttachmentCount = 0;
  2730. sd[0].pPreserveAttachments = NULL;
  2731. const VkPipelineStageFlags graphicsStages = 0
  2732. | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  2733. | VK_PIPELINE_STAGE_VERTEX_INPUT_BIT
  2734. | VK_PIPELINE_STAGE_VERTEX_SHADER_BIT
  2735. | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT
  2736. | VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT
  2737. | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT
  2738. | VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT
  2739. ;
  2740. const VkPipelineStageFlags outsideStages = 0
  2741. | graphicsStages
  2742. | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  2743. | VK_PIPELINE_STAGE_TRANSFER_BIT
  2744. ;
  2745. VkSubpassDependency dep[2];
  2746. dep[0].srcSubpass = VK_SUBPASS_EXTERNAL;
  2747. dep[0].dstSubpass = 0;
  2748. dep[0].srcStageMask = outsideStages;
  2749. dep[0].dstStageMask = graphicsStages;
  2750. dep[0].srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  2751. dep[0].dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  2752. dep[0].dependencyFlags = 0;
  2753. dep[1].srcSubpass = BX_COUNTOF(sd)-1;
  2754. dep[1].dstSubpass = VK_SUBPASS_EXTERNAL;
  2755. dep[1].srcStageMask = graphicsStages;
  2756. dep[1].dstStageMask = outsideStages;
  2757. dep[1].srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  2758. dep[1].dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  2759. dep[1].dependencyFlags = 0;
  2760. VkRenderPassCreateInfo rpi;
  2761. rpi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO;
  2762. rpi.pNext = NULL;
  2763. rpi.flags = 0;
  2764. rpi.attachmentCount = _num + numResolveAr;
  2765. rpi.pAttachments = ad;
  2766. rpi.subpassCount = BX_COUNTOF(sd);
  2767. rpi.pSubpasses = sd;
  2768. rpi.dependencyCount = BX_COUNTOF(dep);
  2769. rpi.pDependencies = dep;
  2770. result = vkCreateRenderPass(m_device, &rpi, m_allocatorCb, &renderPass);
  2771. if (VK_SUCCESS != result)
  2772. {
  2773. BX_TRACE("Create render pass error: vkCreateRenderPass failed %d: %s.", result, getName(result) );
  2774. return result;
  2775. }
  2776. m_renderPassCache.add(hashKey, renderPass);
  2777. *_renderPass = renderPass;
  2778. return result;
  2779. }
  2780. VkResult getRenderPass(uint8_t _num, const Attachment* _attachments, ::VkRenderPass* _renderPass)
  2781. {
  2782. VkFormat formats[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2783. VkImageAspectFlags aspects[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2784. VkSampleCountFlagBits samples = VK_SAMPLE_COUNT_1_BIT;
  2785. for (uint8_t ii = 0; ii < _num; ++ii)
  2786. {
  2787. const TextureVK& texture = m_textures[_attachments[ii].handle.idx];
  2788. formats[ii] = texture.m_format;
  2789. aspects[ii] = texture.m_aspectMask;
  2790. samples = texture.m_sampler.Sample;
  2791. }
  2792. return getRenderPass(_num, formats, aspects, NULL, samples, _renderPass);
  2793. }
  2794. VkResult getRenderPass(const SwapChainVK& swapChain, ::VkRenderPass* _renderPass)
  2795. {
  2796. const VkFormat formats[2] =
  2797. {
  2798. swapChain.m_sci.imageFormat,
  2799. swapChain.m_backBufferDepthStencil.m_format
  2800. };
  2801. const VkImageAspectFlags aspects[2] =
  2802. {
  2803. VK_IMAGE_ASPECT_COLOR_BIT,
  2804. swapChain.m_backBufferDepthStencil.m_aspectMask
  2805. };
  2806. const bool resolve[2] =
  2807. {
  2808. swapChain.m_supportsManualResolve ? false : true,
  2809. false
  2810. };
  2811. const VkSampleCountFlagBits samples = swapChain.m_sampler.Sample;
  2812. return getRenderPass(BX_COUNTOF(formats), formats, aspects, resolve, samples, _renderPass);
  2813. }
  2814. VkSampler getSampler(uint32_t _flags, VkFormat _format, const float _palette[][4])
  2815. {
  2816. uint32_t index = ((_flags & BGFX_SAMPLER_BORDER_COLOR_MASK) >> BGFX_SAMPLER_BORDER_COLOR_SHIFT);
  2817. index = bx::min<uint32_t>(BGFX_CONFIG_MAX_COLOR_PALETTE - 1, index);
  2818. _flags &= BGFX_SAMPLER_BITS_MASK;
  2819. _flags &= ~(m_deviceFeatures.samplerAnisotropy ? 0 : (BGFX_SAMPLER_MIN_ANISOTROPIC | BGFX_SAMPLER_MAG_ANISOTROPIC) );
  2820. // Force both min+max anisotropic, can't be set individually.
  2821. _flags |= 0 != (_flags & (BGFX_SAMPLER_MIN_ANISOTROPIC|BGFX_SAMPLER_MAG_ANISOTROPIC) )
  2822. ? BGFX_SAMPLER_MIN_ANISOTROPIC|BGFX_SAMPLER_MAG_ANISOTROPIC
  2823. : 0
  2824. ;
  2825. const float* rgba = NULL == _palette
  2826. ? NULL
  2827. : _palette[index]
  2828. ;
  2829. const bool needColor = true
  2830. && needBorderColor(_flags)
  2831. && NULL != rgba
  2832. && m_borderColorSupport
  2833. ;
  2834. uint32_t hashKey;
  2835. VkSampler sampler = VK_NULL_HANDLE;
  2836. if (!needColor)
  2837. {
  2838. bx::HashMurmur2A hash;
  2839. hash.begin();
  2840. hash.add(_flags);
  2841. hash.add(-1);
  2842. hash.add(VK_FORMAT_UNDEFINED);
  2843. hashKey = hash.end();
  2844. sampler = m_samplerCache.find(hashKey);
  2845. }
  2846. else
  2847. {
  2848. bx::HashMurmur2A hash;
  2849. hash.begin();
  2850. hash.add(_flags);
  2851. hash.add(index);
  2852. hash.add(_format);
  2853. hashKey = hash.end();
  2854. const uint32_t colorHashKey = m_samplerBorderColorCache.find(hashKey);
  2855. const uint32_t newColorHashKey = bx::hash<bx::HashMurmur2A>(rgba, sizeof(float) * 4);
  2856. if (newColorHashKey == colorHashKey)
  2857. {
  2858. sampler = m_samplerCache.find(hashKey);
  2859. }
  2860. else
  2861. {
  2862. m_samplerBorderColorCache.add(hashKey, newColorHashKey);
  2863. }
  2864. }
  2865. if (VK_NULL_HANDLE != sampler)
  2866. {
  2867. return sampler;
  2868. }
  2869. const uint32_t cmpFunc = (_flags&BGFX_SAMPLER_COMPARE_MASK)>>BGFX_SAMPLER_COMPARE_SHIFT;
  2870. const float maxLodBias = m_deviceProperties.limits.maxSamplerLodBias;
  2871. const float lodBias = bx::clamp(float(BGFX_CONFIG_MIP_LOD_BIAS), -maxLodBias, maxLodBias);
  2872. VkSamplerCreateInfo sci;
  2873. sci.sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO;
  2874. sci.pNext = NULL;
  2875. sci.flags = 0;
  2876. sci.magFilter = _flags & BGFX_SAMPLER_MAG_POINT ? VK_FILTER_NEAREST : VK_FILTER_LINEAR;
  2877. sci.minFilter = _flags & BGFX_SAMPLER_MIN_POINT ? VK_FILTER_NEAREST : VK_FILTER_LINEAR;
  2878. sci.mipmapMode = _flags & BGFX_SAMPLER_MIP_POINT ? VK_SAMPLER_MIPMAP_MODE_NEAREST : VK_SAMPLER_MIPMAP_MODE_LINEAR;
  2879. sci.addressModeU = s_textureAddress[(_flags&BGFX_SAMPLER_U_MASK)>>BGFX_SAMPLER_U_SHIFT];
  2880. sci.addressModeV = s_textureAddress[(_flags&BGFX_SAMPLER_V_MASK)>>BGFX_SAMPLER_V_SHIFT];
  2881. sci.addressModeW = s_textureAddress[(_flags&BGFX_SAMPLER_W_MASK)>>BGFX_SAMPLER_W_SHIFT];
  2882. sci.mipLodBias = lodBias;
  2883. sci.anisotropyEnable = !!(_flags & (BGFX_SAMPLER_MIN_ANISOTROPIC | BGFX_SAMPLER_MAG_ANISOTROPIC) );
  2884. sci.maxAnisotropy = m_maxAnisotropy;
  2885. sci.compareEnable = 0 != cmpFunc;
  2886. sci.compareOp = s_cmpFunc[cmpFunc];
  2887. sci.minLod = 0.0f;
  2888. sci.maxLod = VK_LOD_CLAMP_NONE;
  2889. sci.borderColor = VK_BORDER_COLOR_FLOAT_OPAQUE_BLACK;
  2890. sci.unnormalizedCoordinates = VK_FALSE;
  2891. VkSamplerCustomBorderColorCreateInfoEXT cbcci;
  2892. if (needColor)
  2893. {
  2894. cbcci.sType = VK_STRUCTURE_TYPE_SAMPLER_CUSTOM_BORDER_COLOR_CREATE_INFO_EXT;
  2895. cbcci.pNext = NULL;
  2896. cbcci.format = _format;
  2897. bx::memCopy(cbcci.customBorderColor.float32, rgba, sizeof(cbcci.customBorderColor.float32) );
  2898. sci.pNext = &cbcci;
  2899. sci.borderColor = VK_BORDER_COLOR_FLOAT_CUSTOM_EXT;
  2900. }
  2901. VK_CHECK(vkCreateSampler(m_device, &sci, m_allocatorCb, &sampler) );
  2902. m_samplerCache.add(hashKey, sampler);
  2903. return sampler;
  2904. }
  2905. VkImageView getCachedImageView(TextureHandle _handle, uint32_t _mip, uint32_t _numMips, VkImageViewType _type, bool _stencil = false)
  2906. {
  2907. const TextureVK& texture = m_textures[_handle.idx];
  2908. _stencil = _stencil && !!(texture.m_aspectMask & VK_IMAGE_ASPECT_STENCIL_BIT);
  2909. bx::HashMurmur2A hash;
  2910. hash.begin();
  2911. hash.add(_handle.idx);
  2912. hash.add(_mip);
  2913. hash.add(_numMips);
  2914. hash.add(_type);
  2915. hash.add(_stencil);
  2916. uint32_t hashKey = hash.end();
  2917. VkImageView* viewCached = m_imageViewCache.find(hashKey);
  2918. if (NULL != viewCached)
  2919. {
  2920. return *viewCached;
  2921. }
  2922. const VkImageAspectFlags aspectMask = 0
  2923. | VK_IMAGE_ASPECT_COLOR_BIT
  2924. | ( _stencil ? VK_IMAGE_ASPECT_STENCIL_BIT : VK_IMAGE_ASPECT_DEPTH_BIT)
  2925. ;
  2926. VkImageView view;
  2927. VK_CHECK(texture.createView(0, texture.m_numSides, _mip, _numMips, _type, aspectMask, false, &view) );
  2928. m_imageViewCache.add(hashKey, view, _handle.idx);
  2929. return view;
  2930. }
  2931. VkPipeline getPipeline(ProgramHandle _program)
  2932. {
  2933. ProgramVK& program = m_program[_program.idx];
  2934. bx::HashMurmur2A murmur;
  2935. murmur.begin();
  2936. murmur.add(program.m_vsh->m_hash);
  2937. const uint32_t hash = murmur.end();
  2938. VkPipeline pipeline = m_pipelineStateCache.find(hash);
  2939. if (VK_NULL_HANDLE != pipeline)
  2940. {
  2941. return pipeline;
  2942. }
  2943. VkComputePipelineCreateInfo cpci;
  2944. cpci.sType = VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO;
  2945. cpci.pNext = NULL;
  2946. cpci.flags = 0;
  2947. cpci.stage.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  2948. cpci.stage.pNext = NULL;
  2949. cpci.stage.flags = 0;
  2950. cpci.stage.stage = VK_SHADER_STAGE_COMPUTE_BIT;
  2951. cpci.stage.module = program.m_vsh->m_module;
  2952. cpci.stage.pName = "main";
  2953. cpci.stage.pSpecializationInfo = NULL;
  2954. cpci.layout = program.m_pipelineLayout;
  2955. cpci.basePipelineHandle = VK_NULL_HANDLE;
  2956. cpci.basePipelineIndex = 0;
  2957. VK_CHECK(vkCreateComputePipelines(m_device, m_pipelineCache, 1, &cpci, m_allocatorCb, &pipeline) );
  2958. m_pipelineStateCache.add(hash, pipeline);
  2959. return pipeline;
  2960. }
  2961. VkPipeline getPipeline(uint64_t _state, uint64_t _stencil, uint8_t _numStreams, const VertexLayout** _layouts, ProgramHandle _program, uint8_t _numInstanceData)
  2962. {
  2963. ProgramVK& program = m_program[_program.idx];
  2964. _state &= 0
  2965. | BGFX_STATE_WRITE_MASK
  2966. | BGFX_STATE_DEPTH_TEST_MASK
  2967. | BGFX_STATE_BLEND_MASK
  2968. | BGFX_STATE_BLEND_EQUATION_MASK
  2969. | (g_caps.supported & BGFX_CAPS_BLEND_INDEPENDENT ? BGFX_STATE_BLEND_INDEPENDENT : 0)
  2970. | BGFX_STATE_BLEND_ALPHA_TO_COVERAGE
  2971. | BGFX_STATE_CULL_MASK
  2972. | BGFX_STATE_FRONT_CCW
  2973. | BGFX_STATE_MSAA
  2974. | (m_lineAASupport ? BGFX_STATE_LINEAA : 0)
  2975. | (g_caps.supported & BGFX_CAPS_CONSERVATIVE_RASTER ? BGFX_STATE_CONSERVATIVE_RASTER : 0)
  2976. | BGFX_STATE_PT_MASK
  2977. ;
  2978. _stencil &= packStencil(~BGFX_STENCIL_FUNC_REF_MASK, ~BGFX_STENCIL_FUNC_REF_MASK);
  2979. VertexLayout layout;
  2980. if (0 < _numStreams)
  2981. {
  2982. bx::memCopy(&layout, _layouts[0], sizeof(VertexLayout) );
  2983. const uint16_t* attrMask = program.m_vsh->m_attrMask;
  2984. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  2985. {
  2986. uint16_t mask = attrMask[ii];
  2987. uint16_t attr = (layout.m_attributes[ii] & mask);
  2988. layout.m_attributes[ii] = attr == 0 ? UINT16_MAX : attr == UINT16_MAX ? 0 : attr;
  2989. }
  2990. }
  2991. const FrameBufferVK& frameBuffer = isValid(m_fbh)
  2992. ? m_frameBuffers[m_fbh.idx]
  2993. : m_backBuffer
  2994. ;
  2995. bx::HashMurmur2A murmur;
  2996. murmur.begin();
  2997. murmur.add(_state);
  2998. murmur.add(_stencil);
  2999. murmur.add(program.m_vsh->m_hash);
  3000. murmur.add(program.m_vsh->m_attrMask, sizeof(program.m_vsh->m_attrMask) );
  3001. if (NULL != program.m_fsh)
  3002. {
  3003. murmur.add(program.m_fsh->m_hash);
  3004. }
  3005. for (uint8_t ii = 0; ii < _numStreams; ++ii)
  3006. {
  3007. murmur.add(_layouts[ii]->m_hash);
  3008. }
  3009. murmur.add(layout.m_attributes, sizeof(layout.m_attributes) );
  3010. murmur.add(_numInstanceData);
  3011. murmur.add(frameBuffer.m_renderPass);
  3012. const uint32_t hash = murmur.end();
  3013. VkPipeline pipeline = m_pipelineStateCache.find(hash);
  3014. if (VK_NULL_HANDLE != pipeline)
  3015. {
  3016. return pipeline;
  3017. }
  3018. VkPipelineColorBlendAttachmentState blendAttachmentState[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  3019. VkPipelineColorBlendStateCreateInfo colorBlendState;
  3020. colorBlendState.pAttachments = blendAttachmentState;
  3021. setBlendState(colorBlendState, _state);
  3022. VkPipelineInputAssemblyStateCreateInfo inputAssemblyState;
  3023. inputAssemblyState.sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO;
  3024. inputAssemblyState.pNext = NULL;
  3025. inputAssemblyState.flags = 0;
  3026. inputAssemblyState.topology = s_primInfo[(_state&BGFX_STATE_PT_MASK) >> BGFX_STATE_PT_SHIFT].m_topology;
  3027. inputAssemblyState.primitiveRestartEnable = VK_FALSE;
  3028. VkPipelineRasterizationStateCreateInfo rasterizationState;
  3029. setRasterizerState(rasterizationState, _state, m_wireframe);
  3030. VkBaseInStructure* nextRasterizationState = (VkBaseInStructure*)&rasterizationState;
  3031. VkPipelineRasterizationConservativeStateCreateInfoEXT conservativeRasterizationState;
  3032. if (s_extension[Extension::EXT_conservative_rasterization].m_supported)
  3033. {
  3034. nextRasterizationState->pNext = (VkBaseInStructure*)&conservativeRasterizationState;
  3035. nextRasterizationState = (VkBaseInStructure*)&conservativeRasterizationState;
  3036. setConservativeRasterizerState(conservativeRasterizationState, _state);
  3037. }
  3038. VkPipelineRasterizationLineStateCreateInfoEXT lineRasterizationState;
  3039. if (m_lineAASupport)
  3040. {
  3041. nextRasterizationState->pNext = (VkBaseInStructure*)&lineRasterizationState;
  3042. nextRasterizationState = (VkBaseInStructure*)&lineRasterizationState;
  3043. setLineRasterizerState(lineRasterizationState, _state);
  3044. }
  3045. VkPipelineDepthStencilStateCreateInfo depthStencilState;
  3046. setDepthStencilState(depthStencilState, _state, _stencil);
  3047. VkVertexInputBindingDescription inputBinding[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  3048. VkVertexInputAttributeDescription inputAttrib[Attrib::Count + BGFX_CONFIG_MAX_INSTANCE_DATA_COUNT];
  3049. VkPipelineVertexInputStateCreateInfo vertexInputState;
  3050. vertexInputState.pVertexBindingDescriptions = inputBinding;
  3051. vertexInputState.pVertexAttributeDescriptions = inputAttrib;
  3052. setInputLayout(vertexInputState, _numStreams, _layouts, program, _numInstanceData);
  3053. const VkDynamicState dynamicStates[] =
  3054. {
  3055. VK_DYNAMIC_STATE_VIEWPORT,
  3056. VK_DYNAMIC_STATE_SCISSOR,
  3057. VK_DYNAMIC_STATE_BLEND_CONSTANTS,
  3058. VK_DYNAMIC_STATE_STENCIL_REFERENCE,
  3059. };
  3060. VkPipelineDynamicStateCreateInfo dynamicState;
  3061. dynamicState.sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO;
  3062. dynamicState.pNext = NULL;
  3063. dynamicState.flags = 0;
  3064. dynamicState.dynamicStateCount = BX_COUNTOF(dynamicStates);
  3065. dynamicState.pDynamicStates = dynamicStates;
  3066. VkPipelineShaderStageCreateInfo shaderStages[2];
  3067. shaderStages[0].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  3068. shaderStages[0].pNext = NULL;
  3069. shaderStages[0].flags = 0;
  3070. shaderStages[0].stage = VK_SHADER_STAGE_VERTEX_BIT;
  3071. shaderStages[0].module = program.m_vsh->m_module;
  3072. shaderStages[0].pName = "main";
  3073. shaderStages[0].pSpecializationInfo = NULL;
  3074. if (NULL != program.m_fsh)
  3075. {
  3076. shaderStages[1].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  3077. shaderStages[1].pNext = NULL;
  3078. shaderStages[1].flags = 0;
  3079. shaderStages[1].stage = VK_SHADER_STAGE_FRAGMENT_BIT;
  3080. shaderStages[1].module = program.m_fsh->m_module;
  3081. shaderStages[1].pName = "main";
  3082. shaderStages[1].pSpecializationInfo = NULL;
  3083. }
  3084. VkPipelineViewportStateCreateInfo viewportState;
  3085. viewportState.sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO;
  3086. viewportState.pNext = NULL;
  3087. viewportState.flags = 0;
  3088. viewportState.viewportCount = 1;
  3089. viewportState.pViewports = NULL;
  3090. viewportState.scissorCount = 1;
  3091. viewportState.pScissors = NULL;
  3092. VkPipelineMultisampleStateCreateInfo multisampleState;
  3093. multisampleState.sType = VK_STRUCTURE_TYPE_PIPELINE_MULTISAMPLE_STATE_CREATE_INFO;
  3094. multisampleState.pNext = NULL;
  3095. multisampleState.flags = 0;
  3096. multisampleState.rasterizationSamples = frameBuffer.m_sampler.Sample;
  3097. multisampleState.sampleShadingEnable = VK_FALSE;
  3098. multisampleState.minSampleShading = 0.0f;
  3099. multisampleState.pSampleMask = NULL;
  3100. multisampleState.alphaToCoverageEnable = !!(BGFX_STATE_BLEND_ALPHA_TO_COVERAGE & _state);
  3101. multisampleState.alphaToOneEnable = VK_FALSE;
  3102. VkGraphicsPipelineCreateInfo graphicsPipeline;
  3103. graphicsPipeline.sType = VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO;
  3104. graphicsPipeline.pNext = NULL;
  3105. graphicsPipeline.flags = 0;
  3106. graphicsPipeline.stageCount = NULL == program.m_fsh ? 1 : 2;
  3107. graphicsPipeline.pStages = shaderStages;
  3108. graphicsPipeline.pVertexInputState = &vertexInputState;
  3109. graphicsPipeline.pInputAssemblyState = &inputAssemblyState;
  3110. graphicsPipeline.pTessellationState = NULL;
  3111. graphicsPipeline.pViewportState = &viewportState;
  3112. graphicsPipeline.pRasterizationState = &rasterizationState;
  3113. graphicsPipeline.pMultisampleState = &multisampleState;
  3114. graphicsPipeline.pDepthStencilState = &depthStencilState;
  3115. graphicsPipeline.pColorBlendState = &colorBlendState;
  3116. graphicsPipeline.pDynamicState = &dynamicState;
  3117. graphicsPipeline.layout = program.m_pipelineLayout;
  3118. graphicsPipeline.renderPass = frameBuffer.m_renderPass;
  3119. graphicsPipeline.subpass = 0;
  3120. graphicsPipeline.basePipelineHandle = VK_NULL_HANDLE;
  3121. graphicsPipeline.basePipelineIndex = 0;
  3122. uint32_t length = g_callback->cacheReadSize(hash);
  3123. bool cached = length > 0;
  3124. void* cachedData = NULL;
  3125. VkPipelineCacheCreateInfo pcci;
  3126. pcci.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO;
  3127. pcci.pNext = NULL;
  3128. pcci.flags = 0;
  3129. pcci.initialDataSize = 0;
  3130. pcci.pInitialData = NULL;
  3131. if (cached)
  3132. {
  3133. cachedData = bx::alloc(g_allocator, length);
  3134. if (g_callback->cacheRead(hash, cachedData, length) )
  3135. {
  3136. BX_TRACE("Loading cached pipeline state (size %d).", length);
  3137. bx::MemoryReader reader(cachedData, length);
  3138. pcci.initialDataSize = (size_t)reader.remaining();
  3139. pcci.pInitialData = reader.getDataPtr();
  3140. }
  3141. }
  3142. VkPipelineCache cache;
  3143. VK_CHECK(vkCreatePipelineCache(m_device, &pcci, m_allocatorCb, &cache) );
  3144. VK_CHECK(vkCreateGraphicsPipelines(
  3145. m_device
  3146. , cache
  3147. , 1
  3148. , &graphicsPipeline
  3149. , m_allocatorCb
  3150. , &pipeline
  3151. ) );
  3152. m_pipelineStateCache.add(hash, pipeline);
  3153. size_t dataSize;
  3154. VK_CHECK(vkGetPipelineCacheData(m_device, cache, &dataSize, NULL) );
  3155. if (0 < dataSize)
  3156. {
  3157. if (length < dataSize)
  3158. {
  3159. cachedData = bx::realloc(g_allocator, cachedData, dataSize);
  3160. }
  3161. VK_CHECK(vkGetPipelineCacheData(m_device, cache, &dataSize, cachedData) );
  3162. g_callback->cacheWrite(hash, cachedData, (uint32_t)dataSize);
  3163. }
  3164. VK_CHECK(vkMergePipelineCaches(m_device, m_pipelineCache, 1, &cache) );
  3165. vkDestroy(cache);
  3166. if (NULL != cachedData)
  3167. {
  3168. bx::free(g_allocator, cachedData);
  3169. }
  3170. return pipeline;
  3171. }
  3172. VkDescriptorSet getDescriptorSet(const ProgramVK& program, const RenderBind& renderBind, const ScratchBufferVK& scratchBuffer, const float _palette[][4])
  3173. {
  3174. VkDescriptorSet descriptorSet;
  3175. VkDescriptorSetAllocateInfo dsai;
  3176. dsai.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO;
  3177. dsai.pNext = NULL;
  3178. dsai.descriptorPool = m_descriptorPool;
  3179. dsai.descriptorSetCount = 1;
  3180. dsai.pSetLayouts = &program.m_descriptorSetLayout;
  3181. VK_CHECK(vkAllocateDescriptorSets(m_device, &dsai, &descriptorSet) );
  3182. VkDescriptorImageInfo imageInfo[BGFX_CONFIG_MAX_TEXTURE_SAMPLERS];
  3183. VkDescriptorBufferInfo bufferInfo[BGFX_CONFIG_MAX_TEXTURE_SAMPLERS];
  3184. constexpr uint32_t kMaxDescriptorSets = 2 * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS + 2;
  3185. VkWriteDescriptorSet wds[kMaxDescriptorSets] = {};
  3186. uint32_t wdsCount = 0;
  3187. uint32_t bufferCount = 0;
  3188. uint32_t imageCount = 0;
  3189. for (uint32_t stage = 0; stage < BGFX_CONFIG_MAX_TEXTURE_SAMPLERS; ++stage)
  3190. {
  3191. const Binding& bind = renderBind.m_bind[stage];
  3192. const BindInfo& bindInfo = program.m_bindInfo[stage];
  3193. if (kInvalidHandle != bind.m_idx
  3194. && isValid(bindInfo.uniformHandle) )
  3195. {
  3196. switch (bind.m_type)
  3197. {
  3198. case Binding::Image:
  3199. {
  3200. const bool isImageDescriptor = BindType::Image == bindInfo.type;
  3201. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3202. wds[wdsCount].pNext = NULL;
  3203. wds[wdsCount].dstSet = descriptorSet;
  3204. wds[wdsCount].dstBinding = bindInfo.binding;
  3205. wds[wdsCount].dstArrayElement = 0;
  3206. wds[wdsCount].descriptorCount = 1;
  3207. wds[wdsCount].descriptorType = isImageDescriptor
  3208. ? VK_DESCRIPTOR_TYPE_STORAGE_IMAGE
  3209. : VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE
  3210. ;
  3211. wds[wdsCount].pImageInfo = NULL;
  3212. wds[wdsCount].pBufferInfo = NULL;
  3213. wds[wdsCount].pTexelBufferView = NULL;
  3214. const TextureVK& texture = m_textures[bind.m_idx];
  3215. VkImageViewType type = texture.m_type;
  3216. if (UINT32_MAX != bindInfo.index)
  3217. {
  3218. type = program.m_textures[bindInfo.index].type;
  3219. }
  3220. else if (type == VK_IMAGE_VIEW_TYPE_CUBE
  3221. || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY)
  3222. {
  3223. type = VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  3224. }
  3225. BX_ASSERT(
  3226. texture.m_currentImageLayout == texture.m_sampledLayout
  3227. , "Mismatching image layout. Texture currently used as a framebuffer attachment?"
  3228. );
  3229. imageInfo[imageCount].imageLayout = texture.m_sampledLayout;
  3230. imageInfo[imageCount].sampler = VK_NULL_HANDLE;
  3231. imageInfo[imageCount].imageView = getCachedImageView(
  3232. { bind.m_idx }
  3233. , bind.m_mip
  3234. , 1
  3235. , type
  3236. );
  3237. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3238. ++imageCount;
  3239. ++wdsCount;
  3240. }
  3241. break;
  3242. case Binding::VertexBuffer:
  3243. case Binding::IndexBuffer:
  3244. {
  3245. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3246. wds[wdsCount].pNext = NULL;
  3247. wds[wdsCount].dstSet = descriptorSet;
  3248. wds[wdsCount].dstBinding = bindInfo.binding;
  3249. wds[wdsCount].dstArrayElement = 0;
  3250. wds[wdsCount].descriptorCount = 1;
  3251. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
  3252. wds[wdsCount].pImageInfo = NULL;
  3253. wds[wdsCount].pBufferInfo = NULL;
  3254. wds[wdsCount].pTexelBufferView = NULL;
  3255. const BufferVK& sb = bind.m_type == Binding::VertexBuffer
  3256. ? m_vertexBuffers[bind.m_idx]
  3257. : m_indexBuffers[bind.m_idx]
  3258. ;
  3259. bufferInfo[bufferCount].buffer = sb.m_buffer;
  3260. bufferInfo[bufferCount].offset = 0;
  3261. bufferInfo[bufferCount].range = sb.m_size;
  3262. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3263. ++bufferCount;
  3264. ++wdsCount;
  3265. }
  3266. break;
  3267. case Binding::Texture:
  3268. {
  3269. TextureVK& texture = m_textures[bind.m_idx];
  3270. const uint32_t samplerFlags = 0 == (BGFX_SAMPLER_INTERNAL_DEFAULT & bind.m_samplerFlags)
  3271. ? bind.m_samplerFlags
  3272. : (uint32_t)texture.m_flags
  3273. ;
  3274. const bool sampleStencil = !!(samplerFlags & BGFX_SAMPLER_SAMPLE_STENCIL);
  3275. VkSampler sampler = getSampler(samplerFlags, texture.m_format, _palette);
  3276. const VkImageViewType type = UINT32_MAX == bindInfo.index
  3277. ? texture.m_type
  3278. : program.m_textures[bindInfo.index].type
  3279. ;
  3280. BX_ASSERT(
  3281. texture.m_currentImageLayout == texture.m_sampledLayout
  3282. , "Mismatching image layout. Texture currently used as a framebuffer attachment?"
  3283. );
  3284. imageInfo[imageCount].imageLayout = texture.m_sampledLayout;
  3285. imageInfo[imageCount].sampler = sampler;
  3286. imageInfo[imageCount].imageView = getCachedImageView(
  3287. { bind.m_idx }
  3288. , 0
  3289. , texture.m_numMips
  3290. , type
  3291. , sampleStencil
  3292. );
  3293. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3294. wds[wdsCount].pNext = NULL;
  3295. wds[wdsCount].dstSet = descriptorSet;
  3296. wds[wdsCount].dstBinding = bindInfo.binding;
  3297. wds[wdsCount].dstArrayElement = 0;
  3298. wds[wdsCount].descriptorCount = 1;
  3299. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
  3300. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3301. wds[wdsCount].pBufferInfo = NULL;
  3302. wds[wdsCount].pTexelBufferView = NULL;
  3303. ++wdsCount;
  3304. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3305. wds[wdsCount].pNext = NULL;
  3306. wds[wdsCount].dstSet = descriptorSet;
  3307. wds[wdsCount].dstBinding = bindInfo.samplerBinding;
  3308. wds[wdsCount].dstArrayElement = 0;
  3309. wds[wdsCount].descriptorCount = 1;
  3310. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
  3311. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3312. wds[wdsCount].pBufferInfo = NULL;
  3313. wds[wdsCount].pTexelBufferView = NULL;
  3314. ++wdsCount;
  3315. ++imageCount;
  3316. }
  3317. break;
  3318. }
  3319. }
  3320. }
  3321. const uint32_t vsize = program.m_vsh->m_size;
  3322. const uint32_t fsize = NULL != program.m_fsh ? program.m_fsh->m_size : 0;
  3323. if (vsize > 0)
  3324. {
  3325. bufferInfo[bufferCount].buffer = scratchBuffer.m_buffer;
  3326. bufferInfo[bufferCount].offset = 0;
  3327. bufferInfo[bufferCount].range = vsize;
  3328. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3329. wds[wdsCount].pNext = NULL;
  3330. wds[wdsCount].dstSet = descriptorSet;
  3331. wds[wdsCount].dstBinding = program.m_vsh->m_uniformBinding;
  3332. wds[wdsCount].dstArrayElement = 0;
  3333. wds[wdsCount].descriptorCount = 1;
  3334. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  3335. wds[wdsCount].pImageInfo = NULL;
  3336. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3337. wds[wdsCount].pTexelBufferView = NULL;
  3338. ++wdsCount;
  3339. ++bufferCount;
  3340. }
  3341. if (fsize > 0)
  3342. {
  3343. bufferInfo[bufferCount].buffer = scratchBuffer.m_buffer;
  3344. bufferInfo[bufferCount].offset = 0;
  3345. bufferInfo[bufferCount].range = fsize;
  3346. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3347. wds[wdsCount].pNext = NULL;
  3348. wds[wdsCount].dstSet = descriptorSet;
  3349. wds[wdsCount].dstBinding = program.m_fsh->m_uniformBinding;
  3350. wds[wdsCount].dstArrayElement = 0;
  3351. wds[wdsCount].descriptorCount = 1;
  3352. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  3353. wds[wdsCount].pImageInfo = NULL;
  3354. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3355. wds[wdsCount].pTexelBufferView = NULL;
  3356. ++wdsCount;
  3357. ++bufferCount;
  3358. }
  3359. vkUpdateDescriptorSets(m_device, wdsCount, wds, 0, NULL);
  3360. VkDescriptorSet temp = descriptorSet;
  3361. release(temp);
  3362. return descriptorSet;
  3363. }
  3364. bool isSwapChainReadable(const SwapChainVK& _swapChain)
  3365. {
  3366. return true
  3367. && NULL != _swapChain.m_nwh
  3368. && _swapChain.m_needPresent
  3369. && _swapChain.m_supportsReadback
  3370. && bimg::imageConvert(bimg::TextureFormat::BGRA8, bimg::TextureFormat::Enum(_swapChain.m_colorFormat) )
  3371. ;
  3372. }
  3373. typedef void (*SwapChainReadFunc)(void* /*src*/, uint32_t /*width*/, uint32_t /*height*/, uint32_t /*pitch*/, const void* /*userData*/);
  3374. bool readSwapChain(const SwapChainVK& _swapChain, VkBuffer _buffer, VkDeviceMemory _memory, SwapChainReadFunc _func, const void* _userData = NULL)
  3375. {
  3376. if (isSwapChainReadable(_swapChain) )
  3377. {
  3378. // source for the copy is the last rendered swapchain image
  3379. const VkImage image = _swapChain.m_backBufferColorImage[_swapChain.m_backBufferColorIdx];
  3380. const VkImageLayout layout = _swapChain.m_backBufferColorImageLayout[_swapChain.m_backBufferColorIdx];
  3381. const uint32_t width = _swapChain.m_sci.imageExtent.width;
  3382. const uint32_t height = _swapChain.m_sci.imageExtent.height;
  3383. ReadbackVK readback;
  3384. readback.create(image, width, height, _swapChain.m_colorFormat);
  3385. const uint32_t pitch = readback.pitch();
  3386. readback.copyImageToBuffer(m_commandBuffer, _buffer, layout, VK_IMAGE_ASPECT_COLOR_BIT);
  3387. // stall for commandbuffer to finish
  3388. kick(true);
  3389. uint8_t* src;
  3390. VK_CHECK(vkMapMemory(m_device, _memory, 0, VK_WHOLE_SIZE, 0, (void**)&src) );
  3391. if (_swapChain.m_colorFormat == TextureFormat::RGBA8)
  3392. {
  3393. bimg::imageSwizzleBgra8(src, pitch, width, height, src, pitch);
  3394. _func(src, width, height, pitch, _userData);
  3395. }
  3396. else if (_swapChain.m_colorFormat == TextureFormat::BGRA8)
  3397. {
  3398. _func(src, width, height, pitch, _userData);
  3399. }
  3400. else
  3401. {
  3402. const uint8_t dstBpp = bimg::getBitsPerPixel(bimg::TextureFormat::BGRA8);
  3403. const uint32_t dstPitch = width * dstBpp / 8;
  3404. const uint32_t dstSize = height * dstPitch;
  3405. void* dst = bx::alloc(g_allocator, dstSize);
  3406. bimg::imageConvert(g_allocator, dst, bimg::TextureFormat::BGRA8, src, bimg::TextureFormat::Enum(_swapChain.m_colorFormat), width, height, 1);
  3407. _func(dst, width, height, dstPitch, _userData);
  3408. bx::free(g_allocator, dst);
  3409. }
  3410. vkUnmapMemory(m_device, _memory);
  3411. readback.destroy();
  3412. return true;
  3413. }
  3414. return false;
  3415. }
  3416. void capture()
  3417. {
  3418. if (m_captureSize > 0)
  3419. {
  3420. m_backBuffer.resolve();
  3421. auto callback = [](void* _src, uint32_t /*_width*/, uint32_t _height, uint32_t _pitch, const void* /*_userData*/)
  3422. {
  3423. const uint32_t size = _height * _pitch;
  3424. g_callback->captureFrame(_src, size);
  3425. };
  3426. readSwapChain(m_backBuffer.m_swapChain, m_captureBuffer, m_captureMemory, callback);
  3427. }
  3428. }
  3429. bool isVisible(Frame* _render, OcclusionQueryHandle _handle, bool _visible)
  3430. {
  3431. return _visible == (0 != _render->m_occlusion[_handle.idx]);
  3432. }
  3433. void commit(UniformBuffer& _uniformBuffer)
  3434. {
  3435. _uniformBuffer.reset();
  3436. for (;;)
  3437. {
  3438. uint32_t opcode = _uniformBuffer.read();
  3439. if (UniformType::End == opcode)
  3440. {
  3441. break;
  3442. }
  3443. UniformType::Enum type;
  3444. uint16_t loc;
  3445. uint16_t num;
  3446. uint16_t copy;
  3447. UniformBuffer::decodeOpcode(opcode, type, loc, num, copy);
  3448. const char* data;
  3449. if (copy)
  3450. {
  3451. data = _uniformBuffer.read(g_uniformTypeSize[type]*num);
  3452. }
  3453. else
  3454. {
  3455. UniformHandle handle;
  3456. bx::memCopy(&handle, _uniformBuffer.read(sizeof(UniformHandle) ), sizeof(UniformHandle) );
  3457. data = (const char*)m_uniforms[handle.idx];
  3458. }
  3459. switch ( (uint32_t)type)
  3460. {
  3461. case UniformType::Mat3:
  3462. case UniformType::Mat3|kUniformFragmentBit:
  3463. {
  3464. float* value = (float*)data;
  3465. for (uint32_t ii = 0, count = num/3; ii < count; ++ii, loc += 3*16, value += 9)
  3466. {
  3467. Matrix4 mtx;
  3468. mtx.un.val[ 0] = value[0];
  3469. mtx.un.val[ 1] = value[1];
  3470. mtx.un.val[ 2] = value[2];
  3471. mtx.un.val[ 3] = 0.0f;
  3472. mtx.un.val[ 4] = value[3];
  3473. mtx.un.val[ 5] = value[4];
  3474. mtx.un.val[ 6] = value[5];
  3475. mtx.un.val[ 7] = 0.0f;
  3476. mtx.un.val[ 8] = value[6];
  3477. mtx.un.val[ 9] = value[7];
  3478. mtx.un.val[10] = value[8];
  3479. mtx.un.val[11] = 0.0f;
  3480. setShaderUniform(uint8_t(type), loc, &mtx.un.val[0], 3);
  3481. }
  3482. }
  3483. break;
  3484. case UniformType::Sampler:
  3485. case UniformType::Sampler|kUniformFragmentBit:
  3486. // do nothing, but VkDescriptorSetImageInfo would be set before drawing
  3487. break;
  3488. case UniformType::Vec4:
  3489. case UniformType::Vec4 | kUniformFragmentBit:
  3490. case UniformType::Mat4:
  3491. case UniformType::Mat4 | kUniformFragmentBit:
  3492. {
  3493. setShaderUniform(uint8_t(type), loc, data, num);
  3494. }
  3495. break;
  3496. case UniformType::End:
  3497. break;
  3498. default:
  3499. BX_TRACE("%4d: INVALID 0x%08x, t %d, l %d, n %d, c %d", _uniformBuffer.getPos(), opcode, type, loc, num, copy);
  3500. break;
  3501. }
  3502. }
  3503. }
  3504. void clearQuad(const Rect& _rect, const Clear& _clear, const float _palette[][4])
  3505. {
  3506. VkClearRect rect[1];
  3507. rect[0].rect.offset.x = _rect.m_x;
  3508. rect[0].rect.offset.y = _rect.m_y;
  3509. rect[0].rect.extent.width = _rect.m_width;
  3510. rect[0].rect.extent.height = _rect.m_height;
  3511. rect[0].baseArrayLayer = 0;
  3512. rect[0].layerCount = 1;
  3513. uint32_t numMrt;
  3514. bgfx::TextureFormat::Enum mrtFormat[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  3515. VkImageAspectFlags depthAspectMask;
  3516. const FrameBufferVK& fb = isValid(m_fbh)
  3517. ? m_frameBuffers[m_fbh.idx]
  3518. : m_backBuffer
  3519. ;
  3520. if (NULL == fb.m_nwh)
  3521. {
  3522. numMrt = fb.m_num;
  3523. for (uint8_t ii = 0; ii < fb.m_num; ++ii)
  3524. {
  3525. mrtFormat[ii] = bgfx::TextureFormat::Enum(m_textures[fb.m_texture[ii].idx].m_requestedFormat);
  3526. }
  3527. depthAspectMask = isValid(fb.m_depth) ? m_textures[fb.m_depth.idx].m_aspectMask : 0;
  3528. rect[0].layerCount = fb.m_attachment[0].numLayers;
  3529. }
  3530. else
  3531. {
  3532. numMrt = 1;
  3533. mrtFormat[0] = fb.m_swapChain.m_colorFormat;
  3534. depthAspectMask = fb.m_swapChain.m_backBufferDepthStencil.m_aspectMask;
  3535. }
  3536. VkClearAttachment attachments[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS + 1];
  3537. uint32_t mrt = 0;
  3538. if (BGFX_CLEAR_COLOR & _clear.m_flags)
  3539. {
  3540. for (uint32_t ii = 0; ii < numMrt; ++ii)
  3541. {
  3542. attachments[mrt].colorAttachment = mrt;
  3543. attachments[mrt].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  3544. VkClearColorValue& clearValue = attachments[mrt].clearValue.color;
  3545. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(bimg::TextureFormat::Enum(mrtFormat[ii]) );
  3546. const bx::EncodingType::Enum type = bx::EncodingType::Enum(blockInfo.encoding);
  3547. if (BGFX_CLEAR_COLOR_USE_PALETTE & _clear.m_flags)
  3548. {
  3549. const uint8_t index = bx::min<uint8_t>(BGFX_CONFIG_MAX_COLOR_PALETTE-1, _clear.m_index[ii]);
  3550. switch (type)
  3551. {
  3552. case bx::EncodingType::Int:
  3553. case bx::EncodingType::Uint:
  3554. clearValue.int32[0] = int32_t(_palette[index][0]);
  3555. clearValue.int32[1] = int32_t(_palette[index][1]);
  3556. clearValue.int32[2] = int32_t(_palette[index][2]);
  3557. clearValue.int32[3] = int32_t(_palette[index][3]);
  3558. break;
  3559. default:
  3560. bx::memCopy(&clearValue.float32, _palette[index], sizeof(clearValue.float32) );
  3561. break;
  3562. }
  3563. }
  3564. else
  3565. {
  3566. switch (type)
  3567. {
  3568. case bx::EncodingType::Int:
  3569. case bx::EncodingType::Uint:
  3570. clearValue.uint32[0] = _clear.m_index[0];
  3571. clearValue.uint32[1] = _clear.m_index[1];
  3572. clearValue.uint32[2] = _clear.m_index[2];
  3573. clearValue.uint32[3] = _clear.m_index[3];
  3574. break;
  3575. default:
  3576. bx::unpackRgba8(clearValue.float32, _clear.m_index);
  3577. break;
  3578. }
  3579. }
  3580. ++mrt;
  3581. }
  3582. }
  3583. depthAspectMask &= 0
  3584. | (_clear.m_flags & BGFX_CLEAR_DEPTH ? VK_IMAGE_ASPECT_DEPTH_BIT : 0)
  3585. | (_clear.m_flags & BGFX_CLEAR_STENCIL ? VK_IMAGE_ASPECT_STENCIL_BIT : 0)
  3586. ;
  3587. if (0 != depthAspectMask)
  3588. {
  3589. attachments[mrt].aspectMask = depthAspectMask;
  3590. attachments[mrt].clearValue.depthStencil.stencil = _clear.m_stencil;
  3591. attachments[mrt].clearValue.depthStencil.depth = _clear.m_depth;
  3592. ++mrt;
  3593. }
  3594. if (mrt > 0)
  3595. {
  3596. vkCmdClearAttachments(m_commandBuffer, mrt, attachments, BX_COUNTOF(rect), rect);
  3597. }
  3598. }
  3599. void kick(bool _finishAll = false)
  3600. {
  3601. m_cmd.kick(_finishAll);
  3602. VK_CHECK(m_cmd.alloc(&m_commandBuffer) );
  3603. m_cmd.finish(_finishAll);
  3604. }
  3605. int32_t selectMemoryType(uint32_t _memoryTypeBits, uint32_t _propertyFlags, int32_t _startIndex = 0) const
  3606. {
  3607. for (int32_t ii = _startIndex, num = m_memoryProperties.memoryTypeCount; ii < num; ++ii)
  3608. {
  3609. const VkMemoryType& memType = m_memoryProperties.memoryTypes[ii];
  3610. if ( (0 != ( (1<<ii) & _memoryTypeBits) )
  3611. && ( (memType.propertyFlags & _propertyFlags) == _propertyFlags) )
  3612. {
  3613. return ii;
  3614. }
  3615. }
  3616. BX_TRACE("Failed to find memory that supports flags 0x%08x.", _propertyFlags);
  3617. return -1;
  3618. }
  3619. VkResult allocateMemory(const VkMemoryRequirements* requirements, VkMemoryPropertyFlags propertyFlags, ::VkDeviceMemory* memory) const
  3620. {
  3621. VkMemoryAllocateInfo ma;
  3622. ma.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO;
  3623. ma.pNext = NULL;
  3624. ma.allocationSize = requirements->size;
  3625. VkResult result = VK_ERROR_UNKNOWN;
  3626. int32_t searchIndex = -1;
  3627. do
  3628. {
  3629. searchIndex++;
  3630. searchIndex = selectMemoryType(requirements->memoryTypeBits, propertyFlags, searchIndex);
  3631. if (searchIndex >= 0)
  3632. {
  3633. ma.memoryTypeIndex = searchIndex;
  3634. result = vkAllocateMemory(m_device, &ma, m_allocatorCb, memory);
  3635. }
  3636. }
  3637. while (result != VK_SUCCESS
  3638. && searchIndex >= 0);
  3639. return result;
  3640. }
  3641. VkResult createHostBuffer(uint32_t _size, VkMemoryPropertyFlags _flags, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory, const void* _data = NULL)
  3642. {
  3643. VkResult result = VK_SUCCESS;
  3644. VkBufferCreateInfo bci;
  3645. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3646. bci.pNext = NULL;
  3647. bci.flags = 0;
  3648. bci.size = _size;
  3649. bci.queueFamilyIndexCount = 0;
  3650. bci.pQueueFamilyIndices = NULL;
  3651. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3652. bci.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
  3653. result = vkCreateBuffer(m_device, &bci, m_allocatorCb, _buffer);
  3654. if (VK_SUCCESS != result)
  3655. {
  3656. BX_TRACE("Create host buffer error: vkCreateBuffer failed %d: %s.", result, getName(result) );
  3657. return result;
  3658. }
  3659. VkMemoryRequirements mr;
  3660. vkGetBufferMemoryRequirements(m_device, *_buffer, &mr);
  3661. result = allocateMemory(&mr, _flags, _memory);
  3662. if (VK_SUCCESS != result
  3663. && (_flags & VK_MEMORY_PROPERTY_HOST_CACHED_BIT) )
  3664. {
  3665. result = allocateMemory(&mr, _flags & ~VK_MEMORY_PROPERTY_HOST_CACHED_BIT, _memory);
  3666. }
  3667. if (VK_SUCCESS != result)
  3668. {
  3669. BX_TRACE("Create host buffer error: vkAllocateMemory failed %d: %s.", result, getName(result) );
  3670. return result;
  3671. }
  3672. result = vkBindBufferMemory(m_device, *_buffer, *_memory, 0);
  3673. if (VK_SUCCESS != result)
  3674. {
  3675. BX_TRACE("Create host buffer error: vkBindBufferMemory failed %d: %s.", result, getName(result) );
  3676. return result;
  3677. }
  3678. if (_data != NULL)
  3679. {
  3680. void* dst;
  3681. result = vkMapMemory(m_device, *_memory, 0, _size, 0, &dst);
  3682. if (VK_SUCCESS != result)
  3683. {
  3684. BX_TRACE("Create host buffer error: vkMapMemory failed %d: %s.", result, getName(result) );
  3685. return result;
  3686. }
  3687. bx::memCopy(dst, _data, _size);
  3688. vkUnmapMemory(m_device, *_memory);
  3689. }
  3690. return result;
  3691. }
  3692. VkResult createStagingBuffer(uint32_t _size, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory, const void* _data = NULL)
  3693. {
  3694. const VkMemoryPropertyFlags flags = 0
  3695. | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT
  3696. | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT
  3697. ;
  3698. return createHostBuffer(_size, flags, _buffer, _memory, _data);
  3699. }
  3700. VkResult createReadbackBuffer(uint32_t _size, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory)
  3701. {
  3702. const VkMemoryPropertyFlags flags = 0
  3703. | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT
  3704. | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT
  3705. | VK_MEMORY_PROPERTY_HOST_CACHED_BIT
  3706. ;
  3707. return createHostBuffer(_size, flags, _buffer, _memory, NULL);
  3708. }
  3709. VkAllocationCallbacks* m_allocatorCb;
  3710. VkDebugReportCallbackEXT m_debugReportCallback;
  3711. VkInstance m_instance;
  3712. VkPhysicalDevice m_physicalDevice;
  3713. uint32_t m_instanceApiVersion;
  3714. VkPhysicalDeviceProperties m_deviceProperties;
  3715. VkPhysicalDeviceMemoryProperties m_memoryProperties;
  3716. VkPhysicalDeviceFeatures m_deviceFeatures;
  3717. bool m_lineAASupport;
  3718. bool m_borderColorSupport;
  3719. bool m_timerQuerySupport;
  3720. FrameBufferVK m_backBuffer;
  3721. TextureFormat::Enum m_swapchainFormats[TextureFormat::Count];
  3722. uint16_t m_numWindows;
  3723. FrameBufferHandle m_windows[BGFX_CONFIG_MAX_FRAME_BUFFERS];
  3724. int64_t m_presentElapsed;
  3725. ScratchBufferVK m_scratchBuffer[BGFX_CONFIG_MAX_FRAME_LATENCY];
  3726. uint32_t m_numFramesInFlight;
  3727. CommandQueueVK m_cmd;
  3728. VkCommandBuffer m_commandBuffer;
  3729. VkDevice m_device;
  3730. uint32_t m_globalQueueFamily;
  3731. VkQueue m_globalQueue;
  3732. VkDescriptorPool m_descriptorPool;
  3733. VkPipelineCache m_pipelineCache;
  3734. TimerQueryVK m_gpuTimer;
  3735. OcclusionQueryVK m_occlusionQuery;
  3736. void* m_renderDocDll;
  3737. void* m_vulkan1Dll;
  3738. IndexBufferVK m_indexBuffers[BGFX_CONFIG_MAX_INDEX_BUFFERS];
  3739. VertexBufferVK m_vertexBuffers[BGFX_CONFIG_MAX_VERTEX_BUFFERS];
  3740. ShaderVK m_shaders[BGFX_CONFIG_MAX_SHADERS];
  3741. ProgramVK m_program[BGFX_CONFIG_MAX_PROGRAMS];
  3742. TextureVK m_textures[BGFX_CONFIG_MAX_TEXTURES];
  3743. VertexLayout m_vertexLayouts[BGFX_CONFIG_MAX_VERTEX_LAYOUTS];
  3744. FrameBufferVK m_frameBuffers[BGFX_CONFIG_MAX_FRAME_BUFFERS];
  3745. void* m_uniforms[BGFX_CONFIG_MAX_UNIFORMS];
  3746. Matrix4 m_predefinedUniforms[PredefinedUniform::Count];
  3747. UniformRegistry m_uniformReg;
  3748. StateCacheT<VkPipeline> m_pipelineStateCache;
  3749. StateCacheT<VkDescriptorSetLayout> m_descriptorSetLayoutCache;
  3750. StateCacheT<VkRenderPass> m_renderPassCache;
  3751. StateCacheT<VkSampler> m_samplerCache;
  3752. StateCacheT<uint32_t> m_samplerBorderColorCache;
  3753. StateCacheLru<VkImageView, 1024> m_imageViewCache;
  3754. Resolution m_resolution;
  3755. float m_maxAnisotropy;
  3756. bool m_depthClamp;
  3757. bool m_wireframe;
  3758. VkBuffer m_captureBuffer;
  3759. VkDeviceMemory m_captureMemory;
  3760. uint32_t m_captureSize;
  3761. TextVideoMem m_textVideoMem;
  3762. uint8_t m_fsScratch[64<<10];
  3763. uint8_t m_vsScratch[64<<10];
  3764. FrameBufferHandle m_fbh;
  3765. };
  3766. static RendererContextVK* s_renderVK;
  3767. RendererContextI* rendererCreate(const Init& _init)
  3768. {
  3769. s_renderVK = BX_NEW(g_allocator, RendererContextVK);
  3770. if (!s_renderVK->init(_init) )
  3771. {
  3772. bx::deleteObject(g_allocator, s_renderVK);
  3773. s_renderVK = NULL;
  3774. }
  3775. return s_renderVK;
  3776. }
  3777. void rendererDestroy()
  3778. {
  3779. s_renderVK->shutdown();
  3780. bx::deleteObject(g_allocator, s_renderVK);
  3781. s_renderVK = NULL;
  3782. }
  3783. #define VK_DESTROY_FUNC(_name) \
  3784. void vkDestroy(Vk##_name& _obj) \
  3785. { \
  3786. if (VK_NULL_HANDLE != _obj) \
  3787. { \
  3788. vkDestroy##_name(s_renderVK->m_device, _obj.vk, s_renderVK->m_allocatorCb); \
  3789. _obj = VK_NULL_HANDLE; \
  3790. } \
  3791. } \
  3792. void release(Vk##_name& _obj) \
  3793. { \
  3794. s_renderVK->release(_obj); \
  3795. }
  3796. VK_DESTROY
  3797. #undef VK_DESTROY_FUNC
  3798. void vkDestroy(VkDeviceMemory& _obj)
  3799. {
  3800. if (VK_NULL_HANDLE != _obj)
  3801. {
  3802. vkFreeMemory(s_renderVK->m_device, _obj.vk, s_renderVK->m_allocatorCb);
  3803. _obj = VK_NULL_HANDLE;
  3804. }
  3805. }
  3806. void vkDestroy(VkSurfaceKHR& _obj)
  3807. {
  3808. if (VK_NULL_HANDLE != _obj)
  3809. {
  3810. vkDestroySurfaceKHR(s_renderVK->m_instance, _obj.vk, s_renderVK->m_allocatorCb);
  3811. _obj = VK_NULL_HANDLE;
  3812. }
  3813. }
  3814. void vkDestroy(VkDescriptorSet& _obj)
  3815. {
  3816. if (VK_NULL_HANDLE != _obj)
  3817. {
  3818. vkFreeDescriptorSets(s_renderVK->m_device, s_renderVK->m_descriptorPool, 1, &_obj);
  3819. _obj = VK_NULL_HANDLE;
  3820. }
  3821. }
  3822. void release(VkDeviceMemory& _obj)
  3823. {
  3824. s_renderVK->release(_obj);
  3825. }
  3826. void release(VkSurfaceKHR& _obj)
  3827. {
  3828. s_renderVK->release(_obj);
  3829. }
  3830. void release(VkDescriptorSet& _obj)
  3831. {
  3832. s_renderVK->release(_obj);
  3833. }
  3834. void ScratchBufferVK::create(uint32_t _size, uint32_t _count)
  3835. {
  3836. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  3837. const VkDevice device = s_renderVK->m_device;
  3838. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3839. const uint32_t align = uint32_t(deviceLimits.minUniformBufferOffsetAlignment);
  3840. const uint32_t entrySize = bx::strideAlign(_size, align);
  3841. const uint32_t totalSize = entrySize * _count;
  3842. VkBufferCreateInfo bci;
  3843. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3844. bci.pNext = NULL;
  3845. bci.flags = 0;
  3846. bci.size = totalSize;
  3847. bci.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT;
  3848. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3849. bci.queueFamilyIndexCount = 0;
  3850. bci.pQueueFamilyIndices = NULL;
  3851. VK_CHECK(vkCreateBuffer(
  3852. device
  3853. , &bci
  3854. , allocatorCb
  3855. , &m_buffer
  3856. ) );
  3857. VkMemoryRequirements mr;
  3858. vkGetBufferMemoryRequirements(
  3859. device
  3860. , m_buffer
  3861. , &mr
  3862. );
  3863. VkMemoryPropertyFlags flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
  3864. VkResult result = s_renderVK->allocateMemory(&mr, flags, &m_deviceMem);
  3865. if (VK_SUCCESS != result)
  3866. {
  3867. flags &= ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
  3868. VK_CHECK(s_renderVK->allocateMemory(&mr, flags, &m_deviceMem) );
  3869. }
  3870. m_size = (uint32_t)mr.size;
  3871. m_pos = 0;
  3872. VK_CHECK(vkBindBufferMemory(device, m_buffer, m_deviceMem, 0) );
  3873. VK_CHECK(vkMapMemory(device, m_deviceMem, 0, m_size, 0, (void**)&m_data) );
  3874. }
  3875. void ScratchBufferVK::destroy()
  3876. {
  3877. reset();
  3878. vkUnmapMemory(s_renderVK->m_device, m_deviceMem);
  3879. s_renderVK->release(m_buffer);
  3880. s_renderVK->release(m_deviceMem);
  3881. }
  3882. void ScratchBufferVK::reset()
  3883. {
  3884. m_pos = 0;
  3885. }
  3886. uint32_t ScratchBufferVK::write(const void* _data, uint32_t _size)
  3887. {
  3888. BX_ASSERT(m_pos < m_size, "Out of scratch buffer memory");
  3889. const uint32_t offset = m_pos;
  3890. if (_size > 0)
  3891. {
  3892. bx::memCopy(&m_data[m_pos], _data, _size);
  3893. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3894. const uint32_t align = uint32_t(deviceLimits.minUniformBufferOffsetAlignment);
  3895. const uint32_t alignedSize = bx::strideAlign(_size, align);
  3896. m_pos += alignedSize;
  3897. }
  3898. return offset;
  3899. }
  3900. void ScratchBufferVK::flush()
  3901. {
  3902. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3903. VkDevice device = s_renderVK->m_device;
  3904. const uint32_t align = uint32_t(deviceLimits.nonCoherentAtomSize);
  3905. const uint32_t size = bx::min(bx::strideAlign(m_pos, align), m_size);
  3906. VkMappedMemoryRange range;
  3907. range.sType = VK_STRUCTURE_TYPE_MAPPED_MEMORY_RANGE;
  3908. range.pNext = NULL;
  3909. range.memory = m_deviceMem;
  3910. range.offset = 0;
  3911. range.size = size;
  3912. VK_CHECK(vkFlushMappedMemoryRanges(device, 1, &range) );
  3913. }
  3914. void BufferVK::create(VkCommandBuffer _commandBuffer, uint32_t _size, void* _data, uint16_t _flags, bool _vertex, uint32_t _stride)
  3915. {
  3916. BX_UNUSED(_stride);
  3917. m_size = _size;
  3918. m_flags = _flags;
  3919. m_dynamic = NULL == _data;
  3920. const bool storage = m_flags & BGFX_BUFFER_COMPUTE_READ_WRITE;
  3921. const bool indirect = m_flags & BGFX_BUFFER_DRAW_INDIRECT;
  3922. VkBufferCreateInfo bci;
  3923. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3924. bci.pNext = NULL;
  3925. bci.flags = 0;
  3926. bci.size = _size;
  3927. bci.usage = 0
  3928. | (_vertex ? VK_BUFFER_USAGE_VERTEX_BUFFER_BIT : VK_BUFFER_USAGE_INDEX_BUFFER_BIT)
  3929. | (storage || indirect ? VK_BUFFER_USAGE_STORAGE_BUFFER_BIT : 0)
  3930. | (indirect ? VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT : 0)
  3931. | VK_BUFFER_USAGE_TRANSFER_DST_BIT
  3932. ;
  3933. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3934. bci.queueFamilyIndexCount = 0;
  3935. bci.pQueueFamilyIndices = NULL;
  3936. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  3937. const VkDevice device = s_renderVK->m_device;
  3938. VK_CHECK(vkCreateBuffer(device, &bci, allocatorCb, &m_buffer) );
  3939. VkMemoryRequirements mr;
  3940. vkGetBufferMemoryRequirements(device, m_buffer, &mr);
  3941. VK_CHECK(s_renderVK->allocateMemory(&mr, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_deviceMem) );
  3942. VK_CHECK(vkBindBufferMemory(device, m_buffer, m_deviceMem, 0) );
  3943. if (!m_dynamic)
  3944. {
  3945. update(_commandBuffer, 0, _size, _data);
  3946. }
  3947. }
  3948. void BufferVK::update(VkCommandBuffer _commandBuffer, uint32_t _offset, uint32_t _size, void* _data, bool _discard)
  3949. {
  3950. BGFX_PROFILER_SCOPE("BufferVK::update", kColorFrame);
  3951. BX_UNUSED(_discard);
  3952. VkBuffer stagingBuffer;
  3953. VkDeviceMemory stagingMem;
  3954. VK_CHECK(s_renderVK->createStagingBuffer(_size, &stagingBuffer, &stagingMem, _data) );
  3955. VkBufferCopy region;
  3956. region.srcOffset = 0;
  3957. region.dstOffset = _offset;
  3958. region.size = _size;
  3959. vkCmdCopyBuffer(_commandBuffer, stagingBuffer, m_buffer, 1, &region);
  3960. setMemoryBarrier(
  3961. _commandBuffer
  3962. , VK_PIPELINE_STAGE_TRANSFER_BIT
  3963. , VK_PIPELINE_STAGE_TRANSFER_BIT
  3964. );
  3965. s_renderVK->release(stagingBuffer);
  3966. s_renderVK->release(stagingMem);
  3967. }
  3968. void BufferVK::destroy()
  3969. {
  3970. if (VK_NULL_HANDLE != m_buffer)
  3971. {
  3972. s_renderVK->release(m_buffer);
  3973. s_renderVK->release(m_deviceMem);
  3974. m_dynamic = false;
  3975. }
  3976. }
  3977. void VertexBufferVK::create(VkCommandBuffer _commandBuffer, uint32_t _size, void* _data, VertexLayoutHandle _layoutHandle, uint16_t _flags)
  3978. {
  3979. BufferVK::create(_commandBuffer, _size, _data, _flags, true);
  3980. m_layoutHandle = _layoutHandle;
  3981. }
  3982. void ShaderVK::create(const Memory* _mem)
  3983. {
  3984. bx::MemoryReader reader(_mem->data, _mem->size);
  3985. bx::ErrorAssert err;
  3986. uint32_t magic;
  3987. bx::read(&reader, magic, &err);
  3988. VkShaderStageFlagBits shaderStage = VK_SHADER_STAGE_ALL;
  3989. if (isShaderType(magic, 'C') )
  3990. {
  3991. shaderStage = VK_SHADER_STAGE_COMPUTE_BIT;
  3992. }
  3993. else if (isShaderType(magic, 'F') )
  3994. {
  3995. shaderStage = VK_SHADER_STAGE_FRAGMENT_BIT;
  3996. }
  3997. else if (isShaderType(magic, 'V') )
  3998. {
  3999. shaderStage = VK_SHADER_STAGE_VERTEX_BIT;
  4000. }
  4001. const bool fragment = isShaderType(magic, 'F');
  4002. uint32_t hashIn;
  4003. bx::read(&reader, hashIn, &err);
  4004. uint32_t hashOut;
  4005. if (isShaderVerLess(magic, 6) )
  4006. {
  4007. hashOut = hashIn;
  4008. }
  4009. else
  4010. {
  4011. bx::read(&reader, hashOut, &err);
  4012. }
  4013. uint16_t count;
  4014. bx::read(&reader, count, &err);
  4015. m_numPredefined = 0;
  4016. m_numUniforms = count;
  4017. m_numTextures = 0;
  4018. m_oldBindingModel = isShaderVerLess(magic, 11);
  4019. BX_TRACE("%s Shader consts %d"
  4020. , getShaderTypeName(magic)
  4021. , count
  4022. );
  4023. uint8_t fragmentBit = fragment ? kUniformFragmentBit : 0;
  4024. for (uint32_t ii = 0; ii < BGFX_CONFIG_MAX_TEXTURE_SAMPLERS; ++ii)
  4025. {
  4026. m_bindInfo[ii].uniformHandle = BGFX_INVALID_HANDLE;
  4027. m_bindInfo[ii].type = BindType::Count;
  4028. m_bindInfo[ii].binding = 0;
  4029. m_bindInfo[ii].samplerBinding = 0;
  4030. m_bindInfo[ii].index = UINT32_MAX;
  4031. }
  4032. if (0 < count)
  4033. {
  4034. for (uint32_t ii = 0; ii < count; ++ii)
  4035. {
  4036. uint8_t nameSize = 0;
  4037. bx::read(&reader, nameSize, &err);
  4038. char name[256];
  4039. bx::read(&reader, &name, nameSize, &err);
  4040. name[nameSize] = '\0';
  4041. uint8_t type = 0;
  4042. bx::read(&reader, type, &err);
  4043. uint8_t num;
  4044. bx::read(&reader, num, &err);
  4045. uint16_t regIndex;
  4046. bx::read(&reader, regIndex, &err);
  4047. uint16_t regCount;
  4048. bx::read(&reader, regCount, &err);
  4049. const bool hasTexData = !isShaderVerLess(magic, 8);
  4050. const bool hasTexFormat = !isShaderVerLess(magic, 10);
  4051. uint8_t texComponent = 0;
  4052. uint8_t texDimension = 0;
  4053. uint16_t texFormat = 0;
  4054. if (hasTexData)
  4055. {
  4056. bx::read(&reader, texComponent, &err);
  4057. bx::read(&reader, texDimension, &err);
  4058. }
  4059. if (hasTexFormat)
  4060. {
  4061. bx::read(&reader, texFormat, &err);
  4062. }
  4063. const char* kind = "invalid";
  4064. BX_UNUSED(num, texComponent, texFormat);
  4065. auto textureDimensionToViewType = [](TextureDimension::Enum dimension)
  4066. {
  4067. switch (dimension)
  4068. {
  4069. case TextureDimension::Dimension1D: return VK_IMAGE_VIEW_TYPE_1D;
  4070. case TextureDimension::Dimension2D: return VK_IMAGE_VIEW_TYPE_2D;
  4071. case TextureDimension::Dimension2DArray: return VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  4072. case TextureDimension::DimensionCube: return VK_IMAGE_VIEW_TYPE_CUBE;
  4073. case TextureDimension::DimensionCubeArray: return VK_IMAGE_VIEW_TYPE_CUBE_ARRAY;
  4074. case TextureDimension::Dimension3D: return VK_IMAGE_VIEW_TYPE_3D;
  4075. default: return VK_IMAGE_VIEW_TYPE_MAX_ENUM;
  4076. }
  4077. };
  4078. if (UINT16_MAX != regIndex)
  4079. {
  4080. PredefinedUniform::Enum predefined = nameToPredefinedUniformEnum(name);
  4081. if (PredefinedUniform::Count != predefined)
  4082. {
  4083. kind = "predefined";
  4084. m_predefined[m_numPredefined].m_loc = regIndex;
  4085. m_predefined[m_numPredefined].m_count = regCount;
  4086. m_predefined[m_numPredefined].m_type = uint8_t(predefined|fragmentBit);
  4087. m_numPredefined++;
  4088. }
  4089. else if (UniformType::End == (~kUniformMask & type) )
  4090. {
  4091. // regCount is used for descriptor type
  4092. const bool isBuffer = idToDescriptorType(regCount) == DescriptorType::StorageBuffer;
  4093. if (0 == regIndex)
  4094. {
  4095. continue;
  4096. }
  4097. const uint8_t reverseShift = m_oldBindingModel
  4098. ? (fragment ? kSpirvOldFragmentShift : 0) + (isBuffer ? kSpirvOldBufferShift : kSpirvOldImageShift)
  4099. : kSpirvBindShift;
  4100. const uint16_t stage = regIndex - reverseShift; // regIndex is used for buffer binding index
  4101. m_bindInfo[stage].type = isBuffer ? BindType::Buffer : BindType::Image;
  4102. m_bindInfo[stage].uniformHandle = { 0 };
  4103. m_bindInfo[stage].binding = regIndex;
  4104. if (!isBuffer)
  4105. {
  4106. const VkImageViewType viewType = hasTexData
  4107. ? textureDimensionToViewType(idToTextureDimension(texDimension) )
  4108. : VK_IMAGE_VIEW_TYPE_MAX_ENUM
  4109. ;
  4110. if (VK_IMAGE_VIEW_TYPE_MAX_ENUM != viewType)
  4111. {
  4112. m_bindInfo[stage].index = m_numTextures;
  4113. m_textures[m_numTextures].type = viewType;
  4114. m_numTextures++;
  4115. }
  4116. }
  4117. kind = "storage";
  4118. }
  4119. else if (UniformType::Sampler == (~kUniformMask & type) )
  4120. {
  4121. const uint8_t reverseShift = m_oldBindingModel
  4122. ? (fragment ? kSpirvOldFragmentShift : 0) + kSpirvOldTextureShift
  4123. : kSpirvBindShift;
  4124. const uint16_t stage = regIndex - reverseShift; // regIndex is used for image/sampler binding index
  4125. const UniformRegInfo* info = s_renderVK->m_uniformReg.find(name);
  4126. BX_ASSERT(NULL != info, "User defined uniform '%s' is not found, it won't be set.", name);
  4127. m_bindInfo[stage].uniformHandle = info->m_handle;
  4128. m_bindInfo[stage].type = BindType::Sampler;
  4129. m_bindInfo[stage].binding = regIndex;
  4130. m_bindInfo[stage].samplerBinding = regIndex + kSpirvSamplerShift;
  4131. const VkImageViewType viewType = hasTexData
  4132. ? textureDimensionToViewType(idToTextureDimension(texDimension) )
  4133. : VK_IMAGE_VIEW_TYPE_MAX_ENUM
  4134. ;
  4135. if (VK_IMAGE_VIEW_TYPE_MAX_ENUM != viewType)
  4136. {
  4137. m_bindInfo[stage].index = m_numTextures;
  4138. m_textures[m_numTextures].type = viewType;
  4139. m_numTextures++;
  4140. }
  4141. kind = "sampler";
  4142. }
  4143. else
  4144. {
  4145. const UniformRegInfo* info = s_renderVK->m_uniformReg.find(name);
  4146. BX_ASSERT(NULL != info, "User defined uniform '%s' is not found, it won't be set.", name);
  4147. if (NULL != info)
  4148. {
  4149. if (NULL == m_constantBuffer)
  4150. {
  4151. m_constantBuffer = UniformBuffer::create(1024);
  4152. }
  4153. kind = "user";
  4154. m_constantBuffer->writeUniformHandle( (UniformType::Enum)(type|fragmentBit), regIndex, info->m_handle, regCount);
  4155. }
  4156. }
  4157. }
  4158. BX_TRACE("\t%s: %s (%s), r.index %3d, r.count %2d, r.texComponent %1d, r.texDimension %1d"
  4159. , kind
  4160. , name
  4161. , getUniformTypeName(UniformType::Enum(type&~kUniformMask) )
  4162. , regIndex
  4163. , regCount
  4164. , texComponent
  4165. , texDimension
  4166. );
  4167. BX_UNUSED(kind);
  4168. }
  4169. if (NULL != m_constantBuffer)
  4170. {
  4171. m_constantBuffer->finish();
  4172. }
  4173. }
  4174. uint32_t shaderSize;
  4175. bx::read(&reader, shaderSize, &err);
  4176. const void* code = reader.getDataPtr();
  4177. bx::skip(&reader, shaderSize+1);
  4178. m_code = alloc(shaderSize);
  4179. bx::memCopy(m_code->data, code, shaderSize);
  4180. VkShaderModuleCreateInfo smci;
  4181. smci.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO;
  4182. smci.pNext = NULL;
  4183. smci.flags = 0;
  4184. smci.codeSize = m_code->size;
  4185. smci.pCode = (const uint32_t*)m_code->data;
  4186. // disassemble(bx::getDebugOut(), m_code->data, m_code->size);
  4187. VK_CHECK(vkCreateShaderModule(
  4188. s_renderVK->m_device
  4189. , &smci
  4190. , s_renderVK->m_allocatorCb
  4191. , &m_module
  4192. ) );
  4193. bx::memSet(m_attrMask, 0, sizeof(m_attrMask) );
  4194. bx::memSet(m_attrRemap, 0, sizeof(m_attrRemap) );
  4195. bx::read(&reader, m_numAttrs, &err);
  4196. for (uint8_t ii = 0; ii < m_numAttrs; ++ii)
  4197. {
  4198. uint16_t id;
  4199. bx::read(&reader, id, &err);
  4200. Attrib::Enum attr = idToAttrib(id);
  4201. if (Attrib::Count != attr)
  4202. {
  4203. m_attrMask[attr] = UINT16_MAX;
  4204. m_attrRemap[attr] = ii;
  4205. }
  4206. }
  4207. bx::HashMurmur2A murmur;
  4208. murmur.begin();
  4209. murmur.add(hashIn);
  4210. murmur.add(hashOut);
  4211. murmur.add(m_code->data, m_code->size);
  4212. murmur.add(m_numAttrs);
  4213. murmur.add(m_attrMask, m_numAttrs);
  4214. murmur.add(m_attrRemap, m_numAttrs);
  4215. m_hash = murmur.end();
  4216. bx::read(&reader, m_size, &err);
  4217. // fill binding description with uniform information
  4218. uint16_t bidx = 0;
  4219. if (m_size > 0)
  4220. {
  4221. m_uniformBinding = fragment ? (m_oldBindingModel ? kSpirvOldFragmentBinding : kSpirvFragmentBinding) : 0;
  4222. VkDescriptorSetLayoutBinding& binding = m_bindings[bidx];
  4223. binding.stageFlags = shaderStage;
  4224. binding.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  4225. binding.binding = m_uniformBinding;
  4226. binding.pImmutableSamplers = NULL;
  4227. binding.descriptorCount = 1;
  4228. bidx++;
  4229. }
  4230. for (uint32_t ii = 0; ii < BX_COUNTOF(m_bindInfo); ++ii)
  4231. {
  4232. switch (m_bindInfo[ii].type)
  4233. {
  4234. case BindType::Buffer:
  4235. case BindType::Image:
  4236. {
  4237. VkDescriptorSetLayoutBinding& binding = m_bindings[bidx];
  4238. binding.stageFlags = shaderStage;
  4239. binding.descriptorType = BindType::Buffer == m_bindInfo[ii].type
  4240. ? VK_DESCRIPTOR_TYPE_STORAGE_BUFFER
  4241. : VK_DESCRIPTOR_TYPE_STORAGE_IMAGE
  4242. ;
  4243. binding.binding = m_bindInfo[ii].binding;
  4244. binding.pImmutableSamplers = NULL;
  4245. binding.descriptorCount = 1;
  4246. bidx++;
  4247. }
  4248. break;
  4249. case BindType::Sampler:
  4250. {
  4251. VkDescriptorSetLayoutBinding& textureBinding = m_bindings[bidx];
  4252. textureBinding.stageFlags = shaderStage;
  4253. textureBinding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
  4254. textureBinding.binding = m_bindInfo[ii].binding;
  4255. textureBinding.pImmutableSamplers = NULL;
  4256. textureBinding.descriptorCount = 1;
  4257. bidx++;
  4258. VkDescriptorSetLayoutBinding& samplerBinding = m_bindings[bidx];
  4259. samplerBinding.stageFlags = shaderStage;
  4260. samplerBinding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
  4261. samplerBinding.binding = m_bindInfo[ii].samplerBinding;
  4262. samplerBinding.pImmutableSamplers = NULL;
  4263. samplerBinding.descriptorCount = 1;
  4264. bidx++;
  4265. }
  4266. break;
  4267. default:
  4268. break;
  4269. }
  4270. }
  4271. m_numBindings = bidx;
  4272. }
  4273. void ShaderVK::destroy()
  4274. {
  4275. if (NULL != m_constantBuffer)
  4276. {
  4277. UniformBuffer::destroy(m_constantBuffer);
  4278. m_constantBuffer = NULL;
  4279. }
  4280. m_numPredefined = 0;
  4281. if (NULL != m_code)
  4282. {
  4283. release(m_code);
  4284. m_code = NULL;
  4285. m_hash = 0;
  4286. }
  4287. if (VK_NULL_HANDLE != m_module)
  4288. {
  4289. vkDestroy(m_module);
  4290. }
  4291. }
  4292. void ProgramVK::create(const ShaderVK* _vsh, const ShaderVK* _fsh)
  4293. {
  4294. BX_ASSERT(NULL != _vsh->m_code, "Vertex shader doesn't exist.");
  4295. m_vsh = _vsh;
  4296. bx::memCopy(
  4297. &m_predefined[0]
  4298. , _vsh->m_predefined
  4299. , _vsh->m_numPredefined * sizeof(PredefinedUniform)
  4300. );
  4301. m_numPredefined = _vsh->m_numPredefined;
  4302. if (NULL != _fsh)
  4303. {
  4304. BX_ASSERT(NULL != _fsh->m_code, "Fragment shader doesn't exist.");
  4305. m_fsh = _fsh;
  4306. bx::memCopy(
  4307. &m_predefined[m_numPredefined]
  4308. , _fsh->m_predefined
  4309. , _fsh->m_numPredefined * sizeof(PredefinedUniform)
  4310. );
  4311. m_numPredefined += _fsh->m_numPredefined;
  4312. }
  4313. m_numTextures = 0;
  4314. for (uint8_t stage = 0; stage < BX_COUNTOF(m_bindInfo); ++stage)
  4315. {
  4316. const ShaderVK* shader = NULL;
  4317. if (isValid(m_vsh->m_bindInfo[stage].uniformHandle) )
  4318. {
  4319. shader = _vsh;
  4320. BX_ASSERT(false
  4321. || NULL == m_fsh
  4322. || !isValid(m_fsh->m_bindInfo[stage].uniformHandle)
  4323. || !(m_vsh->m_oldBindingModel || m_fsh->m_oldBindingModel)
  4324. , "Shared vertex/fragment bindings require shader binary version >= 11."
  4325. );
  4326. }
  4327. else if (NULL != m_fsh
  4328. && isValid(m_fsh->m_bindInfo[stage].uniformHandle) )
  4329. {
  4330. shader = _fsh;
  4331. }
  4332. if (NULL != shader)
  4333. {
  4334. m_bindInfo[stage] = shader->m_bindInfo[stage];
  4335. uint32_t& index = m_bindInfo[stage].index;
  4336. if (UINT32_MAX != index)
  4337. {
  4338. m_textures[m_numTextures] = shader->m_textures[index];
  4339. index = m_numTextures;
  4340. m_numTextures++;
  4341. }
  4342. }
  4343. }
  4344. // create exact pipeline layout
  4345. m_descriptorSetLayout = VK_NULL_HANDLE;
  4346. uint32_t numBindings = m_vsh->m_numBindings + (m_fsh ? m_fsh->m_numBindings : 0);
  4347. if (0 < numBindings)
  4348. {
  4349. // generate descriptor set layout hash
  4350. bx::HashMurmur2A murmur;
  4351. murmur.begin();
  4352. murmur.add(m_vsh->m_bindings, sizeof(VkDescriptorSetLayoutBinding) * m_vsh->m_numBindings);
  4353. if (NULL != m_fsh)
  4354. {
  4355. murmur.add(m_fsh->m_bindings, sizeof(VkDescriptorSetLayoutBinding) * m_fsh->m_numBindings);
  4356. }
  4357. uint32_t descriptorSetLayoutHash = murmur.end();
  4358. m_descriptorSetLayout = s_renderVK->m_descriptorSetLayoutCache.find(descriptorSetLayoutHash);
  4359. if (VK_NULL_HANDLE == m_descriptorSetLayout)
  4360. {
  4361. VkDescriptorSetLayoutBinding bindings[2 * BX_COUNTOF(ShaderVK::m_bindings)];
  4362. bx::memCopy(
  4363. bindings
  4364. , m_vsh->m_bindings
  4365. , sizeof(VkDescriptorSetLayoutBinding) * m_vsh->m_numBindings
  4366. );
  4367. numBindings = m_vsh->m_numBindings;
  4368. if (NULL != m_fsh)
  4369. {
  4370. for (uint16_t ii = 0; ii < m_fsh->m_numBindings; ii++)
  4371. {
  4372. const VkDescriptorSetLayoutBinding& fsBinding = m_fsh->m_bindings[ii];
  4373. uint16_t vsBindingIdx = UINT16_MAX;
  4374. for (uint16_t jj = 0; jj < m_vsh->m_numBindings; jj++)
  4375. {
  4376. if (fsBinding.binding == bindings[jj].binding)
  4377. {
  4378. vsBindingIdx = jj;
  4379. break;
  4380. }
  4381. }
  4382. if (UINT16_MAX != vsBindingIdx)
  4383. {
  4384. BX_ASSERT(
  4385. bindings[vsBindingIdx].descriptorType == fsBinding.descriptorType
  4386. , "Mismatching descriptor types. Shaders compiled with different versions of shaderc?"
  4387. );
  4388. bindings[vsBindingIdx].stageFlags |= fsBinding.stageFlags;
  4389. }
  4390. else
  4391. {
  4392. bindings[numBindings] = fsBinding;
  4393. numBindings++;
  4394. }
  4395. }
  4396. }
  4397. VkDescriptorSetLayoutCreateInfo dslci;
  4398. dslci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO;
  4399. dslci.pNext = NULL;
  4400. dslci.flags = 0;
  4401. dslci.bindingCount = numBindings;
  4402. dslci.pBindings = bindings;
  4403. VK_CHECK(vkCreateDescriptorSetLayout(
  4404. s_renderVK->m_device
  4405. , &dslci
  4406. , s_renderVK->m_allocatorCb
  4407. , &m_descriptorSetLayout
  4408. ) );
  4409. s_renderVK->m_descriptorSetLayoutCache.add(descriptorSetLayoutHash, m_descriptorSetLayout);
  4410. }
  4411. }
  4412. VkPipelineLayoutCreateInfo plci;
  4413. plci.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO;
  4414. plci.pNext = NULL;
  4415. plci.flags = 0;
  4416. plci.pushConstantRangeCount = 0;
  4417. plci.pPushConstantRanges = NULL;
  4418. plci.setLayoutCount = (m_descriptorSetLayout == VK_NULL_HANDLE ? 0 : 1);
  4419. plci.pSetLayouts = &m_descriptorSetLayout;
  4420. VK_CHECK(vkCreatePipelineLayout(
  4421. s_renderVK->m_device
  4422. , &plci
  4423. , s_renderVK->m_allocatorCb
  4424. , &m_pipelineLayout
  4425. ) );
  4426. }
  4427. void ProgramVK::destroy()
  4428. {
  4429. s_renderVK->release(m_pipelineLayout);
  4430. m_numPredefined = 0;
  4431. m_vsh = NULL;
  4432. m_fsh = NULL;
  4433. }
  4434. VkResult TimerQueryVK::init()
  4435. {
  4436. BGFX_PROFILER_SCOPE("TimerQueryVK::init", kColorFrame);
  4437. VkResult result = VK_SUCCESS;
  4438. const VkDevice device = s_renderVK->m_device;
  4439. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4440. const uint32_t count = m_control.m_size * 2;
  4441. VkQueryPoolCreateInfo qpci;
  4442. qpci.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
  4443. qpci.pNext = NULL;
  4444. qpci.flags = 0;
  4445. qpci.queryType = VK_QUERY_TYPE_TIMESTAMP;
  4446. qpci.queryCount = count;
  4447. qpci.pipelineStatistics = 0;
  4448. result = vkCreateQueryPool(device, &qpci, s_renderVK->m_allocatorCb, &m_queryPool);
  4449. if (VK_SUCCESS != result)
  4450. {
  4451. BX_TRACE("Create timer query error: vkCreateQueryPool failed %d: %s.", result, getName(result) );
  4452. return result;
  4453. }
  4454. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4455. const uint32_t size = count * sizeof(uint64_t);
  4456. result = s_renderVK->createReadbackBuffer(size, &m_readback, &m_readbackMemory);
  4457. if (VK_SUCCESS != result)
  4458. {
  4459. return result;
  4460. }
  4461. result = vkMapMemory(device, m_readbackMemory, 0, VK_WHOLE_SIZE, 0, (void**)&m_queryResult);
  4462. if (VK_SUCCESS != result)
  4463. {
  4464. BX_TRACE("Create timer query error: vkMapMemory failed %d: %s.", result, getName(result) );
  4465. return result;
  4466. }
  4467. m_frequency = uint64_t(1000000000.0 / double(s_renderVK->m_deviceProperties.limits.timestampPeriod) );
  4468. for (uint32_t ii = 0; ii < BX_COUNTOF(m_result); ++ii)
  4469. {
  4470. m_result[ii].reset();
  4471. }
  4472. m_control.reset();
  4473. return result;
  4474. }
  4475. void TimerQueryVK::shutdown()
  4476. {
  4477. vkDestroy(m_queryPool);
  4478. vkDestroy(m_readback);
  4479. vkUnmapMemory(s_renderVK->m_device, m_readbackMemory);
  4480. vkDestroy(m_readbackMemory);
  4481. }
  4482. uint32_t TimerQueryVK::begin(uint32_t _resultIdx, uint32_t _frameNum)
  4483. {
  4484. BGFX_PROFILER_SCOPE("TimerQueryVK::begin", kColorFrame);
  4485. while (0 == m_control.reserve(1) )
  4486. {
  4487. m_control.consume(1);
  4488. }
  4489. Result& result = m_result[_resultIdx];
  4490. ++result.m_pending;
  4491. const uint32_t idx = m_control.m_current;
  4492. Query& query = m_query[idx];
  4493. query.m_resultIdx = _resultIdx;
  4494. query.m_ready = false;
  4495. query.m_frameNum = _frameNum;
  4496. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4497. const uint32_t offset = idx * 2 + 0;
  4498. vkCmdResetQueryPool(commandBuffer, m_queryPool, offset, 2);
  4499. vkCmdWriteTimestamp(commandBuffer, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT, m_queryPool, offset + 0);
  4500. m_control.commit(1);
  4501. return idx;
  4502. }
  4503. void TimerQueryVK::end(uint32_t _idx)
  4504. {
  4505. BGFX_PROFILER_SCOPE("TimerQueryVK::end", kColorFrame);
  4506. Query& query = m_query[_idx];
  4507. query.m_ready = true;
  4508. query.m_completed = s_renderVK->m_cmd.m_submitted + s_renderVK->m_cmd.m_numFramesInFlight;
  4509. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4510. const uint32_t offset = _idx * 2 + 0;
  4511. vkCmdWriteTimestamp(commandBuffer, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT, m_queryPool, offset + 1);
  4512. vkCmdCopyQueryPoolResults(
  4513. commandBuffer
  4514. , m_queryPool
  4515. , offset
  4516. , 2
  4517. , m_readback
  4518. , offset * sizeof(uint64_t)
  4519. , sizeof(uint64_t)
  4520. , VK_QUERY_RESULT_WAIT_BIT | VK_QUERY_RESULT_64_BIT
  4521. );
  4522. setMemoryBarrier(commandBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT);
  4523. while (update() )
  4524. {
  4525. }
  4526. }
  4527. bool TimerQueryVK::update()
  4528. {
  4529. if (0 != m_control.available() )
  4530. {
  4531. uint32_t idx = m_control.m_read;
  4532. Query& query = m_query[idx];
  4533. if (!query.m_ready)
  4534. {
  4535. return false;
  4536. }
  4537. if (query.m_completed > s_renderVK->m_cmd.m_submitted)
  4538. {
  4539. return false;
  4540. }
  4541. m_control.consume(1);
  4542. Result& result = m_result[query.m_resultIdx];
  4543. --result.m_pending;
  4544. result.m_frameNum = query.m_frameNum;
  4545. uint32_t offset = idx * 2;
  4546. result.m_begin = m_queryResult[offset+0];
  4547. result.m_end = m_queryResult[offset+1];
  4548. return true;
  4549. }
  4550. return false;
  4551. }
  4552. VkResult OcclusionQueryVK::init()
  4553. {
  4554. BGFX_PROFILER_SCOPE("OcclusionQueryVK::init", kColorFrame);
  4555. VkResult result = VK_SUCCESS;
  4556. const VkDevice device = s_renderVK->m_device;
  4557. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4558. const uint32_t count = BX_COUNTOF(m_handle);
  4559. VkQueryPoolCreateInfo qpci;
  4560. qpci.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
  4561. qpci.pNext = NULL;
  4562. qpci.flags = 0;
  4563. qpci.queryType = VK_QUERY_TYPE_OCCLUSION;
  4564. qpci.queryCount = count;
  4565. qpci.pipelineStatistics = 0;
  4566. result = vkCreateQueryPool(device, &qpci, s_renderVK->m_allocatorCb, &m_queryPool);
  4567. if (VK_SUCCESS != result)
  4568. {
  4569. BX_TRACE("Create occlusion query error: vkCreateQueryPool failed %d: %s.", result, getName(result) );
  4570. return result;
  4571. }
  4572. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4573. const uint32_t size = count * sizeof(uint32_t);
  4574. result = s_renderVK->createReadbackBuffer(size, &m_readback, &m_readbackMemory);
  4575. if (VK_SUCCESS != result)
  4576. {
  4577. return result;
  4578. }
  4579. result = vkMapMemory(device, m_readbackMemory, 0, VK_WHOLE_SIZE, 0, (void**)&m_queryResult);
  4580. if (VK_SUCCESS != result)
  4581. {
  4582. BX_TRACE("Create occlusion query error: vkMapMemory failed %d: %s.", result, getName(result) );
  4583. return result;
  4584. }
  4585. m_control.reset();
  4586. return result;
  4587. }
  4588. void OcclusionQueryVK::shutdown()
  4589. {
  4590. vkDestroy(m_queryPool);
  4591. vkDestroy(m_readback);
  4592. vkUnmapMemory(s_renderVK->m_device, m_readbackMemory);
  4593. vkDestroy(m_readbackMemory);
  4594. }
  4595. void OcclusionQueryVK::begin(OcclusionQueryHandle _handle)
  4596. {
  4597. BGFX_PROFILER_SCOPE("OcclusionQueryVK::shutdown", kColorFrame);
  4598. m_control.reserve(1);
  4599. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4600. m_handle[m_control.m_current] = _handle;
  4601. vkCmdBeginQuery(commandBuffer, m_queryPool, _handle.idx, 0);
  4602. }
  4603. void OcclusionQueryVK::end()
  4604. {
  4605. BGFX_PROFILER_SCOPE("OcclusionQueryVK::end", kColorFrame);
  4606. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4607. const OcclusionQueryHandle handle = m_handle[m_control.m_current];
  4608. vkCmdEndQuery(commandBuffer, m_queryPool, handle.idx);
  4609. m_control.commit(1);
  4610. }
  4611. void OcclusionQueryVK::flush(Frame* _render)
  4612. {
  4613. BGFX_PROFILER_SCOPE("OcclusionQueryVK::flush", kColorFrame);
  4614. if (0 < m_control.available() )
  4615. {
  4616. VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4617. const uint32_t size = m_control.m_size;
  4618. // need to copy each result individually because VK_QUERY_RESULT_WAIT_BIT causes
  4619. // vkWaitForFences to hang indefinitely if we copy all results (including unavailable ones)
  4620. for (uint32_t ii = 0, num = m_control.available(); ii < num; ++ii)
  4621. {
  4622. const OcclusionQueryHandle& handle = m_handle[(m_control.m_read + ii) % size];
  4623. if (isValid(handle) )
  4624. {
  4625. vkCmdCopyQueryPoolResults(
  4626. commandBuffer
  4627. , m_queryPool
  4628. , handle.idx
  4629. , 1
  4630. , m_readback
  4631. , handle.idx * sizeof(uint32_t)
  4632. , sizeof(uint32_t)
  4633. , VK_QUERY_RESULT_WAIT_BIT
  4634. );
  4635. }
  4636. }
  4637. setMemoryBarrier(commandBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT);
  4638. s_renderVK->kick(true);
  4639. commandBuffer = s_renderVK->m_commandBuffer;
  4640. // resetting in the new command buffer prevents a false positive validation layer error
  4641. const uint32_t count = BX_COUNTOF(m_handle);
  4642. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4643. resolve(_render);
  4644. }
  4645. }
  4646. void OcclusionQueryVK::resolve(Frame* _render)
  4647. {
  4648. while (0 != m_control.available() )
  4649. {
  4650. OcclusionQueryHandle handle = m_handle[m_control.m_read];
  4651. if (isValid(handle) )
  4652. {
  4653. _render->m_occlusion[handle.idx] = m_queryResult[handle.idx];
  4654. }
  4655. m_control.consume(1);
  4656. }
  4657. }
  4658. void OcclusionQueryVK::invalidate(OcclusionQueryHandle _handle)
  4659. {
  4660. const uint32_t size = m_control.m_size;
  4661. for (uint32_t ii = 0, num = m_control.available(); ii < num; ++ii)
  4662. {
  4663. OcclusionQueryHandle& handle = m_handle[(m_control.m_read + ii) % size];
  4664. if (handle.idx == _handle.idx)
  4665. {
  4666. handle.idx = bgfx::kInvalidHandle;
  4667. }
  4668. }
  4669. }
  4670. void ReadbackVK::create(VkImage _image, uint32_t _width, uint32_t _height, TextureFormat::Enum _format)
  4671. {
  4672. m_image = _image;
  4673. m_width = _width;
  4674. m_height = _height;
  4675. m_format = _format;
  4676. }
  4677. void ReadbackVK::destroy()
  4678. {
  4679. m_image = VK_NULL_HANDLE;
  4680. }
  4681. uint32_t ReadbackVK::pitch(uint8_t _mip) const
  4682. {
  4683. uint32_t mipWidth = bx::uint32_max(1, m_width >> _mip);
  4684. uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_format) );
  4685. return mipWidth * bpp / 8;
  4686. }
  4687. void ReadbackVK::copyImageToBuffer(VkCommandBuffer _commandBuffer, VkBuffer _buffer, VkImageLayout _layout, VkImageAspectFlags _aspect, uint8_t _mip) const
  4688. {
  4689. BGFX_PROFILER_SCOPE("ReadbackVK::copyImageToBuffer", kColorFrame);
  4690. uint32_t mipWidth = bx::uint32_max(1, m_width >> _mip);
  4691. uint32_t mipHeight = bx::uint32_max(1, m_height >> _mip);
  4692. setImageMemoryBarrier(
  4693. _commandBuffer
  4694. , m_image
  4695. , _aspect
  4696. , _layout
  4697. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4698. , _mip
  4699. , 1
  4700. , 0
  4701. , 1
  4702. );
  4703. VkBufferImageCopy bic;
  4704. bic.bufferOffset = 0;
  4705. bic.bufferRowLength = mipWidth;
  4706. bic.bufferImageHeight = mipHeight;
  4707. bic.imageSubresource.aspectMask = _aspect;
  4708. bic.imageSubresource.mipLevel = _mip;
  4709. bic.imageSubresource.baseArrayLayer = 0;
  4710. bic.imageSubresource.layerCount = 1;
  4711. bic.imageOffset = { 0, 0, 0 };
  4712. bic.imageExtent = { mipWidth, mipHeight, 1 };
  4713. vkCmdCopyImageToBuffer(
  4714. _commandBuffer
  4715. , m_image
  4716. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4717. , _buffer
  4718. , 1
  4719. , &bic
  4720. );
  4721. // Make changes to the buffer visible to the host
  4722. setMemoryBarrier(
  4723. _commandBuffer
  4724. , VK_PIPELINE_STAGE_TRANSFER_BIT
  4725. , VK_PIPELINE_STAGE_HOST_BIT
  4726. );
  4727. setImageMemoryBarrier(
  4728. _commandBuffer
  4729. , m_image
  4730. , _aspect
  4731. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4732. , _layout
  4733. , _mip
  4734. , 1
  4735. , 0
  4736. , 1
  4737. );
  4738. }
  4739. void ReadbackVK::readback(VkDeviceMemory _memory, VkDeviceSize _offset, void* _data, uint8_t _mip) const
  4740. {
  4741. if (m_image == VK_NULL_HANDLE)
  4742. {
  4743. return;
  4744. }
  4745. uint32_t mipHeight = bx::uint32_max(1, m_height >> _mip);
  4746. uint32_t rowPitch = pitch(_mip);
  4747. uint8_t* src;
  4748. VK_CHECK(vkMapMemory(s_renderVK->m_device, _memory, 0, VK_WHOLE_SIZE, 0, (void**)&src) );
  4749. src += _offset;
  4750. uint8_t* dst = (uint8_t*)_data;
  4751. for (uint32_t yy = 0; yy < mipHeight; ++yy)
  4752. {
  4753. bx::memCopy(dst, src, rowPitch);
  4754. src += rowPitch;
  4755. dst += rowPitch;
  4756. }
  4757. vkUnmapMemory(s_renderVK->m_device, _memory);
  4758. }
  4759. VkResult TextureVK::create(VkCommandBuffer _commandBuffer, uint32_t _width, uint32_t _height, uint64_t _flags, VkFormat _format)
  4760. {
  4761. BX_ASSERT(0 != (_flags & BGFX_TEXTURE_RT_MASK), "");
  4762. _flags |= BGFX_TEXTURE_RT_WRITE_ONLY;
  4763. m_flags = _flags;
  4764. m_width = _width;
  4765. m_height = _height;
  4766. m_depth = 1;
  4767. m_numLayers = 1;
  4768. m_requestedFormat = uint8_t(bimg::TextureFormat::Count);
  4769. m_textureFormat = uint8_t(bimg::TextureFormat::Count);
  4770. m_format = _format;
  4771. m_components = { VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY };
  4772. m_aspectMask = getAspectMask(m_format);
  4773. m_sampler = s_msaa[bx::uint32_satsub( (m_flags & BGFX_TEXTURE_RT_MSAA_MASK) >> BGFX_TEXTURE_RT_MSAA_SHIFT, 1)];
  4774. m_type = VK_IMAGE_VIEW_TYPE_2D;
  4775. m_numMips = 1;
  4776. m_numSides = 1;
  4777. VkResult result = createImages(_commandBuffer);
  4778. if (VK_SUCCESS == result)
  4779. {
  4780. const VkImageLayout layout = 0 != (m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  4781. ? VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL
  4782. : VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  4783. ;
  4784. setImageMemoryBarrier(_commandBuffer, layout);
  4785. }
  4786. return result;
  4787. }
  4788. VkResult TextureVK::createImages(VkCommandBuffer _commandBuffer)
  4789. {
  4790. VkResult result = VK_SUCCESS;
  4791. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  4792. const VkDevice device = s_renderVK->m_device;
  4793. if (m_sampler.Count > 1)
  4794. {
  4795. BX_ASSERT(VK_IMAGE_VIEW_TYPE_3D != m_type, "Can't create multisample 3D image.");
  4796. BX_ASSERT(m_numMips <= 1, "Can't create multisample image with mip chain.");
  4797. }
  4798. // create texture and allocate its device memory
  4799. VkImageCreateInfo ici;
  4800. ici.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
  4801. ici.pNext = NULL;
  4802. ici.flags = 0
  4803. | (VK_IMAGE_VIEW_TYPE_CUBE == m_type
  4804. ? VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT
  4805. : 0
  4806. )
  4807. | (VK_IMAGE_VIEW_TYPE_3D == m_type
  4808. ? VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT_KHR
  4809. : 0
  4810. )
  4811. ;
  4812. ici.pQueueFamilyIndices = NULL;
  4813. ici.queueFamilyIndexCount = 0;
  4814. ici.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  4815. ici.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  4816. ici.usage = 0
  4817. | VK_IMAGE_USAGE_TRANSFER_SRC_BIT
  4818. | VK_IMAGE_USAGE_TRANSFER_DST_BIT
  4819. | VK_IMAGE_USAGE_SAMPLED_BIT
  4820. | (m_flags & BGFX_TEXTURE_RT_MASK
  4821. ? (m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT)
  4822. ? VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT
  4823. : VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT)
  4824. : 0
  4825. )
  4826. | (m_flags & BGFX_TEXTURE_COMPUTE_WRITE ? VK_IMAGE_USAGE_STORAGE_BIT : 0)
  4827. ;
  4828. ici.format = m_format;
  4829. ici.samples = m_sampler.Sample;
  4830. ici.mipLevels = m_numMips;
  4831. ici.arrayLayers = m_numSides;
  4832. ici.extent.width = m_width;
  4833. ici.extent.height = m_height;
  4834. ici.extent.depth = m_depth;
  4835. ici.imageType = VK_IMAGE_VIEW_TYPE_3D == m_type
  4836. ? VK_IMAGE_TYPE_3D
  4837. : VK_IMAGE_TYPE_2D
  4838. ;
  4839. ici.tiling = VK_IMAGE_TILING_OPTIMAL;
  4840. result = vkCreateImage(device, &ici, allocatorCb, &m_textureImage);
  4841. if (VK_SUCCESS != result)
  4842. {
  4843. BX_TRACE("Create texture image error: vkCreateImage failed %d: %s.", result, getName(result) );
  4844. return result;
  4845. }
  4846. VkMemoryRequirements imageMemReq;
  4847. vkGetImageMemoryRequirements(device, m_textureImage, &imageMemReq);
  4848. result = s_renderVK->allocateMemory(&imageMemReq, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_textureDeviceMem);
  4849. if (VK_SUCCESS != result)
  4850. {
  4851. BX_TRACE("Create texture image error: allocateMemory failed %d: %s.", result, getName(result) );
  4852. return result;
  4853. }
  4854. result = vkBindImageMemory(device, m_textureImage, m_textureDeviceMem, 0);
  4855. if (VK_SUCCESS != result)
  4856. {
  4857. BX_TRACE("Create texture image error: vkBindImageMemory failed %d: %s.", result, getName(result) );
  4858. return result;
  4859. }
  4860. m_sampledLayout = m_flags & BGFX_TEXTURE_COMPUTE_WRITE
  4861. ? VK_IMAGE_LAYOUT_GENERAL
  4862. : VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL
  4863. ;
  4864. const bool needResolve = true
  4865. && 1 < m_sampler.Count
  4866. && 0 != (ici.usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT)
  4867. && 0 == (m_flags & BGFX_TEXTURE_MSAA_SAMPLE)
  4868. && 0 == (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY)
  4869. ;
  4870. if (needResolve)
  4871. {
  4872. VkImageCreateInfo ici_resolve = ici;
  4873. ici_resolve.samples = s_msaa[0].Sample;
  4874. ici_resolve.usage &= ~VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
  4875. ici_resolve.flags &= ~VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
  4876. result = vkCreateImage(device, &ici_resolve, allocatorCb, &m_singleMsaaImage);
  4877. if (VK_SUCCESS != result)
  4878. {
  4879. BX_TRACE("Create texture image error: vkCreateImage failed %d: %s.", result, getName(result) );
  4880. return result;
  4881. }
  4882. VkMemoryRequirements imageMemReq_resolve;
  4883. vkGetImageMemoryRequirements(device, m_singleMsaaImage, &imageMemReq_resolve);
  4884. result = s_renderVK->allocateMemory(&imageMemReq_resolve, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_singleMsaaDeviceMem);
  4885. if (VK_SUCCESS != result)
  4886. {
  4887. BX_TRACE("Create texture image error: allocateMemory failed %d: %s.", result, getName(result) );
  4888. return result;
  4889. }
  4890. result = vkBindImageMemory(device, m_singleMsaaImage, m_singleMsaaDeviceMem, 0);
  4891. if (VK_SUCCESS != result)
  4892. {
  4893. BX_TRACE("Create texture image error: vkBindImageMemory failed %d: %s.", result, getName(result) );
  4894. return result;
  4895. }
  4896. setImageMemoryBarrier(_commandBuffer, m_sampledLayout, true);
  4897. }
  4898. return result;
  4899. }
  4900. void* TextureVK::create(VkCommandBuffer _commandBuffer, const Memory* _mem, uint64_t _flags, uint8_t _skip)
  4901. {
  4902. bimg::ImageContainer imageContainer;
  4903. if (bimg::imageParse(imageContainer, _mem->data, _mem->size) )
  4904. {
  4905. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(imageContainer.m_format);
  4906. const uint8_t startLod = bx::min<uint8_t>(_skip, imageContainer.m_numMips - 1);
  4907. bimg::TextureInfo ti;
  4908. bimg::imageGetSize(
  4909. &ti
  4910. , uint16_t(imageContainer.m_width >> startLod)
  4911. , uint16_t(imageContainer.m_height >> startLod)
  4912. , uint16_t(imageContainer.m_depth >> startLod)
  4913. , imageContainer.m_cubeMap
  4914. , 1 < imageContainer.m_numMips
  4915. , imageContainer.m_numLayers
  4916. , imageContainer.m_format
  4917. );
  4918. ti.numMips = bx::min<uint8_t>(imageContainer.m_numMips - startLod, ti.numMips);
  4919. m_flags = _flags;
  4920. m_width = ti.width;
  4921. m_height = ti.height;
  4922. m_depth = ti.depth;
  4923. m_numLayers = ti.numLayers;
  4924. m_requestedFormat = uint8_t(imageContainer.m_format);
  4925. m_textureFormat = uint8_t(getViableTextureFormat(imageContainer) );
  4926. m_format = bimg::isDepth(bimg::TextureFormat::Enum(m_textureFormat) )
  4927. ? s_textureFormat[m_textureFormat].m_fmtDsv
  4928. : (m_flags & BGFX_TEXTURE_SRGB) ? s_textureFormat[m_textureFormat].m_fmtSrgb : s_textureFormat[m_textureFormat].m_fmt
  4929. ;
  4930. m_components = s_textureFormat[m_textureFormat].m_mapping;
  4931. const bool convert = m_textureFormat != m_requestedFormat;
  4932. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_textureFormat) );
  4933. m_aspectMask = getAspectMask(m_format);
  4934. m_sampler = s_msaa[bx::uint32_satsub( (m_flags & BGFX_TEXTURE_RT_MSAA_MASK) >> BGFX_TEXTURE_RT_MSAA_SHIFT, 1)];
  4935. if (imageContainer.m_cubeMap)
  4936. {
  4937. m_type = imageContainer.m_numLayers > 1
  4938. ? VK_IMAGE_VIEW_TYPE_CUBE_ARRAY
  4939. : VK_IMAGE_VIEW_TYPE_CUBE
  4940. ;
  4941. }
  4942. else if (imageContainer.m_depth > 1)
  4943. {
  4944. m_type = VK_IMAGE_VIEW_TYPE_3D;
  4945. }
  4946. else if (imageContainer.m_numLayers > 1)
  4947. {
  4948. m_type = VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  4949. }
  4950. else
  4951. {
  4952. m_type = VK_IMAGE_VIEW_TYPE_2D;
  4953. }
  4954. m_numMips = ti.numMips;
  4955. m_numSides = ti.numLayers * (imageContainer.m_cubeMap ? 6 : 1);
  4956. const uint16_t numSides = ti.numLayers * (imageContainer.m_cubeMap ? 6 : 1);
  4957. const uint32_t numSrd = numSides * ti.numMips;
  4958. uint32_t kk = 0;
  4959. const bool compressed = bimg::isCompressed(bimg::TextureFormat::Enum(m_textureFormat) );
  4960. const bool swizzle = TextureFormat::BGRA8 == m_textureFormat && 0 != (m_flags & BGFX_TEXTURE_COMPUTE_WRITE);
  4961. const bool writeOnly = 0 != (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY);
  4962. const bool computeWrite = 0 != (m_flags & BGFX_TEXTURE_COMPUTE_WRITE);
  4963. const bool renderTarget = 0 != (m_flags & BGFX_TEXTURE_RT_MASK);
  4964. const bool blit = 0 != (m_flags & BGFX_TEXTURE_BLIT_DST);
  4965. BX_UNUSED(swizzle, writeOnly, computeWrite, renderTarget, blit);
  4966. BX_TRACE(
  4967. "Texture %3d: %s (requested: %s), %dx%dx%d%s RT[%c], BO[%c], CW[%c]%s."
  4968. , (int)(this - s_renderVK->m_textures)
  4969. , getName( (TextureFormat::Enum)m_textureFormat)
  4970. , getName( (TextureFormat::Enum)m_requestedFormat)
  4971. , ti.width
  4972. , ti.height
  4973. , ti.depth
  4974. , imageContainer.m_cubeMap ? "x6" : ""
  4975. , renderTarget ? 'x' : ' '
  4976. , writeOnly ? 'x' : ' '
  4977. , computeWrite ? 'x' : ' '
  4978. , swizzle ? " (swizzle BGRA8 -> RGBA8)" : ""
  4979. );
  4980. VK_CHECK(createImages(_commandBuffer) );
  4981. // decode images
  4982. struct ImageInfo
  4983. {
  4984. uint8_t* data;
  4985. uint32_t width;
  4986. uint32_t height;
  4987. uint32_t depth;
  4988. uint32_t pitch;
  4989. uint32_t slice;
  4990. uint32_t size;
  4991. uint32_t mipLevel;
  4992. uint32_t layer;
  4993. };
  4994. ImageInfo* imageInfos = (ImageInfo*)bx::alloc(g_allocator, sizeof(ImageInfo) * numSrd);
  4995. bx::memSet(imageInfos, 0, sizeof(ImageInfo) * numSrd);
  4996. uint32_t alignment = 1; // tightly aligned buffer
  4997. for (uint16_t side = 0; side < numSides; ++side)
  4998. {
  4999. for (uint8_t lod = 0; lod < ti.numMips; ++lod)
  5000. {
  5001. bimg::ImageMip mip;
  5002. if (bimg::imageGetRawData(imageContainer, side, lod + startLod, _mem->data, _mem->size, mip) )
  5003. {
  5004. if (convert)
  5005. {
  5006. const uint32_t pitch = bx::strideAlign(bx::max<uint32_t>(mip.m_width, 4) * bpp / 8, alignment);
  5007. const uint32_t slice = bx::strideAlign(bx::max<uint32_t>(mip.m_height, 4) * pitch, alignment);
  5008. const uint32_t size = slice * mip.m_depth;
  5009. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5010. bimg::imageDecodeToBgra8(
  5011. g_allocator
  5012. , temp
  5013. , mip.m_data
  5014. , mip.m_width
  5015. , mip.m_height
  5016. , pitch
  5017. , mip.m_format
  5018. );
  5019. imageInfos[kk].data = temp;
  5020. imageInfos[kk].width = mip.m_width;
  5021. imageInfos[kk].height = mip.m_height;
  5022. imageInfos[kk].depth = mip.m_depth;
  5023. imageInfos[kk].pitch = pitch;
  5024. imageInfos[kk].slice = slice;
  5025. imageInfos[kk].size = size;
  5026. imageInfos[kk].mipLevel = lod;
  5027. imageInfos[kk].layer = side;
  5028. }
  5029. else if (compressed)
  5030. {
  5031. const uint32_t pitch = bx::strideAlign( (mip.m_width / blockInfo.blockWidth) * mip.m_blockSize, alignment);
  5032. const uint32_t slice = bx::strideAlign( (mip.m_height / blockInfo.blockHeight) * pitch, alignment);
  5033. const uint32_t size = slice * mip.m_depth;
  5034. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5035. bimg::imageCopy(
  5036. temp
  5037. , mip.m_height / blockInfo.blockHeight
  5038. , (mip.m_width / blockInfo.blockWidth) * mip.m_blockSize
  5039. , mip.m_depth
  5040. , mip.m_data
  5041. , pitch
  5042. );
  5043. imageInfos[kk].data = temp;
  5044. imageInfos[kk].width = mip.m_width;
  5045. imageInfos[kk].height = mip.m_height;
  5046. imageInfos[kk].depth = mip.m_depth;
  5047. imageInfos[kk].pitch = pitch;
  5048. imageInfos[kk].slice = slice;
  5049. imageInfos[kk].size = size;
  5050. imageInfos[kk].mipLevel = lod;
  5051. imageInfos[kk].layer = side;
  5052. }
  5053. else
  5054. {
  5055. const uint32_t pitch = bx::strideAlign(mip.m_width * mip.m_bpp / 8, alignment);
  5056. const uint32_t slice = bx::strideAlign(mip.m_height * pitch, alignment);
  5057. const uint32_t size = slice * mip.m_depth;
  5058. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5059. bimg::imageCopy(
  5060. temp
  5061. , mip.m_height
  5062. , mip.m_width * mip.m_bpp / 8
  5063. , mip.m_depth
  5064. , mip.m_data
  5065. , pitch
  5066. );
  5067. imageInfos[kk].data = temp;
  5068. imageInfos[kk].width = mip.m_width;
  5069. imageInfos[kk].height = mip.m_height;
  5070. imageInfos[kk].depth = mip.m_depth;
  5071. imageInfos[kk].pitch = pitch;
  5072. imageInfos[kk].slice = slice;
  5073. imageInfos[kk].size = size;
  5074. imageInfos[kk].mipLevel = lod;
  5075. imageInfos[kk].layer = side;
  5076. }
  5077. }
  5078. ++kk;
  5079. }
  5080. }
  5081. uint32_t totalMemSize = 0;
  5082. VkBufferImageCopy* bufferCopyInfo = (VkBufferImageCopy*)bx::alloc(g_allocator, sizeof(VkBufferImageCopy) * numSrd);
  5083. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5084. {
  5085. const uint32_t idealWidth = bx::max<uint32_t>(1, m_width >> imageInfos[ii].mipLevel);
  5086. const uint32_t idealHeight = bx::max<uint32_t>(1, m_height >> imageInfos[ii].mipLevel);
  5087. bufferCopyInfo[ii].bufferOffset = totalMemSize;
  5088. bufferCopyInfo[ii].bufferRowLength = 0; // assume that image data are tightly aligned
  5089. bufferCopyInfo[ii].bufferImageHeight = 0; // assume that image data are tightly aligned
  5090. bufferCopyInfo[ii].imageSubresource.aspectMask = m_aspectMask;
  5091. bufferCopyInfo[ii].imageSubresource.mipLevel = imageInfos[ii].mipLevel;
  5092. bufferCopyInfo[ii].imageSubresource.baseArrayLayer = imageInfos[ii].layer;
  5093. bufferCopyInfo[ii].imageSubresource.layerCount = 1;
  5094. bufferCopyInfo[ii].imageOffset = { 0, 0, 0 };
  5095. bufferCopyInfo[ii].imageExtent = { idealWidth, idealHeight, imageInfos[ii].depth };
  5096. totalMemSize += imageInfos[ii].size;
  5097. }
  5098. if (totalMemSize > 0)
  5099. {
  5100. const VkDevice device = s_renderVK->m_device;
  5101. VkBuffer stagingBuffer;
  5102. VkDeviceMemory stagingDeviceMem;
  5103. VK_CHECK(s_renderVK->createStagingBuffer(totalMemSize, &stagingBuffer, &stagingDeviceMem) );
  5104. uint8_t* mappedMemory;
  5105. VK_CHECK(vkMapMemory(
  5106. device
  5107. , stagingDeviceMem
  5108. , 0
  5109. , totalMemSize
  5110. , 0
  5111. , (void**)&mappedMemory
  5112. ) );
  5113. // copy image to staging buffer
  5114. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5115. {
  5116. bx::memCopy(mappedMemory, imageInfos[ii].data, imageInfos[ii].size);
  5117. mappedMemory += imageInfos[ii].size;
  5118. }
  5119. vkUnmapMemory(device, stagingDeviceMem);
  5120. copyBufferToTexture(_commandBuffer, stagingBuffer, numSrd, bufferCopyInfo);
  5121. s_renderVK->release(stagingBuffer);
  5122. s_renderVK->release(stagingDeviceMem);
  5123. }
  5124. else
  5125. {
  5126. setImageMemoryBarrier(_commandBuffer, m_sampledLayout);
  5127. }
  5128. bx::free(g_allocator, bufferCopyInfo);
  5129. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5130. {
  5131. bx::free(g_allocator, imageInfos[ii].data);
  5132. }
  5133. bx::free(g_allocator, imageInfos);
  5134. m_readback.create(m_textureImage, m_width, m_height, TextureFormat::Enum(m_textureFormat) );
  5135. }
  5136. return m_directAccessPtr;
  5137. }
  5138. void TextureVK::destroy()
  5139. {
  5140. m_readback.destroy();
  5141. if (VK_NULL_HANDLE != m_textureImage)
  5142. {
  5143. s_renderVK->release(m_textureImage);
  5144. s_renderVK->release(m_textureDeviceMem);
  5145. }
  5146. if (VK_NULL_HANDLE != m_singleMsaaImage)
  5147. {
  5148. s_renderVK->release(m_singleMsaaImage);
  5149. s_renderVK->release(m_singleMsaaDeviceMem);
  5150. }
  5151. m_currentImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  5152. m_currentSingleMsaaImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  5153. }
  5154. void TextureVK::update(VkCommandBuffer _commandBuffer, uint8_t _side, uint8_t _mip, const Rect& _rect, uint16_t _z, uint16_t _depth, uint16_t _pitch, const Memory* _mem)
  5155. {
  5156. const uint32_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_textureFormat) );
  5157. uint32_t rectpitch = _rect.m_width * bpp / 8;
  5158. uint32_t slicepitch = rectpitch * _rect.m_height;
  5159. if (bimg::isCompressed(bimg::TextureFormat::Enum(m_textureFormat) ) )
  5160. {
  5161. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(bimg::TextureFormat::Enum(m_textureFormat) );
  5162. rectpitch = (_rect.m_width / blockInfo.blockWidth ) * blockInfo.blockSize;
  5163. slicepitch = (_rect.m_height / blockInfo.blockHeight) * rectpitch;
  5164. }
  5165. const uint32_t srcpitch = UINT16_MAX == _pitch ? rectpitch : _pitch;
  5166. const uint32_t size = UINT16_MAX == _pitch ? slicepitch * _depth: _rect.m_height * _pitch * _depth;
  5167. const bool convert = m_textureFormat != m_requestedFormat;
  5168. VkBufferImageCopy region;
  5169. region.bufferOffset = 0;
  5170. region.bufferRowLength = (_pitch == UINT16_MAX ? 0 : _pitch * 8 / bpp);
  5171. region.bufferImageHeight = 0;
  5172. region.imageSubresource.aspectMask = m_aspectMask;
  5173. region.imageSubresource.mipLevel = _mip;
  5174. region.imageSubresource.baseArrayLayer = 0;
  5175. region.imageSubresource.layerCount = 1;
  5176. region.imageOffset = { _rect.m_x, _rect.m_y, 0 };
  5177. region.imageExtent = { _rect.m_width, _rect.m_height, _depth };
  5178. uint8_t* data = _mem->data;
  5179. uint8_t* temp = NULL;
  5180. if (convert)
  5181. {
  5182. temp = (uint8_t*)bx::alloc(g_allocator, slicepitch);
  5183. bimg::imageDecodeToBgra8(g_allocator, temp, data, _rect.m_width, _rect.m_height, srcpitch, bimg::TextureFormat::Enum(m_requestedFormat));
  5184. data = temp;
  5185. region.imageExtent =
  5186. {
  5187. bx::max(1u, m_width >> _mip),
  5188. bx::max(1u, m_height >> _mip),
  5189. _depth,
  5190. };
  5191. }
  5192. VkBuffer stagingBuffer = VK_NULL_HANDLE;
  5193. VkDeviceMemory stagingDeviceMem = VK_NULL_HANDLE;
  5194. VK_CHECK(s_renderVK->createStagingBuffer(size, &stagingBuffer, &stagingDeviceMem, data) );
  5195. if (VK_IMAGE_VIEW_TYPE_3D == m_type)
  5196. {
  5197. region.imageOffset.z = _z;
  5198. }
  5199. else if (VK_IMAGE_VIEW_TYPE_CUBE == m_type
  5200. || VK_IMAGE_VIEW_TYPE_CUBE_ARRAY == m_type)
  5201. {
  5202. region.imageSubresource.baseArrayLayer = _z * 6 + _side;
  5203. }
  5204. else
  5205. {
  5206. region.imageSubresource.baseArrayLayer = _z;
  5207. }
  5208. copyBufferToTexture(_commandBuffer, stagingBuffer, 1, &region);
  5209. s_renderVK->release(stagingBuffer);
  5210. s_renderVK->release(stagingDeviceMem);
  5211. if (NULL != temp)
  5212. {
  5213. bx::free(g_allocator, temp);
  5214. }
  5215. }
  5216. void TextureVK::resolve(VkCommandBuffer _commandBuffer, uint8_t _resolve, uint32_t _layer, uint32_t _numLayers, uint32_t _mip)
  5217. {
  5218. BGFX_PROFILER_SCOPE("TextureVK::resolve", kColorResource);
  5219. const bool needResolve = VK_NULL_HANDLE != m_singleMsaaImage;
  5220. const bool needMipGen = true
  5221. && !needResolve
  5222. && 0 != (m_flags & BGFX_TEXTURE_RT_MASK)
  5223. && 0 == (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY)
  5224. && (_mip + 1) < m_numMips
  5225. && 0 != (_resolve & BGFX_RESOLVE_AUTO_GEN_MIPS);
  5226. const VkImageLayout oldLayout = m_currentImageLayout;
  5227. const VkImageLayout oldSingleMsaaLayout = m_currentSingleMsaaImageLayout;
  5228. const uint32_t numLayers = false
  5229. || m_type == VK_IMAGE_VIEW_TYPE_CUBE
  5230. || m_type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY
  5231. ? m_numSides
  5232. : _numLayers
  5233. ;
  5234. if (needResolve)
  5235. {
  5236. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
  5237. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, true);
  5238. VkImageResolve resolve;
  5239. resolve.srcOffset.x = 0;
  5240. resolve.srcOffset.y = 0;
  5241. resolve.srcOffset.z = 0;
  5242. resolve.dstOffset.x = 0;
  5243. resolve.dstOffset.y = 0;
  5244. resolve.dstOffset.z = 0;
  5245. resolve.srcSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5246. resolve.srcSubresource.mipLevel = _mip;
  5247. resolve.srcSubresource.baseArrayLayer = _layer;
  5248. resolve.srcSubresource.layerCount = numLayers;
  5249. resolve.dstSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5250. resolve.dstSubresource.mipLevel = _mip;
  5251. resolve.dstSubresource.baseArrayLayer = _layer;
  5252. resolve.dstSubresource.layerCount = numLayers;
  5253. resolve.extent.width = m_width;
  5254. resolve.extent.height = m_height;
  5255. resolve.extent.depth = 1;
  5256. vkCmdResolveImage(
  5257. _commandBuffer
  5258. , m_textureImage
  5259. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5260. , m_singleMsaaImage
  5261. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5262. , 1
  5263. , &resolve
  5264. );
  5265. }
  5266. if (needMipGen)
  5267. {
  5268. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  5269. int32_t mipWidth = bx::max<int32_t>(int32_t(m_width) >> _mip, 1);
  5270. int32_t mipHeight = bx::max<int32_t>(int32_t(m_height) >> _mip, 1);
  5271. const VkFilter filter = bimg::isDepth(bimg::TextureFormat::Enum(m_textureFormat) )
  5272. ? VK_FILTER_NEAREST
  5273. : VK_FILTER_LINEAR
  5274. ;
  5275. VkImageBlit blit;
  5276. blit.srcOffsets[0] = { 0, 0, 0 };
  5277. blit.srcOffsets[1] = { mipWidth, mipHeight, 1 };
  5278. blit.srcSubresource.aspectMask = m_aspectMask;
  5279. blit.srcSubresource.mipLevel = 0;
  5280. blit.srcSubresource.baseArrayLayer = _layer;
  5281. blit.srcSubresource.layerCount = numLayers;
  5282. blit.dstOffsets[0] = { 0, 0, 0 };
  5283. blit.dstOffsets[1] = { mipWidth, mipHeight, 1 };
  5284. blit.dstSubresource.aspectMask = m_aspectMask;
  5285. blit.dstSubresource.mipLevel = 0;
  5286. blit.dstSubresource.baseArrayLayer = _layer;
  5287. blit.dstSubresource.layerCount = numLayers;
  5288. for (uint32_t i = _mip + 1; i < m_numMips; i++)
  5289. {
  5290. BGFX_PROFILER_SCOPE("mipmap", kColorResource);
  5291. blit.srcOffsets[1] = { mipWidth, mipHeight, 1 };
  5292. blit.srcSubresource.mipLevel = i - 1;
  5293. mipWidth = bx::uint32_max(mipWidth >> 1, 1);
  5294. mipHeight = bx::uint32_max(mipHeight >> 1, 1);
  5295. blit.dstOffsets[1] = { mipWidth, mipHeight, 1 };
  5296. blit.dstSubresource.mipLevel = i;
  5297. vk::setImageMemoryBarrier(
  5298. _commandBuffer
  5299. , m_textureImage
  5300. , m_aspectMask
  5301. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5302. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5303. , blit.srcSubresource.mipLevel
  5304. , 1
  5305. , _layer
  5306. , numLayers
  5307. );
  5308. vkCmdBlitImage(
  5309. _commandBuffer
  5310. , m_textureImage
  5311. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5312. , m_textureImage
  5313. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5314. , 1
  5315. , &blit
  5316. , filter
  5317. );
  5318. }
  5319. vk::setImageMemoryBarrier(
  5320. _commandBuffer
  5321. , m_textureImage
  5322. , m_aspectMask
  5323. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5324. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5325. , _mip
  5326. , m_numMips - _mip - 1
  5327. , _layer
  5328. , numLayers
  5329. );
  5330. }
  5331. setImageMemoryBarrier(_commandBuffer, oldLayout);
  5332. setImageMemoryBarrier(_commandBuffer, oldSingleMsaaLayout, true);
  5333. }
  5334. void TextureVK::copyBufferToTexture(VkCommandBuffer _commandBuffer, VkBuffer _stagingBuffer, uint32_t _bufferImageCopyCount, VkBufferImageCopy* _bufferImageCopy)
  5335. {
  5336. BGFX_PROFILER_SCOPE("TextureVK::copyBufferToTexture", kColorResource);
  5337. const VkImageLayout oldLayout = m_currentImageLayout == VK_IMAGE_LAYOUT_UNDEFINED
  5338. ? m_sampledLayout
  5339. : m_currentImageLayout
  5340. ;
  5341. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  5342. vkCmdCopyBufferToImage(
  5343. _commandBuffer
  5344. , _stagingBuffer
  5345. , m_textureImage
  5346. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5347. , _bufferImageCopyCount
  5348. , _bufferImageCopy
  5349. );
  5350. setImageMemoryBarrier(_commandBuffer, oldLayout);
  5351. }
  5352. VkImageLayout TextureVK::setImageMemoryBarrier(VkCommandBuffer _commandBuffer, VkImageLayout _newImageLayout, bool _singleMsaaImage)
  5353. {
  5354. if (_singleMsaaImage && VK_NULL_HANDLE == m_singleMsaaImage)
  5355. {
  5356. return VK_IMAGE_LAYOUT_UNDEFINED;
  5357. }
  5358. VkImageLayout& currentLayout = _singleMsaaImage
  5359. ? m_currentSingleMsaaImageLayout
  5360. : m_currentImageLayout
  5361. ;
  5362. const VkImageLayout oldLayout = currentLayout;
  5363. if (currentLayout == _newImageLayout)
  5364. {
  5365. return oldLayout;
  5366. }
  5367. const VkImage image = _singleMsaaImage
  5368. ? m_singleMsaaImage
  5369. : m_textureImage
  5370. ;
  5371. vk::setImageMemoryBarrier(
  5372. _commandBuffer
  5373. , image
  5374. , m_aspectMask
  5375. , currentLayout
  5376. , _newImageLayout
  5377. );
  5378. currentLayout = _newImageLayout;
  5379. return oldLayout;
  5380. }
  5381. VkResult TextureVK::createView(uint32_t _layer, uint32_t _numLayers, uint32_t _mip, uint32_t _numMips, VkImageViewType _type, VkImageAspectFlags _aspectMask, bool _renderTarget, ::VkImageView* _view) const
  5382. {
  5383. VkResult result = VK_SUCCESS;
  5384. if (VK_IMAGE_VIEW_TYPE_3D == m_type)
  5385. {
  5386. BX_ASSERT(false
  5387. || !_renderTarget
  5388. || !(m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  5389. , "3D image can't be a depth attachment"
  5390. );
  5391. }
  5392. if (VK_IMAGE_VIEW_TYPE_CUBE == _type
  5393. || VK_IMAGE_VIEW_TYPE_CUBE_ARRAY == _type)
  5394. {
  5395. BX_ASSERT(_numLayers % 6 == 0, "");
  5396. BX_ASSERT(
  5397. VK_IMAGE_VIEW_TYPE_3D != m_type
  5398. , "3D image can't be aliased as a cube texture"
  5399. );
  5400. }
  5401. VkImageViewCreateInfo viewInfo;
  5402. viewInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
  5403. viewInfo.pNext = NULL;
  5404. viewInfo.flags = 0;
  5405. viewInfo.image = ((VK_NULL_HANDLE != m_singleMsaaImage) && !_renderTarget)
  5406. ? m_singleMsaaImage
  5407. : m_textureImage
  5408. ;
  5409. viewInfo.viewType = _type;
  5410. viewInfo.format = m_format;
  5411. viewInfo.components = m_components;
  5412. viewInfo.subresourceRange.aspectMask = m_aspectMask & _aspectMask;
  5413. viewInfo.subresourceRange.baseMipLevel = _mip;
  5414. viewInfo.subresourceRange.levelCount = _numMips;
  5415. viewInfo.subresourceRange.baseArrayLayer = _layer;
  5416. viewInfo.subresourceRange.layerCount = 1;
  5417. if (VK_IMAGE_VIEW_TYPE_2D != _type
  5418. && VK_IMAGE_VIEW_TYPE_3D != _type)
  5419. {
  5420. viewInfo.subresourceRange.layerCount = VK_IMAGE_VIEW_TYPE_CUBE == _type
  5421. ? 6
  5422. : _numLayers
  5423. ;
  5424. }
  5425. VkImageView view = VK_NULL_HANDLE;
  5426. result = vkCreateImageView(
  5427. s_renderVK->m_device
  5428. , &viewInfo
  5429. , s_renderVK->m_allocatorCb
  5430. , &view
  5431. );
  5432. if (VK_SUCCESS != result)
  5433. {
  5434. BX_TRACE("Create texture view error: vkCreateImageView failed %d: %s.", result, getName(result) );
  5435. return result;
  5436. }
  5437. *_view = view;
  5438. return result;
  5439. }
  5440. VkImageAspectFlags TextureVK::getAspectMask(VkFormat _format)
  5441. {
  5442. switch (_format)
  5443. {
  5444. case VK_FORMAT_S8_UINT:
  5445. return VK_IMAGE_ASPECT_STENCIL_BIT;
  5446. break;
  5447. case VK_FORMAT_D16_UNORM:
  5448. case VK_FORMAT_X8_D24_UNORM_PACK32:
  5449. case VK_FORMAT_D32_SFLOAT:
  5450. return VK_IMAGE_ASPECT_DEPTH_BIT;
  5451. case VK_FORMAT_D16_UNORM_S8_UINT:
  5452. case VK_FORMAT_D24_UNORM_S8_UINT:
  5453. case VK_FORMAT_D32_SFLOAT_S8_UINT:
  5454. return VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT;
  5455. default:
  5456. return VK_IMAGE_ASPECT_COLOR_BIT;
  5457. }
  5458. }
  5459. VkResult SwapChainVK::create(VkCommandBuffer _commandBuffer, void* _nwh, const Resolution& _resolution, TextureFormat::Enum _depthFormat)
  5460. {
  5461. struct ErrorState
  5462. {
  5463. enum Enum
  5464. {
  5465. Default,
  5466. SurfaceCreated,
  5467. SwapChainCreated,
  5468. AttachmentsCreated
  5469. };
  5470. };
  5471. ErrorState::Enum errorState = ErrorState::Default;
  5472. VkResult result = VK_SUCCESS;
  5473. if (NULL == _nwh)
  5474. {
  5475. return result;
  5476. }
  5477. m_nwh = _nwh;
  5478. m_resolution = _resolution;
  5479. m_depthFormat = TextureFormat::Count == _depthFormat ? TextureFormat::D24S8 : _depthFormat;
  5480. m_queue = s_renderVK->m_globalQueue;
  5481. result = createSurface();
  5482. if (VK_SUCCESS != result)
  5483. {
  5484. BX_TRACE("Create swap chain error: creating surface failed %d: %s.", result, getName(result) );
  5485. goto error;
  5486. }
  5487. errorState = ErrorState::SurfaceCreated;
  5488. {
  5489. m_sci.sType = VK_STRUCTURE_TYPE_SWAPCHAIN_CREATE_INFO_KHR;
  5490. m_sci.pNext = NULL;
  5491. m_sci.flags = 0;
  5492. m_sci.imageArrayLayers = 1;
  5493. m_sci.imageSharingMode = VK_SHARING_MODE_EXCLUSIVE;
  5494. m_sci.queueFamilyIndexCount = 0;
  5495. m_sci.pQueueFamilyIndices = NULL;
  5496. m_sci.preTransform = VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR;
  5497. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5498. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  5499. {
  5500. m_backBufferColorImage[ii] = VK_NULL_HANDLE;
  5501. m_backBufferColorImageView[ii] = VK_NULL_HANDLE;
  5502. m_backBufferFrameBuffer[ii] = VK_NULL_HANDLE;
  5503. m_backBufferFence[ii] = VK_NULL_HANDLE;
  5504. m_presentDoneSemaphore[ii] = VK_NULL_HANDLE;
  5505. m_renderDoneSemaphore[ii] = VK_NULL_HANDLE;
  5506. }
  5507. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  5508. m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  5509. result = createSwapChain();
  5510. if (VK_SUCCESS != result)
  5511. {
  5512. BX_TRACE("Create swap chain error: creating swapchain and image views failed %d: %s", result, getName(result) );
  5513. goto error;
  5514. }
  5515. }
  5516. errorState = ErrorState::SwapChainCreated;
  5517. {
  5518. result = createAttachments(_commandBuffer);
  5519. if (VK_SUCCESS != result)
  5520. {
  5521. BX_TRACE("Create swap chain error: creating MSAA/depth attachments failed %d: %s.", result, getName(result) );
  5522. goto error;
  5523. }
  5524. }
  5525. errorState = ErrorState::AttachmentsCreated;
  5526. {
  5527. result = createFrameBuffer();
  5528. if (VK_SUCCESS != result)
  5529. {
  5530. BX_TRACE("Create swap chain error: creating frame buffers failed %d: %s.", result, getName(result) );
  5531. goto error;
  5532. }
  5533. }
  5534. return VK_SUCCESS;
  5535. error:
  5536. BX_TRACE("errorState %d", errorState);
  5537. switch (errorState)
  5538. {
  5539. case ErrorState::AttachmentsCreated:
  5540. releaseAttachments();
  5541. [[fallthrough]];
  5542. case ErrorState::SwapChainCreated:
  5543. releaseSwapChain();
  5544. [[fallthrough]];
  5545. case ErrorState::SurfaceCreated:
  5546. releaseSurface();
  5547. [[fallthrough]];
  5548. case ErrorState::Default:
  5549. break;
  5550. };
  5551. return VK_SUCCESS != result
  5552. ? result
  5553. : VK_ERROR_INITIALIZATION_FAILED
  5554. ;
  5555. }
  5556. void SwapChainVK::destroy()
  5557. {
  5558. if (VK_NULL_HANDLE != m_swapchain)
  5559. {
  5560. releaseFrameBuffer();
  5561. releaseAttachments();
  5562. releaseSwapChain();
  5563. releaseSurface();
  5564. // can't delay-delete the surface, since there can only be one swapchain per surface
  5565. // new framebuffer with the same window would get an error at swapchain creation
  5566. s_renderVK->kick(true);
  5567. }
  5568. m_nwh = NULL;
  5569. }
  5570. void SwapChainVK::update(VkCommandBuffer _commandBuffer, void* _nwh, const Resolution& _resolution)
  5571. {
  5572. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5573. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  5574. m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  5575. const uint64_t recreateSurfaceMask = BGFX_RESET_HIDPI;
  5576. const uint64_t recreateSwapchainMask = BGFX_RESET_VSYNC | BGFX_RESET_SRGB_BACKBUFFER;
  5577. const uint64_t recreateAttachmentsMask = BGFX_RESET_MSAA_MASK;
  5578. const bool recreateSurface = false
  5579. || m_needToRecreateSurface
  5580. || m_nwh != _nwh
  5581. || (m_resolution.reset & recreateSurfaceMask) != (_resolution.reset & recreateSurfaceMask)
  5582. ;
  5583. const bool recreateSwapchain = false
  5584. || m_resolution.format != _resolution.format
  5585. || m_resolution.width != _resolution.width
  5586. || m_resolution.height != _resolution.height
  5587. || (m_resolution.reset & recreateSwapchainMask) != (_resolution.reset & recreateSwapchainMask)
  5588. || recreateSurface
  5589. ;
  5590. const bool recreateAttachments = false
  5591. || (m_resolution.reset & recreateAttachmentsMask) != (_resolution.reset & recreateAttachmentsMask)
  5592. || recreateSwapchain
  5593. ;
  5594. m_nwh = _nwh;
  5595. m_resolution = _resolution;
  5596. if (recreateAttachments)
  5597. {
  5598. releaseFrameBuffer();
  5599. releaseAttachments();
  5600. if (recreateSwapchain)
  5601. {
  5602. releaseSwapChain();
  5603. if (recreateSurface)
  5604. {
  5605. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5606. releaseSurface();
  5607. s_renderVK->kick(true);
  5608. _commandBuffer = s_renderVK->m_commandBuffer;
  5609. VkResult result = createSurface();
  5610. if (VK_SUCCESS != result)
  5611. {
  5612. BX_TRACE("Surface lost.");
  5613. return;
  5614. }
  5615. }
  5616. VkSurfaceCapabilitiesKHR surfaceCapabilities;
  5617. VK_CHECK(vkGetPhysicalDeviceSurfaceCapabilitiesKHR(physicalDevice, m_surface, &surfaceCapabilities) );
  5618. const uint32_t width = bx::clamp<uint32_t>(
  5619. m_resolution.width
  5620. , surfaceCapabilities.minImageExtent.width
  5621. , surfaceCapabilities.maxImageExtent.width
  5622. );
  5623. const uint32_t height = bx::clamp<uint32_t>(
  5624. m_resolution.height
  5625. , surfaceCapabilities.minImageExtent.height
  5626. , surfaceCapabilities.maxImageExtent.height
  5627. );
  5628. // swapchain can't have size 0
  5629. // on some platforms this happens when minimized
  5630. if (width == 0
  5631. || height == 0)
  5632. {
  5633. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5634. s_renderVK->kick(true);
  5635. return;
  5636. }
  5637. VK_CHECK(createSwapChain() );
  5638. }
  5639. VK_CHECK(createAttachments(_commandBuffer) );
  5640. VK_CHECK(createFrameBuffer() );
  5641. }
  5642. }
  5643. VkResult SwapChainVK::createSurface()
  5644. {
  5645. VkResult result = VK_ERROR_INITIALIZATION_FAILED;
  5646. const VkInstance instance = s_renderVK->m_instance;
  5647. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  5648. #if BX_PLATFORM_WINDOWS
  5649. {
  5650. if (NULL != vkCreateWin32SurfaceKHR)
  5651. {
  5652. VkWin32SurfaceCreateInfoKHR sci;
  5653. sci.sType = VK_STRUCTURE_TYPE_WIN32_SURFACE_CREATE_INFO_KHR;
  5654. sci.pNext = NULL;
  5655. sci.flags = 0;
  5656. sci.hinstance = (HINSTANCE)GetModuleHandle(NULL);
  5657. sci.hwnd = (HWND)m_nwh;
  5658. result = vkCreateWin32SurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5659. }
  5660. }
  5661. #elif BX_PLATFORM_ANDROID
  5662. {
  5663. if (NULL != vkCreateAndroidSurfaceKHR)
  5664. {
  5665. VkAndroidSurfaceCreateInfoKHR sci;
  5666. sci.sType = VK_STRUCTURE_TYPE_ANDROID_SURFACE_CREATE_INFO_KHR;
  5667. sci.pNext = NULL;
  5668. sci.flags = 0;
  5669. sci.window = (ANativeWindow*)m_nwh;
  5670. result = vkCreateAndroidSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5671. }
  5672. }
  5673. #elif BX_PLATFORM_LINUX
  5674. {
  5675. #if defined(WL_EGL_PLATFORM)
  5676. if (g_platformData.type == bgfx::NativeWindowHandleType::Wayland)
  5677. {
  5678. VkWaylandSurfaceCreateInfoKHR sci;
  5679. sci.sType = VK_STRUCTURE_TYPE_WAYLAND_SURFACE_CREATE_INFO_KHR;
  5680. sci.pNext = NULL;
  5681. sci.flags = 0;
  5682. sci.display = (wl_display*)g_platformData.ndt;
  5683. sci.surface = (wl_surface*)((wl_egl_window*)m_nwh)->surface;
  5684. result = vkCreateWaylandSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5685. }
  5686. else
  5687. #endif // defined(WL_EGL_PLATFORM)
  5688. {
  5689. if (NULL != vkCreateXlibSurfaceKHR)
  5690. {
  5691. VkXlibSurfaceCreateInfoKHR sci;
  5692. sci.sType = VK_STRUCTURE_TYPE_XLIB_SURFACE_CREATE_INFO_KHR;
  5693. sci.pNext = NULL;
  5694. sci.flags = 0;
  5695. sci.dpy = (Display*)g_platformData.ndt;
  5696. sci.window = (Window)m_nwh;
  5697. result = vkCreateXlibSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5698. }
  5699. if (VK_SUCCESS != result)
  5700. {
  5701. void* xcbdll = bx::dlopen("libX11-xcb.so.1");
  5702. if (NULL != xcbdll
  5703. && NULL != vkCreateXcbSurfaceKHR)
  5704. {
  5705. typedef xcb_connection_t* (*PFN_XGETXCBCONNECTION)(Display*);
  5706. PFN_XGETXCBCONNECTION XGetXCBConnection = (PFN_XGETXCBCONNECTION)bx::dlsym(xcbdll, "XGetXCBConnection");
  5707. union { void* ptr; xcb_window_t window; } cast = { m_nwh };
  5708. VkXcbSurfaceCreateInfoKHR sci;
  5709. sci.sType = VK_STRUCTURE_TYPE_XCB_SURFACE_CREATE_INFO_KHR;
  5710. sci.pNext = NULL;
  5711. sci.flags = 0;
  5712. sci.connection = XGetXCBConnection( (Display*)g_platformData.ndt);
  5713. sci.window = cast.window;
  5714. result = vkCreateXcbSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5715. bx::dlclose(xcbdll);
  5716. }
  5717. }
  5718. }
  5719. }
  5720. #elif BX_PLATFORM_OSX
  5721. {
  5722. if (NULL != vkCreateMacOSSurfaceMVK)
  5723. {
  5724. NSWindow* window = (NSWindow*)(m_nwh);
  5725. CAMetalLayer* layer = (CAMetalLayer*)(m_nwh);
  5726. if ([window isKindOfClass:[NSWindow class]])
  5727. {
  5728. NSView *contentView = (NSView *)window.contentView;
  5729. layer = [CAMetalLayer layer];
  5730. [contentView setWantsLayer : YES];
  5731. [contentView setLayer : layer];
  5732. }
  5733. else if ([layer isKindOfClass:[CAMetalLayer class]])
  5734. {
  5735. NSView *contentView = (NSView *)layer.delegate;
  5736. window = contentView.window;
  5737. }
  5738. else
  5739. {
  5740. BX_WARN(0, "Unable to create MoltenVk surface. Please set platform data window to an NSWindow or CAMetalLayer");
  5741. return result;
  5742. }
  5743. if (m_resolution.reset & BGFX_RESET_HIDPI)
  5744. {
  5745. layer.contentsScale = [window backingScaleFactor];
  5746. }
  5747. VkMacOSSurfaceCreateInfoMVK sci;
  5748. sci.sType = VK_STRUCTURE_TYPE_MACOS_SURFACE_CREATE_INFO_MVK;
  5749. sci.pNext = NULL;
  5750. sci.flags = 0;
  5751. sci.pView = (__bridge void*)layer;
  5752. result = vkCreateMacOSSurfaceMVK(instance, &sci, allocatorCb, &m_surface);
  5753. }
  5754. }
  5755. #else
  5756. # error "Figure out KHR surface..."
  5757. #endif // BX_PLATFORM_
  5758. m_needToRecreateSurface = false;
  5759. if (VK_SUCCESS != result)
  5760. {
  5761. BX_TRACE("Create surface error: vkCreate[Platform]SurfaceKHR failed %d: %s.", result, getName(result) );
  5762. return result;
  5763. }
  5764. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5765. const uint32_t queueFamily = s_renderVK->m_globalQueueFamily;
  5766. VkBool32 surfaceSupported;
  5767. result = vkGetPhysicalDeviceSurfaceSupportKHR(physicalDevice, queueFamily, m_surface, &surfaceSupported);
  5768. if (VK_SUCCESS != result
  5769. || !surfaceSupported)
  5770. {
  5771. BX_TRACE("Create surface error: Presentation to the given surface not supported.");
  5772. return VK_ERROR_INITIALIZATION_FAILED;
  5773. }
  5774. return result;
  5775. }
  5776. void SwapChainVK::releaseSurface()
  5777. {
  5778. release(m_surface);
  5779. }
  5780. VkResult SwapChainVK::createSwapChain()
  5781. {
  5782. VkResult result = VK_SUCCESS;
  5783. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5784. const VkDevice device = s_renderVK->m_device;
  5785. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  5786. VkSurfaceCapabilitiesKHR surfaceCapabilities;
  5787. result = vkGetPhysicalDeviceSurfaceCapabilitiesKHR(physicalDevice, m_surface, &surfaceCapabilities);
  5788. if (VK_SUCCESS != result)
  5789. {
  5790. BX_TRACE("Create swapchain error: vkGetPhysicalDeviceSurfaceCapabilitiesKHR failed %d: %s.", result, getName(result) );
  5791. return result;
  5792. }
  5793. const uint32_t minSwapBufferCount = bx::max<uint32_t>(surfaceCapabilities.minImageCount, 2);
  5794. const uint32_t maxSwapBufferCount = surfaceCapabilities.maxImageCount == 0
  5795. ? kMaxBackBuffers
  5796. : bx::min<uint32_t>(surfaceCapabilities.maxImageCount, kMaxBackBuffers)
  5797. ;
  5798. if (minSwapBufferCount > maxSwapBufferCount)
  5799. {
  5800. BX_TRACE("Create swapchain error: Incompatible swapchain image count (min: %d, max: %d, MaxBackBuffers: %d)."
  5801. , minSwapBufferCount
  5802. , maxSwapBufferCount
  5803. , kMaxBackBuffers
  5804. );
  5805. return VK_ERROR_INITIALIZATION_FAILED;
  5806. }
  5807. const uint32_t swapBufferCount = bx::clamp<uint32_t>(m_resolution.numBackBuffers, minSwapBufferCount, maxSwapBufferCount);
  5808. const VkColorSpaceKHR surfaceColorSpace = VK_COLOR_SPACE_SRGB_NONLINEAR_KHR;
  5809. const bool srgb = !!(m_resolution.reset & BGFX_RESET_SRGB_BACKBUFFER);
  5810. m_colorFormat = findSurfaceFormat(m_resolution.format, surfaceColorSpace, srgb);
  5811. if (TextureFormat::Count == m_colorFormat)
  5812. {
  5813. BX_TRACE("Create swapchain error: Unable to find surface format (srgb: %d).", srgb);
  5814. return VK_ERROR_INITIALIZATION_FAILED;
  5815. }
  5816. const VkFormat surfaceFormat = srgb
  5817. ? s_textureFormat[m_colorFormat].m_fmtSrgb
  5818. : s_textureFormat[m_colorFormat].m_fmt
  5819. ;
  5820. const uint32_t width = bx::clamp<uint32_t>(
  5821. m_resolution.width
  5822. , surfaceCapabilities.minImageExtent.width
  5823. , surfaceCapabilities.maxImageExtent.width
  5824. );
  5825. const uint32_t height = bx::clamp<uint32_t>(
  5826. m_resolution.height
  5827. , surfaceCapabilities.minImageExtent.height
  5828. , surfaceCapabilities.maxImageExtent.height
  5829. );
  5830. VkCompositeAlphaFlagBitsKHR compositeAlpha = VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR;
  5831. if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_INHERIT_BIT_KHR)
  5832. {
  5833. compositeAlpha = VK_COMPOSITE_ALPHA_INHERIT_BIT_KHR;
  5834. }
  5835. else if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_PRE_MULTIPLIED_BIT_KHR)
  5836. {
  5837. compositeAlpha = VK_COMPOSITE_ALPHA_PRE_MULTIPLIED_BIT_KHR;
  5838. }
  5839. else if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_POST_MULTIPLIED_BIT_KHR)
  5840. {
  5841. compositeAlpha = VK_COMPOSITE_ALPHA_POST_MULTIPLIED_BIT_KHR;
  5842. }
  5843. const VkImageUsageFlags imageUsageMask = 0
  5844. | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT
  5845. | VK_IMAGE_USAGE_TRANSFER_SRC_BIT
  5846. | VK_IMAGE_USAGE_TRANSFER_DST_BIT
  5847. ;
  5848. const VkImageUsageFlags imageUsage = surfaceCapabilities.supportedUsageFlags & imageUsageMask;
  5849. m_supportsReadback = 0 != (imageUsage & VK_IMAGE_USAGE_TRANSFER_SRC_BIT);
  5850. m_supportsManualResolve = 0 != (imageUsage & VK_IMAGE_USAGE_TRANSFER_DST_BIT);
  5851. const bool vsync = !!(m_resolution.reset & BGFX_RESET_VSYNC);
  5852. uint32_t presentModeIdx = findPresentMode(vsync);
  5853. if (UINT32_MAX == presentModeIdx)
  5854. {
  5855. BX_TRACE("Create swapchain error: Unable to find present mode (vsync: %d).", vsync);
  5856. return VK_ERROR_INITIALIZATION_FAILED;
  5857. }
  5858. m_sci.surface = m_surface;
  5859. m_sci.minImageCount = swapBufferCount;
  5860. m_sci.imageFormat = surfaceFormat;
  5861. m_sci.imageColorSpace = surfaceColorSpace;
  5862. m_sci.imageExtent.width = width;
  5863. m_sci.imageExtent.height = height;
  5864. m_sci.imageUsage = imageUsage;
  5865. m_sci.compositeAlpha = compositeAlpha;
  5866. m_sci.presentMode = s_presentMode[presentModeIdx].mode;
  5867. m_sci.clipped = VK_FALSE;
  5868. result = vkCreateSwapchainKHR(device, &m_sci, allocatorCb, &m_swapchain);
  5869. if (VK_SUCCESS != result)
  5870. {
  5871. BX_TRACE("Create swapchain error: vkCreateSwapchainKHR failed %d: %s.", result, getName(result) );
  5872. return result;
  5873. }
  5874. m_sci.oldSwapchain = m_swapchain;
  5875. result = vkGetSwapchainImagesKHR(device, m_swapchain, &m_numSwapchainImages, NULL);
  5876. if (VK_SUCCESS != result)
  5877. {
  5878. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR failed %d: %s.", result, getName(result) );
  5879. return result;
  5880. }
  5881. if (m_numSwapchainImages < m_sci.minImageCount)
  5882. {
  5883. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR: numSwapchainImages %d < minImageCount %d."
  5884. , m_numSwapchainImages
  5885. , m_sci.minImageCount
  5886. );
  5887. return VK_ERROR_INITIALIZATION_FAILED;
  5888. }
  5889. if (m_numSwapchainImages > BX_COUNTOF(m_backBufferColorImage) )
  5890. {
  5891. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR: numSwapchainImages %d > countof(m_backBufferColorImage) %d."
  5892. , m_numSwapchainImages
  5893. , BX_COUNTOF(m_backBufferColorImage)
  5894. );
  5895. return VK_ERROR_INITIALIZATION_FAILED;
  5896. }
  5897. result = vkGetSwapchainImagesKHR(device, m_swapchain, &m_numSwapchainImages, &m_backBufferColorImage[0]);
  5898. if (VK_SUCCESS != result && VK_INCOMPLETE != result)
  5899. {
  5900. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR failed %d: %s."
  5901. , result
  5902. , getName(result)
  5903. );
  5904. return result;
  5905. }
  5906. VkImageViewCreateInfo ivci;
  5907. ivci.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
  5908. ivci.pNext = NULL;
  5909. ivci.flags = 0;
  5910. ivci.viewType = VK_IMAGE_VIEW_TYPE_2D;
  5911. ivci.format = m_sci.imageFormat;
  5912. ivci.components.r = VK_COMPONENT_SWIZZLE_IDENTITY;
  5913. ivci.components.g = VK_COMPONENT_SWIZZLE_IDENTITY;
  5914. ivci.components.b = VK_COMPONENT_SWIZZLE_IDENTITY;
  5915. ivci.components.a = VK_COMPONENT_SWIZZLE_IDENTITY;
  5916. ivci.subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5917. ivci.subresourceRange.baseMipLevel = 0;
  5918. ivci.subresourceRange.levelCount = 1;
  5919. ivci.subresourceRange.baseArrayLayer = 0;
  5920. ivci.subresourceRange.layerCount = 1;
  5921. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  5922. {
  5923. ivci.image = m_backBufferColorImage[ii];
  5924. result = vkCreateImageView(device, &ivci, allocatorCb, &m_backBufferColorImageView[ii]);
  5925. if (VK_SUCCESS != result)
  5926. {
  5927. BX_TRACE("Create swapchain error: vkCreateImageView failed %d: %s.", result, getName(result) );
  5928. return result;
  5929. }
  5930. m_backBufferColorImageLayout[ii] = VK_IMAGE_LAYOUT_UNDEFINED;
  5931. }
  5932. VkSemaphoreCreateInfo sci;
  5933. sci.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  5934. sci.pNext = NULL;
  5935. sci.flags = 0;
  5936. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  5937. {
  5938. if (VK_SUCCESS != vkCreateSemaphore(device, &sci, allocatorCb, &m_presentDoneSemaphore[ii])
  5939. || VK_SUCCESS != vkCreateSemaphore(device, &sci, allocatorCb, &m_renderDoneSemaphore[ii]) )
  5940. {
  5941. BX_TRACE("Create swapchain error: vkCreateSemaphore failed %d: %s.", result, getName(result) );
  5942. return result;
  5943. }
  5944. }
  5945. m_backBufferColorIdx = 0;
  5946. m_currentSemaphore = 0;
  5947. m_needPresent = false;
  5948. m_needToRefreshSwapchain = false;
  5949. return result;
  5950. }
  5951. void SwapChainVK::releaseSwapChain()
  5952. {
  5953. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  5954. {
  5955. release(m_backBufferColorImageView[ii]);
  5956. m_backBufferFence[ii] = VK_NULL_HANDLE;
  5957. release(m_presentDoneSemaphore[ii]);
  5958. release(m_renderDoneSemaphore[ii]);
  5959. }
  5960. release(m_swapchain);
  5961. }
  5962. VkResult SwapChainVK::createAttachments(VkCommandBuffer _commandBuffer)
  5963. {
  5964. VkResult result = VK_SUCCESS;
  5965. const uint32_t samplerIndex = (m_resolution.reset & BGFX_RESET_MSAA_MASK) >> BGFX_RESET_MSAA_SHIFT;
  5966. const uint64_t textureFlags = (uint64_t(samplerIndex + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_RT | BGFX_TEXTURE_RT_WRITE_ONLY;
  5967. m_sampler = s_msaa[samplerIndex];
  5968. const uint16_t requiredCaps = m_sampler.Count > 1
  5969. ? BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER_MSAA
  5970. : BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER
  5971. ;
  5972. // the spec guarantees that at least one of D24S8 and D32FS8 is supported
  5973. VkFormat depthFormat = VK_FORMAT_D32_SFLOAT_S8_UINT;
  5974. if (g_caps.formats[m_depthFormat] & requiredCaps)
  5975. {
  5976. depthFormat = s_textureFormat[m_depthFormat].m_fmtDsv;
  5977. }
  5978. else if (g_caps.formats[TextureFormat::D24S8] & requiredCaps)
  5979. {
  5980. depthFormat = s_textureFormat[TextureFormat::D24S8].m_fmtDsv;
  5981. }
  5982. result = m_backBufferDepthStencil.create(
  5983. _commandBuffer
  5984. , m_sci.imageExtent.width
  5985. , m_sci.imageExtent.height
  5986. , textureFlags
  5987. , depthFormat
  5988. );
  5989. if (VK_SUCCESS != result)
  5990. {
  5991. BX_TRACE("Create swapchain error: creating depth stencil image failed %d: %s.", result, getName(result) );
  5992. return result;
  5993. }
  5994. result = m_backBufferDepthStencil.createView(0, 1, 0, 1, VK_IMAGE_VIEW_TYPE_2D, m_backBufferDepthStencil.m_aspectMask, true, &m_backBufferDepthStencilImageView);
  5995. if (VK_SUCCESS != result)
  5996. {
  5997. BX_TRACE("Create swapchain error: creating depth stencil image view failed %d: %s.", result, getName(result) );
  5998. return result;
  5999. }
  6000. if (m_sampler.Count > 1)
  6001. {
  6002. result = m_backBufferColorMsaa.create(
  6003. _commandBuffer
  6004. , m_sci.imageExtent.width
  6005. , m_sci.imageExtent.height
  6006. , textureFlags
  6007. , m_sci.imageFormat
  6008. );
  6009. if (VK_SUCCESS != result)
  6010. {
  6011. BX_TRACE("Create swapchain error: creating MSAA color image failed %d: %s.", result, getName(result) );
  6012. return result;
  6013. }
  6014. result = m_backBufferColorMsaa.createView(0, 1, 0, 1, VK_IMAGE_VIEW_TYPE_2D, m_backBufferColorMsaa.m_aspectMask, true, &m_backBufferColorMsaaImageView);
  6015. if (VK_SUCCESS != result)
  6016. {
  6017. BX_TRACE("Create swapchain error: creating MSAA color image view failed %d: %s.", result, getName(result) );
  6018. return result;
  6019. }
  6020. }
  6021. return result;
  6022. }
  6023. void SwapChainVK::releaseAttachments()
  6024. {
  6025. release(m_backBufferDepthStencilImageView);
  6026. release(m_backBufferColorMsaaImageView);
  6027. m_backBufferDepthStencil.destroy();
  6028. m_backBufferColorMsaa.destroy();
  6029. }
  6030. VkResult SwapChainVK::createFrameBuffer()
  6031. {
  6032. VkResult result = VK_SUCCESS;
  6033. const VkDevice device = s_renderVK->m_device;
  6034. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  6035. VkRenderPass renderPass;
  6036. result = s_renderVK->getRenderPass(*this, &renderPass);
  6037. if (VK_SUCCESS != result)
  6038. {
  6039. return result;
  6040. }
  6041. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  6042. {
  6043. uint32_t numAttachments = 2;
  6044. ::VkImageView attachments[3] =
  6045. {
  6046. m_sampler.Count > 1
  6047. ? m_backBufferColorMsaaImageView
  6048. : m_backBufferColorImageView[ii],
  6049. m_backBufferDepthStencilImageView,
  6050. };
  6051. if (m_sampler.Count > 1 && !m_supportsManualResolve)
  6052. {
  6053. attachments[numAttachments++] = m_backBufferColorImageView[ii];
  6054. }
  6055. VkFramebufferCreateInfo fci;
  6056. fci.sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO;
  6057. fci.pNext = NULL;
  6058. fci.flags = 0;
  6059. fci.renderPass = renderPass;
  6060. fci.attachmentCount = numAttachments;
  6061. fci.pAttachments = attachments;
  6062. fci.width = m_sci.imageExtent.width;
  6063. fci.height = m_sci.imageExtent.height;
  6064. fci.layers = 1;
  6065. result = vkCreateFramebuffer(device, &fci, allocatorCb, &m_backBufferFrameBuffer[ii]);
  6066. if (VK_SUCCESS != result)
  6067. {
  6068. return result;
  6069. }
  6070. }
  6071. return result;
  6072. }
  6073. void SwapChainVK::releaseFrameBuffer()
  6074. {
  6075. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  6076. {
  6077. release(m_backBufferFrameBuffer[ii]);
  6078. }
  6079. }
  6080. uint32_t SwapChainVK::findPresentMode(bool _vsync)
  6081. {
  6082. VkResult result = VK_SUCCESS;
  6083. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  6084. uint32_t numPresentModes;
  6085. result = vkGetPhysicalDeviceSurfacePresentModesKHR(
  6086. physicalDevice
  6087. , m_surface
  6088. , &numPresentModes
  6089. , NULL
  6090. );
  6091. if (VK_SUCCESS != result)
  6092. {
  6093. BX_TRACE("findPresentMode error: vkGetPhysicalDeviceSurfacePresentModesKHR failed %d: %s.", result, getName(result) );
  6094. return UINT32_MAX;
  6095. }
  6096. VkPresentModeKHR presentModes[16];
  6097. numPresentModes = bx::min<uint32_t>(numPresentModes, BX_COUNTOF(presentModes) );
  6098. result = vkGetPhysicalDeviceSurfacePresentModesKHR(
  6099. physicalDevice
  6100. , m_surface
  6101. , &numPresentModes
  6102. , presentModes
  6103. );
  6104. if (VK_SUCCESS != result)
  6105. {
  6106. BX_TRACE("findPresentMode error: vkGetPhysicalDeviceSurfacePresentModesKHR failed %d: %s.", result, getName(result) );
  6107. return UINT32_MAX;
  6108. }
  6109. uint32_t idx = UINT32_MAX;
  6110. for (uint32_t ii = 0; ii < BX_COUNTOF(s_presentMode) && UINT32_MAX == idx; ++ii)
  6111. {
  6112. for (uint32_t jj = 0; jj < numPresentModes; ++jj)
  6113. {
  6114. const PresentMode& pm = s_presentMode[ii];
  6115. if (pm.mode == presentModes[jj]
  6116. && pm.vsync == _vsync)
  6117. {
  6118. idx = ii;
  6119. break;
  6120. }
  6121. }
  6122. }
  6123. if (UINT32_MAX == idx)
  6124. {
  6125. idx = 0;
  6126. BX_TRACE("Present mode not found! Defaulting to %s.", s_presentMode[idx].name);
  6127. }
  6128. return idx;
  6129. }
  6130. TextureFormat::Enum SwapChainVK::findSurfaceFormat(TextureFormat::Enum _format, VkColorSpaceKHR _colorSpace, bool _srgb)
  6131. {
  6132. VkResult result = VK_SUCCESS;
  6133. TextureFormat::Enum selectedFormat = TextureFormat::Count;
  6134. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  6135. uint32_t numSurfaceFormats;
  6136. result = vkGetPhysicalDeviceSurfaceFormatsKHR(physicalDevice, m_surface, &numSurfaceFormats, NULL);
  6137. if (VK_SUCCESS != result)
  6138. {
  6139. BX_TRACE("findSurfaceFormat error: vkGetPhysicalDeviceSurfaceFormatsKHR failed %d: %s.", result, getName(result) );
  6140. return selectedFormat;
  6141. }
  6142. VkSurfaceFormatKHR* surfaceFormats = (VkSurfaceFormatKHR*)bx::alloc(g_allocator, numSurfaceFormats * sizeof(VkSurfaceFormatKHR) );
  6143. result = vkGetPhysicalDeviceSurfaceFormatsKHR(physicalDevice, m_surface, &numSurfaceFormats, surfaceFormats);
  6144. if (VK_SUCCESS != result)
  6145. {
  6146. BX_TRACE("findSurfaceFormat error: vkGetPhysicalDeviceSurfaceFormatsKHR failed %d: %s.", result, getName(result) );
  6147. bx::free(g_allocator, surfaceFormats);
  6148. return selectedFormat;
  6149. }
  6150. const TextureFormat::Enum requestedFormats[] =
  6151. {
  6152. _format,
  6153. TextureFormat::BGRA8,
  6154. TextureFormat::RGBA8,
  6155. };
  6156. for (uint32_t ii = 0; ii < BX_COUNTOF(requestedFormats) && TextureFormat::Count == selectedFormat; ii++)
  6157. {
  6158. const TextureFormat::Enum requested = requestedFormats[ii];
  6159. const VkFormat requestedVkFormat = _srgb
  6160. ? s_textureFormat[requested].m_fmtSrgb
  6161. : s_textureFormat[requested].m_fmt
  6162. ;
  6163. for (uint32_t jj = 0; jj < numSurfaceFormats; jj++)
  6164. {
  6165. if (_colorSpace == surfaceFormats[jj].colorSpace
  6166. && requestedVkFormat == surfaceFormats[jj].format)
  6167. {
  6168. selectedFormat = requested;
  6169. if (0 != ii
  6170. && s_renderVK->m_swapchainFormats[_format] != selectedFormat)
  6171. {
  6172. s_renderVK->m_swapchainFormats[_format] = selectedFormat;
  6173. BX_TRACE(
  6174. "findSurfaceFormat: Surface format %s not found! Defaulting to %s."
  6175. , bimg::getName(bimg::TextureFormat::Enum(_format) )
  6176. , bimg::getName(bimg::TextureFormat::Enum(selectedFormat) )
  6177. );
  6178. }
  6179. break;
  6180. }
  6181. }
  6182. }
  6183. bx::free(g_allocator, surfaceFormats);
  6184. if (TextureFormat::Count == selectedFormat)
  6185. {
  6186. BX_TRACE("findSurfaceFormat error: No supported surface format found.");
  6187. }
  6188. return selectedFormat;
  6189. }
  6190. bool SwapChainVK::acquire(VkCommandBuffer _commandBuffer)
  6191. {
  6192. if (VK_NULL_HANDLE == m_swapchain
  6193. || m_needToRefreshSwapchain)
  6194. {
  6195. return false;
  6196. }
  6197. if (!m_needPresent)
  6198. {
  6199. const VkDevice device = s_renderVK->m_device;
  6200. m_lastImageAcquiredSemaphore = m_presentDoneSemaphore[m_currentSemaphore];
  6201. m_lastImageRenderedSemaphore = m_renderDoneSemaphore[m_currentSemaphore];
  6202. m_currentSemaphore = (m_currentSemaphore + 1) % m_numSwapchainImages;
  6203. VkResult result;
  6204. {
  6205. BGFX_PROFILER_SCOPE("vkAcquireNextImageKHR", kColorFrame);
  6206. result = vkAcquireNextImageKHR(
  6207. device
  6208. , m_swapchain
  6209. , UINT64_MAX
  6210. , m_lastImageAcquiredSemaphore
  6211. , VK_NULL_HANDLE
  6212. , &m_backBufferColorIdx
  6213. );
  6214. }
  6215. switch (result)
  6216. {
  6217. case VK_SUCCESS:
  6218. break;
  6219. case VK_ERROR_SURFACE_LOST_KHR:
  6220. m_needToRecreateSurface = true;
  6221. [[fallthrough]];
  6222. case VK_ERROR_OUT_OF_DATE_KHR:
  6223. case VK_SUBOPTIMAL_KHR:
  6224. m_needToRefreshSwapchain = true;
  6225. return false;
  6226. default:
  6227. BX_ASSERT(VK_SUCCESS == result, "vkAcquireNextImageKHR(...); VK error 0x%x: %s", result, getName(result) );
  6228. return false;
  6229. }
  6230. if (VK_NULL_HANDLE != m_backBufferFence[m_backBufferColorIdx])
  6231. {
  6232. BGFX_PROFILER_SCOPE("vkWaitForFences", kColorFrame);
  6233. VK_CHECK(vkWaitForFences(
  6234. device
  6235. , 1
  6236. , &m_backBufferFence[m_backBufferColorIdx]
  6237. , VK_TRUE
  6238. , UINT64_MAX
  6239. ) );
  6240. }
  6241. transitionImage(_commandBuffer);
  6242. m_needPresent = true;
  6243. }
  6244. return true;
  6245. }
  6246. void SwapChainVK::present()
  6247. {
  6248. BGFX_PROFILER_SCOPE("SwapChainVk::present", kColorFrame);
  6249. if (VK_NULL_HANDLE != m_swapchain
  6250. && m_needPresent)
  6251. {
  6252. VkPresentInfoKHR pi;
  6253. pi.sType = VK_STRUCTURE_TYPE_PRESENT_INFO_KHR;
  6254. pi.pNext = NULL;
  6255. pi.waitSemaphoreCount = 1;
  6256. pi.pWaitSemaphores = &m_lastImageRenderedSemaphore;
  6257. pi.swapchainCount = 1;
  6258. pi.pSwapchains = &m_swapchain;
  6259. pi.pImageIndices = &m_backBufferColorIdx;
  6260. pi.pResults = NULL;
  6261. VkResult result;
  6262. {
  6263. BGFX_PROFILER_SCOPE("vkQueuePresentHKR", kColorFrame);
  6264. result = vkQueuePresentKHR(m_queue, &pi);
  6265. }
  6266. switch (result)
  6267. {
  6268. case VK_ERROR_SURFACE_LOST_KHR:
  6269. m_needToRecreateSurface = true;
  6270. [[fallthrough]];
  6271. case VK_ERROR_OUT_OF_DATE_KHR:
  6272. case VK_SUBOPTIMAL_KHR:
  6273. m_needToRefreshSwapchain = true;
  6274. break;
  6275. default:
  6276. BX_ASSERT(VK_SUCCESS == result, "vkQueuePresentKHR(...); VK error 0x%x: %s", result, getName(result) );
  6277. break;
  6278. }
  6279. m_needPresent = false;
  6280. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  6281. }
  6282. }
  6283. void SwapChainVK::transitionImage(VkCommandBuffer _commandBuffer)
  6284. {
  6285. VkImageLayout& layout = m_backBufferColorImageLayout[m_backBufferColorIdx];
  6286. const bool toPresent = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL == layout;
  6287. const VkImageLayout newLayout = toPresent
  6288. ? VK_IMAGE_LAYOUT_PRESENT_SRC_KHR
  6289. : VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  6290. ;
  6291. layout = toPresent ? layout : VK_IMAGE_LAYOUT_UNDEFINED;
  6292. setImageMemoryBarrier(
  6293. _commandBuffer
  6294. , m_backBufferColorImage[m_backBufferColorIdx]
  6295. , VK_IMAGE_ASPECT_COLOR_BIT
  6296. , layout
  6297. , newLayout
  6298. );
  6299. layout = newLayout;
  6300. }
  6301. void FrameBufferVK::create(uint8_t _num, const Attachment* _attachment)
  6302. {
  6303. m_numTh = _num;
  6304. bx::memCopy(m_attachment, _attachment, sizeof(Attachment) * _num);
  6305. postReset();
  6306. }
  6307. VkResult FrameBufferVK::create(uint16_t _denseIdx, void* _nwh, uint32_t _width, uint32_t _height, TextureFormat::Enum _format, TextureFormat::Enum _depthFormat)
  6308. {
  6309. VkResult result = VK_SUCCESS;
  6310. Resolution resolution = s_renderVK->m_resolution;
  6311. resolution.format = TextureFormat::Count == _format ? resolution.format : _format;
  6312. resolution.width = _width;
  6313. resolution.height = _height;
  6314. if (_denseIdx != UINT16_MAX)
  6315. {
  6316. resolution.reset &= ~BGFX_RESET_MSAA_MASK;
  6317. }
  6318. result = m_swapChain.create(s_renderVK->m_commandBuffer, _nwh, resolution, _depthFormat);
  6319. if (VK_SUCCESS != result)
  6320. {
  6321. return result;
  6322. }
  6323. result = s_renderVK->getRenderPass(m_swapChain, &m_renderPass);
  6324. if (VK_SUCCESS != result)
  6325. {
  6326. return result;
  6327. }
  6328. m_denseIdx = _denseIdx;
  6329. m_nwh = _nwh;
  6330. m_width = _width;
  6331. m_height = _height;
  6332. m_sampler = m_swapChain.m_sampler;
  6333. return result;
  6334. }
  6335. void FrameBufferVK::preReset()
  6336. {
  6337. if (VK_NULL_HANDLE != m_framebuffer)
  6338. {
  6339. s_renderVK->release(m_framebuffer);
  6340. for (uint8_t ii = 0; ii < m_numTh; ++ii)
  6341. {
  6342. s_renderVK->release(m_textureImageViews[ii]);
  6343. }
  6344. }
  6345. }
  6346. void FrameBufferVK::postReset()
  6347. {
  6348. if (m_numTh > 0)
  6349. {
  6350. const VkDevice device = s_renderVK->m_device;
  6351. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  6352. VK_CHECK(s_renderVK->getRenderPass(m_numTh, m_attachment, &m_renderPass) );
  6353. m_depth = BGFX_INVALID_HANDLE;
  6354. m_num = 0;
  6355. for (uint8_t ii = 0; ii < m_numTh; ++ii)
  6356. {
  6357. const Attachment& at = m_attachment[ii];
  6358. const TextureVK& texture = s_renderVK->m_textures[at.handle.idx];
  6359. VK_CHECK(texture.createView(
  6360. at.layer
  6361. , at.numLayers
  6362. , at.mip
  6363. , 1
  6364. , at.numLayers > 1 ? VK_IMAGE_VIEW_TYPE_2D_ARRAY : VK_IMAGE_VIEW_TYPE_2D
  6365. , texture.m_aspectMask
  6366. , true
  6367. , &m_textureImageViews[ii]
  6368. ) );
  6369. if (texture.m_aspectMask & VK_IMAGE_ASPECT_COLOR_BIT)
  6370. {
  6371. m_texture[m_num] = at.handle;
  6372. m_num++;
  6373. }
  6374. else if (texture.m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  6375. {
  6376. m_depth = at.handle;
  6377. }
  6378. }
  6379. const TextureVK& firstTexture = s_renderVK->m_textures[m_attachment[0].handle.idx];
  6380. m_width = bx::uint32_max(firstTexture.m_width >> m_attachment[0].mip, 1);
  6381. m_height = bx::uint32_max(firstTexture.m_height >> m_attachment[0].mip, 1);
  6382. m_sampler = firstTexture.m_sampler;
  6383. VkFramebufferCreateInfo fci;
  6384. fci.sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO;
  6385. fci.pNext = NULL;
  6386. fci.flags = 0;
  6387. fci.renderPass = m_renderPass;
  6388. fci.attachmentCount = m_numTh;
  6389. fci.pAttachments = &m_textureImageViews[0];
  6390. fci.width = m_width;
  6391. fci.height = m_height;
  6392. fci.layers = m_attachment[0].numLayers;
  6393. VK_CHECK(vkCreateFramebuffer(device, &fci, allocatorCb, &m_framebuffer) );
  6394. m_currentFramebuffer = m_framebuffer;
  6395. }
  6396. }
  6397. void FrameBufferVK::update(VkCommandBuffer _commandBuffer, const Resolution& _resolution)
  6398. {
  6399. m_swapChain.update(_commandBuffer, m_nwh, _resolution);
  6400. VK_CHECK(s_renderVK->getRenderPass(m_swapChain, &m_renderPass) );
  6401. m_width = _resolution.width;
  6402. m_height = _resolution.height;
  6403. m_sampler = m_swapChain.m_sampler;
  6404. }
  6405. void FrameBufferVK::resolve()
  6406. {
  6407. if (!m_needResolve)
  6408. {
  6409. return;
  6410. }
  6411. if (NULL == m_nwh)
  6412. {
  6413. for (uint32_t ii = 0; ii < m_numTh; ++ii)
  6414. {
  6415. const Attachment& at = m_attachment[ii];
  6416. if (isValid(at.handle) )
  6417. {
  6418. TextureVK& texture = s_renderVK->m_textures[at.handle.idx];
  6419. texture.resolve(s_renderVK->m_commandBuffer, at.resolve, at.layer, at.numLayers, at.mip);
  6420. }
  6421. }
  6422. }
  6423. else if (isRenderable()
  6424. && m_sampler.Count > 1
  6425. && m_swapChain.m_supportsManualResolve)
  6426. {
  6427. m_swapChain.m_backBufferColorMsaa.m_singleMsaaImage = m_swapChain.m_backBufferColorImage[m_swapChain.m_backBufferColorIdx];
  6428. m_swapChain.m_backBufferColorMsaa.m_currentSingleMsaaImageLayout = m_swapChain.m_backBufferColorImageLayout[m_swapChain.m_backBufferColorIdx];
  6429. m_swapChain.m_backBufferColorMsaa.resolve(s_renderVK->m_commandBuffer, 0, 0, 1, 0);
  6430. m_swapChain.m_backBufferColorMsaa.m_singleMsaaImage = VK_NULL_HANDLE;
  6431. m_swapChain.m_backBufferColorMsaa.m_currentSingleMsaaImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  6432. }
  6433. m_needResolve = false;
  6434. }
  6435. uint16_t FrameBufferVK::destroy()
  6436. {
  6437. preReset();
  6438. if (NULL != m_nwh)
  6439. {
  6440. m_swapChain.destroy();
  6441. m_nwh = NULL;
  6442. m_needPresent = false;
  6443. }
  6444. m_numTh = 0;
  6445. m_num = 0;
  6446. m_depth = BGFX_INVALID_HANDLE;
  6447. m_needResolve = false;
  6448. uint16_t denseIdx = m_denseIdx;
  6449. m_denseIdx = UINT16_MAX;
  6450. return denseIdx;
  6451. }
  6452. bool FrameBufferVK::acquire(VkCommandBuffer _commandBuffer)
  6453. {
  6454. bool acquired = true;
  6455. if (NULL != m_nwh)
  6456. {
  6457. acquired = m_swapChain.acquire(_commandBuffer);
  6458. m_needPresent = m_swapChain.m_needPresent;
  6459. m_currentFramebuffer = m_swapChain.m_backBufferFrameBuffer[m_swapChain.m_backBufferColorIdx];
  6460. }
  6461. m_needResolve = true;
  6462. return acquired;
  6463. }
  6464. void FrameBufferVK::present()
  6465. {
  6466. m_swapChain.present();
  6467. m_needPresent = false;
  6468. }
  6469. bool FrameBufferVK::isRenderable() const
  6470. {
  6471. return false
  6472. || (NULL == m_nwh)
  6473. || m_swapChain.m_needPresent
  6474. ;
  6475. }
  6476. VkResult CommandQueueVK::init(uint32_t _queueFamily, VkQueue _queue, uint32_t _numFramesInFlight)
  6477. {
  6478. m_queueFamily = _queueFamily;
  6479. m_queue = _queue;
  6480. m_numFramesInFlight = bx::clamp<uint32_t>(_numFramesInFlight, 1, BGFX_CONFIG_MAX_FRAME_LATENCY);
  6481. m_activeCommandBuffer = VK_NULL_HANDLE;
  6482. return reset();
  6483. }
  6484. VkResult CommandQueueVK::reset()
  6485. {
  6486. shutdown();
  6487. m_currentFrameInFlight = 0;
  6488. m_consumeIndex = 0;
  6489. m_numSignalSemaphores = 0;
  6490. m_numWaitSemaphores = 0;
  6491. m_activeCommandBuffer = VK_NULL_HANDLE;
  6492. m_currentFence = VK_NULL_HANDLE;
  6493. m_completedFence = VK_NULL_HANDLE;
  6494. m_submitted = 0;
  6495. VkCommandPoolCreateInfo cpci;
  6496. cpci.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  6497. cpci.pNext = NULL;
  6498. cpci.flags = VK_COMMAND_POOL_CREATE_TRANSIENT_BIT;
  6499. cpci.queueFamilyIndex = m_queueFamily;
  6500. VkCommandBufferAllocateInfo cbai;
  6501. cbai.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  6502. cbai.pNext = NULL;
  6503. cbai.level = VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  6504. cbai.commandBufferCount = 1;
  6505. VkFenceCreateInfo fci;
  6506. fci.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  6507. fci.pNext = NULL;
  6508. fci.flags = VK_FENCE_CREATE_SIGNALED_BIT;
  6509. VkResult result = VK_SUCCESS;
  6510. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6511. {
  6512. result = vkCreateCommandPool(
  6513. s_renderVK->m_device
  6514. , &cpci
  6515. , s_renderVK->m_allocatorCb
  6516. , &m_commandList[ii].m_commandPool
  6517. );
  6518. if (VK_SUCCESS != result)
  6519. {
  6520. BX_TRACE("Create command queue error: vkCreateCommandPool failed %d: %s.", result, getName(result) );
  6521. return result;
  6522. }
  6523. cbai.commandPool = m_commandList[ii].m_commandPool;
  6524. result = vkAllocateCommandBuffers(
  6525. s_renderVK->m_device
  6526. , &cbai
  6527. , &m_commandList[ii].m_commandBuffer
  6528. );
  6529. if (VK_SUCCESS != result)
  6530. {
  6531. BX_TRACE("Create command queue error: vkAllocateCommandBuffers failed %d: %s.", result, getName(result) );
  6532. return result;
  6533. }
  6534. result = vkCreateFence(
  6535. s_renderVK->m_device
  6536. , &fci
  6537. , s_renderVK->m_allocatorCb
  6538. , &m_commandList[ii].m_fence
  6539. );
  6540. if (VK_SUCCESS != result)
  6541. {
  6542. BX_TRACE("Create command queue error: vkCreateFence failed %d: %s.", result, getName(result) );
  6543. return result;
  6544. }
  6545. }
  6546. return result;
  6547. }
  6548. void CommandQueueVK::shutdown()
  6549. {
  6550. kick(true);
  6551. finish(true);
  6552. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6553. {
  6554. vkDestroy(m_commandList[ii].m_fence);
  6555. m_commandList[ii].m_commandBuffer = VK_NULL_HANDLE;
  6556. vkDestroy(m_commandList[ii].m_commandPool);
  6557. }
  6558. }
  6559. VkResult CommandQueueVK::alloc(VkCommandBuffer* _commandBuffer)
  6560. {
  6561. VkResult result = VK_SUCCESS;
  6562. if (m_activeCommandBuffer == VK_NULL_HANDLE)
  6563. {
  6564. const VkDevice device = s_renderVK->m_device;
  6565. CommandList& commandList = m_commandList[m_currentFrameInFlight];
  6566. {
  6567. BGFX_PROFILER_SCOPE("vkWaitForFences", kColorFrame);
  6568. result = vkWaitForFences(device, 1, &commandList.m_fence, VK_TRUE, UINT64_MAX);
  6569. }
  6570. if (VK_SUCCESS != result)
  6571. {
  6572. BX_TRACE("Allocate command buffer error: vkWaitForFences failed %d: %s.", result, getName(result) );
  6573. return result;
  6574. }
  6575. result = vkResetCommandPool(device, commandList.m_commandPool, 0);
  6576. if (VK_SUCCESS != result)
  6577. {
  6578. BX_TRACE("Allocate command buffer error: vkResetCommandPool failed %d: %s.", result, getName(result) );
  6579. return result;
  6580. }
  6581. VkCommandBufferBeginInfo cbi;
  6582. cbi.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  6583. cbi.pNext = NULL;
  6584. cbi.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  6585. cbi.pInheritanceInfo = NULL;
  6586. result = vkBeginCommandBuffer(commandList.m_commandBuffer, &cbi);
  6587. if (VK_SUCCESS != result)
  6588. {
  6589. BX_TRACE("Allocate command buffer error: vkBeginCommandBuffer failed %d: %s.", result, getName(result) );
  6590. return result;
  6591. }
  6592. m_activeCommandBuffer = commandList.m_commandBuffer;
  6593. m_currentFence = commandList.m_fence;
  6594. }
  6595. if (NULL != _commandBuffer)
  6596. {
  6597. *_commandBuffer = m_activeCommandBuffer;
  6598. }
  6599. return result;
  6600. }
  6601. void CommandQueueVK::addWaitSemaphore(VkSemaphore _semaphore, VkPipelineStageFlags _waitFlags)
  6602. {
  6603. BX_ASSERT(m_numWaitSemaphores < BX_COUNTOF(m_waitSemaphores), "Too many wait semaphores.");
  6604. m_waitSemaphores[m_numWaitSemaphores] = _semaphore;
  6605. m_waitSemaphoreStages[m_numWaitSemaphores] = _waitFlags;
  6606. m_numWaitSemaphores++;
  6607. }
  6608. void CommandQueueVK::addSignalSemaphore(VkSemaphore _semaphore)
  6609. {
  6610. BX_ASSERT(m_numSignalSemaphores < BX_COUNTOF(m_signalSemaphores), "Too many signal semaphores.");
  6611. m_signalSemaphores[m_numSignalSemaphores] = _semaphore;
  6612. m_numSignalSemaphores++;
  6613. }
  6614. void CommandQueueVK::kick(bool _wait)
  6615. {
  6616. if (VK_NULL_HANDLE != m_activeCommandBuffer)
  6617. {
  6618. const VkDevice device = s_renderVK->m_device;
  6619. setMemoryBarrier(
  6620. m_activeCommandBuffer
  6621. , VK_PIPELINE_STAGE_ALL_COMMANDS_BIT
  6622. , VK_PIPELINE_STAGE_ALL_COMMANDS_BIT
  6623. );
  6624. VK_CHECK(vkEndCommandBuffer(m_activeCommandBuffer) );
  6625. m_completedFence = m_currentFence;
  6626. m_currentFence = VK_NULL_HANDLE;
  6627. VK_CHECK(vkResetFences(device, 1, &m_completedFence) );
  6628. VkSubmitInfo si;
  6629. si.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO;
  6630. si.pNext = NULL;
  6631. si.waitSemaphoreCount = m_numWaitSemaphores;
  6632. si.pWaitSemaphores = &m_waitSemaphores[0];
  6633. si.pWaitDstStageMask = m_waitSemaphoreStages;
  6634. si.commandBufferCount = 1;
  6635. si.pCommandBuffers = &m_activeCommandBuffer;
  6636. si.signalSemaphoreCount = m_numSignalSemaphores;
  6637. si.pSignalSemaphores = &m_signalSemaphores[0];
  6638. m_numWaitSemaphores = 0;
  6639. m_numSignalSemaphores = 0;
  6640. VK_CHECK(vkQueueSubmit(m_queue, 1, &si, m_completedFence) );
  6641. if (_wait)
  6642. {
  6643. BGFX_PROFILER_SCOPE("vkWaitForFences", kColorFrame);
  6644. VK_CHECK(vkWaitForFences(device, 1, &m_completedFence, VK_TRUE, UINT64_MAX) );
  6645. }
  6646. m_activeCommandBuffer = VK_NULL_HANDLE;
  6647. m_currentFrameInFlight = (m_currentFrameInFlight + 1) % m_numFramesInFlight;
  6648. m_submitted++;
  6649. }
  6650. }
  6651. void CommandQueueVK::finish(bool _finishAll)
  6652. {
  6653. if (_finishAll)
  6654. {
  6655. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6656. {
  6657. consume();
  6658. }
  6659. m_consumeIndex = m_currentFrameInFlight;
  6660. }
  6661. else
  6662. {
  6663. consume();
  6664. }
  6665. }
  6666. void CommandQueueVK::release(uint64_t _handle, VkObjectType _type)
  6667. {
  6668. Resource resource;
  6669. resource.m_type = _type;
  6670. resource.m_handle = _handle;
  6671. m_release[m_currentFrameInFlight].push_back(resource);
  6672. }
  6673. void CommandQueueVK::consume()
  6674. {
  6675. m_consumeIndex = (m_consumeIndex + 1) % m_numFramesInFlight;
  6676. for (const Resource& resource : m_release[m_consumeIndex])
  6677. {
  6678. switch (resource.m_type)
  6679. {
  6680. case VK_OBJECT_TYPE_BUFFER: destroy<VkBuffer >(resource.m_handle); break;
  6681. case VK_OBJECT_TYPE_IMAGE_VIEW: destroy<VkImageView >(resource.m_handle); break;
  6682. case VK_OBJECT_TYPE_IMAGE: destroy<VkImage >(resource.m_handle); break;
  6683. case VK_OBJECT_TYPE_FRAMEBUFFER: destroy<VkFramebuffer >(resource.m_handle); break;
  6684. case VK_OBJECT_TYPE_PIPELINE_LAYOUT: destroy<VkPipelineLayout >(resource.m_handle); break;
  6685. case VK_OBJECT_TYPE_PIPELINE: destroy<VkPipeline >(resource.m_handle); break;
  6686. case VK_OBJECT_TYPE_DESCRIPTOR_SET: destroy<VkDescriptorSet >(resource.m_handle); break;
  6687. case VK_OBJECT_TYPE_DESCRIPTOR_SET_LAYOUT: destroy<VkDescriptorSetLayout>(resource.m_handle); break;
  6688. case VK_OBJECT_TYPE_RENDER_PASS: destroy<VkRenderPass >(resource.m_handle); break;
  6689. case VK_OBJECT_TYPE_SAMPLER: destroy<VkSampler >(resource.m_handle); break;
  6690. case VK_OBJECT_TYPE_SEMAPHORE: destroy<VkSemaphore >(resource.m_handle); break;
  6691. case VK_OBJECT_TYPE_SURFACE_KHR: destroy<VkSurfaceKHR >(resource.m_handle); break;
  6692. case VK_OBJECT_TYPE_SWAPCHAIN_KHR: destroy<VkSwapchainKHR >(resource.m_handle); break;
  6693. case VK_OBJECT_TYPE_DEVICE_MEMORY: destroy<VkDeviceMemory >(resource.m_handle); break;
  6694. default:
  6695. BX_ASSERT(false, "Invalid resource type: %d", resource.m_type);
  6696. break;
  6697. }
  6698. }
  6699. m_release[m_consumeIndex].clear();
  6700. }
  6701. void RendererContextVK::submitBlit(BlitState& _bs, uint16_t _view)
  6702. {
  6703. BGFX_PROFILER_SCOPE("RendererContextVK::submitBlit", kColorFrame);
  6704. VkImageLayout srcLayouts[BGFX_CONFIG_MAX_BLIT_ITEMS];
  6705. VkImageLayout dstLayouts[BGFX_CONFIG_MAX_BLIT_ITEMS];
  6706. BlitState bs0 = _bs;
  6707. while (bs0.hasItem(_view) )
  6708. {
  6709. uint16_t item = bs0.m_item;
  6710. const BlitItem& blit = bs0.advance();
  6711. TextureVK& src = m_textures[blit.m_src.idx];
  6712. TextureVK& dst = m_textures[blit.m_dst.idx];
  6713. srcLayouts[item] = VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_currentSingleMsaaImageLayout : src.m_currentImageLayout;
  6714. dstLayouts[item] = dst.m_currentImageLayout;
  6715. }
  6716. bs0 = _bs;
  6717. while (bs0.hasItem(_view) )
  6718. {
  6719. const BlitItem& blit = bs0.advance();
  6720. TextureVK& src = m_textures[blit.m_src.idx];
  6721. TextureVK& dst = m_textures[blit.m_dst.idx];
  6722. src.setImageMemoryBarrier(
  6723. m_commandBuffer
  6724. , blit.m_src.idx == blit.m_dst.idx ? VK_IMAGE_LAYOUT_GENERAL : VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  6725. , VK_NULL_HANDLE != src.m_singleMsaaImage
  6726. );
  6727. if (blit.m_src.idx != blit.m_dst.idx)
  6728. {
  6729. dst.setImageMemoryBarrier(m_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  6730. }
  6731. const uint16_t srcSamples = VK_NULL_HANDLE != src.m_singleMsaaImage ? 1 : src.m_sampler.Count;
  6732. const uint16_t dstSamples = dst.m_sampler.Count;
  6733. BX_UNUSED(srcSamples, dstSamples);
  6734. BX_ASSERT(
  6735. srcSamples == dstSamples
  6736. , "Mismatching texture sample count (%d != %d)."
  6737. , srcSamples
  6738. , dstSamples
  6739. );
  6740. VkImageCopy copyInfo;
  6741. copyInfo.srcSubresource.aspectMask = src.m_aspectMask;
  6742. copyInfo.srcSubresource.mipLevel = blit.m_srcMip;
  6743. copyInfo.srcSubresource.baseArrayLayer = 0;
  6744. copyInfo.srcSubresource.layerCount = 1;
  6745. copyInfo.srcOffset.x = blit.m_srcX;
  6746. copyInfo.srcOffset.y = blit.m_srcY;
  6747. copyInfo.srcOffset.z = 0;
  6748. copyInfo.dstSubresource.aspectMask = dst.m_aspectMask;
  6749. copyInfo.dstSubresource.mipLevel = blit.m_dstMip;
  6750. copyInfo.dstSubresource.baseArrayLayer = 0;
  6751. copyInfo.dstSubresource.layerCount = 1;
  6752. copyInfo.dstOffset.x = blit.m_dstX;
  6753. copyInfo.dstOffset.y = blit.m_dstY;
  6754. copyInfo.dstOffset.z = 0;
  6755. copyInfo.extent.width = blit.m_width;
  6756. copyInfo.extent.height = blit.m_height;
  6757. copyInfo.extent.depth = 1;
  6758. const uint32_t depth = bx::max<uint32_t>(1, blit.m_depth);
  6759. if (VK_IMAGE_VIEW_TYPE_3D == src.m_type)
  6760. {
  6761. BX_ASSERT(VK_IMAGE_VIEW_TYPE_3D == dst.m_type, "Can't blit between 2D and 3D image.");
  6762. copyInfo.srcOffset.z = blit.m_srcZ;
  6763. copyInfo.dstOffset.z = blit.m_dstZ;
  6764. copyInfo.extent.depth = depth;
  6765. }
  6766. else
  6767. {
  6768. copyInfo.srcSubresource.baseArrayLayer = blit.m_srcZ;
  6769. copyInfo.dstSubresource.baseArrayLayer = blit.m_dstZ;
  6770. copyInfo.srcSubresource.layerCount = depth;
  6771. copyInfo.dstSubresource.layerCount = depth;
  6772. }
  6773. vkCmdCopyImage(
  6774. m_commandBuffer
  6775. , VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_singleMsaaImage : src.m_textureImage
  6776. , VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_currentSingleMsaaImageLayout : src.m_currentImageLayout
  6777. , dst.m_textureImage
  6778. , dst.m_currentImageLayout
  6779. , 1
  6780. , &copyInfo
  6781. );
  6782. setMemoryBarrier(
  6783. m_commandBuffer
  6784. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6785. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6786. );
  6787. }
  6788. while (_bs.hasItem(_view) )
  6789. {
  6790. uint16_t item = _bs.m_item;
  6791. const BlitItem& blit = _bs.advance();
  6792. TextureVK& src = m_textures[blit.m_src.idx];
  6793. TextureVK& dst = m_textures[blit.m_dst.idx];
  6794. src.setImageMemoryBarrier(m_commandBuffer, srcLayouts[item], VK_NULL_HANDLE != src.m_singleMsaaImage);
  6795. dst.setImageMemoryBarrier(m_commandBuffer, dstLayouts[item]);
  6796. }
  6797. }
  6798. void RendererContextVK::submit(Frame* _render, ClearQuad& _clearQuad, TextVideoMemBlitter& _textVideoMemBlitter)
  6799. {
  6800. BX_UNUSED(_clearQuad);
  6801. if (updateResolution(_render->m_resolution) )
  6802. {
  6803. return;
  6804. }
  6805. if (_render->m_capture)
  6806. {
  6807. renderDocTriggerCapture();
  6808. }
  6809. BGFX_VK_PROFILER_BEGIN_LITERAL("rendererSubmit", kColorView);
  6810. int64_t timeBegin = bx::getHPCounter();
  6811. int64_t captureElapsed = 0;
  6812. uint32_t frameQueryIdx = UINT32_MAX;
  6813. if (m_timerQuerySupport)
  6814. {
  6815. frameQueryIdx = m_gpuTimer.begin(BGFX_CONFIG_MAX_VIEWS, _render->m_frameNum);
  6816. }
  6817. if (0 < _render->m_iboffset)
  6818. {
  6819. BGFX_PROFILER_SCOPE("bgfx/Update transient index buffer", kColorResource);
  6820. TransientIndexBuffer* ib = _render->m_transientIb;
  6821. m_indexBuffers[ib->handle.idx].update(m_commandBuffer, 0, _render->m_iboffset, ib->data);
  6822. }
  6823. if (0 < _render->m_vboffset)
  6824. {
  6825. BGFX_PROFILER_SCOPE("bgfx/Update transient vertex buffer", kColorResource);
  6826. TransientVertexBuffer* vb = _render->m_transientVb;
  6827. m_vertexBuffers[vb->handle.idx].update(m_commandBuffer, 0, _render->m_vboffset, vb->data);
  6828. }
  6829. _render->sort();
  6830. RenderDraw currentState;
  6831. currentState.clear();
  6832. currentState.m_stateFlags = BGFX_STATE_NONE;
  6833. currentState.m_stencil = packStencil(BGFX_STENCIL_NONE, BGFX_STENCIL_NONE);
  6834. static ViewState viewState;
  6835. viewState.reset(_render);
  6836. bool wireframe = !!(_render->m_debug&BGFX_DEBUG_WIREFRAME);
  6837. setDebugWireframe(wireframe);
  6838. ProgramHandle currentProgram = BGFX_INVALID_HANDLE;
  6839. bool hasPredefined = false;
  6840. VkPipeline currentPipeline = VK_NULL_HANDLE;
  6841. VkDescriptorSet currentDescriptorSet = VK_NULL_HANDLE;
  6842. uint32_t currentBindHash = 0;
  6843. uint32_t descriptorSetCount = 0;
  6844. VkIndexType currentIndexFormat = VK_INDEX_TYPE_MAX_ENUM;
  6845. SortKey key;
  6846. uint16_t view = UINT16_MAX;
  6847. FrameBufferHandle fbh = { BGFX_CONFIG_MAX_FRAME_BUFFERS };
  6848. BlitState bs(_render);
  6849. uint64_t blendFactor = UINT64_MAX;
  6850. bool wasCompute = false;
  6851. bool viewHasScissor = false;
  6852. bool restoreScissor = false;
  6853. Rect viewScissorRect;
  6854. viewScissorRect.clear();
  6855. bool isFrameBufferValid = false;
  6856. uint32_t statsNumPrimsSubmitted[BX_COUNTOF(s_primInfo)] = {};
  6857. uint32_t statsNumPrimsRendered[BX_COUNTOF(s_primInfo)] = {};
  6858. uint32_t statsNumInstances[BX_COUNTOF(s_primInfo)] = {};
  6859. uint32_t statsNumIndices = 0;
  6860. uint32_t statsKeyType[2] = {};
  6861. const uint64_t f0 = BGFX_STATE_BLEND_FACTOR;
  6862. const uint64_t f1 = BGFX_STATE_BLEND_INV_FACTOR;
  6863. const uint64_t f2 = BGFX_STATE_BLEND_FACTOR<<4;
  6864. const uint64_t f3 = BGFX_STATE_BLEND_INV_FACTOR<<4;
  6865. ScratchBufferVK& scratchBuffer = m_scratchBuffer[m_cmd.m_currentFrameInFlight];
  6866. scratchBuffer.reset();
  6867. setMemoryBarrier(
  6868. m_commandBuffer
  6869. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6870. , VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  6871. );
  6872. VkRenderPassBeginInfo rpbi;
  6873. rpbi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  6874. rpbi.pNext = NULL;
  6875. rpbi.clearValueCount = 0;
  6876. rpbi.pClearValues = NULL;
  6877. bool beginRenderPass = false;
  6878. Profiler<TimerQueryVK> profiler(
  6879. _render
  6880. , m_gpuTimer
  6881. , s_viewName
  6882. , m_timerQuerySupport
  6883. );
  6884. m_occlusionQuery.flush(_render);
  6885. if (0 == (_render->m_debug&BGFX_DEBUG_IFH) )
  6886. {
  6887. viewState.m_rect = _render->m_view[0].m_rect;
  6888. int32_t numItems = _render->m_numRenderItems;
  6889. for (int32_t item = 0; item < numItems;)
  6890. {
  6891. const uint64_t encodedKey = _render->m_sortKeys[item];
  6892. const bool isCompute = key.decode(encodedKey, _render->m_viewRemap);
  6893. statsKeyType[isCompute]++;
  6894. const bool viewChanged = 0
  6895. || key.m_view != view
  6896. || item == numItems
  6897. ;
  6898. const uint32_t itemIdx = _render->m_sortValues[item];
  6899. const RenderItem& renderItem = _render->m_renderItem[itemIdx];
  6900. const RenderBind& renderBind = _render->m_renderItemBind[itemIdx];
  6901. ++item;
  6902. if (viewChanged)
  6903. {
  6904. if (beginRenderPass)
  6905. {
  6906. vkCmdEndRenderPass(m_commandBuffer);
  6907. beginRenderPass = false;
  6908. }
  6909. view = key.m_view;
  6910. currentProgram = BGFX_INVALID_HANDLE;
  6911. hasPredefined = false;
  6912. if (item > 1)
  6913. {
  6914. profiler.end();
  6915. }
  6916. BGFX_VK_PROFILER_END();
  6917. setViewType(view, " ");
  6918. BGFX_VK_PROFILER_BEGIN(view, kColorView);
  6919. profiler.begin(view);
  6920. if (_render->m_view[view].m_fbh.idx != fbh.idx)
  6921. {
  6922. fbh = _render->m_view[view].m_fbh;
  6923. setFrameBuffer(fbh);
  6924. }
  6925. const FrameBufferVK& fb = isValid(m_fbh)
  6926. ? m_frameBuffers[m_fbh.idx]
  6927. : m_backBuffer
  6928. ;
  6929. isFrameBufferValid = fb.isRenderable();
  6930. if (isFrameBufferValid)
  6931. {
  6932. viewState.m_rect = _render->m_view[view].m_rect;
  6933. const Rect& rect = _render->m_view[view].m_rect;
  6934. const Rect& scissorRect = _render->m_view[view].m_scissor;
  6935. viewHasScissor = !scissorRect.isZero();
  6936. viewScissorRect = viewHasScissor ? scissorRect : rect;
  6937. restoreScissor = false;
  6938. rpbi.framebuffer = fb.m_currentFramebuffer;
  6939. rpbi.renderPass = fb.m_renderPass;
  6940. rpbi.renderArea.offset.x = rect.m_x;
  6941. rpbi.renderArea.offset.y = rect.m_y;
  6942. rpbi.renderArea.extent.width = rect.m_width;
  6943. rpbi.renderArea.extent.height = rect.m_height;
  6944. VkViewport vp;
  6945. vp.x = float(rect.m_x);
  6946. vp.y = float(rect.m_y + rect.m_height);
  6947. vp.width = float(rect.m_width);
  6948. vp.height = -float(rect.m_height);
  6949. vp.minDepth = 0.0f;
  6950. vp.maxDepth = 1.0f;
  6951. vkCmdSetViewport(m_commandBuffer, 0, 1, &vp);
  6952. VkRect2D rc;
  6953. rc.offset.x = viewScissorRect.m_x;
  6954. rc.offset.y = viewScissorRect.m_y;
  6955. rc.extent.width = viewScissorRect.m_width;
  6956. rc.extent.height = viewScissorRect.m_height;
  6957. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  6958. const Clear& clr = _render->m_view[view].m_clear;
  6959. if (BGFX_CLEAR_NONE != clr.m_flags)
  6960. {
  6961. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  6962. Rect clearRect = rect;
  6963. clearRect.setIntersect(rect, viewScissorRect);
  6964. clearQuad(clearRect, clr, _render->m_colorPalette);
  6965. vkCmdEndRenderPass(m_commandBuffer);
  6966. }
  6967. submitBlit(bs, view);
  6968. }
  6969. }
  6970. if (isCompute)
  6971. {
  6972. if (!wasCompute)
  6973. {
  6974. wasCompute = true;
  6975. currentBindHash = 0;
  6976. BGFX_VK_PROFILER_END();
  6977. setViewType(view, "C");
  6978. BGFX_VK_PROFILER_BEGIN(view, kColorCompute);
  6979. }
  6980. // renderpass external subpass dependencies handle graphics -> compute and compute -> graphics
  6981. // but not compute -> compute (possibly also across views if they contain no draw calls)
  6982. setMemoryBarrier(
  6983. m_commandBuffer
  6984. , VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  6985. , VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  6986. );
  6987. const RenderCompute& compute = renderItem.compute;
  6988. const VkPipeline pipeline = getPipeline(key.m_program);
  6989. if (currentPipeline != pipeline)
  6990. {
  6991. currentPipeline = pipeline;
  6992. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline);
  6993. }
  6994. bool constantsChanged = false;
  6995. if (compute.m_uniformBegin < compute.m_uniformEnd
  6996. || currentProgram.idx != key.m_program.idx)
  6997. {
  6998. rendererUpdateUniforms(this, _render->m_uniformBuffer[compute.m_uniformIdx], compute.m_uniformBegin, compute.m_uniformEnd);
  6999. currentProgram = key.m_program;
  7000. ProgramVK& program = m_program[currentProgram.idx];
  7001. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  7002. if (NULL != vcb)
  7003. {
  7004. commit(*vcb);
  7005. }
  7006. hasPredefined = 0 < program.m_numPredefined;
  7007. constantsChanged = true;
  7008. }
  7009. const ProgramVK& program = m_program[currentProgram.idx];
  7010. if (constantsChanged
  7011. || hasPredefined)
  7012. {
  7013. viewState.setPredefined<4>(this, view, program, _render, compute);
  7014. }
  7015. if (VK_NULL_HANDLE != program.m_descriptorSetLayout)
  7016. {
  7017. const uint32_t vsize = program.m_vsh->m_size;
  7018. uint32_t numOffset = 0;
  7019. uint32_t offset = 0;
  7020. if (constantsChanged
  7021. || hasPredefined)
  7022. {
  7023. if (vsize > 0)
  7024. {
  7025. offset = scratchBuffer.write(m_vsScratch, vsize);
  7026. ++numOffset;
  7027. }
  7028. }
  7029. bx::HashMurmur2A hash;
  7030. hash.begin();
  7031. hash.add(program.m_descriptorSetLayout);
  7032. hash.add(renderBind.m_bind, sizeof(renderBind.m_bind) );
  7033. hash.add(vsize);
  7034. hash.add(0);
  7035. const uint32_t bindHash = hash.end();
  7036. if (currentBindHash != bindHash)
  7037. {
  7038. currentBindHash = bindHash;
  7039. currentDescriptorSet = getDescriptorSet(
  7040. program
  7041. , renderBind
  7042. , scratchBuffer
  7043. , _render->m_colorPalette
  7044. );
  7045. descriptorSetCount++;
  7046. }
  7047. vkCmdBindDescriptorSets(
  7048. m_commandBuffer
  7049. , VK_PIPELINE_BIND_POINT_COMPUTE
  7050. , program.m_pipelineLayout
  7051. , 0
  7052. , 1
  7053. , &currentDescriptorSet
  7054. , numOffset
  7055. , &offset
  7056. );
  7057. }
  7058. if (isValid(compute.m_indirectBuffer) )
  7059. {
  7060. const VertexBufferVK& vb = m_vertexBuffers[compute.m_indirectBuffer.idx];
  7061. uint32_t numDrawIndirect = UINT32_MAX == compute.m_numIndirect
  7062. ? vb.m_size/BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7063. : compute.m_numIndirect
  7064. ;
  7065. uint32_t args = compute.m_startIndirect * BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7066. for (uint32_t ii = 0; ii < numDrawIndirect; ++ii)
  7067. {
  7068. vkCmdDispatchIndirect(m_commandBuffer, vb.m_buffer, args);
  7069. args += BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7070. }
  7071. }
  7072. else
  7073. {
  7074. vkCmdDispatch(m_commandBuffer, compute.m_numX, compute.m_numY, compute.m_numZ);
  7075. }
  7076. continue;
  7077. }
  7078. const RenderDraw& draw = renderItem.draw;
  7079. rendererUpdateUniforms(this, _render->m_uniformBuffer[draw.m_uniformIdx], draw.m_uniformBegin, draw.m_uniformEnd);
  7080. const bool hasOcclusionQuery = 0 != (draw.m_stateFlags & BGFX_STATE_INTERNAL_OCCLUSION_QUERY);
  7081. {
  7082. const bool occluded = true
  7083. && isValid(draw.m_occlusionQuery)
  7084. && !hasOcclusionQuery
  7085. && !isVisible(_render, draw.m_occlusionQuery, 0 != (draw.m_submitFlags & BGFX_SUBMIT_INTERNAL_OCCLUSION_VISIBLE) )
  7086. ;
  7087. if (occluded
  7088. || !isFrameBufferValid
  7089. || 0 == draw.m_streamMask
  7090. || _render->m_frameCache.isZeroArea(viewScissorRect, draw.m_scissor) )
  7091. {
  7092. continue;
  7093. }
  7094. }
  7095. const uint64_t changedFlags = currentState.m_stateFlags ^ draw.m_stateFlags;
  7096. currentState.m_stateFlags = draw.m_stateFlags;
  7097. if (!beginRenderPass)
  7098. {
  7099. if (wasCompute)
  7100. {
  7101. wasCompute = false;
  7102. currentBindHash = 0;
  7103. }
  7104. BGFX_VK_PROFILER_END();
  7105. setViewType(view, " ");
  7106. BGFX_VK_PROFILER_BEGIN(view, kColorDraw);
  7107. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  7108. beginRenderPass = true;
  7109. currentProgram = BGFX_INVALID_HANDLE;
  7110. currentState.m_scissor = !draw.m_scissor;
  7111. }
  7112. if (0 != draw.m_streamMask)
  7113. {
  7114. const bool bindAttribs = hasVertexStreamChanged(currentState, draw);
  7115. currentState.m_streamMask = draw.m_streamMask;
  7116. currentState.m_instanceDataBuffer = draw.m_instanceDataBuffer;
  7117. currentState.m_instanceDataOffset = draw.m_instanceDataOffset;
  7118. currentState.m_instanceDataStride = draw.m_instanceDataStride;
  7119. const VertexLayout* layouts[BGFX_CONFIG_MAX_VERTEX_STREAMS];
  7120. VkBuffer streamBuffers[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  7121. VkDeviceSize streamOffsets[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  7122. uint8_t numStreams = 0;
  7123. uint32_t numVertices = draw.m_numVertices;
  7124. if (UINT8_MAX != draw.m_streamMask)
  7125. {
  7126. for (uint32_t idx = 0, streamMask = draw.m_streamMask
  7127. ; 0 != streamMask
  7128. ; streamMask >>= 1, idx += 1, ++numStreams
  7129. )
  7130. {
  7131. const uint32_t ntz = bx::uint32_cnttz(streamMask);
  7132. streamMask >>= ntz;
  7133. idx += ntz;
  7134. currentState.m_stream[idx] = draw.m_stream[idx];
  7135. const VertexBufferHandle handle = draw.m_stream[idx].m_handle;
  7136. const VertexBufferVK& vb = m_vertexBuffers[handle.idx];
  7137. const uint16_t decl = isValid(draw.m_stream[idx].m_layoutHandle)
  7138. ? draw.m_stream[idx].m_layoutHandle.idx
  7139. : vb.m_layoutHandle.idx
  7140. ;
  7141. const VertexLayout& layout = m_vertexLayouts[decl];
  7142. const uint32_t stride = layout.m_stride;
  7143. streamBuffers[numStreams] = m_vertexBuffers[handle.idx].m_buffer;
  7144. streamOffsets[numStreams] = draw.m_stream[idx].m_startVertex * stride;
  7145. layouts[numStreams] = &layout;
  7146. numVertices = bx::uint32_min(UINT32_MAX == draw.m_numVertices
  7147. ? vb.m_size/stride
  7148. : draw.m_numVertices
  7149. , numVertices
  7150. );
  7151. }
  7152. }
  7153. if (bindAttribs)
  7154. {
  7155. uint32_t numVertexBuffers = numStreams;
  7156. if (isValid(draw.m_instanceDataBuffer) )
  7157. {
  7158. streamOffsets[numVertexBuffers] = draw.m_instanceDataOffset;
  7159. streamBuffers[numVertexBuffers] = m_vertexBuffers[draw.m_instanceDataBuffer.idx].m_buffer;
  7160. numVertexBuffers++;
  7161. }
  7162. if (0 < numVertexBuffers)
  7163. {
  7164. vkCmdBindVertexBuffers(
  7165. m_commandBuffer
  7166. , 0
  7167. , numVertexBuffers
  7168. , &streamBuffers[0]
  7169. , streamOffsets
  7170. );
  7171. }
  7172. }
  7173. const VkPipeline pipeline =
  7174. getPipeline(draw.m_stateFlags
  7175. , draw.m_stencil
  7176. , numStreams
  7177. , layouts
  7178. , key.m_program
  7179. , uint8_t(draw.m_instanceDataStride/16)
  7180. );
  7181. if (currentPipeline != pipeline)
  7182. {
  7183. currentPipeline = pipeline;
  7184. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline);
  7185. }
  7186. const bool hasStencil = 0 != draw.m_stencil;
  7187. if (hasStencil
  7188. && currentState.m_stencil != draw.m_stencil)
  7189. {
  7190. currentState.m_stencil = draw.m_stencil;
  7191. const uint32_t fstencil = unpackStencil(0, draw.m_stencil);
  7192. const uint32_t ref = (fstencil&BGFX_STENCIL_FUNC_REF_MASK)>>BGFX_STENCIL_FUNC_REF_SHIFT;
  7193. vkCmdSetStencilReference(m_commandBuffer, VK_STENCIL_FRONT_AND_BACK, ref);
  7194. }
  7195. const bool hasFactor = 0
  7196. || f0 == (draw.m_stateFlags & f0)
  7197. || f1 == (draw.m_stateFlags & f1)
  7198. || f2 == (draw.m_stateFlags & f2)
  7199. || f3 == (draw.m_stateFlags & f3)
  7200. ;
  7201. if (hasFactor
  7202. && blendFactor != draw.m_rgba)
  7203. {
  7204. blendFactor = draw.m_rgba;
  7205. float bf[4];
  7206. bf[0] = ( (draw.m_rgba>>24) )/255.0f;
  7207. bf[1] = ( (draw.m_rgba>>16)&0xff)/255.0f;
  7208. bf[2] = ( (draw.m_rgba>> 8)&0xff)/255.0f;
  7209. bf[3] = ( (draw.m_rgba )&0xff)/255.0f;
  7210. vkCmdSetBlendConstants(m_commandBuffer, bf);
  7211. }
  7212. const uint16_t scissor = draw.m_scissor;
  7213. if (currentState.m_scissor != scissor)
  7214. {
  7215. currentState.m_scissor = scissor;
  7216. if (UINT16_MAX == scissor)
  7217. {
  7218. if (restoreScissor
  7219. || viewHasScissor)
  7220. {
  7221. restoreScissor = false;
  7222. VkRect2D rc;
  7223. rc.offset.x = viewScissorRect.m_x;
  7224. rc.offset.y = viewScissorRect.m_y;
  7225. rc.extent.width = viewScissorRect.m_width;
  7226. rc.extent.height = viewScissorRect.m_height;
  7227. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  7228. }
  7229. }
  7230. else
  7231. {
  7232. restoreScissor = true;
  7233. Rect scissorRect;
  7234. scissorRect.setIntersect(viewScissorRect, _render->m_frameCache.m_rectCache.m_cache[scissor]);
  7235. VkRect2D rc;
  7236. rc.offset.x = scissorRect.m_x;
  7237. rc.offset.y = scissorRect.m_y;
  7238. rc.extent.width = scissorRect.m_width;
  7239. rc.extent.height = scissorRect.m_height;
  7240. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  7241. }
  7242. }
  7243. bool constantsChanged = false;
  7244. if (draw.m_uniformBegin < draw.m_uniformEnd
  7245. || currentProgram.idx != key.m_program.idx
  7246. || BGFX_STATE_ALPHA_REF_MASK & changedFlags)
  7247. {
  7248. currentProgram = key.m_program;
  7249. ProgramVK& program = m_program[currentProgram.idx];
  7250. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  7251. if (NULL != vcb)
  7252. {
  7253. commit(*vcb);
  7254. }
  7255. if (NULL != program.m_fsh)
  7256. {
  7257. UniformBuffer* fcb = program.m_fsh->m_constantBuffer;
  7258. if (NULL != fcb)
  7259. {
  7260. commit(*fcb);
  7261. }
  7262. }
  7263. hasPredefined = 0 < program.m_numPredefined;
  7264. constantsChanged = true;
  7265. }
  7266. const ProgramVK& program = m_program[currentProgram.idx];
  7267. if (hasPredefined)
  7268. {
  7269. uint32_t ref = (draw.m_stateFlags & BGFX_STATE_ALPHA_REF_MASK) >> BGFX_STATE_ALPHA_REF_SHIFT;
  7270. viewState.m_alphaRef = ref / 255.0f;
  7271. viewState.setPredefined<4>(this, view, program, _render, draw);
  7272. }
  7273. if (VK_NULL_HANDLE != program.m_descriptorSetLayout)
  7274. {
  7275. const uint32_t vsize = program.m_vsh->m_size;
  7276. const uint32_t fsize = NULL != program.m_fsh ? program.m_fsh->m_size : 0;
  7277. uint32_t numOffset = 0;
  7278. uint32_t offsets[2] = { 0, 0 };
  7279. if (constantsChanged
  7280. || hasPredefined)
  7281. {
  7282. if (vsize > 0)
  7283. {
  7284. offsets[numOffset++] = scratchBuffer.write(m_vsScratch, vsize);
  7285. }
  7286. if (fsize > 0)
  7287. {
  7288. offsets[numOffset++] = scratchBuffer.write(m_fsScratch, fsize);
  7289. }
  7290. }
  7291. bx::HashMurmur2A hash;
  7292. hash.begin();
  7293. hash.add(program.m_descriptorSetLayout);
  7294. hash.add(renderBind.m_bind, sizeof(renderBind.m_bind) );
  7295. hash.add(vsize);
  7296. hash.add(fsize);
  7297. const uint32_t bindHash = hash.end();
  7298. if (currentBindHash != bindHash)
  7299. {
  7300. currentBindHash = bindHash;
  7301. currentDescriptorSet = getDescriptorSet(
  7302. program
  7303. , renderBind
  7304. , scratchBuffer
  7305. , _render->m_colorPalette
  7306. );
  7307. descriptorSetCount++;
  7308. }
  7309. vkCmdBindDescriptorSets(
  7310. m_commandBuffer
  7311. , VK_PIPELINE_BIND_POINT_GRAPHICS
  7312. , program.m_pipelineLayout
  7313. , 0
  7314. , 1
  7315. , &currentDescriptorSet
  7316. , numOffset
  7317. , offsets
  7318. );
  7319. }
  7320. VkBuffer bufferIndirect = VK_NULL_HANDLE;
  7321. VkBuffer bufferNumIndirect = VK_NULL_HANDLE;
  7322. uint32_t numDrawIndirect = 0;
  7323. uint32_t bufferOffsetIndirect = 0;
  7324. uint32_t bufferNumOffsetIndirect = 0;
  7325. if (isValid(draw.m_indirectBuffer) )
  7326. {
  7327. const VertexBufferVK& vb = m_vertexBuffers[draw.m_indirectBuffer.idx];
  7328. bufferIndirect = vb.m_buffer;
  7329. numDrawIndirect = UINT32_MAX == draw.m_numIndirect
  7330. ? vb.m_size / BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7331. : draw.m_numIndirect
  7332. ;
  7333. bufferOffsetIndirect = draw.m_startIndirect * BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7334. if (isValid(draw.m_numIndirectBuffer) )
  7335. {
  7336. bufferNumIndirect = m_indexBuffers[draw.m_numIndirectBuffer.idx].m_buffer;
  7337. bufferNumOffsetIndirect = draw.m_numIndirectIndex * sizeof(uint32_t);
  7338. }
  7339. }
  7340. if (hasOcclusionQuery)
  7341. {
  7342. m_occlusionQuery.begin(draw.m_occlusionQuery);
  7343. }
  7344. const uint8_t primIndex = uint8_t((draw.m_stateFlags & BGFX_STATE_PT_MASK) >> BGFX_STATE_PT_SHIFT);
  7345. const PrimInfo& prim = s_primInfo[primIndex];
  7346. uint32_t numPrimsSubmitted = 0;
  7347. uint32_t numIndices = 0;
  7348. if (!isValid(draw.m_indexBuffer) )
  7349. {
  7350. numPrimsSubmitted = numVertices / prim.m_div - prim.m_sub;
  7351. if (isValid(draw.m_indirectBuffer) )
  7352. {
  7353. if (isValid(draw.m_numIndirectBuffer) )
  7354. {
  7355. vkCmdDrawIndirectCountKHR(
  7356. m_commandBuffer
  7357. , bufferIndirect
  7358. , bufferOffsetIndirect
  7359. , bufferNumIndirect
  7360. , bufferNumOffsetIndirect
  7361. , numDrawIndirect
  7362. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7363. );
  7364. }
  7365. else
  7366. {
  7367. vkCmdDrawIndirect(
  7368. m_commandBuffer
  7369. , bufferIndirect
  7370. , bufferOffsetIndirect
  7371. , numDrawIndirect
  7372. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7373. );
  7374. }
  7375. }
  7376. else
  7377. {
  7378. vkCmdDraw(
  7379. m_commandBuffer
  7380. , numVertices
  7381. , draw.m_numInstances
  7382. , 0
  7383. , 0
  7384. );
  7385. }
  7386. }
  7387. else
  7388. {
  7389. const bool isIndex16 = draw.isIndex16();
  7390. const uint32_t indexSize = isIndex16 ? 2 : 4;
  7391. const VkIndexType indexFormat = isIndex16 ? VK_INDEX_TYPE_UINT16 : VK_INDEX_TYPE_UINT32;
  7392. const BufferVK& ib = m_indexBuffers[draw.m_indexBuffer.idx];
  7393. numIndices = UINT32_MAX == draw.m_numIndices
  7394. ? ib.m_size / indexSize
  7395. : draw.m_numIndices
  7396. ;
  7397. numPrimsSubmitted = numIndices / prim.m_div - prim.m_sub;
  7398. if (currentState.m_indexBuffer.idx != draw.m_indexBuffer.idx
  7399. || currentIndexFormat != indexFormat)
  7400. {
  7401. currentState.m_indexBuffer = draw.m_indexBuffer;
  7402. currentIndexFormat = indexFormat;
  7403. vkCmdBindIndexBuffer(
  7404. m_commandBuffer
  7405. , m_indexBuffers[draw.m_indexBuffer.idx].m_buffer
  7406. , 0
  7407. , indexFormat
  7408. );
  7409. }
  7410. if (isValid(draw.m_indirectBuffer) )
  7411. {
  7412. if (isValid(draw.m_numIndirectBuffer) )
  7413. {
  7414. vkCmdDrawIndexedIndirectCountKHR(
  7415. m_commandBuffer
  7416. , bufferIndirect
  7417. , bufferOffsetIndirect
  7418. , bufferNumIndirect
  7419. , bufferNumOffsetIndirect
  7420. , numDrawIndirect
  7421. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7422. );
  7423. }
  7424. else
  7425. {
  7426. vkCmdDrawIndexedIndirect(
  7427. m_commandBuffer
  7428. , bufferIndirect
  7429. , bufferOffsetIndirect
  7430. , numDrawIndirect
  7431. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7432. );
  7433. }
  7434. }
  7435. else
  7436. {
  7437. vkCmdDrawIndexed(
  7438. m_commandBuffer
  7439. , numIndices
  7440. , draw.m_numInstances
  7441. , draw.m_startIndex
  7442. , 0
  7443. , 0
  7444. );
  7445. }
  7446. }
  7447. uint32_t numPrimsRendered = numPrimsSubmitted*draw.m_numInstances;
  7448. statsNumPrimsSubmitted[primIndex] += numPrimsSubmitted;
  7449. statsNumPrimsRendered[primIndex] += numPrimsRendered;
  7450. statsNumInstances[primIndex] += draw.m_numInstances;
  7451. statsNumIndices += numIndices;
  7452. if (hasOcclusionQuery)
  7453. {
  7454. m_occlusionQuery.end();
  7455. }
  7456. }
  7457. }
  7458. if (beginRenderPass)
  7459. {
  7460. vkCmdEndRenderPass(m_commandBuffer);
  7461. beginRenderPass = false;
  7462. }
  7463. if (wasCompute)
  7464. {
  7465. setViewType(view, "C");
  7466. BGFX_VK_PROFILER_END();
  7467. BGFX_VK_PROFILER_BEGIN(view, kColorCompute);
  7468. }
  7469. submitBlit(bs, BGFX_CONFIG_MAX_VIEWS);
  7470. if (0 < _render->m_numRenderItems)
  7471. {
  7472. captureElapsed = -bx::getHPCounter();
  7473. capture();
  7474. captureElapsed += bx::getHPCounter();
  7475. profiler.end();
  7476. }
  7477. }
  7478. BGFX_VK_PROFILER_END();
  7479. int64_t timeEnd = bx::getHPCounter();
  7480. int64_t frameTime = timeEnd - timeBegin;
  7481. static int64_t min = frameTime;
  7482. static int64_t max = frameTime;
  7483. min = bx::min<int64_t>(min, frameTime);
  7484. max = bx::max<int64_t>(max, frameTime);
  7485. static uint32_t maxGpuLatency = 0;
  7486. static double maxGpuElapsed = 0.0f;
  7487. double elapsedGpuMs = 0.0;
  7488. static int64_t presentMin = m_presentElapsed;
  7489. static int64_t presentMax = m_presentElapsed;
  7490. presentMin = bx::min<int64_t>(presentMin, m_presentElapsed);
  7491. presentMax = bx::max<int64_t>(presentMax, m_presentElapsed);
  7492. if (UINT32_MAX != frameQueryIdx)
  7493. {
  7494. m_gpuTimer.end(frameQueryIdx);
  7495. const TimerQueryVK::Result& result = m_gpuTimer.m_result[BGFX_CONFIG_MAX_VIEWS];
  7496. double toGpuMs = 1000.0 / double(m_gpuTimer.m_frequency);
  7497. elapsedGpuMs = (result.m_end - result.m_begin) * toGpuMs;
  7498. maxGpuElapsed = elapsedGpuMs > maxGpuElapsed ? elapsedGpuMs : maxGpuElapsed;
  7499. maxGpuLatency = bx::uint32_imax(maxGpuLatency, result.m_pending-1);
  7500. }
  7501. maxGpuLatency = bx::uint32_imax(maxGpuLatency, m_gpuTimer.m_control.available()-1);
  7502. const int64_t timerFreq = bx::getHPFrequency();
  7503. VkPhysicalDeviceMemoryBudgetPropertiesEXT dmbp;
  7504. dmbp.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_BUDGET_PROPERTIES_EXT;
  7505. dmbp.pNext = NULL;
  7506. int64_t gpuMemoryAvailable = -INT64_MAX;
  7507. int64_t gpuMemoryUsed = -INT64_MAX;
  7508. if (s_extension[Extension::EXT_memory_budget].m_supported)
  7509. {
  7510. VkPhysicalDeviceMemoryProperties2 pdmp2;
  7511. pdmp2.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_PROPERTIES_2;
  7512. pdmp2.pNext = &dmbp;
  7513. vkGetPhysicalDeviceMemoryProperties2KHR(m_physicalDevice, &pdmp2);
  7514. gpuMemoryAvailable = 0;
  7515. gpuMemoryUsed = 0;
  7516. for (uint32_t ii = 0; ii < m_memoryProperties.memoryHeapCount; ++ii)
  7517. {
  7518. if (!!(m_memoryProperties.memoryHeaps[ii].flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) )
  7519. {
  7520. gpuMemoryAvailable += dmbp.heapBudget[ii];
  7521. gpuMemoryUsed += dmbp.heapUsage[ii];
  7522. }
  7523. }
  7524. }
  7525. Stats& perfStats = _render->m_perfStats;
  7526. perfStats.cpuTimeBegin = timeBegin;
  7527. perfStats.cpuTimeEnd = timeEnd;
  7528. perfStats.cpuTimerFreq = timerFreq;
  7529. const TimerQueryVK::Result& result = m_gpuTimer.m_result[BGFX_CONFIG_MAX_VIEWS];
  7530. perfStats.gpuTimeBegin = result.m_begin;
  7531. perfStats.gpuTimeEnd = result.m_end;
  7532. perfStats.gpuTimerFreq = m_gpuTimer.m_frequency;
  7533. perfStats.numDraw = statsKeyType[0];
  7534. perfStats.numCompute = statsKeyType[1];
  7535. perfStats.numBlit = _render->m_numBlitItems;
  7536. perfStats.maxGpuLatency = maxGpuLatency;
  7537. perfStats.gpuFrameNum = result.m_frameNum;
  7538. bx::memCopy(perfStats.numPrims, statsNumPrimsRendered, sizeof(perfStats.numPrims) );
  7539. perfStats.gpuMemoryMax = gpuMemoryAvailable;
  7540. perfStats.gpuMemoryUsed = gpuMemoryUsed;
  7541. if (_render->m_debug & (BGFX_DEBUG_IFH|BGFX_DEBUG_STATS) )
  7542. {
  7543. BGFX_VK_PROFILER_BEGIN_LITERAL("debugstats", kColorFrame);
  7544. TextVideoMem& tvm = m_textVideoMem;
  7545. static int64_t next = timeEnd;
  7546. if (timeEnd >= next)
  7547. {
  7548. next = timeEnd + timerFreq;
  7549. double freq = double(timerFreq);
  7550. double toMs = 1000.0 / freq;
  7551. tvm.clear();
  7552. uint16_t pos = 0;
  7553. tvm.printf(0, pos++, BGFX_CONFIG_DEBUG ? 0x8c : 0x8f
  7554. , " %s / " BX_COMPILER_NAME
  7555. " / " BX_CPU_NAME
  7556. " / " BX_ARCH_NAME
  7557. " / " BX_PLATFORM_NAME
  7558. " / Version 1.%d.%d (commit: " BGFX_REV_SHA1 ")"
  7559. , getRendererName()
  7560. , BGFX_API_VERSION
  7561. , BGFX_REV_NUMBER
  7562. );
  7563. const VkPhysicalDeviceProperties& pdp = m_deviceProperties;
  7564. tvm.printf(0, pos++, 0x8f, " Device: %s (%s)"
  7565. , pdp.deviceName
  7566. , getName(pdp.deviceType)
  7567. );
  7568. if (0 <= gpuMemoryAvailable && 0 <= gpuMemoryUsed)
  7569. {
  7570. for (uint32_t ii = 0; ii < m_memoryProperties.memoryHeapCount; ++ii)
  7571. {
  7572. char budget[16];
  7573. bx::prettify(budget, BX_COUNTOF(budget), dmbp.heapBudget[ii]);
  7574. char usage[16];
  7575. bx::prettify(usage, BX_COUNTOF(usage), dmbp.heapUsage[ii]);
  7576. const bool local = (!!(m_memoryProperties.memoryHeaps[ii].flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) );
  7577. tvm.printf(0, pos++, 0x8f, " Memory %d %s - Budget: %12s, Usage: %12s"
  7578. , ii
  7579. , local ? "(local) " : "(non-local)"
  7580. , budget
  7581. , usage
  7582. );
  7583. }
  7584. }
  7585. pos = 10;
  7586. tvm.printf(10, pos++, 0x8b, " Frame: % 7.3f, % 7.3f \x1f, % 7.3f \x1e [ms] / % 6.2f FPS "
  7587. , double(frameTime)*toMs
  7588. , double(min)*toMs
  7589. , double(max)*toMs
  7590. , freq/frameTime
  7591. );
  7592. tvm.printf(10, pos++, 0x8b, " Present: % 7.3f, % 7.3f \x1f, % 7.3f \x1e [ms] "
  7593. , double(m_presentElapsed)*toMs
  7594. , double(presentMin)*toMs
  7595. , double(presentMax)*toMs
  7596. );
  7597. const uint32_t msaa = (m_resolution.reset&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT;
  7598. tvm.printf(10, pos++, 0x8b, " Reset flags: [%c] vsync, [%c] MSAAx%d, [%c] MaxAnisotropy "
  7599. , !!(m_resolution.reset&BGFX_RESET_VSYNC) ? '\xfe' : ' '
  7600. , 0 != msaa ? '\xfe' : ' '
  7601. , 1<<msaa
  7602. , !!(m_resolution.reset&BGFX_RESET_MAXANISOTROPY) ? '\xfe' : ' '
  7603. );
  7604. double elapsedCpuMs = double(frameTime)*toMs;
  7605. tvm.printf(10, pos++, 0x8b, " Submitted: %5d (draw %5d, compute %4d) / CPU %7.4f [ms] "
  7606. , _render->m_numRenderItems
  7607. , statsKeyType[0]
  7608. , statsKeyType[1]
  7609. , elapsedCpuMs
  7610. );
  7611. for (uint32_t ii = 0; ii < Topology::Count; ++ii)
  7612. {
  7613. tvm.printf(10, pos++, 0x8b, " %9s: %7d (#inst: %5d), submitted: %7d "
  7614. , getName(Topology::Enum(ii) )
  7615. , statsNumPrimsRendered[ii]
  7616. , statsNumInstances[ii]
  7617. , statsNumPrimsSubmitted[ii]
  7618. );
  7619. }
  7620. if (NULL != m_renderDocDll)
  7621. {
  7622. tvm.printf(tvm.m_width-27, 0, 0x4f, " [F11 - RenderDoc capture] ");
  7623. }
  7624. tvm.printf(10, pos++, 0x8b, " Indices: %7d ", statsNumIndices);
  7625. // tvm.printf(10, pos++, 0x8b, " Uniform size: %7d, Max: %7d ", _render->m_uniformEnd, _render->m_uniformMax);
  7626. tvm.printf(10, pos++, 0x8b, " DVB size: %7d ", _render->m_vboffset);
  7627. tvm.printf(10, pos++, 0x8b, " DIB size: %7d ", _render->m_iboffset);
  7628. pos++;
  7629. tvm.printf(10, pos++, 0x8b, " Occlusion queries: %3d ", m_occlusionQuery.m_control.available() );
  7630. pos++;
  7631. tvm.printf(10, pos++, 0x8b, " State cache: ");
  7632. tvm.printf(10, pos++, 0x8b, " PSO | DSL | DS ");
  7633. tvm.printf(10, pos++, 0x8b, " %6d | %6d | %6d "
  7634. , m_pipelineStateCache.getCount()
  7635. , m_descriptorSetLayoutCache.getCount()
  7636. , descriptorSetCount
  7637. );
  7638. pos++;
  7639. double captureMs = double(captureElapsed)*toMs;
  7640. tvm.printf(10, pos++, 0x8b, " Capture: %7.4f [ms] ", captureMs);
  7641. uint8_t attr[2] = { 0x8c, 0x8a };
  7642. uint8_t attrIndex = _render->m_waitSubmit < _render->m_waitRender;
  7643. tvm.printf(10, pos++, attr[attrIndex&1], " Submit wait: %7.4f [ms] ", _render->m_waitSubmit*toMs);
  7644. tvm.printf(10, pos++, attr[(attrIndex+1)&1], " Render wait: %7.4f [ms] ", _render->m_waitRender*toMs);
  7645. min = frameTime;
  7646. max = frameTime;
  7647. presentMin = m_presentElapsed;
  7648. presentMax = m_presentElapsed;
  7649. }
  7650. blit(this, _textVideoMemBlitter, tvm);
  7651. BGFX_VK_PROFILER_END();
  7652. }
  7653. else if (_render->m_debug & BGFX_DEBUG_TEXT)
  7654. {
  7655. BGFX_VK_PROFILER_BEGIN_LITERAL("debugtext", kColorFrame);
  7656. blit(this, _textVideoMemBlitter, _render->m_textVideoMem);
  7657. BGFX_VK_PROFILER_END();
  7658. }
  7659. m_presentElapsed = 0;
  7660. scratchBuffer.flush();
  7661. for (uint16_t ii = 0; ii < m_numWindows; ++ii)
  7662. {
  7663. FrameBufferVK& fb = isValid(m_windows[ii])
  7664. ? m_frameBuffers[m_windows[ii].idx]
  7665. : m_backBuffer
  7666. ;
  7667. if (fb.m_needPresent)
  7668. {
  7669. fb.resolve();
  7670. fb.m_swapChain.transitionImage(m_commandBuffer);
  7671. m_cmd.addWaitSemaphore(fb.m_swapChain.m_lastImageAcquiredSemaphore, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  7672. m_cmd.addSignalSemaphore(fb.m_swapChain.m_lastImageRenderedSemaphore);
  7673. fb.m_swapChain.m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  7674. fb.m_swapChain.m_backBufferFence[fb.m_swapChain.m_backBufferColorIdx] = m_cmd.m_currentFence;
  7675. }
  7676. }
  7677. kick();
  7678. }
  7679. } /* namespace vk */ } // namespace bgfx
  7680. #else
  7681. namespace bgfx { namespace vk
  7682. {
  7683. RendererContextI* rendererCreate(const Init& _init)
  7684. {
  7685. BX_UNUSED(_init);
  7686. return NULL;
  7687. }
  7688. void rendererDestroy()
  7689. {
  7690. }
  7691. } /* namespace vk */ } // namespace bgfx
  7692. #endif // BGFX_CONFIG_RENDERER_VULKAN