renderer_vk.cpp 285 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900490149024903490449054906490749084909491049114912491349144915491649174918491949204921492249234924492549264927492849294930493149324933493449354936493749384939494049414942494349444945494649474948494949504951495249534954495549564957495849594960496149624963496449654966496749684969497049714972497349744975497649774978497949804981498249834984498549864987498849894990499149924993499449954996499749984999500050015002500350045005500650075008500950105011501250135014501550165017501850195020502150225023502450255026502750285029503050315032503350345035503650375038503950405041504250435044504550465047504850495050505150525053505450555056505750585059506050615062506350645065506650675068506950705071507250735074507550765077507850795080508150825083508450855086508750885089509050915092509350945095509650975098509951005101510251035104510551065107510851095110511151125113511451155116511751185119512051215122512351245125512651275128512951305131513251335134513551365137513851395140514151425143514451455146514751485149515051515152515351545155515651575158515951605161516251635164516551665167516851695170517151725173517451755176517751785179518051815182518351845185518651875188518951905191519251935194519551965197519851995200520152025203520452055206520752085209521052115212521352145215521652175218521952205221522252235224522552265227522852295230523152325233523452355236523752385239524052415242524352445245524652475248524952505251525252535254525552565257525852595260526152625263526452655266526752685269527052715272527352745275527652775278527952805281528252835284528552865287528852895290529152925293529452955296529752985299530053015302530353045305530653075308530953105311531253135314531553165317531853195320532153225323532453255326532753285329533053315332533353345335533653375338533953405341534253435344534553465347534853495350535153525353535453555356535753585359536053615362536353645365536653675368536953705371537253735374537553765377537853795380538153825383538453855386538753885389539053915392539353945395539653975398539954005401540254035404540554065407540854095410541154125413541454155416541754185419542054215422542354245425542654275428542954305431543254335434543554365437543854395440544154425443544454455446544754485449545054515452545354545455545654575458545954605461546254635464546554665467546854695470547154725473547454755476547754785479548054815482548354845485548654875488548954905491549254935494549554965497549854995500550155025503550455055506550755085509551055115512551355145515551655175518551955205521552255235524552555265527552855295530553155325533553455355536553755385539554055415542554355445545554655475548554955505551555255535554555555565557555855595560556155625563556455655566556755685569557055715572557355745575557655775578557955805581558255835584558555865587558855895590559155925593559455955596559755985599560056015602560356045605560656075608560956105611561256135614561556165617561856195620562156225623562456255626562756285629563056315632563356345635563656375638563956405641564256435644564556465647564856495650565156525653565456555656565756585659566056615662566356645665566656675668566956705671567256735674567556765677567856795680568156825683568456855686568756885689569056915692569356945695569656975698569957005701570257035704570557065707570857095710571157125713571457155716571757185719572057215722572357245725572657275728572957305731573257335734573557365737573857395740574157425743574457455746574757485749575057515752575357545755575657575758575957605761576257635764576557665767576857695770577157725773577457755776577757785779578057815782578357845785578657875788578957905791579257935794579557965797579857995800580158025803580458055806580758085809581058115812581358145815581658175818581958205821582258235824582558265827582858295830583158325833583458355836583758385839584058415842584358445845584658475848584958505851585258535854585558565857585858595860586158625863586458655866586758685869587058715872587358745875587658775878587958805881588258835884588558865887588858895890589158925893589458955896589758985899590059015902590359045905590659075908590959105911591259135914591559165917591859195920592159225923592459255926592759285929593059315932593359345935593659375938593959405941594259435944594559465947594859495950595159525953595459555956595759585959596059615962596359645965596659675968596959705971597259735974597559765977597859795980598159825983598459855986598759885989599059915992599359945995599659975998599960006001600260036004600560066007600860096010601160126013601460156016601760186019602060216022602360246025602660276028602960306031603260336034603560366037603860396040604160426043604460456046604760486049605060516052605360546055605660576058605960606061606260636064606560666067606860696070607160726073607460756076607760786079608060816082608360846085608660876088608960906091609260936094609560966097609860996100610161026103610461056106610761086109611061116112611361146115611661176118611961206121612261236124612561266127612861296130613161326133613461356136613761386139614061416142614361446145614661476148614961506151615261536154615561566157615861596160616161626163616461656166616761686169617061716172617361746175617661776178617961806181618261836184618561866187618861896190619161926193619461956196619761986199620062016202620362046205620662076208620962106211621262136214621562166217621862196220622162226223622462256226622762286229623062316232623362346235623662376238623962406241624262436244624562466247624862496250625162526253625462556256625762586259626062616262626362646265626662676268626962706271627262736274627562766277627862796280628162826283628462856286628762886289629062916292629362946295629662976298629963006301630263036304630563066307630863096310631163126313631463156316631763186319632063216322632363246325632663276328632963306331633263336334633563366337633863396340634163426343634463456346634763486349635063516352635363546355635663576358635963606361636263636364636563666367636863696370637163726373637463756376637763786379638063816382638363846385638663876388638963906391639263936394639563966397639863996400640164026403640464056406640764086409641064116412641364146415641664176418641964206421642264236424642564266427642864296430643164326433643464356436643764386439644064416442644364446445644664476448644964506451645264536454645564566457645864596460646164626463646464656466646764686469647064716472647364746475647664776478647964806481648264836484648564866487648864896490649164926493649464956496649764986499650065016502650365046505650665076508650965106511651265136514651565166517651865196520652165226523652465256526652765286529653065316532653365346535653665376538653965406541654265436544654565466547654865496550655165526553655465556556655765586559656065616562656365646565656665676568656965706571657265736574657565766577657865796580658165826583658465856586658765886589659065916592659365946595659665976598659966006601660266036604660566066607660866096610661166126613661466156616661766186619662066216622662366246625662666276628662966306631663266336634663566366637663866396640664166426643664466456646664766486649665066516652665366546655665666576658665966606661666266636664666566666667666866696670667166726673667466756676667766786679668066816682668366846685668666876688668966906691669266936694669566966697669866996700670167026703670467056706670767086709671067116712671367146715671667176718671967206721672267236724672567266727672867296730673167326733673467356736673767386739674067416742674367446745674667476748674967506751675267536754675567566757675867596760676167626763676467656766676767686769677067716772677367746775677667776778677967806781678267836784678567866787678867896790679167926793679467956796679767986799680068016802680368046805680668076808680968106811681268136814681568166817681868196820682168226823682468256826682768286829683068316832683368346835683668376838683968406841684268436844684568466847684868496850685168526853685468556856685768586859686068616862686368646865686668676868686968706871687268736874687568766877687868796880688168826883688468856886688768886889689068916892689368946895689668976898689969006901690269036904690569066907690869096910691169126913691469156916691769186919692069216922692369246925692669276928692969306931693269336934693569366937693869396940694169426943694469456946694769486949695069516952695369546955695669576958695969606961696269636964696569666967696869696970697169726973697469756976697769786979698069816982698369846985698669876988698969906991699269936994699569966997699869997000700170027003700470057006700770087009701070117012701370147015701670177018701970207021702270237024702570267027702870297030703170327033703470357036703770387039704070417042704370447045704670477048704970507051705270537054705570567057705870597060706170627063706470657066706770687069707070717072707370747075707670777078707970807081708270837084708570867087708870897090709170927093709470957096709770987099710071017102710371047105710671077108710971107111711271137114711571167117711871197120712171227123712471257126712771287129713071317132713371347135713671377138713971407141714271437144714571467147714871497150715171527153715471557156715771587159716071617162716371647165716671677168716971707171717271737174717571767177717871797180718171827183718471857186718771887189719071917192719371947195719671977198719972007201720272037204720572067207720872097210721172127213721472157216721772187219722072217222722372247225722672277228722972307231723272337234723572367237723872397240724172427243724472457246724772487249725072517252725372547255725672577258725972607261726272637264726572667267726872697270727172727273727472757276727772787279728072817282728372847285728672877288728972907291729272937294729572967297729872997300730173027303730473057306730773087309731073117312731373147315731673177318731973207321732273237324732573267327732873297330733173327333733473357336733773387339734073417342734373447345734673477348734973507351735273537354735573567357735873597360736173627363736473657366736773687369737073717372737373747375737673777378737973807381738273837384738573867387738873897390739173927393739473957396739773987399740074017402740374047405740674077408740974107411741274137414741574167417741874197420742174227423742474257426742774287429743074317432743374347435743674377438743974407441744274437444744574467447744874497450745174527453745474557456745774587459746074617462746374647465746674677468746974707471747274737474747574767477747874797480748174827483748474857486748774887489749074917492749374947495749674977498749975007501750275037504750575067507750875097510751175127513751475157516751775187519752075217522752375247525752675277528752975307531753275337534753575367537753875397540754175427543754475457546754775487549755075517552755375547555755675577558755975607561756275637564756575667567756875697570757175727573757475757576757775787579758075817582758375847585758675877588758975907591759275937594759575967597759875997600760176027603760476057606760776087609761076117612761376147615761676177618761976207621762276237624762576267627762876297630763176327633763476357636763776387639764076417642764376447645764676477648764976507651765276537654765576567657765876597660766176627663766476657666766776687669767076717672767376747675767676777678767976807681768276837684768576867687768876897690769176927693769476957696769776987699770077017702770377047705770677077708770977107711771277137714771577167717771877197720772177227723772477257726772777287729773077317732773377347735773677377738773977407741774277437744774577467747774877497750775177527753775477557756775777587759776077617762776377647765776677677768776977707771777277737774777577767777777877797780778177827783778477857786778777887789779077917792779377947795779677977798779978007801780278037804780578067807780878097810781178127813781478157816781778187819782078217822782378247825782678277828782978307831783278337834783578367837783878397840784178427843784478457846784778487849785078517852785378547855785678577858785978607861786278637864786578667867786878697870787178727873787478757876787778787879788078817882788378847885788678877888788978907891789278937894789578967897789878997900790179027903790479057906790779087909791079117912791379147915791679177918791979207921792279237924792579267927792879297930793179327933793479357936793779387939794079417942794379447945794679477948794979507951795279537954795579567957795879597960796179627963796479657966796779687969797079717972797379747975797679777978797979807981798279837984798579867987798879897990799179927993799479957996799779987999800080018002800380048005800680078008800980108011801280138014801580168017801880198020802180228023802480258026802780288029803080318032803380348035803680378038803980408041804280438044804580468047804880498050805180528053805480558056805780588059806080618062806380648065806680678068806980708071807280738074807580768077807880798080808180828083808480858086808780888089809080918092809380948095809680978098809981008101810281038104810581068107810881098110811181128113811481158116811781188119812081218122812381248125812681278128812981308131813281338134813581368137813881398140814181428143814481458146814781488149815081518152815381548155815681578158815981608161816281638164816581668167816881698170817181728173817481758176817781788179818081818182818381848185818681878188818981908191819281938194819581968197819881998200820182028203820482058206820782088209821082118212821382148215821682178218821982208221822282238224822582268227822882298230823182328233823482358236823782388239824082418242824382448245824682478248824982508251825282538254825582568257825882598260826182628263826482658266826782688269827082718272827382748275827682778278827982808281828282838284828582868287828882898290829182928293829482958296829782988299830083018302830383048305830683078308830983108311831283138314831583168317831883198320832183228323832483258326832783288329833083318332833383348335833683378338833983408341834283438344834583468347834883498350835183528353835483558356835783588359836083618362836383648365836683678368836983708371837283738374837583768377837883798380838183828383838483858386838783888389839083918392839383948395839683978398839984008401840284038404840584068407840884098410841184128413841484158416841784188419842084218422842384248425842684278428842984308431843284338434843584368437843884398440844184428443844484458446844784488449845084518452845384548455845684578458845984608461846284638464846584668467846884698470847184728473847484758476847784788479848084818482848384848485848684878488848984908491849284938494849584968497849884998500850185028503850485058506850785088509851085118512851385148515851685178518851985208521852285238524852585268527852885298530853185328533853485358536853785388539854085418542854385448545854685478548854985508551855285538554855585568557855885598560856185628563856485658566856785688569857085718572857385748575857685778578857985808581858285838584858585868587858885898590859185928593859485958596859785988599860086018602860386048605860686078608860986108611861286138614861586168617861886198620862186228623862486258626862786288629863086318632863386348635863686378638863986408641864286438644864586468647864886498650865186528653865486558656865786588659866086618662866386648665866686678668866986708671867286738674867586768677867886798680868186828683868486858686868786888689869086918692869386948695869686978698869987008701870287038704870587068707870887098710871187128713871487158716871787188719872087218722872387248725872687278728872987308731873287338734873587368737873887398740874187428743874487458746874787488749875087518752875387548755875687578758875987608761876287638764876587668767876887698770877187728773877487758776877787788779878087818782878387848785878687878788878987908791879287938794879587968797879887998800880188028803880488058806880788088809881088118812881388148815881688178818881988208821882288238824882588268827882888298830883188328833883488358836883788388839884088418842884388448845884688478848884988508851885288538854885588568857885888598860886188628863886488658866886788688869887088718872887388748875887688778878887988808881888288838884888588868887888888898890889188928893889488958896889788988899890089018902890389048905890689078908890989108911891289138914891589168917891889198920892189228923892489258926892789288929893089318932893389348935893689378938893989408941894289438944894589468947894889498950895189528953895489558956895789588959896089618962896389648965896689678968896989708971897289738974897589768977897889798980898189828983898489858986898789888989899089918992899389948995899689978998899990009001900290039004900590069007900890099010901190129013901490159016901790189019902090219022902390249025902690279028902990309031903290339034903590369037903890399040904190429043904490459046904790489049905090519052905390549055905690579058905990609061906290639064906590669067906890699070907190729073907490759076907790789079908090819082908390849085908690879088908990909091909290939094909590969097909890999100910191029103910491059106910791089109911091119112911391149115911691179118911991209121912291239124912591269127912891299130913191329133913491359136913791389139914091419142914391449145914691479148914991509151915291539154915591569157915891599160916191629163916491659166916791689169917091719172917391749175917691779178917991809181918291839184918591869187918891899190919191929193919491959196919791989199920092019202920392049205920692079208920992109211921292139214921592169217921892199220
  1. /*
  2. * Copyright 2011-2023 Branimir Karadzic. All rights reserved.
  3. * License: https://github.com/bkaradzic/bgfx/blob/master/LICENSE
  4. */
  5. #include "bgfx_p.h"
  6. #if BGFX_CONFIG_RENDERER_VULKAN
  7. # include <bx/pixelformat.h>
  8. # include "renderer_vk.h"
  9. # include "shader_spirv.h"
  10. #if BX_PLATFORM_OSX
  11. # import <Cocoa/Cocoa.h>
  12. # import <Foundation/Foundation.h>
  13. # import <QuartzCore/QuartzCore.h>
  14. # import <Metal/Metal.h>
  15. #endif // BX_PLATFORM_OSX
  16. #if defined(WL_EGL_PLATFORM)
  17. # include <wayland-egl-backend.h>
  18. #endif // defined(WL_EGL_PLATFORM)
  19. namespace bgfx { namespace vk
  20. {
  21. static char s_viewName[BGFX_CONFIG_MAX_VIEWS][BGFX_CONFIG_MAX_VIEW_NAME];
  22. inline void setViewType(ViewId _view, const bx::StringView _str)
  23. {
  24. if (BX_ENABLED(BGFX_CONFIG_DEBUG_ANNOTATION || BGFX_CONFIG_PROFILER) )
  25. {
  26. bx::memCopy(&s_viewName[_view][3], _str.getPtr(), _str.getLength() );
  27. }
  28. }
  29. struct PrimInfo
  30. {
  31. VkPrimitiveTopology m_topology;
  32. uint32_t m_min;
  33. uint32_t m_div;
  34. uint32_t m_sub;
  35. };
  36. static const PrimInfo s_primInfo[] =
  37. {
  38. { VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST, 3, 3, 0 },
  39. { VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP, 3, 1, 2 },
  40. { VK_PRIMITIVE_TOPOLOGY_LINE_LIST, 2, 2, 0 },
  41. { VK_PRIMITIVE_TOPOLOGY_LINE_STRIP, 2, 1, 1 },
  42. { VK_PRIMITIVE_TOPOLOGY_POINT_LIST, 1, 1, 0 },
  43. { VK_PRIMITIVE_TOPOLOGY_MAX_ENUM, 0, 0, 0 },
  44. };
  45. BX_STATIC_ASSERT(Topology::Count == BX_COUNTOF(s_primInfo)-1);
  46. static MsaaSamplerVK s_msaa[] =
  47. {
  48. { 1, VK_SAMPLE_COUNT_1_BIT },
  49. { 2, VK_SAMPLE_COUNT_2_BIT },
  50. { 4, VK_SAMPLE_COUNT_4_BIT },
  51. { 8, VK_SAMPLE_COUNT_8_BIT },
  52. { 16, VK_SAMPLE_COUNT_16_BIT },
  53. };
  54. static const VkBlendFactor s_blendFactor[][2] =
  55. {
  56. { VkBlendFactor(0), VkBlendFactor(0) }, // ignored
  57. { VK_BLEND_FACTOR_ZERO, VK_BLEND_FACTOR_ZERO }, // ZERO
  58. { VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_ONE }, // ONE
  59. { VK_BLEND_FACTOR_SRC_COLOR, VK_BLEND_FACTOR_SRC_ALPHA }, // SRC_COLOR
  60. { VK_BLEND_FACTOR_ONE_MINUS_SRC_COLOR, VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA }, // INV_SRC_COLOR
  61. { VK_BLEND_FACTOR_SRC_ALPHA, VK_BLEND_FACTOR_SRC_ALPHA }, // SRC_ALPHA
  62. { VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA, VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA }, // INV_SRC_ALPHA
  63. { VK_BLEND_FACTOR_DST_ALPHA, VK_BLEND_FACTOR_DST_ALPHA }, // DST_ALPHA
  64. { VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA, VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA }, // INV_DST_ALPHA
  65. { VK_BLEND_FACTOR_DST_COLOR, VK_BLEND_FACTOR_DST_ALPHA }, // DST_COLOR
  66. { VK_BLEND_FACTOR_ONE_MINUS_DST_COLOR, VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA }, // INV_DST_COLOR
  67. { VK_BLEND_FACTOR_SRC_ALPHA, VK_BLEND_FACTOR_ONE }, // SRC_ALPHA_SAT
  68. { VK_BLEND_FACTOR_CONSTANT_COLOR, VK_BLEND_FACTOR_CONSTANT_COLOR }, // FACTOR
  69. { VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR, VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR }, // INV_FACTOR
  70. };
  71. static const VkBlendOp s_blendEquation[] =
  72. {
  73. VK_BLEND_OP_ADD,
  74. VK_BLEND_OP_SUBTRACT,
  75. VK_BLEND_OP_REVERSE_SUBTRACT,
  76. VK_BLEND_OP_MIN,
  77. VK_BLEND_OP_MAX,
  78. };
  79. static const VkCompareOp s_cmpFunc[] =
  80. {
  81. VkCompareOp(0), // ignored
  82. VK_COMPARE_OP_LESS,
  83. VK_COMPARE_OP_LESS_OR_EQUAL,
  84. VK_COMPARE_OP_EQUAL,
  85. VK_COMPARE_OP_GREATER_OR_EQUAL,
  86. VK_COMPARE_OP_GREATER,
  87. VK_COMPARE_OP_NOT_EQUAL,
  88. VK_COMPARE_OP_NEVER,
  89. VK_COMPARE_OP_ALWAYS,
  90. };
  91. static const VkStencilOp s_stencilOp[] =
  92. {
  93. VK_STENCIL_OP_ZERO,
  94. VK_STENCIL_OP_KEEP,
  95. VK_STENCIL_OP_REPLACE,
  96. VK_STENCIL_OP_INCREMENT_AND_WRAP,
  97. VK_STENCIL_OP_INCREMENT_AND_CLAMP,
  98. VK_STENCIL_OP_DECREMENT_AND_WRAP,
  99. VK_STENCIL_OP_DECREMENT_AND_CLAMP,
  100. VK_STENCIL_OP_INVERT,
  101. };
  102. static const VkCullModeFlagBits s_cullMode[] =
  103. {
  104. VK_CULL_MODE_NONE,
  105. VK_CULL_MODE_FRONT_BIT,
  106. VK_CULL_MODE_BACK_BIT,
  107. };
  108. static const VkSamplerAddressMode s_textureAddress[] =
  109. {
  110. VK_SAMPLER_ADDRESS_MODE_REPEAT,
  111. VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT,
  112. VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE,
  113. VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER,
  114. };
  115. struct PresentMode
  116. {
  117. VkPresentModeKHR mode;
  118. bool vsync;
  119. const char* name;
  120. };
  121. static const PresentMode s_presentMode[] =
  122. {
  123. { VK_PRESENT_MODE_FIFO_KHR, true, "VK_PRESENT_MODE_FIFO_KHR" },
  124. { VK_PRESENT_MODE_FIFO_RELAXED_KHR, true, "VK_PRESENT_MODE_FIFO_RELAXED_KHR" },
  125. { VK_PRESENT_MODE_MAILBOX_KHR, true, "VK_PRESENT_MODE_MAILBOX_KHR" },
  126. { VK_PRESENT_MODE_IMMEDIATE_KHR, false, "VK_PRESENT_MODE_IMMEDIATE_KHR" },
  127. };
  128. #define VK_IMPORT_FUNC(_optional, _func) PFN_##_func _func
  129. #define VK_IMPORT_INSTANCE_FUNC VK_IMPORT_FUNC
  130. #define VK_IMPORT_DEVICE_FUNC VK_IMPORT_FUNC
  131. VK_IMPORT
  132. VK_IMPORT_INSTANCE
  133. VK_IMPORT_DEVICE
  134. #undef VK_IMPORT_DEVICE_FUNC
  135. #undef VK_IMPORT_INSTANCE_FUNC
  136. #undef VK_IMPORT_FUNC
  137. struct TextureFormatInfo
  138. {
  139. VkFormat m_fmt;
  140. VkFormat m_fmtSrv;
  141. VkFormat m_fmtDsv;
  142. VkFormat m_fmtSrgb;
  143. VkComponentMapping m_mapping;
  144. };
  145. static const TextureFormatInfo s_textureFormat[] =
  146. {
  147. #define $_ VK_COMPONENT_SWIZZLE_IDENTITY
  148. #define $0 VK_COMPONENT_SWIZZLE_ZERO
  149. #define $1 VK_COMPONENT_SWIZZLE_ONE
  150. #define $R VK_COMPONENT_SWIZZLE_R
  151. #define $G VK_COMPONENT_SWIZZLE_G
  152. #define $B VK_COMPONENT_SWIZZLE_B
  153. #define $A VK_COMPONENT_SWIZZLE_A
  154. { VK_FORMAT_BC1_RGB_UNORM_BLOCK, VK_FORMAT_BC1_RGB_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC1_RGB_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC1
  155. { VK_FORMAT_BC2_UNORM_BLOCK, VK_FORMAT_BC2_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC2_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC2
  156. { VK_FORMAT_BC3_UNORM_BLOCK, VK_FORMAT_BC3_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC3_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC3
  157. { VK_FORMAT_BC4_UNORM_BLOCK, VK_FORMAT_BC4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC4
  158. { VK_FORMAT_BC5_UNORM_BLOCK, VK_FORMAT_BC5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC5
  159. { VK_FORMAT_BC6H_SFLOAT_BLOCK, VK_FORMAT_BC6H_SFLOAT_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BC6H
  160. { VK_FORMAT_BC7_UNORM_BLOCK, VK_FORMAT_BC7_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_BC7_SRGB_BLOCK, { $_, $_, $_, $_ } }, // BC7
  161. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ETC1
  162. { VK_FORMAT_ETC2_R8G8B8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2
  163. { VK_FORMAT_ETC2_R8G8B8A8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8A8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2A
  164. { VK_FORMAT_ETC2_R8G8B8A1_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_ETC2_R8G8B8A1_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ETC2A1
  165. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC12
  166. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC14
  167. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC12A
  168. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC14A
  169. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC22
  170. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // PTC24
  171. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATC
  172. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATCE
  173. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // ATCI
  174. { VK_FORMAT_ASTC_4x4_UNORM_BLOCK, VK_FORMAT_ASTC_4x4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_4x4_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC4x4
  175. { VK_FORMAT_ASTC_5x4_UNORM_BLOCK, VK_FORMAT_ASTC_5x4_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_5x4_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC5x4
  176. { VK_FORMAT_ASTC_5x5_UNORM_BLOCK, VK_FORMAT_ASTC_5x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_5x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC5x5
  177. { VK_FORMAT_ASTC_6x5_UNORM_BLOCK, VK_FORMAT_ASTC_6x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_6x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC6x5
  178. { VK_FORMAT_ASTC_6x6_UNORM_BLOCK, VK_FORMAT_ASTC_6x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_6x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC6x6
  179. { VK_FORMAT_ASTC_8x5_UNORM_BLOCK, VK_FORMAT_ASTC_8x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x5
  180. { VK_FORMAT_ASTC_8x6_UNORM_BLOCK, VK_FORMAT_ASTC_8x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x6
  181. { VK_FORMAT_ASTC_8x8_UNORM_BLOCK, VK_FORMAT_ASTC_8x8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_8x8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC8x8
  182. { VK_FORMAT_ASTC_10x5_UNORM_BLOCK, VK_FORMAT_ASTC_10x5_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x5_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x5
  183. { VK_FORMAT_ASTC_10x6_UNORM_BLOCK, VK_FORMAT_ASTC_10x6_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x6_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x6
  184. { VK_FORMAT_ASTC_10x8_UNORM_BLOCK, VK_FORMAT_ASTC_10x8_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x8_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x8
  185. { VK_FORMAT_ASTC_10x10_UNORM_BLOCK, VK_FORMAT_ASTC_10x10_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_10x10_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC10x10
  186. { VK_FORMAT_ASTC_12x10_UNORM_BLOCK, VK_FORMAT_ASTC_12x10_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_12x10_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC12x10
  187. { VK_FORMAT_ASTC_12x12_UNORM_BLOCK, VK_FORMAT_ASTC_12x12_UNORM_BLOCK, VK_FORMAT_UNDEFINED, VK_FORMAT_ASTC_12x12_SRGB_BLOCK, { $_, $_, $_, $_ } }, // ASTC12x12
  188. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // Unknown
  189. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R1
  190. { VK_FORMAT_R8_UNORM, VK_FORMAT_R8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $0, $0, $0, $R } }, // A8
  191. { VK_FORMAT_R8_UNORM, VK_FORMAT_R8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8_SRGB, { $_, $_, $_, $_ } }, // R8
  192. { VK_FORMAT_R8_SINT, VK_FORMAT_R8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8I
  193. { VK_FORMAT_R8_UINT, VK_FORMAT_R8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8U
  194. { VK_FORMAT_R8_SNORM, VK_FORMAT_R8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R8S
  195. { VK_FORMAT_R16_UNORM, VK_FORMAT_R16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16
  196. { VK_FORMAT_R16_SINT, VK_FORMAT_R16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16I
  197. { VK_FORMAT_R16_UINT, VK_FORMAT_R16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16U
  198. { VK_FORMAT_R16_SFLOAT, VK_FORMAT_R16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16F
  199. { VK_FORMAT_R16_SNORM, VK_FORMAT_R16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R16S
  200. { VK_FORMAT_R32_SINT, VK_FORMAT_R32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32I
  201. { VK_FORMAT_R32_UINT, VK_FORMAT_R32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32U
  202. { VK_FORMAT_R32_SFLOAT, VK_FORMAT_R32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R32F
  203. { VK_FORMAT_R8G8_UNORM, VK_FORMAT_R8G8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8_SRGB, { $_, $_, $_, $_ } }, // RG8
  204. { VK_FORMAT_R8G8_SINT, VK_FORMAT_R8G8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8I
  205. { VK_FORMAT_R8G8_UINT, VK_FORMAT_R8G8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8U
  206. { VK_FORMAT_R8G8_SNORM, VK_FORMAT_R8G8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG8S
  207. { VK_FORMAT_R16G16_UNORM, VK_FORMAT_R16G16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16
  208. { VK_FORMAT_R16G16_SINT, VK_FORMAT_R16G16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16I
  209. { VK_FORMAT_R16G16_UINT, VK_FORMAT_R16G16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16U
  210. { VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16F
  211. { VK_FORMAT_R16G16_SNORM, VK_FORMAT_R16G16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG16S
  212. { VK_FORMAT_R32G32_SINT, VK_FORMAT_R32G32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32I
  213. { VK_FORMAT_R32G32_UINT, VK_FORMAT_R32G32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32U
  214. { VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG32F
  215. { VK_FORMAT_R8G8B8_UNORM, VK_FORMAT_R8G8B8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8
  216. { VK_FORMAT_R8G8B8_SINT, VK_FORMAT_R8G8B8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8I
  217. { VK_FORMAT_R8G8B8_UINT, VK_FORMAT_R8G8B8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8_SRGB, { $_, $_, $_, $_ } }, // RGB8U
  218. { VK_FORMAT_R8G8B8_SNORM, VK_FORMAT_R8G8B8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGB8S
  219. { VK_FORMAT_E5B9G9R9_UFLOAT_PACK32, VK_FORMAT_E5B9G9R9_UFLOAT_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGB9E5F
  220. { VK_FORMAT_B8G8R8A8_UNORM, VK_FORMAT_B8G8R8A8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_B8G8R8A8_SRGB, { $_, $_, $_, $_ } }, // BGRA8
  221. { VK_FORMAT_R8G8B8A8_UNORM, VK_FORMAT_R8G8B8A8_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8
  222. { VK_FORMAT_R8G8B8A8_SINT, VK_FORMAT_R8G8B8A8_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8I
  223. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_R8G8B8A8_SRGB, { $_, $_, $_, $_ } }, // RGBA8U
  224. { VK_FORMAT_R8G8B8A8_SNORM, VK_FORMAT_R8G8B8A8_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA8S
  225. { VK_FORMAT_R16G16B16A16_UNORM, VK_FORMAT_R16G16B16A16_UNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16
  226. { VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16I
  227. { VK_FORMAT_R16G16B16A16_UINT, VK_FORMAT_R16G16B16A16_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16U
  228. { VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16F
  229. { VK_FORMAT_R16G16B16A16_SNORM, VK_FORMAT_R16G16B16A16_SNORM, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA16S
  230. { VK_FORMAT_R32G32B32A32_SINT, VK_FORMAT_R32G32B32A32_SINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32I
  231. { VK_FORMAT_R32G32B32A32_UINT, VK_FORMAT_R32G32B32A32_UINT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32U
  232. { VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RGBA32F
  233. { VK_FORMAT_R5G6B5_UNORM_PACK16, VK_FORMAT_R5G6B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // B5G6R5
  234. { VK_FORMAT_B5G6R5_UNORM_PACK16, VK_FORMAT_B5G6R5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // R5G6B5
  235. { VK_FORMAT_B4G4R4A4_UNORM_PACK16, VK_FORMAT_B4G4R4A4_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $G, $R, $A, $B } }, // BGRA4
  236. { VK_FORMAT_R4G4B4A4_UNORM_PACK16, VK_FORMAT_R4G4B4A4_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $A, $B, $G, $R } }, // RGBA4
  237. { VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // BGR5A1
  238. { VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_A1R5G5B5_UNORM_PACK16, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $B, $G, $R, $A } }, // RGB5A1
  239. { VK_FORMAT_A2R10G10B10_UNORM_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $B, $G, $R, $A } }, // RGB10A2
  240. { VK_FORMAT_B10G11R11_UFLOAT_PACK32, VK_FORMAT_B10G11R11_UFLOAT_PACK32, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // RG11B10F
  241. { VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // UnknownDepth
  242. { VK_FORMAT_UNDEFINED, VK_FORMAT_R16_UNORM, VK_FORMAT_D16_UNORM, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D16
  243. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24
  244. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24S8
  245. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D32
  246. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D16F
  247. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D24F
  248. { VK_FORMAT_UNDEFINED, VK_FORMAT_R32_SFLOAT, VK_FORMAT_D32_SFLOAT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D32F
  249. { VK_FORMAT_UNDEFINED, VK_FORMAT_R8_UINT, VK_FORMAT_S8_UINT, VK_FORMAT_UNDEFINED, { $_, $_, $_, $_ } }, // D0S8
  250. #undef $_
  251. #undef $0
  252. #undef $1
  253. #undef $R
  254. #undef $G
  255. #undef $B
  256. #undef $A
  257. };
  258. BX_STATIC_ASSERT(TextureFormat::Count == BX_COUNTOF(s_textureFormat) );
  259. struct LayerInfo
  260. {
  261. bool m_supported;
  262. bool m_initialize;
  263. };
  264. struct Layer
  265. {
  266. enum Enum
  267. {
  268. VK_LAYER_LUNARG_standard_validation,
  269. VK_LAYER_KHRONOS_validation,
  270. Count
  271. };
  272. const char* m_name;
  273. uint32_t m_minVersion;
  274. LayerInfo m_instance;
  275. LayerInfo m_device;
  276. };
  277. // Layer registry
  278. //
  279. static Layer s_layer[] =
  280. {
  281. { "VK_LAYER_LUNARG_standard_validation", 1, { false, false }, { false, false } },
  282. { "VK_LAYER_KHRONOS_validation", 1, { false, false }, { false, false } },
  283. { "", 0, { false, false }, { false, false } },
  284. };
  285. BX_STATIC_ASSERT(Layer::Count == BX_COUNTOF(s_layer)-1);
  286. void updateLayer(const char* _name, uint32_t _version, bool _instanceLayer)
  287. {
  288. bx::StringView layerName(_name);
  289. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  290. {
  291. Layer& layer = s_layer[ii];
  292. LayerInfo& layerInfo = _instanceLayer
  293. ? layer.m_instance
  294. : layer.m_device
  295. ;
  296. if (!layerInfo.m_supported && layerInfo.m_initialize)
  297. {
  298. if ( 0 == bx::strCmp(layerName, layer.m_name)
  299. && _version >= layer.m_minVersion)
  300. {
  301. layerInfo.m_supported = true;
  302. break;
  303. }
  304. }
  305. }
  306. }
  307. struct Extension
  308. {
  309. enum Enum
  310. {
  311. EXT_debug_utils,
  312. EXT_debug_report,
  313. EXT_memory_budget,
  314. KHR_get_physical_device_properties2,
  315. EXT_conservative_rasterization,
  316. EXT_line_rasterization,
  317. EXT_shader_viewport_index_layer,
  318. EXT_custom_border_color,
  319. KHR_draw_indirect_count,
  320. Count
  321. };
  322. const char* m_name;
  323. uint32_t m_minVersion;
  324. bool m_instanceExt;
  325. bool m_supported;
  326. bool m_initialize;
  327. Layer::Enum m_layer;
  328. };
  329. // Extension registry
  330. //
  331. static Extension s_extension[] =
  332. {
  333. { "VK_EXT_debug_utils", 1, false, false, BGFX_CONFIG_DEBUG_OBJECT_NAME || BGFX_CONFIG_DEBUG_ANNOTATION, Layer::Count },
  334. { "VK_EXT_debug_report", 1, false, false, false , Layer::Count },
  335. { "VK_EXT_memory_budget", 1, false, false, true , Layer::Count },
  336. { "VK_KHR_get_physical_device_properties2", 1, false, false, true , Layer::Count },
  337. { "VK_EXT_conservative_rasterization", 1, false, false, true , Layer::Count },
  338. { "VK_EXT_line_rasterization", 1, false, false, true , Layer::Count },
  339. { "VK_EXT_shader_viewport_index_layer", 1, false, false, true , Layer::Count },
  340. { "VK_EXT_custom_border_color", 1, false, false, true , Layer::Count },
  341. { "VK_KHR_draw_indirect_count", 1, false, false, true , Layer::Count },
  342. };
  343. BX_STATIC_ASSERT(Extension::Count == BX_COUNTOF(s_extension) );
  344. bool updateExtension(const char* _name, uint32_t _version, bool _instanceExt, Extension _extensions[Extension::Count])
  345. {
  346. bool supported = false;
  347. if (BX_ENABLED(BGFX_CONFIG_RENDERER_USE_EXTENSIONS) )
  348. {
  349. const bx::StringView ext(_name);
  350. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  351. {
  352. Extension& extension = _extensions[ii];
  353. const LayerInfo& layerInfo = _instanceExt
  354. ? s_layer[extension.m_layer].m_instance
  355. : s_layer[extension.m_layer].m_device
  356. ;
  357. if (!extension.m_supported
  358. && extension.m_initialize
  359. && (extension.m_layer == Layer::Count || layerInfo.m_supported) )
  360. {
  361. if ( 0 == bx::strCmp(ext, extension.m_name)
  362. && _version >= extension.m_minVersion)
  363. {
  364. extension.m_supported = true;
  365. extension.m_instanceExt = _instanceExt;
  366. supported = true;
  367. break;
  368. }
  369. }
  370. }
  371. }
  372. return supported;
  373. }
  374. static const VkFormat s_attribType[][4][2] =
  375. {
  376. { // Uint8
  377. { VK_FORMAT_R8_UINT, VK_FORMAT_R8_UNORM },
  378. { VK_FORMAT_R8G8_UINT, VK_FORMAT_R8G8_UNORM },
  379. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UNORM },
  380. { VK_FORMAT_R8G8B8A8_UINT, VK_FORMAT_R8G8B8A8_UNORM },
  381. },
  382. { // Uint10
  383. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  384. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  385. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  386. { VK_FORMAT_A2R10G10B10_UINT_PACK32, VK_FORMAT_A2R10G10B10_UNORM_PACK32 },
  387. },
  388. { // Int16
  389. { VK_FORMAT_R16_SINT, VK_FORMAT_R16_SNORM },
  390. { VK_FORMAT_R16G16_SINT, VK_FORMAT_R16G16_SNORM },
  391. { VK_FORMAT_R16G16B16_SINT, VK_FORMAT_R16G16B16_SNORM },
  392. { VK_FORMAT_R16G16B16A16_SINT, VK_FORMAT_R16G16B16A16_SNORM },
  393. },
  394. { // Half
  395. { VK_FORMAT_R16_SFLOAT, VK_FORMAT_R16_SFLOAT },
  396. { VK_FORMAT_R16G16_SFLOAT, VK_FORMAT_R16G16_SFLOAT },
  397. { VK_FORMAT_R16G16B16_SFLOAT, VK_FORMAT_R16G16B16_SFLOAT },
  398. { VK_FORMAT_R16G16B16A16_SFLOAT, VK_FORMAT_R16G16B16A16_SFLOAT },
  399. },
  400. { // Float
  401. { VK_FORMAT_R32_SFLOAT, VK_FORMAT_R32_SFLOAT },
  402. { VK_FORMAT_R32G32_SFLOAT, VK_FORMAT_R32G32_SFLOAT },
  403. { VK_FORMAT_R32G32B32_SFLOAT, VK_FORMAT_R32G32B32_SFLOAT },
  404. { VK_FORMAT_R32G32B32A32_SFLOAT, VK_FORMAT_R32G32B32A32_SFLOAT },
  405. },
  406. };
  407. BX_STATIC_ASSERT(AttribType::Count == BX_COUNTOF(s_attribType) );
  408. void fillVertexLayout(const ShaderVK* _vsh, VkPipelineVertexInputStateCreateInfo& _vertexInputState, const VertexLayout& _layout)
  409. {
  410. uint32_t numBindings = _vertexInputState.vertexBindingDescriptionCount;
  411. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  412. VkVertexInputBindingDescription* inputBinding = const_cast<VkVertexInputBindingDescription*>(_vertexInputState.pVertexBindingDescriptions + numBindings);
  413. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  414. inputBinding->binding = numBindings;
  415. inputBinding->stride = _layout.m_stride;
  416. inputBinding->inputRate = VK_VERTEX_INPUT_RATE_VERTEX;
  417. for (uint32_t attr = 0; attr < Attrib::Count; ++attr)
  418. {
  419. if (UINT16_MAX != _layout.m_attributes[attr])
  420. {
  421. inputAttrib->location = _vsh->m_attrRemap[attr];
  422. inputAttrib->binding = numBindings;
  423. uint8_t num;
  424. AttribType::Enum type;
  425. bool normalized;
  426. bool asInt;
  427. _layout.decode(Attrib::Enum(attr), num, type, normalized, asInt);
  428. inputAttrib->format = s_attribType[type][num-1][normalized];
  429. inputAttrib->offset = _layout.m_offset[attr];
  430. ++inputAttrib;
  431. ++numAttribs;
  432. }
  433. }
  434. _vertexInputState.vertexBindingDescriptionCount = numBindings + 1;
  435. _vertexInputState.vertexAttributeDescriptionCount = numAttribs;
  436. }
  437. void fillInstanceBinding(const ShaderVK* _vsh, VkPipelineVertexInputStateCreateInfo& _vertexInputState, uint32_t _numInstanceData)
  438. {
  439. BX_UNUSED(_vsh);
  440. uint32_t numBindings = _vertexInputState.vertexBindingDescriptionCount;
  441. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  442. VkVertexInputBindingDescription* inputBinding = const_cast<VkVertexInputBindingDescription*>(_vertexInputState.pVertexBindingDescriptions + numBindings);
  443. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  444. inputBinding->binding = numBindings;
  445. inputBinding->stride = _numInstanceData * 16;
  446. inputBinding->inputRate = VK_VERTEX_INPUT_RATE_INSTANCE;
  447. for (uint32_t inst = 0; inst < _numInstanceData; ++inst)
  448. {
  449. inputAttrib->location = numAttribs;
  450. inputAttrib->binding = numBindings;
  451. inputAttrib->format = VK_FORMAT_R32G32B32A32_SFLOAT;
  452. inputAttrib->offset = inst * 16;
  453. ++numAttribs;
  454. ++inputAttrib;
  455. }
  456. _vertexInputState.vertexBindingDescriptionCount = numBindings + 1;
  457. _vertexInputState.vertexAttributeDescriptionCount = numAttribs;
  458. }
  459. static const char* s_deviceTypeName[] =
  460. {
  461. "Other",
  462. "Integrated GPU",
  463. "Discrete GPU",
  464. "Virtual GPU",
  465. "CPU",
  466. "Unknown?!"
  467. };
  468. const char* getName(VkPhysicalDeviceType _type)
  469. {
  470. return s_deviceTypeName[bx::min<int32_t>(_type, BX_COUNTOF(s_deviceTypeName)-1 )];
  471. }
  472. static const char* s_allocScopeName[] =
  473. {
  474. "vkCommand",
  475. "vkObject",
  476. "vkCache",
  477. "vkDevice",
  478. "vkInstance",
  479. };
  480. BX_STATIC_ASSERT(VK_SYSTEM_ALLOCATION_SCOPE_INSTANCE == BX_COUNTOF(s_allocScopeName)-1);
  481. constexpr size_t kMinAlignment = 8;
  482. static void* VKAPI_PTR allocationFunction(void* _userData, size_t _size, size_t _alignment, VkSystemAllocationScope _allocationScope)
  483. {
  484. BX_UNUSED(_userData);
  485. return bx::alignedAlloc(g_allocator, _size, bx::max(kMinAlignment, _alignment), bx::Location(s_allocScopeName[_allocationScope], 0) );
  486. }
  487. static void* VKAPI_PTR reallocationFunction(void* _userData, void* _original, size_t _size, size_t _alignment, VkSystemAllocationScope _allocationScope)
  488. {
  489. BX_UNUSED(_userData);
  490. if (_size == 0) {
  491. bx::alignedFree(g_allocator, _original, 0);
  492. return NULL;
  493. }
  494. return bx::alignedRealloc(g_allocator, _original, _size, bx::max(kMinAlignment, _alignment), bx::Location(s_allocScopeName[_allocationScope], 0) );
  495. }
  496. static void VKAPI_PTR freeFunction(void* _userData, void* _memory)
  497. {
  498. BX_UNUSED(_userData);
  499. if (NULL == _memory)
  500. {
  501. return;
  502. }
  503. bx::alignedFree(g_allocator, _memory, 0);
  504. }
  505. static void VKAPI_PTR internalAllocationNotification(void* _userData, size_t _size, VkInternalAllocationType _allocationType, VkSystemAllocationScope _allocationScope)
  506. {
  507. BX_UNUSED(_userData, _size, _allocationType, _allocationScope);
  508. }
  509. static void VKAPI_PTR internalFreeNotification(void* _userData, size_t _size, VkInternalAllocationType _allocationType, VkSystemAllocationScope _allocationScope)
  510. {
  511. BX_UNUSED(_userData, _size, _allocationType, _allocationScope);
  512. }
  513. static VkAllocationCallbacks s_allocationCb =
  514. {
  515. NULL,
  516. allocationFunction,
  517. reallocationFunction,
  518. freeFunction,
  519. internalAllocationNotification,
  520. internalFreeNotification,
  521. };
  522. VkResult VKAPI_PTR stubSetDebugUtilsObjectNameEXT(VkDevice _device, const VkDebugUtilsObjectNameInfoEXT* _nameInfo)
  523. {
  524. BX_UNUSED(_device, _nameInfo);
  525. return VK_SUCCESS;
  526. }
  527. void VKAPI_PTR stubCmdInsertDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer, const VkDebugUtilsLabelEXT* _labelInfo)
  528. {
  529. BX_UNUSED(_commandBuffer, _labelInfo);
  530. }
  531. void VKAPI_PTR stubCmdBeginDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer, const VkDebugUtilsLabelEXT* _labelInfo)
  532. {
  533. BX_UNUSED(_commandBuffer, _labelInfo);
  534. }
  535. void VKAPI_PTR stubCmdEndDebugUtilsLabelEXT(VkCommandBuffer _commandBuffer)
  536. {
  537. BX_UNUSED(_commandBuffer);
  538. }
  539. static const char* s_debugReportObjectType[] =
  540. {
  541. "Unknown",
  542. "Instance",
  543. "PhysicalDevice",
  544. "Device",
  545. "Queue",
  546. "Semaphore",
  547. "CommandBuffer",
  548. "Fence",
  549. "DeviceMemory",
  550. "Buffer",
  551. "Image",
  552. "Event",
  553. "QueryPool",
  554. "BufferView",
  555. "ImageView",
  556. "ShaderModule",
  557. "PipelineCache",
  558. "PipelineLayout",
  559. "RenderPass",
  560. "Pipeline",
  561. "DescriptorSetLayout",
  562. "Sampler",
  563. "DescriptorPool",
  564. "DescriptorSet",
  565. "Framebuffer",
  566. "CommandPool",
  567. "SurfaceKHR",
  568. "SwapchainKHR",
  569. "DebugReport",
  570. };
  571. VkBool32 VKAPI_PTR debugReportCb(
  572. VkDebugReportFlagsEXT _flags,
  573. VkDebugReportObjectTypeEXT _objectType,
  574. uint64_t _object,
  575. size_t _location,
  576. int32_t _messageCode,
  577. const char* _layerPrefix,
  578. const char* _message,
  579. void* _userData
  580. )
  581. {
  582. BX_UNUSED(_flags
  583. , _objectType
  584. , _object
  585. , _location
  586. , _messageCode
  587. , _layerPrefix
  588. , _message
  589. , _userData
  590. , s_debugReportObjectType
  591. );
  592. // For more info about 'VUID-VkSwapchainCreateInfoKHR-imageExtent-01274'
  593. // check https://github.com/KhronosGroup/Vulkan-Docs/issues/1144
  594. if (!bx::strFind(_message, "PointSizeMissing").isEmpty()
  595. || !bx::strFind(_message, "SwapchainTooManyImages").isEmpty()
  596. || !bx::strFind(_message, "SwapchainImageNotAcquired").isEmpty()
  597. || !bx::strFind(_message, "VUID-VkSwapchainCreateInfoKHR-imageExtent-01274").isEmpty() )
  598. {
  599. return VK_FALSE;
  600. }
  601. BX_TRACE("%c%c%c%c%c %19s, %s, %d: %s"
  602. , 0 != (_flags & VK_DEBUG_REPORT_INFORMATION_BIT_EXT ) ? 'I' : '-'
  603. , 0 != (_flags & VK_DEBUG_REPORT_WARNING_BIT_EXT ) ? 'W' : '-'
  604. , 0 != (_flags & VK_DEBUG_REPORT_PERFORMANCE_WARNING_BIT_EXT) ? 'P' : '-'
  605. , 0 != (_flags & VK_DEBUG_REPORT_ERROR_BIT_EXT ) ? 'E' : '-'
  606. , 0 != (_flags & VK_DEBUG_REPORT_DEBUG_BIT_EXT ) ? 'D' : '-'
  607. , s_debugReportObjectType[_objectType]
  608. , _layerPrefix
  609. , _messageCode
  610. , _message
  611. );
  612. return VK_FALSE;
  613. }
  614. VkResult enumerateLayerProperties(VkPhysicalDevice _physicalDevice, uint32_t* _propertyCount, VkLayerProperties* _properties)
  615. {
  616. return (VK_NULL_HANDLE == _physicalDevice)
  617. ? vkEnumerateInstanceLayerProperties(_propertyCount, _properties)
  618. : vkEnumerateDeviceLayerProperties(_physicalDevice, _propertyCount, _properties)
  619. ;
  620. }
  621. VkResult enumerateExtensionProperties(VkPhysicalDevice _physicalDevice, const char* _layerName, uint32_t* _propertyCount, VkExtensionProperties* _properties)
  622. {
  623. return (VK_NULL_HANDLE == _physicalDevice)
  624. ? vkEnumerateInstanceExtensionProperties(_layerName, _propertyCount, _properties)
  625. : vkEnumerateDeviceExtensionProperties(_physicalDevice, _layerName, _propertyCount, _properties)
  626. ;
  627. }
  628. void dumpExtensions(VkPhysicalDevice _physicalDevice, Extension _extensions[Extension::Count])
  629. {
  630. { // Global extensions.
  631. uint32_t numExtensionProperties;
  632. VkResult result = enumerateExtensionProperties(_physicalDevice
  633. , NULL
  634. , &numExtensionProperties
  635. , NULL
  636. );
  637. if (VK_SUCCESS == result
  638. && 0 < numExtensionProperties)
  639. {
  640. VkExtensionProperties* extensionProperties = (VkExtensionProperties*)bx::alloc(g_allocator, numExtensionProperties * sizeof(VkExtensionProperties) );
  641. result = enumerateExtensionProperties(_physicalDevice
  642. , NULL
  643. , &numExtensionProperties
  644. , extensionProperties
  645. );
  646. BX_TRACE("Global extensions (%d):"
  647. , numExtensionProperties
  648. );
  649. for (uint32_t extension = 0; extension < numExtensionProperties; ++extension)
  650. {
  651. bool supported = updateExtension(
  652. extensionProperties[extension].extensionName
  653. , extensionProperties[extension].specVersion
  654. , VK_NULL_HANDLE == _physicalDevice
  655. , _extensions
  656. );
  657. BX_TRACE("\tv%-3d %s%s"
  658. , extensionProperties[extension].specVersion
  659. , extensionProperties[extension].extensionName
  660. , supported ? " (supported)" : "", extensionProperties[extension].extensionName
  661. );
  662. BX_UNUSED(supported);
  663. }
  664. bx::free(g_allocator, extensionProperties);
  665. }
  666. }
  667. // Layer extensions.
  668. uint32_t numLayerProperties;
  669. VkResult result = enumerateLayerProperties(_physicalDevice, &numLayerProperties, NULL);
  670. if (VK_SUCCESS == result
  671. && 0 < numLayerProperties)
  672. {
  673. VkLayerProperties* layerProperties = (VkLayerProperties*)bx::alloc(g_allocator, numLayerProperties * sizeof(VkLayerProperties) );
  674. result = enumerateLayerProperties(_physicalDevice, &numLayerProperties, layerProperties);
  675. char indent = VK_NULL_HANDLE == _physicalDevice ? '\0' : '\t';
  676. BX_UNUSED(indent);
  677. BX_TRACE("%cLayer extensions (%d):"
  678. , indent
  679. , numLayerProperties
  680. );
  681. for (uint32_t layer = 0; layer < numLayerProperties; ++layer)
  682. {
  683. updateLayer(
  684. layerProperties[layer].layerName
  685. , layerProperties[layer].implementationVersion
  686. , VK_NULL_HANDLE == _physicalDevice
  687. );
  688. BX_TRACE("%c\t%s (s: 0x%08x, i: 0x%08x), %s"
  689. , indent
  690. , layerProperties[layer].layerName
  691. , layerProperties[layer].specVersion
  692. , layerProperties[layer].implementationVersion
  693. , layerProperties[layer].description
  694. );
  695. uint32_t numExtensionProperties;
  696. result = enumerateExtensionProperties(_physicalDevice
  697. , layerProperties[layer].layerName
  698. , &numExtensionProperties
  699. , NULL
  700. );
  701. if (VK_SUCCESS == result
  702. && 0 < numExtensionProperties)
  703. {
  704. VkExtensionProperties* extensionProperties = (VkExtensionProperties*)bx::alloc(g_allocator, numExtensionProperties * sizeof(VkExtensionProperties) );
  705. result = enumerateExtensionProperties(_physicalDevice
  706. , layerProperties[layer].layerName
  707. , &numExtensionProperties
  708. , extensionProperties
  709. );
  710. for (uint32_t extension = 0; extension < numExtensionProperties; ++extension)
  711. {
  712. bool supported = updateExtension(
  713. extensionProperties[extension].extensionName
  714. , extensionProperties[extension].specVersion
  715. , VK_NULL_HANDLE == _physicalDevice
  716. , _extensions
  717. );
  718. BX_TRACE("%c\t\t%s (s: 0x%08x)"
  719. , indent
  720. , extensionProperties[extension].extensionName
  721. , extensionProperties[extension].specVersion
  722. , supported ? " (supported)" : "", extensionProperties[extension].extensionName
  723. );
  724. BX_UNUSED(supported);
  725. }
  726. bx::free(g_allocator, extensionProperties);
  727. }
  728. }
  729. bx::free(g_allocator, layerProperties);
  730. }
  731. }
  732. const char* getName(VkResult _result)
  733. {
  734. switch (_result)
  735. {
  736. #define VKENUM(_ty) case _ty: return #_ty
  737. VKENUM(VK_SUCCESS);
  738. VKENUM(VK_NOT_READY);
  739. VKENUM(VK_TIMEOUT);
  740. VKENUM(VK_EVENT_SET);
  741. VKENUM(VK_EVENT_RESET);
  742. VKENUM(VK_INCOMPLETE);
  743. VKENUM(VK_ERROR_OUT_OF_HOST_MEMORY);
  744. VKENUM(VK_ERROR_OUT_OF_DEVICE_MEMORY);
  745. VKENUM(VK_ERROR_OUT_OF_POOL_MEMORY);
  746. VKENUM(VK_ERROR_FRAGMENTED_POOL);
  747. VKENUM(VK_ERROR_INITIALIZATION_FAILED);
  748. VKENUM(VK_ERROR_DEVICE_LOST);
  749. VKENUM(VK_ERROR_MEMORY_MAP_FAILED);
  750. VKENUM(VK_ERROR_LAYER_NOT_PRESENT);
  751. VKENUM(VK_ERROR_EXTENSION_NOT_PRESENT);
  752. VKENUM(VK_ERROR_FEATURE_NOT_PRESENT);
  753. VKENUM(VK_ERROR_INCOMPATIBLE_DRIVER);
  754. VKENUM(VK_ERROR_TOO_MANY_OBJECTS);
  755. VKENUM(VK_ERROR_FORMAT_NOT_SUPPORTED);
  756. VKENUM(VK_ERROR_SURFACE_LOST_KHR);
  757. VKENUM(VK_ERROR_NATIVE_WINDOW_IN_USE_KHR);
  758. VKENUM(VK_SUBOPTIMAL_KHR);
  759. VKENUM(VK_ERROR_OUT_OF_DATE_KHR);
  760. VKENUM(VK_ERROR_INCOMPATIBLE_DISPLAY_KHR);
  761. VKENUM(VK_ERROR_VALIDATION_FAILED_EXT);
  762. #undef VKENUM
  763. default: break;
  764. }
  765. BX_WARN(false, "Unknown VkResult? %x", _result);
  766. return "<VkResult?>";
  767. }
  768. template<typename Ty>
  769. constexpr VkObjectType getType();
  770. template<> VkObjectType getType<VkBuffer >() { return VK_OBJECT_TYPE_BUFFER; }
  771. template<> VkObjectType getType<VkCommandPool >() { return VK_OBJECT_TYPE_COMMAND_POOL; }
  772. template<> VkObjectType getType<VkDescriptorPool >() { return VK_OBJECT_TYPE_DESCRIPTOR_POOL; }
  773. template<> VkObjectType getType<VkDescriptorSet >() { return VK_OBJECT_TYPE_DESCRIPTOR_SET; }
  774. template<> VkObjectType getType<VkDescriptorSetLayout>() { return VK_OBJECT_TYPE_DESCRIPTOR_SET_LAYOUT; }
  775. template<> VkObjectType getType<VkDeviceMemory >() { return VK_OBJECT_TYPE_DEVICE_MEMORY; }
  776. template<> VkObjectType getType<VkFence >() { return VK_OBJECT_TYPE_FENCE; }
  777. template<> VkObjectType getType<VkFramebuffer >() { return VK_OBJECT_TYPE_FRAMEBUFFER; }
  778. template<> VkObjectType getType<VkImage >() { return VK_OBJECT_TYPE_IMAGE; }
  779. template<> VkObjectType getType<VkImageView >() { return VK_OBJECT_TYPE_IMAGE_VIEW; }
  780. template<> VkObjectType getType<VkPipeline >() { return VK_OBJECT_TYPE_PIPELINE; }
  781. template<> VkObjectType getType<VkPipelineCache >() { return VK_OBJECT_TYPE_PIPELINE_CACHE; }
  782. template<> VkObjectType getType<VkPipelineLayout >() { return VK_OBJECT_TYPE_PIPELINE_LAYOUT; }
  783. template<> VkObjectType getType<VkQueryPool >() { return VK_OBJECT_TYPE_QUERY_POOL; }
  784. template<> VkObjectType getType<VkRenderPass >() { return VK_OBJECT_TYPE_RENDER_PASS; }
  785. template<> VkObjectType getType<VkSampler >() { return VK_OBJECT_TYPE_SAMPLER; }
  786. template<> VkObjectType getType<VkSemaphore >() { return VK_OBJECT_TYPE_SEMAPHORE; }
  787. template<> VkObjectType getType<VkShaderModule >() { return VK_OBJECT_TYPE_SHADER_MODULE; }
  788. template<> VkObjectType getType<VkSurfaceKHR >() { return VK_OBJECT_TYPE_SURFACE_KHR; }
  789. template<> VkObjectType getType<VkSwapchainKHR >() { return VK_OBJECT_TYPE_SWAPCHAIN_KHR; }
  790. template<typename Ty>
  791. static BX_NO_INLINE void setDebugObjectName(VkDevice _device, Ty _object, const char* _format, ...)
  792. {
  793. if (BX_ENABLED(BGFX_CONFIG_DEBUG_OBJECT_NAME)
  794. && s_extension[Extension::EXT_debug_utils].m_supported)
  795. {
  796. char temp[2048];
  797. va_list argList;
  798. va_start(argList, _format);
  799. int32_t size = bx::min<int32_t>(sizeof(temp)-1, bx::vsnprintf(temp, sizeof(temp), _format, argList) );
  800. va_end(argList);
  801. temp[size] = '\0';
  802. VkDebugUtilsObjectNameInfoEXT ni;
  803. ni.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_OBJECT_NAME_INFO_EXT;
  804. ni.pNext = NULL;
  805. ni.objectType = getType<Ty>();
  806. ni.objectHandle = uint64_t(_object.vk);
  807. ni.pObjectName = temp;
  808. VK_CHECK(vkSetDebugUtilsObjectNameEXT(_device, &ni) );
  809. }
  810. }
  811. void setMemoryBarrier(
  812. VkCommandBuffer _commandBuffer
  813. , VkPipelineStageFlags _srcStages
  814. , VkPipelineStageFlags _dstStages
  815. )
  816. {
  817. VkMemoryBarrier mb;
  818. mb.sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER;
  819. mb.pNext = NULL;
  820. mb.srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  821. mb.dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  822. vkCmdPipelineBarrier(
  823. _commandBuffer
  824. , _srcStages
  825. , _dstStages
  826. , 0
  827. , 1
  828. , &mb
  829. , 0
  830. , NULL
  831. , 0
  832. , NULL
  833. );
  834. }
  835. void setImageMemoryBarrier(
  836. VkCommandBuffer _commandBuffer
  837. , VkImage _image
  838. , VkImageAspectFlags _aspectMask
  839. , VkImageLayout _oldLayout
  840. , VkImageLayout _newLayout
  841. , uint32_t _baseMipLevel = 0
  842. , uint32_t _levelCount = VK_REMAINING_MIP_LEVELS
  843. , uint32_t _baseArrayLayer = 0
  844. , uint32_t _layerCount = VK_REMAINING_ARRAY_LAYERS
  845. )
  846. {
  847. BX_ASSERT(true
  848. && _newLayout != VK_IMAGE_LAYOUT_UNDEFINED
  849. && _newLayout != VK_IMAGE_LAYOUT_PREINITIALIZED
  850. , "_newLayout cannot use VK_IMAGE_LAYOUT_UNDEFINED or VK_IMAGE_LAYOUT_PREINITIALIZED."
  851. );
  852. constexpr VkPipelineStageFlags depthStageMask = 0
  853. | VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT
  854. | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT
  855. ;
  856. constexpr VkPipelineStageFlags sampledStageMask = 0
  857. | VK_PIPELINE_STAGE_VERTEX_SHADER_BIT
  858. | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT
  859. | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  860. ;
  861. VkPipelineStageFlags srcStageMask = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
  862. VkPipelineStageFlags dstStageMask = VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT;
  863. VkAccessFlags srcAccessMask = 0;
  864. VkAccessFlags dstAccessMask = 0;
  865. switch (_oldLayout)
  866. {
  867. case VK_IMAGE_LAYOUT_UNDEFINED:
  868. break;
  869. case VK_IMAGE_LAYOUT_GENERAL:
  870. srcStageMask = VK_PIPELINE_STAGE_ALL_COMMANDS_BIT;
  871. srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  872. break;
  873. case VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL:
  874. srcStageMask = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  875. srcAccessMask = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
  876. break;
  877. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL:
  878. srcStageMask = depthStageMask;
  879. srcAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
  880. break;
  881. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL:
  882. srcStageMask = depthStageMask | sampledStageMask;
  883. break;
  884. case VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL:
  885. srcStageMask = sampledStageMask;
  886. break;
  887. case VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL:
  888. srcStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  889. break;
  890. case VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL:
  891. srcStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  892. srcAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
  893. break;
  894. case VK_IMAGE_LAYOUT_PREINITIALIZED:
  895. srcStageMask = VK_PIPELINE_STAGE_HOST_BIT;
  896. srcAccessMask = VK_ACCESS_HOST_WRITE_BIT;
  897. break;
  898. case VK_IMAGE_LAYOUT_PRESENT_SRC_KHR:
  899. break;
  900. default:
  901. BX_ASSERT(false, "Unknown image layout.");
  902. break;
  903. }
  904. switch (_newLayout)
  905. {
  906. case VK_IMAGE_LAYOUT_GENERAL:
  907. dstStageMask = VK_PIPELINE_STAGE_ALL_COMMANDS_BIT;
  908. dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  909. break;
  910. case VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL:
  911. dstStageMask = VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT;
  912. dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT;
  913. break;
  914. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL:
  915. dstStageMask = depthStageMask;
  916. dstAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT;
  917. break;
  918. case VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL:
  919. dstStageMask = depthStageMask | sampledStageMask;
  920. dstAccessMask = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_READ_BIT | VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_INPUT_ATTACHMENT_READ_BIT;
  921. break;
  922. case VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL:
  923. dstStageMask = sampledStageMask;
  924. dstAccessMask = VK_ACCESS_SHADER_READ_BIT | VK_ACCESS_INPUT_ATTACHMENT_READ_BIT;
  925. break;
  926. case VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL:
  927. dstStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  928. dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT;
  929. break;
  930. case VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL:
  931. dstStageMask = VK_PIPELINE_STAGE_TRANSFER_BIT;
  932. dstAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT;
  933. break;
  934. case VK_IMAGE_LAYOUT_PRESENT_SRC_KHR:
  935. // vkQueuePresentKHR performs automatic visibility operations
  936. break;
  937. default:
  938. BX_ASSERT(false, "Unknown image layout.");
  939. break;
  940. }
  941. VkImageMemoryBarrier imb;
  942. imb.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
  943. imb.pNext = NULL;
  944. imb.srcAccessMask = srcAccessMask;
  945. imb.dstAccessMask = dstAccessMask;
  946. imb.oldLayout = _oldLayout;
  947. imb.newLayout = _newLayout;
  948. imb.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  949. imb.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
  950. imb.image = _image;
  951. imb.subresourceRange.aspectMask = _aspectMask;
  952. imb.subresourceRange.baseMipLevel = _baseMipLevel;
  953. imb.subresourceRange.levelCount = _levelCount;
  954. imb.subresourceRange.baseArrayLayer = _baseArrayLayer;
  955. imb.subresourceRange.layerCount = _layerCount;
  956. vkCmdPipelineBarrier(_commandBuffer
  957. , srcStageMask
  958. , dstStageMask
  959. , 0
  960. , 0
  961. , NULL
  962. , 0
  963. , NULL
  964. , 1
  965. , &imb
  966. );
  967. }
  968. #define MAX_DESCRIPTOR_SETS (1024 * BGFX_CONFIG_MAX_FRAME_LATENCY)
  969. struct RendererContextVK : public RendererContextI
  970. {
  971. RendererContextVK()
  972. : m_allocatorCb(NULL)
  973. , m_renderDocDll(NULL)
  974. , m_vulkan1Dll(NULL)
  975. , m_maxAnisotropy(1.0f)
  976. , m_depthClamp(false)
  977. , m_wireframe(false)
  978. , m_captureBuffer(VK_NULL_HANDLE)
  979. , m_captureMemory(VK_NULL_HANDLE)
  980. , m_captureSize(0)
  981. {
  982. }
  983. ~RendererContextVK()
  984. {
  985. }
  986. bool init(const Init& _init)
  987. {
  988. struct ErrorState
  989. {
  990. enum Enum
  991. {
  992. Default,
  993. LoadedVulkan1,
  994. InstanceCreated,
  995. DeviceCreated,
  996. CommandQueueCreated,
  997. SwapChainCreated,
  998. DescriptorCreated,
  999. TimerQueryCreated,
  1000. };
  1001. };
  1002. ErrorState::Enum errorState = ErrorState::Default;
  1003. const bool headless = NULL == g_platformData.nwh;
  1004. const void* nextFeatures = NULL;
  1005. VkPhysicalDeviceLineRasterizationFeaturesEXT lineRasterizationFeatures;
  1006. VkPhysicalDeviceCustomBorderColorFeaturesEXT customBorderColorFeatures;
  1007. bx::memSet(&lineRasterizationFeatures, 0, sizeof(lineRasterizationFeatures) );
  1008. bx::memSet(&customBorderColorFeatures, 0, sizeof(customBorderColorFeatures) );
  1009. m_fbh.idx = kInvalidHandle;
  1010. bx::memSet(m_uniforms, 0, sizeof(m_uniforms) );
  1011. bx::memSet(&m_resolution, 0, sizeof(m_resolution) );
  1012. bool imported = true;
  1013. VkResult result;
  1014. m_globalQueueFamily = UINT32_MAX;
  1015. if (_init.debug
  1016. || _init.profile)
  1017. {
  1018. m_renderDocDll = loadRenderDoc();
  1019. }
  1020. setGraphicsDebuggerPresent(NULL != m_renderDocDll);
  1021. m_vulkan1Dll = bx::dlopen(
  1022. #if BX_PLATFORM_WINDOWS
  1023. "vulkan-1.dll"
  1024. #elif BX_PLATFORM_ANDROID
  1025. "libvulkan.so"
  1026. #elif BX_PLATFORM_OSX
  1027. "libvulkan.dylib"
  1028. #else
  1029. "libvulkan.so.1"
  1030. #endif // BX_PLATFORM_*
  1031. );
  1032. if (NULL == m_vulkan1Dll)
  1033. {
  1034. BX_TRACE("Init error: Failed to load vulkan dynamic library.");
  1035. goto error;
  1036. }
  1037. errorState = ErrorState::LoadedVulkan1;
  1038. BX_TRACE("Shared library functions:");
  1039. #define VK_IMPORT_FUNC(_optional, _func) \
  1040. _func = (PFN_##_func)bx::dlsym(m_vulkan1Dll, #_func); \
  1041. BX_TRACE("\t%p " #_func, _func); \
  1042. imported &= _optional || NULL != _func
  1043. VK_IMPORT
  1044. #undef VK_IMPORT_FUNC
  1045. if (!imported)
  1046. {
  1047. BX_TRACE("Init error: Failed to load shared library functions.");
  1048. goto error;
  1049. }
  1050. {
  1051. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_device.m_initialize = _init.debug;
  1052. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_instance.m_initialize = _init.debug;
  1053. s_layer[Layer::VK_LAYER_KHRONOS_validation ].m_device.m_initialize = _init.debug;
  1054. s_layer[Layer::VK_LAYER_KHRONOS_validation ].m_instance.m_initialize = _init.debug;
  1055. s_extension[Extension::EXT_debug_report].m_initialize = _init.debug;
  1056. s_extension[Extension::EXT_shader_viewport_index_layer].m_initialize = !!(_init.capabilities & BGFX_CAPS_VIEWPORT_LAYER_ARRAY);
  1057. s_extension[Extension::EXT_conservative_rasterization ].m_initialize = !!(_init.capabilities & BGFX_CAPS_CONSERVATIVE_RASTER );
  1058. s_extension[Extension::KHR_draw_indirect_count ].m_initialize = !!(_init.capabilities & BGFX_CAPS_DRAW_INDIRECT_COUNT );
  1059. dumpExtensions(VK_NULL_HANDLE, s_extension);
  1060. if (s_layer[Layer::VK_LAYER_KHRONOS_validation].m_device.m_supported
  1061. || s_layer[Layer::VK_LAYER_KHRONOS_validation].m_instance.m_supported)
  1062. {
  1063. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_device.m_supported = false;
  1064. s_layer[Layer::VK_LAYER_LUNARG_standard_validation].m_instance.m_supported = false;
  1065. }
  1066. uint32_t numEnabledLayers = 0;
  1067. const char* enabledLayer[Layer::Count];
  1068. BX_TRACE("Enabled instance layers:");
  1069. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  1070. {
  1071. const Layer& layer = s_layer[ii];
  1072. if (layer.m_instance.m_supported
  1073. && layer.m_instance.m_initialize)
  1074. {
  1075. enabledLayer[numEnabledLayers++] = layer.m_name;
  1076. BX_TRACE("\t%s", layer.m_name);
  1077. }
  1078. }
  1079. #if BX_PLATFORM_OSX || defined(WL_EGL_PLATFORM)
  1080. uint32_t numEnabledExtensions = headless ? 0 : 3;
  1081. const char* enabledExtension[Extension::Count + 3] =
  1082. #else
  1083. uint32_t numEnabledExtensions = headless ? 0 : 2;
  1084. const char* enabledExtension[Extension::Count + 2] =
  1085. #endif
  1086. {
  1087. VK_KHR_SURFACE_EXTENSION_NAME,
  1088. KHR_SURFACE_EXTENSION_NAME,
  1089. #if BX_PLATFORM_OSX
  1090. VK_KHR_PORTABILITY_ENUMERATION_EXTENSION_NAME,
  1091. #endif
  1092. };
  1093. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  1094. {
  1095. const Extension& extension = s_extension[ii];
  1096. const LayerInfo& layerInfo = s_layer[extension.m_layer].m_instance;
  1097. const bool layerEnabled = false
  1098. || extension.m_layer == Layer::Count
  1099. || (layerInfo.m_supported && layerInfo.m_initialize)
  1100. ;
  1101. if (extension.m_supported
  1102. && extension.m_initialize
  1103. && extension.m_instanceExt
  1104. && layerEnabled)
  1105. {
  1106. enabledExtension[numEnabledExtensions++] = extension.m_name;
  1107. }
  1108. }
  1109. BX_TRACE("Enabled instance extensions:");
  1110. for (uint32_t ii = 0; ii < numEnabledExtensions; ++ii)
  1111. {
  1112. BX_TRACE("\t%s", enabledExtension[ii]);
  1113. }
  1114. uint32_t vulkanApiVersionSelector;
  1115. if (NULL != vkEnumerateInstanceVersion)
  1116. {
  1117. result = vkEnumerateInstanceVersion(&vulkanApiVersionSelector);
  1118. if (VK_SUCCESS != result)
  1119. {
  1120. BX_TRACE(
  1121. "Init error: vkEnumerateInstanceVersion failed %d: %s."
  1122. , result
  1123. , getName(result)
  1124. );
  1125. goto error;
  1126. }
  1127. }
  1128. else
  1129. {
  1130. vulkanApiVersionSelector = VK_API_VERSION_1_0;
  1131. }
  1132. VkApplicationInfo appInfo;
  1133. appInfo.sType = VK_STRUCTURE_TYPE_APPLICATION_INFO;
  1134. appInfo.pNext = NULL;
  1135. appInfo.pApplicationName = "bgfx";
  1136. appInfo.applicationVersion = BGFX_API_VERSION;
  1137. appInfo.pEngineName = "bgfx";
  1138. appInfo.engineVersion = BGFX_API_VERSION;
  1139. appInfo.apiVersion = vulkanApiVersionSelector;
  1140. VkInstanceCreateInfo ici;
  1141. ici.sType = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO;
  1142. ici.pNext = NULL;
  1143. #if BX_PLATFORM_OSX
  1144. ici.flags = 0 | VK_INSTANCE_CREATE_ENUMERATE_PORTABILITY_BIT_KHR;
  1145. #else
  1146. ici.flags = 0;
  1147. #endif
  1148. ici.pApplicationInfo = &appInfo;
  1149. ici.enabledLayerCount = numEnabledLayers;
  1150. ici.ppEnabledLayerNames = enabledLayer;
  1151. ici.enabledExtensionCount = numEnabledExtensions;
  1152. ici.ppEnabledExtensionNames = enabledExtension;
  1153. if (BX_ENABLED(BGFX_CONFIG_DEBUG) )
  1154. {
  1155. m_allocatorCb = &s_allocationCb;
  1156. BX_UNUSED(s_allocationCb);
  1157. }
  1158. result = vkCreateInstance(
  1159. &ici
  1160. , m_allocatorCb
  1161. , &m_instance
  1162. );
  1163. if (VK_SUCCESS != result)
  1164. {
  1165. BX_TRACE("Init error: vkCreateInstance failed %d: %s.", result, getName(result) );
  1166. goto error;
  1167. }
  1168. m_instanceApiVersion = vulkanApiVersionSelector;
  1169. BX_TRACE("Instance API version: %d.%d.%d"
  1170. , VK_API_VERSION_MAJOR(m_instanceApiVersion)
  1171. , VK_API_VERSION_MINOR(m_instanceApiVersion)
  1172. , VK_API_VERSION_PATCH(m_instanceApiVersion)
  1173. );
  1174. BX_TRACE("Instance variant: %d", VK_API_VERSION_VARIANT(m_instanceApiVersion) );
  1175. }
  1176. errorState = ErrorState::InstanceCreated;
  1177. BX_TRACE("Instance functions:");
  1178. #define VK_IMPORT_INSTANCE_FUNC(_optional, _func) \
  1179. _func = (PFN_##_func)vkGetInstanceProcAddr(m_instance, #_func); \
  1180. BX_TRACE("\t%p " #_func, _func); \
  1181. imported &= _optional || NULL != _func
  1182. VK_IMPORT_INSTANCE
  1183. #undef VK_IMPORT_INSTANCE_FUNC
  1184. if (!imported)
  1185. {
  1186. BX_TRACE("Init error: Failed to load instance functions.");
  1187. goto error;
  1188. }
  1189. m_debugReportCallback = VK_NULL_HANDLE;
  1190. if (s_extension[Extension::EXT_debug_report].m_supported)
  1191. {
  1192. VkDebugReportCallbackCreateInfoEXT drcb;
  1193. drcb.sType = VK_STRUCTURE_TYPE_DEBUG_REPORT_CREATE_INFO_EXT;
  1194. drcb.pNext = NULL;
  1195. drcb.pfnCallback = debugReportCb;
  1196. drcb.pUserData = NULL;
  1197. drcb.flags = 0
  1198. | VK_DEBUG_REPORT_ERROR_BIT_EXT
  1199. | VK_DEBUG_REPORT_WARNING_BIT_EXT
  1200. ;
  1201. result = vkCreateDebugReportCallbackEXT(m_instance
  1202. , &drcb
  1203. , m_allocatorCb
  1204. , &m_debugReportCallback
  1205. );
  1206. BX_WARN(VK_SUCCESS == result, "vkCreateDebugReportCallbackEXT failed %d: %s.", result, getName(result) );
  1207. }
  1208. {
  1209. BX_TRACE("---");
  1210. uint32_t numPhysicalDevices;
  1211. result = vkEnumeratePhysicalDevices(m_instance
  1212. , &numPhysicalDevices
  1213. , NULL
  1214. );
  1215. if (VK_SUCCESS != result)
  1216. {
  1217. BX_TRACE("Init error: vkEnumeratePhysicalDevices failed %d: %s.", result, getName(result) );
  1218. goto error;
  1219. }
  1220. VkPhysicalDevice physicalDevices[4];
  1221. numPhysicalDevices = bx::min<uint32_t>(numPhysicalDevices, BX_COUNTOF(physicalDevices) );
  1222. result = vkEnumeratePhysicalDevices(m_instance
  1223. , &numPhysicalDevices
  1224. , physicalDevices
  1225. );
  1226. if (VK_SUCCESS != result)
  1227. {
  1228. BX_TRACE("Init error: vkEnumeratePhysicalDevices failed %d: %s.", result, getName(result) );
  1229. goto error;
  1230. }
  1231. Extension physicalDeviceExtensions[4][Extension::Count];
  1232. uint32_t physicalDeviceIdx = UINT32_MAX;
  1233. uint32_t fallbackPhysicalDeviceIdx = UINT32_MAX;
  1234. for (uint32_t ii = 0; ii < numPhysicalDevices; ++ii)
  1235. {
  1236. VkPhysicalDeviceProperties pdp;
  1237. vkGetPhysicalDeviceProperties(physicalDevices[ii], &pdp);
  1238. BX_TRACE("Physical device %d:", ii);
  1239. BX_TRACE("\t Name: %s", pdp.deviceName);
  1240. BX_TRACE("\t API version: %d.%d.%d"
  1241. , VK_API_VERSION_MAJOR(pdp.apiVersion)
  1242. , VK_API_VERSION_MINOR(pdp.apiVersion)
  1243. , VK_API_VERSION_PATCH(pdp.apiVersion)
  1244. );
  1245. BX_TRACE("\t API variant: %d", VK_API_VERSION_VARIANT(pdp.apiVersion) );
  1246. BX_TRACE("\tDriver version: %x", pdp.driverVersion);
  1247. BX_TRACE("\t VendorId: %x", pdp.vendorID);
  1248. BX_TRACE("\t DeviceId: %x", pdp.deviceID);
  1249. BX_TRACE("\t Type: %d", pdp.deviceType);
  1250. if (VK_PHYSICAL_DEVICE_TYPE_CPU == pdp.deviceType)
  1251. {
  1252. pdp.vendorID = BGFX_PCI_ID_SOFTWARE_RASTERIZER;
  1253. }
  1254. g_caps.gpu[ii].vendorId = uint16_t(pdp.vendorID);
  1255. g_caps.gpu[ii].deviceId = uint16_t(pdp.deviceID);
  1256. ++g_caps.numGPUs;
  1257. if ( (BGFX_PCI_ID_NONE != g_caps.vendorId || 0 != g_caps.deviceId)
  1258. && (BGFX_PCI_ID_NONE == g_caps.vendorId || pdp.vendorID == g_caps.vendorId)
  1259. && ( 0 == g_caps.deviceId || pdp.deviceID == g_caps.deviceId) )
  1260. {
  1261. if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU
  1262. || pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU)
  1263. {
  1264. fallbackPhysicalDeviceIdx = ii;
  1265. }
  1266. physicalDeviceIdx = ii;
  1267. }
  1268. else if (UINT32_MAX == physicalDeviceIdx)
  1269. {
  1270. if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU)
  1271. {
  1272. fallbackPhysicalDeviceIdx = ii;
  1273. }
  1274. else if (pdp.deviceType == VK_PHYSICAL_DEVICE_TYPE_DISCRETE_GPU)
  1275. {
  1276. physicalDeviceIdx = ii;
  1277. }
  1278. }
  1279. VkPhysicalDeviceMemoryProperties pdmp;
  1280. vkGetPhysicalDeviceMemoryProperties(physicalDevices[ii], &pdmp);
  1281. BX_TRACE("\tMemory type count: %d", pdmp.memoryTypeCount);
  1282. for (uint32_t jj = 0; jj < pdmp.memoryTypeCount; ++jj)
  1283. {
  1284. BX_TRACE("\t%3d: flags 0x%08x, index %d"
  1285. , jj
  1286. , pdmp.memoryTypes[jj].propertyFlags
  1287. , pdmp.memoryTypes[jj].heapIndex
  1288. );
  1289. }
  1290. BX_TRACE("\tMemory heap count: %d", pdmp.memoryHeapCount);
  1291. for (uint32_t jj = 0; jj < pdmp.memoryHeapCount; ++jj)
  1292. {
  1293. char size[16];
  1294. bx::prettify(size, BX_COUNTOF(size), pdmp.memoryHeaps[jj].size);
  1295. BX_TRACE("\t%3d: flags 0x%08x, size %10s"
  1296. , jj
  1297. , pdmp.memoryHeaps[jj].flags
  1298. , size
  1299. );
  1300. }
  1301. bx::memCopy(&physicalDeviceExtensions[ii][0], &s_extension[0], sizeof(s_extension) );
  1302. dumpExtensions(physicalDevices[ii], physicalDeviceExtensions[ii]);
  1303. }
  1304. if (UINT32_MAX == physicalDeviceIdx)
  1305. {
  1306. physicalDeviceIdx = UINT32_MAX == fallbackPhysicalDeviceIdx
  1307. ? 0
  1308. : fallbackPhysicalDeviceIdx
  1309. ;
  1310. }
  1311. m_physicalDevice = physicalDevices[physicalDeviceIdx];
  1312. bx::memCopy(&s_extension[0], &physicalDeviceExtensions[physicalDeviceIdx][0], sizeof(s_extension) );
  1313. vkGetPhysicalDeviceProperties(m_physicalDevice, &m_deviceProperties);
  1314. g_caps.vendorId = uint16_t(m_deviceProperties.vendorID);
  1315. g_caps.deviceId = uint16_t(m_deviceProperties.deviceID);
  1316. BX_TRACE("Using physical device %d: %s", physicalDeviceIdx, m_deviceProperties.deviceName);
  1317. VkPhysicalDeviceFeatures supportedFeatures;
  1318. if (s_extension[Extension::KHR_get_physical_device_properties2].m_supported)
  1319. {
  1320. VkPhysicalDeviceFeatures2KHR deviceFeatures2;
  1321. deviceFeatures2.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_FEATURES_2_KHR;
  1322. deviceFeatures2.pNext = NULL;
  1323. VkBaseOutStructure* next = (VkBaseOutStructure*)&deviceFeatures2;
  1324. if (s_extension[Extension::EXT_line_rasterization].m_supported)
  1325. {
  1326. next->pNext = (VkBaseOutStructure*)&lineRasterizationFeatures;
  1327. next = (VkBaseOutStructure*)&lineRasterizationFeatures;
  1328. lineRasterizationFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_LINE_RASTERIZATION_FEATURES_EXT;
  1329. lineRasterizationFeatures.pNext = NULL;
  1330. }
  1331. if (s_extension[Extension::EXT_custom_border_color].m_supported)
  1332. {
  1333. next->pNext = (VkBaseOutStructure*)&customBorderColorFeatures;
  1334. next = (VkBaseOutStructure*)&customBorderColorFeatures;
  1335. customBorderColorFeatures.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_CUSTOM_BORDER_COLOR_FEATURES_EXT;
  1336. customBorderColorFeatures.pNext = NULL;
  1337. }
  1338. nextFeatures = deviceFeatures2.pNext;
  1339. vkGetPhysicalDeviceFeatures2KHR(m_physicalDevice, &deviceFeatures2);
  1340. supportedFeatures = deviceFeatures2.features;
  1341. }
  1342. else
  1343. {
  1344. vkGetPhysicalDeviceFeatures(m_physicalDevice, &supportedFeatures);
  1345. }
  1346. bx::memSet(&m_deviceFeatures, 0, sizeof(m_deviceFeatures) );
  1347. m_deviceFeatures.fullDrawIndexUint32 = supportedFeatures.fullDrawIndexUint32;
  1348. m_deviceFeatures.imageCubeArray = supportedFeatures.imageCubeArray && (_init.capabilities & BGFX_CAPS_TEXTURE_CUBE_ARRAY);
  1349. m_deviceFeatures.independentBlend = supportedFeatures.independentBlend && (_init.capabilities & BGFX_CAPS_BLEND_INDEPENDENT);
  1350. m_deviceFeatures.multiDrawIndirect = supportedFeatures.multiDrawIndirect && (_init.capabilities & BGFX_CAPS_DRAW_INDIRECT);
  1351. m_deviceFeatures.drawIndirectFirstInstance = supportedFeatures.drawIndirectFirstInstance && (_init.capabilities & BGFX_CAPS_DRAW_INDIRECT);
  1352. m_deviceFeatures.depthClamp = supportedFeatures.depthClamp;
  1353. m_deviceFeatures.fillModeNonSolid = supportedFeatures.fillModeNonSolid;
  1354. m_deviceFeatures.largePoints = supportedFeatures.largePoints;
  1355. m_deviceFeatures.samplerAnisotropy = supportedFeatures.samplerAnisotropy;
  1356. m_deviceFeatures.textureCompressionETC2 = supportedFeatures.textureCompressionETC2;
  1357. m_deviceFeatures.textureCompressionBC = supportedFeatures.textureCompressionBC;
  1358. m_deviceFeatures.vertexPipelineStoresAndAtomics = supportedFeatures.vertexPipelineStoresAndAtomics;
  1359. m_deviceFeatures.fragmentStoresAndAtomics = supportedFeatures.fragmentStoresAndAtomics;
  1360. m_deviceFeatures.shaderImageGatherExtended = supportedFeatures.shaderImageGatherExtended;
  1361. m_deviceFeatures.shaderStorageImageExtendedFormats = supportedFeatures.shaderStorageImageExtendedFormats;
  1362. m_deviceFeatures.shaderClipDistance = supportedFeatures.shaderClipDistance;
  1363. m_deviceFeatures.shaderCullDistance = supportedFeatures.shaderCullDistance;
  1364. m_deviceFeatures.shaderResourceMinLod = supportedFeatures.shaderResourceMinLod;
  1365. m_deviceFeatures.geometryShader = supportedFeatures.geometryShader;
  1366. m_lineAASupport = true
  1367. && s_extension[Extension::EXT_line_rasterization].m_supported
  1368. && lineRasterizationFeatures.smoothLines
  1369. ;
  1370. m_borderColorSupport = true
  1371. && s_extension[Extension::EXT_custom_border_color].m_supported
  1372. && customBorderColorFeatures.customBorderColors
  1373. ;
  1374. m_timerQuerySupport = m_deviceProperties.limits.timestampComputeAndGraphics;
  1375. const bool indirectDrawSupport = true
  1376. && m_deviceFeatures.multiDrawIndirect
  1377. && m_deviceFeatures.drawIndirectFirstInstance
  1378. ;
  1379. g_caps.supported |= ( 0
  1380. | BGFX_CAPS_ALPHA_TO_COVERAGE
  1381. | (m_deviceFeatures.independentBlend ? BGFX_CAPS_BLEND_INDEPENDENT : 0)
  1382. | BGFX_CAPS_COMPUTE
  1383. | (indirectDrawSupport ? BGFX_CAPS_DRAW_INDIRECT : 0)
  1384. | BGFX_CAPS_FRAGMENT_DEPTH
  1385. | BGFX_CAPS_IMAGE_RW
  1386. | (m_deviceFeatures.fullDrawIndexUint32 ? BGFX_CAPS_INDEX32 : 0)
  1387. | BGFX_CAPS_INSTANCING
  1388. | BGFX_CAPS_OCCLUSION_QUERY
  1389. | (!headless ? BGFX_CAPS_SWAP_CHAIN : 0)
  1390. | BGFX_CAPS_TEXTURE_2D_ARRAY
  1391. | BGFX_CAPS_TEXTURE_3D
  1392. | BGFX_CAPS_TEXTURE_BLIT
  1393. | BGFX_CAPS_TEXTURE_COMPARE_ALL
  1394. | (m_deviceFeatures.imageCubeArray ? BGFX_CAPS_TEXTURE_CUBE_ARRAY : 0)
  1395. | BGFX_CAPS_TEXTURE_READ_BACK
  1396. | BGFX_CAPS_VERTEX_ATTRIB_HALF
  1397. | BGFX_CAPS_VERTEX_ATTRIB_UINT10
  1398. | BGFX_CAPS_VERTEX_ID
  1399. | (m_deviceFeatures.geometryShader ? BGFX_CAPS_PRIMITIVE_ID : 0)
  1400. );
  1401. g_caps.supported |= 0
  1402. | (s_extension[Extension::EXT_conservative_rasterization ].m_supported ? BGFX_CAPS_CONSERVATIVE_RASTER : 0)
  1403. | (s_extension[Extension::EXT_shader_viewport_index_layer].m_supported ? BGFX_CAPS_VIEWPORT_LAYER_ARRAY : 0)
  1404. | (s_extension[Extension::KHR_draw_indirect_count ].m_supported && indirectDrawSupport ? BGFX_CAPS_DRAW_INDIRECT_COUNT : 0)
  1405. ;
  1406. const uint32_t maxAttachments = bx::min<uint32_t>(m_deviceProperties.limits.maxFragmentOutputAttachments, m_deviceProperties.limits.maxColorAttachments);
  1407. g_caps.limits.maxTextureSize = m_deviceProperties.limits.maxImageDimension2D;
  1408. g_caps.limits.maxTextureLayers = m_deviceProperties.limits.maxImageArrayLayers;
  1409. g_caps.limits.maxFBAttachments = bx::min<uint32_t>(maxAttachments, BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS);
  1410. g_caps.limits.maxTextureSamplers = bx::min<uint32_t>(m_deviceProperties.limits.maxPerStageResources, BGFX_CONFIG_MAX_TEXTURE_SAMPLERS);
  1411. g_caps.limits.maxComputeBindings = bx::min<uint32_t>(m_deviceProperties.limits.maxPerStageResources, BGFX_MAX_COMPUTE_BINDINGS);
  1412. g_caps.limits.maxVertexStreams = bx::min<uint32_t>(m_deviceProperties.limits.maxVertexInputBindings, BGFX_CONFIG_MAX_VERTEX_STREAMS);
  1413. {
  1414. const VkSampleCountFlags sampleMask = ~0
  1415. & m_deviceProperties.limits.framebufferColorSampleCounts
  1416. & m_deviceProperties.limits.framebufferDepthSampleCounts
  1417. ;
  1418. for (uint16_t ii = 0, last = 0; ii < BX_COUNTOF(s_msaa); ii++)
  1419. {
  1420. const VkSampleCountFlags sampleBit = s_msaa[ii].Sample;
  1421. if (sampleBit & sampleMask)
  1422. {
  1423. last = ii;
  1424. }
  1425. else
  1426. {
  1427. s_msaa[ii] = s_msaa[last];
  1428. }
  1429. }
  1430. }
  1431. {
  1432. struct ImageTest
  1433. {
  1434. VkImageType type;
  1435. VkImageUsageFlags usage;
  1436. VkImageCreateFlags flags;
  1437. uint32_t formatCaps[2];
  1438. };
  1439. const ImageTest imageTest[] =
  1440. {
  1441. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_SAMPLED_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_2D, BGFX_CAPS_FORMAT_TEXTURE_2D_SRGB } },
  1442. { VK_IMAGE_TYPE_3D, VK_IMAGE_USAGE_SAMPLED_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_3D, BGFX_CAPS_FORMAT_TEXTURE_3D_SRGB } },
  1443. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_SAMPLED_BIT, VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT, { BGFX_CAPS_FORMAT_TEXTURE_CUBE, BGFX_CAPS_FORMAT_TEXTURE_CUBE_SRGB } },
  1444. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER, 0 } },
  1445. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER, 0 } },
  1446. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_STORAGE_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_IMAGE_READ, 0 } },
  1447. { VK_IMAGE_TYPE_2D, VK_IMAGE_USAGE_STORAGE_BIT, 0, { BGFX_CAPS_FORMAT_TEXTURE_IMAGE_WRITE, 0 } },
  1448. };
  1449. for (uint32_t ii = 0; ii < TextureFormat::Count; ++ii)
  1450. {
  1451. uint16_t support = BGFX_CAPS_FORMAT_TEXTURE_NONE;
  1452. const bool depth = bimg::isDepth(bimg::TextureFormat::Enum(ii) );
  1453. VkFormat fmt = depth
  1454. ? s_textureFormat[ii].m_fmtDsv
  1455. : s_textureFormat[ii].m_fmt
  1456. ;
  1457. for (uint32_t jj = 0, num = depth ? 1 : 2; jj < num; ++jj)
  1458. {
  1459. if (VK_FORMAT_UNDEFINED != fmt)
  1460. {
  1461. for (uint32_t test = 0; test < BX_COUNTOF(imageTest); ++test)
  1462. {
  1463. const ImageTest& it = imageTest[test];
  1464. VkImageFormatProperties ifp;
  1465. result = vkGetPhysicalDeviceImageFormatProperties(m_physicalDevice
  1466. , fmt
  1467. , it.type
  1468. , VK_IMAGE_TILING_OPTIMAL
  1469. , it.usage
  1470. , it.flags
  1471. , &ifp
  1472. );
  1473. if (VK_SUCCESS == result)
  1474. {
  1475. support |= it.formatCaps[jj];
  1476. const bool multisample = VK_SAMPLE_COUNT_1_BIT < ifp.sampleCounts;
  1477. if (it.usage & VK_IMAGE_USAGE_SAMPLED_BIT)
  1478. {
  1479. support |= 0
  1480. | BGFX_CAPS_FORMAT_TEXTURE_VERTEX
  1481. | (multisample ? BGFX_CAPS_FORMAT_TEXTURE_MSAA : 0)
  1482. ;
  1483. }
  1484. if (it.usage & (VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT) )
  1485. {
  1486. support |= 0
  1487. | BGFX_CAPS_FORMAT_TEXTURE_MIP_AUTOGEN
  1488. | (multisample ? BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER_MSAA : 0)
  1489. ;
  1490. }
  1491. }
  1492. }
  1493. }
  1494. fmt = s_textureFormat[ii].m_fmtSrgb;
  1495. }
  1496. g_caps.formats[ii] = support;
  1497. }
  1498. }
  1499. vkGetPhysicalDeviceMemoryProperties(m_physicalDevice, &m_memoryProperties);
  1500. }
  1501. {
  1502. BX_TRACE("---");
  1503. uint32_t queueFamilyPropertyCount = 0;
  1504. vkGetPhysicalDeviceQueueFamilyProperties(
  1505. m_physicalDevice
  1506. , &queueFamilyPropertyCount
  1507. , NULL
  1508. );
  1509. VkQueueFamilyProperties* queueFamilyPropertices = (VkQueueFamilyProperties*)bx::alloc(g_allocator, queueFamilyPropertyCount * sizeof(VkQueueFamilyProperties) );
  1510. vkGetPhysicalDeviceQueueFamilyProperties(
  1511. m_physicalDevice
  1512. , &queueFamilyPropertyCount
  1513. , queueFamilyPropertices
  1514. );
  1515. for (uint32_t ii = 0; ii < queueFamilyPropertyCount; ++ii)
  1516. {
  1517. const VkQueueFamilyProperties& qfp = queueFamilyPropertices[ii];
  1518. BX_TRACE("Queue family property %d:", ii);
  1519. BX_TRACE("\t Queue flags: 0x%08x", qfp.queueFlags);
  1520. BX_TRACE("\t Queue count: %d", qfp.queueCount);
  1521. BX_TRACE("\tTS valid bits: 0x%08x", qfp.timestampValidBits);
  1522. BX_TRACE("\t Min image: %d x %d x %d"
  1523. , qfp.minImageTransferGranularity.width
  1524. , qfp.minImageTransferGranularity.height
  1525. , qfp.minImageTransferGranularity.depth
  1526. );
  1527. constexpr VkQueueFlags requiredFlags = VK_QUEUE_GRAPHICS_BIT | VK_QUEUE_COMPUTE_BIT;
  1528. if (UINT32_MAX == m_globalQueueFamily
  1529. && requiredFlags == (requiredFlags & qfp.queueFlags) )
  1530. {
  1531. m_globalQueueFamily = ii;
  1532. }
  1533. }
  1534. bx::free(g_allocator, queueFamilyPropertices);
  1535. if (UINT32_MAX == m_globalQueueFamily)
  1536. {
  1537. BX_TRACE("Init error: Unable to find combined graphics and compute queue.");
  1538. goto error;
  1539. }
  1540. }
  1541. {
  1542. uint32_t numEnabledLayers = 0;
  1543. const char* enabledLayer[Layer::Count];
  1544. BX_TRACE("Enabled device layers:");
  1545. for (uint32_t ii = 0; ii < Layer::Count; ++ii)
  1546. {
  1547. const Layer& layer = s_layer[ii];
  1548. if (layer.m_device.m_supported
  1549. && layer.m_device.m_initialize)
  1550. {
  1551. enabledLayer[numEnabledLayers++] = layer.m_name;
  1552. BX_TRACE("\t%s", layer.m_name);
  1553. }
  1554. }
  1555. #if BX_PLATFORM_OSX
  1556. uint32_t numEnabledExtensions = headless ? 1 : 3;
  1557. const char* enabledExtension[Extension::Count + 3] =
  1558. #else
  1559. uint32_t numEnabledExtensions = headless ? 1 : 2;
  1560. const char* enabledExtension[Extension::Count + 2] =
  1561. #endif
  1562. {
  1563. VK_KHR_MAINTENANCE1_EXTENSION_NAME,
  1564. VK_KHR_SWAPCHAIN_EXTENSION_NAME,
  1565. #if BX_PLATFORM_OSX
  1566. "VK_KHR_portability_subset",
  1567. #endif
  1568. };
  1569. for (uint32_t ii = 0; ii < Extension::Count; ++ii)
  1570. {
  1571. const Extension& extension = s_extension[ii];
  1572. bool layerEnabled = extension.m_layer == Layer::Count ||
  1573. (s_layer[extension.m_layer].m_device.m_supported &&
  1574. s_layer[extension.m_layer].m_device.m_initialize);
  1575. if (extension.m_supported
  1576. && extension.m_initialize
  1577. && !extension.m_instanceExt
  1578. && layerEnabled)
  1579. {
  1580. enabledExtension[numEnabledExtensions++] = extension.m_name;
  1581. }
  1582. }
  1583. BX_TRACE("Enabled device extensions:");
  1584. for (uint32_t ii = 0; ii < numEnabledExtensions; ++ii)
  1585. {
  1586. BX_TRACE("\t%s", enabledExtension[ii]);
  1587. }
  1588. float queuePriorities[1] = { 0.0f };
  1589. VkDeviceQueueCreateInfo dcqi;
  1590. dcqi.sType = VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO;
  1591. dcqi.pNext = NULL;
  1592. dcqi.flags = 0;
  1593. dcqi.queueFamilyIndex = m_globalQueueFamily;
  1594. dcqi.queueCount = 1;
  1595. dcqi.pQueuePriorities = queuePriorities;
  1596. VkDeviceCreateInfo dci;
  1597. dci.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO;
  1598. dci.pNext = nextFeatures;
  1599. dci.flags = 0;
  1600. dci.queueCreateInfoCount = 1;
  1601. dci.pQueueCreateInfos = &dcqi;
  1602. dci.enabledLayerCount = numEnabledLayers;
  1603. dci.ppEnabledLayerNames = enabledLayer;
  1604. dci.enabledExtensionCount = numEnabledExtensions;
  1605. dci.ppEnabledExtensionNames = enabledExtension;
  1606. dci.pEnabledFeatures = &m_deviceFeatures;
  1607. result = vkCreateDevice(
  1608. m_physicalDevice
  1609. , &dci
  1610. , m_allocatorCb
  1611. , &m_device
  1612. );
  1613. if (VK_SUCCESS != result)
  1614. {
  1615. BX_TRACE("Init error: vkCreateDevice failed %d: %s.", result, getName(result) );
  1616. goto error;
  1617. }
  1618. }
  1619. errorState = ErrorState::DeviceCreated;
  1620. BX_TRACE("Device functions:");
  1621. #define VK_IMPORT_DEVICE_FUNC(_optional, _func) \
  1622. _func = (PFN_##_func)vkGetDeviceProcAddr(m_device, #_func); \
  1623. BX_TRACE("\t%p " #_func, _func); \
  1624. imported &= _optional || NULL != _func
  1625. VK_IMPORT_DEVICE
  1626. #undef VK_IMPORT_DEVICE_FUNC
  1627. if (!imported)
  1628. {
  1629. BX_TRACE("Init error: Failed to load device functions.");
  1630. goto error;
  1631. }
  1632. vkGetDeviceQueue(m_device, m_globalQueueFamily, 0, &m_globalQueue);
  1633. {
  1634. m_numFramesInFlight = _init.resolution.maxFrameLatency == 0
  1635. ? BGFX_CONFIG_MAX_FRAME_LATENCY
  1636. : _init.resolution.maxFrameLatency
  1637. ;
  1638. result = m_cmd.init(m_globalQueueFamily, m_globalQueue, m_numFramesInFlight);
  1639. if (VK_SUCCESS != result)
  1640. {
  1641. BX_TRACE("Init error: creating command queue failed %d: %s.", result, getName(result) );
  1642. goto error;
  1643. }
  1644. result = m_cmd.alloc(&m_commandBuffer);
  1645. if (VK_SUCCESS != result)
  1646. {
  1647. BX_TRACE("Init error: allocating command buffer failed %d: %s.", result, getName(result) );
  1648. goto error;
  1649. }
  1650. }
  1651. errorState = ErrorState::CommandQueueCreated;
  1652. m_presentElapsed = 0;
  1653. {
  1654. m_resolution = _init.resolution;
  1655. m_resolution.reset &= ~BGFX_RESET_INTERNAL_FORCE;
  1656. m_numWindows = 0;
  1657. if (!headless)
  1658. {
  1659. m_textVideoMem.resize(false, _init.resolution.width, _init.resolution.height);
  1660. m_textVideoMem.clear();
  1661. for (uint8_t ii = 0; ii < BX_COUNTOF(m_swapchainFormats); ++ii)
  1662. {
  1663. m_swapchainFormats[ii] = TextureFormat::Enum(ii);
  1664. }
  1665. result = m_backBuffer.create(UINT16_MAX, g_platformData.nwh, m_resolution.width, m_resolution.height, m_resolution.format);
  1666. if (VK_SUCCESS != result)
  1667. {
  1668. BX_TRACE("Init error: creating swap chain failed %d: %s.", result, getName(result) );
  1669. goto error;
  1670. }
  1671. m_windows[0] = BGFX_INVALID_HANDLE;
  1672. m_numWindows++;
  1673. postReset();
  1674. }
  1675. }
  1676. errorState = ErrorState::SwapChainCreated;
  1677. {
  1678. VkDescriptorPoolSize dps[] =
  1679. {
  1680. { VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1681. { VK_DESCRIPTOR_TYPE_SAMPLER, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1682. { VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC, MAX_DESCRIPTOR_SETS * 2 },
  1683. { VK_DESCRIPTOR_TYPE_STORAGE_BUFFER, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1684. { VK_DESCRIPTOR_TYPE_STORAGE_IMAGE, MAX_DESCRIPTOR_SETS * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS },
  1685. };
  1686. VkDescriptorPoolCreateInfo dpci;
  1687. dpci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_POOL_CREATE_INFO;
  1688. dpci.pNext = NULL;
  1689. dpci.flags = VK_DESCRIPTOR_POOL_CREATE_FREE_DESCRIPTOR_SET_BIT;
  1690. dpci.maxSets = MAX_DESCRIPTOR_SETS;
  1691. dpci.poolSizeCount = BX_COUNTOF(dps);
  1692. dpci.pPoolSizes = dps;
  1693. result = vkCreateDescriptorPool(m_device, &dpci, m_allocatorCb, &m_descriptorPool);
  1694. if (VK_SUCCESS != result)
  1695. {
  1696. BX_TRACE("Init error: vkCreateDescriptorPool failed %d: %s.", result, getName(result) );
  1697. goto error;
  1698. }
  1699. VkPipelineCacheCreateInfo pcci;
  1700. pcci.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO;
  1701. pcci.pNext = NULL;
  1702. pcci.flags = 0;
  1703. pcci.initialDataSize = 0;
  1704. pcci.pInitialData = NULL;
  1705. result = vkCreatePipelineCache(m_device, &pcci, m_allocatorCb, &m_pipelineCache);
  1706. if (VK_SUCCESS != result)
  1707. {
  1708. BX_TRACE("Init error: vkCreatePipelineCache failed %d: %s.", result, getName(result) );
  1709. goto error;
  1710. }
  1711. }
  1712. {
  1713. const uint32_t size = 128;
  1714. const uint32_t count = BGFX_CONFIG_MAX_DRAW_CALLS;
  1715. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1716. {
  1717. BX_TRACE("Create scratch buffer %d", ii);
  1718. m_scratchBuffer[ii].create(size, count);
  1719. }
  1720. }
  1721. errorState = ErrorState::DescriptorCreated;
  1722. if (NULL == vkSetDebugUtilsObjectNameEXT)
  1723. {
  1724. vkSetDebugUtilsObjectNameEXT = stubSetDebugUtilsObjectNameEXT;
  1725. }
  1726. if (NULL == vkCmdBeginDebugUtilsLabelEXT
  1727. || NULL == vkCmdEndDebugUtilsLabelEXT)
  1728. {
  1729. vkCmdBeginDebugUtilsLabelEXT = stubCmdBeginDebugUtilsLabelEXT;
  1730. vkCmdEndDebugUtilsLabelEXT = stubCmdEndDebugUtilsLabelEXT;
  1731. }
  1732. if (NULL == vkCmdInsertDebugUtilsLabelEXT)
  1733. {
  1734. vkCmdInsertDebugUtilsLabelEXT = stubCmdInsertDebugUtilsLabelEXT;
  1735. }
  1736. // Init reserved part of view name.
  1737. for (uint32_t ii = 0; ii < BGFX_CONFIG_MAX_VIEWS; ++ii)
  1738. {
  1739. bx::snprintf(s_viewName[ii], BGFX_CONFIG_MAX_VIEW_NAME_RESERVED+1, "%3d ", ii);
  1740. }
  1741. if (m_timerQuerySupport)
  1742. {
  1743. result = m_gpuTimer.init();
  1744. if (VK_SUCCESS != result)
  1745. {
  1746. BX_TRACE("Init error: creating GPU timer failed %d: %s.", result, getName(result) );
  1747. goto error;
  1748. }
  1749. }
  1750. errorState = ErrorState::TimerQueryCreated;
  1751. result = m_occlusionQuery.init();
  1752. if (VK_SUCCESS != result)
  1753. {
  1754. BX_TRACE("Init error: creating occlusion query failed %d: %s.", result, getName(result) );
  1755. goto error;
  1756. }
  1757. g_internalData.context = m_device;
  1758. return true;
  1759. error:
  1760. BX_TRACE("errorState %d", errorState);
  1761. switch (errorState)
  1762. {
  1763. case ErrorState::TimerQueryCreated:
  1764. if (m_timerQuerySupport)
  1765. {
  1766. m_gpuTimer.shutdown();
  1767. }
  1768. BX_FALLTHROUGH;
  1769. case ErrorState::DescriptorCreated:
  1770. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1771. {
  1772. m_scratchBuffer[ii].destroy();
  1773. }
  1774. vkDestroy(m_pipelineCache);
  1775. vkDestroy(m_descriptorPool);
  1776. BX_FALLTHROUGH;
  1777. case ErrorState::SwapChainCreated:
  1778. m_backBuffer.destroy();
  1779. BX_FALLTHROUGH;
  1780. case ErrorState::CommandQueueCreated:
  1781. m_cmd.shutdown();
  1782. BX_FALLTHROUGH;
  1783. case ErrorState::DeviceCreated:
  1784. vkDestroyDevice(m_device, m_allocatorCb);
  1785. BX_FALLTHROUGH;
  1786. case ErrorState::InstanceCreated:
  1787. if (VK_NULL_HANDLE != m_debugReportCallback)
  1788. {
  1789. vkDestroyDebugReportCallbackEXT(m_instance, m_debugReportCallback, m_allocatorCb);
  1790. }
  1791. vkDestroyInstance(m_instance, m_allocatorCb);
  1792. BX_FALLTHROUGH;
  1793. case ErrorState::LoadedVulkan1:
  1794. bx::dlclose(m_vulkan1Dll);
  1795. m_vulkan1Dll = NULL;
  1796. m_allocatorCb = NULL;
  1797. unloadRenderDoc(m_renderDocDll);
  1798. BX_FALLTHROUGH;
  1799. case ErrorState::Default:
  1800. break;
  1801. };
  1802. return false;
  1803. }
  1804. void shutdown()
  1805. {
  1806. VK_CHECK(vkDeviceWaitIdle(m_device) );
  1807. if (m_timerQuerySupport)
  1808. {
  1809. m_gpuTimer.shutdown();
  1810. }
  1811. m_occlusionQuery.shutdown();
  1812. preReset();
  1813. m_pipelineStateCache.invalidate();
  1814. m_descriptorSetLayoutCache.invalidate();
  1815. m_renderPassCache.invalidate();
  1816. m_samplerCache.invalidate();
  1817. m_samplerBorderColorCache.invalidate();
  1818. m_imageViewCache.invalidate();
  1819. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  1820. {
  1821. m_scratchBuffer[ii].destroy();
  1822. }
  1823. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  1824. {
  1825. m_frameBuffers[ii].destroy();
  1826. }
  1827. for (uint32_t ii = 0; ii < BX_COUNTOF(m_indexBuffers); ++ii)
  1828. {
  1829. m_indexBuffers[ii].destroy();
  1830. }
  1831. for (uint32_t ii = 0; ii < BX_COUNTOF(m_vertexBuffers); ++ii)
  1832. {
  1833. m_vertexBuffers[ii].destroy();
  1834. }
  1835. for (uint32_t ii = 0; ii < BX_COUNTOF(m_shaders); ++ii)
  1836. {
  1837. m_shaders[ii].destroy();
  1838. }
  1839. for (uint32_t ii = 0; ii < BX_COUNTOF(m_textures); ++ii)
  1840. {
  1841. m_textures[ii].destroy();
  1842. }
  1843. m_backBuffer.destroy();
  1844. m_cmd.shutdown();
  1845. vkDestroy(m_pipelineCache);
  1846. vkDestroy(m_descriptorPool);
  1847. vkDestroyDevice(m_device, m_allocatorCb);
  1848. if (VK_NULL_HANDLE != m_debugReportCallback)
  1849. {
  1850. vkDestroyDebugReportCallbackEXT(m_instance, m_debugReportCallback, m_allocatorCb);
  1851. }
  1852. vkDestroyInstance(m_instance, m_allocatorCb);
  1853. bx::dlclose(m_vulkan1Dll);
  1854. m_vulkan1Dll = NULL;
  1855. m_allocatorCb = NULL;
  1856. unloadRenderDoc(m_renderDocDll);
  1857. }
  1858. RendererType::Enum getRendererType() const override
  1859. {
  1860. return RendererType::Vulkan;
  1861. }
  1862. const char* getRendererName() const override
  1863. {
  1864. return BGFX_RENDERER_VULKAN_NAME;
  1865. }
  1866. bool isDeviceRemoved() override
  1867. {
  1868. return false;
  1869. }
  1870. void flip() override
  1871. {
  1872. int64_t start = bx::getHPCounter();
  1873. for (uint16_t ii = 0; ii < m_numWindows; ++ii)
  1874. {
  1875. FrameBufferVK& fb = isValid(m_windows[ii])
  1876. ? m_frameBuffers[m_windows[ii].idx]
  1877. : m_backBuffer
  1878. ;
  1879. fb.present();
  1880. }
  1881. int64_t now = bx::getHPCounter();
  1882. m_presentElapsed += now - start;
  1883. }
  1884. void createIndexBuffer(IndexBufferHandle _handle, const Memory* _mem, uint16_t _flags) override
  1885. {
  1886. m_indexBuffers[_handle.idx].create(m_commandBuffer, _mem->size, _mem->data, _flags, false);
  1887. }
  1888. void destroyIndexBuffer(IndexBufferHandle _handle) override
  1889. {
  1890. m_indexBuffers[_handle.idx].destroy();
  1891. }
  1892. void createVertexLayout(VertexLayoutHandle _handle, const VertexLayout& _layout) override
  1893. {
  1894. VertexLayout& layout = m_vertexLayouts[_handle.idx];
  1895. bx::memCopy(&layout, &_layout, sizeof(VertexLayout) );
  1896. dump(layout);
  1897. }
  1898. void destroyVertexLayout(VertexLayoutHandle /*_handle*/) override
  1899. {
  1900. }
  1901. void createVertexBuffer(VertexBufferHandle _handle, const Memory* _mem, VertexLayoutHandle _layoutHandle, uint16_t _flags) override
  1902. {
  1903. m_vertexBuffers[_handle.idx].create(m_commandBuffer, _mem->size, _mem->data, _layoutHandle, _flags);
  1904. }
  1905. void destroyVertexBuffer(VertexBufferHandle _handle) override
  1906. {
  1907. m_vertexBuffers[_handle.idx].destroy();
  1908. }
  1909. void createDynamicIndexBuffer(IndexBufferHandle _handle, uint32_t _size, uint16_t _flags) override
  1910. {
  1911. m_indexBuffers[_handle.idx].create(m_commandBuffer, _size, NULL, _flags, false);
  1912. }
  1913. void updateDynamicIndexBuffer(IndexBufferHandle _handle, uint32_t _offset, uint32_t _size, const Memory* _mem) override
  1914. {
  1915. m_indexBuffers[_handle.idx].update(m_commandBuffer, _offset, bx::min<uint32_t>(_size, _mem->size), _mem->data);
  1916. }
  1917. void destroyDynamicIndexBuffer(IndexBufferHandle _handle) override
  1918. {
  1919. m_indexBuffers[_handle.idx].destroy();
  1920. }
  1921. void createDynamicVertexBuffer(VertexBufferHandle _handle, uint32_t _size, uint16_t _flags) override
  1922. {
  1923. VertexLayoutHandle layoutHandle = BGFX_INVALID_HANDLE;
  1924. m_vertexBuffers[_handle.idx].create(m_commandBuffer, _size, NULL, layoutHandle, _flags);
  1925. }
  1926. void updateDynamicVertexBuffer(VertexBufferHandle _handle, uint32_t _offset, uint32_t _size, const Memory* _mem) override
  1927. {
  1928. m_vertexBuffers[_handle.idx].update(m_commandBuffer, _offset, bx::min<uint32_t>(_size, _mem->size), _mem->data);
  1929. }
  1930. void destroyDynamicVertexBuffer(VertexBufferHandle _handle) override
  1931. {
  1932. m_vertexBuffers[_handle.idx].destroy();
  1933. }
  1934. void createShader(ShaderHandle _handle, const Memory* _mem) override
  1935. {
  1936. m_shaders[_handle.idx].create(_mem);
  1937. }
  1938. void destroyShader(ShaderHandle _handle) override
  1939. {
  1940. m_shaders[_handle.idx].destroy();
  1941. }
  1942. void createProgram(ProgramHandle _handle, ShaderHandle _vsh, ShaderHandle _fsh) override
  1943. {
  1944. m_program[_handle.idx].create(&m_shaders[_vsh.idx], isValid(_fsh) ? &m_shaders[_fsh.idx] : NULL);
  1945. }
  1946. void destroyProgram(ProgramHandle _handle) override
  1947. {
  1948. m_program[_handle.idx].destroy();
  1949. }
  1950. void* createTexture(TextureHandle _handle, const Memory* _mem, uint64_t _flags, uint8_t _skip) override
  1951. {
  1952. return m_textures[_handle.idx].create(m_commandBuffer, _mem, _flags, _skip);
  1953. }
  1954. void updateTextureBegin(TextureHandle /*_handle*/, uint8_t /*_side*/, uint8_t /*_mip*/) override
  1955. {
  1956. }
  1957. void updateTexture(TextureHandle _handle, uint8_t _side, uint8_t _mip, const Rect& _rect, uint16_t _z, uint16_t _depth, uint16_t _pitch, const Memory* _mem) override
  1958. {
  1959. m_textures[_handle.idx].update(m_commandBuffer, _side, _mip, _rect, _z, _depth, _pitch, _mem);
  1960. }
  1961. void updateTextureEnd() override
  1962. {
  1963. }
  1964. void readTexture(TextureHandle _handle, void* _data, uint8_t _mip) override
  1965. {
  1966. TextureVK& texture = m_textures[_handle.idx];
  1967. uint32_t height = bx::uint32_max(1, texture.m_height >> _mip);
  1968. uint32_t pitch = texture.m_readback.pitch(_mip);
  1969. uint32_t size = height * pitch;
  1970. VkDeviceMemory stagingMemory;
  1971. VkBuffer stagingBuffer;
  1972. VK_CHECK(createReadbackBuffer(size, &stagingBuffer, &stagingMemory) );
  1973. texture.m_readback.copyImageToBuffer(
  1974. m_commandBuffer
  1975. , stagingBuffer
  1976. , texture.m_currentImageLayout
  1977. , texture.m_aspectMask
  1978. , _mip
  1979. );
  1980. kick(true);
  1981. texture.m_readback.readback(stagingMemory, 0, _data, _mip);
  1982. vkDestroy(stagingBuffer);
  1983. vkDestroy(stagingMemory);
  1984. }
  1985. void resizeTexture(TextureHandle _handle, uint16_t _width, uint16_t _height, uint8_t _numMips, uint16_t _numLayers) override
  1986. {
  1987. const TextureVK& texture = m_textures[_handle.idx];
  1988. const TextureFormat::Enum format = TextureFormat::Enum(texture.m_requestedFormat);
  1989. const uint64_t flags = texture.m_flags;
  1990. const uint32_t size = sizeof(uint32_t) + sizeof(TextureCreate);
  1991. const Memory* mem = alloc(size);
  1992. bx::StaticMemoryBlockWriter writer(mem->data, mem->size);
  1993. uint32_t magic = BGFX_CHUNK_MAGIC_TEX;
  1994. bx::write(&writer, magic, bx::ErrorAssert{});
  1995. TextureCreate tc;
  1996. tc.m_width = _width;
  1997. tc.m_height = _height;
  1998. tc.m_depth = 0;
  1999. tc.m_numLayers = _numLayers;
  2000. tc.m_numMips = _numMips;
  2001. tc.m_format = format;
  2002. tc.m_cubeMap = false;
  2003. tc.m_mem = NULL;
  2004. bx::write(&writer, tc, bx::ErrorAssert{});
  2005. destroyTexture(_handle);
  2006. createTexture(_handle, mem, flags, 0);
  2007. bgfx::release(mem);
  2008. }
  2009. void overrideInternal(TextureHandle /*_handle*/, uintptr_t /*_ptr*/) override
  2010. {
  2011. }
  2012. uintptr_t getInternal(TextureHandle /*_handle*/) override
  2013. {
  2014. return 0;
  2015. }
  2016. void destroyTexture(TextureHandle _handle) override
  2017. {
  2018. m_imageViewCache.invalidateWithParent(_handle.idx);
  2019. m_textures[_handle.idx].destroy();
  2020. }
  2021. void createFrameBuffer(FrameBufferHandle _handle, uint8_t _num, const Attachment* _attachment) override
  2022. {
  2023. m_frameBuffers[_handle.idx].create(_num, _attachment);
  2024. }
  2025. void createFrameBuffer(FrameBufferHandle _handle, void* _nwh, uint32_t _width, uint32_t _height, TextureFormat::Enum _format, TextureFormat::Enum _depthFormat) override
  2026. {
  2027. for (uint32_t ii = 0, num = m_numWindows; ii < num; ++ii)
  2028. {
  2029. FrameBufferHandle handle = m_windows[ii];
  2030. if (isValid(handle)
  2031. && m_frameBuffers[handle.idx].m_nwh == _nwh)
  2032. {
  2033. destroyFrameBuffer(handle);
  2034. }
  2035. }
  2036. uint16_t denseIdx = m_numWindows++;
  2037. m_windows[denseIdx] = _handle;
  2038. VK_CHECK(m_frameBuffers[_handle.idx].create(denseIdx, _nwh, _width, _height, _format, _depthFormat) );
  2039. }
  2040. void destroyFrameBuffer(FrameBufferHandle _handle) override
  2041. {
  2042. FrameBufferVK& frameBuffer = m_frameBuffers[_handle.idx];
  2043. if (_handle.idx == m_fbh.idx)
  2044. {
  2045. setFrameBuffer(BGFX_INVALID_HANDLE, false);
  2046. }
  2047. uint16_t denseIdx = frameBuffer.destroy();
  2048. if (UINT16_MAX != denseIdx)
  2049. {
  2050. --m_numWindows;
  2051. if (m_numWindows > 1)
  2052. {
  2053. FrameBufferHandle handle = m_windows[m_numWindows];
  2054. m_windows[m_numWindows] = {kInvalidHandle};
  2055. if (m_numWindows != denseIdx)
  2056. {
  2057. m_windows[denseIdx] = handle;
  2058. m_frameBuffers[handle.idx].m_denseIdx = denseIdx;
  2059. }
  2060. }
  2061. }
  2062. }
  2063. void createUniform(UniformHandle _handle, UniformType::Enum _type, uint16_t _num, const char* _name) override
  2064. {
  2065. if (NULL != m_uniforms[_handle.idx])
  2066. {
  2067. bx::free(g_allocator, m_uniforms[_handle.idx]);
  2068. }
  2069. const uint32_t size = bx::alignUp(g_uniformTypeSize[_type] * _num, 16);
  2070. void* data = bx::alloc(g_allocator, size);
  2071. bx::memSet(data, 0, size);
  2072. m_uniforms[_handle.idx] = data;
  2073. m_uniformReg.add(_handle, _name);
  2074. }
  2075. void destroyUniform(UniformHandle _handle) override
  2076. {
  2077. bx::free(g_allocator, m_uniforms[_handle.idx]);
  2078. m_uniforms[_handle.idx] = NULL;
  2079. }
  2080. void requestScreenShot(FrameBufferHandle _fbh, const char* _filePath) override
  2081. {
  2082. const FrameBufferVK& frameBuffer = isValid(_fbh)
  2083. ? m_frameBuffers[_fbh.idx]
  2084. : m_backBuffer
  2085. ;
  2086. const SwapChainVK& swapChain = frameBuffer.m_swapChain;
  2087. if (!isSwapChainReadable(swapChain) )
  2088. {
  2089. BX_TRACE("Unable to capture screenshot %s.", _filePath);
  2090. return;
  2091. }
  2092. auto callback = [](void* _src, uint32_t _width, uint32_t _height, uint32_t _pitch, const void* _userData)
  2093. {
  2094. const char* filePath = (const char*)_userData;
  2095. g_callback->screenShot(
  2096. filePath
  2097. , _width
  2098. , _height
  2099. , _pitch
  2100. , _src
  2101. , _height * _pitch
  2102. , false
  2103. );
  2104. };
  2105. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(swapChain.m_colorFormat) );
  2106. const uint32_t size = frameBuffer.m_width * frameBuffer.m_height * bpp / 8;
  2107. VkDeviceMemory stagingMemory;
  2108. VkBuffer stagingBuffer;
  2109. VK_CHECK(createReadbackBuffer(size, &stagingBuffer, &stagingMemory) );
  2110. readSwapChain(swapChain, stagingBuffer, stagingMemory, callback, _filePath);
  2111. vkDestroy(stagingBuffer);
  2112. vkDestroy(stagingMemory);
  2113. }
  2114. void updateViewName(ViewId _id, const char* _name) override
  2115. {
  2116. bx::strCopy(&s_viewName[_id][BGFX_CONFIG_MAX_VIEW_NAME_RESERVED]
  2117. , BX_COUNTOF(s_viewName[0]) - BGFX_CONFIG_MAX_VIEW_NAME_RESERVED
  2118. , _name
  2119. );
  2120. }
  2121. void updateUniform(uint16_t _loc, const void* _data, uint32_t _size) override
  2122. {
  2123. bx::memCopy(m_uniforms[_loc], _data, _size);
  2124. }
  2125. void invalidateOcclusionQuery(OcclusionQueryHandle _handle) override
  2126. {
  2127. m_occlusionQuery.invalidate(_handle);
  2128. }
  2129. void setMarker(const char* _marker, uint16_t _len) override
  2130. {
  2131. if (BX_ENABLED(BGFX_CONFIG_DEBUG_ANNOTATION) )
  2132. {
  2133. BX_UNUSED(_len);
  2134. const uint32_t abgr = kColorMarker;
  2135. VkDebugUtilsLabelEXT dul;
  2136. dul.sType = VK_STRUCTURE_TYPE_DEBUG_UTILS_LABEL_EXT;
  2137. dul.pNext = NULL;
  2138. dul.pLabelName = _marker;
  2139. dul.color[0] = ((abgr >> 24) & 0xff) / 255.0f;
  2140. dul.color[1] = ((abgr >> 16) & 0xff) / 255.0f;
  2141. dul.color[2] = ((abgr >> 8) & 0xff) / 255.0f;
  2142. dul.color[3] = ((abgr >> 0) & 0xff) / 255.0f;
  2143. vkCmdInsertDebugUtilsLabelEXT(m_commandBuffer, &dul);
  2144. }
  2145. }
  2146. virtual void setName(Handle _handle, const char* _name, uint16_t _len) override
  2147. {
  2148. switch (_handle.type)
  2149. {
  2150. case Handle::IndexBuffer:
  2151. setDebugObjectName(m_device, m_indexBuffers[_handle.idx].m_buffer, "%.*s", _len, _name);
  2152. break;
  2153. case Handle::Shader:
  2154. setDebugObjectName(m_device, m_shaders[_handle.idx].m_module, "%.*s", _len, _name);
  2155. break;
  2156. case Handle::Texture:
  2157. setDebugObjectName(m_device, m_textures[_handle.idx].m_textureImage, "%.*s", _len, _name);
  2158. if (VK_NULL_HANDLE != m_textures[_handle.idx].m_singleMsaaImage)
  2159. {
  2160. setDebugObjectName(m_device, m_textures[_handle.idx].m_singleMsaaImage, "%.*s", _len, _name);
  2161. }
  2162. break;
  2163. case Handle::VertexBuffer:
  2164. setDebugObjectName(m_device, m_vertexBuffers[_handle.idx].m_buffer, "%.*s", _len, _name);
  2165. break;
  2166. default:
  2167. BX_ASSERT(false, "Invalid handle type?! %d", _handle.type);
  2168. break;
  2169. }
  2170. }
  2171. template<typename Ty>
  2172. void release(Ty& _object)
  2173. {
  2174. if (VK_NULL_HANDLE != _object)
  2175. {
  2176. m_cmd.release(uint64_t(_object.vk), getType<Ty>() );
  2177. _object = VK_NULL_HANDLE;
  2178. }
  2179. }
  2180. void submitBlit(BlitState& _bs, uint16_t _view);
  2181. void submit(Frame* _render, ClearQuad& _clearQuad, TextVideoMemBlitter& _textVideoMemBlitter) override;
  2182. void blitSetup(TextVideoMemBlitter& _blitter) override
  2183. {
  2184. const uint32_t width = m_backBuffer.m_width;
  2185. const uint32_t height = m_backBuffer.m_height;
  2186. setFrameBuffer(BGFX_INVALID_HANDLE);
  2187. VkViewport vp;
  2188. vp.x = 0.0f;
  2189. vp.y = float(height);
  2190. vp.width = float(width);
  2191. vp.height = -float(height);
  2192. vp.minDepth = 0.0f;
  2193. vp.maxDepth = 1.0f;
  2194. vkCmdSetViewport(m_commandBuffer, 0, 1, &vp);
  2195. VkRect2D rc;
  2196. rc.offset.x = 0;
  2197. rc.offset.y = 0;
  2198. rc.extent.width = width;
  2199. rc.extent.height = height;
  2200. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  2201. const uint64_t state = 0
  2202. | BGFX_STATE_WRITE_RGB
  2203. | BGFX_STATE_WRITE_A
  2204. | BGFX_STATE_DEPTH_TEST_ALWAYS
  2205. | BGFX_STATE_MSAA
  2206. ;
  2207. const VertexLayout* layout = &m_vertexLayouts[_blitter.m_vb->layoutHandle.idx];
  2208. VkPipeline pso = getPipeline(state
  2209. , packStencil(BGFX_STENCIL_DEFAULT, BGFX_STENCIL_DEFAULT)
  2210. , 1
  2211. , &layout
  2212. , _blitter.m_program
  2213. , 0
  2214. );
  2215. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pso);
  2216. ProgramVK& program = m_program[_blitter.m_program.idx];
  2217. float proj[16];
  2218. bx::mtxOrtho(proj, 0.0f, (float)width, (float)height, 0.0f, 0.0f, 1000.0f, 0.0f, false);
  2219. PredefinedUniform& predefined = m_program[_blitter.m_program.idx].m_predefined[0];
  2220. uint8_t flags = predefined.m_type;
  2221. setShaderUniform(flags, predefined.m_loc, proj, 4);
  2222. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  2223. if (NULL != vcb)
  2224. {
  2225. commit(*vcb);
  2226. }
  2227. ScratchBufferVK& scratchBuffer = m_scratchBuffer[m_cmd.m_currentFrameInFlight];
  2228. const uint32_t bufferOffset = scratchBuffer.write(m_vsScratch, program.m_vsh->m_size);
  2229. const TextureVK& texture = m_textures[_blitter.m_texture.idx];
  2230. RenderBind bind;
  2231. bind.clear();
  2232. bind.m_bind[0].m_type = Binding::Texture;
  2233. bind.m_bind[0].m_idx = _blitter.m_texture.idx;
  2234. bind.m_bind[0].m_samplerFlags = (uint32_t)(texture.m_flags & BGFX_SAMPLER_BITS_MASK);
  2235. const VkDescriptorSet descriptorSet = getDescriptorSet(program, bind, scratchBuffer, NULL);
  2236. vkCmdBindDescriptorSets(
  2237. m_commandBuffer
  2238. , VK_PIPELINE_BIND_POINT_GRAPHICS
  2239. , program.m_pipelineLayout
  2240. , 0
  2241. , 1
  2242. , &descriptorSet
  2243. , 1
  2244. , &bufferOffset
  2245. );
  2246. const VertexBufferVK& vb = m_vertexBuffers[_blitter.m_vb->handle.idx];
  2247. const VkDeviceSize offset = 0;
  2248. vkCmdBindVertexBuffers(m_commandBuffer, 0, 1, &vb.m_buffer, &offset);
  2249. const BufferVK& ib = m_indexBuffers[_blitter.m_ib->handle.idx];
  2250. vkCmdBindIndexBuffer(
  2251. m_commandBuffer
  2252. , ib.m_buffer
  2253. , 0
  2254. , VK_INDEX_TYPE_UINT16
  2255. );
  2256. }
  2257. void blitRender(TextVideoMemBlitter& _blitter, uint32_t _numIndices) override
  2258. {
  2259. const uint32_t numVertices = _numIndices*4/6;
  2260. if (0 < numVertices && m_backBuffer.isRenderable() )
  2261. {
  2262. m_indexBuffers[_blitter.m_ib->handle.idx].update(m_commandBuffer, 0, _numIndices*2, _blitter.m_ib->data);
  2263. m_vertexBuffers[_blitter.m_vb->handle.idx].update(m_commandBuffer, 0, numVertices*_blitter.m_layout.m_stride, _blitter.m_vb->data, true);
  2264. VkRenderPassBeginInfo rpbi;
  2265. rpbi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  2266. rpbi.pNext = NULL;
  2267. rpbi.renderPass = m_backBuffer.m_renderPass;
  2268. rpbi.framebuffer = m_backBuffer.m_currentFramebuffer;
  2269. rpbi.renderArea.offset.x = 0;
  2270. rpbi.renderArea.offset.y = 0;
  2271. rpbi.renderArea.extent.width = m_backBuffer.m_width;
  2272. rpbi.renderArea.extent.height = m_backBuffer.m_height;
  2273. rpbi.clearValueCount = 0;
  2274. rpbi.pClearValues = NULL;
  2275. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  2276. vkCmdDrawIndexed(m_commandBuffer, _numIndices, 1, 0, 0, 0);
  2277. vkCmdEndRenderPass(m_commandBuffer);
  2278. }
  2279. }
  2280. void preReset()
  2281. {
  2282. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  2283. {
  2284. m_frameBuffers[ii].preReset();
  2285. }
  2286. if (m_captureSize > 0)
  2287. {
  2288. g_callback->captureEnd();
  2289. release(m_captureBuffer);
  2290. release(m_captureMemory);
  2291. m_captureSize = 0;
  2292. }
  2293. }
  2294. void postReset()
  2295. {
  2296. for (uint32_t ii = 0; ii < BX_COUNTOF(m_frameBuffers); ++ii)
  2297. {
  2298. m_frameBuffers[ii].postReset();
  2299. }
  2300. if (m_resolution.reset & BGFX_RESET_CAPTURE)
  2301. {
  2302. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_backBuffer.m_swapChain.m_colorFormat) );
  2303. const uint32_t captureSize = m_backBuffer.m_width * m_backBuffer.m_height * bpp / 8;
  2304. const uint8_t dstBpp = bimg::getBitsPerPixel(bimg::TextureFormat::BGRA8);
  2305. const uint32_t dstPitch = m_backBuffer.m_width * dstBpp / 8;
  2306. if (captureSize > m_captureSize)
  2307. {
  2308. release(m_captureBuffer);
  2309. release(m_captureMemory);
  2310. m_captureSize = captureSize;
  2311. VK_CHECK(createReadbackBuffer(m_captureSize, &m_captureBuffer, &m_captureMemory) );
  2312. }
  2313. g_callback->captureBegin(m_resolution.width, m_resolution.height, dstPitch, TextureFormat::BGRA8, false);
  2314. }
  2315. }
  2316. bool updateResolution(const Resolution& _resolution)
  2317. {
  2318. const bool suspended = !!(_resolution.reset & BGFX_RESET_SUSPEND);
  2319. float maxAnisotropy = 1.0f;
  2320. if (!!(_resolution.reset & BGFX_RESET_MAXANISOTROPY) )
  2321. {
  2322. maxAnisotropy = m_deviceProperties.limits.maxSamplerAnisotropy;
  2323. }
  2324. if (m_maxAnisotropy != maxAnisotropy)
  2325. {
  2326. m_maxAnisotropy = maxAnisotropy;
  2327. m_samplerCache.invalidate();
  2328. m_samplerBorderColorCache.invalidate();
  2329. }
  2330. bool depthClamp = m_deviceFeatures.depthClamp && !!(_resolution.reset & BGFX_RESET_DEPTH_CLAMP);
  2331. if (m_depthClamp != depthClamp)
  2332. {
  2333. m_depthClamp = depthClamp;
  2334. m_pipelineStateCache.invalidate();
  2335. }
  2336. if (NULL == m_backBuffer.m_nwh)
  2337. {
  2338. return suspended;
  2339. }
  2340. uint32_t flags = _resolution.reset & ~(0
  2341. | BGFX_RESET_SUSPEND
  2342. | BGFX_RESET_MAXANISOTROPY
  2343. | BGFX_RESET_DEPTH_CLAMP
  2344. );
  2345. // Note: m_needToRefreshSwapchain is deliberately ignored when deciding whether to recreate the swapchain
  2346. // because it can happen several frames before submit is called with the new resolution.
  2347. // Instead, vkAcquireNextImageKHR and all draws to the backbuffer are skipped until the window size is updated.
  2348. // That also fixes a related issue where VK_ERROR_OUT_OF_DATE_KHR is returned from
  2349. // vkQueuePresentKHR when the window doesn't exist anymore, and vkGetPhysicalDeviceSurfaceCapabilitiesKHR
  2350. // fails with VK_ERROR_SURFACE_LOST_KHR.
  2351. if (false
  2352. || m_resolution.format != _resolution.format
  2353. || m_resolution.width != _resolution.width
  2354. || m_resolution.height != _resolution.height
  2355. || m_resolution.reset != flags
  2356. || m_backBuffer.m_swapChain.m_needToRecreateSurface)
  2357. {
  2358. flags &= ~BGFX_RESET_INTERNAL_FORCE;
  2359. if (m_backBuffer.m_nwh != g_platformData.nwh)
  2360. {
  2361. m_backBuffer.m_nwh = g_platformData.nwh;
  2362. }
  2363. m_resolution = _resolution;
  2364. m_resolution.reset = flags;
  2365. m_textVideoMem.resize(false, _resolution.width, _resolution.height);
  2366. m_textVideoMem.clear();
  2367. preReset();
  2368. m_backBuffer.update(m_commandBuffer, m_resolution);
  2369. postReset();
  2370. }
  2371. return suspended;
  2372. }
  2373. void setShaderUniform(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2374. {
  2375. if (_flags & kUniformFragmentBit)
  2376. {
  2377. bx::memCopy(&m_fsScratch[_regIndex], _val, _numRegs*16);
  2378. }
  2379. else
  2380. {
  2381. bx::memCopy(&m_vsScratch[_regIndex], _val, _numRegs*16);
  2382. }
  2383. }
  2384. void setShaderUniform4f(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2385. {
  2386. setShaderUniform(_flags, _regIndex, _val, _numRegs);
  2387. }
  2388. void setShaderUniform4x4f(uint8_t _flags, uint32_t _regIndex, const void* _val, uint32_t _numRegs)
  2389. {
  2390. setShaderUniform(_flags, _regIndex, _val, _numRegs);
  2391. }
  2392. void setFrameBuffer(FrameBufferHandle _fbh, bool _acquire = true)
  2393. {
  2394. BX_ASSERT(false
  2395. || isValid(_fbh)
  2396. || NULL != m_backBuffer.m_nwh
  2397. , "Rendering to backbuffer in headless mode."
  2398. );
  2399. FrameBufferVK& newFrameBuffer = isValid(_fbh)
  2400. ? m_frameBuffers[_fbh.idx]
  2401. : m_backBuffer
  2402. ;
  2403. FrameBufferVK& oldFrameBuffer = isValid(m_fbh)
  2404. ? m_frameBuffers[m_fbh.idx]
  2405. : m_backBuffer
  2406. ;
  2407. if (NULL == oldFrameBuffer.m_nwh
  2408. && m_fbh.idx != _fbh.idx)
  2409. {
  2410. oldFrameBuffer.resolve();
  2411. for (uint8_t ii = 0, num = oldFrameBuffer.m_num; ii < num; ++ii)
  2412. {
  2413. TextureVK& texture = m_textures[oldFrameBuffer.m_texture[ii].idx];
  2414. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout);
  2415. if (VK_NULL_HANDLE != texture.m_singleMsaaImage)
  2416. {
  2417. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout, true);
  2418. }
  2419. }
  2420. if (isValid(oldFrameBuffer.m_depth) )
  2421. {
  2422. TextureVK& texture = m_textures[oldFrameBuffer.m_depth.idx];
  2423. const bool writeOnly = 0 != (texture.m_flags&BGFX_TEXTURE_RT_WRITE_ONLY);
  2424. if (!writeOnly)
  2425. {
  2426. texture.setImageMemoryBarrier(m_commandBuffer, texture.m_sampledLayout);
  2427. }
  2428. }
  2429. }
  2430. if (NULL == newFrameBuffer.m_nwh)
  2431. {
  2432. for (uint8_t ii = 0, num = newFrameBuffer.m_num; ii < num; ++ii)
  2433. {
  2434. TextureVK& texture = m_textures[newFrameBuffer.m_texture[ii].idx];
  2435. texture.setImageMemoryBarrier(
  2436. m_commandBuffer
  2437. , VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  2438. );
  2439. }
  2440. if (isValid(newFrameBuffer.m_depth) )
  2441. {
  2442. TextureVK& texture = m_textures[newFrameBuffer.m_depth.idx];
  2443. texture.setImageMemoryBarrier(
  2444. m_commandBuffer
  2445. , VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL
  2446. );
  2447. }
  2448. newFrameBuffer.acquire(m_commandBuffer);
  2449. }
  2450. if (_acquire)
  2451. {
  2452. int64_t start = bx::getHPCounter();
  2453. newFrameBuffer.acquire(m_commandBuffer);
  2454. int64_t now = bx::getHPCounter();
  2455. if (NULL != newFrameBuffer.m_nwh)
  2456. {
  2457. m_presentElapsed += now - start;
  2458. }
  2459. }
  2460. m_fbh = _fbh;
  2461. }
  2462. void setDebugWireframe(bool _wireframe)
  2463. {
  2464. const bool wireframe = m_deviceFeatures.fillModeNonSolid && _wireframe;
  2465. if (m_wireframe != wireframe)
  2466. {
  2467. m_wireframe = wireframe;
  2468. m_pipelineStateCache.invalidate();
  2469. }
  2470. }
  2471. void setBlendState(VkPipelineColorBlendStateCreateInfo& _desc, uint64_t _state, uint32_t _rgba = 0)
  2472. {
  2473. VkPipelineColorBlendAttachmentState* bas = const_cast<VkPipelineColorBlendAttachmentState*>(_desc.pAttachments);
  2474. uint8_t writeMask = 0;
  2475. writeMask |= (_state & BGFX_STATE_WRITE_R) ? VK_COLOR_COMPONENT_R_BIT : 0;
  2476. writeMask |= (_state & BGFX_STATE_WRITE_G) ? VK_COLOR_COMPONENT_G_BIT : 0;
  2477. writeMask |= (_state & BGFX_STATE_WRITE_B) ? VK_COLOR_COMPONENT_B_BIT : 0;
  2478. writeMask |= (_state & BGFX_STATE_WRITE_A) ? VK_COLOR_COMPONENT_A_BIT : 0;
  2479. bas->blendEnable = !!(BGFX_STATE_BLEND_MASK & _state);
  2480. {
  2481. const uint32_t blend = uint32_t( (_state & BGFX_STATE_BLEND_MASK ) >> BGFX_STATE_BLEND_SHIFT);
  2482. const uint32_t equation = uint32_t( (_state & BGFX_STATE_BLEND_EQUATION_MASK) >> BGFX_STATE_BLEND_EQUATION_SHIFT);
  2483. const uint32_t srcRGB = (blend ) & 0xf;
  2484. const uint32_t dstRGB = (blend >> 4) & 0xf;
  2485. const uint32_t srcA = (blend >> 8) & 0xf;
  2486. const uint32_t dstA = (blend >> 12) & 0xf;
  2487. const uint32_t equRGB = (equation ) & 0x7;
  2488. const uint32_t equA = (equation >> 3) & 0x7;
  2489. bas->srcColorBlendFactor = s_blendFactor[srcRGB][0];
  2490. bas->dstColorBlendFactor = s_blendFactor[dstRGB][0];
  2491. bas->colorBlendOp = s_blendEquation[equRGB];
  2492. bas->srcAlphaBlendFactor = s_blendFactor[srcA][1];
  2493. bas->dstAlphaBlendFactor = s_blendFactor[dstA][1];
  2494. bas->alphaBlendOp = s_blendEquation[equA];
  2495. bas->colorWriteMask = writeMask;
  2496. }
  2497. const FrameBufferVK& frameBuffer = isValid(m_fbh)
  2498. ? m_frameBuffers[m_fbh.idx]
  2499. : m_backBuffer
  2500. ;
  2501. const uint32_t numAttachments = NULL == frameBuffer.m_nwh
  2502. ? frameBuffer.m_num
  2503. : 1
  2504. ;
  2505. if (!!(BGFX_STATE_BLEND_INDEPENDENT & _state)
  2506. && m_deviceFeatures.independentBlend )
  2507. {
  2508. for (uint32_t ii = 1, rgba = _rgba; ii < numAttachments; ++ii, rgba >>= 11)
  2509. {
  2510. ++bas;
  2511. bas->blendEnable = 0 != (rgba & 0x7ff);
  2512. const uint32_t src = (rgba ) & 0xf;
  2513. const uint32_t dst = (rgba >> 4) & 0xf;
  2514. const uint32_t equation = (rgba >> 8) & 0x7;
  2515. bas->srcColorBlendFactor = s_blendFactor[src][0];
  2516. bas->dstColorBlendFactor = s_blendFactor[dst][0];
  2517. bas->colorBlendOp = s_blendEquation[equation];
  2518. bas->srcAlphaBlendFactor = s_blendFactor[src][1];
  2519. bas->dstAlphaBlendFactor = s_blendFactor[dst][1];
  2520. bas->alphaBlendOp = s_blendEquation[equation];
  2521. bas->colorWriteMask = writeMask;
  2522. }
  2523. }
  2524. else
  2525. {
  2526. for (uint32_t ii = 1; ii < numAttachments; ++ii)
  2527. {
  2528. bx::memCopy(&bas[ii], bas, sizeof(VkPipelineColorBlendAttachmentState) );
  2529. }
  2530. }
  2531. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_COLOR_BLEND_STATE_CREATE_INFO;
  2532. _desc.pNext = NULL;
  2533. _desc.flags = 0;
  2534. _desc.logicOpEnable = VK_FALSE;
  2535. _desc.logicOp = VK_LOGIC_OP_CLEAR;
  2536. _desc.attachmentCount = numAttachments;
  2537. _desc.blendConstants[0] = 0.0f;
  2538. _desc.blendConstants[1] = 0.0f;
  2539. _desc.blendConstants[2] = 0.0f;
  2540. _desc.blendConstants[3] = 0.0f;
  2541. }
  2542. void setRasterizerState(VkPipelineRasterizationStateCreateInfo& _desc, uint64_t _state, bool _wireframe = false)
  2543. {
  2544. const uint32_t cull = (_state&BGFX_STATE_CULL_MASK) >> BGFX_STATE_CULL_SHIFT;
  2545. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_STATE_CREATE_INFO;
  2546. _desc.pNext = NULL;
  2547. _desc.flags = 0;
  2548. _desc.depthClampEnable = m_deviceFeatures.depthClamp && m_depthClamp;
  2549. _desc.rasterizerDiscardEnable = VK_FALSE;
  2550. _desc.polygonMode = m_deviceFeatures.fillModeNonSolid && _wireframe
  2551. ? VK_POLYGON_MODE_LINE
  2552. : VK_POLYGON_MODE_FILL
  2553. ;
  2554. _desc.cullMode = s_cullMode[cull];
  2555. _desc.frontFace = (_state&BGFX_STATE_FRONT_CCW) ? VK_FRONT_FACE_COUNTER_CLOCKWISE : VK_FRONT_FACE_CLOCKWISE;
  2556. _desc.depthBiasEnable = VK_FALSE;
  2557. _desc.depthBiasConstantFactor = 0.0f;
  2558. _desc.depthBiasClamp = 0.0f;
  2559. _desc.depthBiasSlopeFactor = 0.0f;
  2560. _desc.lineWidth = 1.0f;
  2561. }
  2562. void setConservativeRasterizerState(VkPipelineRasterizationConservativeStateCreateInfoEXT& _desc, uint64_t _state)
  2563. {
  2564. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_CONSERVATIVE_STATE_CREATE_INFO_EXT;
  2565. _desc.pNext = NULL;
  2566. _desc.flags = 0;
  2567. _desc.conservativeRasterizationMode = (_state&BGFX_STATE_CONSERVATIVE_RASTER)
  2568. ? VK_CONSERVATIVE_RASTERIZATION_MODE_OVERESTIMATE_EXT
  2569. : VK_CONSERVATIVE_RASTERIZATION_MODE_DISABLED_EXT
  2570. ;
  2571. _desc.extraPrimitiveOverestimationSize = 0.0f;
  2572. }
  2573. void setLineRasterizerState(VkPipelineRasterizationLineStateCreateInfoEXT& _desc, uint64_t _state)
  2574. {
  2575. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_RASTERIZATION_LINE_STATE_CREATE_INFO_EXT;
  2576. _desc.pNext = NULL;
  2577. _desc.lineRasterizationMode = (_state & BGFX_STATE_LINEAA)
  2578. ? VK_LINE_RASTERIZATION_MODE_RECTANGULAR_SMOOTH_EXT
  2579. : VK_LINE_RASTERIZATION_MODE_DEFAULT_EXT
  2580. ;
  2581. _desc.stippledLineEnable = VK_FALSE;
  2582. _desc.lineStippleFactor = 0;
  2583. _desc.lineStipplePattern = 0;
  2584. }
  2585. void setDepthStencilState(VkPipelineDepthStencilStateCreateInfo& _desc, uint64_t _state, uint64_t _stencil = 0)
  2586. {
  2587. const uint32_t fstencil = unpackStencil(0, _stencil);
  2588. uint32_t func = (_state&BGFX_STATE_DEPTH_TEST_MASK)>>BGFX_STATE_DEPTH_TEST_SHIFT;
  2589. _desc.sType = VK_STRUCTURE_TYPE_PIPELINE_DEPTH_STENCIL_STATE_CREATE_INFO;
  2590. _desc.pNext = NULL;
  2591. _desc.flags = 0;
  2592. _desc.depthTestEnable = 0 != func;
  2593. _desc.depthWriteEnable = !!(BGFX_STATE_WRITE_Z & _state);
  2594. _desc.depthCompareOp = s_cmpFunc[func];
  2595. _desc.depthBoundsTestEnable = VK_FALSE;
  2596. _desc.stencilTestEnable = 0 != _stencil;
  2597. uint32_t bstencil = unpackStencil(1, _stencil);
  2598. uint32_t frontAndBack = bstencil != BGFX_STENCIL_NONE && bstencil != fstencil;
  2599. bstencil = frontAndBack ? bstencil : fstencil;
  2600. _desc.front.failOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_FAIL_S_MASK) >> BGFX_STENCIL_OP_FAIL_S_SHIFT];
  2601. _desc.front.passOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_PASS_Z_MASK) >> BGFX_STENCIL_OP_PASS_Z_SHIFT];
  2602. _desc.front.depthFailOp = s_stencilOp[(fstencil & BGFX_STENCIL_OP_FAIL_Z_MASK) >> BGFX_STENCIL_OP_FAIL_Z_SHIFT];
  2603. _desc.front.compareOp = s_cmpFunc[(fstencil & BGFX_STENCIL_TEST_MASK) >> BGFX_STENCIL_TEST_SHIFT];
  2604. _desc.front.compareMask = UINT32_MAX;
  2605. _desc.front.writeMask = UINT32_MAX;
  2606. _desc.front.reference = 0;
  2607. _desc.back.failOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_FAIL_S_MASK) >> BGFX_STENCIL_OP_FAIL_S_SHIFT];
  2608. _desc.back.passOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_PASS_Z_MASK) >> BGFX_STENCIL_OP_PASS_Z_SHIFT];
  2609. _desc.back.depthFailOp = s_stencilOp[(bstencil & BGFX_STENCIL_OP_FAIL_Z_MASK) >> BGFX_STENCIL_OP_FAIL_Z_SHIFT];
  2610. _desc.back.compareOp = s_cmpFunc[(bstencil&BGFX_STENCIL_TEST_MASK) >> BGFX_STENCIL_TEST_SHIFT];
  2611. _desc.back.compareMask = UINT32_MAX;
  2612. _desc.back.writeMask = UINT32_MAX;
  2613. _desc.back.reference = 0;
  2614. _desc.minDepthBounds = 0.0f;
  2615. _desc.maxDepthBounds = 1.0f;
  2616. }
  2617. void setInputLayout(VkPipelineVertexInputStateCreateInfo& _vertexInputState, uint8_t _numStream, const VertexLayout** _layout, const ProgramVK& _program, uint8_t _numInstanceData)
  2618. {
  2619. _vertexInputState.sType = VK_STRUCTURE_TYPE_PIPELINE_VERTEX_INPUT_STATE_CREATE_INFO;
  2620. _vertexInputState.pNext = NULL;
  2621. _vertexInputState.flags = 0;
  2622. _vertexInputState.vertexBindingDescriptionCount = 0;
  2623. _vertexInputState.vertexAttributeDescriptionCount = 0;
  2624. uint16_t unsettedAttr[Attrib::Count];
  2625. bx::memCopy(unsettedAttr, _program.m_vsh->m_attrMask, sizeof(uint16_t) * Attrib::Count);
  2626. for (uint8_t stream = 0; stream < _numStream; ++stream)
  2627. {
  2628. VertexLayout layout;
  2629. bx::memCopy(&layout, _layout[stream], sizeof(VertexLayout) );
  2630. const uint16_t* attrMask = _program.m_vsh->m_attrMask;
  2631. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  2632. {
  2633. uint16_t mask = attrMask[ii];
  2634. uint16_t attr = (layout.m_attributes[ii] & mask);
  2635. layout.m_attributes[ii] = attr == 0 || attr == UINT16_MAX ? UINT16_MAX : attr;
  2636. if (unsettedAttr[ii] && attr != UINT16_MAX)
  2637. {
  2638. unsettedAttr[ii] = 0;
  2639. }
  2640. }
  2641. fillVertexLayout(_program.m_vsh, _vertexInputState, layout);
  2642. }
  2643. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  2644. {
  2645. if (0 < unsettedAttr[ii])
  2646. {
  2647. uint32_t numAttribs = _vertexInputState.vertexAttributeDescriptionCount;
  2648. VkVertexInputAttributeDescription* inputAttrib = const_cast<VkVertexInputAttributeDescription*>(_vertexInputState.pVertexAttributeDescriptions + numAttribs);
  2649. inputAttrib->location = _program.m_vsh->m_attrRemap[ii];
  2650. inputAttrib->binding = 0;
  2651. inputAttrib->format = VK_FORMAT_R32G32B32_SFLOAT;
  2652. inputAttrib->offset = 0;
  2653. _vertexInputState.vertexAttributeDescriptionCount++;
  2654. }
  2655. }
  2656. if (0 < _numInstanceData)
  2657. {
  2658. fillInstanceBinding(_program.m_vsh, _vertexInputState, _numInstanceData);
  2659. }
  2660. }
  2661. VkResult getRenderPass(uint8_t _num, const VkFormat* _formats, const VkImageAspectFlags* _aspects, const bool* _resolve, VkSampleCountFlagBits _samples, ::VkRenderPass* _renderPass)
  2662. {
  2663. VkResult result = VK_SUCCESS;
  2664. if (VK_SAMPLE_COUNT_1_BIT == _samples)
  2665. {
  2666. _resolve = NULL;
  2667. }
  2668. bx::HashMurmur2A hash;
  2669. hash.begin();
  2670. hash.add(_samples);
  2671. hash.add(_formats, sizeof(VkFormat) * _num);
  2672. if (NULL != _resolve)
  2673. {
  2674. hash.add(_resolve, sizeof(bool) * _num);
  2675. }
  2676. uint32_t hashKey = hash.end();
  2677. VkRenderPass renderPass = m_renderPassCache.find(hashKey);
  2678. if (VK_NULL_HANDLE != renderPass)
  2679. {
  2680. *_renderPass = renderPass;
  2681. return result;
  2682. }
  2683. VkAttachmentDescription ad[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS * 2];
  2684. for (uint8_t ii = 0; ii < (_num * 2); ++ii)
  2685. {
  2686. ad[ii].flags = 0;
  2687. ad[ii].format = VK_FORMAT_UNDEFINED;
  2688. ad[ii].samples = _samples;
  2689. ad[ii].loadOp = VK_ATTACHMENT_LOAD_OP_LOAD;
  2690. ad[ii].storeOp = VK_ATTACHMENT_STORE_OP_STORE;
  2691. ad[ii].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
  2692. ad[ii].stencilStoreOp = VK_ATTACHMENT_STORE_OP_DONT_CARE;
  2693. ad[ii].initialLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2694. ad[ii].finalLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2695. }
  2696. VkAttachmentReference colorAr[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2697. VkAttachmentReference resolveAr[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2698. VkAttachmentReference depthAr;
  2699. uint32_t numColorAr = 0;
  2700. uint32_t numResolveAr = 0;
  2701. colorAr[0].attachment = VK_ATTACHMENT_UNUSED;
  2702. colorAr[0].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2703. resolveAr[0].attachment = VK_ATTACHMENT_UNUSED;
  2704. resolveAr[0].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2705. depthAr.attachment = VK_ATTACHMENT_UNUSED;
  2706. depthAr.layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2707. for (uint8_t ii = 0; ii < _num; ++ii)
  2708. {
  2709. ad[ii].format = _formats[ii];
  2710. if (_aspects[ii] & VK_IMAGE_ASPECT_COLOR_BIT)
  2711. {
  2712. colorAr[numColorAr].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2713. colorAr[numColorAr].attachment = ii;
  2714. resolveAr[numColorAr].layout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL;
  2715. resolveAr[numColorAr].attachment = VK_ATTACHMENT_UNUSED;
  2716. if (NULL != _resolve
  2717. && _resolve[ii])
  2718. {
  2719. const uint32_t resolve = _num + numResolveAr;
  2720. ad[resolve].format = _formats[ii];
  2721. ad[resolve].samples = VK_SAMPLE_COUNT_1_BIT;
  2722. ad[resolve].loadOp = VK_ATTACHMENT_LOAD_OP_DONT_CARE;
  2723. resolveAr[numColorAr].attachment = resolve;
  2724. numResolveAr++;
  2725. }
  2726. numColorAr++;
  2727. }
  2728. else if (_aspects[ii] & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  2729. {
  2730. ad[ii].stencilLoadOp = VK_ATTACHMENT_LOAD_OP_LOAD;
  2731. ad[ii].stencilStoreOp = VK_ATTACHMENT_STORE_OP_STORE;
  2732. ad[ii].initialLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2733. ad[ii].finalLayout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2734. depthAr.layout = VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL;
  2735. depthAr.attachment = ii;
  2736. }
  2737. }
  2738. VkSubpassDescription sd[1];
  2739. sd[0].flags = 0;
  2740. sd[0].pipelineBindPoint = VK_PIPELINE_BIND_POINT_GRAPHICS;
  2741. sd[0].inputAttachmentCount = 0;
  2742. sd[0].pInputAttachments = NULL;
  2743. sd[0].colorAttachmentCount = bx::max<uint32_t>(numColorAr, 1);
  2744. sd[0].pColorAttachments = colorAr;
  2745. sd[0].pResolveAttachments = resolveAr;
  2746. sd[0].pDepthStencilAttachment = &depthAr;
  2747. sd[0].preserveAttachmentCount = 0;
  2748. sd[0].pPreserveAttachments = NULL;
  2749. const VkPipelineStageFlags graphicsStages = 0
  2750. | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  2751. | VK_PIPELINE_STAGE_VERTEX_INPUT_BIT
  2752. | VK_PIPELINE_STAGE_VERTEX_SHADER_BIT
  2753. | VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT
  2754. | VK_PIPELINE_STAGE_EARLY_FRAGMENT_TESTS_BIT
  2755. | VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT
  2756. | VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT
  2757. ;
  2758. const VkPipelineStageFlags outsideStages = 0
  2759. | graphicsStages
  2760. | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  2761. | VK_PIPELINE_STAGE_TRANSFER_BIT
  2762. ;
  2763. VkSubpassDependency dep[2];
  2764. dep[0].srcSubpass = VK_SUBPASS_EXTERNAL;
  2765. dep[0].dstSubpass = 0;
  2766. dep[0].srcStageMask = outsideStages;
  2767. dep[0].dstStageMask = graphicsStages;
  2768. dep[0].srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  2769. dep[0].dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  2770. dep[0].dependencyFlags = 0;
  2771. dep[1].srcSubpass = BX_COUNTOF(sd)-1;
  2772. dep[1].dstSubpass = VK_SUBPASS_EXTERNAL;
  2773. dep[1].srcStageMask = graphicsStages;
  2774. dep[1].dstStageMask = outsideStages;
  2775. dep[1].srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT;
  2776. dep[1].dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT;
  2777. dep[1].dependencyFlags = 0;
  2778. VkRenderPassCreateInfo rpi;
  2779. rpi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_CREATE_INFO;
  2780. rpi.pNext = NULL;
  2781. rpi.flags = 0;
  2782. rpi.attachmentCount = _num + numResolveAr;
  2783. rpi.pAttachments = ad;
  2784. rpi.subpassCount = BX_COUNTOF(sd);
  2785. rpi.pSubpasses = sd;
  2786. rpi.dependencyCount = BX_COUNTOF(dep);
  2787. rpi.pDependencies = dep;
  2788. result = vkCreateRenderPass(m_device, &rpi, m_allocatorCb, &renderPass);
  2789. if (VK_SUCCESS != result)
  2790. {
  2791. BX_TRACE("Create render pass error: vkCreateRenderPass failed %d: %s.", result, getName(result) );
  2792. return result;
  2793. }
  2794. m_renderPassCache.add(hashKey, renderPass);
  2795. *_renderPass = renderPass;
  2796. return result;
  2797. }
  2798. VkResult getRenderPass(uint8_t _num, const Attachment* _attachments, ::VkRenderPass* _renderPass)
  2799. {
  2800. VkFormat formats[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2801. VkImageAspectFlags aspects[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  2802. VkSampleCountFlagBits samples = VK_SAMPLE_COUNT_1_BIT;
  2803. for (uint8_t ii = 0; ii < _num; ++ii)
  2804. {
  2805. const TextureVK& texture = m_textures[_attachments[ii].handle.idx];
  2806. formats[ii] = texture.m_format;
  2807. aspects[ii] = texture.m_aspectMask;
  2808. samples = texture.m_sampler.Sample;
  2809. }
  2810. return getRenderPass(_num, formats, aspects, NULL, samples, _renderPass);
  2811. }
  2812. VkResult getRenderPass(const SwapChainVK& swapChain, ::VkRenderPass* _renderPass)
  2813. {
  2814. const VkFormat formats[2] =
  2815. {
  2816. swapChain.m_sci.imageFormat,
  2817. swapChain.m_backBufferDepthStencil.m_format
  2818. };
  2819. const VkImageAspectFlags aspects[2] =
  2820. {
  2821. VK_IMAGE_ASPECT_COLOR_BIT,
  2822. swapChain.m_backBufferDepthStencil.m_aspectMask
  2823. };
  2824. const bool resolve[2] =
  2825. {
  2826. swapChain.m_supportsManualResolve ? false : true,
  2827. false
  2828. };
  2829. const VkSampleCountFlagBits samples = swapChain.m_sampler.Sample;
  2830. return getRenderPass(BX_COUNTOF(formats), formats, aspects, resolve, samples, _renderPass);
  2831. }
  2832. VkSampler getSampler(uint32_t _flags, VkFormat _format, const float _palette[][4])
  2833. {
  2834. uint32_t index = ((_flags & BGFX_SAMPLER_BORDER_COLOR_MASK) >> BGFX_SAMPLER_BORDER_COLOR_SHIFT);
  2835. index = bx::min<uint32_t>(BGFX_CONFIG_MAX_COLOR_PALETTE - 1, index);
  2836. _flags &= BGFX_SAMPLER_BITS_MASK;
  2837. _flags &= ~(m_deviceFeatures.samplerAnisotropy ? 0 : (BGFX_SAMPLER_MIN_ANISOTROPIC | BGFX_SAMPLER_MAG_ANISOTROPIC) );
  2838. // Force both min+max anisotropic, can't be set individually.
  2839. _flags |= 0 != (_flags & (BGFX_SAMPLER_MIN_ANISOTROPIC|BGFX_SAMPLER_MAG_ANISOTROPIC) )
  2840. ? BGFX_SAMPLER_MIN_ANISOTROPIC|BGFX_SAMPLER_MAG_ANISOTROPIC
  2841. : 0
  2842. ;
  2843. const float* rgba = NULL == _palette
  2844. ? NULL
  2845. : _palette[index]
  2846. ;
  2847. const bool needColor = true
  2848. && needBorderColor(_flags)
  2849. && NULL != rgba
  2850. && m_borderColorSupport
  2851. ;
  2852. uint32_t hashKey;
  2853. VkSampler sampler = VK_NULL_HANDLE;
  2854. if (!needColor)
  2855. {
  2856. bx::HashMurmur2A hash;
  2857. hash.begin();
  2858. hash.add(_flags);
  2859. hash.add(-1);
  2860. hash.add(VK_FORMAT_UNDEFINED);
  2861. hashKey = hash.end();
  2862. sampler = m_samplerCache.find(hashKey);
  2863. }
  2864. else
  2865. {
  2866. bx::HashMurmur2A hash;
  2867. hash.begin();
  2868. hash.add(_flags);
  2869. hash.add(index);
  2870. hash.add(_format);
  2871. hashKey = hash.end();
  2872. const uint32_t colorHashKey = m_samplerBorderColorCache.find(hashKey);
  2873. const uint32_t newColorHashKey = bx::hash<bx::HashMurmur2A>(rgba, sizeof(float) * 4);
  2874. if (newColorHashKey == colorHashKey)
  2875. {
  2876. sampler = m_samplerCache.find(hashKey);
  2877. }
  2878. else
  2879. {
  2880. m_samplerBorderColorCache.add(hashKey, newColorHashKey);
  2881. }
  2882. }
  2883. if (VK_NULL_HANDLE != sampler)
  2884. {
  2885. return sampler;
  2886. }
  2887. const uint32_t cmpFunc = (_flags&BGFX_SAMPLER_COMPARE_MASK)>>BGFX_SAMPLER_COMPARE_SHIFT;
  2888. const float maxLodBias = m_deviceProperties.limits.maxSamplerLodBias;
  2889. const float lodBias = bx::clamp(float(BGFX_CONFIG_MIP_LOD_BIAS), -maxLodBias, maxLodBias);
  2890. VkSamplerCreateInfo sci;
  2891. sci.sType = VK_STRUCTURE_TYPE_SAMPLER_CREATE_INFO;
  2892. sci.pNext = NULL;
  2893. sci.flags = 0;
  2894. sci.magFilter = _flags & BGFX_SAMPLER_MAG_POINT ? VK_FILTER_NEAREST : VK_FILTER_LINEAR;
  2895. sci.minFilter = _flags & BGFX_SAMPLER_MIN_POINT ? VK_FILTER_NEAREST : VK_FILTER_LINEAR;
  2896. sci.mipmapMode = _flags & BGFX_SAMPLER_MIP_POINT ? VK_SAMPLER_MIPMAP_MODE_NEAREST : VK_SAMPLER_MIPMAP_MODE_LINEAR;
  2897. sci.addressModeU = s_textureAddress[(_flags&BGFX_SAMPLER_U_MASK)>>BGFX_SAMPLER_U_SHIFT];
  2898. sci.addressModeV = s_textureAddress[(_flags&BGFX_SAMPLER_V_MASK)>>BGFX_SAMPLER_V_SHIFT];
  2899. sci.addressModeW = s_textureAddress[(_flags&BGFX_SAMPLER_W_MASK)>>BGFX_SAMPLER_W_SHIFT];
  2900. sci.mipLodBias = lodBias;
  2901. sci.anisotropyEnable = !!(_flags & (BGFX_SAMPLER_MIN_ANISOTROPIC | BGFX_SAMPLER_MAG_ANISOTROPIC) );
  2902. sci.maxAnisotropy = m_maxAnisotropy;
  2903. sci.compareEnable = 0 != cmpFunc;
  2904. sci.compareOp = s_cmpFunc[cmpFunc];
  2905. sci.minLod = 0.0f;
  2906. sci.maxLod = VK_LOD_CLAMP_NONE;
  2907. sci.borderColor = VK_BORDER_COLOR_FLOAT_OPAQUE_BLACK;
  2908. sci.unnormalizedCoordinates = VK_FALSE;
  2909. VkSamplerCustomBorderColorCreateInfoEXT cbcci;
  2910. if (needColor)
  2911. {
  2912. cbcci.sType = VK_STRUCTURE_TYPE_SAMPLER_CUSTOM_BORDER_COLOR_CREATE_INFO_EXT;
  2913. cbcci.pNext = NULL;
  2914. cbcci.format = _format;
  2915. bx::memCopy(cbcci.customBorderColor.float32, rgba, sizeof(cbcci.customBorderColor.float32) );
  2916. sci.pNext = &cbcci;
  2917. sci.borderColor = VK_BORDER_COLOR_FLOAT_CUSTOM_EXT;
  2918. }
  2919. VK_CHECK(vkCreateSampler(m_device, &sci, m_allocatorCb, &sampler) );
  2920. m_samplerCache.add(hashKey, sampler);
  2921. return sampler;
  2922. }
  2923. VkImageView getCachedImageView(TextureHandle _handle, uint32_t _mip, uint32_t _numMips, VkImageViewType _type, bool _stencil = false)
  2924. {
  2925. const TextureVK& texture = m_textures[_handle.idx];
  2926. _stencil = _stencil && !!(texture.m_aspectMask & VK_IMAGE_ASPECT_STENCIL_BIT);
  2927. bx::HashMurmur2A hash;
  2928. hash.begin();
  2929. hash.add(_handle.idx);
  2930. hash.add(_mip);
  2931. hash.add(_numMips);
  2932. hash.add(_type);
  2933. hash.add(_stencil);
  2934. uint32_t hashKey = hash.end();
  2935. VkImageView* viewCached = m_imageViewCache.find(hashKey);
  2936. if (NULL != viewCached)
  2937. {
  2938. return *viewCached;
  2939. }
  2940. const VkImageAspectFlags aspectMask = 0
  2941. | VK_IMAGE_ASPECT_COLOR_BIT
  2942. | ( _stencil ? VK_IMAGE_ASPECT_STENCIL_BIT : VK_IMAGE_ASPECT_DEPTH_BIT)
  2943. ;
  2944. VkImageView view;
  2945. VK_CHECK(texture.createView(0, texture.m_numSides, _mip, _numMips, _type, aspectMask, false, &view) );
  2946. m_imageViewCache.add(hashKey, view, _handle.idx);
  2947. return view;
  2948. }
  2949. VkPipeline getPipeline(ProgramHandle _program)
  2950. {
  2951. ProgramVK& program = m_program[_program.idx];
  2952. bx::HashMurmur2A murmur;
  2953. murmur.begin();
  2954. murmur.add(program.m_vsh->m_hash);
  2955. const uint32_t hash = murmur.end();
  2956. VkPipeline pipeline = m_pipelineStateCache.find(hash);
  2957. if (VK_NULL_HANDLE != pipeline)
  2958. {
  2959. return pipeline;
  2960. }
  2961. VkComputePipelineCreateInfo cpci;
  2962. cpci.sType = VK_STRUCTURE_TYPE_COMPUTE_PIPELINE_CREATE_INFO;
  2963. cpci.pNext = NULL;
  2964. cpci.flags = 0;
  2965. cpci.stage.sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  2966. cpci.stage.pNext = NULL;
  2967. cpci.stage.flags = 0;
  2968. cpci.stage.stage = VK_SHADER_STAGE_COMPUTE_BIT;
  2969. cpci.stage.module = program.m_vsh->m_module;
  2970. cpci.stage.pName = "main";
  2971. cpci.stage.pSpecializationInfo = NULL;
  2972. cpci.layout = program.m_pipelineLayout;
  2973. cpci.basePipelineHandle = VK_NULL_HANDLE;
  2974. cpci.basePipelineIndex = 0;
  2975. VK_CHECK(vkCreateComputePipelines(m_device, m_pipelineCache, 1, &cpci, m_allocatorCb, &pipeline) );
  2976. m_pipelineStateCache.add(hash, pipeline);
  2977. return pipeline;
  2978. }
  2979. VkPipeline getPipeline(uint64_t _state, uint64_t _stencil, uint8_t _numStreams, const VertexLayout** _layouts, ProgramHandle _program, uint8_t _numInstanceData)
  2980. {
  2981. ProgramVK& program = m_program[_program.idx];
  2982. _state &= 0
  2983. | BGFX_STATE_WRITE_MASK
  2984. | BGFX_STATE_DEPTH_TEST_MASK
  2985. | BGFX_STATE_BLEND_MASK
  2986. | BGFX_STATE_BLEND_EQUATION_MASK
  2987. | (g_caps.supported & BGFX_CAPS_BLEND_INDEPENDENT ? BGFX_STATE_BLEND_INDEPENDENT : 0)
  2988. | BGFX_STATE_BLEND_ALPHA_TO_COVERAGE
  2989. | BGFX_STATE_CULL_MASK
  2990. | BGFX_STATE_FRONT_CCW
  2991. | BGFX_STATE_MSAA
  2992. | (m_lineAASupport ? BGFX_STATE_LINEAA : 0)
  2993. | (g_caps.supported & BGFX_CAPS_CONSERVATIVE_RASTER ? BGFX_STATE_CONSERVATIVE_RASTER : 0)
  2994. | BGFX_STATE_PT_MASK
  2995. ;
  2996. _stencil &= packStencil(~BGFX_STENCIL_FUNC_REF_MASK, ~BGFX_STENCIL_FUNC_REF_MASK);
  2997. VertexLayout layout;
  2998. if (0 < _numStreams)
  2999. {
  3000. bx::memCopy(&layout, _layouts[0], sizeof(VertexLayout) );
  3001. const uint16_t* attrMask = program.m_vsh->m_attrMask;
  3002. for (uint32_t ii = 0; ii < Attrib::Count; ++ii)
  3003. {
  3004. uint16_t mask = attrMask[ii];
  3005. uint16_t attr = (layout.m_attributes[ii] & mask);
  3006. layout.m_attributes[ii] = attr == 0 ? UINT16_MAX : attr == UINT16_MAX ? 0 : attr;
  3007. }
  3008. }
  3009. const FrameBufferVK& frameBuffer = isValid(m_fbh)
  3010. ? m_frameBuffers[m_fbh.idx]
  3011. : m_backBuffer
  3012. ;
  3013. bx::HashMurmur2A murmur;
  3014. murmur.begin();
  3015. murmur.add(_state);
  3016. murmur.add(_stencil);
  3017. murmur.add(program.m_vsh->m_hash);
  3018. murmur.add(program.m_vsh->m_attrMask, sizeof(program.m_vsh->m_attrMask) );
  3019. if (NULL != program.m_fsh)
  3020. {
  3021. murmur.add(program.m_fsh->m_hash);
  3022. }
  3023. for (uint8_t ii = 0; ii < _numStreams; ++ii)
  3024. {
  3025. murmur.add(_layouts[ii]->m_hash);
  3026. }
  3027. murmur.add(layout.m_attributes, sizeof(layout.m_attributes) );
  3028. murmur.add(_numInstanceData);
  3029. murmur.add(frameBuffer.m_renderPass);
  3030. const uint32_t hash = murmur.end();
  3031. VkPipeline pipeline = m_pipelineStateCache.find(hash);
  3032. if (VK_NULL_HANDLE != pipeline)
  3033. {
  3034. return pipeline;
  3035. }
  3036. VkPipelineColorBlendAttachmentState blendAttachmentState[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  3037. VkPipelineColorBlendStateCreateInfo colorBlendState;
  3038. colorBlendState.pAttachments = blendAttachmentState;
  3039. setBlendState(colorBlendState, _state);
  3040. VkPipelineInputAssemblyStateCreateInfo inputAssemblyState;
  3041. inputAssemblyState.sType = VK_STRUCTURE_TYPE_PIPELINE_INPUT_ASSEMBLY_STATE_CREATE_INFO;
  3042. inputAssemblyState.pNext = NULL;
  3043. inputAssemblyState.flags = 0;
  3044. inputAssemblyState.topology = s_primInfo[(_state&BGFX_STATE_PT_MASK) >> BGFX_STATE_PT_SHIFT].m_topology;
  3045. inputAssemblyState.primitiveRestartEnable = VK_FALSE;
  3046. VkPipelineRasterizationStateCreateInfo rasterizationState;
  3047. setRasterizerState(rasterizationState, _state, m_wireframe);
  3048. VkBaseInStructure* nextRasterizationState = (VkBaseInStructure*)&rasterizationState;
  3049. VkPipelineRasterizationConservativeStateCreateInfoEXT conservativeRasterizationState;
  3050. if (s_extension[Extension::EXT_conservative_rasterization].m_supported)
  3051. {
  3052. nextRasterizationState->pNext = (VkBaseInStructure*)&conservativeRasterizationState;
  3053. nextRasterizationState = (VkBaseInStructure*)&conservativeRasterizationState;
  3054. setConservativeRasterizerState(conservativeRasterizationState, _state);
  3055. }
  3056. VkPipelineRasterizationLineStateCreateInfoEXT lineRasterizationState;
  3057. if (m_lineAASupport)
  3058. {
  3059. nextRasterizationState->pNext = (VkBaseInStructure*)&lineRasterizationState;
  3060. nextRasterizationState = (VkBaseInStructure*)&lineRasterizationState;
  3061. setLineRasterizerState(lineRasterizationState, _state);
  3062. }
  3063. VkPipelineDepthStencilStateCreateInfo depthStencilState;
  3064. setDepthStencilState(depthStencilState, _state, _stencil);
  3065. VkVertexInputBindingDescription inputBinding[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  3066. VkVertexInputAttributeDescription inputAttrib[Attrib::Count + BGFX_CONFIG_MAX_INSTANCE_DATA_COUNT];
  3067. VkPipelineVertexInputStateCreateInfo vertexInputState;
  3068. vertexInputState.pVertexBindingDescriptions = inputBinding;
  3069. vertexInputState.pVertexAttributeDescriptions = inputAttrib;
  3070. setInputLayout(vertexInputState, _numStreams, _layouts, program, _numInstanceData);
  3071. const VkDynamicState dynamicStates[] =
  3072. {
  3073. VK_DYNAMIC_STATE_VIEWPORT,
  3074. VK_DYNAMIC_STATE_SCISSOR,
  3075. VK_DYNAMIC_STATE_BLEND_CONSTANTS,
  3076. VK_DYNAMIC_STATE_STENCIL_REFERENCE,
  3077. };
  3078. VkPipelineDynamicStateCreateInfo dynamicState;
  3079. dynamicState.sType = VK_STRUCTURE_TYPE_PIPELINE_DYNAMIC_STATE_CREATE_INFO;
  3080. dynamicState.pNext = NULL;
  3081. dynamicState.flags = 0;
  3082. dynamicState.dynamicStateCount = BX_COUNTOF(dynamicStates);
  3083. dynamicState.pDynamicStates = dynamicStates;
  3084. VkPipelineShaderStageCreateInfo shaderStages[2];
  3085. shaderStages[0].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  3086. shaderStages[0].pNext = NULL;
  3087. shaderStages[0].flags = 0;
  3088. shaderStages[0].stage = VK_SHADER_STAGE_VERTEX_BIT;
  3089. shaderStages[0].module = program.m_vsh->m_module;
  3090. shaderStages[0].pName = "main";
  3091. shaderStages[0].pSpecializationInfo = NULL;
  3092. if (NULL != program.m_fsh)
  3093. {
  3094. shaderStages[1].sType = VK_STRUCTURE_TYPE_PIPELINE_SHADER_STAGE_CREATE_INFO;
  3095. shaderStages[1].pNext = NULL;
  3096. shaderStages[1].flags = 0;
  3097. shaderStages[1].stage = VK_SHADER_STAGE_FRAGMENT_BIT;
  3098. shaderStages[1].module = program.m_fsh->m_module;
  3099. shaderStages[1].pName = "main";
  3100. shaderStages[1].pSpecializationInfo = NULL;
  3101. }
  3102. VkPipelineViewportStateCreateInfo viewportState;
  3103. viewportState.sType = VK_STRUCTURE_TYPE_PIPELINE_VIEWPORT_STATE_CREATE_INFO;
  3104. viewportState.pNext = NULL;
  3105. viewportState.flags = 0;
  3106. viewportState.viewportCount = 1;
  3107. viewportState.pViewports = NULL;
  3108. viewportState.scissorCount = 1;
  3109. viewportState.pScissors = NULL;
  3110. VkPipelineMultisampleStateCreateInfo multisampleState;
  3111. multisampleState.sType = VK_STRUCTURE_TYPE_PIPELINE_MULTISAMPLE_STATE_CREATE_INFO;
  3112. multisampleState.pNext = NULL;
  3113. multisampleState.flags = 0;
  3114. multisampleState.rasterizationSamples = frameBuffer.m_sampler.Sample;
  3115. multisampleState.sampleShadingEnable = VK_FALSE;
  3116. multisampleState.minSampleShading = 0.0f;
  3117. multisampleState.pSampleMask = NULL;
  3118. multisampleState.alphaToCoverageEnable = !!(BGFX_STATE_BLEND_ALPHA_TO_COVERAGE & _state);
  3119. multisampleState.alphaToOneEnable = VK_FALSE;
  3120. VkGraphicsPipelineCreateInfo graphicsPipeline;
  3121. graphicsPipeline.sType = VK_STRUCTURE_TYPE_GRAPHICS_PIPELINE_CREATE_INFO;
  3122. graphicsPipeline.pNext = NULL;
  3123. graphicsPipeline.flags = 0;
  3124. graphicsPipeline.stageCount = NULL == program.m_fsh ? 1 : 2;
  3125. graphicsPipeline.pStages = shaderStages;
  3126. graphicsPipeline.pVertexInputState = &vertexInputState;
  3127. graphicsPipeline.pInputAssemblyState = &inputAssemblyState;
  3128. graphicsPipeline.pTessellationState = NULL;
  3129. graphicsPipeline.pViewportState = &viewportState;
  3130. graphicsPipeline.pRasterizationState = &rasterizationState;
  3131. graphicsPipeline.pMultisampleState = &multisampleState;
  3132. graphicsPipeline.pDepthStencilState = &depthStencilState;
  3133. graphicsPipeline.pColorBlendState = &colorBlendState;
  3134. graphicsPipeline.pDynamicState = &dynamicState;
  3135. graphicsPipeline.layout = program.m_pipelineLayout;
  3136. graphicsPipeline.renderPass = frameBuffer.m_renderPass;
  3137. graphicsPipeline.subpass = 0;
  3138. graphicsPipeline.basePipelineHandle = VK_NULL_HANDLE;
  3139. graphicsPipeline.basePipelineIndex = 0;
  3140. uint32_t length = g_callback->cacheReadSize(hash);
  3141. bool cached = length > 0;
  3142. void* cachedData = NULL;
  3143. VkPipelineCacheCreateInfo pcci;
  3144. pcci.sType = VK_STRUCTURE_TYPE_PIPELINE_CACHE_CREATE_INFO;
  3145. pcci.pNext = NULL;
  3146. pcci.flags = 0;
  3147. pcci.initialDataSize = 0;
  3148. pcci.pInitialData = NULL;
  3149. if (cached)
  3150. {
  3151. cachedData = bx::alloc(g_allocator, length);
  3152. if (g_callback->cacheRead(hash, cachedData, length) )
  3153. {
  3154. BX_TRACE("Loading cached pipeline state (size %d).", length);
  3155. bx::MemoryReader reader(cachedData, length);
  3156. pcci.initialDataSize = (size_t)reader.remaining();
  3157. pcci.pInitialData = reader.getDataPtr();
  3158. }
  3159. }
  3160. VkPipelineCache cache;
  3161. VK_CHECK(vkCreatePipelineCache(m_device, &pcci, m_allocatorCb, &cache) );
  3162. VK_CHECK(vkCreateGraphicsPipelines(
  3163. m_device
  3164. , cache
  3165. , 1
  3166. , &graphicsPipeline
  3167. , m_allocatorCb
  3168. , &pipeline
  3169. ) );
  3170. m_pipelineStateCache.add(hash, pipeline);
  3171. size_t dataSize;
  3172. VK_CHECK(vkGetPipelineCacheData(m_device, cache, &dataSize, NULL) );
  3173. if (0 < dataSize)
  3174. {
  3175. if (length < dataSize)
  3176. {
  3177. cachedData = bx::realloc(g_allocator, cachedData, dataSize);
  3178. }
  3179. VK_CHECK(vkGetPipelineCacheData(m_device, cache, &dataSize, cachedData) );
  3180. g_callback->cacheWrite(hash, cachedData, (uint32_t)dataSize);
  3181. }
  3182. VK_CHECK(vkMergePipelineCaches(m_device, m_pipelineCache, 1, &cache) );
  3183. vkDestroy(cache);
  3184. if (NULL != cachedData)
  3185. {
  3186. bx::free(g_allocator, cachedData);
  3187. }
  3188. return pipeline;
  3189. }
  3190. VkDescriptorSet getDescriptorSet(const ProgramVK& program, const RenderBind& renderBind, const ScratchBufferVK& scratchBuffer, const float _palette[][4])
  3191. {
  3192. VkDescriptorSet descriptorSet;
  3193. VkDescriptorSetAllocateInfo dsai;
  3194. dsai.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_ALLOCATE_INFO;
  3195. dsai.pNext = NULL;
  3196. dsai.descriptorPool = m_descriptorPool;
  3197. dsai.descriptorSetCount = 1;
  3198. dsai.pSetLayouts = &program.m_descriptorSetLayout;
  3199. VK_CHECK(vkAllocateDescriptorSets(m_device, &dsai, &descriptorSet) );
  3200. VkDescriptorImageInfo imageInfo[BGFX_CONFIG_MAX_TEXTURE_SAMPLERS];
  3201. VkDescriptorBufferInfo bufferInfo[BGFX_CONFIG_MAX_TEXTURE_SAMPLERS];
  3202. constexpr uint32_t kMaxDescriptorSets = 2 * BGFX_CONFIG_MAX_TEXTURE_SAMPLERS + 2;
  3203. VkWriteDescriptorSet wds[kMaxDescriptorSets] = {};
  3204. uint32_t wdsCount = 0;
  3205. uint32_t bufferCount = 0;
  3206. uint32_t imageCount = 0;
  3207. for (uint32_t stage = 0; stage < BGFX_CONFIG_MAX_TEXTURE_SAMPLERS; ++stage)
  3208. {
  3209. const Binding& bind = renderBind.m_bind[stage];
  3210. const BindInfo& bindInfo = program.m_bindInfo[stage];
  3211. if (kInvalidHandle != bind.m_idx
  3212. && isValid(bindInfo.uniformHandle) )
  3213. {
  3214. switch (bind.m_type)
  3215. {
  3216. case Binding::Image:
  3217. {
  3218. const bool isImageDescriptor = BindType::Image == bindInfo.type;
  3219. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3220. wds[wdsCount].pNext = NULL;
  3221. wds[wdsCount].dstSet = descriptorSet;
  3222. wds[wdsCount].dstBinding = bindInfo.binding;
  3223. wds[wdsCount].dstArrayElement = 0;
  3224. wds[wdsCount].descriptorCount = 1;
  3225. wds[wdsCount].descriptorType = isImageDescriptor
  3226. ? VK_DESCRIPTOR_TYPE_STORAGE_IMAGE
  3227. : VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE
  3228. ;
  3229. wds[wdsCount].pImageInfo = NULL;
  3230. wds[wdsCount].pBufferInfo = NULL;
  3231. wds[wdsCount].pTexelBufferView = NULL;
  3232. const TextureVK& texture = m_textures[bind.m_idx];
  3233. VkImageViewType type = texture.m_type;
  3234. if (UINT32_MAX != bindInfo.index)
  3235. {
  3236. type = program.m_textures[bindInfo.index].type;
  3237. }
  3238. else if (type == VK_IMAGE_VIEW_TYPE_CUBE
  3239. || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY)
  3240. {
  3241. type = VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  3242. }
  3243. BX_ASSERT(
  3244. texture.m_currentImageLayout == texture.m_sampledLayout
  3245. , "Mismatching image layout. Texture currently used as a framebuffer attachment?"
  3246. );
  3247. imageInfo[imageCount].imageLayout = texture.m_sampledLayout;
  3248. imageInfo[imageCount].sampler = VK_NULL_HANDLE;
  3249. imageInfo[imageCount].imageView = getCachedImageView(
  3250. { bind.m_idx }
  3251. , bind.m_mip
  3252. , 1
  3253. , type
  3254. );
  3255. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3256. ++imageCount;
  3257. ++wdsCount;
  3258. }
  3259. break;
  3260. case Binding::VertexBuffer:
  3261. case Binding::IndexBuffer:
  3262. {
  3263. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3264. wds[wdsCount].pNext = NULL;
  3265. wds[wdsCount].dstSet = descriptorSet;
  3266. wds[wdsCount].dstBinding = bindInfo.binding;
  3267. wds[wdsCount].dstArrayElement = 0;
  3268. wds[wdsCount].descriptorCount = 1;
  3269. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_STORAGE_BUFFER;
  3270. wds[wdsCount].pImageInfo = NULL;
  3271. wds[wdsCount].pBufferInfo = NULL;
  3272. wds[wdsCount].pTexelBufferView = NULL;
  3273. const BufferVK& sb = bind.m_type == Binding::VertexBuffer
  3274. ? m_vertexBuffers[bind.m_idx]
  3275. : m_indexBuffers[bind.m_idx]
  3276. ;
  3277. bufferInfo[bufferCount].buffer = sb.m_buffer;
  3278. bufferInfo[bufferCount].offset = 0;
  3279. bufferInfo[bufferCount].range = sb.m_size;
  3280. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3281. ++bufferCount;
  3282. ++wdsCount;
  3283. }
  3284. break;
  3285. case Binding::Texture:
  3286. {
  3287. TextureVK& texture = m_textures[bind.m_idx];
  3288. const uint32_t samplerFlags = 0 == (BGFX_SAMPLER_INTERNAL_DEFAULT & bind.m_samplerFlags)
  3289. ? bind.m_samplerFlags
  3290. : (uint32_t)texture.m_flags
  3291. ;
  3292. const bool sampleStencil = !!(samplerFlags & BGFX_SAMPLER_SAMPLE_STENCIL);
  3293. VkSampler sampler = getSampler(samplerFlags, texture.m_format, _palette);
  3294. const VkImageViewType type = UINT32_MAX == bindInfo.index
  3295. ? texture.m_type
  3296. : program.m_textures[bindInfo.index].type
  3297. ;
  3298. BX_ASSERT(
  3299. texture.m_currentImageLayout == texture.m_sampledLayout
  3300. , "Mismatching image layout. Texture currently used as a framebuffer attachment?"
  3301. );
  3302. imageInfo[imageCount].imageLayout = texture.m_sampledLayout;
  3303. imageInfo[imageCount].sampler = sampler;
  3304. imageInfo[imageCount].imageView = getCachedImageView(
  3305. { bind.m_idx }
  3306. , 0
  3307. , texture.m_numMips
  3308. , type
  3309. , sampleStencil
  3310. );
  3311. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3312. wds[wdsCount].pNext = NULL;
  3313. wds[wdsCount].dstSet = descriptorSet;
  3314. wds[wdsCount].dstBinding = bindInfo.binding;
  3315. wds[wdsCount].dstArrayElement = 0;
  3316. wds[wdsCount].descriptorCount = 1;
  3317. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
  3318. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3319. wds[wdsCount].pBufferInfo = NULL;
  3320. wds[wdsCount].pTexelBufferView = NULL;
  3321. ++wdsCount;
  3322. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3323. wds[wdsCount].pNext = NULL;
  3324. wds[wdsCount].dstSet = descriptorSet;
  3325. wds[wdsCount].dstBinding = bindInfo.samplerBinding;
  3326. wds[wdsCount].dstArrayElement = 0;
  3327. wds[wdsCount].descriptorCount = 1;
  3328. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
  3329. wds[wdsCount].pImageInfo = &imageInfo[imageCount];
  3330. wds[wdsCount].pBufferInfo = NULL;
  3331. wds[wdsCount].pTexelBufferView = NULL;
  3332. ++wdsCount;
  3333. ++imageCount;
  3334. }
  3335. break;
  3336. }
  3337. }
  3338. }
  3339. const uint32_t vsize = program.m_vsh->m_size;
  3340. const uint32_t fsize = NULL != program.m_fsh ? program.m_fsh->m_size : 0;
  3341. if (vsize > 0)
  3342. {
  3343. bufferInfo[bufferCount].buffer = scratchBuffer.m_buffer;
  3344. bufferInfo[bufferCount].offset = 0;
  3345. bufferInfo[bufferCount].range = vsize;
  3346. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3347. wds[wdsCount].pNext = NULL;
  3348. wds[wdsCount].dstSet = descriptorSet;
  3349. wds[wdsCount].dstBinding = program.m_vsh->m_uniformBinding;
  3350. wds[wdsCount].dstArrayElement = 0;
  3351. wds[wdsCount].descriptorCount = 1;
  3352. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  3353. wds[wdsCount].pImageInfo = NULL;
  3354. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3355. wds[wdsCount].pTexelBufferView = NULL;
  3356. ++wdsCount;
  3357. ++bufferCount;
  3358. }
  3359. if (fsize > 0)
  3360. {
  3361. bufferInfo[bufferCount].buffer = scratchBuffer.m_buffer;
  3362. bufferInfo[bufferCount].offset = 0;
  3363. bufferInfo[bufferCount].range = fsize;
  3364. wds[wdsCount].sType = VK_STRUCTURE_TYPE_WRITE_DESCRIPTOR_SET;
  3365. wds[wdsCount].pNext = NULL;
  3366. wds[wdsCount].dstSet = descriptorSet;
  3367. wds[wdsCount].dstBinding = program.m_fsh->m_uniformBinding;
  3368. wds[wdsCount].dstArrayElement = 0;
  3369. wds[wdsCount].descriptorCount = 1;
  3370. wds[wdsCount].descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  3371. wds[wdsCount].pImageInfo = NULL;
  3372. wds[wdsCount].pBufferInfo = &bufferInfo[bufferCount];
  3373. wds[wdsCount].pTexelBufferView = NULL;
  3374. ++wdsCount;
  3375. ++bufferCount;
  3376. }
  3377. vkUpdateDescriptorSets(m_device, wdsCount, wds, 0, NULL);
  3378. VkDescriptorSet temp = descriptorSet;
  3379. release(temp);
  3380. return descriptorSet;
  3381. }
  3382. bool isSwapChainReadable(const SwapChainVK& _swapChain)
  3383. {
  3384. return true
  3385. && NULL != _swapChain.m_nwh
  3386. && _swapChain.m_needPresent
  3387. && _swapChain.m_supportsReadback
  3388. && bimg::imageConvert(bimg::TextureFormat::BGRA8, bimg::TextureFormat::Enum(_swapChain.m_colorFormat) )
  3389. ;
  3390. }
  3391. typedef void (*SwapChainReadFunc)(void* /*src*/, uint32_t /*width*/, uint32_t /*height*/, uint32_t /*pitch*/, const void* /*userData*/);
  3392. bool readSwapChain(const SwapChainVK& _swapChain, VkBuffer _buffer, VkDeviceMemory _memory, SwapChainReadFunc _func, const void* _userData = NULL)
  3393. {
  3394. if (isSwapChainReadable(_swapChain) )
  3395. {
  3396. // source for the copy is the last rendered swapchain image
  3397. const VkImage image = _swapChain.m_backBufferColorImage[_swapChain.m_backBufferColorIdx];
  3398. const VkImageLayout layout = _swapChain.m_backBufferColorImageLayout[_swapChain.m_backBufferColorIdx];
  3399. const uint32_t width = _swapChain.m_sci.imageExtent.width;
  3400. const uint32_t height = _swapChain.m_sci.imageExtent.height;
  3401. ReadbackVK readback;
  3402. readback.create(image, width, height, _swapChain.m_colorFormat);
  3403. const uint32_t pitch = readback.pitch();
  3404. readback.copyImageToBuffer(m_commandBuffer, _buffer, layout, VK_IMAGE_ASPECT_COLOR_BIT);
  3405. // stall for commandbuffer to finish
  3406. kick(true);
  3407. uint8_t* src;
  3408. VK_CHECK(vkMapMemory(m_device, _memory, 0, VK_WHOLE_SIZE, 0, (void**)&src) );
  3409. if (_swapChain.m_colorFormat == TextureFormat::RGBA8)
  3410. {
  3411. bimg::imageSwizzleBgra8(src, pitch, width, height, src, pitch);
  3412. _func(src, width, height, pitch, _userData);
  3413. }
  3414. else if (_swapChain.m_colorFormat == TextureFormat::BGRA8)
  3415. {
  3416. _func(src, width, height, pitch, _userData);
  3417. }
  3418. else
  3419. {
  3420. const uint8_t dstBpp = bimg::getBitsPerPixel(bimg::TextureFormat::BGRA8);
  3421. const uint32_t dstPitch = width * dstBpp / 8;
  3422. const uint32_t dstSize = height * dstPitch;
  3423. void* dst = bx::alloc(g_allocator, dstSize);
  3424. bimg::imageConvert(g_allocator, dst, bimg::TextureFormat::BGRA8, src, bimg::TextureFormat::Enum(_swapChain.m_colorFormat), width, height, 1);
  3425. _func(dst, width, height, dstPitch, _userData);
  3426. bx::free(g_allocator, dst);
  3427. }
  3428. vkUnmapMemory(m_device, _memory);
  3429. readback.destroy();
  3430. return true;
  3431. }
  3432. return false;
  3433. }
  3434. void capture()
  3435. {
  3436. if (m_captureSize > 0)
  3437. {
  3438. m_backBuffer.resolve();
  3439. auto callback = [](void* _src, uint32_t /*_width*/, uint32_t _height, uint32_t _pitch, const void* /*_userData*/)
  3440. {
  3441. const uint32_t size = _height * _pitch;
  3442. g_callback->captureFrame(_src, size);
  3443. };
  3444. readSwapChain(m_backBuffer.m_swapChain, m_captureBuffer, m_captureMemory, callback);
  3445. }
  3446. }
  3447. bool isVisible(Frame* _render, OcclusionQueryHandle _handle, bool _visible)
  3448. {
  3449. return _visible == (0 != _render->m_occlusion[_handle.idx]);
  3450. }
  3451. void commit(UniformBuffer& _uniformBuffer)
  3452. {
  3453. _uniformBuffer.reset();
  3454. for (;;)
  3455. {
  3456. uint32_t opcode = _uniformBuffer.read();
  3457. if (UniformType::End == opcode)
  3458. {
  3459. break;
  3460. }
  3461. UniformType::Enum type;
  3462. uint16_t loc;
  3463. uint16_t num;
  3464. uint16_t copy;
  3465. UniformBuffer::decodeOpcode(opcode, type, loc, num, copy);
  3466. const char* data;
  3467. if (copy)
  3468. {
  3469. data = _uniformBuffer.read(g_uniformTypeSize[type]*num);
  3470. }
  3471. else
  3472. {
  3473. UniformHandle handle;
  3474. bx::memCopy(&handle, _uniformBuffer.read(sizeof(UniformHandle) ), sizeof(UniformHandle) );
  3475. data = (const char*)m_uniforms[handle.idx];
  3476. }
  3477. switch ( (uint32_t)type)
  3478. {
  3479. case UniformType::Mat3:
  3480. case UniformType::Mat3|kUniformFragmentBit:
  3481. {
  3482. float* value = (float*)data;
  3483. for (uint32_t ii = 0, count = num/3; ii < count; ++ii, loc += 3*16, value += 9)
  3484. {
  3485. Matrix4 mtx;
  3486. mtx.un.val[ 0] = value[0];
  3487. mtx.un.val[ 1] = value[1];
  3488. mtx.un.val[ 2] = value[2];
  3489. mtx.un.val[ 3] = 0.0f;
  3490. mtx.un.val[ 4] = value[3];
  3491. mtx.un.val[ 5] = value[4];
  3492. mtx.un.val[ 6] = value[5];
  3493. mtx.un.val[ 7] = 0.0f;
  3494. mtx.un.val[ 8] = value[6];
  3495. mtx.un.val[ 9] = value[7];
  3496. mtx.un.val[10] = value[8];
  3497. mtx.un.val[11] = 0.0f;
  3498. setShaderUniform(uint8_t(type), loc, &mtx.un.val[0], 3);
  3499. }
  3500. }
  3501. break;
  3502. case UniformType::Sampler:
  3503. case UniformType::Sampler|kUniformFragmentBit:
  3504. // do nothing, but VkDescriptorSetImageInfo would be set before drawing
  3505. break;
  3506. case UniformType::Vec4:
  3507. case UniformType::Vec4 | kUniformFragmentBit:
  3508. case UniformType::Mat4:
  3509. case UniformType::Mat4 | kUniformFragmentBit:
  3510. {
  3511. setShaderUniform(uint8_t(type), loc, data, num);
  3512. }
  3513. break;
  3514. case UniformType::End:
  3515. break;
  3516. default:
  3517. BX_TRACE("%4d: INVALID 0x%08x, t %d, l %d, n %d, c %d", _uniformBuffer.getPos(), opcode, type, loc, num, copy);
  3518. break;
  3519. }
  3520. }
  3521. }
  3522. void clearQuad(const Rect& _rect, const Clear& _clear, const float _palette[][4])
  3523. {
  3524. VkClearRect rect[1];
  3525. rect[0].rect.offset.x = _rect.m_x;
  3526. rect[0].rect.offset.y = _rect.m_y;
  3527. rect[0].rect.extent.width = _rect.m_width;
  3528. rect[0].rect.extent.height = _rect.m_height;
  3529. rect[0].baseArrayLayer = 0;
  3530. rect[0].layerCount = 1;
  3531. uint32_t numMrt;
  3532. bgfx::TextureFormat::Enum mrtFormat[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS];
  3533. VkImageAspectFlags depthAspectMask;
  3534. const FrameBufferVK& fb = isValid(m_fbh)
  3535. ? m_frameBuffers[m_fbh.idx]
  3536. : m_backBuffer
  3537. ;
  3538. if (NULL == fb.m_nwh)
  3539. {
  3540. numMrt = fb.m_num;
  3541. for (uint8_t ii = 0; ii < fb.m_num; ++ii)
  3542. {
  3543. mrtFormat[ii] = bgfx::TextureFormat::Enum(m_textures[fb.m_texture[ii].idx].m_requestedFormat);
  3544. }
  3545. depthAspectMask = isValid(fb.m_depth) ? m_textures[fb.m_depth.idx].m_aspectMask : 0;
  3546. rect[0].layerCount = fb.m_attachment[0].numLayers;
  3547. }
  3548. else
  3549. {
  3550. numMrt = 1;
  3551. mrtFormat[0] = fb.m_swapChain.m_colorFormat;
  3552. depthAspectMask = fb.m_swapChain.m_backBufferDepthStencil.m_aspectMask;
  3553. }
  3554. VkClearAttachment attachments[BGFX_CONFIG_MAX_FRAME_BUFFER_ATTACHMENTS + 1];
  3555. uint32_t mrt = 0;
  3556. if (BGFX_CLEAR_COLOR & _clear.m_flags)
  3557. {
  3558. for (uint32_t ii = 0; ii < numMrt; ++ii)
  3559. {
  3560. attachments[mrt].colorAttachment = mrt;
  3561. attachments[mrt].aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  3562. VkClearColorValue& clearValue = attachments[mrt].clearValue.color;
  3563. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(bimg::TextureFormat::Enum(mrtFormat[ii]) );
  3564. const bx::EncodingType::Enum type = bx::EncodingType::Enum(blockInfo.encoding);
  3565. if (BGFX_CLEAR_COLOR_USE_PALETTE & _clear.m_flags)
  3566. {
  3567. const uint8_t index = bx::min<uint8_t>(BGFX_CONFIG_MAX_COLOR_PALETTE-1, _clear.m_index[ii]);
  3568. switch (type)
  3569. {
  3570. case bx::EncodingType::Int:
  3571. case bx::EncodingType::Uint:
  3572. clearValue.int32[0] = int32_t(_palette[index][0]);
  3573. clearValue.int32[1] = int32_t(_palette[index][1]);
  3574. clearValue.int32[2] = int32_t(_palette[index][2]);
  3575. clearValue.int32[3] = int32_t(_palette[index][3]);
  3576. break;
  3577. default:
  3578. bx::memCopy(&clearValue.float32, _palette[index], sizeof(clearValue.float32) );
  3579. break;
  3580. }
  3581. }
  3582. else
  3583. {
  3584. switch (type)
  3585. {
  3586. case bx::EncodingType::Int:
  3587. case bx::EncodingType::Uint:
  3588. clearValue.uint32[0] = _clear.m_index[0];
  3589. clearValue.uint32[1] = _clear.m_index[1];
  3590. clearValue.uint32[2] = _clear.m_index[2];
  3591. clearValue.uint32[3] = _clear.m_index[3];
  3592. break;
  3593. default:
  3594. bx::unpackRgba8(clearValue.float32, _clear.m_index);
  3595. break;
  3596. }
  3597. }
  3598. ++mrt;
  3599. }
  3600. }
  3601. depthAspectMask &= 0
  3602. | (_clear.m_flags & BGFX_CLEAR_DEPTH ? VK_IMAGE_ASPECT_DEPTH_BIT : 0)
  3603. | (_clear.m_flags & BGFX_CLEAR_STENCIL ? VK_IMAGE_ASPECT_STENCIL_BIT : 0)
  3604. ;
  3605. if (0 != depthAspectMask)
  3606. {
  3607. attachments[mrt].aspectMask = depthAspectMask;
  3608. attachments[mrt].clearValue.depthStencil.stencil = _clear.m_stencil;
  3609. attachments[mrt].clearValue.depthStencil.depth = _clear.m_depth;
  3610. ++mrt;
  3611. }
  3612. if (mrt > 0)
  3613. {
  3614. vkCmdClearAttachments(m_commandBuffer, mrt, attachments, BX_COUNTOF(rect), rect);
  3615. }
  3616. }
  3617. void kick(bool _finishAll = false)
  3618. {
  3619. m_cmd.kick(_finishAll);
  3620. VK_CHECK(m_cmd.alloc(&m_commandBuffer) );
  3621. m_cmd.finish(_finishAll);
  3622. }
  3623. int32_t selectMemoryType(uint32_t _memoryTypeBits, uint32_t _propertyFlags, int32_t _startIndex = 0) const
  3624. {
  3625. for (int32_t ii = _startIndex, num = m_memoryProperties.memoryTypeCount; ii < num; ++ii)
  3626. {
  3627. const VkMemoryType& memType = m_memoryProperties.memoryTypes[ii];
  3628. if ( (0 != ( (1<<ii) & _memoryTypeBits) )
  3629. && ( (memType.propertyFlags & _propertyFlags) == _propertyFlags) )
  3630. {
  3631. return ii;
  3632. }
  3633. }
  3634. BX_TRACE("Failed to find memory that supports flags 0x%08x.", _propertyFlags);
  3635. return -1;
  3636. }
  3637. VkResult allocateMemory(const VkMemoryRequirements* requirements, VkMemoryPropertyFlags propertyFlags, ::VkDeviceMemory* memory) const
  3638. {
  3639. VkMemoryAllocateInfo ma;
  3640. ma.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO;
  3641. ma.pNext = NULL;
  3642. ma.allocationSize = requirements->size;
  3643. VkResult result = VK_ERROR_UNKNOWN;
  3644. int32_t searchIndex = -1;
  3645. do
  3646. {
  3647. searchIndex++;
  3648. searchIndex = selectMemoryType(requirements->memoryTypeBits, propertyFlags, searchIndex);
  3649. if (searchIndex >= 0)
  3650. {
  3651. ma.memoryTypeIndex = searchIndex;
  3652. result = vkAllocateMemory(m_device, &ma, m_allocatorCb, memory);
  3653. }
  3654. }
  3655. while (result != VK_SUCCESS
  3656. && searchIndex >= 0);
  3657. return result;
  3658. }
  3659. VkResult createHostBuffer(uint32_t _size, VkMemoryPropertyFlags _flags, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory, const void* _data = NULL)
  3660. {
  3661. VkResult result = VK_SUCCESS;
  3662. VkBufferCreateInfo bci;
  3663. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3664. bci.pNext = NULL;
  3665. bci.flags = 0;
  3666. bci.size = _size;
  3667. bci.queueFamilyIndexCount = 0;
  3668. bci.pQueueFamilyIndices = NULL;
  3669. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3670. bci.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT | VK_BUFFER_USAGE_TRANSFER_DST_BIT;
  3671. result = vkCreateBuffer(m_device, &bci, m_allocatorCb, _buffer);
  3672. if (VK_SUCCESS != result)
  3673. {
  3674. BX_TRACE("Create host buffer error: vkCreateBuffer failed %d: %s.", result, getName(result) );
  3675. return result;
  3676. }
  3677. VkMemoryRequirements mr;
  3678. vkGetBufferMemoryRequirements(m_device, *_buffer, &mr);
  3679. result = allocateMemory(&mr, _flags, _memory);
  3680. if (VK_SUCCESS != result
  3681. && (_flags & VK_MEMORY_PROPERTY_HOST_CACHED_BIT) )
  3682. {
  3683. result = allocateMemory(&mr, _flags & ~VK_MEMORY_PROPERTY_HOST_CACHED_BIT, _memory);
  3684. }
  3685. if (VK_SUCCESS != result)
  3686. {
  3687. BX_TRACE("Create host buffer error: vkAllocateMemory failed %d: %s.", result, getName(result) );
  3688. return result;
  3689. }
  3690. result = vkBindBufferMemory(m_device, *_buffer, *_memory, 0);
  3691. if (VK_SUCCESS != result)
  3692. {
  3693. BX_TRACE("Create host buffer error: vkBindBufferMemory failed %d: %s.", result, getName(result) );
  3694. return result;
  3695. }
  3696. if (_data != NULL)
  3697. {
  3698. void* dst;
  3699. result = vkMapMemory(m_device, *_memory, 0, _size, 0, &dst);
  3700. if (VK_SUCCESS != result)
  3701. {
  3702. BX_TRACE("Create host buffer error: vkMapMemory failed %d: %s.", result, getName(result) );
  3703. return result;
  3704. }
  3705. bx::memCopy(dst, _data, _size);
  3706. vkUnmapMemory(m_device, *_memory);
  3707. }
  3708. return result;
  3709. }
  3710. VkResult createStagingBuffer(uint32_t _size, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory, const void* _data = NULL)
  3711. {
  3712. const VkMemoryPropertyFlags flags = 0
  3713. | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT
  3714. | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT
  3715. ;
  3716. return createHostBuffer(_size, flags, _buffer, _memory, _data);
  3717. }
  3718. VkResult createReadbackBuffer(uint32_t _size, ::VkBuffer* _buffer, ::VkDeviceMemory* _memory)
  3719. {
  3720. const VkMemoryPropertyFlags flags = 0
  3721. | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT
  3722. | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT
  3723. | VK_MEMORY_PROPERTY_HOST_CACHED_BIT
  3724. ;
  3725. return createHostBuffer(_size, flags, _buffer, _memory, NULL);
  3726. }
  3727. VkAllocationCallbacks* m_allocatorCb;
  3728. VkDebugReportCallbackEXT m_debugReportCallback;
  3729. VkInstance m_instance;
  3730. VkPhysicalDevice m_physicalDevice;
  3731. uint32_t m_instanceApiVersion;
  3732. VkPhysicalDeviceProperties m_deviceProperties;
  3733. VkPhysicalDeviceMemoryProperties m_memoryProperties;
  3734. VkPhysicalDeviceFeatures m_deviceFeatures;
  3735. bool m_lineAASupport;
  3736. bool m_borderColorSupport;
  3737. bool m_timerQuerySupport;
  3738. FrameBufferVK m_backBuffer;
  3739. TextureFormat::Enum m_swapchainFormats[TextureFormat::Count];
  3740. uint16_t m_numWindows;
  3741. FrameBufferHandle m_windows[BGFX_CONFIG_MAX_FRAME_BUFFERS];
  3742. int64_t m_presentElapsed;
  3743. ScratchBufferVK m_scratchBuffer[BGFX_CONFIG_MAX_FRAME_LATENCY];
  3744. uint32_t m_numFramesInFlight;
  3745. CommandQueueVK m_cmd;
  3746. VkCommandBuffer m_commandBuffer;
  3747. VkDevice m_device;
  3748. uint32_t m_globalQueueFamily;
  3749. VkQueue m_globalQueue;
  3750. VkDescriptorPool m_descriptorPool;
  3751. VkPipelineCache m_pipelineCache;
  3752. TimerQueryVK m_gpuTimer;
  3753. OcclusionQueryVK m_occlusionQuery;
  3754. void* m_renderDocDll;
  3755. void* m_vulkan1Dll;
  3756. IndexBufferVK m_indexBuffers[BGFX_CONFIG_MAX_INDEX_BUFFERS];
  3757. VertexBufferVK m_vertexBuffers[BGFX_CONFIG_MAX_VERTEX_BUFFERS];
  3758. ShaderVK m_shaders[BGFX_CONFIG_MAX_SHADERS];
  3759. ProgramVK m_program[BGFX_CONFIG_MAX_PROGRAMS];
  3760. TextureVK m_textures[BGFX_CONFIG_MAX_TEXTURES];
  3761. VertexLayout m_vertexLayouts[BGFX_CONFIG_MAX_VERTEX_LAYOUTS];
  3762. FrameBufferVK m_frameBuffers[BGFX_CONFIG_MAX_FRAME_BUFFERS];
  3763. void* m_uniforms[BGFX_CONFIG_MAX_UNIFORMS];
  3764. Matrix4 m_predefinedUniforms[PredefinedUniform::Count];
  3765. UniformRegistry m_uniformReg;
  3766. StateCacheT<VkPipeline> m_pipelineStateCache;
  3767. StateCacheT<VkDescriptorSetLayout> m_descriptorSetLayoutCache;
  3768. StateCacheT<VkRenderPass> m_renderPassCache;
  3769. StateCacheT<VkSampler> m_samplerCache;
  3770. StateCacheT<uint32_t> m_samplerBorderColorCache;
  3771. StateCacheLru<VkImageView, 1024> m_imageViewCache;
  3772. Resolution m_resolution;
  3773. float m_maxAnisotropy;
  3774. bool m_depthClamp;
  3775. bool m_wireframe;
  3776. VkBuffer m_captureBuffer;
  3777. VkDeviceMemory m_captureMemory;
  3778. uint32_t m_captureSize;
  3779. TextVideoMem m_textVideoMem;
  3780. uint8_t m_fsScratch[64<<10];
  3781. uint8_t m_vsScratch[64<<10];
  3782. FrameBufferHandle m_fbh;
  3783. };
  3784. static RendererContextVK* s_renderVK;
  3785. RendererContextI* rendererCreate(const Init& _init)
  3786. {
  3787. s_renderVK = BX_NEW(g_allocator, RendererContextVK);
  3788. if (!s_renderVK->init(_init) )
  3789. {
  3790. bx::deleteObject(g_allocator, s_renderVK);
  3791. s_renderVK = NULL;
  3792. }
  3793. return s_renderVK;
  3794. }
  3795. void rendererDestroy()
  3796. {
  3797. s_renderVK->shutdown();
  3798. bx::deleteObject(g_allocator, s_renderVK);
  3799. s_renderVK = NULL;
  3800. }
  3801. #define VK_DESTROY_FUNC(_name) \
  3802. void vkDestroy(Vk##_name& _obj) \
  3803. { \
  3804. if (VK_NULL_HANDLE != _obj) \
  3805. { \
  3806. vkDestroy##_name(s_renderVK->m_device, _obj.vk, s_renderVK->m_allocatorCb); \
  3807. _obj = VK_NULL_HANDLE; \
  3808. } \
  3809. } \
  3810. void release(Vk##_name& _obj) \
  3811. { \
  3812. s_renderVK->release(_obj); \
  3813. }
  3814. VK_DESTROY
  3815. #undef VK_DESTROY_FUNC
  3816. void vkDestroy(VkDeviceMemory& _obj)
  3817. {
  3818. if (VK_NULL_HANDLE != _obj)
  3819. {
  3820. vkFreeMemory(s_renderVK->m_device, _obj.vk, s_renderVK->m_allocatorCb);
  3821. _obj = VK_NULL_HANDLE;
  3822. }
  3823. }
  3824. void vkDestroy(VkSurfaceKHR& _obj)
  3825. {
  3826. if (VK_NULL_HANDLE != _obj)
  3827. {
  3828. vkDestroySurfaceKHR(s_renderVK->m_instance, _obj.vk, s_renderVK->m_allocatorCb);
  3829. _obj = VK_NULL_HANDLE;
  3830. }
  3831. }
  3832. void vkDestroy(VkDescriptorSet& _obj)
  3833. {
  3834. if (VK_NULL_HANDLE != _obj)
  3835. {
  3836. vkFreeDescriptorSets(s_renderVK->m_device, s_renderVK->m_descriptorPool, 1, &_obj);
  3837. _obj = VK_NULL_HANDLE;
  3838. }
  3839. }
  3840. void release(VkDeviceMemory& _obj)
  3841. {
  3842. s_renderVK->release(_obj);
  3843. }
  3844. void release(VkSurfaceKHR& _obj)
  3845. {
  3846. s_renderVK->release(_obj);
  3847. }
  3848. void release(VkDescriptorSet& _obj)
  3849. {
  3850. s_renderVK->release(_obj);
  3851. }
  3852. void ScratchBufferVK::create(uint32_t _size, uint32_t _count)
  3853. {
  3854. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  3855. const VkDevice device = s_renderVK->m_device;
  3856. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3857. const uint32_t align = uint32_t(deviceLimits.minUniformBufferOffsetAlignment);
  3858. const uint32_t entrySize = bx::strideAlign(_size, align);
  3859. const uint32_t totalSize = entrySize * _count;
  3860. VkBufferCreateInfo bci;
  3861. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3862. bci.pNext = NULL;
  3863. bci.flags = 0;
  3864. bci.size = totalSize;
  3865. bci.usage = VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT;
  3866. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3867. bci.queueFamilyIndexCount = 0;
  3868. bci.pQueueFamilyIndices = NULL;
  3869. VK_CHECK(vkCreateBuffer(
  3870. device
  3871. , &bci
  3872. , allocatorCb
  3873. , &m_buffer
  3874. ) );
  3875. VkMemoryRequirements mr;
  3876. vkGetBufferMemoryRequirements(
  3877. device
  3878. , m_buffer
  3879. , &mr
  3880. );
  3881. VkMemoryPropertyFlags flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
  3882. VkResult result = s_renderVK->allocateMemory(&mr, flags, &m_deviceMem);
  3883. if (VK_SUCCESS != result)
  3884. {
  3885. flags &= ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
  3886. VK_CHECK(s_renderVK->allocateMemory(&mr, flags, &m_deviceMem) );
  3887. }
  3888. m_size = (uint32_t)mr.size;
  3889. m_pos = 0;
  3890. VK_CHECK(vkBindBufferMemory(device, m_buffer, m_deviceMem, 0) );
  3891. VK_CHECK(vkMapMemory(device, m_deviceMem, 0, m_size, 0, (void**)&m_data) );
  3892. }
  3893. void ScratchBufferVK::destroy()
  3894. {
  3895. reset();
  3896. vkUnmapMemory(s_renderVK->m_device, m_deviceMem);
  3897. s_renderVK->release(m_buffer);
  3898. s_renderVK->release(m_deviceMem);
  3899. }
  3900. void ScratchBufferVK::reset()
  3901. {
  3902. m_pos = 0;
  3903. }
  3904. uint32_t ScratchBufferVK::write(const void* _data, uint32_t _size)
  3905. {
  3906. BX_ASSERT(m_pos < m_size, "Out of scratch buffer memory");
  3907. const uint32_t offset = m_pos;
  3908. if (_size > 0)
  3909. {
  3910. bx::memCopy(&m_data[m_pos], _data, _size);
  3911. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3912. const uint32_t align = uint32_t(deviceLimits.minUniformBufferOffsetAlignment);
  3913. const uint32_t alignedSize = bx::strideAlign(_size, align);
  3914. m_pos += alignedSize;
  3915. }
  3916. return offset;
  3917. }
  3918. void ScratchBufferVK::flush()
  3919. {
  3920. const VkPhysicalDeviceLimits& deviceLimits = s_renderVK->m_deviceProperties.limits;
  3921. VkDevice device = s_renderVK->m_device;
  3922. const uint32_t align = uint32_t(deviceLimits.nonCoherentAtomSize);
  3923. const uint32_t size = bx::min(bx::strideAlign(m_pos, align), m_size);
  3924. VkMappedMemoryRange range;
  3925. range.sType = VK_STRUCTURE_TYPE_MAPPED_MEMORY_RANGE;
  3926. range.pNext = NULL;
  3927. range.memory = m_deviceMem;
  3928. range.offset = 0;
  3929. range.size = size;
  3930. VK_CHECK(vkFlushMappedMemoryRanges(device, 1, &range) );
  3931. }
  3932. void BufferVK::create(VkCommandBuffer _commandBuffer, uint32_t _size, void* _data, uint16_t _flags, bool _vertex, uint32_t _stride)
  3933. {
  3934. BX_UNUSED(_stride);
  3935. m_size = _size;
  3936. m_flags = _flags;
  3937. m_dynamic = NULL == _data;
  3938. const bool storage = m_flags & BGFX_BUFFER_COMPUTE_READ_WRITE;
  3939. const bool indirect = m_flags & BGFX_BUFFER_DRAW_INDIRECT;
  3940. VkBufferCreateInfo bci;
  3941. bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
  3942. bci.pNext = NULL;
  3943. bci.flags = 0;
  3944. bci.size = _size;
  3945. bci.usage = 0
  3946. | (_vertex ? VK_BUFFER_USAGE_VERTEX_BUFFER_BIT : VK_BUFFER_USAGE_INDEX_BUFFER_BIT)
  3947. | (storage || indirect ? VK_BUFFER_USAGE_STORAGE_BUFFER_BIT : 0)
  3948. | (indirect ? VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT : 0)
  3949. | VK_BUFFER_USAGE_TRANSFER_DST_BIT
  3950. ;
  3951. bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  3952. bci.queueFamilyIndexCount = 0;
  3953. bci.pQueueFamilyIndices = NULL;
  3954. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  3955. const VkDevice device = s_renderVK->m_device;
  3956. VK_CHECK(vkCreateBuffer(device, &bci, allocatorCb, &m_buffer) );
  3957. VkMemoryRequirements mr;
  3958. vkGetBufferMemoryRequirements(device, m_buffer, &mr);
  3959. VK_CHECK(s_renderVK->allocateMemory(&mr, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_deviceMem) );
  3960. VK_CHECK(vkBindBufferMemory(device, m_buffer, m_deviceMem, 0) );
  3961. if (!m_dynamic)
  3962. {
  3963. update(_commandBuffer, 0, _size, _data);
  3964. }
  3965. }
  3966. void BufferVK::update(VkCommandBuffer _commandBuffer, uint32_t _offset, uint32_t _size, void* _data, bool _discard)
  3967. {
  3968. BX_UNUSED(_discard);
  3969. VkBuffer stagingBuffer;
  3970. VkDeviceMemory stagingMem;
  3971. VK_CHECK(s_renderVK->createStagingBuffer(_size, &stagingBuffer, &stagingMem, _data) );
  3972. VkBufferCopy region;
  3973. region.srcOffset = 0;
  3974. region.dstOffset = _offset;
  3975. region.size = _size;
  3976. vkCmdCopyBuffer(_commandBuffer, stagingBuffer, m_buffer, 1, &region);
  3977. setMemoryBarrier(
  3978. _commandBuffer
  3979. , VK_PIPELINE_STAGE_TRANSFER_BIT
  3980. , VK_PIPELINE_STAGE_TRANSFER_BIT
  3981. );
  3982. s_renderVK->release(stagingBuffer);
  3983. s_renderVK->release(stagingMem);
  3984. }
  3985. void BufferVK::destroy()
  3986. {
  3987. if (VK_NULL_HANDLE != m_buffer)
  3988. {
  3989. s_renderVK->release(m_buffer);
  3990. s_renderVK->release(m_deviceMem);
  3991. m_dynamic = false;
  3992. }
  3993. }
  3994. void VertexBufferVK::create(VkCommandBuffer _commandBuffer, uint32_t _size, void* _data, VertexLayoutHandle _layoutHandle, uint16_t _flags)
  3995. {
  3996. BufferVK::create(_commandBuffer, _size, _data, _flags, true);
  3997. m_layoutHandle = _layoutHandle;
  3998. }
  3999. void ShaderVK::create(const Memory* _mem)
  4000. {
  4001. bx::MemoryReader reader(_mem->data, _mem->size);
  4002. bx::ErrorAssert err;
  4003. uint32_t magic;
  4004. bx::read(&reader, magic, &err);
  4005. VkShaderStageFlagBits shaderStage = VK_SHADER_STAGE_ALL;
  4006. if (isShaderType(magic, 'C') )
  4007. {
  4008. shaderStage = VK_SHADER_STAGE_COMPUTE_BIT;
  4009. }
  4010. else if (isShaderType(magic, 'F') )
  4011. {
  4012. shaderStage = VK_SHADER_STAGE_FRAGMENT_BIT;
  4013. }
  4014. else if (isShaderType(magic, 'V') )
  4015. {
  4016. shaderStage = VK_SHADER_STAGE_VERTEX_BIT;
  4017. }
  4018. const bool fragment = isShaderType(magic, 'F');
  4019. uint32_t hashIn;
  4020. bx::read(&reader, hashIn, &err);
  4021. uint32_t hashOut;
  4022. if (isShaderVerLess(magic, 6) )
  4023. {
  4024. hashOut = hashIn;
  4025. }
  4026. else
  4027. {
  4028. bx::read(&reader, hashOut, &err);
  4029. }
  4030. uint16_t count;
  4031. bx::read(&reader, count, &err);
  4032. m_numPredefined = 0;
  4033. m_numUniforms = count;
  4034. m_numTextures = 0;
  4035. m_oldBindingModel = isShaderVerLess(magic, 11);
  4036. BX_TRACE("%s Shader consts %d"
  4037. , getShaderTypeName(magic)
  4038. , count
  4039. );
  4040. uint8_t fragmentBit = fragment ? kUniformFragmentBit : 0;
  4041. for (uint32_t ii = 0; ii < BGFX_CONFIG_MAX_TEXTURE_SAMPLERS; ++ii)
  4042. {
  4043. m_bindInfo[ii].uniformHandle = BGFX_INVALID_HANDLE;
  4044. m_bindInfo[ii].type = BindType::Count;
  4045. m_bindInfo[ii].binding = 0;
  4046. m_bindInfo[ii].samplerBinding = 0;
  4047. m_bindInfo[ii].index = UINT32_MAX;
  4048. }
  4049. if (0 < count)
  4050. {
  4051. for (uint32_t ii = 0; ii < count; ++ii)
  4052. {
  4053. uint8_t nameSize = 0;
  4054. bx::read(&reader, nameSize, &err);
  4055. char name[256];
  4056. bx::read(&reader, &name, nameSize, &err);
  4057. name[nameSize] = '\0';
  4058. uint8_t type = 0;
  4059. bx::read(&reader, type, &err);
  4060. uint8_t num;
  4061. bx::read(&reader, num, &err);
  4062. uint16_t regIndex;
  4063. bx::read(&reader, regIndex, &err);
  4064. uint16_t regCount;
  4065. bx::read(&reader, regCount, &err);
  4066. const bool hasTexData = !isShaderVerLess(magic, 8);
  4067. const bool hasTexFormat = !isShaderVerLess(magic, 10);
  4068. uint8_t texComponent = 0;
  4069. uint8_t texDimension = 0;
  4070. uint16_t texFormat = 0;
  4071. if (hasTexData)
  4072. {
  4073. bx::read(&reader, texComponent, &err);
  4074. bx::read(&reader, texDimension, &err);
  4075. }
  4076. if (hasTexFormat)
  4077. {
  4078. bx::read(&reader, texFormat, &err);
  4079. }
  4080. const char* kind = "invalid";
  4081. BX_UNUSED(num, texComponent, texFormat);
  4082. auto textureDimensionToViewType = [](TextureDimension::Enum dimension)
  4083. {
  4084. switch (dimension)
  4085. {
  4086. case TextureDimension::Dimension1D: return VK_IMAGE_VIEW_TYPE_1D;
  4087. case TextureDimension::Dimension2D: return VK_IMAGE_VIEW_TYPE_2D;
  4088. case TextureDimension::Dimension2DArray: return VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  4089. case TextureDimension::DimensionCube: return VK_IMAGE_VIEW_TYPE_CUBE;
  4090. case TextureDimension::DimensionCubeArray: return VK_IMAGE_VIEW_TYPE_CUBE_ARRAY;
  4091. case TextureDimension::Dimension3D: return VK_IMAGE_VIEW_TYPE_3D;
  4092. default: return VK_IMAGE_VIEW_TYPE_MAX_ENUM;
  4093. }
  4094. };
  4095. if (UINT16_MAX != regIndex)
  4096. {
  4097. PredefinedUniform::Enum predefined = nameToPredefinedUniformEnum(name);
  4098. if (PredefinedUniform::Count != predefined)
  4099. {
  4100. kind = "predefined";
  4101. m_predefined[m_numPredefined].m_loc = regIndex;
  4102. m_predefined[m_numPredefined].m_count = regCount;
  4103. m_predefined[m_numPredefined].m_type = uint8_t(predefined|fragmentBit);
  4104. m_numPredefined++;
  4105. }
  4106. else if (UniformType::End == (~kUniformMask & type) )
  4107. {
  4108. // regCount is used for descriptor type
  4109. const bool isBuffer = idToDescriptorType(regCount) == DescriptorType::StorageBuffer;
  4110. if (0 == regIndex)
  4111. {
  4112. continue;
  4113. }
  4114. const uint8_t reverseShift = m_oldBindingModel
  4115. ? (fragment ? kSpirvOldFragmentShift : 0) + (isBuffer ? kSpirvOldBufferShift : kSpirvOldImageShift)
  4116. : kSpirvBindShift;
  4117. const uint16_t stage = regIndex - reverseShift; // regIndex is used for buffer binding index
  4118. m_bindInfo[stage].type = isBuffer ? BindType::Buffer : BindType::Image;
  4119. m_bindInfo[stage].uniformHandle = { 0 };
  4120. m_bindInfo[stage].binding = regIndex;
  4121. if (!isBuffer)
  4122. {
  4123. const VkImageViewType viewType = hasTexData
  4124. ? textureDimensionToViewType(idToTextureDimension(texDimension) )
  4125. : VK_IMAGE_VIEW_TYPE_MAX_ENUM
  4126. ;
  4127. if (VK_IMAGE_VIEW_TYPE_MAX_ENUM != viewType)
  4128. {
  4129. m_bindInfo[stage].index = m_numTextures;
  4130. m_textures[m_numTextures].type = viewType;
  4131. m_numTextures++;
  4132. }
  4133. }
  4134. kind = "storage";
  4135. }
  4136. else if (UniformType::Sampler == (~kUniformMask & type) )
  4137. {
  4138. const uint8_t reverseShift = m_oldBindingModel
  4139. ? (fragment ? kSpirvOldFragmentShift : 0) + kSpirvOldTextureShift
  4140. : kSpirvBindShift;
  4141. const uint16_t stage = regIndex - reverseShift; // regIndex is used for image/sampler binding index
  4142. const UniformRegInfo* info = s_renderVK->m_uniformReg.find(name);
  4143. BX_ASSERT(NULL != info, "User defined uniform '%s' is not found, it won't be set.", name);
  4144. m_bindInfo[stage].uniformHandle = info->m_handle;
  4145. m_bindInfo[stage].type = BindType::Sampler;
  4146. m_bindInfo[stage].binding = regIndex;
  4147. m_bindInfo[stage].samplerBinding = regIndex + kSpirvSamplerShift;
  4148. const VkImageViewType viewType = hasTexData
  4149. ? textureDimensionToViewType(idToTextureDimension(texDimension) )
  4150. : VK_IMAGE_VIEW_TYPE_MAX_ENUM
  4151. ;
  4152. if (VK_IMAGE_VIEW_TYPE_MAX_ENUM != viewType)
  4153. {
  4154. m_bindInfo[stage].index = m_numTextures;
  4155. m_textures[m_numTextures].type = viewType;
  4156. m_numTextures++;
  4157. }
  4158. kind = "sampler";
  4159. }
  4160. else
  4161. {
  4162. const UniformRegInfo* info = s_renderVK->m_uniformReg.find(name);
  4163. BX_ASSERT(NULL != info, "User defined uniform '%s' is not found, it won't be set.", name);
  4164. if (NULL != info)
  4165. {
  4166. if (NULL == m_constantBuffer)
  4167. {
  4168. m_constantBuffer = UniformBuffer::create(1024);
  4169. }
  4170. kind = "user";
  4171. m_constantBuffer->writeUniformHandle( (UniformType::Enum)(type|fragmentBit), regIndex, info->m_handle, regCount);
  4172. }
  4173. }
  4174. }
  4175. BX_TRACE("\t%s: %s (%s), r.index %3d, r.count %2d, r.texComponent %1d, r.texDimension %1d"
  4176. , kind
  4177. , name
  4178. , getUniformTypeName(UniformType::Enum(type&~kUniformMask) )
  4179. , regIndex
  4180. , regCount
  4181. , texComponent
  4182. , texDimension
  4183. );
  4184. BX_UNUSED(kind);
  4185. }
  4186. if (NULL != m_constantBuffer)
  4187. {
  4188. m_constantBuffer->finish();
  4189. }
  4190. }
  4191. uint32_t shaderSize;
  4192. bx::read(&reader, shaderSize, &err);
  4193. const void* code = reader.getDataPtr();
  4194. bx::skip(&reader, shaderSize+1);
  4195. m_code = alloc(shaderSize);
  4196. bx::memCopy(m_code->data, code, shaderSize);
  4197. VkShaderModuleCreateInfo smci;
  4198. smci.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO;
  4199. smci.pNext = NULL;
  4200. smci.flags = 0;
  4201. smci.codeSize = m_code->size;
  4202. smci.pCode = (const uint32_t*)m_code->data;
  4203. // disassemble(bx::getDebugOut(), m_code->data, m_code->size);
  4204. VK_CHECK(vkCreateShaderModule(
  4205. s_renderVK->m_device
  4206. , &smci
  4207. , s_renderVK->m_allocatorCb
  4208. , &m_module
  4209. ) );
  4210. bx::memSet(m_attrMask, 0, sizeof(m_attrMask) );
  4211. bx::memSet(m_attrRemap, 0, sizeof(m_attrRemap) );
  4212. bx::read(&reader, m_numAttrs, &err);
  4213. for (uint8_t ii = 0; ii < m_numAttrs; ++ii)
  4214. {
  4215. uint16_t id;
  4216. bx::read(&reader, id, &err);
  4217. Attrib::Enum attr = idToAttrib(id);
  4218. if (Attrib::Count != attr)
  4219. {
  4220. m_attrMask[attr] = UINT16_MAX;
  4221. m_attrRemap[attr] = ii;
  4222. }
  4223. }
  4224. bx::HashMurmur2A murmur;
  4225. murmur.begin();
  4226. murmur.add(hashIn);
  4227. murmur.add(hashOut);
  4228. murmur.add(m_code->data, m_code->size);
  4229. murmur.add(m_numAttrs);
  4230. murmur.add(m_attrMask, m_numAttrs);
  4231. murmur.add(m_attrRemap, m_numAttrs);
  4232. m_hash = murmur.end();
  4233. bx::read(&reader, m_size, &err);
  4234. // fill binding description with uniform information
  4235. uint16_t bidx = 0;
  4236. if (m_size > 0)
  4237. {
  4238. m_uniformBinding = fragment ? (m_oldBindingModel ? kSpirvOldFragmentBinding : kSpirvFragmentBinding) : 0;
  4239. VkDescriptorSetLayoutBinding& binding = m_bindings[bidx];
  4240. binding.stageFlags = shaderStage;
  4241. binding.descriptorType = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
  4242. binding.binding = m_uniformBinding;
  4243. binding.pImmutableSamplers = NULL;
  4244. binding.descriptorCount = 1;
  4245. bidx++;
  4246. }
  4247. for (uint32_t ii = 0; ii < BX_COUNTOF(m_bindInfo); ++ii)
  4248. {
  4249. switch (m_bindInfo[ii].type)
  4250. {
  4251. case BindType::Buffer:
  4252. case BindType::Image:
  4253. {
  4254. VkDescriptorSetLayoutBinding& binding = m_bindings[bidx];
  4255. binding.stageFlags = shaderStage;
  4256. binding.descriptorType = BindType::Buffer == m_bindInfo[ii].type
  4257. ? VK_DESCRIPTOR_TYPE_STORAGE_BUFFER
  4258. : VK_DESCRIPTOR_TYPE_STORAGE_IMAGE
  4259. ;
  4260. binding.binding = m_bindInfo[ii].binding;
  4261. binding.pImmutableSamplers = NULL;
  4262. binding.descriptorCount = 1;
  4263. bidx++;
  4264. }
  4265. break;
  4266. case BindType::Sampler:
  4267. {
  4268. VkDescriptorSetLayoutBinding& textureBinding = m_bindings[bidx];
  4269. textureBinding.stageFlags = shaderStage;
  4270. textureBinding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
  4271. textureBinding.binding = m_bindInfo[ii].binding;
  4272. textureBinding.pImmutableSamplers = NULL;
  4273. textureBinding.descriptorCount = 1;
  4274. bidx++;
  4275. VkDescriptorSetLayoutBinding& samplerBinding = m_bindings[bidx];
  4276. samplerBinding.stageFlags = shaderStage;
  4277. samplerBinding.descriptorType = VK_DESCRIPTOR_TYPE_SAMPLER;
  4278. samplerBinding.binding = m_bindInfo[ii].samplerBinding;
  4279. samplerBinding.pImmutableSamplers = NULL;
  4280. samplerBinding.descriptorCount = 1;
  4281. bidx++;
  4282. }
  4283. break;
  4284. default:
  4285. break;
  4286. }
  4287. }
  4288. m_numBindings = bidx;
  4289. }
  4290. void ShaderVK::destroy()
  4291. {
  4292. if (NULL != m_constantBuffer)
  4293. {
  4294. UniformBuffer::destroy(m_constantBuffer);
  4295. m_constantBuffer = NULL;
  4296. }
  4297. m_numPredefined = 0;
  4298. if (NULL != m_code)
  4299. {
  4300. release(m_code);
  4301. m_code = NULL;
  4302. m_hash = 0;
  4303. }
  4304. if (VK_NULL_HANDLE != m_module)
  4305. {
  4306. vkDestroy(m_module);
  4307. }
  4308. }
  4309. void ProgramVK::create(const ShaderVK* _vsh, const ShaderVK* _fsh)
  4310. {
  4311. BX_ASSERT(NULL != _vsh->m_code, "Vertex shader doesn't exist.");
  4312. m_vsh = _vsh;
  4313. bx::memCopy(
  4314. &m_predefined[0]
  4315. , _vsh->m_predefined
  4316. , _vsh->m_numPredefined * sizeof(PredefinedUniform)
  4317. );
  4318. m_numPredefined = _vsh->m_numPredefined;
  4319. if (NULL != _fsh)
  4320. {
  4321. BX_ASSERT(NULL != _fsh->m_code, "Fragment shader doesn't exist.");
  4322. m_fsh = _fsh;
  4323. bx::memCopy(
  4324. &m_predefined[m_numPredefined]
  4325. , _fsh->m_predefined
  4326. , _fsh->m_numPredefined * sizeof(PredefinedUniform)
  4327. );
  4328. m_numPredefined += _fsh->m_numPredefined;
  4329. }
  4330. m_numTextures = 0;
  4331. for (uint8_t stage = 0; stage < BX_COUNTOF(m_bindInfo); ++stage)
  4332. {
  4333. const ShaderVK* shader = NULL;
  4334. if (isValid(m_vsh->m_bindInfo[stage].uniformHandle) )
  4335. {
  4336. shader = _vsh;
  4337. BX_ASSERT(false
  4338. || NULL == m_fsh
  4339. || !isValid(m_fsh->m_bindInfo[stage].uniformHandle)
  4340. || !(m_vsh->m_oldBindingModel || m_fsh->m_oldBindingModel)
  4341. , "Shared vertex/fragment bindings require shader binary version >= 11."
  4342. );
  4343. }
  4344. else if (NULL != m_fsh
  4345. && isValid(m_fsh->m_bindInfo[stage].uniformHandle) )
  4346. {
  4347. shader = _fsh;
  4348. }
  4349. if (NULL != shader)
  4350. {
  4351. m_bindInfo[stage] = shader->m_bindInfo[stage];
  4352. uint32_t& index = m_bindInfo[stage].index;
  4353. if (UINT32_MAX != index)
  4354. {
  4355. m_textures[m_numTextures] = shader->m_textures[index];
  4356. index = m_numTextures;
  4357. m_numTextures++;
  4358. }
  4359. }
  4360. }
  4361. // create exact pipeline layout
  4362. m_descriptorSetLayout = VK_NULL_HANDLE;
  4363. uint32_t numBindings = m_vsh->m_numBindings + (m_fsh ? m_fsh->m_numBindings : 0);
  4364. if (0 < numBindings)
  4365. {
  4366. // generate descriptor set layout hash
  4367. bx::HashMurmur2A murmur;
  4368. murmur.begin();
  4369. murmur.add(m_vsh->m_bindings, sizeof(VkDescriptorSetLayoutBinding) * m_vsh->m_numBindings);
  4370. if (NULL != m_fsh)
  4371. {
  4372. murmur.add(m_fsh->m_bindings, sizeof(VkDescriptorSetLayoutBinding) * m_fsh->m_numBindings);
  4373. }
  4374. uint32_t descriptorSetLayoutHash = murmur.end();
  4375. m_descriptorSetLayout = s_renderVK->m_descriptorSetLayoutCache.find(descriptorSetLayoutHash);
  4376. if (VK_NULL_HANDLE == m_descriptorSetLayout)
  4377. {
  4378. VkDescriptorSetLayoutBinding bindings[2 * BX_COUNTOF(ShaderVK::m_bindings)];
  4379. bx::memCopy(
  4380. bindings
  4381. , m_vsh->m_bindings
  4382. , sizeof(VkDescriptorSetLayoutBinding) * m_vsh->m_numBindings
  4383. );
  4384. numBindings = m_vsh->m_numBindings;
  4385. if (NULL != m_fsh)
  4386. {
  4387. for (uint16_t ii = 0; ii < m_fsh->m_numBindings; ii++)
  4388. {
  4389. const VkDescriptorSetLayoutBinding& fsBinding = m_fsh->m_bindings[ii];
  4390. uint16_t vsBindingIdx = UINT16_MAX;
  4391. for (uint16_t jj = 0; jj < m_vsh->m_numBindings; jj++)
  4392. {
  4393. if (fsBinding.binding == bindings[jj].binding)
  4394. {
  4395. vsBindingIdx = jj;
  4396. break;
  4397. }
  4398. }
  4399. if (UINT16_MAX != vsBindingIdx)
  4400. {
  4401. BX_ASSERT(
  4402. bindings[vsBindingIdx].descriptorType == fsBinding.descriptorType
  4403. , "Mismatching descriptor types. Shaders compiled with different versions of shaderc?"
  4404. );
  4405. bindings[vsBindingIdx].stageFlags |= fsBinding.stageFlags;
  4406. }
  4407. else
  4408. {
  4409. bindings[numBindings] = fsBinding;
  4410. numBindings++;
  4411. }
  4412. }
  4413. }
  4414. VkDescriptorSetLayoutCreateInfo dslci;
  4415. dslci.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO;
  4416. dslci.pNext = NULL;
  4417. dslci.flags = 0;
  4418. dslci.bindingCount = numBindings;
  4419. dslci.pBindings = bindings;
  4420. VK_CHECK(vkCreateDescriptorSetLayout(
  4421. s_renderVK->m_device
  4422. , &dslci
  4423. , s_renderVK->m_allocatorCb
  4424. , &m_descriptorSetLayout
  4425. ) );
  4426. s_renderVK->m_descriptorSetLayoutCache.add(descriptorSetLayoutHash, m_descriptorSetLayout);
  4427. }
  4428. }
  4429. VkPipelineLayoutCreateInfo plci;
  4430. plci.sType = VK_STRUCTURE_TYPE_PIPELINE_LAYOUT_CREATE_INFO;
  4431. plci.pNext = NULL;
  4432. plci.flags = 0;
  4433. plci.pushConstantRangeCount = 0;
  4434. plci.pPushConstantRanges = NULL;
  4435. plci.setLayoutCount = (m_descriptorSetLayout == VK_NULL_HANDLE ? 0 : 1);
  4436. plci.pSetLayouts = &m_descriptorSetLayout;
  4437. VK_CHECK(vkCreatePipelineLayout(
  4438. s_renderVK->m_device
  4439. , &plci
  4440. , s_renderVK->m_allocatorCb
  4441. , &m_pipelineLayout
  4442. ) );
  4443. }
  4444. void ProgramVK::destroy()
  4445. {
  4446. s_renderVK->release(m_pipelineLayout);
  4447. m_numPredefined = 0;
  4448. m_vsh = NULL;
  4449. m_fsh = NULL;
  4450. }
  4451. VkResult TimerQueryVK::init()
  4452. {
  4453. VkResult result = VK_SUCCESS;
  4454. const VkDevice device = s_renderVK->m_device;
  4455. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4456. const uint32_t count = m_control.m_size * 2;
  4457. VkQueryPoolCreateInfo qpci;
  4458. qpci.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
  4459. qpci.pNext = NULL;
  4460. qpci.flags = 0;
  4461. qpci.queryType = VK_QUERY_TYPE_TIMESTAMP;
  4462. qpci.queryCount = count;
  4463. qpci.pipelineStatistics = 0;
  4464. result = vkCreateQueryPool(device, &qpci, s_renderVK->m_allocatorCb, &m_queryPool);
  4465. if (VK_SUCCESS != result)
  4466. {
  4467. BX_TRACE("Create timer query error: vkCreateQueryPool failed %d: %s.", result, getName(result) );
  4468. return result;
  4469. }
  4470. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4471. const uint32_t size = count * sizeof(uint64_t);
  4472. result = s_renderVK->createReadbackBuffer(size, &m_readback, &m_readbackMemory);
  4473. if (VK_SUCCESS != result)
  4474. {
  4475. return result;
  4476. }
  4477. result = vkMapMemory(device, m_readbackMemory, 0, VK_WHOLE_SIZE, 0, (void**)&m_queryResult);
  4478. if (VK_SUCCESS != result)
  4479. {
  4480. BX_TRACE("Create timer query error: vkMapMemory failed %d: %s.", result, getName(result) );
  4481. return result;
  4482. }
  4483. m_frequency = uint64_t(1000000000.0 / double(s_renderVK->m_deviceProperties.limits.timestampPeriod) );
  4484. for (uint32_t ii = 0; ii < BX_COUNTOF(m_result); ++ii)
  4485. {
  4486. m_result[ii].reset();
  4487. }
  4488. m_control.reset();
  4489. return result;
  4490. }
  4491. void TimerQueryVK::shutdown()
  4492. {
  4493. vkDestroy(m_queryPool);
  4494. vkDestroy(m_readback);
  4495. vkUnmapMemory(s_renderVK->m_device, m_readbackMemory);
  4496. vkDestroy(m_readbackMemory);
  4497. }
  4498. uint32_t TimerQueryVK::begin(uint32_t _resultIdx, uint32_t _frameNum)
  4499. {
  4500. while (0 == m_control.reserve(1) )
  4501. {
  4502. m_control.consume(1);
  4503. }
  4504. Result& result = m_result[_resultIdx];
  4505. ++result.m_pending;
  4506. const uint32_t idx = m_control.m_current;
  4507. Query& query = m_query[idx];
  4508. query.m_resultIdx = _resultIdx;
  4509. query.m_ready = false;
  4510. query.m_frameNum = _frameNum;
  4511. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4512. const uint32_t offset = idx * 2 + 0;
  4513. vkCmdResetQueryPool(commandBuffer, m_queryPool, offset, 2);
  4514. vkCmdWriteTimestamp(commandBuffer, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT, m_queryPool, offset + 0);
  4515. m_control.commit(1);
  4516. return idx;
  4517. }
  4518. void TimerQueryVK::end(uint32_t _idx)
  4519. {
  4520. Query& query = m_query[_idx];
  4521. query.m_ready = true;
  4522. query.m_completed = s_renderVK->m_cmd.m_submitted + s_renderVK->m_cmd.m_numFramesInFlight;
  4523. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4524. const uint32_t offset = _idx * 2 + 0;
  4525. vkCmdWriteTimestamp(commandBuffer, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT, m_queryPool, offset + 1);
  4526. vkCmdCopyQueryPoolResults(
  4527. commandBuffer
  4528. , m_queryPool
  4529. , offset
  4530. , 2
  4531. , m_readback
  4532. , offset * sizeof(uint64_t)
  4533. , sizeof(uint64_t)
  4534. , VK_QUERY_RESULT_WAIT_BIT | VK_QUERY_RESULT_64_BIT
  4535. );
  4536. setMemoryBarrier(commandBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT);
  4537. while (update() )
  4538. {
  4539. }
  4540. }
  4541. bool TimerQueryVK::update()
  4542. {
  4543. if (0 != m_control.available() )
  4544. {
  4545. uint32_t idx = m_control.m_read;
  4546. Query& query = m_query[idx];
  4547. if (!query.m_ready)
  4548. {
  4549. return false;
  4550. }
  4551. if (query.m_completed > s_renderVK->m_cmd.m_submitted)
  4552. {
  4553. return false;
  4554. }
  4555. m_control.consume(1);
  4556. Result& result = m_result[query.m_resultIdx];
  4557. --result.m_pending;
  4558. result.m_frameNum = query.m_frameNum;
  4559. uint32_t offset = idx * 2;
  4560. result.m_begin = m_queryResult[offset+0];
  4561. result.m_end = m_queryResult[offset+1];
  4562. return true;
  4563. }
  4564. return false;
  4565. }
  4566. VkResult OcclusionQueryVK::init()
  4567. {
  4568. VkResult result = VK_SUCCESS;
  4569. const VkDevice device = s_renderVK->m_device;
  4570. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4571. const uint32_t count = BX_COUNTOF(m_handle);
  4572. VkQueryPoolCreateInfo qpci;
  4573. qpci.sType = VK_STRUCTURE_TYPE_QUERY_POOL_CREATE_INFO;
  4574. qpci.pNext = NULL;
  4575. qpci.flags = 0;
  4576. qpci.queryType = VK_QUERY_TYPE_OCCLUSION;
  4577. qpci.queryCount = count;
  4578. qpci.pipelineStatistics = 0;
  4579. result = vkCreateQueryPool(device, &qpci, s_renderVK->m_allocatorCb, &m_queryPool);
  4580. if (VK_SUCCESS != result)
  4581. {
  4582. BX_TRACE("Create occlusion query error: vkCreateQueryPool failed %d: %s.", result, getName(result) );
  4583. return result;
  4584. }
  4585. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4586. const uint32_t size = count * sizeof(uint32_t);
  4587. result = s_renderVK->createReadbackBuffer(size, &m_readback, &m_readbackMemory);
  4588. if (VK_SUCCESS != result)
  4589. {
  4590. return result;
  4591. }
  4592. result = vkMapMemory(device, m_readbackMemory, 0, VK_WHOLE_SIZE, 0, (void**)&m_queryResult);
  4593. if (VK_SUCCESS != result)
  4594. {
  4595. BX_TRACE("Create occlusion query error: vkMapMemory failed %d: %s.", result, getName(result) );
  4596. return result;
  4597. }
  4598. m_control.reset();
  4599. return result;
  4600. }
  4601. void OcclusionQueryVK::shutdown()
  4602. {
  4603. vkDestroy(m_queryPool);
  4604. vkDestroy(m_readback);
  4605. vkUnmapMemory(s_renderVK->m_device, m_readbackMemory);
  4606. vkDestroy(m_readbackMemory);
  4607. }
  4608. void OcclusionQueryVK::begin(OcclusionQueryHandle _handle)
  4609. {
  4610. m_control.reserve(1);
  4611. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4612. m_handle[m_control.m_current] = _handle;
  4613. vkCmdBeginQuery(commandBuffer, m_queryPool, _handle.idx, 0);
  4614. }
  4615. void OcclusionQueryVK::end()
  4616. {
  4617. const VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4618. const OcclusionQueryHandle handle = m_handle[m_control.m_current];
  4619. vkCmdEndQuery(commandBuffer, m_queryPool, handle.idx);
  4620. m_control.commit(1);
  4621. }
  4622. void OcclusionQueryVK::flush(Frame* _render)
  4623. {
  4624. if (0 < m_control.available() )
  4625. {
  4626. VkCommandBuffer commandBuffer = s_renderVK->m_commandBuffer;
  4627. const uint32_t size = m_control.m_size;
  4628. // need to copy each result individually because VK_QUERY_RESULT_WAIT_BIT causes
  4629. // vkWaitForFences to hang indefinitely if we copy all results (including unavailable ones)
  4630. for (uint32_t ii = 0, num = m_control.available(); ii < num; ++ii)
  4631. {
  4632. const OcclusionQueryHandle& handle = m_handle[(m_control.m_read + ii) % size];
  4633. if (isValid(handle) )
  4634. {
  4635. vkCmdCopyQueryPoolResults(
  4636. commandBuffer
  4637. , m_queryPool
  4638. , handle.idx
  4639. , 1
  4640. , m_readback
  4641. , handle.idx * sizeof(uint32_t)
  4642. , sizeof(uint32_t)
  4643. , VK_QUERY_RESULT_WAIT_BIT
  4644. );
  4645. }
  4646. }
  4647. setMemoryBarrier(commandBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT);
  4648. s_renderVK->kick(true);
  4649. commandBuffer = s_renderVK->m_commandBuffer;
  4650. // resetting in the new command buffer prevents a false positive validation layer error
  4651. const uint32_t count = BX_COUNTOF(m_handle);
  4652. vkCmdResetQueryPool(commandBuffer, m_queryPool, 0, count);
  4653. resolve(_render);
  4654. }
  4655. }
  4656. void OcclusionQueryVK::resolve(Frame* _render)
  4657. {
  4658. while (0 != m_control.available() )
  4659. {
  4660. OcclusionQueryHandle handle = m_handle[m_control.m_read];
  4661. if (isValid(handle) )
  4662. {
  4663. _render->m_occlusion[handle.idx] = m_queryResult[handle.idx];
  4664. }
  4665. m_control.consume(1);
  4666. }
  4667. }
  4668. void OcclusionQueryVK::invalidate(OcclusionQueryHandle _handle)
  4669. {
  4670. const uint32_t size = m_control.m_size;
  4671. for (uint32_t ii = 0, num = m_control.available(); ii < num; ++ii)
  4672. {
  4673. OcclusionQueryHandle& handle = m_handle[(m_control.m_read + ii) % size];
  4674. if (handle.idx == _handle.idx)
  4675. {
  4676. handle.idx = bgfx::kInvalidHandle;
  4677. }
  4678. }
  4679. }
  4680. void ReadbackVK::create(VkImage _image, uint32_t _width, uint32_t _height, TextureFormat::Enum _format)
  4681. {
  4682. m_image = _image;
  4683. m_width = _width;
  4684. m_height = _height;
  4685. m_format = _format;
  4686. }
  4687. void ReadbackVK::destroy()
  4688. {
  4689. m_image = VK_NULL_HANDLE;
  4690. }
  4691. uint32_t ReadbackVK::pitch(uint8_t _mip) const
  4692. {
  4693. uint32_t mipWidth = bx::uint32_max(1, m_width >> _mip);
  4694. uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_format) );
  4695. return mipWidth * bpp / 8;
  4696. }
  4697. void ReadbackVK::copyImageToBuffer(VkCommandBuffer _commandBuffer, VkBuffer _buffer, VkImageLayout _layout, VkImageAspectFlags _aspect, uint8_t _mip) const
  4698. {
  4699. uint32_t mipWidth = bx::uint32_max(1, m_width >> _mip);
  4700. uint32_t mipHeight = bx::uint32_max(1, m_height >> _mip);
  4701. setImageMemoryBarrier(
  4702. _commandBuffer
  4703. , m_image
  4704. , _aspect
  4705. , _layout
  4706. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4707. , _mip
  4708. , 1
  4709. , 0
  4710. , 1
  4711. );
  4712. VkBufferImageCopy bic;
  4713. bic.bufferOffset = 0;
  4714. bic.bufferRowLength = mipWidth;
  4715. bic.bufferImageHeight = mipHeight;
  4716. bic.imageSubresource.aspectMask = _aspect;
  4717. bic.imageSubresource.mipLevel = _mip;
  4718. bic.imageSubresource.baseArrayLayer = 0;
  4719. bic.imageSubresource.layerCount = 1;
  4720. bic.imageOffset = { 0, 0, 0 };
  4721. bic.imageExtent = { mipWidth, mipHeight, 1 };
  4722. vkCmdCopyImageToBuffer(
  4723. _commandBuffer
  4724. , m_image
  4725. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4726. , _buffer
  4727. , 1
  4728. , &bic
  4729. );
  4730. // Make changes to the buffer visible to the host
  4731. setMemoryBarrier(
  4732. _commandBuffer
  4733. , VK_PIPELINE_STAGE_TRANSFER_BIT
  4734. , VK_PIPELINE_STAGE_HOST_BIT
  4735. );
  4736. setImageMemoryBarrier(
  4737. _commandBuffer
  4738. , m_image
  4739. , _aspect
  4740. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  4741. , _layout
  4742. , _mip
  4743. , 1
  4744. , 0
  4745. , 1
  4746. );
  4747. }
  4748. void ReadbackVK::readback(VkDeviceMemory _memory, VkDeviceSize _offset, void* _data, uint8_t _mip) const
  4749. {
  4750. if (m_image == VK_NULL_HANDLE)
  4751. {
  4752. return;
  4753. }
  4754. uint32_t mipHeight = bx::uint32_max(1, m_height >> _mip);
  4755. uint32_t rowPitch = pitch(_mip);
  4756. uint8_t* src;
  4757. VK_CHECK(vkMapMemory(s_renderVK->m_device, _memory, 0, VK_WHOLE_SIZE, 0, (void**)&src) );
  4758. src += _offset;
  4759. uint8_t* dst = (uint8_t*)_data;
  4760. for (uint32_t yy = 0; yy < mipHeight; ++yy)
  4761. {
  4762. bx::memCopy(dst, src, rowPitch);
  4763. src += rowPitch;
  4764. dst += rowPitch;
  4765. }
  4766. vkUnmapMemory(s_renderVK->m_device, _memory);
  4767. }
  4768. VkResult TextureVK::create(VkCommandBuffer _commandBuffer, uint32_t _width, uint32_t _height, uint64_t _flags, VkFormat _format)
  4769. {
  4770. BX_ASSERT(0 != (_flags & BGFX_TEXTURE_RT_MASK), "");
  4771. _flags |= BGFX_TEXTURE_RT_WRITE_ONLY;
  4772. m_flags = _flags;
  4773. m_width = _width;
  4774. m_height = _height;
  4775. m_depth = 1;
  4776. m_numLayers = 1;
  4777. m_requestedFormat = uint8_t(bimg::TextureFormat::Count);
  4778. m_textureFormat = uint8_t(bimg::TextureFormat::Count);
  4779. m_format = _format;
  4780. m_components = { VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY, VK_COMPONENT_SWIZZLE_IDENTITY };
  4781. m_aspectMask = getAspectMask(m_format);
  4782. m_sampler = s_msaa[bx::uint32_satsub( (m_flags & BGFX_TEXTURE_RT_MSAA_MASK) >> BGFX_TEXTURE_RT_MSAA_SHIFT, 1)];
  4783. m_type = VK_IMAGE_VIEW_TYPE_2D;
  4784. m_numMips = 1;
  4785. m_numSides = 1;
  4786. VkResult result = createImages(_commandBuffer);
  4787. if (VK_SUCCESS == result)
  4788. {
  4789. const VkImageLayout layout = 0 != (m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  4790. ? VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL
  4791. : VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  4792. ;
  4793. setImageMemoryBarrier(_commandBuffer, layout);
  4794. }
  4795. return result;
  4796. }
  4797. VkResult TextureVK::createImages(VkCommandBuffer _commandBuffer)
  4798. {
  4799. VkResult result = VK_SUCCESS;
  4800. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  4801. const VkDevice device = s_renderVK->m_device;
  4802. if (m_sampler.Count > 1)
  4803. {
  4804. BX_ASSERT(VK_IMAGE_VIEW_TYPE_3D != m_type, "Can't create multisample 3D image.");
  4805. BX_ASSERT(m_numMips <= 1, "Can't create multisample image with mip chain.");
  4806. }
  4807. // create texture and allocate its device memory
  4808. VkImageCreateInfo ici;
  4809. ici.sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
  4810. ici.pNext = NULL;
  4811. ici.flags = 0
  4812. | (VK_IMAGE_VIEW_TYPE_CUBE == m_type
  4813. ? VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT
  4814. : 0
  4815. )
  4816. | (VK_IMAGE_VIEW_TYPE_3D == m_type
  4817. ? VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT_KHR
  4818. : 0
  4819. )
  4820. ;
  4821. ici.pQueueFamilyIndices = NULL;
  4822. ici.queueFamilyIndexCount = 0;
  4823. ici.initialLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  4824. ici.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
  4825. ici.usage = 0
  4826. | VK_IMAGE_USAGE_TRANSFER_SRC_BIT
  4827. | VK_IMAGE_USAGE_TRANSFER_DST_BIT
  4828. | VK_IMAGE_USAGE_SAMPLED_BIT
  4829. | (m_flags & BGFX_TEXTURE_RT_MASK
  4830. ? (m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT)
  4831. ? VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT
  4832. : VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT)
  4833. : 0
  4834. )
  4835. | (m_flags & BGFX_TEXTURE_COMPUTE_WRITE ? VK_IMAGE_USAGE_STORAGE_BIT : 0)
  4836. ;
  4837. ici.format = m_format;
  4838. ici.samples = m_sampler.Sample;
  4839. ici.mipLevels = m_numMips;
  4840. ici.arrayLayers = m_numSides;
  4841. ici.extent.width = m_width;
  4842. ici.extent.height = m_height;
  4843. ici.extent.depth = m_depth;
  4844. ici.imageType = VK_IMAGE_VIEW_TYPE_3D == m_type
  4845. ? VK_IMAGE_TYPE_3D
  4846. : VK_IMAGE_TYPE_2D
  4847. ;
  4848. ici.tiling = VK_IMAGE_TILING_OPTIMAL;
  4849. result = vkCreateImage(device, &ici, allocatorCb, &m_textureImage);
  4850. if (VK_SUCCESS != result)
  4851. {
  4852. BX_TRACE("Create texture image error: vkCreateImage failed %d: %s.", result, getName(result) );
  4853. return result;
  4854. }
  4855. VkMemoryRequirements imageMemReq;
  4856. vkGetImageMemoryRequirements(device, m_textureImage, &imageMemReq);
  4857. result = s_renderVK->allocateMemory(&imageMemReq, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_textureDeviceMem);
  4858. if (VK_SUCCESS != result)
  4859. {
  4860. BX_TRACE("Create texture image error: allocateMemory failed %d: %s.", result, getName(result) );
  4861. return result;
  4862. }
  4863. result = vkBindImageMemory(device, m_textureImage, m_textureDeviceMem, 0);
  4864. if (VK_SUCCESS != result)
  4865. {
  4866. BX_TRACE("Create texture image error: vkBindImageMemory failed %d: %s.", result, getName(result) );
  4867. return result;
  4868. }
  4869. m_sampledLayout = m_flags & BGFX_TEXTURE_COMPUTE_WRITE
  4870. ? VK_IMAGE_LAYOUT_GENERAL
  4871. : VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL
  4872. ;
  4873. const bool needResolve = true
  4874. && 1 < m_sampler.Count
  4875. && 0 != (ici.usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT)
  4876. && 0 == (m_flags & BGFX_TEXTURE_MSAA_SAMPLE)
  4877. && 0 == (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY)
  4878. ;
  4879. if (needResolve)
  4880. {
  4881. VkImageCreateInfo ici_resolve = ici;
  4882. ici_resolve.samples = s_msaa[0].Sample;
  4883. ici_resolve.usage &= ~VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
  4884. ici_resolve.flags &= ~VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
  4885. result = vkCreateImage(device, &ici_resolve, allocatorCb, &m_singleMsaaImage);
  4886. if (VK_SUCCESS != result)
  4887. {
  4888. BX_TRACE("Create texture image error: vkCreateImage failed %d: %s.", result, getName(result) );
  4889. return result;
  4890. }
  4891. VkMemoryRequirements imageMemReq_resolve;
  4892. vkGetImageMemoryRequirements(device, m_singleMsaaImage, &imageMemReq_resolve);
  4893. result = s_renderVK->allocateMemory(&imageMemReq_resolve, VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, &m_singleMsaaDeviceMem);
  4894. if (VK_SUCCESS != result)
  4895. {
  4896. BX_TRACE("Create texture image error: allocateMemory failed %d: %s.", result, getName(result) );
  4897. return result;
  4898. }
  4899. result = vkBindImageMemory(device, m_singleMsaaImage, m_singleMsaaDeviceMem, 0);
  4900. if (VK_SUCCESS != result)
  4901. {
  4902. BX_TRACE("Create texture image error: vkBindImageMemory failed %d: %s.", result, getName(result) );
  4903. return result;
  4904. }
  4905. setImageMemoryBarrier(_commandBuffer, m_sampledLayout, true);
  4906. }
  4907. return result;
  4908. }
  4909. void* TextureVK::create(VkCommandBuffer _commandBuffer, const Memory* _mem, uint64_t _flags, uint8_t _skip)
  4910. {
  4911. bimg::ImageContainer imageContainer;
  4912. if (bimg::imageParse(imageContainer, _mem->data, _mem->size) )
  4913. {
  4914. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(imageContainer.m_format);
  4915. const uint8_t startLod = bx::min<uint8_t>(_skip, imageContainer.m_numMips - 1);
  4916. bimg::TextureInfo ti;
  4917. bimg::imageGetSize(
  4918. &ti
  4919. , uint16_t(imageContainer.m_width >> startLod)
  4920. , uint16_t(imageContainer.m_height >> startLod)
  4921. , uint16_t(imageContainer.m_depth >> startLod)
  4922. , imageContainer.m_cubeMap
  4923. , 1 < imageContainer.m_numMips
  4924. , imageContainer.m_numLayers
  4925. , imageContainer.m_format
  4926. );
  4927. ti.numMips = bx::min<uint8_t>(imageContainer.m_numMips - startLod, ti.numMips);
  4928. m_flags = _flags;
  4929. m_width = ti.width;
  4930. m_height = ti.height;
  4931. m_depth = ti.depth;
  4932. m_numLayers = ti.numLayers;
  4933. m_requestedFormat = uint8_t(imageContainer.m_format);
  4934. m_textureFormat = uint8_t(getViableTextureFormat(imageContainer) );
  4935. m_format = bimg::isDepth(bimg::TextureFormat::Enum(m_textureFormat) )
  4936. ? s_textureFormat[m_textureFormat].m_fmtDsv
  4937. : (m_flags & BGFX_TEXTURE_SRGB) ? s_textureFormat[m_textureFormat].m_fmtSrgb : s_textureFormat[m_textureFormat].m_fmt
  4938. ;
  4939. m_components = s_textureFormat[m_textureFormat].m_mapping;
  4940. const bool convert = m_textureFormat != m_requestedFormat;
  4941. const uint8_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_textureFormat) );
  4942. m_aspectMask = getAspectMask(m_format);
  4943. m_sampler = s_msaa[bx::uint32_satsub( (m_flags & BGFX_TEXTURE_RT_MSAA_MASK) >> BGFX_TEXTURE_RT_MSAA_SHIFT, 1)];
  4944. if (imageContainer.m_cubeMap)
  4945. {
  4946. m_type = imageContainer.m_numLayers > 1
  4947. ? VK_IMAGE_VIEW_TYPE_CUBE_ARRAY
  4948. : VK_IMAGE_VIEW_TYPE_CUBE
  4949. ;
  4950. }
  4951. else if (imageContainer.m_depth > 1)
  4952. {
  4953. m_type = VK_IMAGE_VIEW_TYPE_3D;
  4954. }
  4955. else if (imageContainer.m_numLayers > 1)
  4956. {
  4957. m_type = VK_IMAGE_VIEW_TYPE_2D_ARRAY;
  4958. }
  4959. else
  4960. {
  4961. m_type = VK_IMAGE_VIEW_TYPE_2D;
  4962. }
  4963. m_numMips = ti.numMips;
  4964. m_numSides = ti.numLayers * (imageContainer.m_cubeMap ? 6 : 1);
  4965. const uint16_t numSides = ti.numLayers * (imageContainer.m_cubeMap ? 6 : 1);
  4966. const uint32_t numSrd = numSides * ti.numMips;
  4967. uint32_t kk = 0;
  4968. const bool compressed = bimg::isCompressed(bimg::TextureFormat::Enum(m_textureFormat) );
  4969. const bool swizzle = TextureFormat::BGRA8 == m_textureFormat && 0 != (m_flags & BGFX_TEXTURE_COMPUTE_WRITE);
  4970. const bool writeOnly = 0 != (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY);
  4971. const bool computeWrite = 0 != (m_flags & BGFX_TEXTURE_COMPUTE_WRITE);
  4972. const bool renderTarget = 0 != (m_flags & BGFX_TEXTURE_RT_MASK);
  4973. const bool blit = 0 != (m_flags & BGFX_TEXTURE_BLIT_DST);
  4974. BX_UNUSED(swizzle, writeOnly, computeWrite, renderTarget, blit);
  4975. BX_TRACE(
  4976. "Texture %3d: %s (requested: %s), %dx%dx%d%s RT[%c], BO[%c], CW[%c]%s."
  4977. , (int)(this - s_renderVK->m_textures)
  4978. , getName( (TextureFormat::Enum)m_textureFormat)
  4979. , getName( (TextureFormat::Enum)m_requestedFormat)
  4980. , ti.width
  4981. , ti.height
  4982. , ti.depth
  4983. , imageContainer.m_cubeMap ? "x6" : ""
  4984. , renderTarget ? 'x' : ' '
  4985. , writeOnly ? 'x' : ' '
  4986. , computeWrite ? 'x' : ' '
  4987. , swizzle ? " (swizzle BGRA8 -> RGBA8)" : ""
  4988. );
  4989. VK_CHECK(createImages(_commandBuffer) );
  4990. // decode images
  4991. struct ImageInfo
  4992. {
  4993. uint8_t* data;
  4994. uint32_t width;
  4995. uint32_t height;
  4996. uint32_t depth;
  4997. uint32_t pitch;
  4998. uint32_t slice;
  4999. uint32_t size;
  5000. uint32_t mipLevel;
  5001. uint32_t layer;
  5002. };
  5003. ImageInfo* imageInfos = (ImageInfo*)bx::alloc(g_allocator, sizeof(ImageInfo) * numSrd);
  5004. bx::memSet(imageInfos, 0, sizeof(ImageInfo) * numSrd);
  5005. uint32_t alignment = 1; // tightly aligned buffer
  5006. for (uint16_t side = 0; side < numSides; ++side)
  5007. {
  5008. for (uint8_t lod = 0; lod < ti.numMips; ++lod)
  5009. {
  5010. bimg::ImageMip mip;
  5011. if (bimg::imageGetRawData(imageContainer, side, lod + startLod, _mem->data, _mem->size, mip) )
  5012. {
  5013. if (convert)
  5014. {
  5015. const uint32_t pitch = bx::strideAlign(bx::max<uint32_t>(mip.m_width, 4) * bpp / 8, alignment);
  5016. const uint32_t slice = bx::strideAlign(bx::max<uint32_t>(mip.m_height, 4) * pitch, alignment);
  5017. const uint32_t size = slice * mip.m_depth;
  5018. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5019. bimg::imageDecodeToBgra8(
  5020. g_allocator
  5021. , temp
  5022. , mip.m_data
  5023. , mip.m_width
  5024. , mip.m_height
  5025. , pitch
  5026. , mip.m_format
  5027. );
  5028. imageInfos[kk].data = temp;
  5029. imageInfos[kk].width = mip.m_width;
  5030. imageInfos[kk].height = mip.m_height;
  5031. imageInfos[kk].depth = mip.m_depth;
  5032. imageInfos[kk].pitch = pitch;
  5033. imageInfos[kk].slice = slice;
  5034. imageInfos[kk].size = size;
  5035. imageInfos[kk].mipLevel = lod;
  5036. imageInfos[kk].layer = side;
  5037. }
  5038. else if (compressed)
  5039. {
  5040. const uint32_t pitch = bx::strideAlign( (mip.m_width / blockInfo.blockWidth) * mip.m_blockSize, alignment);
  5041. const uint32_t slice = bx::strideAlign( (mip.m_height / blockInfo.blockHeight) * pitch, alignment);
  5042. const uint32_t size = slice * mip.m_depth;
  5043. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5044. bimg::imageCopy(
  5045. temp
  5046. , mip.m_height / blockInfo.blockHeight
  5047. , (mip.m_width / blockInfo.blockWidth) * mip.m_blockSize
  5048. , mip.m_depth
  5049. , mip.m_data
  5050. , pitch
  5051. );
  5052. imageInfos[kk].data = temp;
  5053. imageInfos[kk].width = mip.m_width;
  5054. imageInfos[kk].height = mip.m_height;
  5055. imageInfos[kk].depth = mip.m_depth;
  5056. imageInfos[kk].pitch = pitch;
  5057. imageInfos[kk].slice = slice;
  5058. imageInfos[kk].size = size;
  5059. imageInfos[kk].mipLevel = lod;
  5060. imageInfos[kk].layer = side;
  5061. }
  5062. else
  5063. {
  5064. const uint32_t pitch = bx::strideAlign(mip.m_width * mip.m_bpp / 8, alignment);
  5065. const uint32_t slice = bx::strideAlign(mip.m_height * pitch, alignment);
  5066. const uint32_t size = slice * mip.m_depth;
  5067. uint8_t* temp = (uint8_t*)bx::alloc(g_allocator, size);
  5068. bimg::imageCopy(
  5069. temp
  5070. , mip.m_height
  5071. , mip.m_width * mip.m_bpp / 8
  5072. , mip.m_depth
  5073. , mip.m_data
  5074. , pitch
  5075. );
  5076. imageInfos[kk].data = temp;
  5077. imageInfos[kk].width = mip.m_width;
  5078. imageInfos[kk].height = mip.m_height;
  5079. imageInfos[kk].depth = mip.m_depth;
  5080. imageInfos[kk].pitch = pitch;
  5081. imageInfos[kk].slice = slice;
  5082. imageInfos[kk].size = size;
  5083. imageInfos[kk].mipLevel = lod;
  5084. imageInfos[kk].layer = side;
  5085. }
  5086. }
  5087. ++kk;
  5088. }
  5089. }
  5090. uint32_t totalMemSize = 0;
  5091. VkBufferImageCopy* bufferCopyInfo = (VkBufferImageCopy*)bx::alloc(g_allocator, sizeof(VkBufferImageCopy) * numSrd);
  5092. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5093. {
  5094. const uint32_t idealWidth = bx::max<uint32_t>(1, m_width >> imageInfos[ii].mipLevel);
  5095. const uint32_t idealHeight = bx::max<uint32_t>(1, m_height >> imageInfos[ii].mipLevel);
  5096. bufferCopyInfo[ii].bufferOffset = totalMemSize;
  5097. bufferCopyInfo[ii].bufferRowLength = 0; // assume that image data are tightly aligned
  5098. bufferCopyInfo[ii].bufferImageHeight = 0; // assume that image data are tightly aligned
  5099. bufferCopyInfo[ii].imageSubresource.aspectMask = m_aspectMask;
  5100. bufferCopyInfo[ii].imageSubresource.mipLevel = imageInfos[ii].mipLevel;
  5101. bufferCopyInfo[ii].imageSubresource.baseArrayLayer = imageInfos[ii].layer;
  5102. bufferCopyInfo[ii].imageSubresource.layerCount = 1;
  5103. bufferCopyInfo[ii].imageOffset = { 0, 0, 0 };
  5104. bufferCopyInfo[ii].imageExtent = { idealWidth, idealHeight, imageInfos[ii].depth };
  5105. totalMemSize += imageInfos[ii].size;
  5106. }
  5107. if (totalMemSize > 0)
  5108. {
  5109. const VkDevice device = s_renderVK->m_device;
  5110. VkBuffer stagingBuffer;
  5111. VkDeviceMemory stagingDeviceMem;
  5112. VK_CHECK(s_renderVK->createStagingBuffer(totalMemSize, &stagingBuffer, &stagingDeviceMem) );
  5113. uint8_t* mappedMemory;
  5114. VK_CHECK(vkMapMemory(
  5115. device
  5116. , stagingDeviceMem
  5117. , 0
  5118. , totalMemSize
  5119. , 0
  5120. , (void**)&mappedMemory
  5121. ) );
  5122. // copy image to staging buffer
  5123. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5124. {
  5125. bx::memCopy(mappedMemory, imageInfos[ii].data, imageInfos[ii].size);
  5126. mappedMemory += imageInfos[ii].size;
  5127. }
  5128. vkUnmapMemory(device, stagingDeviceMem);
  5129. copyBufferToTexture(_commandBuffer, stagingBuffer, numSrd, bufferCopyInfo);
  5130. s_renderVK->release(stagingBuffer);
  5131. s_renderVK->release(stagingDeviceMem);
  5132. }
  5133. else
  5134. {
  5135. setImageMemoryBarrier(_commandBuffer, m_sampledLayout);
  5136. }
  5137. bx::free(g_allocator, bufferCopyInfo);
  5138. for (uint32_t ii = 0; ii < numSrd; ++ii)
  5139. {
  5140. bx::free(g_allocator, imageInfos[ii].data);
  5141. }
  5142. bx::free(g_allocator, imageInfos);
  5143. m_readback.create(m_textureImage, m_width, m_height, TextureFormat::Enum(m_textureFormat) );
  5144. }
  5145. return m_directAccessPtr;
  5146. }
  5147. void TextureVK::destroy()
  5148. {
  5149. m_readback.destroy();
  5150. if (VK_NULL_HANDLE != m_textureImage)
  5151. {
  5152. s_renderVK->release(m_textureImage);
  5153. s_renderVK->release(m_textureDeviceMem);
  5154. }
  5155. if (VK_NULL_HANDLE != m_singleMsaaImage)
  5156. {
  5157. s_renderVK->release(m_singleMsaaImage);
  5158. s_renderVK->release(m_singleMsaaDeviceMem);
  5159. }
  5160. m_currentImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  5161. m_currentSingleMsaaImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  5162. }
  5163. void TextureVK::update(VkCommandBuffer _commandBuffer, uint8_t _side, uint8_t _mip, const Rect& _rect, uint16_t _z, uint16_t _depth, uint16_t _pitch, const Memory* _mem)
  5164. {
  5165. const uint32_t bpp = bimg::getBitsPerPixel(bimg::TextureFormat::Enum(m_textureFormat) );
  5166. uint32_t rectpitch = _rect.m_width * bpp / 8;
  5167. uint32_t slicepitch = rectpitch * _rect.m_height;
  5168. if (bimg::isCompressed(bimg::TextureFormat::Enum(m_textureFormat) ) )
  5169. {
  5170. const bimg::ImageBlockInfo& blockInfo = bimg::getBlockInfo(bimg::TextureFormat::Enum(m_textureFormat) );
  5171. rectpitch = (_rect.m_width / blockInfo.blockWidth ) * blockInfo.blockSize;
  5172. slicepitch = (_rect.m_height / blockInfo.blockHeight) * rectpitch;
  5173. }
  5174. const uint32_t srcpitch = UINT16_MAX == _pitch ? rectpitch : _pitch;
  5175. const uint32_t size = UINT16_MAX == _pitch ? slicepitch * _depth: _rect.m_height * _pitch * _depth;
  5176. const bool convert = m_textureFormat != m_requestedFormat;
  5177. VkBufferImageCopy region;
  5178. region.bufferOffset = 0;
  5179. region.bufferRowLength = (_pitch == UINT16_MAX ? 0 : _pitch * 8 / bpp);
  5180. region.bufferImageHeight = 0;
  5181. region.imageSubresource.aspectMask = m_aspectMask;
  5182. region.imageSubresource.mipLevel = _mip;
  5183. region.imageSubresource.baseArrayLayer = 0;
  5184. region.imageSubresource.layerCount = 1;
  5185. region.imageOffset = { _rect.m_x, _rect.m_y, 0 };
  5186. region.imageExtent = { _rect.m_width, _rect.m_height, _depth };
  5187. uint8_t* data = _mem->data;
  5188. uint8_t* temp = NULL;
  5189. if (convert)
  5190. {
  5191. temp = (uint8_t*)bx::alloc(g_allocator, slicepitch);
  5192. bimg::imageDecodeToBgra8(g_allocator, temp, data, _rect.m_width, _rect.m_height, srcpitch, bimg::TextureFormat::Enum(m_requestedFormat));
  5193. data = temp;
  5194. region.imageExtent =
  5195. {
  5196. bx::max(1u, m_width >> _mip),
  5197. bx::max(1u, m_height >> _mip),
  5198. _depth,
  5199. };
  5200. }
  5201. VkBuffer stagingBuffer = VK_NULL_HANDLE;
  5202. VkDeviceMemory stagingDeviceMem = VK_NULL_HANDLE;
  5203. VK_CHECK(s_renderVK->createStagingBuffer(size, &stagingBuffer, &stagingDeviceMem, data) );
  5204. if (VK_IMAGE_VIEW_TYPE_3D == m_type)
  5205. {
  5206. region.imageOffset.z = _z;
  5207. }
  5208. else if (VK_IMAGE_VIEW_TYPE_CUBE == m_type
  5209. || VK_IMAGE_VIEW_TYPE_CUBE_ARRAY == m_type)
  5210. {
  5211. region.imageSubresource.baseArrayLayer = _z * 6 + _side;
  5212. }
  5213. else
  5214. {
  5215. region.imageSubresource.baseArrayLayer = _z;
  5216. }
  5217. copyBufferToTexture(_commandBuffer, stagingBuffer, 1, &region);
  5218. s_renderVK->release(stagingBuffer);
  5219. s_renderVK->release(stagingDeviceMem);
  5220. if (NULL != temp)
  5221. {
  5222. bx::free(g_allocator, temp);
  5223. }
  5224. }
  5225. void TextureVK::resolve(VkCommandBuffer _commandBuffer, uint8_t _resolve, uint32_t _layer, uint32_t _numLayers, uint32_t _mip)
  5226. {
  5227. const bool needResolve = VK_NULL_HANDLE != m_singleMsaaImage;
  5228. const bool needMipGen = true
  5229. && !needResolve
  5230. && 0 != (m_flags & BGFX_TEXTURE_RT_MASK)
  5231. && 0 == (m_flags & BGFX_TEXTURE_RT_WRITE_ONLY)
  5232. && (_mip + 1) < m_numMips
  5233. && 0 != (_resolve & BGFX_RESOLVE_AUTO_GEN_MIPS);
  5234. const VkImageLayout oldLayout = m_currentImageLayout;
  5235. const VkImageLayout oldSingleMsaaLayout = m_currentSingleMsaaImageLayout;
  5236. if (needResolve)
  5237. {
  5238. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
  5239. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, true);
  5240. VkImageResolve resolve;
  5241. resolve.srcOffset.x = 0;
  5242. resolve.srcOffset.y = 0;
  5243. resolve.srcOffset.z = 0;
  5244. resolve.dstOffset.x = 0;
  5245. resolve.dstOffset.y = 0;
  5246. resolve.dstOffset.z = 0;
  5247. resolve.srcSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5248. resolve.srcSubresource.mipLevel = _mip;
  5249. resolve.srcSubresource.baseArrayLayer = _layer;
  5250. resolve.srcSubresource.layerCount = _numLayers;
  5251. resolve.dstSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5252. resolve.dstSubresource.mipLevel = _mip;
  5253. resolve.dstSubresource.baseArrayLayer = _layer;
  5254. resolve.dstSubresource.layerCount = _numLayers;
  5255. resolve.extent.width = m_width;
  5256. resolve.extent.height = m_height;
  5257. resolve.extent.depth = 1;
  5258. vkCmdResolveImage(
  5259. _commandBuffer
  5260. , m_textureImage
  5261. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5262. , m_singleMsaaImage
  5263. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5264. , 1
  5265. , &resolve
  5266. );
  5267. }
  5268. if (needMipGen)
  5269. {
  5270. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  5271. int32_t mipWidth = bx::max<int32_t>(int32_t(m_width) >> _mip, 1);
  5272. int32_t mipHeight = bx::max<int32_t>(int32_t(m_height) >> _mip, 1);
  5273. const VkFilter filter = bimg::isDepth(bimg::TextureFormat::Enum(m_textureFormat) )
  5274. ? VK_FILTER_NEAREST
  5275. : VK_FILTER_LINEAR
  5276. ;
  5277. VkImageBlit blit;
  5278. blit.srcOffsets[0] = { 0, 0, 0 };
  5279. blit.srcOffsets[1] = { mipWidth, mipHeight, 1 };
  5280. blit.srcSubresource.aspectMask = m_aspectMask;
  5281. blit.srcSubresource.mipLevel = 0;
  5282. blit.srcSubresource.baseArrayLayer = _layer;
  5283. blit.srcSubresource.layerCount = _numLayers;
  5284. blit.dstOffsets[0] = { 0, 0, 0 };
  5285. blit.dstOffsets[1] = { mipWidth, mipHeight, 1 };
  5286. blit.dstSubresource.aspectMask = m_aspectMask;
  5287. blit.dstSubresource.mipLevel = 0;
  5288. blit.dstSubresource.baseArrayLayer = _layer;
  5289. blit.dstSubresource.layerCount = _numLayers;
  5290. for (uint32_t i = _mip + 1; i < m_numMips; i++)
  5291. {
  5292. blit.srcOffsets[1] = { mipWidth, mipHeight, 1 };
  5293. blit.srcSubresource.mipLevel = i - 1;
  5294. mipWidth = bx::uint32_max(mipWidth >> 1, 1);
  5295. mipHeight = bx::uint32_max(mipHeight >> 1, 1);
  5296. blit.dstOffsets[1] = { mipWidth, mipHeight, 1 };
  5297. blit.dstSubresource.mipLevel = i;
  5298. vk::setImageMemoryBarrier(
  5299. _commandBuffer
  5300. , m_textureImage
  5301. , m_aspectMask
  5302. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5303. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5304. , blit.srcSubresource.mipLevel
  5305. , 1
  5306. , _layer
  5307. , _numLayers
  5308. );
  5309. vkCmdBlitImage(
  5310. _commandBuffer
  5311. , m_textureImage
  5312. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5313. , m_textureImage
  5314. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5315. , 1
  5316. , &blit
  5317. , filter
  5318. );
  5319. }
  5320. vk::setImageMemoryBarrier(
  5321. _commandBuffer
  5322. , m_textureImage
  5323. , m_aspectMask
  5324. , VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  5325. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5326. , _mip
  5327. , m_numMips - _mip - 1
  5328. , _layer
  5329. , _numLayers
  5330. );
  5331. }
  5332. setImageMemoryBarrier(_commandBuffer, oldLayout);
  5333. setImageMemoryBarrier(_commandBuffer, oldSingleMsaaLayout, true);
  5334. }
  5335. void TextureVK::copyBufferToTexture(VkCommandBuffer _commandBuffer, VkBuffer _stagingBuffer, uint32_t _bufferImageCopyCount, VkBufferImageCopy* _bufferImageCopy)
  5336. {
  5337. const VkImageLayout oldLayout = m_currentImageLayout == VK_IMAGE_LAYOUT_UNDEFINED
  5338. ? m_sampledLayout
  5339. : m_currentImageLayout
  5340. ;
  5341. setImageMemoryBarrier(_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  5342. vkCmdCopyBufferToImage(
  5343. _commandBuffer
  5344. , _stagingBuffer
  5345. , m_textureImage
  5346. , VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL
  5347. , _bufferImageCopyCount
  5348. , _bufferImageCopy
  5349. );
  5350. setImageMemoryBarrier(_commandBuffer, oldLayout);
  5351. }
  5352. VkImageLayout TextureVK::setImageMemoryBarrier(VkCommandBuffer _commandBuffer, VkImageLayout _newImageLayout, bool _singleMsaaImage)
  5353. {
  5354. if (_singleMsaaImage && VK_NULL_HANDLE == m_singleMsaaImage)
  5355. {
  5356. return VK_IMAGE_LAYOUT_UNDEFINED;
  5357. }
  5358. VkImageLayout& currentLayout = _singleMsaaImage
  5359. ? m_currentSingleMsaaImageLayout
  5360. : m_currentImageLayout
  5361. ;
  5362. const VkImageLayout oldLayout = currentLayout;
  5363. if (currentLayout == _newImageLayout)
  5364. {
  5365. return oldLayout;
  5366. }
  5367. const VkImage image = _singleMsaaImage
  5368. ? m_singleMsaaImage
  5369. : m_textureImage
  5370. ;
  5371. vk::setImageMemoryBarrier(
  5372. _commandBuffer
  5373. , image
  5374. , m_aspectMask
  5375. , currentLayout
  5376. , _newImageLayout
  5377. );
  5378. currentLayout = _newImageLayout;
  5379. return oldLayout;
  5380. }
  5381. VkResult TextureVK::createView(uint32_t _layer, uint32_t _numLayers, uint32_t _mip, uint32_t _numMips, VkImageViewType _type, VkImageAspectFlags _aspectMask, bool _renderTarget, ::VkImageView* _view) const
  5382. {
  5383. VkResult result = VK_SUCCESS;
  5384. if (VK_IMAGE_VIEW_TYPE_3D == m_type)
  5385. {
  5386. BX_ASSERT(false
  5387. || !_renderTarget
  5388. || !(m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  5389. , "3D image can't be a depth attachment"
  5390. );
  5391. }
  5392. if (VK_IMAGE_VIEW_TYPE_CUBE == _type
  5393. || VK_IMAGE_VIEW_TYPE_CUBE_ARRAY == _type)
  5394. {
  5395. BX_ASSERT(_numLayers % 6 == 0, "");
  5396. BX_ASSERT(
  5397. VK_IMAGE_VIEW_TYPE_3D != m_type
  5398. , "3D image can't be aliased as a cube texture"
  5399. );
  5400. }
  5401. VkImageViewCreateInfo viewInfo;
  5402. viewInfo.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
  5403. viewInfo.pNext = NULL;
  5404. viewInfo.flags = 0;
  5405. viewInfo.image = ((VK_NULL_HANDLE != m_singleMsaaImage) && !_renderTarget)
  5406. ? m_singleMsaaImage
  5407. : m_textureImage
  5408. ;
  5409. viewInfo.viewType = _type;
  5410. viewInfo.format = m_format;
  5411. viewInfo.components = m_components;
  5412. viewInfo.subresourceRange.aspectMask = m_aspectMask & _aspectMask;
  5413. viewInfo.subresourceRange.baseMipLevel = _mip;
  5414. viewInfo.subresourceRange.levelCount = _numMips;
  5415. viewInfo.subresourceRange.baseArrayLayer = _layer;
  5416. viewInfo.subresourceRange.layerCount = 1;
  5417. if (VK_IMAGE_VIEW_TYPE_2D != _type
  5418. && VK_IMAGE_VIEW_TYPE_3D != _type)
  5419. {
  5420. viewInfo.subresourceRange.layerCount = VK_IMAGE_VIEW_TYPE_CUBE == _type
  5421. ? 6
  5422. : _numLayers
  5423. ;
  5424. }
  5425. VkImageView view = VK_NULL_HANDLE;
  5426. result = vkCreateImageView(
  5427. s_renderVK->m_device
  5428. , &viewInfo
  5429. , s_renderVK->m_allocatorCb
  5430. , &view
  5431. );
  5432. if (VK_SUCCESS != result)
  5433. {
  5434. BX_TRACE("Create texture view error: vkCreateImageView failed %d: %s.", result, getName(result) );
  5435. return result;
  5436. }
  5437. *_view = view;
  5438. return result;
  5439. }
  5440. VkImageAspectFlags TextureVK::getAspectMask(VkFormat _format)
  5441. {
  5442. switch (_format)
  5443. {
  5444. case VK_FORMAT_S8_UINT:
  5445. return VK_IMAGE_ASPECT_STENCIL_BIT;
  5446. break;
  5447. case VK_FORMAT_D16_UNORM:
  5448. case VK_FORMAT_X8_D24_UNORM_PACK32:
  5449. case VK_FORMAT_D32_SFLOAT:
  5450. return VK_IMAGE_ASPECT_DEPTH_BIT;
  5451. case VK_FORMAT_D16_UNORM_S8_UINT:
  5452. case VK_FORMAT_D24_UNORM_S8_UINT:
  5453. case VK_FORMAT_D32_SFLOAT_S8_UINT:
  5454. return VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT;
  5455. default:
  5456. return VK_IMAGE_ASPECT_COLOR_BIT;
  5457. }
  5458. }
  5459. VkResult SwapChainVK::create(VkCommandBuffer _commandBuffer, void* _nwh, const Resolution& _resolution, TextureFormat::Enum _depthFormat)
  5460. {
  5461. struct ErrorState
  5462. {
  5463. enum Enum
  5464. {
  5465. Default,
  5466. SurfaceCreated,
  5467. SwapChainCreated,
  5468. AttachmentsCreated
  5469. };
  5470. };
  5471. ErrorState::Enum errorState = ErrorState::Default;
  5472. VkResult result = VK_SUCCESS;
  5473. if (NULL == _nwh)
  5474. {
  5475. return result;
  5476. }
  5477. m_nwh = _nwh;
  5478. m_resolution = _resolution;
  5479. m_depthFormat = TextureFormat::Count == _depthFormat ? TextureFormat::D24S8 : _depthFormat;
  5480. m_queue = s_renderVK->m_globalQueue;
  5481. result = createSurface();
  5482. if (VK_SUCCESS != result)
  5483. {
  5484. BX_TRACE("Create swap chain error: creating surface failed %d: %s.", result, getName(result) );
  5485. goto error;
  5486. }
  5487. errorState = ErrorState::SurfaceCreated;
  5488. {
  5489. m_sci.sType = VK_STRUCTURE_TYPE_SWAPCHAIN_CREATE_INFO_KHR;
  5490. m_sci.pNext = NULL;
  5491. m_sci.flags = 0;
  5492. m_sci.imageArrayLayers = 1;
  5493. m_sci.imageSharingMode = VK_SHARING_MODE_EXCLUSIVE;
  5494. m_sci.queueFamilyIndexCount = 0;
  5495. m_sci.pQueueFamilyIndices = NULL;
  5496. m_sci.preTransform = VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR;
  5497. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5498. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  5499. {
  5500. m_backBufferColorImage[ii] = VK_NULL_HANDLE;
  5501. m_backBufferColorImageView[ii] = VK_NULL_HANDLE;
  5502. m_backBufferFrameBuffer[ii] = VK_NULL_HANDLE;
  5503. m_backBufferFence[ii] = VK_NULL_HANDLE;
  5504. m_presentDoneSemaphore[ii] = VK_NULL_HANDLE;
  5505. m_renderDoneSemaphore[ii] = VK_NULL_HANDLE;
  5506. }
  5507. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  5508. m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  5509. result = createSwapChain();
  5510. if (VK_SUCCESS != result)
  5511. {
  5512. BX_TRACE("Create swap chain error: creating swapchain and image views failed %d: %s", result, getName(result) );
  5513. goto error;
  5514. }
  5515. }
  5516. errorState = ErrorState::SwapChainCreated;
  5517. {
  5518. result = createAttachments(_commandBuffer);
  5519. if (VK_SUCCESS != result)
  5520. {
  5521. BX_TRACE("Create swap chain error: creating MSAA/depth attachments failed %d: %s.", result, getName(result) );
  5522. goto error;
  5523. }
  5524. }
  5525. errorState = ErrorState::AttachmentsCreated;
  5526. {
  5527. result = createFrameBuffer();
  5528. if (VK_SUCCESS != result)
  5529. {
  5530. BX_TRACE("Create swap chain error: creating frame buffers failed %d: %s.", result, getName(result) );
  5531. goto error;
  5532. }
  5533. }
  5534. return VK_SUCCESS;
  5535. error:
  5536. BX_TRACE("errorState %d", errorState);
  5537. switch (errorState)
  5538. {
  5539. case ErrorState::AttachmentsCreated:
  5540. releaseAttachments();
  5541. BX_FALLTHROUGH;
  5542. case ErrorState::SwapChainCreated:
  5543. releaseSwapChain();
  5544. BX_FALLTHROUGH;
  5545. case ErrorState::SurfaceCreated:
  5546. releaseSurface();
  5547. BX_FALLTHROUGH;
  5548. case ErrorState::Default:
  5549. break;
  5550. };
  5551. return VK_SUCCESS != result
  5552. ? result
  5553. : VK_ERROR_INITIALIZATION_FAILED
  5554. ;
  5555. }
  5556. void SwapChainVK::destroy()
  5557. {
  5558. if (VK_NULL_HANDLE != m_swapchain)
  5559. {
  5560. releaseFrameBuffer();
  5561. releaseAttachments();
  5562. releaseSwapChain();
  5563. releaseSurface();
  5564. // can't delay-delete the surface, since there can only be one swapchain per surface
  5565. // new framebuffer with the same window would get an error at swapchain creation
  5566. s_renderVK->kick(true);
  5567. }
  5568. m_nwh = NULL;
  5569. }
  5570. void SwapChainVK::update(VkCommandBuffer _commandBuffer, void* _nwh, const Resolution& _resolution)
  5571. {
  5572. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5573. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  5574. m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  5575. const uint64_t recreateSurfaceMask = BGFX_RESET_HIDPI;
  5576. const uint64_t recreateSwapchainMask = BGFX_RESET_VSYNC | BGFX_RESET_SRGB_BACKBUFFER;
  5577. const uint64_t recreateAttachmentsMask = BGFX_RESET_MSAA_MASK;
  5578. const bool recreateSurface = false
  5579. || m_needToRecreateSurface
  5580. || m_nwh != _nwh
  5581. || (m_resolution.reset & recreateSurfaceMask) != (_resolution.reset & recreateSurfaceMask)
  5582. ;
  5583. const bool recreateSwapchain = false
  5584. || m_resolution.format != _resolution.format
  5585. || m_resolution.width != _resolution.width
  5586. || m_resolution.height != _resolution.height
  5587. || (m_resolution.reset & recreateSwapchainMask) != (_resolution.reset & recreateSwapchainMask)
  5588. || recreateSurface
  5589. ;
  5590. const bool recreateAttachments = false
  5591. || (m_resolution.reset & recreateAttachmentsMask) != (_resolution.reset & recreateAttachmentsMask)
  5592. || recreateSwapchain
  5593. ;
  5594. m_nwh = _nwh;
  5595. m_resolution = _resolution;
  5596. if (recreateAttachments)
  5597. {
  5598. releaseFrameBuffer();
  5599. releaseAttachments();
  5600. if (recreateSwapchain)
  5601. {
  5602. releaseSwapChain();
  5603. if (recreateSurface)
  5604. {
  5605. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5606. releaseSurface();
  5607. s_renderVK->kick(true);
  5608. _commandBuffer = s_renderVK->m_commandBuffer;
  5609. VkResult result = createSurface();
  5610. if (VK_SUCCESS != result)
  5611. {
  5612. BX_TRACE("Surface lost.");
  5613. return;
  5614. }
  5615. }
  5616. VkSurfaceCapabilitiesKHR surfaceCapabilities;
  5617. VK_CHECK(vkGetPhysicalDeviceSurfaceCapabilitiesKHR(physicalDevice, m_surface, &surfaceCapabilities) );
  5618. const uint32_t width = bx::clamp<uint32_t>(
  5619. m_resolution.width
  5620. , surfaceCapabilities.minImageExtent.width
  5621. , surfaceCapabilities.maxImageExtent.width
  5622. );
  5623. const uint32_t height = bx::clamp<uint32_t>(
  5624. m_resolution.height
  5625. , surfaceCapabilities.minImageExtent.height
  5626. , surfaceCapabilities.maxImageExtent.height
  5627. );
  5628. // swapchain can't have size 0
  5629. // on some platforms this happens when minimized
  5630. if (width == 0
  5631. || height == 0)
  5632. {
  5633. m_sci.oldSwapchain = VK_NULL_HANDLE;
  5634. s_renderVK->kick(true);
  5635. return;
  5636. }
  5637. VK_CHECK(createSwapChain() );
  5638. }
  5639. VK_CHECK(createAttachments(_commandBuffer) );
  5640. VK_CHECK(createFrameBuffer() );
  5641. }
  5642. }
  5643. VkResult SwapChainVK::createSurface()
  5644. {
  5645. VkResult result = VK_ERROR_INITIALIZATION_FAILED;
  5646. const VkInstance instance = s_renderVK->m_instance;
  5647. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  5648. #if BX_PLATFORM_WINDOWS
  5649. {
  5650. if (NULL != vkCreateWin32SurfaceKHR)
  5651. {
  5652. VkWin32SurfaceCreateInfoKHR sci;
  5653. sci.sType = VK_STRUCTURE_TYPE_WIN32_SURFACE_CREATE_INFO_KHR;
  5654. sci.pNext = NULL;
  5655. sci.flags = 0;
  5656. sci.hinstance = (HINSTANCE)GetModuleHandle(NULL);
  5657. sci.hwnd = (HWND)m_nwh;
  5658. result = vkCreateWin32SurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5659. }
  5660. }
  5661. #elif BX_PLATFORM_ANDROID
  5662. {
  5663. if (NULL != vkCreateAndroidSurfaceKHR)
  5664. {
  5665. VkAndroidSurfaceCreateInfoKHR sci;
  5666. sci.sType = VK_STRUCTURE_TYPE_ANDROID_SURFACE_CREATE_INFO_KHR;
  5667. sci.pNext = NULL;
  5668. sci.flags = 0;
  5669. sci.window = (ANativeWindow*)m_nwh;
  5670. result = vkCreateAndroidSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5671. }
  5672. }
  5673. #elif BX_PLATFORM_LINUX
  5674. {
  5675. #if defined(WL_EGL_PLATFORM)
  5676. if (g_platformData.type == bgfx::NativeWindowHandleType::Wayland)
  5677. {
  5678. VkWaylandSurfaceCreateInfoKHR sci;
  5679. sci.sType = VK_STRUCTURE_TYPE_WAYLAND_SURFACE_CREATE_INFO_KHR;
  5680. sci.pNext = NULL;
  5681. sci.flags = 0;
  5682. sci.display = (wl_display*)g_platformData.ndt;
  5683. sci.surface = (wl_surface*)((wl_egl_window*)m_nwh)->surface;
  5684. result = vkCreateWaylandSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5685. }
  5686. else
  5687. #endif // defined(WL_EGL_PLATFORM)
  5688. {
  5689. if (NULL != vkCreateXlibSurfaceKHR)
  5690. {
  5691. VkXlibSurfaceCreateInfoKHR sci;
  5692. sci.sType = VK_STRUCTURE_TYPE_XLIB_SURFACE_CREATE_INFO_KHR;
  5693. sci.pNext = NULL;
  5694. sci.flags = 0;
  5695. sci.dpy = (Display*)g_platformData.ndt;
  5696. sci.window = (Window)m_nwh;
  5697. result = vkCreateXlibSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5698. }
  5699. if (VK_SUCCESS != result)
  5700. {
  5701. void* xcbdll = bx::dlopen("libX11-xcb.so.1");
  5702. if (NULL != xcbdll
  5703. && NULL != vkCreateXcbSurfaceKHR)
  5704. {
  5705. typedef xcb_connection_t* (*PFN_XGETXCBCONNECTION)(Display*);
  5706. PFN_XGETXCBCONNECTION XGetXCBConnection = (PFN_XGETXCBCONNECTION)bx::dlsym(xcbdll, "XGetXCBConnection");
  5707. union { void* ptr; xcb_window_t window; } cast = { m_nwh };
  5708. VkXcbSurfaceCreateInfoKHR sci;
  5709. sci.sType = VK_STRUCTURE_TYPE_XCB_SURFACE_CREATE_INFO_KHR;
  5710. sci.pNext = NULL;
  5711. sci.flags = 0;
  5712. sci.connection = XGetXCBConnection( (Display*)g_platformData.ndt);
  5713. sci.window = cast.window;
  5714. result = vkCreateXcbSurfaceKHR(instance, &sci, allocatorCb, &m_surface);
  5715. bx::dlclose(xcbdll);
  5716. }
  5717. }
  5718. }
  5719. }
  5720. #elif BX_PLATFORM_OSX
  5721. {
  5722. if (NULL != vkCreateMacOSSurfaceMVK)
  5723. {
  5724. NSWindow* window = (NSWindow*)(m_nwh);
  5725. CAMetalLayer* layer = (CAMetalLayer*)(m_nwh);
  5726. if ([window isKindOfClass:[NSWindow class]])
  5727. {
  5728. NSView *contentView = (NSView *)window.contentView;
  5729. layer = [CAMetalLayer layer];
  5730. [contentView setWantsLayer : YES];
  5731. [contentView setLayer : layer];
  5732. }
  5733. else if ([layer isKindOfClass:[CAMetalLayer class]])
  5734. {
  5735. NSView *contentView = (NSView *)layer.delegate;
  5736. window = contentView.window;
  5737. }
  5738. else
  5739. {
  5740. BX_WARN(0, "Unable to create MoltenVk surface. Please set platform data window to an NSWindow or CAMetalLayer");
  5741. return result;
  5742. }
  5743. if (m_resolution.reset & BGFX_RESET_HIDPI)
  5744. {
  5745. layer.contentsScale = [window backingScaleFactor];
  5746. }
  5747. VkMacOSSurfaceCreateInfoMVK sci;
  5748. sci.sType = VK_STRUCTURE_TYPE_MACOS_SURFACE_CREATE_INFO_MVK;
  5749. sci.pNext = NULL;
  5750. sci.flags = 0;
  5751. sci.pView = (__bridge void*)layer;
  5752. result = vkCreateMacOSSurfaceMVK(instance, &sci, allocatorCb, &m_surface);
  5753. }
  5754. }
  5755. #else
  5756. # error "Figure out KHR surface..."
  5757. #endif // BX_PLATFORM_
  5758. m_needToRecreateSurface = false;
  5759. if (VK_SUCCESS != result)
  5760. {
  5761. BX_TRACE("Create surface error: vkCreate[Platform]SurfaceKHR failed %d: %s.", result, getName(result) );
  5762. return result;
  5763. }
  5764. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5765. const uint32_t queueFamily = s_renderVK->m_globalQueueFamily;
  5766. VkBool32 surfaceSupported;
  5767. result = vkGetPhysicalDeviceSurfaceSupportKHR(physicalDevice, queueFamily, m_surface, &surfaceSupported);
  5768. if (VK_SUCCESS != result
  5769. || !surfaceSupported)
  5770. {
  5771. BX_TRACE("Create surface error: Presentation to the given surface not supported.");
  5772. return VK_ERROR_INITIALIZATION_FAILED;
  5773. }
  5774. return result;
  5775. }
  5776. void SwapChainVK::releaseSurface()
  5777. {
  5778. release(m_surface);
  5779. }
  5780. VkResult SwapChainVK::createSwapChain()
  5781. {
  5782. VkResult result = VK_SUCCESS;
  5783. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  5784. const VkDevice device = s_renderVK->m_device;
  5785. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  5786. VkSurfaceCapabilitiesKHR surfaceCapabilities;
  5787. result = vkGetPhysicalDeviceSurfaceCapabilitiesKHR(physicalDevice, m_surface, &surfaceCapabilities);
  5788. if (VK_SUCCESS != result)
  5789. {
  5790. BX_TRACE("Create swapchain error: vkGetPhysicalDeviceSurfaceCapabilitiesKHR failed %d: %s.", result, getName(result) );
  5791. return result;
  5792. }
  5793. const uint32_t minSwapBufferCount = bx::max<uint32_t>(surfaceCapabilities.minImageCount, 2);
  5794. const uint32_t maxSwapBufferCount = surfaceCapabilities.maxImageCount == 0
  5795. ? kMaxBackBuffers
  5796. : bx::min<uint32_t>(surfaceCapabilities.maxImageCount, kMaxBackBuffers)
  5797. ;
  5798. if (minSwapBufferCount > maxSwapBufferCount)
  5799. {
  5800. BX_TRACE("Create swapchain error: Incompatible swapchain image count (min: %d, max: %d, MaxBackBuffers: %d)."
  5801. , minSwapBufferCount
  5802. , maxSwapBufferCount
  5803. , kMaxBackBuffers
  5804. );
  5805. return VK_ERROR_INITIALIZATION_FAILED;
  5806. }
  5807. const uint32_t swapBufferCount = bx::clamp<uint32_t>(m_resolution.numBackBuffers, minSwapBufferCount, maxSwapBufferCount);
  5808. const VkColorSpaceKHR surfaceColorSpace = VK_COLOR_SPACE_SRGB_NONLINEAR_KHR;
  5809. const bool srgb = !!(m_resolution.reset & BGFX_RESET_SRGB_BACKBUFFER);
  5810. m_colorFormat = findSurfaceFormat(m_resolution.format, surfaceColorSpace, srgb);
  5811. if (TextureFormat::Count == m_colorFormat)
  5812. {
  5813. BX_TRACE("Create swapchain error: Unable to find surface format (srgb: %d).", srgb);
  5814. return VK_ERROR_INITIALIZATION_FAILED;
  5815. }
  5816. const VkFormat surfaceFormat = srgb
  5817. ? s_textureFormat[m_colorFormat].m_fmtSrgb
  5818. : s_textureFormat[m_colorFormat].m_fmt
  5819. ;
  5820. const uint32_t width = bx::clamp<uint32_t>(
  5821. m_resolution.width
  5822. , surfaceCapabilities.minImageExtent.width
  5823. , surfaceCapabilities.maxImageExtent.width
  5824. );
  5825. const uint32_t height = bx::clamp<uint32_t>(
  5826. m_resolution.height
  5827. , surfaceCapabilities.minImageExtent.height
  5828. , surfaceCapabilities.maxImageExtent.height
  5829. );
  5830. VkCompositeAlphaFlagBitsKHR compositeAlpha = VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR;
  5831. if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_INHERIT_BIT_KHR)
  5832. {
  5833. compositeAlpha = VK_COMPOSITE_ALPHA_INHERIT_BIT_KHR;
  5834. }
  5835. else if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_PRE_MULTIPLIED_BIT_KHR)
  5836. {
  5837. compositeAlpha = VK_COMPOSITE_ALPHA_PRE_MULTIPLIED_BIT_KHR;
  5838. }
  5839. else if (surfaceCapabilities.supportedCompositeAlpha & VK_COMPOSITE_ALPHA_POST_MULTIPLIED_BIT_KHR)
  5840. {
  5841. compositeAlpha = VK_COMPOSITE_ALPHA_POST_MULTIPLIED_BIT_KHR;
  5842. }
  5843. const VkImageUsageFlags imageUsageMask = 0
  5844. | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT
  5845. | VK_IMAGE_USAGE_TRANSFER_SRC_BIT
  5846. | VK_IMAGE_USAGE_TRANSFER_DST_BIT
  5847. ;
  5848. const VkImageUsageFlags imageUsage = surfaceCapabilities.supportedUsageFlags & imageUsageMask;
  5849. m_supportsReadback = 0 != (imageUsage & VK_IMAGE_USAGE_TRANSFER_SRC_BIT);
  5850. m_supportsManualResolve = 0 != (imageUsage & VK_IMAGE_USAGE_TRANSFER_DST_BIT);
  5851. const bool vsync = !!(m_resolution.reset & BGFX_RESET_VSYNC);
  5852. uint32_t presentModeIdx = findPresentMode(vsync);
  5853. if (UINT32_MAX == presentModeIdx)
  5854. {
  5855. BX_TRACE("Create swapchain error: Unable to find present mode (vsync: %d).", vsync);
  5856. return VK_ERROR_INITIALIZATION_FAILED;
  5857. }
  5858. m_sci.surface = m_surface;
  5859. m_sci.minImageCount = swapBufferCount;
  5860. m_sci.imageFormat = surfaceFormat;
  5861. m_sci.imageColorSpace = surfaceColorSpace;
  5862. m_sci.imageExtent.width = width;
  5863. m_sci.imageExtent.height = height;
  5864. m_sci.imageUsage = imageUsage;
  5865. m_sci.compositeAlpha = compositeAlpha;
  5866. m_sci.presentMode = s_presentMode[presentModeIdx].mode;
  5867. m_sci.clipped = VK_FALSE;
  5868. result = vkCreateSwapchainKHR(device, &m_sci, allocatorCb, &m_swapchain);
  5869. if (VK_SUCCESS != result)
  5870. {
  5871. BX_TRACE("Create swapchain error: vkCreateSwapchainKHR failed %d: %s.", result, getName(result) );
  5872. return result;
  5873. }
  5874. m_sci.oldSwapchain = m_swapchain;
  5875. result = vkGetSwapchainImagesKHR(device, m_swapchain, &m_numSwapchainImages, NULL);
  5876. if (VK_SUCCESS != result)
  5877. {
  5878. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR failed %d: %s.", result, getName(result) );
  5879. return result;
  5880. }
  5881. if (m_numSwapchainImages < m_sci.minImageCount)
  5882. {
  5883. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR: numSwapchainImages %d < minImageCount %d."
  5884. , m_numSwapchainImages
  5885. , m_sci.minImageCount
  5886. );
  5887. return VK_ERROR_INITIALIZATION_FAILED;
  5888. }
  5889. if (m_numSwapchainImages > BX_COUNTOF(m_backBufferColorImage) )
  5890. {
  5891. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR: numSwapchainImages %d > countof(m_backBufferColorImage) %d."
  5892. , m_numSwapchainImages
  5893. , BX_COUNTOF(m_backBufferColorImage)
  5894. );
  5895. return VK_ERROR_INITIALIZATION_FAILED;
  5896. }
  5897. result = vkGetSwapchainImagesKHR(device, m_swapchain, &m_numSwapchainImages, &m_backBufferColorImage[0]);
  5898. if (VK_SUCCESS != result && VK_INCOMPLETE != result)
  5899. {
  5900. BX_TRACE("Create swapchain error: vkGetSwapchainImagesKHR failed %d: %s."
  5901. , result
  5902. , getName(result)
  5903. );
  5904. return result;
  5905. }
  5906. VkImageViewCreateInfo ivci;
  5907. ivci.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
  5908. ivci.pNext = NULL;
  5909. ivci.flags = 0;
  5910. ivci.viewType = VK_IMAGE_VIEW_TYPE_2D;
  5911. ivci.format = m_sci.imageFormat;
  5912. ivci.components.r = VK_COMPONENT_SWIZZLE_IDENTITY;
  5913. ivci.components.g = VK_COMPONENT_SWIZZLE_IDENTITY;
  5914. ivci.components.b = VK_COMPONENT_SWIZZLE_IDENTITY;
  5915. ivci.components.a = VK_COMPONENT_SWIZZLE_IDENTITY;
  5916. ivci.subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
  5917. ivci.subresourceRange.baseMipLevel = 0;
  5918. ivci.subresourceRange.levelCount = 1;
  5919. ivci.subresourceRange.baseArrayLayer = 0;
  5920. ivci.subresourceRange.layerCount = 1;
  5921. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  5922. {
  5923. ivci.image = m_backBufferColorImage[ii];
  5924. result = vkCreateImageView(device, &ivci, allocatorCb, &m_backBufferColorImageView[ii]);
  5925. if (VK_SUCCESS != result)
  5926. {
  5927. BX_TRACE("Create swapchain error: vkCreateImageView failed %d: %s.", result, getName(result) );
  5928. return result;
  5929. }
  5930. m_backBufferColorImageLayout[ii] = VK_IMAGE_LAYOUT_UNDEFINED;
  5931. }
  5932. VkSemaphoreCreateInfo sci;
  5933. sci.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
  5934. sci.pNext = NULL;
  5935. sci.flags = 0;
  5936. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  5937. {
  5938. if (VK_SUCCESS != vkCreateSemaphore(device, &sci, allocatorCb, &m_presentDoneSemaphore[ii])
  5939. || VK_SUCCESS != vkCreateSemaphore(device, &sci, allocatorCb, &m_renderDoneSemaphore[ii]) )
  5940. {
  5941. BX_TRACE("Create swapchain error: vkCreateSemaphore failed %d: %s.", result, getName(result) );
  5942. return result;
  5943. }
  5944. }
  5945. m_backBufferColorIdx = 0;
  5946. m_currentSemaphore = 0;
  5947. m_needPresent = false;
  5948. m_needToRefreshSwapchain = false;
  5949. return result;
  5950. }
  5951. void SwapChainVK::releaseSwapChain()
  5952. {
  5953. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  5954. {
  5955. release(m_backBufferColorImageView[ii]);
  5956. m_backBufferFence[ii] = VK_NULL_HANDLE;
  5957. release(m_presentDoneSemaphore[ii]);
  5958. release(m_renderDoneSemaphore[ii]);
  5959. }
  5960. release(m_swapchain);
  5961. }
  5962. VkResult SwapChainVK::createAttachments(VkCommandBuffer _commandBuffer)
  5963. {
  5964. VkResult result = VK_SUCCESS;
  5965. const uint32_t samplerIndex = (m_resolution.reset & BGFX_RESET_MSAA_MASK) >> BGFX_RESET_MSAA_SHIFT;
  5966. const uint64_t textureFlags = (uint64_t(samplerIndex + 1) << BGFX_TEXTURE_RT_MSAA_SHIFT) | BGFX_TEXTURE_RT | BGFX_TEXTURE_RT_WRITE_ONLY;
  5967. m_sampler = s_msaa[samplerIndex];
  5968. const uint16_t requiredCaps = m_sampler.Count > 1
  5969. ? BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER_MSAA
  5970. : BGFX_CAPS_FORMAT_TEXTURE_FRAMEBUFFER
  5971. ;
  5972. // the spec guarantees that at least one of D24S8 and D32FS8 is supported
  5973. VkFormat depthFormat = VK_FORMAT_D32_SFLOAT_S8_UINT;
  5974. if (g_caps.formats[m_depthFormat] & requiredCaps)
  5975. {
  5976. depthFormat = s_textureFormat[m_depthFormat].m_fmtDsv;
  5977. }
  5978. else if (g_caps.formats[TextureFormat::D24S8] & requiredCaps)
  5979. {
  5980. depthFormat = s_textureFormat[TextureFormat::D24S8].m_fmtDsv;
  5981. }
  5982. result = m_backBufferDepthStencil.create(
  5983. _commandBuffer
  5984. , m_sci.imageExtent.width
  5985. , m_sci.imageExtent.height
  5986. , textureFlags
  5987. , depthFormat
  5988. );
  5989. if (VK_SUCCESS != result)
  5990. {
  5991. BX_TRACE("Create swapchain error: creating depth stencil image failed %d: %s.", result, getName(result) );
  5992. return result;
  5993. }
  5994. result = m_backBufferDepthStencil.createView(0, 1, 0, 1, VK_IMAGE_VIEW_TYPE_2D, m_backBufferDepthStencil.m_aspectMask, true, &m_backBufferDepthStencilImageView);
  5995. if (VK_SUCCESS != result)
  5996. {
  5997. BX_TRACE("Create swapchain error: creating depth stencil image view failed %d: %s.", result, getName(result) );
  5998. return result;
  5999. }
  6000. if (m_sampler.Count > 1)
  6001. {
  6002. result = m_backBufferColorMsaa.create(
  6003. _commandBuffer
  6004. , m_sci.imageExtent.width
  6005. , m_sci.imageExtent.height
  6006. , textureFlags
  6007. , m_sci.imageFormat
  6008. );
  6009. if (VK_SUCCESS != result)
  6010. {
  6011. BX_TRACE("Create swapchain error: creating MSAA color image failed %d: %s.", result, getName(result) );
  6012. return result;
  6013. }
  6014. result = m_backBufferColorMsaa.createView(0, 1, 0, 1, VK_IMAGE_VIEW_TYPE_2D, m_backBufferColorMsaa.m_aspectMask, true, &m_backBufferColorMsaaImageView);
  6015. if (VK_SUCCESS != result)
  6016. {
  6017. BX_TRACE("Create swapchain error: creating MSAA color image view failed %d: %s.", result, getName(result) );
  6018. return result;
  6019. }
  6020. }
  6021. return result;
  6022. }
  6023. void SwapChainVK::releaseAttachments()
  6024. {
  6025. release(m_backBufferDepthStencilImageView);
  6026. release(m_backBufferColorMsaaImageView);
  6027. m_backBufferDepthStencil.destroy();
  6028. m_backBufferColorMsaa.destroy();
  6029. }
  6030. VkResult SwapChainVK::createFrameBuffer()
  6031. {
  6032. VkResult result = VK_SUCCESS;
  6033. const VkDevice device = s_renderVK->m_device;
  6034. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  6035. VkRenderPass renderPass;
  6036. result = s_renderVK->getRenderPass(*this, &renderPass);
  6037. if (VK_SUCCESS != result)
  6038. {
  6039. return result;
  6040. }
  6041. for (uint32_t ii = 0; ii < m_numSwapchainImages; ++ii)
  6042. {
  6043. uint32_t numAttachments = 2;
  6044. ::VkImageView attachments[3] =
  6045. {
  6046. m_sampler.Count > 1
  6047. ? m_backBufferColorMsaaImageView
  6048. : m_backBufferColorImageView[ii],
  6049. m_backBufferDepthStencilImageView,
  6050. };
  6051. if (m_sampler.Count > 1 && !m_supportsManualResolve)
  6052. {
  6053. attachments[numAttachments++] = m_backBufferColorImageView[ii];
  6054. }
  6055. VkFramebufferCreateInfo fci;
  6056. fci.sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO;
  6057. fci.pNext = NULL;
  6058. fci.flags = 0;
  6059. fci.renderPass = renderPass;
  6060. fci.attachmentCount = numAttachments;
  6061. fci.pAttachments = attachments;
  6062. fci.width = m_sci.imageExtent.width;
  6063. fci.height = m_sci.imageExtent.height;
  6064. fci.layers = 1;
  6065. result = vkCreateFramebuffer(device, &fci, allocatorCb, &m_backBufferFrameBuffer[ii]);
  6066. if (VK_SUCCESS != result)
  6067. {
  6068. return result;
  6069. }
  6070. }
  6071. return result;
  6072. }
  6073. void SwapChainVK::releaseFrameBuffer()
  6074. {
  6075. for (uint32_t ii = 0; ii < BX_COUNTOF(m_backBufferColorImageView); ++ii)
  6076. {
  6077. release(m_backBufferFrameBuffer[ii]);
  6078. }
  6079. }
  6080. uint32_t SwapChainVK::findPresentMode(bool _vsync)
  6081. {
  6082. VkResult result = VK_SUCCESS;
  6083. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  6084. uint32_t numPresentModes;
  6085. result = vkGetPhysicalDeviceSurfacePresentModesKHR(
  6086. physicalDevice
  6087. , m_surface
  6088. , &numPresentModes
  6089. , NULL
  6090. );
  6091. if (VK_SUCCESS != result)
  6092. {
  6093. BX_TRACE("findPresentMode error: vkGetPhysicalDeviceSurfacePresentModesKHR failed %d: %s.", result, getName(result) );
  6094. return UINT32_MAX;
  6095. }
  6096. VkPresentModeKHR presentModes[16];
  6097. numPresentModes = bx::min<uint32_t>(numPresentModes, BX_COUNTOF(presentModes) );
  6098. result = vkGetPhysicalDeviceSurfacePresentModesKHR(
  6099. physicalDevice
  6100. , m_surface
  6101. , &numPresentModes
  6102. , presentModes
  6103. );
  6104. if (VK_SUCCESS != result)
  6105. {
  6106. BX_TRACE("findPresentMode error: vkGetPhysicalDeviceSurfacePresentModesKHR failed %d: %s.", result, getName(result) );
  6107. return UINT32_MAX;
  6108. }
  6109. uint32_t idx = UINT32_MAX;
  6110. for (uint32_t ii = 0; ii < BX_COUNTOF(s_presentMode) && UINT32_MAX == idx; ++ii)
  6111. {
  6112. for (uint32_t jj = 0; jj < numPresentModes; ++jj)
  6113. {
  6114. const PresentMode& pm = s_presentMode[ii];
  6115. if (pm.mode == presentModes[jj]
  6116. && pm.vsync == _vsync)
  6117. {
  6118. idx = ii;
  6119. break;
  6120. }
  6121. }
  6122. }
  6123. if (UINT32_MAX == idx)
  6124. {
  6125. idx = 0;
  6126. BX_TRACE("Present mode not found! Defaulting to %s.", s_presentMode[idx].name);
  6127. }
  6128. return idx;
  6129. }
  6130. TextureFormat::Enum SwapChainVK::findSurfaceFormat(TextureFormat::Enum _format, VkColorSpaceKHR _colorSpace, bool _srgb)
  6131. {
  6132. VkResult result = VK_SUCCESS;
  6133. TextureFormat::Enum selectedFormat = TextureFormat::Count;
  6134. const VkPhysicalDevice physicalDevice = s_renderVK->m_physicalDevice;
  6135. uint32_t numSurfaceFormats;
  6136. result = vkGetPhysicalDeviceSurfaceFormatsKHR(physicalDevice, m_surface, &numSurfaceFormats, NULL);
  6137. if (VK_SUCCESS != result)
  6138. {
  6139. BX_TRACE("findSurfaceFormat error: vkGetPhysicalDeviceSurfaceFormatsKHR failed %d: %s.", result, getName(result) );
  6140. return selectedFormat;
  6141. }
  6142. VkSurfaceFormatKHR* surfaceFormats = (VkSurfaceFormatKHR*)bx::alloc(g_allocator, numSurfaceFormats * sizeof(VkSurfaceFormatKHR) );
  6143. result = vkGetPhysicalDeviceSurfaceFormatsKHR(physicalDevice, m_surface, &numSurfaceFormats, surfaceFormats);
  6144. if (VK_SUCCESS != result)
  6145. {
  6146. BX_TRACE("findSurfaceFormat error: vkGetPhysicalDeviceSurfaceFormatsKHR failed %d: %s.", result, getName(result) );
  6147. bx::free(g_allocator, surfaceFormats);
  6148. return selectedFormat;
  6149. }
  6150. const TextureFormat::Enum requestedFormats[] =
  6151. {
  6152. _format,
  6153. TextureFormat::BGRA8,
  6154. TextureFormat::RGBA8,
  6155. };
  6156. for (uint32_t ii = 0; ii < BX_COUNTOF(requestedFormats) && TextureFormat::Count == selectedFormat; ii++)
  6157. {
  6158. const TextureFormat::Enum requested = requestedFormats[ii];
  6159. const VkFormat requestedVkFormat = _srgb
  6160. ? s_textureFormat[requested].m_fmtSrgb
  6161. : s_textureFormat[requested].m_fmt
  6162. ;
  6163. for (uint32_t jj = 0; jj < numSurfaceFormats; jj++)
  6164. {
  6165. if (_colorSpace == surfaceFormats[jj].colorSpace
  6166. && requestedVkFormat == surfaceFormats[jj].format)
  6167. {
  6168. selectedFormat = requested;
  6169. if (0 != ii
  6170. && s_renderVK->m_swapchainFormats[_format] != selectedFormat)
  6171. {
  6172. s_renderVK->m_swapchainFormats[_format] = selectedFormat;
  6173. BX_TRACE(
  6174. "findSurfaceFormat: Surface format %s not found! Defaulting to %s."
  6175. , bimg::getName(bimg::TextureFormat::Enum(_format) )
  6176. , bimg::getName(bimg::TextureFormat::Enum(selectedFormat) )
  6177. );
  6178. }
  6179. break;
  6180. }
  6181. }
  6182. }
  6183. bx::free(g_allocator, surfaceFormats);
  6184. if (TextureFormat::Count == selectedFormat)
  6185. {
  6186. BX_TRACE("findSurfaceFormat error: No supported surface format found.");
  6187. }
  6188. return selectedFormat;
  6189. }
  6190. bool SwapChainVK::acquire(VkCommandBuffer _commandBuffer)
  6191. {
  6192. if (VK_NULL_HANDLE == m_swapchain
  6193. || m_needToRefreshSwapchain)
  6194. {
  6195. return false;
  6196. }
  6197. if (!m_needPresent)
  6198. {
  6199. const VkDevice device = s_renderVK->m_device;
  6200. m_lastImageAcquiredSemaphore = m_presentDoneSemaphore[m_currentSemaphore];
  6201. m_lastImageRenderedSemaphore = m_renderDoneSemaphore[m_currentSemaphore];
  6202. m_currentSemaphore = (m_currentSemaphore + 1) % m_numSwapchainImages;
  6203. VkResult result = vkAcquireNextImageKHR(
  6204. device
  6205. , m_swapchain
  6206. , UINT64_MAX
  6207. , m_lastImageAcquiredSemaphore
  6208. , VK_NULL_HANDLE
  6209. , &m_backBufferColorIdx
  6210. );
  6211. switch (result)
  6212. {
  6213. case VK_SUCCESS:
  6214. break;
  6215. case VK_ERROR_SURFACE_LOST_KHR:
  6216. m_needToRecreateSurface = true;
  6217. BX_FALLTHROUGH;
  6218. case VK_ERROR_OUT_OF_DATE_KHR:
  6219. case VK_SUBOPTIMAL_KHR:
  6220. m_needToRefreshSwapchain = true;
  6221. return false;
  6222. default:
  6223. BX_ASSERT(VK_SUCCESS == result, "vkAcquireNextImageKHR(...); VK error 0x%x: %s", result, getName(result) );
  6224. return false;
  6225. }
  6226. if (VK_NULL_HANDLE != m_backBufferFence[m_backBufferColorIdx])
  6227. {
  6228. VK_CHECK(vkWaitForFences(
  6229. device
  6230. , 1
  6231. , &m_backBufferFence[m_backBufferColorIdx]
  6232. , VK_TRUE
  6233. , UINT64_MAX
  6234. ) );
  6235. }
  6236. transitionImage(_commandBuffer);
  6237. m_needPresent = true;
  6238. }
  6239. return true;
  6240. }
  6241. void SwapChainVK::present()
  6242. {
  6243. if (VK_NULL_HANDLE != m_swapchain
  6244. && m_needPresent)
  6245. {
  6246. VkPresentInfoKHR pi;
  6247. pi.sType = VK_STRUCTURE_TYPE_PRESENT_INFO_KHR;
  6248. pi.pNext = NULL;
  6249. pi.waitSemaphoreCount = 1;
  6250. pi.pWaitSemaphores = &m_lastImageRenderedSemaphore;
  6251. pi.swapchainCount = 1;
  6252. pi.pSwapchains = &m_swapchain;
  6253. pi.pImageIndices = &m_backBufferColorIdx;
  6254. pi.pResults = NULL;
  6255. VkResult result = vkQueuePresentKHR(m_queue, &pi);
  6256. switch (result)
  6257. {
  6258. case VK_ERROR_SURFACE_LOST_KHR:
  6259. m_needToRecreateSurface = true;
  6260. BX_FALLTHROUGH;
  6261. case VK_ERROR_OUT_OF_DATE_KHR:
  6262. case VK_SUBOPTIMAL_KHR:
  6263. m_needToRefreshSwapchain = true;
  6264. break;
  6265. default:
  6266. BX_ASSERT(VK_SUCCESS == result, "vkQueuePresentKHR(...); VK error 0x%x: %s", result, getName(result) );
  6267. break;
  6268. }
  6269. m_needPresent = false;
  6270. m_lastImageRenderedSemaphore = VK_NULL_HANDLE;
  6271. }
  6272. }
  6273. void SwapChainVK::transitionImage(VkCommandBuffer _commandBuffer)
  6274. {
  6275. VkImageLayout& layout = m_backBufferColorImageLayout[m_backBufferColorIdx];
  6276. const bool toPresent = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL == layout;
  6277. const VkImageLayout newLayout = toPresent
  6278. ? VK_IMAGE_LAYOUT_PRESENT_SRC_KHR
  6279. : VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL
  6280. ;
  6281. layout = toPresent ? layout : VK_IMAGE_LAYOUT_UNDEFINED;
  6282. setImageMemoryBarrier(
  6283. _commandBuffer
  6284. , m_backBufferColorImage[m_backBufferColorIdx]
  6285. , VK_IMAGE_ASPECT_COLOR_BIT
  6286. , layout
  6287. , newLayout
  6288. );
  6289. layout = newLayout;
  6290. }
  6291. void FrameBufferVK::create(uint8_t _num, const Attachment* _attachment)
  6292. {
  6293. m_numTh = _num;
  6294. bx::memCopy(m_attachment, _attachment, sizeof(Attachment) * _num);
  6295. postReset();
  6296. }
  6297. VkResult FrameBufferVK::create(uint16_t _denseIdx, void* _nwh, uint32_t _width, uint32_t _height, TextureFormat::Enum _format, TextureFormat::Enum _depthFormat)
  6298. {
  6299. VkResult result = VK_SUCCESS;
  6300. Resolution resolution = s_renderVK->m_resolution;
  6301. resolution.format = TextureFormat::Count == _format ? resolution.format : _format;
  6302. resolution.width = _width;
  6303. resolution.height = _height;
  6304. if (_denseIdx != UINT16_MAX)
  6305. {
  6306. resolution.reset &= ~BGFX_RESET_MSAA_MASK;
  6307. }
  6308. result = m_swapChain.create(s_renderVK->m_commandBuffer, _nwh, resolution, _depthFormat);
  6309. if (VK_SUCCESS != result)
  6310. {
  6311. return result;
  6312. }
  6313. result = s_renderVK->getRenderPass(m_swapChain, &m_renderPass);
  6314. if (VK_SUCCESS != result)
  6315. {
  6316. return result;
  6317. }
  6318. m_denseIdx = _denseIdx;
  6319. m_nwh = _nwh;
  6320. m_width = _width;
  6321. m_height = _height;
  6322. m_sampler = m_swapChain.m_sampler;
  6323. return result;
  6324. }
  6325. void FrameBufferVK::preReset()
  6326. {
  6327. if (VK_NULL_HANDLE != m_framebuffer)
  6328. {
  6329. s_renderVK->release(m_framebuffer);
  6330. for (uint8_t ii = 0; ii < m_numTh; ++ii)
  6331. {
  6332. s_renderVK->release(m_textureImageViews[ii]);
  6333. }
  6334. }
  6335. }
  6336. void FrameBufferVK::postReset()
  6337. {
  6338. if (m_numTh > 0)
  6339. {
  6340. const VkDevice device = s_renderVK->m_device;
  6341. const VkAllocationCallbacks* allocatorCb = s_renderVK->m_allocatorCb;
  6342. VK_CHECK(s_renderVK->getRenderPass(m_numTh, m_attachment, &m_renderPass) );
  6343. m_depth = BGFX_INVALID_HANDLE;
  6344. m_num = 0;
  6345. for (uint8_t ii = 0; ii < m_numTh; ++ii)
  6346. {
  6347. const Attachment& at = m_attachment[ii];
  6348. const TextureVK& texture = s_renderVK->m_textures[at.handle.idx];
  6349. VK_CHECK(texture.createView(
  6350. at.layer
  6351. , at.numLayers
  6352. , at.mip
  6353. , 1
  6354. , at.numLayers > 1 ? VK_IMAGE_VIEW_TYPE_2D_ARRAY : VK_IMAGE_VIEW_TYPE_2D
  6355. , texture.m_aspectMask
  6356. , true
  6357. , &m_textureImageViews[ii]
  6358. ) );
  6359. if (texture.m_aspectMask & VK_IMAGE_ASPECT_COLOR_BIT)
  6360. {
  6361. m_texture[m_num] = at.handle;
  6362. m_num++;
  6363. }
  6364. else if (texture.m_aspectMask & (VK_IMAGE_ASPECT_DEPTH_BIT | VK_IMAGE_ASPECT_STENCIL_BIT) )
  6365. {
  6366. m_depth = at.handle;
  6367. }
  6368. }
  6369. const TextureVK& firstTexture = s_renderVK->m_textures[m_attachment[0].handle.idx];
  6370. m_width = bx::uint32_max(firstTexture.m_width >> m_attachment[0].mip, 1);
  6371. m_height = bx::uint32_max(firstTexture.m_height >> m_attachment[0].mip, 1);
  6372. m_sampler = firstTexture.m_sampler;
  6373. VkFramebufferCreateInfo fci;
  6374. fci.sType = VK_STRUCTURE_TYPE_FRAMEBUFFER_CREATE_INFO;
  6375. fci.pNext = NULL;
  6376. fci.flags = 0;
  6377. fci.renderPass = m_renderPass;
  6378. fci.attachmentCount = m_numTh;
  6379. fci.pAttachments = &m_textureImageViews[0];
  6380. fci.width = m_width;
  6381. fci.height = m_height;
  6382. fci.layers = m_attachment[0].numLayers;
  6383. VK_CHECK(vkCreateFramebuffer(device, &fci, allocatorCb, &m_framebuffer) );
  6384. m_currentFramebuffer = m_framebuffer;
  6385. }
  6386. }
  6387. void FrameBufferVK::update(VkCommandBuffer _commandBuffer, const Resolution& _resolution)
  6388. {
  6389. m_swapChain.update(_commandBuffer, m_nwh, _resolution);
  6390. VK_CHECK(s_renderVK->getRenderPass(m_swapChain, &m_renderPass) );
  6391. m_width = _resolution.width;
  6392. m_height = _resolution.height;
  6393. m_sampler = m_swapChain.m_sampler;
  6394. }
  6395. void FrameBufferVK::resolve()
  6396. {
  6397. if (!m_needResolve)
  6398. {
  6399. return;
  6400. }
  6401. if (NULL == m_nwh)
  6402. {
  6403. for (uint32_t ii = 0; ii < m_numTh; ++ii)
  6404. {
  6405. const Attachment& at = m_attachment[ii];
  6406. if (isValid(at.handle) )
  6407. {
  6408. TextureVK& texture = s_renderVK->m_textures[at.handle.idx];
  6409. texture.resolve(s_renderVK->m_commandBuffer, at.resolve, at.layer, at.numLayers, at.mip);
  6410. }
  6411. }
  6412. }
  6413. else if (isRenderable()
  6414. && m_sampler.Count > 1
  6415. && m_swapChain.m_supportsManualResolve)
  6416. {
  6417. m_swapChain.m_backBufferColorMsaa.m_singleMsaaImage = m_swapChain.m_backBufferColorImage[m_swapChain.m_backBufferColorIdx];
  6418. m_swapChain.m_backBufferColorMsaa.m_currentSingleMsaaImageLayout = m_swapChain.m_backBufferColorImageLayout[m_swapChain.m_backBufferColorIdx];
  6419. m_swapChain.m_backBufferColorMsaa.resolve(s_renderVK->m_commandBuffer, 0, 0, 1, 0);
  6420. m_swapChain.m_backBufferColorMsaa.m_singleMsaaImage = VK_NULL_HANDLE;
  6421. m_swapChain.m_backBufferColorMsaa.m_currentSingleMsaaImageLayout = VK_IMAGE_LAYOUT_UNDEFINED;
  6422. }
  6423. m_needResolve = false;
  6424. }
  6425. uint16_t FrameBufferVK::destroy()
  6426. {
  6427. preReset();
  6428. if (NULL != m_nwh)
  6429. {
  6430. m_swapChain.destroy();
  6431. m_nwh = NULL;
  6432. m_needPresent = false;
  6433. }
  6434. m_numTh = 0;
  6435. m_num = 0;
  6436. m_depth = BGFX_INVALID_HANDLE;
  6437. m_needResolve = false;
  6438. uint16_t denseIdx = m_denseIdx;
  6439. m_denseIdx = UINT16_MAX;
  6440. return denseIdx;
  6441. }
  6442. bool FrameBufferVK::acquire(VkCommandBuffer _commandBuffer)
  6443. {
  6444. bool acquired = true;
  6445. if (NULL != m_nwh)
  6446. {
  6447. acquired = m_swapChain.acquire(_commandBuffer);
  6448. m_needPresent = m_swapChain.m_needPresent;
  6449. m_currentFramebuffer = m_swapChain.m_backBufferFrameBuffer[m_swapChain.m_backBufferColorIdx];
  6450. }
  6451. m_needResolve = true;
  6452. return acquired;
  6453. }
  6454. void FrameBufferVK::present()
  6455. {
  6456. m_swapChain.present();
  6457. m_needPresent = false;
  6458. }
  6459. bool FrameBufferVK::isRenderable() const
  6460. {
  6461. return false
  6462. || (NULL == m_nwh)
  6463. || m_swapChain.m_needPresent
  6464. ;
  6465. }
  6466. VkResult CommandQueueVK::init(uint32_t _queueFamily, VkQueue _queue, uint32_t _numFramesInFlight)
  6467. {
  6468. m_queueFamily = _queueFamily;
  6469. m_queue = _queue;
  6470. m_numFramesInFlight = bx::clamp<uint32_t>(_numFramesInFlight, 1, BGFX_CONFIG_MAX_FRAME_LATENCY);
  6471. m_activeCommandBuffer = VK_NULL_HANDLE;
  6472. return reset();
  6473. }
  6474. VkResult CommandQueueVK::reset()
  6475. {
  6476. shutdown();
  6477. m_currentFrameInFlight = 0;
  6478. m_consumeIndex = 0;
  6479. m_numSignalSemaphores = 0;
  6480. m_numWaitSemaphores = 0;
  6481. m_activeCommandBuffer = VK_NULL_HANDLE;
  6482. m_currentFence = VK_NULL_HANDLE;
  6483. m_completedFence = VK_NULL_HANDLE;
  6484. m_submitted = 0;
  6485. VkCommandPoolCreateInfo cpci;
  6486. cpci.sType = VK_STRUCTURE_TYPE_COMMAND_POOL_CREATE_INFO;
  6487. cpci.pNext = NULL;
  6488. cpci.flags = VK_COMMAND_POOL_CREATE_TRANSIENT_BIT;
  6489. cpci.queueFamilyIndex = m_queueFamily;
  6490. VkCommandBufferAllocateInfo cbai;
  6491. cbai.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO;
  6492. cbai.pNext = NULL;
  6493. cbai.level = VK_COMMAND_BUFFER_LEVEL_PRIMARY;
  6494. cbai.commandBufferCount = 1;
  6495. VkFenceCreateInfo fci;
  6496. fci.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
  6497. fci.pNext = NULL;
  6498. fci.flags = VK_FENCE_CREATE_SIGNALED_BIT;
  6499. VkResult result = VK_SUCCESS;
  6500. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6501. {
  6502. result = vkCreateCommandPool(
  6503. s_renderVK->m_device
  6504. , &cpci
  6505. , s_renderVK->m_allocatorCb
  6506. , &m_commandList[ii].m_commandPool
  6507. );
  6508. if (VK_SUCCESS != result)
  6509. {
  6510. BX_TRACE("Create command queue error: vkCreateCommandPool failed %d: %s.", result, getName(result) );
  6511. return result;
  6512. }
  6513. cbai.commandPool = m_commandList[ii].m_commandPool;
  6514. result = vkAllocateCommandBuffers(
  6515. s_renderVK->m_device
  6516. , &cbai
  6517. , &m_commandList[ii].m_commandBuffer
  6518. );
  6519. if (VK_SUCCESS != result)
  6520. {
  6521. BX_TRACE("Create command queue error: vkAllocateCommandBuffers failed %d: %s.", result, getName(result) );
  6522. return result;
  6523. }
  6524. result = vkCreateFence(
  6525. s_renderVK->m_device
  6526. , &fci
  6527. , s_renderVK->m_allocatorCb
  6528. , &m_commandList[ii].m_fence
  6529. );
  6530. if (VK_SUCCESS != result)
  6531. {
  6532. BX_TRACE("Create command queue error: vkCreateFence failed %d: %s.", result, getName(result) );
  6533. return result;
  6534. }
  6535. }
  6536. return result;
  6537. }
  6538. void CommandQueueVK::shutdown()
  6539. {
  6540. kick(true);
  6541. finish(true);
  6542. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6543. {
  6544. vkDestroy(m_commandList[ii].m_fence);
  6545. m_commandList[ii].m_commandBuffer = VK_NULL_HANDLE;
  6546. vkDestroy(m_commandList[ii].m_commandPool);
  6547. }
  6548. }
  6549. VkResult CommandQueueVK::alloc(VkCommandBuffer* _commandBuffer)
  6550. {
  6551. VkResult result = VK_SUCCESS;
  6552. if (m_activeCommandBuffer == VK_NULL_HANDLE)
  6553. {
  6554. const VkDevice device = s_renderVK->m_device;
  6555. CommandList& commandList = m_commandList[m_currentFrameInFlight];
  6556. result = vkWaitForFences(device, 1, &commandList.m_fence, VK_TRUE, UINT64_MAX);
  6557. if (VK_SUCCESS != result)
  6558. {
  6559. BX_TRACE("Allocate command buffer error: vkWaitForFences failed %d: %s.", result, getName(result) );
  6560. return result;
  6561. }
  6562. result = vkResetCommandPool(device, commandList.m_commandPool, 0);
  6563. if (VK_SUCCESS != result)
  6564. {
  6565. BX_TRACE("Allocate command buffer error: vkResetCommandPool failed %d: %s.", result, getName(result) );
  6566. return result;
  6567. }
  6568. VkCommandBufferBeginInfo cbi;
  6569. cbi.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO;
  6570. cbi.pNext = NULL;
  6571. cbi.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT;
  6572. cbi.pInheritanceInfo = NULL;
  6573. result = vkBeginCommandBuffer(commandList.m_commandBuffer, &cbi);
  6574. if (VK_SUCCESS != result)
  6575. {
  6576. BX_TRACE("Allocate command buffer error: vkBeginCommandBuffer failed %d: %s.", result, getName(result) );
  6577. return result;
  6578. }
  6579. m_activeCommandBuffer = commandList.m_commandBuffer;
  6580. m_currentFence = commandList.m_fence;
  6581. }
  6582. if (NULL != _commandBuffer)
  6583. {
  6584. *_commandBuffer = m_activeCommandBuffer;
  6585. }
  6586. return result;
  6587. }
  6588. void CommandQueueVK::addWaitSemaphore(VkSemaphore _semaphore, VkPipelineStageFlags _waitFlags)
  6589. {
  6590. BX_ASSERT(m_numWaitSemaphores < BX_COUNTOF(m_waitSemaphores), "Too many wait semaphores.");
  6591. m_waitSemaphores[m_numWaitSemaphores] = _semaphore;
  6592. m_waitSemaphoreStages[m_numWaitSemaphores] = _waitFlags;
  6593. m_numWaitSemaphores++;
  6594. }
  6595. void CommandQueueVK::addSignalSemaphore(VkSemaphore _semaphore)
  6596. {
  6597. BX_ASSERT(m_numSignalSemaphores < BX_COUNTOF(m_signalSemaphores), "Too many signal semaphores.");
  6598. m_signalSemaphores[m_numSignalSemaphores] = _semaphore;
  6599. m_numSignalSemaphores++;
  6600. }
  6601. void CommandQueueVK::kick(bool _wait)
  6602. {
  6603. if (VK_NULL_HANDLE != m_activeCommandBuffer)
  6604. {
  6605. const VkDevice device = s_renderVK->m_device;
  6606. setMemoryBarrier(
  6607. m_activeCommandBuffer
  6608. , VK_PIPELINE_STAGE_ALL_COMMANDS_BIT
  6609. , VK_PIPELINE_STAGE_ALL_COMMANDS_BIT
  6610. );
  6611. VK_CHECK(vkEndCommandBuffer(m_activeCommandBuffer) );
  6612. m_completedFence = m_currentFence;
  6613. m_currentFence = VK_NULL_HANDLE;
  6614. VK_CHECK(vkResetFences(device, 1, &m_completedFence) );
  6615. VkSubmitInfo si;
  6616. si.sType = VK_STRUCTURE_TYPE_SUBMIT_INFO;
  6617. si.pNext = NULL;
  6618. si.waitSemaphoreCount = m_numWaitSemaphores;
  6619. si.pWaitSemaphores = &m_waitSemaphores[0];
  6620. si.pWaitDstStageMask = m_waitSemaphoreStages;
  6621. si.commandBufferCount = 1;
  6622. si.pCommandBuffers = &m_activeCommandBuffer;
  6623. si.signalSemaphoreCount = m_numSignalSemaphores;
  6624. si.pSignalSemaphores = &m_signalSemaphores[0];
  6625. m_numWaitSemaphores = 0;
  6626. m_numSignalSemaphores = 0;
  6627. VK_CHECK(vkQueueSubmit(m_queue, 1, &si, m_completedFence) );
  6628. if (_wait)
  6629. {
  6630. VK_CHECK(vkWaitForFences(device, 1, &m_completedFence, VK_TRUE, UINT64_MAX) );
  6631. }
  6632. m_activeCommandBuffer = VK_NULL_HANDLE;
  6633. m_currentFrameInFlight = (m_currentFrameInFlight + 1) % m_numFramesInFlight;
  6634. m_submitted++;
  6635. }
  6636. }
  6637. void CommandQueueVK::finish(bool _finishAll)
  6638. {
  6639. if (_finishAll)
  6640. {
  6641. for (uint32_t ii = 0; ii < m_numFramesInFlight; ++ii)
  6642. {
  6643. consume();
  6644. }
  6645. m_consumeIndex = m_currentFrameInFlight;
  6646. }
  6647. else
  6648. {
  6649. consume();
  6650. }
  6651. }
  6652. void CommandQueueVK::release(uint64_t _handle, VkObjectType _type)
  6653. {
  6654. Resource resource;
  6655. resource.m_type = _type;
  6656. resource.m_handle = _handle;
  6657. m_release[m_currentFrameInFlight].push_back(resource);
  6658. }
  6659. void CommandQueueVK::consume()
  6660. {
  6661. m_consumeIndex = (m_consumeIndex + 1) % m_numFramesInFlight;
  6662. for (const Resource& resource : m_release[m_consumeIndex])
  6663. {
  6664. switch (resource.m_type)
  6665. {
  6666. case VK_OBJECT_TYPE_BUFFER: destroy<VkBuffer >(resource.m_handle); break;
  6667. case VK_OBJECT_TYPE_IMAGE_VIEW: destroy<VkImageView >(resource.m_handle); break;
  6668. case VK_OBJECT_TYPE_IMAGE: destroy<VkImage >(resource.m_handle); break;
  6669. case VK_OBJECT_TYPE_FRAMEBUFFER: destroy<VkFramebuffer >(resource.m_handle); break;
  6670. case VK_OBJECT_TYPE_PIPELINE_LAYOUT: destroy<VkPipelineLayout >(resource.m_handle); break;
  6671. case VK_OBJECT_TYPE_PIPELINE: destroy<VkPipeline >(resource.m_handle); break;
  6672. case VK_OBJECT_TYPE_DESCRIPTOR_SET: destroy<VkDescriptorSet >(resource.m_handle); break;
  6673. case VK_OBJECT_TYPE_DESCRIPTOR_SET_LAYOUT: destroy<VkDescriptorSetLayout>(resource.m_handle); break;
  6674. case VK_OBJECT_TYPE_RENDER_PASS: destroy<VkRenderPass >(resource.m_handle); break;
  6675. case VK_OBJECT_TYPE_SAMPLER: destroy<VkSampler >(resource.m_handle); break;
  6676. case VK_OBJECT_TYPE_SEMAPHORE: destroy<VkSemaphore >(resource.m_handle); break;
  6677. case VK_OBJECT_TYPE_SURFACE_KHR: destroy<VkSurfaceKHR >(resource.m_handle); break;
  6678. case VK_OBJECT_TYPE_SWAPCHAIN_KHR: destroy<VkSwapchainKHR >(resource.m_handle); break;
  6679. case VK_OBJECT_TYPE_DEVICE_MEMORY: destroy<VkDeviceMemory >(resource.m_handle); break;
  6680. default:
  6681. BX_ASSERT(false, "Invalid resource type: %d", resource.m_type);
  6682. break;
  6683. }
  6684. }
  6685. m_release[m_consumeIndex].clear();
  6686. }
  6687. void RendererContextVK::submitBlit(BlitState& _bs, uint16_t _view)
  6688. {
  6689. VkImageLayout srcLayouts[BGFX_CONFIG_MAX_BLIT_ITEMS];
  6690. VkImageLayout dstLayouts[BGFX_CONFIG_MAX_BLIT_ITEMS];
  6691. BlitState bs0 = _bs;
  6692. while (bs0.hasItem(_view) )
  6693. {
  6694. uint16_t item = bs0.m_item;
  6695. const BlitItem& blit = bs0.advance();
  6696. TextureVK& src = m_textures[blit.m_src.idx];
  6697. TextureVK& dst = m_textures[blit.m_dst.idx];
  6698. srcLayouts[item] = VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_currentSingleMsaaImageLayout : src.m_currentImageLayout;
  6699. dstLayouts[item] = dst.m_currentImageLayout;
  6700. }
  6701. bs0 = _bs;
  6702. while (bs0.hasItem(_view) )
  6703. {
  6704. const BlitItem& blit = bs0.advance();
  6705. TextureVK& src = m_textures[blit.m_src.idx];
  6706. TextureVK& dst = m_textures[blit.m_dst.idx];
  6707. src.setImageMemoryBarrier(
  6708. m_commandBuffer
  6709. , blit.m_src.idx == blit.m_dst.idx ? VK_IMAGE_LAYOUT_GENERAL : VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL
  6710. , VK_NULL_HANDLE != src.m_singleMsaaImage
  6711. );
  6712. if (blit.m_src.idx != blit.m_dst.idx)
  6713. {
  6714. dst.setImageMemoryBarrier(m_commandBuffer, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
  6715. }
  6716. const uint16_t srcSamples = VK_NULL_HANDLE != src.m_singleMsaaImage ? 1 : src.m_sampler.Count;
  6717. const uint16_t dstSamples = dst.m_sampler.Count;
  6718. BX_UNUSED(srcSamples, dstSamples);
  6719. BX_ASSERT(
  6720. srcSamples == dstSamples
  6721. , "Mismatching texture sample count (%d != %d)."
  6722. , srcSamples
  6723. , dstSamples
  6724. );
  6725. VkImageCopy copyInfo;
  6726. copyInfo.srcSubresource.aspectMask = src.m_aspectMask;
  6727. copyInfo.srcSubresource.mipLevel = blit.m_srcMip;
  6728. copyInfo.srcSubresource.baseArrayLayer = 0;
  6729. copyInfo.srcSubresource.layerCount = 1;
  6730. copyInfo.srcOffset.x = blit.m_srcX;
  6731. copyInfo.srcOffset.y = blit.m_srcY;
  6732. copyInfo.srcOffset.z = 0;
  6733. copyInfo.dstSubresource.aspectMask = dst.m_aspectMask;
  6734. copyInfo.dstSubresource.mipLevel = blit.m_dstMip;
  6735. copyInfo.dstSubresource.baseArrayLayer = 0;
  6736. copyInfo.dstSubresource.layerCount = 1;
  6737. copyInfo.dstOffset.x = blit.m_dstX;
  6738. copyInfo.dstOffset.y = blit.m_dstY;
  6739. copyInfo.dstOffset.z = 0;
  6740. copyInfo.extent.width = blit.m_width;
  6741. copyInfo.extent.height = blit.m_height;
  6742. copyInfo.extent.depth = 1;
  6743. const uint32_t depth = bx::max<uint32_t>(1, blit.m_depth);
  6744. if (VK_IMAGE_VIEW_TYPE_3D == src.m_type)
  6745. {
  6746. BX_ASSERT(VK_IMAGE_VIEW_TYPE_3D == dst.m_type, "Can't blit between 2D and 3D image.");
  6747. copyInfo.srcOffset.z = blit.m_srcZ;
  6748. copyInfo.dstOffset.z = blit.m_dstZ;
  6749. copyInfo.extent.depth = depth;
  6750. }
  6751. else
  6752. {
  6753. copyInfo.srcSubresource.baseArrayLayer = blit.m_srcZ;
  6754. copyInfo.dstSubresource.baseArrayLayer = blit.m_dstZ;
  6755. copyInfo.srcSubresource.layerCount = depth;
  6756. copyInfo.dstSubresource.layerCount = depth;
  6757. }
  6758. vkCmdCopyImage(
  6759. m_commandBuffer
  6760. , VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_singleMsaaImage : src.m_textureImage
  6761. , VK_NULL_HANDLE != src.m_singleMsaaImage ? src.m_currentSingleMsaaImageLayout : src.m_currentImageLayout
  6762. , dst.m_textureImage
  6763. , dst.m_currentImageLayout
  6764. , 1
  6765. , &copyInfo
  6766. );
  6767. setMemoryBarrier(
  6768. m_commandBuffer
  6769. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6770. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6771. );
  6772. }
  6773. while (_bs.hasItem(_view) )
  6774. {
  6775. uint16_t item = _bs.m_item;
  6776. const BlitItem& blit = _bs.advance();
  6777. TextureVK& src = m_textures[blit.m_src.idx];
  6778. TextureVK& dst = m_textures[blit.m_dst.idx];
  6779. src.setImageMemoryBarrier(m_commandBuffer, srcLayouts[item], VK_NULL_HANDLE != src.m_singleMsaaImage);
  6780. dst.setImageMemoryBarrier(m_commandBuffer, dstLayouts[item]);
  6781. }
  6782. }
  6783. void RendererContextVK::submit(Frame* _render, ClearQuad& _clearQuad, TextVideoMemBlitter& _textVideoMemBlitter)
  6784. {
  6785. BX_UNUSED(_clearQuad);
  6786. if (updateResolution(_render->m_resolution) )
  6787. {
  6788. return;
  6789. }
  6790. if (_render->m_capture)
  6791. {
  6792. renderDocTriggerCapture();
  6793. }
  6794. BGFX_VK_PROFILER_BEGIN_LITERAL("rendererSubmit", kColorView);
  6795. int64_t timeBegin = bx::getHPCounter();
  6796. int64_t captureElapsed = 0;
  6797. uint32_t frameQueryIdx = UINT32_MAX;
  6798. if (m_timerQuerySupport)
  6799. {
  6800. frameQueryIdx = m_gpuTimer.begin(BGFX_CONFIG_MAX_VIEWS, _render->m_frameNum);
  6801. }
  6802. if (0 < _render->m_iboffset)
  6803. {
  6804. BGFX_PROFILER_SCOPE("bgfx/Update transient index buffer", kColorResource);
  6805. TransientIndexBuffer* ib = _render->m_transientIb;
  6806. m_indexBuffers[ib->handle.idx].update(m_commandBuffer, 0, _render->m_iboffset, ib->data);
  6807. }
  6808. if (0 < _render->m_vboffset)
  6809. {
  6810. BGFX_PROFILER_SCOPE("bgfx/Update transient vertex buffer", kColorResource);
  6811. TransientVertexBuffer* vb = _render->m_transientVb;
  6812. m_vertexBuffers[vb->handle.idx].update(m_commandBuffer, 0, _render->m_vboffset, vb->data);
  6813. }
  6814. _render->sort();
  6815. RenderDraw currentState;
  6816. currentState.clear();
  6817. currentState.m_stateFlags = BGFX_STATE_NONE;
  6818. currentState.m_stencil = packStencil(BGFX_STENCIL_NONE, BGFX_STENCIL_NONE);
  6819. static ViewState viewState;
  6820. viewState.reset(_render);
  6821. bool wireframe = !!(_render->m_debug&BGFX_DEBUG_WIREFRAME);
  6822. setDebugWireframe(wireframe);
  6823. ProgramHandle currentProgram = BGFX_INVALID_HANDLE;
  6824. bool hasPredefined = false;
  6825. VkPipeline currentPipeline = VK_NULL_HANDLE;
  6826. VkDescriptorSet currentDescriptorSet = VK_NULL_HANDLE;
  6827. uint32_t currentBindHash = 0;
  6828. uint32_t descriptorSetCount = 0;
  6829. VkIndexType currentIndexFormat = VK_INDEX_TYPE_MAX_ENUM;
  6830. SortKey key;
  6831. uint16_t view = UINT16_MAX;
  6832. FrameBufferHandle fbh = { BGFX_CONFIG_MAX_FRAME_BUFFERS };
  6833. BlitState bs(_render);
  6834. uint64_t blendFactor = UINT64_MAX;
  6835. bool wasCompute = false;
  6836. bool viewHasScissor = false;
  6837. bool restoreScissor = false;
  6838. Rect viewScissorRect;
  6839. viewScissorRect.clear();
  6840. bool isFrameBufferValid = false;
  6841. uint32_t statsNumPrimsSubmitted[BX_COUNTOF(s_primInfo)] = {};
  6842. uint32_t statsNumPrimsRendered[BX_COUNTOF(s_primInfo)] = {};
  6843. uint32_t statsNumInstances[BX_COUNTOF(s_primInfo)] = {};
  6844. uint32_t statsNumIndices = 0;
  6845. uint32_t statsKeyType[2] = {};
  6846. const uint64_t f0 = BGFX_STATE_BLEND_FACTOR;
  6847. const uint64_t f1 = BGFX_STATE_BLEND_INV_FACTOR;
  6848. const uint64_t f2 = BGFX_STATE_BLEND_FACTOR<<4;
  6849. const uint64_t f3 = BGFX_STATE_BLEND_INV_FACTOR<<4;
  6850. ScratchBufferVK& scratchBuffer = m_scratchBuffer[m_cmd.m_currentFrameInFlight];
  6851. scratchBuffer.reset();
  6852. setMemoryBarrier(
  6853. m_commandBuffer
  6854. , VK_PIPELINE_STAGE_TRANSFER_BIT
  6855. , VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT | VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  6856. );
  6857. VkRenderPassBeginInfo rpbi;
  6858. rpbi.sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO;
  6859. rpbi.pNext = NULL;
  6860. rpbi.clearValueCount = 0;
  6861. rpbi.pClearValues = NULL;
  6862. bool beginRenderPass = false;
  6863. Profiler<TimerQueryVK> profiler(
  6864. _render
  6865. , m_gpuTimer
  6866. , s_viewName
  6867. , m_timerQuerySupport
  6868. );
  6869. m_occlusionQuery.flush(_render);
  6870. if (0 == (_render->m_debug&BGFX_DEBUG_IFH) )
  6871. {
  6872. viewState.m_rect = _render->m_view[0].m_rect;
  6873. int32_t numItems = _render->m_numRenderItems;
  6874. for (int32_t item = 0; item < numItems;)
  6875. {
  6876. const uint64_t encodedKey = _render->m_sortKeys[item];
  6877. const bool isCompute = key.decode(encodedKey, _render->m_viewRemap);
  6878. statsKeyType[isCompute]++;
  6879. const bool viewChanged = 0
  6880. || key.m_view != view
  6881. || item == numItems
  6882. ;
  6883. const uint32_t itemIdx = _render->m_sortValues[item];
  6884. const RenderItem& renderItem = _render->m_renderItem[itemIdx];
  6885. const RenderBind& renderBind = _render->m_renderItemBind[itemIdx];
  6886. ++item;
  6887. if (viewChanged)
  6888. {
  6889. if (beginRenderPass)
  6890. {
  6891. vkCmdEndRenderPass(m_commandBuffer);
  6892. beginRenderPass = false;
  6893. }
  6894. view = key.m_view;
  6895. currentProgram = BGFX_INVALID_HANDLE;
  6896. hasPredefined = false;
  6897. if (item > 1)
  6898. {
  6899. profiler.end();
  6900. }
  6901. BGFX_VK_PROFILER_END();
  6902. setViewType(view, " ");
  6903. BGFX_VK_PROFILER_BEGIN(view, kColorView);
  6904. profiler.begin(view);
  6905. if (_render->m_view[view].m_fbh.idx != fbh.idx)
  6906. {
  6907. fbh = _render->m_view[view].m_fbh;
  6908. setFrameBuffer(fbh);
  6909. }
  6910. const FrameBufferVK& fb = isValid(m_fbh)
  6911. ? m_frameBuffers[m_fbh.idx]
  6912. : m_backBuffer
  6913. ;
  6914. isFrameBufferValid = fb.isRenderable();
  6915. if (isFrameBufferValid)
  6916. {
  6917. viewState.m_rect = _render->m_view[view].m_rect;
  6918. const Rect& rect = _render->m_view[view].m_rect;
  6919. const Rect& scissorRect = _render->m_view[view].m_scissor;
  6920. viewHasScissor = !scissorRect.isZero();
  6921. viewScissorRect = viewHasScissor ? scissorRect : rect;
  6922. restoreScissor = false;
  6923. rpbi.framebuffer = fb.m_currentFramebuffer;
  6924. rpbi.renderPass = fb.m_renderPass;
  6925. rpbi.renderArea.offset.x = rect.m_x;
  6926. rpbi.renderArea.offset.y = rect.m_y;
  6927. rpbi.renderArea.extent.width = rect.m_width;
  6928. rpbi.renderArea.extent.height = rect.m_height;
  6929. VkViewport vp;
  6930. vp.x = float(rect.m_x);
  6931. vp.y = float(rect.m_y + rect.m_height);
  6932. vp.width = float(rect.m_width);
  6933. vp.height = -float(rect.m_height);
  6934. vp.minDepth = 0.0f;
  6935. vp.maxDepth = 1.0f;
  6936. vkCmdSetViewport(m_commandBuffer, 0, 1, &vp);
  6937. VkRect2D rc;
  6938. rc.offset.x = viewScissorRect.m_x;
  6939. rc.offset.y = viewScissorRect.m_y;
  6940. rc.extent.width = viewScissorRect.m_width;
  6941. rc.extent.height = viewScissorRect.m_height;
  6942. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  6943. const Clear& clr = _render->m_view[view].m_clear;
  6944. if (BGFX_CLEAR_NONE != clr.m_flags)
  6945. {
  6946. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  6947. Rect clearRect = rect;
  6948. clearRect.setIntersect(rect, viewScissorRect);
  6949. clearQuad(clearRect, clr, _render->m_colorPalette);
  6950. vkCmdEndRenderPass(m_commandBuffer);
  6951. }
  6952. submitBlit(bs, view);
  6953. }
  6954. }
  6955. if (isCompute)
  6956. {
  6957. if (!wasCompute)
  6958. {
  6959. wasCompute = true;
  6960. currentBindHash = 0;
  6961. BGFX_VK_PROFILER_END();
  6962. setViewType(view, "C");
  6963. BGFX_VK_PROFILER_BEGIN(view, kColorCompute);
  6964. }
  6965. // renderpass external subpass dependencies handle graphics -> compute and compute -> graphics
  6966. // but not compute -> compute (possibly also across views if they contain no draw calls)
  6967. setMemoryBarrier(
  6968. m_commandBuffer
  6969. , VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT
  6970. , VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT | VK_PIPELINE_STAGE_DRAW_INDIRECT_BIT
  6971. );
  6972. const RenderCompute& compute = renderItem.compute;
  6973. const VkPipeline pipeline = getPipeline(key.m_program);
  6974. if (currentPipeline != pipeline)
  6975. {
  6976. currentPipeline = pipeline;
  6977. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, pipeline);
  6978. }
  6979. bool constantsChanged = false;
  6980. if (compute.m_uniformBegin < compute.m_uniformEnd
  6981. || currentProgram.idx != key.m_program.idx)
  6982. {
  6983. rendererUpdateUniforms(this, _render->m_uniformBuffer[compute.m_uniformIdx], compute.m_uniformBegin, compute.m_uniformEnd);
  6984. currentProgram = key.m_program;
  6985. ProgramVK& program = m_program[currentProgram.idx];
  6986. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  6987. if (NULL != vcb)
  6988. {
  6989. commit(*vcb);
  6990. }
  6991. hasPredefined = 0 < program.m_numPredefined;
  6992. constantsChanged = true;
  6993. }
  6994. const ProgramVK& program = m_program[currentProgram.idx];
  6995. if (constantsChanged
  6996. || hasPredefined)
  6997. {
  6998. viewState.setPredefined<4>(this, view, program, _render, compute);
  6999. }
  7000. if (VK_NULL_HANDLE != program.m_descriptorSetLayout)
  7001. {
  7002. const uint32_t vsize = program.m_vsh->m_size;
  7003. uint32_t numOffset = 0;
  7004. uint32_t offset = 0;
  7005. if (constantsChanged
  7006. || hasPredefined)
  7007. {
  7008. if (vsize > 0)
  7009. {
  7010. offset = scratchBuffer.write(m_vsScratch, vsize);
  7011. ++numOffset;
  7012. }
  7013. }
  7014. bx::HashMurmur2A hash;
  7015. hash.begin();
  7016. hash.add(program.m_descriptorSetLayout);
  7017. hash.add(renderBind.m_bind, sizeof(renderBind.m_bind) );
  7018. hash.add(vsize);
  7019. hash.add(0);
  7020. const uint32_t bindHash = hash.end();
  7021. if (currentBindHash != bindHash)
  7022. {
  7023. currentBindHash = bindHash;
  7024. currentDescriptorSet = getDescriptorSet(
  7025. program
  7026. , renderBind
  7027. , scratchBuffer
  7028. , _render->m_colorPalette
  7029. );
  7030. descriptorSetCount++;
  7031. }
  7032. vkCmdBindDescriptorSets(
  7033. m_commandBuffer
  7034. , VK_PIPELINE_BIND_POINT_COMPUTE
  7035. , program.m_pipelineLayout
  7036. , 0
  7037. , 1
  7038. , &currentDescriptorSet
  7039. , numOffset
  7040. , &offset
  7041. );
  7042. }
  7043. if (isValid(compute.m_indirectBuffer) )
  7044. {
  7045. const VertexBufferVK& vb = m_vertexBuffers[compute.m_indirectBuffer.idx];
  7046. uint32_t numDrawIndirect = UINT16_MAX == compute.m_numIndirect
  7047. ? vb.m_size/BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7048. : compute.m_numIndirect
  7049. ;
  7050. uint32_t args = compute.m_startIndirect * BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7051. for (uint32_t ii = 0; ii < numDrawIndirect; ++ii)
  7052. {
  7053. vkCmdDispatchIndirect(m_commandBuffer, vb.m_buffer, args);
  7054. args += BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7055. }
  7056. }
  7057. else
  7058. {
  7059. vkCmdDispatch(m_commandBuffer, compute.m_numX, compute.m_numY, compute.m_numZ);
  7060. }
  7061. continue;
  7062. }
  7063. const RenderDraw& draw = renderItem.draw;
  7064. rendererUpdateUniforms(this, _render->m_uniformBuffer[draw.m_uniformIdx], draw.m_uniformBegin, draw.m_uniformEnd);
  7065. const bool hasOcclusionQuery = 0 != (draw.m_stateFlags & BGFX_STATE_INTERNAL_OCCLUSION_QUERY);
  7066. {
  7067. const bool occluded = true
  7068. && isValid(draw.m_occlusionQuery)
  7069. && !hasOcclusionQuery
  7070. && !isVisible(_render, draw.m_occlusionQuery, 0 != (draw.m_submitFlags & BGFX_SUBMIT_INTERNAL_OCCLUSION_VISIBLE) )
  7071. ;
  7072. if (occluded
  7073. || !isFrameBufferValid
  7074. || 0 == draw.m_streamMask
  7075. || _render->m_frameCache.isZeroArea(viewScissorRect, draw.m_scissor) )
  7076. {
  7077. continue;
  7078. }
  7079. }
  7080. const uint64_t changedFlags = currentState.m_stateFlags ^ draw.m_stateFlags;
  7081. currentState.m_stateFlags = draw.m_stateFlags;
  7082. if (!beginRenderPass)
  7083. {
  7084. if (wasCompute)
  7085. {
  7086. wasCompute = false;
  7087. currentBindHash = 0;
  7088. }
  7089. BGFX_VK_PROFILER_END();
  7090. setViewType(view, " ");
  7091. BGFX_VK_PROFILER_BEGIN(view, kColorDraw);
  7092. vkCmdBeginRenderPass(m_commandBuffer, &rpbi, VK_SUBPASS_CONTENTS_INLINE);
  7093. beginRenderPass = true;
  7094. currentProgram = BGFX_INVALID_HANDLE;
  7095. currentState.m_scissor = !draw.m_scissor;
  7096. }
  7097. if (0 != draw.m_streamMask)
  7098. {
  7099. const bool bindAttribs = hasVertexStreamChanged(currentState, draw);
  7100. currentState.m_streamMask = draw.m_streamMask;
  7101. currentState.m_instanceDataBuffer = draw.m_instanceDataBuffer;
  7102. currentState.m_instanceDataOffset = draw.m_instanceDataOffset;
  7103. currentState.m_instanceDataStride = draw.m_instanceDataStride;
  7104. const VertexLayout* layouts[BGFX_CONFIG_MAX_VERTEX_STREAMS];
  7105. VkBuffer streamBuffers[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  7106. VkDeviceSize streamOffsets[BGFX_CONFIG_MAX_VERTEX_STREAMS + 1];
  7107. uint8_t numStreams = 0;
  7108. uint32_t numVertices = draw.m_numVertices;
  7109. if (UINT8_MAX != draw.m_streamMask)
  7110. {
  7111. for (uint32_t idx = 0, streamMask = draw.m_streamMask
  7112. ; 0 != streamMask
  7113. ; streamMask >>= 1, idx += 1, ++numStreams
  7114. )
  7115. {
  7116. const uint32_t ntz = bx::uint32_cnttz(streamMask);
  7117. streamMask >>= ntz;
  7118. idx += ntz;
  7119. currentState.m_stream[idx] = draw.m_stream[idx];
  7120. const VertexBufferHandle handle = draw.m_stream[idx].m_handle;
  7121. const VertexBufferVK& vb = m_vertexBuffers[handle.idx];
  7122. const uint16_t decl = isValid(draw.m_stream[idx].m_layoutHandle)
  7123. ? draw.m_stream[idx].m_layoutHandle.idx
  7124. : vb.m_layoutHandle.idx
  7125. ;
  7126. const VertexLayout& layout = m_vertexLayouts[decl];
  7127. const uint32_t stride = layout.m_stride;
  7128. streamBuffers[numStreams] = m_vertexBuffers[handle.idx].m_buffer;
  7129. streamOffsets[numStreams] = draw.m_stream[idx].m_startVertex * stride;
  7130. layouts[numStreams] = &layout;
  7131. numVertices = bx::uint32_min(UINT32_MAX == draw.m_numVertices
  7132. ? vb.m_size/stride
  7133. : draw.m_numVertices
  7134. , numVertices
  7135. );
  7136. }
  7137. }
  7138. if (bindAttribs)
  7139. {
  7140. uint32_t numVertexBuffers = numStreams;
  7141. if (isValid(draw.m_instanceDataBuffer) )
  7142. {
  7143. streamOffsets[numVertexBuffers] = draw.m_instanceDataOffset;
  7144. streamBuffers[numVertexBuffers] = m_vertexBuffers[draw.m_instanceDataBuffer.idx].m_buffer;
  7145. numVertexBuffers++;
  7146. }
  7147. if (0 < numVertexBuffers)
  7148. {
  7149. vkCmdBindVertexBuffers(
  7150. m_commandBuffer
  7151. , 0
  7152. , numVertexBuffers
  7153. , &streamBuffers[0]
  7154. , streamOffsets
  7155. );
  7156. }
  7157. }
  7158. const VkPipeline pipeline =
  7159. getPipeline(draw.m_stateFlags
  7160. , draw.m_stencil
  7161. , numStreams
  7162. , layouts
  7163. , key.m_program
  7164. , uint8_t(draw.m_instanceDataStride/16)
  7165. );
  7166. if (currentPipeline != pipeline)
  7167. {
  7168. currentPipeline = pipeline;
  7169. vkCmdBindPipeline(m_commandBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline);
  7170. }
  7171. const bool hasStencil = 0 != draw.m_stencil;
  7172. if (hasStencil
  7173. && currentState.m_stencil != draw.m_stencil)
  7174. {
  7175. currentState.m_stencil = draw.m_stencil;
  7176. const uint32_t fstencil = unpackStencil(0, draw.m_stencil);
  7177. const uint32_t ref = (fstencil&BGFX_STENCIL_FUNC_REF_MASK)>>BGFX_STENCIL_FUNC_REF_SHIFT;
  7178. vkCmdSetStencilReference(m_commandBuffer, VK_STENCIL_FRONT_AND_BACK, ref);
  7179. }
  7180. const bool hasFactor = 0
  7181. || f0 == (draw.m_stateFlags & f0)
  7182. || f1 == (draw.m_stateFlags & f1)
  7183. || f2 == (draw.m_stateFlags & f2)
  7184. || f3 == (draw.m_stateFlags & f3)
  7185. ;
  7186. if (hasFactor
  7187. && blendFactor != draw.m_rgba)
  7188. {
  7189. blendFactor = draw.m_rgba;
  7190. float bf[4];
  7191. bf[0] = ( (draw.m_rgba>>24) )/255.0f;
  7192. bf[1] = ( (draw.m_rgba>>16)&0xff)/255.0f;
  7193. bf[2] = ( (draw.m_rgba>> 8)&0xff)/255.0f;
  7194. bf[3] = ( (draw.m_rgba )&0xff)/255.0f;
  7195. vkCmdSetBlendConstants(m_commandBuffer, bf);
  7196. }
  7197. const uint16_t scissor = draw.m_scissor;
  7198. if (currentState.m_scissor != scissor)
  7199. {
  7200. currentState.m_scissor = scissor;
  7201. if (UINT16_MAX == scissor)
  7202. {
  7203. if (restoreScissor
  7204. || viewHasScissor)
  7205. {
  7206. restoreScissor = false;
  7207. VkRect2D rc;
  7208. rc.offset.x = viewScissorRect.m_x;
  7209. rc.offset.y = viewScissorRect.m_y;
  7210. rc.extent.width = viewScissorRect.m_width;
  7211. rc.extent.height = viewScissorRect.m_height;
  7212. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  7213. }
  7214. }
  7215. else
  7216. {
  7217. restoreScissor = true;
  7218. Rect scissorRect;
  7219. scissorRect.setIntersect(viewScissorRect, _render->m_frameCache.m_rectCache.m_cache[scissor]);
  7220. VkRect2D rc;
  7221. rc.offset.x = scissorRect.m_x;
  7222. rc.offset.y = scissorRect.m_y;
  7223. rc.extent.width = scissorRect.m_width;
  7224. rc.extent.height = scissorRect.m_height;
  7225. vkCmdSetScissor(m_commandBuffer, 0, 1, &rc);
  7226. }
  7227. }
  7228. bool constantsChanged = false;
  7229. if (draw.m_uniformBegin < draw.m_uniformEnd
  7230. || currentProgram.idx != key.m_program.idx
  7231. || BGFX_STATE_ALPHA_REF_MASK & changedFlags)
  7232. {
  7233. currentProgram = key.m_program;
  7234. ProgramVK& program = m_program[currentProgram.idx];
  7235. UniformBuffer* vcb = program.m_vsh->m_constantBuffer;
  7236. if (NULL != vcb)
  7237. {
  7238. commit(*vcb);
  7239. }
  7240. if (NULL != program.m_fsh)
  7241. {
  7242. UniformBuffer* fcb = program.m_fsh->m_constantBuffer;
  7243. if (NULL != fcb)
  7244. {
  7245. commit(*fcb);
  7246. }
  7247. }
  7248. hasPredefined = 0 < program.m_numPredefined;
  7249. constantsChanged = true;
  7250. }
  7251. const ProgramVK& program = m_program[currentProgram.idx];
  7252. if (hasPredefined)
  7253. {
  7254. uint32_t ref = (draw.m_stateFlags & BGFX_STATE_ALPHA_REF_MASK) >> BGFX_STATE_ALPHA_REF_SHIFT;
  7255. viewState.m_alphaRef = ref / 255.0f;
  7256. viewState.setPredefined<4>(this, view, program, _render, draw);
  7257. }
  7258. if (VK_NULL_HANDLE != program.m_descriptorSetLayout)
  7259. {
  7260. const uint32_t vsize = program.m_vsh->m_size;
  7261. const uint32_t fsize = NULL != program.m_fsh ? program.m_fsh->m_size : 0;
  7262. uint32_t numOffset = 0;
  7263. uint32_t offsets[2] = { 0, 0 };
  7264. if (constantsChanged
  7265. || hasPredefined)
  7266. {
  7267. if (vsize > 0)
  7268. {
  7269. offsets[numOffset++] = scratchBuffer.write(m_vsScratch, vsize);
  7270. }
  7271. if (fsize > 0)
  7272. {
  7273. offsets[numOffset++] = scratchBuffer.write(m_fsScratch, fsize);
  7274. }
  7275. }
  7276. bx::HashMurmur2A hash;
  7277. hash.begin();
  7278. hash.add(program.m_descriptorSetLayout);
  7279. hash.add(renderBind.m_bind, sizeof(renderBind.m_bind) );
  7280. hash.add(vsize);
  7281. hash.add(fsize);
  7282. const uint32_t bindHash = hash.end();
  7283. if (currentBindHash != bindHash)
  7284. {
  7285. currentBindHash = bindHash;
  7286. currentDescriptorSet = getDescriptorSet(
  7287. program
  7288. , renderBind
  7289. , scratchBuffer
  7290. , _render->m_colorPalette
  7291. );
  7292. descriptorSetCount++;
  7293. }
  7294. vkCmdBindDescriptorSets(
  7295. m_commandBuffer
  7296. , VK_PIPELINE_BIND_POINT_GRAPHICS
  7297. , program.m_pipelineLayout
  7298. , 0
  7299. , 1
  7300. , &currentDescriptorSet
  7301. , numOffset
  7302. , offsets
  7303. );
  7304. }
  7305. VkBuffer bufferIndirect = VK_NULL_HANDLE;
  7306. VkBuffer bufferNumIndirect = VK_NULL_HANDLE;
  7307. uint32_t numDrawIndirect = 0;
  7308. uint32_t bufferOffsetIndirect = 0;
  7309. uint32_t bufferNumOffsetIndirect = 0;
  7310. if (isValid(draw.m_indirectBuffer) )
  7311. {
  7312. const VertexBufferVK& vb = m_vertexBuffers[draw.m_indirectBuffer.idx];
  7313. bufferIndirect = vb.m_buffer;
  7314. numDrawIndirect = UINT16_MAX == draw.m_numIndirect
  7315. ? vb.m_size / BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7316. : draw.m_numIndirect
  7317. ;
  7318. bufferOffsetIndirect = draw.m_startIndirect * BGFX_CONFIG_DRAW_INDIRECT_STRIDE;
  7319. if (isValid(draw.m_numIndirectBuffer) )
  7320. {
  7321. bufferNumIndirect = m_indexBuffers[draw.m_numIndirectBuffer.idx].m_buffer;
  7322. bufferNumOffsetIndirect = draw.m_numIndirectIndex * sizeof(uint32_t);
  7323. }
  7324. }
  7325. if (hasOcclusionQuery)
  7326. {
  7327. m_occlusionQuery.begin(draw.m_occlusionQuery);
  7328. }
  7329. const uint8_t primIndex = uint8_t((draw.m_stateFlags & BGFX_STATE_PT_MASK) >> BGFX_STATE_PT_SHIFT);
  7330. const PrimInfo& prim = s_primInfo[primIndex];
  7331. uint32_t numPrimsSubmitted = 0;
  7332. uint32_t numIndices = 0;
  7333. if (!isValid(draw.m_indexBuffer) )
  7334. {
  7335. numPrimsSubmitted = numVertices / prim.m_div - prim.m_sub;
  7336. if (isValid(draw.m_indirectBuffer) )
  7337. {
  7338. if (isValid(draw.m_numIndirectBuffer) )
  7339. {
  7340. vkCmdDrawIndirectCountKHR(
  7341. m_commandBuffer
  7342. , bufferIndirect
  7343. , bufferOffsetIndirect
  7344. , bufferNumIndirect
  7345. , bufferNumOffsetIndirect
  7346. , numDrawIndirect
  7347. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7348. );
  7349. }
  7350. else
  7351. {
  7352. vkCmdDrawIndirect(
  7353. m_commandBuffer
  7354. , bufferIndirect
  7355. , bufferOffsetIndirect
  7356. , numDrawIndirect
  7357. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7358. );
  7359. }
  7360. }
  7361. else
  7362. {
  7363. vkCmdDraw(
  7364. m_commandBuffer
  7365. , numVertices
  7366. , draw.m_numInstances
  7367. , 0
  7368. , 0
  7369. );
  7370. }
  7371. }
  7372. else
  7373. {
  7374. const bool isIndex16 = draw.isIndex16();
  7375. const uint32_t indexSize = isIndex16 ? 2 : 4;
  7376. const VkIndexType indexFormat = isIndex16 ? VK_INDEX_TYPE_UINT16 : VK_INDEX_TYPE_UINT32;
  7377. const BufferVK& ib = m_indexBuffers[draw.m_indexBuffer.idx];
  7378. numIndices = UINT32_MAX == draw.m_numIndices
  7379. ? ib.m_size / indexSize
  7380. : draw.m_numIndices
  7381. ;
  7382. numPrimsSubmitted = numIndices / prim.m_div - prim.m_sub;
  7383. if (currentState.m_indexBuffer.idx != draw.m_indexBuffer.idx
  7384. || currentIndexFormat != indexFormat)
  7385. {
  7386. currentState.m_indexBuffer = draw.m_indexBuffer;
  7387. currentIndexFormat = indexFormat;
  7388. vkCmdBindIndexBuffer(
  7389. m_commandBuffer
  7390. , m_indexBuffers[draw.m_indexBuffer.idx].m_buffer
  7391. , 0
  7392. , indexFormat
  7393. );
  7394. }
  7395. if (isValid(draw.m_indirectBuffer) )
  7396. {
  7397. if (isValid(draw.m_numIndirectBuffer) )
  7398. {
  7399. vkCmdDrawIndexedIndirectCountKHR(
  7400. m_commandBuffer
  7401. , bufferIndirect
  7402. , bufferOffsetIndirect
  7403. , bufferNumIndirect
  7404. , bufferNumOffsetIndirect
  7405. , numDrawIndirect
  7406. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7407. );
  7408. }
  7409. else
  7410. {
  7411. vkCmdDrawIndexedIndirect(
  7412. m_commandBuffer
  7413. , bufferIndirect
  7414. , bufferOffsetIndirect
  7415. , numDrawIndirect
  7416. , BGFX_CONFIG_DRAW_INDIRECT_STRIDE
  7417. );
  7418. }
  7419. }
  7420. else
  7421. {
  7422. vkCmdDrawIndexed(
  7423. m_commandBuffer
  7424. , numIndices
  7425. , draw.m_numInstances
  7426. , draw.m_startIndex
  7427. , 0
  7428. , 0
  7429. );
  7430. }
  7431. }
  7432. uint32_t numPrimsRendered = numPrimsSubmitted*draw.m_numInstances;
  7433. statsNumPrimsSubmitted[primIndex] += numPrimsSubmitted;
  7434. statsNumPrimsRendered[primIndex] += numPrimsRendered;
  7435. statsNumInstances[primIndex] += draw.m_numInstances;
  7436. statsNumIndices += numIndices;
  7437. if (hasOcclusionQuery)
  7438. {
  7439. m_occlusionQuery.end();
  7440. }
  7441. }
  7442. }
  7443. if (beginRenderPass)
  7444. {
  7445. vkCmdEndRenderPass(m_commandBuffer);
  7446. beginRenderPass = false;
  7447. }
  7448. if (wasCompute)
  7449. {
  7450. setViewType(view, "C");
  7451. BGFX_VK_PROFILER_END();
  7452. BGFX_VK_PROFILER_BEGIN(view, kColorCompute);
  7453. }
  7454. submitBlit(bs, BGFX_CONFIG_MAX_VIEWS);
  7455. if (0 < _render->m_numRenderItems)
  7456. {
  7457. captureElapsed = -bx::getHPCounter();
  7458. capture();
  7459. captureElapsed += bx::getHPCounter();
  7460. profiler.end();
  7461. }
  7462. }
  7463. BGFX_VK_PROFILER_END();
  7464. int64_t timeEnd = bx::getHPCounter();
  7465. int64_t frameTime = timeEnd - timeBegin;
  7466. static int64_t min = frameTime;
  7467. static int64_t max = frameTime;
  7468. min = bx::min<int64_t>(min, frameTime);
  7469. max = bx::max<int64_t>(max, frameTime);
  7470. static uint32_t maxGpuLatency = 0;
  7471. static double maxGpuElapsed = 0.0f;
  7472. double elapsedGpuMs = 0.0;
  7473. static int64_t presentMin = m_presentElapsed;
  7474. static int64_t presentMax = m_presentElapsed;
  7475. presentMin = bx::min<int64_t>(presentMin, m_presentElapsed);
  7476. presentMax = bx::max<int64_t>(presentMax, m_presentElapsed);
  7477. if (UINT32_MAX != frameQueryIdx)
  7478. {
  7479. m_gpuTimer.end(frameQueryIdx);
  7480. const TimerQueryVK::Result& result = m_gpuTimer.m_result[BGFX_CONFIG_MAX_VIEWS];
  7481. double toGpuMs = 1000.0 / double(m_gpuTimer.m_frequency);
  7482. elapsedGpuMs = (result.m_end - result.m_begin) * toGpuMs;
  7483. maxGpuElapsed = elapsedGpuMs > maxGpuElapsed ? elapsedGpuMs : maxGpuElapsed;
  7484. maxGpuLatency = bx::uint32_imax(maxGpuLatency, result.m_pending-1);
  7485. }
  7486. maxGpuLatency = bx::uint32_imax(maxGpuLatency, m_gpuTimer.m_control.available()-1);
  7487. const int64_t timerFreq = bx::getHPFrequency();
  7488. VkPhysicalDeviceMemoryBudgetPropertiesEXT dmbp;
  7489. dmbp.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_BUDGET_PROPERTIES_EXT;
  7490. dmbp.pNext = NULL;
  7491. int64_t gpuMemoryAvailable = -INT64_MAX;
  7492. int64_t gpuMemoryUsed = -INT64_MAX;
  7493. if (s_extension[Extension::EXT_memory_budget].m_supported)
  7494. {
  7495. VkPhysicalDeviceMemoryProperties2 pdmp2;
  7496. pdmp2.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_MEMORY_PROPERTIES_2;
  7497. pdmp2.pNext = &dmbp;
  7498. vkGetPhysicalDeviceMemoryProperties2KHR(m_physicalDevice, &pdmp2);
  7499. gpuMemoryAvailable = 0;
  7500. gpuMemoryUsed = 0;
  7501. for (uint32_t ii = 0; ii < m_memoryProperties.memoryHeapCount; ++ii)
  7502. {
  7503. if (!!(m_memoryProperties.memoryHeaps[ii].flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) )
  7504. {
  7505. gpuMemoryAvailable += dmbp.heapBudget[ii];
  7506. gpuMemoryUsed += dmbp.heapUsage[ii];
  7507. }
  7508. }
  7509. }
  7510. Stats& perfStats = _render->m_perfStats;
  7511. perfStats.cpuTimeBegin = timeBegin;
  7512. perfStats.cpuTimeEnd = timeEnd;
  7513. perfStats.cpuTimerFreq = timerFreq;
  7514. const TimerQueryVK::Result& result = m_gpuTimer.m_result[BGFX_CONFIG_MAX_VIEWS];
  7515. perfStats.gpuTimeBegin = result.m_begin;
  7516. perfStats.gpuTimeEnd = result.m_end;
  7517. perfStats.gpuTimerFreq = m_gpuTimer.m_frequency;
  7518. perfStats.numDraw = statsKeyType[0];
  7519. perfStats.numCompute = statsKeyType[1];
  7520. perfStats.numBlit = _render->m_numBlitItems;
  7521. perfStats.maxGpuLatency = maxGpuLatency;
  7522. perfStats.gpuFrameNum = result.m_frameNum;
  7523. bx::memCopy(perfStats.numPrims, statsNumPrimsRendered, sizeof(perfStats.numPrims) );
  7524. perfStats.gpuMemoryMax = gpuMemoryAvailable;
  7525. perfStats.gpuMemoryUsed = gpuMemoryUsed;
  7526. if (_render->m_debug & (BGFX_DEBUG_IFH|BGFX_DEBUG_STATS) )
  7527. {
  7528. BGFX_VK_PROFILER_BEGIN_LITERAL("debugstats", kColorFrame);
  7529. TextVideoMem& tvm = m_textVideoMem;
  7530. static int64_t next = timeEnd;
  7531. if (timeEnd >= next)
  7532. {
  7533. next = timeEnd + timerFreq;
  7534. double freq = double(timerFreq);
  7535. double toMs = 1000.0 / freq;
  7536. tvm.clear();
  7537. uint16_t pos = 0;
  7538. tvm.printf(0, pos++, BGFX_CONFIG_DEBUG ? 0x8c : 0x8f
  7539. , " %s / " BX_COMPILER_NAME
  7540. " / " BX_CPU_NAME
  7541. " / " BX_ARCH_NAME
  7542. " / " BX_PLATFORM_NAME
  7543. " / Version 1.%d.%d (commit: " BGFX_REV_SHA1 ")"
  7544. , getRendererName()
  7545. , BGFX_API_VERSION
  7546. , BGFX_REV_NUMBER
  7547. );
  7548. const VkPhysicalDeviceProperties& pdp = m_deviceProperties;
  7549. tvm.printf(0, pos++, 0x8f, " Device: %s (%s)"
  7550. , pdp.deviceName
  7551. , getName(pdp.deviceType)
  7552. );
  7553. if (0 <= gpuMemoryAvailable && 0 <= gpuMemoryUsed)
  7554. {
  7555. for (uint32_t ii = 0; ii < m_memoryProperties.memoryHeapCount; ++ii)
  7556. {
  7557. char budget[16];
  7558. bx::prettify(budget, BX_COUNTOF(budget), dmbp.heapBudget[ii]);
  7559. char usage[16];
  7560. bx::prettify(usage, BX_COUNTOF(usage), dmbp.heapUsage[ii]);
  7561. const bool local = (!!(m_memoryProperties.memoryHeaps[ii].flags & VK_MEMORY_HEAP_DEVICE_LOCAL_BIT) );
  7562. tvm.printf(0, pos++, 0x8f, " Memory %d %s - Budget: %12s, Usage: %12s"
  7563. , ii
  7564. , local ? "(local) " : "(non-local)"
  7565. , budget
  7566. , usage
  7567. );
  7568. }
  7569. }
  7570. pos = 10;
  7571. tvm.printf(10, pos++, 0x8b, " Frame: % 7.3f, % 7.3f \x1f, % 7.3f \x1e [ms] / % 6.2f FPS "
  7572. , double(frameTime)*toMs
  7573. , double(min)*toMs
  7574. , double(max)*toMs
  7575. , freq/frameTime
  7576. );
  7577. tvm.printf(10, pos++, 0x8b, " Present: % 7.3f, % 7.3f \x1f, % 7.3f \x1e [ms] "
  7578. , double(m_presentElapsed)*toMs
  7579. , double(presentMin)*toMs
  7580. , double(presentMax)*toMs
  7581. );
  7582. const uint32_t msaa = (m_resolution.reset&BGFX_RESET_MSAA_MASK)>>BGFX_RESET_MSAA_SHIFT;
  7583. tvm.printf(10, pos++, 0x8b, " Reset flags: [%c] vsync, [%c] MSAAx%d, [%c] MaxAnisotropy "
  7584. , !!(m_resolution.reset&BGFX_RESET_VSYNC) ? '\xfe' : ' '
  7585. , 0 != msaa ? '\xfe' : ' '
  7586. , 1<<msaa
  7587. , !!(m_resolution.reset&BGFX_RESET_MAXANISOTROPY) ? '\xfe' : ' '
  7588. );
  7589. double elapsedCpuMs = double(frameTime)*toMs;
  7590. tvm.printf(10, pos++, 0x8b, " Submitted: %5d (draw %5d, compute %4d) / CPU %7.4f [ms] "
  7591. , _render->m_numRenderItems
  7592. , statsKeyType[0]
  7593. , statsKeyType[1]
  7594. , elapsedCpuMs
  7595. );
  7596. for (uint32_t ii = 0; ii < Topology::Count; ++ii)
  7597. {
  7598. tvm.printf(10, pos++, 0x8b, " %9s: %7d (#inst: %5d), submitted: %7d "
  7599. , getName(Topology::Enum(ii) )
  7600. , statsNumPrimsRendered[ii]
  7601. , statsNumInstances[ii]
  7602. , statsNumPrimsSubmitted[ii]
  7603. );
  7604. }
  7605. if (NULL != m_renderDocDll)
  7606. {
  7607. tvm.printf(tvm.m_width-27, 0, 0x4f, " [F11 - RenderDoc capture] ");
  7608. }
  7609. tvm.printf(10, pos++, 0x8b, " Indices: %7d ", statsNumIndices);
  7610. // tvm.printf(10, pos++, 0x8b, " Uniform size: %7d, Max: %7d ", _render->m_uniformEnd, _render->m_uniformMax);
  7611. tvm.printf(10, pos++, 0x8b, " DVB size: %7d ", _render->m_vboffset);
  7612. tvm.printf(10, pos++, 0x8b, " DIB size: %7d ", _render->m_iboffset);
  7613. pos++;
  7614. tvm.printf(10, pos++, 0x8b, " Occlusion queries: %3d ", m_occlusionQuery.m_control.available() );
  7615. pos++;
  7616. tvm.printf(10, pos++, 0x8b, " State cache: ");
  7617. tvm.printf(10, pos++, 0x8b, " PSO | DSL | DS ");
  7618. tvm.printf(10, pos++, 0x8b, " %6d | %6d | %6d "
  7619. , m_pipelineStateCache.getCount()
  7620. , m_descriptorSetLayoutCache.getCount()
  7621. , descriptorSetCount
  7622. );
  7623. pos++;
  7624. double captureMs = double(captureElapsed)*toMs;
  7625. tvm.printf(10, pos++, 0x8b, " Capture: %7.4f [ms] ", captureMs);
  7626. uint8_t attr[2] = { 0x8c, 0x8a };
  7627. uint8_t attrIndex = _render->m_waitSubmit < _render->m_waitRender;
  7628. tvm.printf(10, pos++, attr[attrIndex&1], " Submit wait: %7.4f [ms] ", _render->m_waitSubmit*toMs);
  7629. tvm.printf(10, pos++, attr[(attrIndex+1)&1], " Render wait: %7.4f [ms] ", _render->m_waitRender*toMs);
  7630. min = frameTime;
  7631. max = frameTime;
  7632. presentMin = m_presentElapsed;
  7633. presentMax = m_presentElapsed;
  7634. }
  7635. blit(this, _textVideoMemBlitter, tvm);
  7636. BGFX_VK_PROFILER_END();
  7637. }
  7638. else if (_render->m_debug & BGFX_DEBUG_TEXT)
  7639. {
  7640. BGFX_VK_PROFILER_BEGIN_LITERAL("debugtext", kColorFrame);
  7641. blit(this, _textVideoMemBlitter, _render->m_textVideoMem);
  7642. BGFX_VK_PROFILER_END();
  7643. }
  7644. m_presentElapsed = 0;
  7645. scratchBuffer.flush();
  7646. for (uint16_t ii = 0; ii < m_numWindows; ++ii)
  7647. {
  7648. FrameBufferVK& fb = isValid(m_windows[ii])
  7649. ? m_frameBuffers[m_windows[ii].idx]
  7650. : m_backBuffer
  7651. ;
  7652. if (fb.m_needPresent)
  7653. {
  7654. fb.resolve();
  7655. fb.m_swapChain.transitionImage(m_commandBuffer);
  7656. m_cmd.addWaitSemaphore(fb.m_swapChain.m_lastImageAcquiredSemaphore, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT);
  7657. m_cmd.addSignalSemaphore(fb.m_swapChain.m_lastImageRenderedSemaphore);
  7658. fb.m_swapChain.m_lastImageAcquiredSemaphore = VK_NULL_HANDLE;
  7659. fb.m_swapChain.m_backBufferFence[fb.m_swapChain.m_backBufferColorIdx] = m_cmd.m_currentFence;
  7660. }
  7661. }
  7662. kick();
  7663. }
  7664. } /* namespace vk */ } // namespace bgfx
  7665. #else
  7666. namespace bgfx { namespace vk
  7667. {
  7668. RendererContextI* rendererCreate(const Init& _init)
  7669. {
  7670. BX_UNUSED(_init);
  7671. return NULL;
  7672. }
  7673. void rendererDestroy()
  7674. {
  7675. }
  7676. } /* namespace vk */ } // namespace bgfx
  7677. #endif // BGFX_CONFIG_RENDERER_VULKAN