| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223222422252226222722282229223022312232223322342235223622372238223922402241224222432244224522462247224822492250225122522253225422552256225722582259226022612262226322642265226622672268226922702271227222732274227522762277227822792280228122822283228422852286228722882289229022912292229322942295229622972298229923002301230223032304230523062307230823092310231123122313231423152316231723182319232023212322232323242325232623272328232923302331233223332334233523362337233823392340234123422343234423452346234723482349235023512352235323542355235623572358235923602361236223632364236523662367236823692370237123722373237423752376237723782379238023812382238323842385238623872388238923902391239223932394239523962397239823992400240124022403240424052406240724082409241024112412241324142415241624172418241924202421242224232424242524262427242824292430243124322433243424352436243724382439244024412442244324442445244624472448244924502451245224532454245524562457245824592460246124622463246424652466246724682469247024712472247324742475247624772478247924802481248224832484248524862487248824892490249124922493249424952496249724982499250025012502250325042505250625072508250925102511251225132514251525162517251825192520252125222523252425252526252725282529253025312532253325342535253625372538253925402541254225432544254525462547254825492550255125522553255425552556255725582559256025612562256325642565256625672568256925702571257225732574257525762577257825792580258125822583258425852586258725882589259025912592259325942595259625972598259926002601260226032604260526062607260826092610261126122613261426152616261726182619262026212622262326242625262626272628262926302631263226332634263526362637263826392640264126422643264426452646264726482649265026512652265326542655265626572658265926602661266226632664266526662667266826692670267126722673267426752676267726782679268026812682268326842685268626872688268926902691269226932694269526962697269826992700270127022703270427052706270727082709271027112712271327142715271627172718271927202721272227232724272527262727272827292730273127322733273427352736273727382739274027412742274327442745274627472748274927502751275227532754275527562757275827592760276127622763276427652766276727682769277027712772277327742775277627772778277927802781278227832784278527862787278827892790279127922793279427952796279727982799280028012802280328042805280628072808280928102811281228132814281528162817281828192820282128222823282428252826282728282829283028312832283328342835283628372838283928402841284228432844284528462847284828492850285128522853285428552856285728582859286028612862286328642865286628672868286928702871287228732874287528762877287828792880288128822883288428852886288728882889289028912892289328942895289628972898289929002901290229032904290529062907290829092910291129122913291429152916291729182919292029212922292329242925292629272928292929302931293229332934293529362937293829392940294129422943294429452946294729482949295029512952295329542955295629572958295929602961296229632964296529662967296829692970297129722973297429752976297729782979298029812982298329842985298629872988298929902991299229932994299529962997299829993000300130023003300430053006300730083009301030113012301330143015301630173018301930203021302230233024302530263027302830293030303130323033303430353036303730383039304030413042304330443045304630473048304930503051305230533054305530563057305830593060306130623063306430653066306730683069307030713072307330743075307630773078307930803081308230833084308530863087308830893090309130923093309430953096309730983099310031013102310331043105310631073108310931103111311231133114311531163117311831193120312131223123312431253126312731283129313031313132313331343135313631373138313931403141314231433144314531463147314831493150315131523153315431553156315731583159316031613162316331643165316631673168316931703171317231733174317531763177317831793180318131823183318431853186318731883189319031913192319331943195319631973198319932003201320232033204320532063207320832093210321132123213321432153216321732183219322032213222322332243225322632273228322932303231323232333234323532363237323832393240324132423243324432453246324732483249325032513252325332543255325632573258325932603261326232633264326532663267326832693270327132723273327432753276327732783279328032813282328332843285328632873288328932903291329232933294329532963297329832993300330133023303330433053306330733083309331033113312331333143315331633173318331933203321332233233324332533263327332833293330333133323333333433353336333733383339334033413342334333443345334633473348334933503351335233533354335533563357335833593360336133623363336433653366336733683369337033713372337333743375337633773378337933803381338233833384338533863387338833893390339133923393339433953396339733983399340034013402340334043405340634073408340934103411341234133414341534163417341834193420342134223423342434253426342734283429343034313432343334343435343634373438343934403441344234433444344534463447344834493450345134523453345434553456345734583459346034613462346334643465346634673468346934703471347234733474347534763477347834793480348134823483348434853486348734883489349034913492349334943495349634973498349935003501350235033504350535063507350835093510351135123513351435153516351735183519352035213522352335243525352635273528352935303531353235333534353535363537353835393540354135423543354435453546354735483549355035513552355335543555355635573558355935603561356235633564356535663567356835693570357135723573357435753576357735783579358035813582358335843585358635873588358935903591359235933594359535963597359835993600360136023603360436053606360736083609361036113612361336143615361636173618361936203621362236233624362536263627362836293630363136323633363436353636363736383639364036413642364336443645364636473648364936503651365236533654365536563657365836593660366136623663366436653666366736683669367036713672367336743675367636773678367936803681368236833684368536863687368836893690369136923693369436953696369736983699370037013702370337043705370637073708370937103711371237133714371537163717371837193720372137223723372437253726372737283729373037313732373337343735373637373738373937403741374237433744374537463747374837493750375137523753375437553756375737583759376037613762376337643765376637673768376937703771377237733774377537763777377837793780378137823783378437853786378737883789379037913792379337943795379637973798379938003801380238033804380538063807380838093810381138123813381438153816381738183819382038213822382338243825382638273828382938303831383238333834383538363837383838393840384138423843384438453846384738483849385038513852385338543855385638573858385938603861386238633864386538663867386838693870387138723873387438753876387738783879388038813882388338843885388638873888388938903891389238933894389538963897389838993900390139023903390439053906390739083909391039113912391339143915391639173918391939203921392239233924392539263927392839293930393139323933393439353936393739383939394039413942394339443945394639473948394939503951395239533954395539563957395839593960396139623963396439653966396739683969397039713972397339743975397639773978397939803981398239833984398539863987398839893990399139923993399439953996399739983999400040014002400340044005400640074008400940104011401240134014401540164017401840194020402140224023402440254026402740284029403040314032403340344035403640374038403940404041404240434044404540464047404840494050405140524053405440554056405740584059406040614062406340644065406640674068406940704071407240734074407540764077407840794080408140824083408440854086408740884089409040914092409340944095409640974098409941004101410241034104410541064107410841094110411141124113411441154116411741184119412041214122412341244125412641274128412941304131413241334134413541364137413841394140414141424143414441454146414741484149415041514152415341544155415641574158415941604161416241634164416541664167416841694170417141724173417441754176417741784179418041814182418341844185418641874188418941904191419241934194419541964197419841994200420142024203420442054206420742084209421042114212421342144215421642174218421942204221422242234224422542264227422842294230423142324233423442354236423742384239424042414242424342444245424642474248424942504251425242534254425542564257425842594260426142624263426442654266426742684269427042714272427342744275427642774278427942804281428242834284428542864287428842894290429142924293429442954296429742984299430043014302430343044305430643074308430943104311431243134314431543164317431843194320432143224323432443254326432743284329433043314332433343344335433643374338433943404341434243434344434543464347434843494350435143524353435443554356435743584359436043614362436343644365436643674368436943704371437243734374437543764377437843794380438143824383438443854386438743884389439043914392439343944395439643974398439944004401440244034404440544064407440844094410441144124413441444154416441744184419442044214422442344244425442644274428442944304431443244334434443544364437443844394440444144424443444444454446444744484449445044514452445344544455445644574458445944604461446244634464446544664467446844694470447144724473447444754476447744784479448044814482448344844485448644874488448944904491449244934494449544964497449844994500450145024503450445054506450745084509451045114512451345144515451645174518451945204521452245234524452545264527452845294530453145324533453445354536453745384539454045414542454345444545454645474548454945504551455245534554455545564557455845594560456145624563456445654566456745684569457045714572457345744575457645774578457945804581458245834584458545864587458845894590459145924593459445954596459745984599460046014602460346044605460646074608460946104611461246134614461546164617461846194620462146224623462446254626462746284629463046314632463346344635463646374638463946404641464246434644464546464647464846494650465146524653465446554656465746584659466046614662466346644665466646674668466946704671467246734674467546764677467846794680468146824683468446854686468746884689469046914692469346944695469646974698469947004701470247034704470547064707470847094710471147124713471447154716471747184719472047214722472347244725472647274728472947304731473247334734473547364737473847394740474147424743474447454746474747484749475047514752475347544755475647574758475947604761476247634764476547664767476847694770477147724773477447754776477747784779478047814782478347844785478647874788478947904791479247934794479547964797479847994800480148024803480448054806480748084809481048114812481348144815481648174818481948204821482248234824482548264827482848294830483148324833483448354836483748384839484048414842484348444845484648474848484948504851485248534854485548564857485848594860486148624863486448654866486748684869487048714872487348744875487648774878487948804881488248834884488548864887488848894890489148924893489448954896489748984899490049014902490349044905490649074908490949104911491249134914491549164917491849194920492149224923492449254926492749284929493049314932493349344935493649374938493949404941494249434944494549464947494849494950495149524953495449554956495749584959496049614962496349644965496649674968496949704971497249734974497549764977497849794980498149824983498449854986498749884989499049914992499349944995499649974998499950005001500250035004500550065007500850095010501150125013501450155016501750185019502050215022502350245025502650275028502950305031503250335034503550365037503850395040504150425043504450455046504750485049505050515052505350545055505650575058505950605061506250635064506550665067506850695070507150725073507450755076507750785079508050815082508350845085508650875088508950905091509250935094509550965097509850995100510151025103510451055106510751085109511051115112511351145115511651175118511951205121512251235124512551265127512851295130513151325133513451355136513751385139514051415142514351445145514651475148514951505151515251535154515551565157515851595160516151625163516451655166516751685169517051715172517351745175517651775178517951805181518251835184518551865187518851895190519151925193519451955196519751985199520052015202520352045205520652075208520952105211521252135214521552165217521852195220522152225223522452255226522752285229523052315232523352345235523652375238523952405241524252435244524552465247524852495250525152525253525452555256525752585259526052615262526352645265526652675268526952705271527252735274527552765277527852795280528152825283528452855286528752885289529052915292529352945295529652975298529953005301530253035304530553065307530853095310531153125313531453155316531753185319532053215322532353245325532653275328532953305331533253335334533553365337533853395340534153425343534453455346534753485349535053515352535353545355535653575358535953605361536253635364536553665367536853695370537153725373537453755376537753785379538053815382538353845385538653875388538953905391539253935394539553965397539853995400540154025403540454055406540754085409541054115412541354145415541654175418541954205421542254235424542554265427542854295430543154325433543454355436543754385439544054415442544354445445544654475448544954505451545254535454545554565457545854595460546154625463546454655466546754685469547054715472547354745475547654775478547954805481548254835484548554865487548854895490549154925493549454955496549754985499550055015502550355045505550655075508550955105511551255135514551555165517551855195520552155225523552455255526552755285529553055315532553355345535553655375538553955405541554255435544554555465547554855495550555155525553555455555556555755585559556055615562556355645565556655675568556955705571557255735574557555765577557855795580558155825583558455855586558755885589559055915592559355945595559655975598559956005601560256035604560556065607560856095610561156125613561456155616561756185619562056215622562356245625562656275628562956305631563256335634563556365637563856395640564156425643564456455646564756485649565056515652565356545655565656575658565956605661566256635664566556665667566856695670567156725673567456755676567756785679568056815682568356845685568656875688568956905691569256935694569556965697569856995700570157025703570457055706570757085709571057115712571357145715571657175718571957205721572257235724572557265727572857295730573157325733573457355736573757385739574057415742574357445745574657475748574957505751575257535754575557565757575857595760576157625763576457655766576757685769577057715772577357745775577657775778577957805781578257835784578557865787578857895790579157925793579457955796579757985799580058015802580358045805580658075808580958105811581258135814581558165817581858195820582158225823582458255826582758285829583058315832583358345835583658375838583958405841584258435844584558465847584858495850585158525853585458555856585758585859586058615862586358645865586658675868586958705871587258735874587558765877587858795880588158825883588458855886588758885889589058915892589358945895589658975898589959005901590259035904590559065907590859095910591159125913591459155916591759185919592059215922592359245925592659275928592959305931593259335934593559365937593859395940594159425943594459455946594759485949595059515952595359545955595659575958595959605961596259635964596559665967596859695970597159725973597459755976597759785979598059815982598359845985598659875988598959905991599259935994599559965997599859996000600160026003600460056006600760086009601060116012601360146015601660176018601960206021602260236024602560266027602860296030603160326033603460356036603760386039604060416042604360446045604660476048604960506051605260536054605560566057605860596060606160626063606460656066606760686069607060716072607360746075607660776078607960806081608260836084608560866087608860896090609160926093609460956096609760986099610061016102610361046105610661076108 |
- tring';
- import { Readable } from 'node:stream';
- import { finished } from 'node:stream/promises';
- import { URL } from 'node:url';
- import util from 'node:util';
- const exec = util.promisify(child_process.exec);
- import { Readability } from '@mozilla/readability';
- import FileCookieStore from '@root/file-cookie-store';
- import merge from 'deepmerge';
- import { createCursor, getRandomPagePoint } from 'ghost-cursor';
- import { JSDOM, VirtualConsole } from 'jsdom';
- import mime from 'mime-types';
- import ToughCookie from 'tough-cookie';
- import unzip from 'unzip-crx-3';
- import puppeteer from 'puppeteer';
- import { Browser, Page, Cookie, HTTPResponse } from 'puppeteer';
- import { Cluster } from 'puppeteer-cluster';
- import PupeteerExtra from "puppeteer-extra";
- import Stealth#!/usr/bin/env node --env-file .env
- // https://gist.github.com/pirate/d9a350e83025a1e6cf452cddd815d0d4
- // npm install request node-request minimist deepmerge mime-types decompress puppeteer-extra puppeteer-extra-plugin-repl puppeteer-extra-plugin-user-preferences puppeteer-extra-plugin-recaptcha puppeteer-extra-plugin-stealth puppeteer-screen-recorder puppeteer-cluster ghost-cursor @mozilla/readability jsdom unzip-crx-3 node-fetch@2
- import assert from 'node:assert/strict';
- import { Buffer } from 'node:buffer';
- import child_process from 'node:child_process';
- import crypto from 'node:crypto';
- import fs from 'node:fs';
- import { createServer } from 'node:http';
- import os from 'node:os';
- import path from 'node:path';
- import querystring from 'node:querysPlugin from "puppeteer-extra-plugin-stealth";
- import PrefsPlugin from 'puppeteer-extra-plugin-user-preferences';
- import { PuppeteerScreenRecorder } from 'puppeteer-screen-recorder';
- // import RecaptchaPlugin from 'puppeteer-extra-plugin-recaptcha';
- // import ReplPlugin from 'puppeteer-extra-plugin-repl';
- const __dirname = import.meta.dirname
- import { getDatabase } from './models/init-models.js';
- const { Tag, Snapshot, ArchiveResult } = await getDatabase({ dbpath: './index.sqlite3' })
- // move mitm CA cert into /usr/local/share/ca-certificates/mitmproxy-ca-cert.crt
- // update-ca-certificates
- const ANSI = {
- reset: "\x1b[0m",
- blue: "\x1b[34m",
- black: "\x1b[30m",
- }
- /************************* Main Input Arguments *******************************/
- let URLS = [
- // 'chrome://about',
- // 'chrome://system/#chrome_root_store',
- 'https://facebook.com/815781663692514/?comment_id=1508571679703640',
- 'https://www.instagram.com/p/CrTY1fENHr5/',
- 'https://www.tiktok.com/@zemmour_eric/video/7342474065598319904?cid=7343316616878490400',
- 'https://twitter.com/DZasken68678/status/1799833933271687304',
- 'https://t.me/IONONMIARRENDOGROUP/13598',
- 'https://www.youtube.com/watch?v=rpD0qgzlCms',
- 'https://www.aap.com.au/factcheck/aboriginal-lands-claim-a-total-abdication-of-facts/',
- 'https://gologin.com/check-browser',
- 'https://arh.antoinevastel.com/bots/areyouheadless',
- 'https://2captcha.com/demo/hcaptcha',
- 'https://2captcha.com/demo/cloudflare-turnstile',
- 'https://2captcha.com/demo/recaptcha-v3',
- 'https://ipinfo.io/',
- // 'https://2captcha.com/demo/recaptcha-v2',
- // 'https://2captcha.com/demo/keycaptcha',
- // 'https://browserleaks.com/canvas',
- // 'https://bot.incolumitas.com/#botChallenge',
- // 'https://infosimples.github.io/detect-headless/',
- // 'https://coveryourtracks.eff.org/',
- // 'https://fingerprint.com/demo/',
- // 'https://nowsecure.nl',
- // 'https://abrahamjuliot.github.io/creepjs/',
- // 'https://scrapfly.io/web-scraping-tools/http2-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/browser-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/ja3-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/canvas-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/webgl-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/audio-fingerprint',
- // 'https://scrapfly.io/web-scraping-tools/screen-fingerprint',
- // 'https://web-scraping.dev/',
- // 'https://example.com',
- // 'https://www.okta.com/',
- // 'https://www.webflow.com/',
- // 'https://docker-compose.archivebox.io',
- // 'https://www.reddit.com/r/AskReddit/comments/1br0q9b/what_was_ok_10_years_ago_but_isnt_today/',
- // 'https://www.quora.com/Is-the-website-2Captcha-true-or-fake-with-paying-money-for-working-on-it',
- // 'https://x.com/yawnzzcalo7/status/1747853178849435894',
- // 'https://twitter.com/yawnzzcalo7/status/1747853178849435894',
- // 'https://rachdele.substack.com/p/is-the-job-market-dying',
- // 'https://www.flowradar.com/cloneables/mouse-image-trail-effect',
- // 'https://wrong.host.badssl.com/',
- // 'http://docker-compose.archivebox.io',
- // 'https://pptr.dev/api/puppeteer.page.setrequestinterception',
- // 'https://blog.sweeting.me#Writing',
- // 'https://github.com/yarnpkg/yarn/issues/9005',
- // 'https://archive.md/739Oc',
- // 'https://archive.md/Oc72d',
- // 'https://archive.vn/fPUBe',
- // 'https://archive.vn/mRz4P',
- // 'https://archive.vn/Qct6Y',
- // 'https://archive.vn/sv50h',
- // 'https://facebook.com/815781663692514/?comment_id=1508571679703640',
- // 'https://facebook.com/815781663692514/?comment_id=924451748966499',
- // 'https://www.facebook.com/wayne.brennan.528/posts/pfbid02fvxFppng2WsHMavhBa62cXizCBGdmPQRH3CMhac79qzS5C1ADaSNC587d3u6qVbkl',
- // 'https://www.facebook.com/wildeprods/posts/pfbid02YEPfoB7pZqMNzE4y2MpYSQbRAzASquvHyEMzHqrNngJCSL7onEg2jnsqS6epcQHWl',
- // 'https://t.me/aubontouite_francais/9493',
- // 'https://t.me/BC_BLACKMIROR/5044',
- // 'https://t.me/IONONMIARRENDOGROUP/14004',
- // 'https://t.me/newsfactory_pl/51014',
- // 'https://t.me/oliverjanich/132574',
- // 'https://t.me/tomaszgryguc/10449',
- // 'https://t.me/amigosDisidentes/123177',
- // 'https://twitter.com/1nfiltr4do_NN/status/1767238399943991389',
- // 'https://twitter.com/4lmondcookie/status/1748519205438111914',
- // 'https://twitter.com/4olll1ke/status/1753796944827199766',
- // 'https://twitter.com/yeokiloss/status/1754908226179502345',
- // 'https://twitter.com/YoungWaifLover/status/1735667278090297561',
- // 'https://twitter.com/Z_Pour_Demain/status/1766133730278605182',
- // 'https://www.aap.com.au/factcheck/aboriginal-lands-claim-a-total-abdication-of-facts/',
- // 'https://www.aap.com.au/factcheck/absurd-albanese-clip-fools-voice-voters/',
- // 'https://www.instagram.com/_the.forgotten.ones/p/CQQDyoqhsF6/',
- // 'https://www.instagram.com/p/CqSM_f9MR4b/',
- // 'https://www.instagram.com/p/CqSQgf1sv8B/',
- // 'https://instagram.com/p/B-Q22Z_pxyC/',
- // 'https://www.tiktok.com/@zitatezurzeit/photo/7342474065598319904?cid=7343316616878490400',
- // 'https://tiktok.com/@zitatezurzeit/photo/7342474065598319904?cid=7343316616878490400',
- // 'https://www.youtube.com/watch?v=rpD0qgzlCms',
- ]
- const isTruthy = (env_value) => ['1', 'yes', 'true'].includes(env_value?.toLowerCase() || 'false')
- /********************** Config: General High-Level Options ********************/
- const PASSIVE_ARCHIVING = isTruthy(process.env.PASSIVE_ARCHIVING)
- const CHROME_CLUSTER = isTruthy(process.env.CHROME_CLUSTER)
- const CHROME_CLUSTER_WORKERS = 4
- const API_SERVER_HOST = '0.0.0.0'
- const API_SERVER_PORT = 9595
- const CHROME_DEBUG_PORT = 9222 // 9222 is default, or use 0 for random port
- /********************** Config: Keys & Secrets ********************************/
- const API_KEY_2CAPTCHA = process.env.API_KEY_2CAPTCHA || 'YOUR_API_KEY_HERE'
- const FLARESOLVERR_API_ENDPOINT = process.env.FLARESOLVERR_API_ENDPOINT || "http://localhost:8191/v1"
- const ACTIVE_PERSONA = process.env.ACTIVE_PERSONA || 'Default'
- const CHROME_PROFILE_USER = process.env.CHROME_PROFILE_USER || 'Default'
- const LOAD_AUTH_STORAGE = isTruthy(process.env.LOAD_AUTH_STORAGE)
- const SAVE_AUTH_STORAGE = isTruthy(process.env.SAVE_AUTH_STORAGE)
- /********************** Config: Data Dir Locations ****************************/
- const SRC_DIR = path.resolve(__dirname)
- const DATA_DIR = process.env.DATA_DIR || await fs.promises.realpath(path.join(SRC_DIR, 'data'))
- const INDEXES_DIR = path.join(DATA_DIR, 'index')
- const ARCHIVE_DIR = path.join(DATA_DIR, 'archive')
- if (!fs.existsSync(ARCHIVE_DIR))
- throw 'Could not find data/archive, are you running in the right pwd?'
- const PERSONA_DIR = path.join(DATA_DIR, 'personas', ACTIVE_PERSONA)
- const CHROME_PROFILE_PATH = path.join(PERSONA_DIR, 'chrome_profile')
- const CHROME_DOWNLOADS_DIR = path.join(PERSONA_DIR, 'chrome_downloads')
- const CHROME_EXTENSIONS_DIR = path.join(PERSONA_DIR, 'chrome_extensions')
- const CHROME_EXTENSIONS_JSON_PATH = path.join(CHROME_EXTENSIONS_DIR, 'extensions.json')
- const AUTH_JSON_PATH = path.join(PERSONA_DIR, 'auth.json')
- const COOKIES_TXT_PATH = path.join(PERSONA_DIR, 'cookies.txt')
- const SPEEDTESTS_DIR = path.join(PERSONA_DIR, 'speedtests')
- // const CHROME_PROFILE_IMPORT_USER = 'Profile 1'
- // const CHROME_PROFILE_IMPORT_PATH = '/Volumes/NVME/Users/squash/Library/Application Support/Google/Chrome'
- // chrome profile / persona directories
- fs.mkdirSync(PERSONA_DIR, {recursive: true})
- fs.mkdirSync(SPEEDTESTS_DIR, {recursive: true})
- fs.mkdirSync(CHROME_PROFILE_PATH, {recursive: true})
- fs.mkdirSync(CHROME_EXTENSIONS_DIR, {recursive: true})
- fs.mkdirSync(CHROME_DOWNLOADS_DIR, {recursive: true})
- // cruft directories
- const ORPHANS_DIR = path.join(DATA_DIR, 'orphans')
- const PARTIALS_DIR = path.join(DATA_DIR, 'partials')
- const DUPLICATES_DIR = path.join(DATA_DIR, 'duplicates')
- await fs.promises.mkdir(ORPHANS_DIR, {recursive: true})
- await fs.promises.mkdir(PARTIALS_DIR, {recursive: true})
- await fs.promises.mkdir(DUPLICATES_DIR, {recursive: true})
- /********************** Config: Viewport Setup Opts ***************************/
- // Config: Viewport
- const DEFAULT_TIMEOUT = 20_000
- const DEFAULT_GEOLOCATION = {latitude: 59.95, longitude: 30.31667}
- const DEFAULT_USER_AGENT = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36'
- const DEFAULT_ASPECT_RAIO = 16/9 // recommended: 16:9 (most common desktop window aspect ratio)
- const SCREENSHOT_ASPECT_RATIO = 4/3 // recommended: 4:3 (easier to use as thumbnails when square-ish)
- const DEFAULT_WINDOW_WIDTH = 1920 // recommended: 1920x1080p (1080p screenshots)
- const DEFAULT_WINDOW_HEIGHT = Math.floor(DEFAULT_WINDOW_WIDTH/DEFAULT_ASPECT_RAIO)
- const DEFAULT_VIEWPORT = {
- width: DEFAULT_WINDOW_WIDTH,
- height: DEFAULT_WINDOW_HEIGHT,
- deviceScaleFactor: 2, // 2 gives much sharper text in screenshots/pdfs/etc but uses more CPU/GPU
- isMobile: false,
- hasTouch: false,
- isLandscape: false,
- }
- const DEFAULT_COLOR_SCHEME = 'light'
- const DEFAULT_HEADERS = {
- // requires frequent tweaking to remain undetected by cloudflare/recaptcha/etc.
- // 'accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7',
- // 'accept-encoding': 'gzip, deflate, br, zstd',
- // 'accept-language': accept_language,
- // 'cache-Control': no_cache ? 'no-cache' : '',
- // 'dnt': '1',
- 'sec-ch-ua': '"Google Chrome";v="122", "Not:A-Brand";v="8", "Chromium";v="122"',
- 'sec-ch-ua-mobile': '?0',
- 'sec-ch-ua-platform': '"macOS"',
- 'connection-rtt': '50',
- // 'pragma': no_cache ? 'no-cache' : '',
- // 'sec-fetch-dest': 'document',
- // 'sec-fetch-mode': 'navigate',
- // 'sec-fetch-site': 'none',
- // 'sec-fetch-user': '?1',
- // // 'upgrade-insecure-requests': '1', // breaks some sites, e.g. https://www.flowradar.com/cloneables/mouse-image-trail-effect
- // 'user-agent': user_agent,
- }
- const DEFAULT_REFERRERS = ["https://www.google.com", "https://www.facebook.com", "https://www.instagram.com"]
- /****************** Config: Human Behavior Emulation **************************/
- const SCROLL_LIMIT = 20; // e.g. 30 = 30 * (1000px/2s) => 30,000px scrolled in 60sec
- const SCROLL_DELAY = 1350; // interval per scroll, e.g. 2000 = 2sec to travel 1 * SCROLL_DISTANCE
- const SCROLL_DISTANCE = DEFAULT_VIEWPORT.height - 100; // make sure this is slightly less than viewport height so there is some overlap to make stitching easier
- /********************** Config: URL Rewriting *********************************/
- const URL_REWRITES = [
- // replacements should come first
- // {
- // idx: 0,
- // pattern: /\/\/(www\.)?x\.com/gi,
- // replacement: '//$1twitter.com/',
- // // TODO: scope: 'hostname',
- // },
- // {
- // idx: 1,
- // pattern: /\/\/(www\.)?twitter\.com/gi,
- // replacement: '//$1nitter.net',
- // // TODO: scope: 'hostname',
- // },
- // // blocks should come at the end
- // {
- // idx: 999,
- // pattern: /\/\/(www\.)?notallowed\.com/gi,
- // replacement: '',
- // // TODO: scope: 'href',
- // },
- ]
- const URL_SCHEMES_IGNORED = [
- '', // no scheme is also invalid (e.g. opening a new tab page without any url yet)
- 'chrome',
- 'chrome-extension',
- 'chrome-untrusted',
- 'file',
- 'data',
- 'about',
- ]
- /**************** Load existing data/archive/<timestamp> snapshots *************/
- const snapshots = await Snapshot.findAll({ attributes: ['id', 'timestamp', 'url'] }) // include: { model: ArchiveResult, as: 'archiveresults' }, });
- const results = await ArchiveResult.findAll({ attributes: ['id', 'snapshot_id', 'extractor', 'start_ts'] }) // include: { model: Snapshot, as: 'snapshot' }, });
- globalThis.snapshots = snapshots
- globalThis.results = results
- console.log(`[💿] Found ${snapshots.length} existing snapshots in index.sqlite3...`)
- console.log(`[💿] Found ${results.length} existing results in index.sqlite3...`)
- // debugger;
- const locateExistingSnapshots = (archive_dir) => {
- const urls_to_dirs = {}
- // for each data/archive/<timestamp>/index.json found, store {url: data/archive/<timestamp>}
- for (const snapshot_dir of fs.readdirSync(archive_dir)) {
- const snapshot_json = path.join(archive_dir, snapshot_dir, 'index.json')
- if (fs.existsSync(snapshot_json)) {
- const {url, archive_path} = JSON.parse(fs.readFileSync(snapshot_json, 'utf-8'))
- if (!snapshot_dir.includes(archive_path.replace('archive/', '')))
- throw 'Found incorrect index.json inside snapshot dir' + snapshot_dir
- if (url && url.includes('://')) {
- urls_to_dirs[url] = path.join(archive_dir, snapshot_dir)
- }
- }
- }
- return urls_to_dirs
- }
- let SNAPSHOT_DIRS_BY_URL = locateExistingSnapshots(ARCHIVE_DIR)
- let all_snap_dirs = (await fs.promises.readdir(ARCHIVE_DIR))
- // const orphan_snap_dirs = all_snap_dirs.filter(dirname => dirname.startsWith('19999'))
- // // scan through existing snapshot dirs, move orphans to orphans/ or correct archive/<snapid>
- // for (const snap_id of orphan_snap_dirs) {
- // if (snap_id.startsWith('.')) continue
- // const src_dir = path.join(ARCHIVE_DIR, snap_id)
- // let src_path = src_dir
- // assert((await fs.promises.stat(src_dir)).isDirectory())
- // let dest_path = null
- // const orphan_metrics_path = path.join(src_dir, 'metrics.json')
- // if (fs.existsSync(orphan_metrics_path)) {
- // const orphan_metrics = JSON.parse(await fs.promises.readFile(orphan_metrics_path, 'utf-8'))
- // const url = orphan_metrics.url || orphan_metrics.URL
- // const version = orphan_metrics.VERSION || versionStrFromDate(orphan_metrics.start_time)
-
- // // move all bare files into ./versions/YYYYMMDD/* and symlink ./* to latest version
- // await symlinkBestSnapshotResults(src_dir)
- // dest_path = SNAPSHOT_DIRS_BY_URL[url]
- // const dest_id = dest_path?.split('/').at(-1)
-
- // if (dest_id && (dest_id != snap_id)) {
- // if (fs.existsSync(dest_path)) {
- // console.log(` - moving duplicate snap_dir ${src_dir} -> ${dest_path}`)
- // } else {
- // console.log(` - moving valid snap_dir ${src_dir} -> ${dest_path}`)
- // }
- // } else if (dest_id == snap_id) {
- // continue
- // } else {
- // dest_path = path.join(ORPHANS_DIR, snap_id)
- // console.log(` - moving orphan snap_dir ${src_dir} -> ${dest_path}`)
- // }
- // } else {
- // // corrupt/par
- // dest_path = path.join(PARTIALS_DIR, snap_id)
- // console.log(` - moving parial snap_dir ${src_dir} -> ${dest_path}`)
- // }
- // if (dest_path) {
- // for (const version_dir of (await fs.promises.readdir(path.join(src_path, 'versions')))) {
- // const version_src = path.join(src_path, 'versions', version_dir)
- // const version_dst = path.join(dest_path, 'versions', version_dir)
- // // move all bare files into ./versions/YYYYMMDD/* and symlink ./* to latest version
- // await symlinkBestSnapshotResults(dest_path)
- // assert(!fs.existsSync(version_dst))
- // await fs.promises.rename(version_src, version_dst)
- // console.log(' - ', version_src, '--->', version_dst)
- // }
- // await fs.promises.rename(src_dir, path.join(PARTIALS_DIR, snap_id))
- // await symlinkBestSnapshotResults(dest_path)
- // }
- // }
- // const duplicate_snap_dirs = (await fs.promises.readdir(DUPLICATES_DIR)).filter(dirname => dirname.startsWith('19999'))
- // for (const snap_id of duplicate_snap_dirs) {
- // const src_dir = path.join(DUPLICATES_DIR, snap_id)
- // const metrics = JSON.parse(await fs.promises.readFile(path.join(src_dir, 'metrics.json'), 'utf-8'))
- // }
- // all_snap_dirs = (await fs.promises.readdir(ARCHIVE_DIR))
- // for (const snap_id of all_snap_dirs) {
- // if (snap_id.startsWith('.')) continue
- // const snap_dir = path.join(ARCHIVE_DIR, snap_id)
- // const metrics_path = path.join(snap_dir, 'metrics.json')
- // if (fs.existsSync(metrics_path)) {
- // // console.log(' - updating snap_dir', snap_dir)
- // await symlinkBestSnapshotResults(snap_dir)
- // }
- // }
- // SNAPSHOT_DIRS_BY_URL = locateExistingSnapshots(ARCHIVE_DIR)
- fs.writeFileSync(path.join(DATA_DIR, 'queue.csv'), '')
- const snapIdFromDir = (dir_path) =>
- dir_path.split('/archive/').at(-1)
- const snapshot_dir_list = (
- Object.entries(SNAPSHOT_DIRS_BY_URL)
- .sort(([_ak, a], [_bk, b]) =>
- Number(snapIdFromDir(b)) - Number(snapIdFromDir(a)))
- .reverse())
- for (const [existing_url, snapshot_dir] of snapshot_dir_list) {
- // if (existing_url.startsWith('https://www.facebook.com/')) {
- const is_desired_url = !(existing_url.includes('facebook.com/') || existing_url.includes('instagram.com/'))
- const already_archived = false // fs.existsSync(path.join(SNAPSHOT_DIRS_BY_URL[existing_url], 'versions'))
- if (is_desired_url && !already_archived) {
- // URLS.push(existing_url)
- fs.appendFileSync(
- path.join(DATA_DIR, 'queue.csv'),
- `${SNAPSHOT_DIRS_BY_URL[existing_url]},${existing_url}\n`,
- 'utf-8',
- )
- }
- }
- URLS = [...new Set(URLS)]
- console.log('[+] Added', URLS.length, 'existing urls to queue...')
- /********************** Config: Output Paths **********************************/
- // const TASK_PATH = (url) => path.join(DATA_DIR, 'results', `${hashCode(url)}`)
- const TASK_PATH = (url) => SNAPSHOT_DIRS_BY_URL[url] || path.join(ARCHIVE_DIR, `1999999999.${hashCode(url)}`)
- // const TASK_PATH = (url) => {
- // const existing_snap_dir = SNAPSHOT_DIRS_BY_URL[url]
- // assert(existing_snap_dir, `Could not find existing snapshot dir for ${url}`)
- // return existing_snap_dir
- // }
- const OUTPUT_PATH = (page, filename, extname='') =>
- path.join(TASK_PATH(page._original_url), `${filename}${extname}`)
- const SSL_PATH = (page) => OUTPUT_PATH(page, 'ssl.json')
- const CONSOLELOG_PATH = (page) => OUTPUT_PATH(page, 'console.log')
- const HEADERS_PATH = (page) => OUTPUT_PATH(page, 'headers.json')
- const REDIRECTS_PATH = (page) => OUTPUT_PATH(page, 'redirects.json')
- const REQUESTS_PATH = (page) => OUTPUT_PATH(page, 'requests.json')
- const TRACE_PATH = (page) => OUTPUT_PATH(page, 'trace.json')
- const METRICS_PATH = (page) => OUTPUT_PATH(page, 'metrics.json')
- const OUTLINKS_PATH = (page) => OUTPUT_PATH(page, 'outlinks.json')
- const SEO_PATH = (page) => OUTPUT_PATH(page, 'seo.json')
- const FAVICON_PATH = (page) => OUTPUT_PATH(page, 'favicon.json')
- const TITLE_PATH = (page) => OUTPUT_PATH(page, 'title.txt')
- const BODYTEXT_PATH = (page) => OUTPUT_PATH(page, 'body.txt')
- const PANDOC_PATH = (page) => OUTPUT_PATH(page, 'pandoc.md')
- const READABILITY_PATH = (page) => OUTPUT_PATH(page, 'readability.json')
- const ACCESIBILITY_PATH = (page) => OUTPUT_PATH(page, 'accessibility.json')
- const DOM_PATH = (page) => OUTPUT_PATH(page, 'dom.html')
- const PDF_PATH = (page) => OUTPUT_PATH(page, 'output.pdf')
- const SCREENSHOT_PATH = (page) => OUTPUT_PATH(page, 'screenshot.png')
- const SCREENSHOT_JPG_PATH = (page) => OUTPUT_PATH(page, 'screenshot.jpg')
- const AIQA_PATH = (page) => OUTPUT_PATH(page, 'aiqa.json')
- const SINGLEFILE_PATH = (page) => OUTPUT_PATH(page, 'singlefile.html')
- const YTDLP_PATH = (page) => OUTPUT_PATH(page, 'media/')
- const GALLERYDL_PATH = (page) => OUTPUT_PATH(page, 'photos/')
- const SCREENRECORDING_PATH = (page) => OUTPUT_PATH(page, 'screenrecording.mp4')
- const SCREENRECORDGIF_PATH = (page) => OUTPUT_PATH(page, 'screenrecording.gif')
- const RESPONSES_PATH = (page) => OUTPUT_PATH(page, 'responses')
- const RAW_PATH = (page) => OUTPUT_PATH(page, 'raw')
- /********************** Config: Chrome Extensions *****************************/
- interface ChromeExtension {
- name: string
- webstore_id: string
- }
- interface LoadedChromeExtension extends ChromeExtension {
- id?: string
- webstore_url?: string
- crx_url?: string
- crx_path?: string
- unpacked_path?: string
- read_manifest?: () => any
- read_version?: () => string | null
- }
- const CHROME_EXTENSIONS: LoadedChromeExtension[] = [
- // Content access / unblocking / blocking plugins
- {webstore_id: 'ifibfemgeogfhoebkmokieepdoobkbpo', name: 'twocaptcha'}, // https://2captcha.com/blog/how-to-use-2captcha-solver-extension-in-puppeteer
- {webstore_id: 'edibdbjcniadpccecjdfdjjppcpchdlm', name: 'istilldontcareaboutcookies'},
- {webstore_id: 'cjpalhdlnbpafiamejdnhcphjbkeiagm', name: 'ublock'},
- // {webstore_id: 'mlomiejdfkolichcflejclcbmpeaniij', name: 'ghostery'},
- // {webstore_id: 'mnjggcdmjocbbbhaepdhchncahnbgone', name: 'sponsorblock'},
- // {webstore_id: 'iplffkdpngmdjhlpjmppncnlhomiipha', name: 'unpaywall'},
- // {webstore_id: 'gofocbepaccnkpphbgjpolififgcakhn', name: 'spaywallnews'},
-
- // Archiving plugins
- {webstore_id: 'mpiodijhokgodhhofbcjdecpffjipkle', name: 'singlefile'},
- // {webstore_id: 'fpeoodllldobpkbkabpblcfaogecpndd', name: 'archivewebpage'},
- // {webstore_id: 'niloccemoadcdkdjlinkgdfekeahmflj', name: 'pocket'},
- // {webstore_id: 'kenncghfghgolcbmckhiljgaabnpcaaa', name: 'warcreate'},
- // {webstore_id: 'jjndjgheafjngoipoacpjgeicjeomjli', name: 'puppeteerstream'},
- // Utilities for humans setting up/viewing/debugging the archiving session
- // {webstore_id: 'aeblfdkhhhdcdjpifhhbdiojplfjncoa', name: '1password'},
- // {webstore_id: 'fngmhnnpilhplaeedifhccceomclgfbg', name: 'editthiscookie'},
- // {webstore_id: 'cgfpgnepljlgenjclbekbjdlgcodfmjp', name: 'simpletabsorter'},
-
- // Scripting/automation plugins
- // {webstore_id: 'jinjaccalgkegednnccohejagnlnfdag', name: 'violentmonkey'},
- // {webstore_id: 'infppggnoaenmfagbfknfkancpbljcca', name: 'automa'},
- // {webstore_id: 'pfegffhjcgkneoemnlniggnhkfioidjg', name: 'screenscraper'},
- ]
- /******************** Config: Chrome Profile Preferences **********************/
- // https://niek.github.io/chrome-features/
- const CHROME_DISABLED_COMPONENTS = [
- 'Translate',
- 'AcceptCHFrame',
- 'OptimizationHints',
- 'ProcessPerSiteUpToMainFrameThreshold',
- 'InterestFeedContentSuggestions',
- 'CalculateNativeWinOcclusion',
- 'BackForwardCache',
- 'HeavyAdPrivacyMitigations',
- 'LazyFrameLoading',
- 'ImprovedCookieControls',
- 'PrivacySandboxSettings4',
- 'AutofillServerCommunication',
- 'CertificateTransparencyComponentUpdater',
- 'DestroyProfileOnBrowserClose',
- 'CrashReporting',
- 'OverscrollHistoryNavigation',
- 'InfiniteSessionRestore',
- //'LockProfileCookieDatabase', // disabling allows multiple chrome instances to concurrently modify profile, but might make chrome much slower https://github.com/yt-dlp/yt-dlp/issues/7271 https://issues.chromium.org/issues/40901624
- ]
- const CHROME_PREFERENCES_EXTRA = {}
- const CHROME_PREFERENCES_DEFAULT = {
- // https://chromium.googlesource.com/chromium/src/+/32352ad08ee673a4d43e8593ce988b224f6482d3/chrome/common/pref_names.cc
- homepage: 'about:blank', // doesn't work here, managed by Secure Preferences
- homepage_is_newtabpage: false, // doesn't work here, managed by Secure Preferences
- session: { // doesn't work here, managed by Secure Preferences
- restore_on_startup: 4, // doesn't work here, managed by Secure Preferences
- startup_urls: 'about:blank', // doesn't work here, managed by Secure Preferences
- },
- default_apps: 'noinstall',
- browser: {
- confirm_to_quit: false,
- enable_spellchecking: false,
- check_default_browser: false,
- show_update_promotion_info_bar: false,
- },
- profile: {
- // name: 'ArchiveBox Persona: Default', // doesnt work to change display name, not sure why
- // using_default_name: false,
- exited_cleanly: true,
- default_content_setting_values: {
- automatic_downloads: 1,
- },
- },
- bookmark_bar: {show_on_all_tabs: false},
- safebrowsing: {enabled: false},
- search: {suggest_enabled: false},
- download: {
- prompt_for_download: false,
- open_pdf_in_system_reader: true,
- // default_directory: CHROME_DOWNLOADS_DIR || path.join(__dirname, 'downloads'),
- },
- select_file_dialogs: {allowed: false},
- autofill: {save_data: false},
- printing: {enabled: false},
- message_center: {welcome_notification_dismissed_local: true},
- extensions: {
- ui: {
- developer_mode: true,
- dismissed_adt_promo: true,
- },
- // pinned_extensions: CHROME_EXTENSIONS?.map(({id}) => id) || [],
- },
- webkit: {
- webprefs: {
- javascript_enabled: true,
- minimum_font_size: 9,
- // default_font_size: 12,
- // web_security_enabled: false,
- // allow_displaying_insecure_content: true,
- // allow_running_insecure_content: true,
- java_enabled: true,
- loads_images_automatically: true,
- },
- },
- settings: {
- multi_profile_never_show_intro: true,
- multi_profile_warning_show_dismissed: true,
- first_run_tutorial_shown: true,
- },
- plugins: {
- always_open_pdf_externally: true,
- },
- }
- const CHROME_PREFERENCES_PATH = path.join(CHROME_PROFILE_PATH, 'Default', 'Preferences')
- const getChromePreferences = ({CHROME_PREFERENCES_DEFAULT, CHROME_PREFERENCES_EXTRA, CHROME_EXTENSIONS, CHROME_DOWNLOADS_DIR}) =>
- merge.all([CHROME_PREFERENCES_DEFAULT, CHROME_PREFERENCES_EXTRA, {
- extensions: {
- pinned_extensions: CHROME_EXTENSIONS?.map(({id}) => id) || [],
- },
- download: {
- default_directory: CHROME_DOWNLOADS_DIR || path.join(__dirname, 'downloads'),
- },
- }])
- function applyChromePreferences(puppeteer, prefs_path, preferences) {
- if (fs.existsSync(prefs_path)) {
- const preferences_existing = JSON.parse(fs.readFileSync(prefs_path, 'utf-8'))
- const preferences_merged = merge(preferences_existing, preferences)
- // console.log(JSON.stringify(preferences_merged, null, 4))
- fs.writeFileSync(prefs_path, JSON.stringify(preferences_merged))
- } else {
- // otherwise profile has not been created yet, use plugin instead (plugin only works on first creation)
- puppeteer.use(PrefsPlugin({userPrefs: preferences}))
- }
- return puppeteer
- }
- /******************** Config: Chrome Launch Args ******************************/
- const CHROME_ARGS_DEFAULT = [
- // Headless behavior tuning, determinstic behavior settings
- // '--headless=new',
- '--test-type',
- '--test-type=gpu', // https://github.com/puppeteer/puppeteer/issues/10516
- '--deterministic-mode',
- '--js-flags=--random-seed=1157259159', // make all JS random numbers deterministic by providing a seed
- '--allow-pre-commit-input', // allow JS mutations before page rendering is complete
- '--disable-blink-features=AutomationControlled', // hide the signatures that announce browser is being remote-controlled
- '--enable-automation', // <- DONT USE THIS, it makes you easily detectable / blocked by cloudflare
- // `--proxy-server=https://43.159.28.126:2334:u7ce652b7568805c4-zone-custom-region-us-session-szGWq3FRU-sessTime-60:u7ce652b7568805c4`, // send all network traffic through a proxy https://2captcha.com/proxy
- // `--proxy-bypass-list=127.0.0.1`,
- // Docker-specific options
- // https://github.com/GoogleChrome/lighthouse-ci/tree/main/docs/recipes/docker-client#--no-sandbox-issues-explained
- // '--no-sandbox', // rely on docker sandboxing in docker, otherwise we need cap_add: SYS_ADM to use host sandboxing
- // '--disable-gpu-sandbox',
- // '--disable-setuid-sandbox',
- // '--disable-dev-shm-usage', // docker 75mb default shm size is not big enough, disabling just uses /tmp instead
- // '--no-xshm',
- // Profile data dir setup
- // chrome://profile-internals
- `--user-data-dir=${CHROME_PROFILE_PATH}`,
- `--profile-directory=${CHROME_PROFILE_USER}`,
- '--password-store=basic', // use mock keychain instead of OS-provided keychain (we manage auth.json instead)
- '--use-mock-keychain',
- '--disable-cookie-encryption', // we need to be able to write unencrypted cookies to save/load auth.json
- // '--disable-sync', // don't try to use Google account sync features
- // Extensions
- // chrome://inspect/#extensions
- // `--load-extension=${CHROME_EXTENSIONS.map(({unpacked_path}) => unpacked_path).join(',')}`, // not needed when using existing profile that already has extensions installed
- `--allowlisted-extension-id=${CHROME_EXTENSIONS.map(({ webstore_id }) => webstore_id).join(',')}`,
- '--allow-legacy-extension-manifests',
- // Browser window and viewport setup
- // chrome://version
- // `--user-agent="${DEFAULT_USER_AGENT}"`,
- // `--window-size=${DEFAULT_VIEWPORT.width},${DEFAULT_VIEWPORT.height}`,
- '--window-position=0,0',
- '--hide-scrollbars', // hide scrollbars because otherwise they show up in screenshots
- '--install-autogenerated-theme=169,32,85', // red border makes it easier to see which chrome window is archivebox's
- '--autoplay-policy=no-user-gesture-required', // auto-start videos so they trigger network requests + show up in outputs
- '--disable-gesture-requirement-for-media-playback',
- '--lang=en-US,en;q=0.9',
- // DANGER: JS isolation security features (to allow easier tampering with pages during archiving)
- // chrome://net-internals
- // '--disable-web-security', // <- WARNING, breaks some sites that expect/enforce strict CORS headers (try webflow.com)
- // '--disable-features=IsolateOrigins,site-per-process', // useful for injecting JS, but some very strict sites can panic / show error pages when isolation is disabled (e.g. webflow.com)
- // '--allow-running-insecure-content', // Breaks CORS/CSRF/HSTS etc., useful sometimes but very easy to detect
- // '--allow-file-access-from-files', // <- WARNING, dangerous, allows JS to read filesystem using file:// URLs
- // // DANGER: Disable HTTPS verification
- // '--ignore-certificate-errors',
- // '--ignore-ssl-errors',
- // '--ignore-certificate-errors-spki-list',
- // '--allow-insecure-localhost',
- // IO: stdin/stdout, debug port config
- // chrome://inspect
- '--log-level=2', // 1=DEBUG 2=WARNING 3=ERROR
- '--enable-logging=stderr',
- '--remote-debugging-address=0.0.0.0',
- `--remote-debugging-port=${CHROME_DEBUG_PORT}`,
- // GPU, canvas, text, and pdf rendering config
- // chrome://gpu
- '--enable-webgl', // enable web-gl graphics support
- '--font-render-hinting=none', // make rendering more deterministic by ignoring OS font hints, may also need css override, try: * {text-rendering: geometricprecision !important; -webkit-font-smoothing: antialiased;}
- '--force-color-profile=srgb', // make rendering more deterministic by using consitent color profile, if browser looks weird, try: generic-rgb
- '--disable-partial-raster', // make rendering more deterministic (TODO: verify if still needed)
- '--disable-skia-runtime-opts', // make rendering more deterministic by avoiding Skia hot path runtime optimizations
- '--disable-2d-canvas-clip-aa', // make rendering more deterministic by disabling antialiasing on 2d canvas clips
- // '--disable-gpu', // falls back to more consistent software renderer
- // // '--use-gl=swiftshader', <- DO NOT USE, breaks M1 ARM64. it makes rendering more deterministic by using simpler CPU renderer instead of OS GPU renderer bug: https://groups.google.com/a/chromium.org/g/chromium-dev/c/8eR2GctzGuw
- // // '--disable-software-rasterizer', <- DO NOT USE, harmless, used in tandem with --disable-gpu
- // // '--run-all-compositor-stages-before-draw', <- DO NOT USE, makes headful chrome hang on startup (tested v121 Google Chrome.app on macOS)
- // // '--disable-gl-drawing-for-tests', <- DO NOT USE, disables gl output (makes tests run faster if you dont care about canvas)
- // // '--blink-settings=imagesEnabled=false', <- DO NOT USE, disables images entirely (only sometimes useful to speed up loading)
- // Process management & performance tuning
- // chrome://process-internals
- '--disable-lazy-loading', // make rendering more deterministic by loading all content up-front instead of on-focus
- '--disable-renderer-backgrounding', // dont throttle tab rendering based on focus/visibility
- '--disable-background-networking', // dont throttle tab networking based on focus/visibility
- '--disable-background-timer-throttling', // dont throttle tab timers based on focus/visibility
- '--disable-backgrounding-occluded-windows', // dont throttle tab window based on focus/visibility
- '--disable-ipc-flooding-protection', // dont throttle ipc traffic or accessing big request/response/buffer/etc. objects will fail
- '--disable-extensions-http-throttling', // dont throttle http traffic based on runtime heuristics
- '--disable-field-trial-config', // disable shared field trial state between browser processes
- '--disable-back-forward-cache', // disable browsing navigation cache
- // '--in-process-gpu', <- DONT USE THIS, makes headful startup time ~5-10s slower (tested v121 Google Chrome.app on macOS)
- // '--disable-component-extensions-with-background-pages', // TODO: check this, disables chrome components that only run in background (could lower startup time)
- // uncomment to disable hardware camera/mic/speaker access + present fake devices to websites
- // (faster to disable, but disabling breaks recording browser audio in puppeteer-stream screenrecordings)
- // '--use-fake-device-for-media-stream',
- // '--use-fake-ui-for-media-stream',
- // '--disable-features=GlobalMediaControls,MediaRouter,DialMediaRouteProvider',
-
- // // Output format options (PDF, screenshot, etc.)
- '--export-tagged-pdf', // include table on contents and tags in printed PDFs
- '--generate-pdf-document-outline',
- // Suppress first-run features, popups, hints, updates, etc.
- // chrome://system
- '--no-pings',
- '--no-first-run',
- '--no-default-browser-check',
- '--disable-default-apps',
- '--ash-no-nudges',
- '--disable-infobars',
- '--disable-search-engine-choice-screen',
- '--disable-session-crashed-bubble',
- '--simulate-outdated-no-au="Tue, 31 Dec 2099 23:59:59 GMT"',
- '--hide-crash-restore-bubble',
- '--suppress-message-center-popups',
- '--disable-client-side-phishing-detection',
- '--disable-domain-reliability',
- '--disable-component-update',
- '--disable-datasaver-prompt',
- '--disable-hang-monitor',
- '--disable-session-crashed-bubble',
- '--disable-speech-synthesis-api',
- '--disable-speech-api',
- '--disable-print-preview',
- '--safebrowsing-disable-auto-update',
- '--deny-permission-prompts',
- '--disable-external-intent-requests',
- '--disable-notifications',
- '--disable-desktop-notifications',
- '--noerrdialogs',
- '--disable-popup-blocking',
- '--disable-prompt-on-repost',
- '--silent-debugger-extension-api',
- '--block-new-web-contents',
- '--metrics-recording-only',
- '--disable-breakpad',
-
- // other feature flags
- // chrome://flags chrome://components
- `--disable-features=${CHROME_DISABLED_COMPONENTS.join(',')}`,
- '--enable-features=NetworkService',
- ]
- const CHROME_ARGS_EXTRA = []
- const CHROME_LAUNCH_OPTIONS = {
- CHROME_PROFILE_PATH,
- CHROME_PROFILE_USER,
- CHROME_EXTENSIONS,
- CHROME_DEBUG_PORT,
- CHROME_DISABLED_COMPONENTS,
- DEFAULT_VIEWPORT,
- CHROME_ARGS_DEFAULT,
- CHROME_ARGS_EXTRA,
- }
- /* Chrome CLI Args Documentation
- - https://github.com/GoogleChrome/chrome-launcher/blob/main/docs/chrome-flags-for-tools.md
- - https://chromium.googlesource.com/chromium/chromium/+/master/content/public/common/content_switches.cc
- - https://jtway.co/optimize-your-chrome-options-for-testing-to-get-x1-25-impact-4f19f071bf45
- - https://peter.sh/experiments/chromium-command-line-switches/
- - https://www.chromium.org/developers/how-tos/run-chromium-with-flags/
- - https://github.com/manoj9788/Chrome-Driver-arguments/blob/master/README.md
- */
- const getChromeArgs = ({CHROME_ARGS_DEFAULT, CHROME_ARGS_EXTRA,
- CHROME_PROFILE_PATH, CHROME_PROFILE_USER,
- CHROME_EXTENSIONS,
- CHROME_DEBUG_PORT,
- CHROME_DISABLED_COMPONENTS,
- DEFAULT_VIEWPORT}=CHROME_LAUNCH_OPTIONS) =>
- [
- ...CHROME_ARGS_DEFAULT,
- `--user-data-dir=${CHROME_PROFILE_PATH}`,
- `--profile-directory=${CHROME_PROFILE_USER}`,
- `--load-extension=${CHROME_EXTENSIONS.map(({unpacked_path}) => unpacked_path).join(',')}`,
- `--allowlisted-extension-id=${CHROME_EXTENSIONS.map(({id}) => id).join(',')}`,
- `--window-size=${DEFAULT_VIEWPORT.width},${DEFAULT_VIEWPORT.height}`,
- `--remote-debugging-port=${CHROME_DEBUG_PORT}`,
- `--disable-features=${CHROME_DISABLED_COMPONENTS.join(',')}`,
- ...CHROME_ARGS_EXTRA,
- ]
- /******************** Chrome Extension Management *****************************/
- function getExtensionId(unpacked_path) {
- const manifest_path = path.join(unpacked_path, 'manifest.json')
- if (!fs.existsSync(manifest_path)) return null
- // chrome uses a SHA256 hash of the unpacked extension directory path to compute a dynamic id
- const hash = crypto.createHash('sha256');
- hash.update(Buffer.from(unpacked_path, 'utf-8'));
- const detected_extension_id = Array.from(hash.digest('hex'))
- .slice(0, 32) // Convert each hexadecimal character to a character in the range 'a'-'p'
- .map(i => String.fromCharCode(parseInt(i, 16) + 'a'.charCodeAt(0)))
- .join('');
- return detected_extension_id
- }
- async function installExtension(extension) {
- const manifest_path = path.join(extension.unpacked_path, 'manifest.json')
- // Download extensions using:
- // curl -fsSL 'https://clients2.google.com/service/update2/crx?response=redirect&prodversion=1230&acceptformat=crx3&x=id%3D$EXTENSION_ID%26uc' > extensionname.crx
- // unzip -d extensionname extensionname.zip
- if (!fs.existsSync(manifest_path) && !fs.existsSync(extension.crx_path)) {
- console.log("[🛠️] Downloading missing extension", extension.name, extension.webstore_id, '->', extension.crx_path);
- // Download crx file from ext.crx_url -> ext.crx_path
- const response = await fetch(extension.crx_url) as Response
- const crx_file = fs.createWriteStream(extension.crx_path);
- if (response.headers.get("content-length") && response.body) {
- // @ts-ignore
- const crx_stream = Readable.fromWeb(response.body)
- await finished(crx_stream.pipe(crx_file))
- } else {
- console.warn('[⚠️] Failed to download extension', extension.name, extension.webstore_id)
- }
- }
-
- var {stdout, stderr} = {stdout: '', stderr: ''}
- // Unzip crx file from ext.crx_url -> ext.unpacked_path
- await fs.promises.mkdir(extension.unpacked_path, {recursive: true})
- try {
- var {stdout, stderr} = await exec(`/usr/bin/unzip ${extension.crx_path} -d ${extension.unpacked_path}`)
- } catch(err1) {
- try {
- await unzip(extension.crx_path, extension.unpacked_path)
- } catch(err2) {
- // console.error(`[❌] Failed to install ${extension.crx_path}: could not unzip crx`, err1, err2)
- // return false
- }
- }
- if (!fs.existsSync(manifest_path))
- console.error(`[❌] Failed to install ${extension.crx_path}: could not find manifest.json in unpacked_path`, stdout, stderr)
- return fs.existsSync(manifest_path)
- }
- async function loadOrInstallExtension(ext) {
- if (!(ext.webstore_id || ext.unpacked_path))
- throw 'Extension must have either {webstore_id} or {unpacked_path}'
- // Set statically computable extension metadata
- ext.webstore_id = ext.webstore_id || ext.id
- ext.name = ext.name || ext.webstore_id
- ext.webstore_url = ext.webstore_url || `https://chromewebstore.google.com/detail/${ext.webstore_id}`
- ext.crx_url = ext.crx_url || `https://clients2.google.com/service/update2/crx?response=redirect&prodversion=1230&acceptformat=crx3&x=id%3D${ext.webstore_id}%26uc`
- ext.crx_path = ext.crx_path || path.join(CHROME_EXTENSIONS_DIR, `${ext.webstore_id}__${ext.name}.crx`)
- ext.unpacked_path = ext.unpacked_path || path.join(CHROME_EXTENSIONS_DIR, `${ext.webstore_id}__${ext.name}`)
-
- const manifest_path = path.join(ext.unpacked_path, 'manifest.json')
- ext.read_manifest = () => JSON.parse(fs.readFileSync(manifest_path, 'utf-8'))
- ext.read_version = () => fs.existsSync(manifest_path) && ext.read_manifest()?.version || null
- // if extension is not installed, download and unpack it
- if (!ext.read_version()) {
- await installExtension(ext)
- }
- // autodetect id from filesystem path (unpacked extensions dont have stable IDs)
- ext.id = getExtensionId(ext.unpacked_path)
- ext.version = ext.read_version()
- if (!ext.version) {
- console.warn('[❌] Unable to detect ID and version of installed extension', prettyPath(ext.unpacked_path))
- } else {
- console.log(`[➕] Installed extension ${ext.name} (${ext.version})...`.padEnd(82), prettyPath(ext.unpacked_path))
- }
- return ext
- }
- async function isTargetExtension(target) {
- let target_type
- let target_ctx
- let target_url
- try {
- target_type = target.type()
- target_ctx = (await target.worker()) || (await target.page()) || null
- target_url = target.url() || target_ctx?.url() || null
- } catch(err) {
- if (String(err).includes('No target with given id found')) {
- // because this runs on initial browser startup, we sometimes race with closing the initial
- // new tab page. it will throw a harmless error if we try to check a target that's already closed,
- // ignore it and return null since that page is definitely not an extension's bg page anyway
- target_type = 'closed'
- target_ctx = null
- target_url = 'about:closed'
- } else {
- throw err
- }
- }
-
- const target_is_bg = ['service_worker', 'background_page'].includes(target_type)
- const target_is_extension = target_url?.startsWith('chrome-extension://')
- const extension_id = (target_is_extension && target_url.split('://')[1].split('/')[0]) || null
- const manifest_version = target_type === 'service_worker' ? '3' : '2'
- return {
- target_type,
- target_ctx,
- target_url,
- target_is_bg,
- target_is_extension,
- extension_id,
- manifest_version,
- }
- }
- async function loadExtensionFromTarget(extensions, target) {
- const {
- target_is_bg,
- target_is_extension,
- target_type,
- target_ctx,
- target_url,
- extension_id,
- manifest_version,
- } = await isTargetExtension(target)
-
- if (!(target_is_bg && extension_id && target_ctx))
- return null
- const manifest = await target_ctx.evaluate(() =>
- // @ts-ignore
- chrome.runtime.getManifest())
-
- const { name, version, homepage_url, options_page, options_ui } = manifest
- if (!version || !extension_id)
- return null
-
- const options_url = await target_ctx.evaluate(
- (options_page) => chrome.runtime.getURL(options_page),
- options_page || options_ui?.page || 'options.html',
- )
-
- const commands = await target_ctx.evaluate(async () =>
- (await new Promise((resolve, reject) => {
- if (chrome.commands)
- chrome.commands.getAll(resolve)
- else
- resolve({})
- }))
- )
- // console.log(`[+] Found Manifest V${manifest_version} Extension:`, extension_id, name, target_url, Object.keys(commands).length)
- let dispatchEval = async (...args) =>
- await target_ctx.evaluate(...args)
- let dispatchPopup = async () =>
- await target_ctx.evaluate('chrome.action?.openPopup() || chrome.tabs.create({url: chrome.runtime.getURL("popup.html")})')
-
- let dispatchAction
- let dispatchMessage
- let dispatchCommand
- if (manifest_version === '3') {
- dispatchAction = async (tab) => {
- // https://developer.chrome.com/docs/extensions/reference/api/action#event-onClicked
- return await target_ctx.evaluate(async (tab) => {
- tab = tab || (await new Promise((resolve) =>
- chrome.tabs.query({currentWindow: true, active: true}, ([tab]) => resolve(tab))))
- // @ts-ignore
- return await chrome.action.onClicked.dispatch(tab)
- }, tab)
- }
- dispatchMessage = async (message, options) => {
- // https://developer.chrome.com/docs/extensions/reference/api/runtime
- return await target_ctx.evaluate(async (extension_id, message, options) => {
- return await chrome.runtime.sendMessage(extension_id, message, options)
- }, extension_id, message, options)
- }
- dispatchCommand = async (command, tab) => {
- // https://developer.chrome.com/docs/extensions/reference/api/commands#event-onCommand
- return await target_ctx.evaluate(async (command, tab) => {
- // @ts-ignore
- return await chrome.commands.onCommand.dispatch(command, tab)
- }, command, tab)
- }
- } else if (manifest_version === '2') {
- dispatchAction = async (tab) => {
- // https://developer.chrome.com/docs/extensions/mv2/reference/browserAction#event-onClicked
- return await target_ctx.evaluate(async (tab) => {
- tab = tab || (await new Promise((resolve) =>
- chrome.tabs.query({currentWindow: true, active: true}, ([tab]) => resolve(tab))))
- // @ts-ignore
- return await chrome.browserAction.onClicked.dispatch(tab)
- }, tab)
- }
- dispatchMessage = async (message, options) => {
- // https://developer.chrome.com/docs/extensions/mv2/reference/runtime#method-sendMessage
- return await target_ctx.evaluate(async (extension_id, message, options) => {
- return await new Promise((resolve) =>
- chrome.runtime.sendMessage(extension_id, message, options, resolve)
- )
- }, extension_id, message, options)
- }
- dispatchCommand = async (command, tab) => {
- // https://developer.chrome.com/docs/extensions/mv2/reference/commands#event-onCommand
- return await target_ctx.evaluate(async (command, tab) => {
- return await new Promise((resolve) =>
- // @ts-ignore
- chrome.commands.onCommand.dispatch(command, tab, resolve)
- )
- }, command, tab)
- }
- }
- const existing_extension = extensions.filter(({id}) => id === extension_id)[0] || {}
- const new_extension = {
- ...existing_extension,
- id: extension_id,
- webstore_name: name,
- target,
- target_ctx,
- target_type,
- target_url,
- manifest_version,
- manifest,
- version,
- homepage_url,
- options_url,
- dispatchEval, // run some JS in the extension's service worker context
- dispatchPopup, // open the extension popup
- dispatchAction, // trigger an extension menubar icon click
- dispatchMessage, // send a chrome runtime message in the service worker context
- dispatchCommand, // trigger an extension keyboard shortcut command
- }
- console.log(`[➕] Loaded extension ${name.substring(0, 32)} (${version}) ${target_type}...`.padEnd(82), target_url)
- Object.assign(existing_extension, new_extension)
- return new_extension
- }
- async function getChromeExtensionsFromPersona({CHROME_EXTENSIONS, CHROME_EXTENSIONS_DIR}) {
- console.log('*************************************************************************')
- console.log(`[⚙️] Installing ${CHROME_EXTENSIONS.length} chrome extensions from CHROME_EXTENSIONS...`)
- try {
- // read extension metadata from filesystem (installing from Chrome webstore if extension is missing)
- for (const extension of CHROME_EXTENSIONS) {
- Object.assign(extension, await loadOrInstallExtension(extension))
- }
- // for easier debugging, write parsed extension info to filesystem
- await overwriteFile(
- CHROME_EXTENSIONS_JSON_PATH.replace('.json', '.present.json'),
- CHROME_EXTENSIONS,
- )
- } catch(err) {
- console.error(err)
- }
- console.log('*************************************************************************')
- return CHROME_EXTENSIONS
- }
- let _EXTENSIONS_CACHE = null
- async function getChromeExtensionsFromCache({browser, extensions=CHROME_EXTENSIONS, extensions_dir=CHROME_EXTENSIONS_DIR}) {
- if (_EXTENSIONS_CACHE === null) {
- console.log(`[⚙️] Loading ${CHROME_EXTENSIONS.length} chrome extensions from CHROME_EXTENSIONS...`)
- // find loaded Extensions at runtime / browser launch time & connect handlers
- // looks at all the open targets for extension service workers / bg pages
- for (const target of browser.targets()) {
- // mutates extensions object in-place to add metadata loaded from filesystem persona dir
- await loadExtensionFromTarget(extensions, target)
- }
- _EXTENSIONS_CACHE = extensions
- // write installed extension metadata to filesystem extensions.json for easier debugging
- await overwriteFile(
- CHROME_EXTENSIONS_JSON_PATH.replace('.json', '.loaded.json'),
- extensions,
- )
- await overwriteSymlink(
- CHROME_EXTENSIONS_JSON_PATH.replace('.json', '.loaded.json'),
- CHROME_EXTENSIONS_JSON_PATH,
- )
- }
-
- return _EXTENSIONS_CACHE
- }
- async function setup2CaptchaExtension({browser, extensions}) {
- let page = null
- try {
- // open a new tab to finish setting up the 2captcha extension manually using its extension options page
- page = await browser.newPage()
- const { options_url } = extensions.filter(ext => ext.name === 'twocaptcha')[0]
- await page.goto(options_url)
- await wait(2_500)
- await page.bringToFront()
-
- // type in the API key and click the Login button (and auto-close success modal after it pops up)
- await page.evaluate(() => {
- const elem = document.querySelector("input[name=apiKey]") as HTMLInputElement
- elem.value = ""
- })
- await page.type('input[name=apiKey]', API_KEY_2CAPTCHA, { delay: 25 })
-
- // toggle all the important switches to ON
- await page.evaluate(() => {
- const checkboxes = Array.from(document.querySelectorAll<HTMLInputElement>('input#isPluginEnabled, input[name*=enabledFor], input[name*=autoSolve]'));
- for (const checkbox of checkboxes) {
- if (!checkbox.checked) checkbox.click()
- }
- })
-
- let dialog_opened = false
- page.on('dialog', async (dialog) => {
- setTimeout(async () => {
- await dialog.accept();
- dialog_opened = true
- }, 500);
- })
- await page.click('button#connect')
- await wait(2_500)
- if (!dialog_opened) {
- throw `2captcha extension login confirmation dialog never opened, please check its options page manually: ${options_url}`
- }
- console.log('[🔑] Configured the 2captcha extension using its options page...')
- } catch(err) {
- console.warn(`[❌] Failed to configure the 2captcha extension using its options page!`, err)
- }
- if (page) await page.close()
- }
- async function speedtest({browser, page, measureUpload=true, timeout=25000}: {browser?: Browser, page?: Page, measureUpload?: boolean, timeout?: number}) {
- // run a speedtest using fast.com, printing results once per second
- browser = browser || await page.browser()
- page = page || await browser.newPage()
- // save one speedtest_<date>.json result per day
- const today = versionStrFromDate(new Date(), {withDate: true, withTime: false})
- const SPEEDTEST_PATH = path.join(SPEEDTESTS_DIR, `speedtest_${today}.json`)
- // check if we've already run one today, if so return earlier results and skip running again
- try {
- return JSON.parse(await fs.promises.readFile(SPEEDTEST_PATH, 'utf-8'))
- } catch(err) {
- // otherwise speedtest does not exist yet for today, continue onwards...
- }
- console.log('[🚤] Running Speedtest using Fast.com...'.padEnd(82), prettyPath(SPEEDTEST_PATH))
- await page.goto('https://fast.com', {timeout, waitUntil: 'domcontentloaded'});
- await page.waitForSelector('#speed-value', {timeout})
- let result = null
- let loop_idx = 0
- while (loop_idx < 100) {
- result = await page.evaluate(() => {
- const $ = document.querySelector.bind(document);
- return {
- downloadSpeed: Number($('#speed-value').textContent),
- downloadUnit: $('#speed-units').textContent.trim(),
- downloaded: Number($('#down-mb-value').textContent.trim()),
- uploadSpeed: Number($('#upload-value').textContent),
- uploadUnit: $('#upload-units').textContent.trim(),
- uploaded: Number($('#up-mb-value').textContent.trim()),
- latency: Number($('#latency-value').textContent.trim()),
- bufferBloat: Number($('#bufferbloat-value').textContent.trim()),
- userLocation: $('#user-location').textContent.trim(),
- userIp: $('#user-ip').textContent.trim(),
- isDone: Boolean($('#speed-value.succeeded') && $('#upload-value.succeeded')),
- };
- })
- if (result.downloadSpeed > 0) {
- // console.log(JSON.stringify(result).replaceAll('"', '').replaceAll(',', ' ').replaceAll('{', '').replaceAll('}', ''))
- }
- if (result.isDone || (!measureUpload && result.uploadSpeed)) {
- break
- }
- await wait(500)
- loop_idx++
- }
- await Promise.allSettled([
- page.close(),
- overwriteFile(SPEEDTEST_PATH, result)
- ])
- return result
- }
- /******************************************************************************/
- /******************************************************************************/
- const ALREADY_ARCHIVED = new Set(['', 'about:blank', 'chrome://newtab', 'chrome://version'])
- const TASKS_PER_RUN_LIMIT = 200
- async function botArchiveTask({page, data, url=''}) {
- url = url || data // puppeteer-cluster passes in the url value via the data: arg
- const is_unarchivable_url = URL_SCHEMES_IGNORED.includes(url.split(':')[0])
- const is_already_archived = ALREADY_ARCHIVED.has(url.slice(0, 4096))
- if (is_unarchivable_url || is_already_archived) return null
- ALREADY_ARCHIVED.add(url.slice(0, 4096))
- if (ALREADY_ARCHIVED.size > TASKS_PER_RUN_LIMIT) {
- console.warn('[❌] Hit maximum URLs archived per browser session, exiting to free memory.')
- console.warn(' Run this process again to continue with the next batch...')
- process.exit(21)
- }
- const browser = await page.browser()
- const client = await page.target().createCDPSession()
- const extensions = await getChromeExtensionsFromCache({browser})
- const browser_version = await browser.version()
- const original_url = url.toString()
- const start_time = (new Date())
-
- console.log('[0/4]-------------------------------------------------------------------------')
- const snapshot_dir = await setupSnapshotDir({original_url, start_time})
- const snapshot = await setupSnapshotDB({original_url, start_time, snapshot_dir})
- console.log('[1/4]-------------------------------------------------------------------------')
- console.log(`[🪟] Starting page & viewport setup (${browser_version} ${DEFAULT_VIEWPORT.isMobile ? 'mobile' : 'desktop'} ${DEFAULT_VIEWPORT.width}x${DEFAULT_VIEWPORT.height}px)...`)
- const page_state = {
- // global static state
- browser,
- client,
- browser_version,
- extensions,
- // per-page static metadata
- original_url,
- snapshot,
- snapshot_dir,
- start_time: start_time.toISOString(),
- start_ts: Number(start_time),
- version: versionStrFromDate(start_time),
- // per-page mutable archiving state
- main_response: null,
- recorder: null,
- console_log: [],
- traffic_log: {},
- redirects: {},
- }
- page._original_url = original_url
-
- try {
- // run all page setup functions in parallel
- const results = await Promise.allSettled([
- // loadAuthStorage(page, page_state, { apply: true }),
- startMetadataRecording(page, page_state),
- setupURLRewriting(page, page_state),
- // setupViewport(page, page_state),
- setupModalAutoClosing(page, page_state),
- loadCloudflareCookie(page, page_state),
- startResponseSaving(page, page_state),
- saveYTDLP(page, page_state),
- saveGALLERYDL(page, page_state),
- // saveSourceMaps(page, page_state),
- // TODO: someday setup https://github.com/osnr/TabFS ?
- ]);
- // run all page setup functions in parallel
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason);
- if (rejected.length) console.warn('[⚠️] Partial failures during page setup:', rejected);
- } catch(err) {
- console.error('[❌] PAGE SETUP ERROR', JSON.stringify(err, null, 4))
- return
- }
- console.log('[2/4]-------------------------------------------------------------------------')
- console.log('[➡️] NAVIGATION[INI]', ANSI.blue + url + ANSI.reset)
- const startrecording_promise = startScreenrecording(page, page_state)
- page_state.main_response = await page.goto(url, {waitUntil: 'load', timeout: 40_000})
- try {
- const results = await Promise.allSettled([
- startrecording_promise,
- page.bringToFront(),
- page.waitForNetworkIdle({concurrency: 0, idleTime: 900, timeout: 20_000}),
- ])
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason)
- if (rejected.length) console.warn('[⚠️] Parial failures during page load:', rejected)
- } catch(err) {
- console.error('[❌] ERROR DURING PAGE LOAD', JSON.stringify(err, null, 4))
- return
- }
- if (page_state.main_response === null) {
- page_state.main_response = await page.waitForResponse(() => true)
- }
- assert(page_state.main_response)
- if (page_state.main_response.status() == 429) {
- throw `[⚠️] Got 429 rate-limit response, skipping this URL for now...`
- }
- // emulate human browsing behavior
- // await disableAnimations(page, page_state);
- await jiggleMouse(page, page_state);
- await solveCaptchas(page, page_state);
- await blockRedirects(page, page_state);
- await scrollDown(page, page_state);
- // await expandComments(page, page_state);
- await submitForm(page, page_state);
- // await blockJSExecution(page, page_state);
- console.log('[3/4]-------------------------------------------------------------------------')
-
- // stop tampering with page requests & JS / recording metadata / traffic log
- await stopMetadataRecording(page, page_state)
- // do all synchonous archiving steps that need exclusive use of the whole page while doing stuff
- const saveScreenrecording_promise = saveScreenrecording(page, page_state);
- await saveScreenshot(page, page_state);
- await savePDF(page, page_state);
- console.log('[4/4]-------------------------------------------------------------------------')
- // do all async archiving steps that can be run at the same time
- await inlineShadowDOM(page, page_state);
- const results = await Promise.allSettled([
- saveTitle(page, page_state),
- saveSEO(page, page_state),
- saveFavicon(page, page_state),
- saveSSL(page, page_state),
- saveRequests(page, page_state),
- saveRedirects(page, page_state),
- saveHeaders(page, page_state),
- saveRaw(page, page_state),
- saveDOM(page, page_state),
- saveBodyText(page, page_state),
- // savePandoc(page, page_state),
- saveReadability(page, page_state),
- saveAccessibility(page, page_state),
- saveOutlinks(page, page_state),
- // saveAuthStorage(page, page_state),
- saveAIQualityAssuranceResult(page, page_state),
- ]);
- // do all sync archiving steps that require browser extensions at the very end (they are the buggiest)
- const bg_results = Promise.allSettled([
- saveScreenrecording_promise,
- saveSinglefile(page, page_state),
- // saveArchiveWebPage(page, page_state),
- // savePocket(page, page_state),
- ])
- const {duration} = await saveMetrics(page, page_state);
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason) // not sure why this has a ts-error, .reason does exist on rejected promises
- if (rejected.length)
- console.warn('[⚠️] Parial failures during archiving:', rejected)
- // Start an interactive REPL here with the `page` instance.
- // https://github.com/berstend/puppeteer-extra/tree/master/packages/puppeteer-extra-plugin-repl
- // await page.repl()
- // await page.browser().repl()
- console.log(`[✅] ${ANSI.blue}Finished archiving in ${duration/1000}s.${ANSI.reset}`)
-
- try {
- const rejected = (await bg_results)
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason) // not sure why this has a ts-error, .reason does exist on rejected promises
- if (rejected.length)
- console.warn('[⚠️] Parial failures during wrap-up tasks:', rejected)
-
- console.log('[🗑️] Resetting to about:blank to ensure memory is freed...')
- await page.goto('about:blank')
- await page.close()
- } catch(err) {
- console.log(err)
- }
- // symlink the best results from across all the versions/ into the snapshot dir root
- await symlinkBestSnapshotResults(snapshot_dir)
- // display latest version screenshot GIF
- console.log()
- try {
- const latest_version_gif = path.join(snapshot_dir, 'versions', page_state.version, path.basename(SCREENRECORDGIF_PATH(page)))
- const dirent = await blockUntilExists(latest_version_gif, {min_bytes: 100, timeout: 15_000})
- child_process.spawn('/Users/squash/.iterm2/imgcat', [dirent.abspath], {stdio: [null, 'inherit', 'inherit']})
- } catch(err) {
- console.warn('[⚠️] Failed to display screenrecording.gif...', err)
- console.log()
- }
- // determine whether task succeeded or failed based on AI QA score
- const latest_version_aiqa = path.join(snapshot_dir, 'versions', page_state.version, path.basename(AIQA_PATH(page)))
- const qa_results = JSON.parse((await fs.promises.readFile(latest_version_aiqa)).toString())
- if (qa_results.pct_visible < 50) {
- throw `[❌] Task completed with problems, got AI QA score of ${qa_results.pct_visible}%! ${qa_results.warnings.join(', ')} ${qa_results.error_text || ''}`
- } else {
- console.log(`[💫] Task completed succesfully: ${qa_results.pct_visible}% ${qa_results.warnings.join(', ') || ''}`)
- console.log(` Summary: ${(qa_results.main_content_title || qa_results.description || 'No title/description detected').substring(0, 80)}... ${qa_results.main_content_author || ''} ${qa_results.main_content_date || ''}`)
- return true
- }
- }
- async function passiveArchiveTask({browser, page, url}) {
- // archive passively (e.g. a tab that was opened already by a human), without changing the active page
- const is_unarchivable_url = URL_SCHEMES_IGNORED.includes(url.split(':')[0])
- const is_already_archived = ALREADY_ARCHIVED.has(url.slice(0, 4096))
- if (is_unarchivable_url || is_already_archived) return null
- ALREADY_ARCHIVED.add(url.slice(0, 4096))
- // these have to be as early as possible because we're racing with the page load (we might even be too late)
- // jk nevermind, we now re-open a new bg tab for every tab that's created to re-capture the initial request
- // await page.setRequestInterception(true);
- // await page.setCacheEnabled(false);
- const original_url = url.toString()
- const start_time = (new Date())
- const browser_version = await browser.version()
-
- console.log('------------------------------------------------------------------------------')
- console.log('[➕] Starting archive of new tab opened in driver browser...', await browser.version())
- const snapshot_dir = await setupSnapshotDir({original_url, start_time})
- const snapshot = await setupSnapshotDB({ original_url, start_time, snapshot_dir })
- console.log('------------------------------------------------------------------------------')
- console.log(`[🪟] Starting page & viewport setup (${browser_version} ${DEFAULT_VIEWPORT.isMobile ? 'mobile' : 'desktop'} ${DEFAULT_VIEWPORT.width}x${DEFAULT_VIEWPORT.height}px)...`)
- // create a new page in the background for archiving
- const old_page = page
- page = await browser.newPage()
- await old_page.bringToFront()
- const client = await page.target().createCDPSession()
- const extensions = await getChromeExtensionsFromCache({ browser })
- const page_state = {
- // global static state
- browser,
- client,
- browser_version,
- extensions,
- // per-page static metadata
- original_url,
- snapshot,
- snapshot_dir,
- start_time: start_time.toISOString(),
- start_ts: Number(start_time),
- version: versionStrFromDate(start_time),
- // per-page mutable archiving state
- main_response: null,
- recorder: null,
- console_log: [],
- traffic_log: {},
- redirects: {},
- }
- page._original_url = original_url
- try {
-
- // run all page setup functions in parallel
- const results = await Promise.allSettled([
- // loadAuthStorage(page, page_state, {apply: true}),
- startMetadataRecording(page, page_state),
- setupURLRewriting(page, page_state),
- startResponseSaving(page, page_state),
- saveYTDLP(page, page_state),
- saveGALLERYDL(page, page_state),
- // saveSourceMaps(page, page_state),
- ]);
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason)
- if (rejected.length) console.warn('[⚠️] Parial failures during page setup:', rejected)
- } catch(err) {
- console.warn('[❌] ERROR DURING PAGE SETUP', JSON.stringify(err, null, 4))
- return
- }
- // load the url in the background page, then switch to it once its loaded and close the original tab
- console.log('[➡️] NAVIGATION[INI]', ANSI.blue + url + ANSI.reset)
- const startrecording_promise = startScreenrecording(page, page_state)
- page_state.main_response = await page.goto(url, {waitUntil: 'load', timeout: 40_000})
-
- // for debugging
- globalThis.page = page
- globalThis.page_state = page_state
- // start loading the page, start screenrecording, close the old page, and wait for loading to finish (all at once, fine for these to race)
- try {
- const results = await Promise.allSettled([
- startrecording_promise,
- page.bringToFront(),
- old_page.close(),
- page.waitForNetworkIdle({concurrency: 0, idleTime: 900, timeout: 20_000}),
- ])
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason)
- if (rejected.length) console.warn('[⚠️] Parial failures during [age load:', rejected)
- } catch(err) {
- console.warn('[❌] ERROR DURING PAGE LOAD', JSON.stringify(err, null, 4))
- return
- }
- if (page_state.main_response === null) {
- page_state.main_response = await page.waitForResponse(() => true)
- }
- assert(page_state.main_response)
- if (page_state.main_response.status() == 429) {
- throw `[⚠️] Got 429 rate-limit response, skipping this URL for now...`
- }
- // resume page if paused by waitForDebuggerOnStart/dev tools debugger/backgrounding
- try {
- await client.send('Page.enable');
- await client.send('Page.setWebLifecycleState', {state: 'active'});
- await client.send('Runtime.runIfWaitingForDebugger')
- } catch(err) { /* console.warn(err) */ }
- // wait a couple seconds for page to finish loading
- await wait(5_000)
- // emulate human browsing behavior
- // await disableAnimations(page, page_state);
- // await jiggleMouse(page, page_state);
- await solveCaptchas(page, page_state);
- // await blockRedirects(page, page_state);
- // await scrollDown(page, page_state);
- // await expandComments(page, page_state);
- await submitForm(page, page_state);
- // await blockJSExecution(page, page_state);
- await stopMetadataRecording(page, page_state) // stop tampering with page requests & JS
- console.log('[3/4]-------------------------------------------------------------------------')
- // do all synchonous archiving steps that need exclusive use of the whole page while doing stuff
- const saveScreenrecording_promise = saveScreenrecording(page, page_state);
- await saveScreenshot(page, page_state);
- await savePDF(page, page_state);
- console.log('[4/4]-------------------------------------------------------------------------')
- // do all async archiving steps that can be run at the same time
- await inlineShadowDOM(page, page_state);
- const results = await Promise.allSettled([
- saveTitle(page, page_state),
- saveSEO(page, page_state),
- saveFavicon(page, page_state),
- saveSSL(page, page_state),
- saveRequests(page, page_state),
- saveRedirects(page, page_state),
- saveHeaders(page, page_state),
- saveRaw(page, page_state),
- saveDOM(page, page_state),
- saveBodyText(page, page_state),
- // savePandoc(page, page_state),
- saveReadability(page, page_state),
- saveAccessibility(page, page_state),
- saveOutlinks(page, page_state),
- // saveAuthStorage(page, page_state),
- saveAIQualityAssuranceResult(page, page_state),
- ]);
- // do all sync archiving steps that require browser extensions at the very end (they are the buggiest)
- const bg_results = Promise.allSettled([
- saveScreenrecording_promise,
- saveSinglefile(page, page_state),
- // saveArchiveWebPage(page, page_state),
- // savePocket(page, page_state),
- ])
- const {duration} = await saveMetrics(page, page_state);
- const rejected = results
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason)
- if (rejected.length)
- console.warn('[⚠️] Parial failures during page archiving:', rejected)
- // Start an interactive REPL here with the `page` instance.
- // https://github.com/berstend/puppeteer-extra/tree/master/packages/puppeteer-extra-plugin-repl
- // await page.repl()
- // await page.browser().repl()
- console.log(`[✅] Finished archiving in ${duration/1000}s.`,)
-
- // await page.tracing.stop();
- try {
- const rejected = (await bg_results)
- .filter(result => result.status === 'rejected')
- .map(result => (result as PromiseRejectedResult).reason)
- if (rejected.length)
- console.warn('[⚠️] Parial failures during page wrap-up tasks:', rejected)
- } catch(err) {
- console.log(err)
- }
- await symlinkBestSnapshotResults(snapshot_dir)
- }
- /******************************************************************************/
- /************************* Page Setup Tasks ***********************************/
- async function setupSnapshotDir({original_url, start_time, snapshot_dir=null}) {
- // setup archive/<id> snapshot output folder, move old files into versions/<date>/* + clear any existing symlinks
- const snap_dir = snapshot_dir || TASK_PATH(original_url)
-
- console.log()
- console.log()
- console.log(ANSI.blue + original_url + ANSI.reset)
- console.log(ANSI.black + snap_dir + ANSI.reset)
- console.log()
- console.log('[📂] Setting up Snapshot output directory...'.padEnd(82), prettyPath(snap_dir))
- // check for existing data at old legacy paths e.g. ./data/archive/1999999999.1723425
- const hacky_dir = path.join(ARCHIVE_DIR, `1999999999.${hashCode(original_url)}`)
- const known_dir = SNAPSHOT_DIRS_BY_URL[original_url]
- const known_dir_exists = fs.existsSync(known_dir)
- const hacky_dir_exists = fs.existsSync(hacky_dir)
- if (snap_dir == hacky_dir) {
- if (known_dir_exists) {
- throw `Tried to create snapshot in ${snap_dir} but potential duplicate exists in ${known_dir}!`
- }
- } else if (snap_dir == known_dir) {
- if (hacky_dir_exists) {
- throw `Tried to create snapshot in ${snap_dir} but potential duplicate exists in ${hacky_dir}!`
- }
- } else {
- if (known_dir_exists) {
- throw `Tried to create snapshot in ${snap_dir} but potential duplicate exists in ${known_dir}!`
- } else if (hacky_dir_exists) {
- throw `Tried to create snapshot in ${snap_dir} but potential duplicate exists in ${hacky_dir}!`
- } else {
- throw `Tried to create snapshot in ${snap_dir} but its not a recognized snapshot dir path:\n - ${known_dir}\n - ${hacky_dir}`
- }
- }
-
- // mkdir -p ./data/archive/<snap_id>/versions && cd ./data/archive/<snap_id>
- await fs.promises.mkdir(path.join(snap_dir, 'versions'), {recursive: true})
- process.chdir(snap_dir)
- // clear any /data/archive/<snap_id>/*.* symlinks pointing to existing ./versions/<versionid>/*.* files
- await clearSnapshotDirSymlinks(snap_dir)
- // move /data/archive/<snap_id>/*.* loose output files from any prior run into ./versions/<versionid>/*.*
- await collectSnapshotDirVersionFiles(snap_dir)
-
- // update /data/indexes/<index_name>/* to include references to /data/archive/<snap_id> as-needed
- await updateSnapshotDirIndexes(snap_dir, {original_url, start_time})
- // assert /data/archive/<snap_id>/ contains no invalid/partial files + is empty/ready to receive new files
- await assertSnapshotDirIsValid(snap_dir, {is_empty: true})
- return snap_dir
- }
- // ./index/<index_name> : index_getter(page_state) => "<index_key_str>"
- const INDEXES = {
- snapshots_by_day: ({start_time}) =>
- versionStrFromDate(start_time, {withDate: true, withTime: false}),
- snapshots_by_domain: ({original_url}) =>
- (new URL(original_url)).hostname || '', // hostname does not include :port
- }
- async function updateSnapshotDirIndexes(snap_dir, page_state, indexes=INDEXES, indexes_dir=INDEXES_DIR) {
- assert(indexes)
- console.log(`[🔎] Linking Snapshot in indexes (${Object.keys(indexes).join(', ')})...`)
- // const {snapshot_dir, original_url, start_ts} = page_state
- for (const [index_name, index_key_getter] of Object.entries(indexes)) {
- const index_entry = await indexSnapshotDir(snap_dir, {index_name, index_key_getter, indexes_dir}, page_state)
- }
- }
- async function indexSnapshotDir(snap_dir, {index_name, index_key_getter, indexes_dir=INDEXES_DIR}, page_state) {
- // place symlinks to this snapshot in any /indexes/<index_name/ indexes as-needed
- // const snap_id = snap_dir.split('/').at(-1)
- const index_dir = path.join(indexes_dir, index_name) // /data/index/snapshots_by_day
- await fs.promises.mkdir(index_dir, {recursive: true})
- // calculate the index key, e.g. "200101231" or "example.com"
- assert(index_name && index_key_getter)
- assert(page_state)
- const index_key = String(index_key_getter(page_state)) // '20010131'
- assert(index_key)
- const snap_id = path.parse(snap_dir).base // '19999999.23423523'
- assert(snap_id)
- const index_entries_dir = path.join(index_dir, index_key) // /data/index/snapshots_by_day/20010131
- await fs.promises.mkdir(index_entries_dir, {recursive: true})
- const symlink_path = path.join(index_entries_dir, snap_id) // /data/index/snapshots_by_day/20010131/19999999.23423523
- // create symlink index/snapshots_by_day/<YYYYMMDD>/<snap id> -> ./archive/<snap_id> symlink
- const {symlink_abspath} = await overwriteSymlink(snap_dir, symlink_path, {relative: true, mkdirs: false})
- }
- async function collectSnapshotDirVersionFiles(snap_dir) {
- // move archive/<id>/*.* snapshot output files into archive/<id>/versions/<date>/* dated version folder
- // detect start time / version info from previous result metrics.json
- const snap_id = snap_dir.split('/archive/').at(-1)
- const existing_metrics = path.join(snap_dir, 'metrics.json')
- let {start_time, VERSION} = {start_time: '1970-01-01T00:00:00.000Z', VERSION: '19700101000000'}
- try {
- ;({start_time, VERSION} = JSON.parse(await fs.promises.readFile(existing_metrics, 'utf-8')));
- } catch(err) {
- // continue normally, overwriting existing files is fine if they're broken to begin with
- }
-
- // create new version folder based on metrics.json start_time (or epoch time as fallback for legacy output)
- const version_dir_name = VERSION || versionStrFromDate(start_time)
- const version_dir = path.join(snap_dir, 'versions', version_dir_name)
- await fs.promises.mkdir(version_dir, {recursive: true})
- // move all result files from snapshot_dir root into version folder
- const existing_snapshot_files =
- (await fs.promises.readdir(snap_dir, {withFileTypes: true}))
- .filter(dirent => {
- if (dirent.name.startsWith('.')) return false // ignore hidden files, dont version them
- if (dirent.name == 'versions') return false // dont try to move versions folder into itself
- if (dirent.isSymbolicLink()) return false // skip existing symbolic links
- return (dirent.isFile() || dirent.isDirectory()) // dont try to version sockets/FIFOs/devs etc.
- })
- if (existing_snapshot_files.length) {
- console.log(`[📅] Moving snapshot results into version dir: ./data/archive/${snap_id}/* ->`.padEnd(82), `./data/archive/${snap_id}/versions/${VERSION}/`)
- }
-
- const snapshot_files = await getDirInfo(snap_dir, {withRoot: false, filter: ({relpath}) => !relpath.startsWith('versions')})
- const version_files = await getDirInfo(version_dir, {withRoot: false})
- for (const {name} of existing_snapshot_files) {
- const snapdir_entry_abspath = path.join(snap_dir, name)
- const versioned_entry_abspath = path.join(version_dir, name)
- const snapshot_entry = snapshot_files[name]
- const version_entry = version_files[name]
- if (snapshot_entry && version_entry) {
- // a conflicting file/dir already exists in the destination path
- // we have a few options here, we can try to merge them, or we can create a new version
- if (snapshot_entry.sha256 == version_entry.sha256) {
- // both are the same already, delete the duplicate (leaving the copy inside the version dir)
- // if (snapshot_entry.is_dir) {
- // await fs.promises.rmdir(snapshot_entry.abspath, {recursive: true})
- // } else {
- // await fs.promises.unlink(snapshot_entry.abspath)
- // }
- // console.warn(`[!] Found harmless exact duplicate files, leaving as is: ${snapshot_entry.summary} and ${version_entry.summary}`)
- } else {
- // both are different,
- if (snapshot_entry.num_bytes > version_entry.num_bytes) {
- // snapshot entry is bigger, keep it and delete version entry?
- } else {
- // version entry is bigger, keep it and delete snapshot entry
- }
- console.warn(' ', snapshot_entry.summary)
- console.warn(' ', version_entry.summary)
- // throw `Found conflicting duplicate files with different contents: ${name}`
- }
- } else {
- // mv ./data/archive/<snap_id>/example.txt -> ./data/archive/<snap_id>/versions/<version_id>/example.txt
- await fs.promises.rename(snapdir_entry_abspath, versioned_entry_abspath)
- console.log(` ↣ ${prettyPath(snapdir_entry_abspath)} ->`.padEnd(82), prettyPath(versioned_entry_abspath))
- }
- }
- }
- // Extractor definition
- // {
- // phase: setup | load | sync1 | async1 | sync2 | close
- // name: 'media' | 'photos', 'wget', 'singlefile'
- //
- // shouldRun(page, page_state)
-
- // pageSetup
- // pageLoad
- // pageInteraction clicking around/scrolling
- // archivePhase1 sync
- // archivePhase2 async
- // archivePhase3 async
- // pageClose
- // execute(page, page_state)
- // validateResult(page, page_state)
- // }
- async function clearSnapshotDirSymlinks(snap_dir) {
- // delete all archive/<id>/* symlinks in preparation for new snapshot output to be placed there
- const existing_symlinks =
- (await fs.promises.readdir(snap_dir, {withFileTypes: true}))
- .filter(dirent => {
- if (dirent.name.startsWith('.')) return false // ignore hidden files, dont version them
- if (dirent.name == 'versions') return false // dont try to move versions folder into itself
- return dirent.isSymbolicLink()
- })
- for (const {name: existing_symlink} of existing_symlinks) {
- await fs.promises.unlink(path.join(snap_dir, existing_symlink))
- // if symlinks are not cleared before starting, it can cause issues with outputs writing into previous versions folders
- // e.g. screerecording saves to ./media which could be pointing to previous version's ./versions/<olddate>/media
- }
- }
- async function symlinkBestSnapshotResults(snap_dir) {
- // move any existing files into versions/<date> folder (clear out main folder)
- // symlink latest files from versions/<date>/* into main folder
-
- await fs.promises.mkdir(path.join(snap_dir, 'versions'), {recursive: true})
- process.chdir(snap_dir)
- const metrics_file = path.join(snap_dir, 'metrics.json')
- // if (!fs.existsSync(metrics_file) || (await fs.promises.lstat(metrics_file)).isSymbolicLink()) {
- // console.warn('[⚠️] Warning, found partial dirty snapshot state (did the snapshot get interrupted?)', snap_dir)
- // }
- // move output files into versioned folder
- await collectSnapshotDirVersionFiles(snap_dir)
- // clear any existing symlinks
- await clearSnapshotDirSymlinks(snap_dir)
- // assert task dir is empty and contains no bare files that might get overwritten, also asserts version dirs are valid
- await assertSnapshotDirIsValid(snap_dir, {is_empty: true})
- const version_dirs = (await fs.promises.readdir(path.join(snap_dir, 'versions'))).sort() // earliest to latest
- const most_recent = version_dirs.at(-1)
- // for each version dir in versions/ (oldest -> newest)
- for (const version_dir of version_dirs) {
- if (version_dir.startsWith('.')) continue
- const version_dir_abspath = path.join(snap_dir, 'versions', version_dir)
- const version_dir_files = (
- (await fs.promises.readdir(version_dir_abspath))
- .filter(filename => !filename.startsWith('.')))
- // iterate through all the files/folders in the version dir
- for (const filename of version_dir_files) {
- const snapdir_entry = path.join(snap_dir, filename) // ./data/archive/<snapid>/filename
- const versiondir_entry = path.join(snap_dir, 'versions', version_dir, filename) // ./data/archive/<snapid>/versions/<versionid>/filename
-
- if (fs.existsSync(snapdir_entry)) {
- // if an entry already exists in the snapshot root for this filename
- if ((await fs.promises.lstat(snapdir_entry)).isSymbolicLink()) {
- // if a symlink already exists in the root with the same name,
- // check if the version file we're looking at is a better candidate to replace it
- const existing_abspath = await fs.promises.realpath(snapdir_entry)
- const desired_abspath = path.join(version_dir_abspath, filename)
- if (existing_abspath != desired_abspath) {
- // check if the new candidate is larger or if the existing symlink is larger (largest file = most likely to be highest quality capture data)
- const largest_path = await getLargestPath(existing_abspath, desired_abspath)
- if (largest_path != (await fs.promises.realpath(existing_abspath))) {
- const larger_version = path.basename(path.dirname(largest_path))
- const larger_abspath = path.join(snap_dir, 'versions', larger_version, filename)
-
- // console.log(' - swapping for larger file:', filename, '->', larger_abspath.split('/archive/').at(-1))
- await overwriteSymlink(larger_abspath, snapdir_entry, {search_limit: snap_dir})
- } else {
- // console.log(' - leaving larger file:', largest_path.split('/archive/').at(-1))
- }
- } else {
- // leave existing symlink pointing to current version file, nothing to change
- // console.log(' - leaving current file:', existing_abspath.split('/archive/').at(-1))
- }
- } else {
- // clearSnapshotDirSymlinks() should have already cleared these files out!
- throw `Non-symlink file found in root of snapshot dir! Refusing to overwrite: ${prettyPath(snapdir_entry)}`
- }
- } else {
- // no entry exists in the snapshot root for this filename, create one by linking to the version file
- await overwriteSymlink(versiondir_entry, snapdir_entry, {search_limit: snap_dir})
- }
- // if (version_dir == most_recent) {
- // // only log most recent links even though we link older ones too (otherwise its too noisy)
- // console.log(` 🔗 ./${filename} -> ./${versiondir_entry} linking...`)
- // }
- }
- }
- return snap_dir
- }
- async function assertSnapshotDirIsValid(snap_dir, {is_empty=false}={}) {
- process.chdir(snap_dir)
- console.log()
- console.log(`[☑️] Checking that snapshot records are valid...`)
- // get all directory entries in archive/<snapshot_id>/*
- const snapshot_dir_entries =
- (await fs.promises.readdir(snap_dir, {withFileTypes: true}))
- .filter(dirent => {
- if (dirent.name.startsWith('.')) return false
- if (dirent.name == 'versions') return false
- })
- // assert versions folder exists and is not a symbolic link
- const versions_dir = path.join(snap_dir, 'versions')
- assert(fs.existsSync(versions_dir))
- assert(!(await fs.promises.lstat(versions_dir)).isSymbolicLink())
- // if it should be empty, check that no loose files exist
- if (is_empty) {
- assert(!snapshot_dir_entries.length, `Found loose files in snapshot-dir that shouldn't be there! ${snap_dir}`)
- }
- // assert all non-hidden files in snapshot dir are symbolic links to actual data in versions/<date>/*
- for (const snapshot_dir_entry of snapshot_dir_entries) {
- if (snapshot_dir_entry.name.startsWith('.')) continue
- if (snapshot_dir_entry.name == 'versions') continue
- assert(snapshot_dir_entry.isSymbolicLink(), `Found non-symbolic link in root of snapshot dir! ${snap_dir}/${snapshot_dir_entry.name}`)
- assert(fs.existsSync(snapshot_dir_entry.name), `Found broken symbolic link in root of snapshot dir! ${snap_dir}/${snapshot_dir_entry.name}`)
- }
- const version_entries = (
- (await fs.promises.readdir(versions_dir))
- .filter(foldername => !foldername.startsWith('.'))
- .sort())
- console.log(` √ ${prettyPath(versions_dir)}`, version_entries.length)
- for (const version_dir of version_entries) {
- await assertVersionDirIsValid(path.join(versions_dir, version_dir))
- }
- // write snapshot dir file listing w/ sizes & hashes to .files.json
- const directory_info = await getDirInfo(snap_dir, {withRoot: true, withHelpers: false, maxdepth: 3})
- await overwriteFile(path.join(snap_dir, '.files.json'), directory_info)
- }
- async function assertVersionDirIsValid(version_dir) {
- const dirname = path.parse(version_dir).name
- assert(fs.existsSync(version_dir), `Version dir does not exist: ${prettyPath(version_dir)}`)
- const dirent = await fs.promises.lstat(version_dir)
- assert(dirent.isDirectory() && !dirent.isSymbolicLink(), `Found non-directory in versions dir! ${prettyPath(version_dir)}`)
-
- const unix_epoch = '19700101000000'
- const is_name_valid_datestr = /^\d+$/.test(dirname) && (dirname.length == 14) && (dirname.startsWith('2') || dirname == unix_epoch) && parseVersionDateStr(dirname)
- assert(is_name_valid_datestr, `Version directories must be a 14-character long date string like 20251231235959! ${dirname}`)
- // get all directory entries in archive/<snapshot_id>/versions/<version_id>/*
- const version_dir_entries = (
- (await fs.promises.readdir(version_dir, {withFileTypes: true}))
- .filter((dirent) => !dirent.name.startsWith('.')))
- // assert version dir contains only actual snapshot output files (not-symbolic links or other version dirs)
- for (const version_dir_entry of version_dir_entries) {
- assert(version_dir_entry.name != 'versions', `Version dir cannot contain another versions folder! ${prettyPath(version_dir)}/versions`)
- assert(!version_dir_entry.isSymbolicLink(), `Version dir cannot contain symbolic link! ${prettyPath(version_dir)}/${version_dir_entry.name}`)
- }
- // color highlight the unix epoch version in black, and any version created today in blue
- let pretty_dirname = dirname
- if (dirname == unix_epoch) {
- pretty_dirname = ANSI.black + unix_epoch + ANSI.reset
- }
- const today = versionStrFromDate(new Date(), {withDate: true, withTime: false})
- if (dirname.startsWith(today)) {
- pretty_dirname = ANSI.blue + dirname + ANSI.reset
- }
- // write version dir file listing w/ sizes & hashes to .files.json
- const directory_info = await getDirInfo(version_dir, { withRoot: true, withHelpers: false, maxdepth: 3 })
- await overwriteFile(path.join(version_dir, '.files.json'), directory_info)
- console.log(` √ ./versions/${pretty_dirname} contains`, version_dir_entries.length, 'results')
- }
- async function setupSnapshotDB({ original_url, start_time, snapshot_dir }) {
- // setup Snapshot database row, finding it if it already exists or creating a new one
- const timestamp = snapshot_dir.split('/').at(-1)
- const search_attrs = { url: original_url, timestamp }
- const update_attrs = { url: original_url, timestamp, added: start_time, title: null }
- let snapshot = await Snapshot.findOne({ where: search_attrs });
- let created = false
- if (!snapshot) {
- snapshot = await Snapshot.findOne({ where: {url: original_url} });
- if (snapshot) {
- // console.warn(`[X] Found DB Snapshot [${timestamp}](${original_url.substring(0, 30)}...) that has different timestamp from existing dir ${prettyPath(snapshot_dir)}!`)
- // throw 'Snapshot DB record does not match filesystem path!'
- } else {
- console.log(`[+] Creating new DB Snapshot [${timestamp}](${original_url.substring(0, 30)}...) for ${prettyPath(snapshot_dir)}...`)
- // ;([snapshot, created] = await Snapshot.findOrCreate({where: search_attrs, defaults: update_attrs }));
- // throw 'Wanted to create new Snapshot but refusing to modify DB during testing!'
- }
- }
- // assert(snapshot && (snapshot instanceof Snapshot))
- return snapshot
- }
- async function setupViewport(page, _page_state) {
- // setup viewport
- await page.setViewport(DEFAULT_VIEWPORT);
- await page.setGeolocation(DEFAULT_GEOLOCATION);
- // await page.setBypassCSP(true); // bypass CSP restrictions (requires --disable-web-security)
- page.setDefaultTimeout(DEFAULT_TIMEOUT);
- // Optional: emulate a mobile device
- // await page.emulate(puppeteer.devices['iPhone 6']);
- // Configure light mode/dark mode & accessibility reduced motion preferences
- await page.emulateMediaFeatures([
- {name: 'prefers-color-scheme', value: DEFAULT_COLOR_SCHEME},
- {name: 'prefers-reduced-motion', value: 'reduce'},
- ]);
- // Setup headers & deterministically chose a random referrer based on URL
- const rand_idx = hashCode(await page.url()) % DEFAULT_REFERRERS.length
- await page.setExtraHTTPHeaders({
- ...DEFAULT_HEADERS,
- referrer: DEFAULT_REFERRERS[rand_idx],
- })
- // Setup alert to trigger if site tries to sniff whether we are a bot
- function sniffDetector() {
- const userAgent = window.navigator.userAgent;
- const platform = window.navigator.platform;
- // @ts-ignore
- window.navigator.__defineGetter__('userAgent', function () {
- // @ts-ignore
- window.navigator.sniffed = true;
- return userAgent;
- });
- // @ts-ignore
- window.navigator.__defineGetter__('platform', function () {
- // @ts-ignore
- window.navigator.sniffed = true;
- return platform;
- });
- }
- await page.evaluateOnNewDocument(sniffDetector);
- // @ts-ignore
- const was_sniffed = await page.evaluate(() => (!!window.navigator.sniffed))
- if (was_sniffed) {
- console.warn('[⚠️] Site tried to sniff if we are a bot! Site may be difficult to archive.')
- }
-
- return page
- }
- async function setupModalAutoClosing(page, page_state, {timeout=1_250}={}) {
- page.on('dialog', (dialog) => {
- console.log(`[👆] Auto-closing modal that popped up: ${dialog.message()}...`)
- setTimeout(() => {try { dialog.accept() } catch(err) {}}, timeout);
- })
- // if you expect a file-upload dialog, use this to catch it instead:
- // const [fileChooser] = await Promise.all([
- // page.waitForFileChooser(),
- // ]);
- // await fileChooser.accept(['/tmp/myfile.pdf']);
- page.on('close', () => {
- try {
- page.off('dialog')
- } catch(err) {}
- })
- }
- async function startScreenrecording(page, page_state, {duration_limit=60, codec='libx264'}={}) {
- await fs.promises.mkdir(path.dirname(SCREENRECORDING_PATH(page)), {recursive: true})
- // console.log(`[🎬] Starting screen-recording stream...`.padEnd(82), prettyPath(SCREENRECORDING_PATH(page)))
-
- // alternative: interact with low-level puppeteer screencast API directly
- // using puppeteer.page.screencast: https://pptr.dev/api/puppeteer.page.screencast
- // const recorder = await page.screencast({path: SCREENRECORDING_PATH(page)});
- // alternative: use puppeteer-stream for .webm/.mp4 screen recordings with tab audio included
- // works sometimes but has a few issues, e.g.: https://github.com/SamuelScheit/puppeteer-stream/issues/8
- // alternative: puppeteer-screen-recorder (most compatible/stable but doesn't include tab audio output)
- const recorder = new PuppeteerScreenRecorder(page, {
- followNewTab: false,
- recordDurationLimit: duration_limit,
- // fps: 25,
- // ffmpeg_Path: '<path of ffmpeg_path>' || null,
- // videoFrame: {
- // width: 1024,
- // height: 768,
- // },
- // videoCrf: 18,
- videoCodec: codec,
- // videoPreset: 'ultrafast',
- // videoBitrate: 1000,
- // autopad: {
- // color: 'black' | '#35A5FF',
- // },
- // aspectRatio: '4:3',
- });
- page_state.recorder = recorder
- await recorder.start(SCREENRECORDING_PATH(page))
- page.on('close', async () => {await saveScreenrecording(page, page_state)});
- return page_state
- }
- async function startResponseSaving(page, page_state) {
- const dir = RESPONSES_PATH(page)
- await fs.promises.mkdir(dir, {recursive: true})
- console.log(`[🌄] Starting raw response bytes recording...`.padEnd(82), prettyPath(dir) + '/')
- // Document, Stylesheet, Image, Media, Font, Script, TextTrack, XHR, Fetch, Prefetch, EventSource, WebSocket, Manifest, SignedExchange, Ping, CSPViolationReport, Preflight, Other
- const types_to_save = [
- // 'document',
- 'script',
- 'stylesheet',
- 'font',
- 'image',
- 'media',
- 'xhr',
- 'websocket',
- ]
- // reset responses index file to empty
- const responses_log_path = path.join(dir, 'index.jsonl')
- await overwriteFile(responses_log_path, '')
- // add handler to save all image repsonses into output directory
- page.on('response', async (response) => {
- try {
- const timestamp = versionStrFromDate(new Date(), {withDate: true, withTime: true, withSeconds: true, withMilliseconds: true})
- if (!page_state.main_response && (response.request().url() == page_state.original_url)) {
- // save first response as main page response (if we havent already caught it earlier)
- page_state.main_response = response
- }
- const status = response.status()
- if ((status >= 300) && (status < 500)) {
- // console.log('Got bad response from', response.url(), 'to', response.headers()['location'])
- return
- }
- const request = response.request()
- const resourceType = request.resourceType()
- const url_scheme = (response.url() || request.url()).split(':')[0].toLowerCase()
- const method = (url_scheme === 'data') ? 'DATA' : request.method()
- // console.log(' ', resourceType, response.url())
- if (types_to_save.includes(resourceType)) {
- // create ./responses/xhr/www.facebook.com/static/images/icons/ subdir based on hostname + path
- const resource_type_dir = path.join(dir, resourceType)
- const url = new URL(response.url())
- let subdir = resource_type_dir
- const url_path = (url.pathname || '').slice(0, 250).endsWith('/')
- ? (url.pathname || '').slice(0, 250)
- : path.dirname((url.pathname || '').slice(0, 250))
- // determine subdirectory based on url type (handles http:,https:,file:,data:,chrome-extension:,about:,etc.)
- if (!URL_SCHEMES_IGNORED.includes(url_scheme)) {
- // is a normal http:// or https:// url, use the domain + path to construct subdirectory
- subdir = path.join(resource_type_dir, (url.hostname || 'data').slice(0, 250), url_path)
- } else if (url_scheme == 'data') {
- // is a data:... url, store in ./data subdirectory
- subdir = path.join(resource_type_dir, 'data')
- } else {
- // is a chrome-extension:// or other special url, use the extension id + path to construct subdirectory
- const url_path = path.dirname((url.pathname || '').slice(0, 999))
- subdir = path.join(resource_type_dir, url_scheme, (url.hostname || 'data').slice(0, 250), url_path)
- }
- // write response to responses/all/1716861056899__https%3A%2F%2Fwww.instagram.com%2Fgraphql%2Fquery.json
- let abspath = null
- let resp_mimetype = null
- let extension = ''
- let uniq_filename = null
- let uniq_abspath = null
- let symlink_abspath = null
- let responseSha256 = null
- try {
- await fs.promises.mkdir(path.join(dir, 'all'), {recursive: true})
- try {
- await fs.promises.mkdir(subdir, {recursive: true})
- } catch(err) {
- subdir = subdir + '.dir' // TODO: apply this workaround to parent path entries too
- try {
- await fs.promises.mkdir(subdir, {recursive: true})
- } catch(err) {
- subdir = path.join(resource_type_dir, 'data')
- await fs.promises.mkdir(subdir, {recursive: true})
- }
- }
- ;({abspath: symlink_abspath, resp_mimetype, extension} = await detectFilename({page, response, dir: subdir, resourceType}))
-
- // responses/all/1716861056899__https%3A%2F%2Fwww.instagram.com%2Fgraphql%2Fquery.json
- uniq_filename = `${timestamp}__${method}__` + [encodeURIComponent(url.href).slice(0, 64).replaceAll('/', '_').replace(new RegExp(`.${extension}$`), ''), extension].filter(s => s.length).join('.')
- uniq_abspath = path.join(dir, 'all', uniq_filename)
- let bytesBuffer = null
- try {
- bytesBuffer = await response.buffer()
- } catch(err) {
- if (String(err).includes("Cannot read properties of undefined (reading 'body')")) {
- // not sure why it's happening but seems to be too late to caputre body sometimes? possible race condition
- } else {
- console.warn('[⚠️] Failed to save response bytes for:', response.request().url(), err)
- }
- }
- if (bytesBuffer) {
- // write response data into ./all/<TS>__<METHOD>__<URL>.<EXT>
- await overwriteFile(uniq_abspath, bytesBuffer)
-
- responseSha256 = crypto.createHash('sha256').update(bytesBuffer).digest('hex')
- // write symlink file to ./<TYPE>/<DOMAIN>/...<PATH>/<FILENAME>.<EXT> -> ./all/<TS>__<METHOD>__<URL>.<EXT>
- await overwriteSymlink(uniq_abspath, symlink_abspath, {relative: dir, mkdirs: true, search_limit: dir})
- }
- // console.log(' ->', symlink_abspath)
- } catch(err) {
- // dont do anything for redirectresponses, error responses, etc.
- console.warn(err)
- }
- const urlSha256 = crypto.createHash('sha256').update(String(request.url())).digest('hex')
- // const headersSha256 = crypto.createHash('sha256').update(String(request.headers())) // someday we may want to save headers hashes too
- const truncated_url = (method == 'DATA') ? request.url().slice(0, 128) : request.url() // don't duplicate bytes in data: urls (we already saved them in the file)
- // this is essentially replicating the functionality of a WARC file, but in directory + index.jsonl form
- await fs.promises.appendFile(
- responses_log_path,
- JSON.stringify({
- ts: timestamp,
- method,
- url: truncated_url,
- urlSha256,
- postData: request.postData(),
- response_url: ((method != 'DATA') && (url.href != request.url())) ? url.href : undefined,
- status,
- resourceType,
- mimeType: resp_mimetype,
- responseSha256,
- path: uniq_abspath?.replace(dir, '.'),
- symlink_path: symlink_abspath?.replace(dir, '.'),
- extension,
- }) + '\n',
- 'utf-8',
- )
- }
- } catch(err) {
- // we should never throw hard errors here because there's nothing above us to catch it
- // and we dont want to crash the entire CDP session / browser / main node process
- console.warn('[❌] Error in response handler (set in startResponseSaving):', err)
- }
- });
- // handled by stopMetadataRecording():
- // page.on('close', () => {
- // page.off('response')
- // })
- }
- function dedupeCookies(cookies) {
- const len_before = cookies.length
- const allowed_cookie_attrs = ['domain', 'path', 'name', 'value', 'expires', 'sameSite', 'sourceScheme', 'url', 'priority', 'secure', 'httpOnly']
- const deduped_cookies = {}
- for (const cookie of cookies) {
- try {
- const unique_id = `${cookie.domain}${cookie.path}${cookie.name}`
- deduped_cookies[unique_id] = {
- ...(deduped_cookies[unique_id] || {}),
- ...cookie,
- expires: 2147483640, // max allowed expiry time (2038-01-18)
- session: false, // make sure cookies dont expire at browser close time
- secure: false, // make cookie restrictions more lax (for archiving scripts)
- httpOnly: false, // make it easier to tamper with cookies from JS (for archiving scripts)
-
- // "path": "/",
- // "expires": 2147483641,
- // "size": 194,
- // "httpOnly": false,
- // "secure": false,
- // "session": false,
- // "priority": "High",
- // "sameParty": false,
- // "sourceScheme": "Secure",
- // "sourcePort": 443
-
- // and more... https://pptr.dev/api/puppeteer.cookieparam
- } as Cookie
- if (!deduped_cookies[unique_id].value) {
- delete deduped_cookies[unique_id]
- continue
- }
- if (deduped_cookies[unique_id].name.startsWith('__')) {
- // cookies that start with __ must be secure, see https://github.com/puppeteer/puppeteer/issues/6806
- deduped_cookies[unique_id].secure = true
- deduped_cookies[unique_id].sourceScheme = 'Secure'
- }
- if (deduped_cookies[unique_id].domain.startsWith('.')) {
- deduped_cookies[unique_id].sameParty = false
- deduped_cookies[unique_id].domain = deduped_cookies[unique_id].domain.slice(1)
- }
-
- for (const key of Object.keys(deduped_cookies[unique_id])) {
- if (!allowed_cookie_attrs.includes(key)) {
- delete deduped_cookies[unique_id][key]
- }
- }
- } catch(err) {
- console.error('[❌] Failed to parse cookie during deduping', cookie)
- throw err
- }
- }
- // console.log(`[🍪] Deduped ${len_before} cookies to ${Object.keys(deduped_cookies).length}...`)
- return Object.values(deduped_cookies) as Cookie[]
- }
- async function loadCookiesTxt() {
- const cookies = [] as Cookie[]
- return cookies // write-only from chrome -> files for now
- if (fs.existsSync(COOKIES_TXT_PATH)) {
- // console.log(`[🍪] Loading cookies/localStorage/sessionStorage from ${COOKIES_TXT_PATH}...`)
- // Read from to cookies.txt file using tough-cookie + @root/file-cookie-store
- const cookies_store = new FileCookieStore(COOKIES_TXT_PATH, {auto_sync: false, lockfile: false});
- cookies_store.getAllCookiesAsync = util.promisify(cookies_store.getAllCookies);
- const exported_cookies = await cookies_store.getAllCookiesAsync()
- for (const cookie of exported_cookies) {
- const cookie_from_tough = cookie.toJSON()
- const domain = cookie_from_tough.hostOnly ? `.${cookie_from_tough.domain}` : cookie_from_tough.domain
- const cookie_for_puppeteer: Cookie = {
- domain,
- name: cookie_from_tough.key,
- path: cookie_from_tough.path,
- value: cookie_from_tough.value,
- secure: cookie_from_tough.secure || false,
- httpOnly: cookie_from_tough.httpOnly || false,
- session: false,
- expires: (new Date(cookie_from_tough.expires)).valueOf()/1000,
- size: undefined,
- }
- // console.log('COOKIE_FROM_TOUGH_TXT', cookie_from_tough, cookie_for_puppeteer)
- cookies.push(cookie_for_puppeteer)
- }
- }
- }
- type AuthJSON = {
- cookies: Cookie[],
- sessionStorage: any,
- localStorage: any,
- }
- async function loadAuthStorage(page, {client}, {apply=true}={}) {
- var {
- cookies,
- sessionStorage,
- localStorage,
- }: AuthJSON = {cookies: [], sessionStorage: {}, localStorage: {}}
-
- if (!LOAD_AUTH_STORAGE) {
- // dont read auth from filesystem auth.json/cookies.txt, just rely on existing cookies in chrome profile
- return {cookies, sessionStorage, localStorage}
- }
- if (fs.existsSync(COOKIES_TXT_PATH)) {
- try {
- cookies = await loadCookiesTxt()
- } catch(err) {
- console.warn('[⚠️] Loaded invalid cookies.txt, moved it to cookies.txt.corrupted (did two processes try to change it at the same time?)')
- await fs.promises.rename(COOKIES_TXT_PATH, COOKIES_TXT_PATH + '.corrupted')
- }
- // console.log(`[🍪] Loading cookies from cookies.txt...`, cookies.length)
- }
- if (fs.existsSync(AUTH_JSON_PATH)) {
- try {
- var {
- cookies: auth_json_cookies,
- sessionStorage,
- localStorage,
- } = JSON.parse(await fs.promises.readFile(AUTH_JSON_PATH, 'utf-8'));
- cookies = [...cookies, ...auth_json_cookies]
- // console.log(`[🍪] Loading cookies from auth.json...`, auth_json_cookies.length)
- } catch(err) {
- console.warn('[⚠️] Loaded invalid auth.json, moved it to auth.json.corrupted (did two processes try to change it at the same time?)')
- await fs.promises.rename(AUTH_JSON_PATH, AUTH_JSON_PATH + '.corrupted')
- }
- }
- cookies = dedupeCookies(cookies)
- if (apply) {
- console.log(`[🍪] Loading stored cookies/localStorage/sessionStorage into session...`, cookies.length)
- // if (cookies?.length) {
- // try {
- // // try setting all at once first (much faster)
- // await page.setCookie(...cookies)
- // } catch(err) {
- // // if any errors, fall back to setting one-by-one so that individual error can be caught
- // for (const cookie of cookies) {
- // try {
- // await page.setCookie(cookie);
- // } catch(err) {
- // console.error('[❌] Failed to set cookie', cookie)
- // throw err
- // }
- // }
- // }
- // }
- const origin = await page.evaluate(() => window.location.origin)
- await page.evaluate((savedSessionStorage) => {
- for (const [key, value] of Object.entries(savedSessionStorage)) {
- sessionStorage[key] = value;
- }
- }, sessionStorage[origin] || {});
-
- await page.evaluate((savedLocalStorage) => {
- for (const [key, value] of Object.entries(savedLocalStorage)) {
- localStorage[key] = value;
- }
- }, localStorage[origin] || {});
- // origin/auth context changes when we do page.goto so we have to hook pageload and apply it then as well
- // https://stackoverflow.com/questions/51789038/set-localstorage-items-before-page-loads-in-puppeteer
- await page.evaluateOnNewDocument(({sessionStorage, localStorage}) => {
- const origin = window.location.origin;
- for (const [key, value] of Object.entries(sessionStorage[origin] || {})) {
- window.sessionStorage.setItem(key, value as string)
- }
- for (const [key, value] of Object.entries(localStorage[origin] || {})) {
- window.localStorage.setItem(key, value as string)
- }
-
- }, {sessionStorage, localStorage});
- }
- return {cookies, sessionStorage, localStorage}
- }
- async function loadCloudflareCookie(page, {original_url}, {timeout=20_000}={}) {
- // make request to FlareSolverr server to get magic cookies that let us bypass cloudflare bot detection
- // docker run -p 8191:8191 -e LOG_LEVEL=info ghcr.io/flaresolverr/flaresolverr
- // alternatives if this stops working:
- // - https://github.com/omkarcloud/botasaurus
- // - https://github.com/ultrafunkamsterdam/nodriver
- // - https://github.com/Akmal-CloudFreed/CloudFreed-CloudFlare-bypass
- // - https://github.com/VeNoMouS/cloudscraper
- const query = { url: original_url, cmd: "request.get", maxTimeout: timeout }
- try {
- const response = await fetch(FLARESOLVERR_API_ENDPOINT, {
- method: 'POST',
- headers: {'Content-Type': 'application/json'},
- body: JSON.stringify(query),
- });
- const data = await response.json();
- const new_cookies = (data?.solution?.cookies || []).map(cookie => ({
- ...cookie,
- 'expires': 2147483640, // overwrite expiration to 32bit maximum timestamp (2038-01-18)
- 'secure': false, // cookie value is plain text (not encrypted/encoded)
- }))
- if (new_cookies.length) {
- console.log(`[☑️] Got Cloudflare bypass cookies (${new_cookies.length}) from FlareSolverr API...`)
- await page.setCookie(...new_cookies);
- return new_cookies
- } else {
- const error_str = JSON.stringify(data?.message || data, null, 4)
- throw `Bad FlareSolverr Response: ${error_str}`
- }
- } catch (error) {
- if (JSON.stringify(error).includes('Challenge not detected')) {
- console.log('[☑️] Page is accessible without FlareSolverr Cloudflare bypass.')
- } else {
- console.warn('[❌] Failed to get Cloudflare bypass cookies from FlareSolverr API.', error)
- }
- }
- return []
- }
- async function setupURLRewriting(page, page_state) {
- await page.setRequestInterception(true);
- const rewrites = URL_REWRITES.sort((a, b) => (a.idx || 0) - (b.idx || 0))
- page.on('request', interceptedRequest => {
- if (interceptedRequest.isInterceptResolutionHandled()) return;
- const original_url = interceptedRequest.url()
- // apply all the rewrites in order to the request URL
- let url = original_url
- for (const rewrite of rewrites) {
- const new_url = url.replace(rewrite.pattern, rewrite.replacement)
- // console.log(rewrite, url, new_url)
-
- // if url is rewritten to an emptystring, abort the request
- if (!new_url) {
- console.warn('[🟥] Request blocked', rewrite.pattern, ':', url)
- interceptedRequest.abort()
- return
- }
- else if (new_url && new_url != url) {
- // console.warn('[📳] Request rewritten', rewrite.pattern, rewrite.replacement, ':', url, '->', new_url)
- console.warn('[📳] Request rewritten', rewrite.pattern, ':', new_url)
- url = new_url
- }
- }
- if (url == original_url) {
- // if url is unchanged, continue request flow as-is
- interceptedRequest.continue()
- } else {
- // otherwise redirect the browser to our rewritten version
- interceptedRequest.respond({
- status: 302,
- headers: {
- location: url,
- 'x-redirect-by': 'ArchiveBox.setupURLRewriting',
- },
- })
- }
- });
- // handled by stopMetadataRecording():
- // page.on('close', () => {
- // page.off('request')
- // page.setRequestInterception(false)
- // })
- }
- async function startMetadataRecording(page, {original_url, version, client, traffic_log, console_log, redirects}) {
- // update helper state on page
- page._original_url = (original_url || (await page.url())).toString()
- // DEBUGGING: helpers for repl() debugging, dont rely on these (global state is badd mmkay)
- // page._client = client || page._client || await page.target().createCDPSession()
- // page._redirects = redirects
- // page._traffic_log = traffic_log
- // add initial entry to page redirect log
- redirects[original_url] = {
- idx: 0,
- url: original_url,
- src: null,
- type: 'Initial',
- wallTime: Date.now()/1000,
- frameId: page.mainFrame()._id,
- requestId: null,
- initiator: {type: "user"},
- isMainFrame: true,
- }
-
- // DEBUGGING: record optional chrome debug trace with screenshots (heavy)
- // try {
- // await page.tracing.stop()
- // await wait(200)
- // } catch(err) {}
- // try {
- // await page.tracing.start({path: TRACE_PATH(page), screenshots: true});
- // } catch(err) {}
- let last_main_frame_url = original_url
- // setup network request intercepts handler
- const addCDPRequestDataListener = (eventName) => {
- client.on(eventName, event => {
- try {
- // save any HTTP/JS redirects to redirects for saveRedirects(page) to use later on
- const new_url = event.documentURL
- const http_status = event.redirectResponse?.status || 0
- const is_new_url = (new_url !== original_url) && !redirects[new_url]
- const is_main_frame_navigation = (event.frameId == page.mainFrame()._id)
- const is_http_redirect = (300 < http_status) && (http_status < 400)
- if (new_url && is_new_url && (is_main_frame_navigation || is_http_redirect) && event.type == 'Document') {
- const new_redirect_entry = {
- url: new_url,
- src: event.redirectResponse?.url || last_main_frame_url,
- type: http_status || 'JS',
- wallTime: Date.now()/1000,
- frameId: event.frameId,
- requestId: event.requestId,
- initiator: event.initiator,
- idx: Object.keys(redirects).length,
- isMainFrame: is_main_frame_navigation,
- }
- redirects[new_url] = new_redirect_entry
- if (is_main_frame_navigation) {
- ALREADY_ARCHIVED.add(new_redirect_entry.url.slice(0, 4096)) // we're already archiving this tab as it redirects, dont create a duplicate archive for the destination
- console.warn(`[➡️] NAVIGATION[${new_redirect_entry.type}]${ANSI.blue} ${last_main_frame_url} ${ANSI.reset}\n ->${ANSI.blue} ${new_redirect_entry.url} ${ANSI.reset}`)
- last_main_frame_url = new_url
- }
- }
- if (event.loaderId) {
- traffic_log[event.loaderId] = traffic_log[event.loaderId] || {} // make sure loader is also in requests list first
- // sometimes it's not in the list if we start archiving too late / after a page's initial request was already made
- }
- // save to traffic_log as {8BC2087A2CCEF28017099C0E10E87440: {Network.eventWillBeSent: {eventId,loaderId, request|response, ...}}
- // https://stackoverflow.com/questions/47078655/missing-request-headers-in-puppeteer?noredirect=1&lq=1
- traffic_log[event.requestId] = traffic_log[event.requestId] || {}
- Object.assign(traffic_log[event.requestId], { [eventName]: event })
-
- // DEBUGGING: log page visits and navigation events to console
- // if (event?.response?.status) {
- // // if we're expecting an HTML response, then we assume it's a page visit & log it to console
- // const acceptMimeType = traffic_log[event.requestId]['Network.requestWillBeSentExtraInfo']?.headers?.accept
- // if (acceptMimeType && acceptMimeType.includes('text/html')) {
- // // log any HTML page responses (less noisy)
- // console.log(`[>] GOT ${event.documentURL}: ${event.response.status} ${event.response.url} (${event.response.mimeType})`)
- // } else {
- // // log ALL responses, inclusing JS,CSS,Images,etc. (very noisy)
- // // console.log(` > ${event.response.status} ${event.response.url} (${event.response.mimeType})`)
- // }
- // }
- } catch(err) {
- console.warn('[X] Error during request/response handler (startMetadataRecording.addCDPRequestDataListener)')
- console.warn(err)
- }
- })
- }
- addCDPRequestDataListener('Network.requestWillBeSent')
- addCDPRequestDataListener('Network.requestWillBeSentExtraInfo')
- addCDPRequestDataListener('Network.responseReceived')
- addCDPRequestDataListener('Network.responseReceivedExtraInfo')
- // clear any existing log entries
- const consolelog_info = {
- TYPE: 'console',
- VERSION: version,
- URL: original_url,
- }
- await overwriteFile(CONSOLELOG_PATH(page), JSON.stringify(consolelog_info) + '\n')
- // record console logs from page
- const appendConsoleLog = async (line) => {
- if (!line) return
- console_log.push(line)
- await fs.promises.appendFile(
- CONSOLELOG_PATH(page),
- line + '\n',
- 'utf-8',
- )
- }
- page.on('console', async(message) =>
- await appendConsoleLog(`${message.type().toUpperCase()} ${message.location()} ${JSON.stringify(message.text())}`))
- page.on('pageerror', async (error) =>
- await appendConsoleLog(error.message || JSON.stringify(error)))
- page.on('requestfailed', async (request) =>
- await appendConsoleLog(`${request.failure()?.errorText} ${request.url() || JSON.stringify(request)}`))
-
- // set puppeteer options on page
- await client.send('Network.enable') // enable network tampering API
- await client.send('Emulation.clearDeviceMetricsOverride'); // clear timing statistics
- await client.send('Page.setDownloadBehavior', {
- behavior: 'allow',
- downloadPath: CHROME_DOWNLOADS_DIR,
- })
- // handled by stopMetadataRecording():
- // page.on('close', () => {
- // try {
- // page.off('request')
- // page.off('console')
- // page.off('pageerror')
- // page.off('requestfailed')
- // page.setRequestInterception(false)
- // } catch(err) {
- // // some versions of puppeteer have had race conditions here where page is already closed by now
- // console.warn('[X] Error in page close handler', err)
- // }
- // })
- return {original_url, client, redirects, traffic_log, console_log}
- }
- async function stopMetadataRecording(page, _page_state) {
- console.log('[🪝] Stopping CDP event hooks and request interception...')
- try {
- page.off('request')
- page.off('response')
- page.off('console')
- page.off('pageerror')
- page.off('requestfailed')
- page.off('hashchange')
- page.setRequestInterception(false)
- // page.tracing.stop()
- } catch(err) {
- // some versions of puppeteer have had race conditions here where page is already closed by now
- console.warn('[X] Error in page close handler', err)
- }
- }
- /********************** Human Behavior Emulation ******************************/
- async function solveCaptchas(page, page_state, {timeout=90_000}={}) {
- // using puppeteer-extra-plugin-recaptcha auto-solver
- // await page.solveRecaptchas()
- // using 2captcha-solver extension auto-solver
- try {
- // console.log('[🕑] Waiting for CAPTCHA to appear...')
- await page.waitForSelector('.captcha-solver', {timeout: 5_000})
- console.log('[🤖] CAPTCHA challenge found, submitting to 2Captcha for solving...')
- await page.click('.captcha-solver')
- console.log(`[🧠] Waiting up to ${timeout/1000}s for CAPTCHA to be solved...`)
- await page.waitForSelector(`.captcha-solver[data-state="solved"]`, {timeout})
- console.log('[🔓] CAPTCHA solution retrieved from 2captcha.')
- } catch(err) {
- console.log('[☑️] No CATPCHA challenges found, site thinks we are human.')
- }
- }
- async function jiggleMouse(page, page_state, {timeout=600}={}) {
- console.log(`[🐁] Moving mouse around randomly for ${timeout/1000}s...`)
- const randomPoint = await getRandomPagePoint(page)
- const cursor = createCursor(page, randomPoint, true)
- cursor.toggleRandomMove(true)
- await wait(timeout/2);
- await cursor.moveTo({x: DEFAULT_VIEWPORT.width/2, y: DEFAULT_VIEWPORT.height/2});
- await wait(timeout/2);
- cursor.toggleRandomMove(false)
- }
- async function blockRedirects(page, {original_url}) {
- page.on('request', req => {
- if (req.isInterceptResolutionHandled()) return;
- // if it's a top-level navigation event to a new url
- if (req.isNavigationRequest() && req.frame() === page.mainFrame() && req.url() !== original_url) {
- req.abort('aborted');
- console.warn('[🟥] Blocked page attempt to naviage to new URL', req.url())
- } else {
- req.continue();
- }
- });
- // handled by stopMetadataRecording():
- // page.on('close', () => {
- // page.off('request')
- // page.setRequestInterception(false)
- // })
- await page.setRequestInterception(true);
- }
- async function blockJSExecution(page, _page_state) {
- console.warn('[🟥] Stopping all JS execution on page...')
- await page.evaluate(() => {
- debugger;
- })
- // OR alternatively this (more buggy, breaks many sites):
- // const html = await page.content();
- // page.setJavaScriptEnabled(false);
- // await page.setContent(html, { waitUntil: 'networkidle0' }); // 4
- }
- async function scrollDown(page, _page_state, {timeout=120_000, scroll_delay=SCROLL_DELAY, scroll_distance=SCROLL_DISTANCE, scroll_limit=SCROLL_LIMIT}={}) {
- const starting_height = await page.evaluate('document.body.scrollHeight');
- let last_height = starting_height
- let scroll_count = 0;
- let scroll_position = scroll_count * scroll_distance
- // await page.bringToFront()
- // scroll to top
- await page.evaluate(() => { window.scrollTo({ top: 0, left: 0, behavior: 'smooth' }); });
- while ((scroll_count < scroll_limit) && ((scroll_delay * scroll_count) < timeout)) {
- console.log(`[⬇️] Scrolling down ${scroll_count}x 1000px... (${scroll_position}/${last_height})`)
- await page.evaluate((y_offset) => { window.scrollTo({ top: y_offset, left: 0, behavior: 'smooth' }); }, scroll_position);
- scroll_count++
- scroll_position = scroll_count * scroll_distance
- // check if any new content was added / if we are infiniscrolling
- let new_height = await page.evaluate('document.body.scrollHeight')
- const added_px = new_height - last_height
- if (added_px > 0) {
- console.log('[✚] Detected infini-scrolling...', `${last_height}+${added_px} => ${new_height}`)
- } else if (scroll_position >= new_height + scroll_distance) {
- // we've reached the bottom, condition isn't true until we've tried to go n+1 past the end (which is fine)
- if (scroll_count > 2)
- break
- }
- last_height = new_height
-
- // sleep 2s, perform the smooth scroll down by 1000px, and increment the counter
- await wait(scroll_delay);
- // facebook watch pages infiniscroll (more and more recommendations forever), stop them after 3 pages
- if (page._original_url.startsWith('https://www.facebook.com/watch/?v') && scroll_count > 3) break
- }
- // scroll to bottom
- if (scroll_position < last_height) {
- await page.evaluate(() => { window.scrollTo({ top: document.body.scrollHeight, left: 0, behavior: 'smooth' }); });
- await wait(scroll_delay)
- await page.evaluate(() => { window.scrollTo({ top: document.body.scrollHeight, left: 0, behavior: 'smooth' }); });
- }
- // Always wait an additional 2sec at the end for scroll animations / loading / rendering to settle down
- console.log('[📉] Reached bottom of the page.', `(${scroll_position}/${last_height})`)
- await wait(scroll_delay);
- await page.evaluate(() => { window.scrollTo({ top: 0, left: 0, behavior: 'smooth' }); });
- await wait(scroll_delay);
- return last_height
- }
- async function disableAnimations(page, _page_state) {
- console.log(`[⛄️] Disabling all animations using CSS override...`)
- // https://stackoverflow.com/questions/53167644/injecting-css-into-site-with-puppeteer
- const css_override = `*, *::before, *::after {
- -moz-animation: none !important;
- -moz-transition: none !important;
- animation: none !important;
- transition: none !important;
- caret-color: transparent !important;
- }`
- // inject override into current page
- await page.addStyleTag({content: css_override});
- // inject override into any subsequently navigated pages
- await page.evaluateOnNewDocument((css_override) => {
- const style_tag = document.createElement('style')
- style_tag.type = 'text/css'
- style_tag.innerHTML = css_override
- document.getElementsByTagName('head')[0].appendChild(style_tag)
- }, css_override);
- }
- async function expandComments(page, _page_state, {timeout=120_000, limit=15_000, delay=650}={}) {
- console.log(`[🗃️] Expanding up to ${limit} comments every ${delay}ms...`)
-
- // expand all <details> sections in Github READMEs, HedgeDoc pages, etc.
- await page.$$eval('pierce/article details', elem => {elem.open = true}) // expand Github README details sections
- await page.$$eval('pierce/div.js-discussion details:not(.details-overlay)', elem => {elem.open = true}) // expand Github issue discussion hidden comments
- await page.$$eval('pierce/.markdown-body details', elem => {elem.open = true}) // expand HedgeDoc Markdown details sections
- await page.exposeFunction('onHashChange', url => page.emit('hashchange', url));
- await page.evaluateOnNewDocument(() => {
- // @ts-ignore
- addEventListener('hashchange', (e) => onHashChange(location.href));
- });
- // Listen for hashchange events in node Puppeteer code.
- page.on('hashchange', url => console.log('Page tried to navigate to:', new URL(url)));
- const num_expanded = await page.evaluate(async ({timeout, limit, delay}) => {
- function getElementsByXPath(xpath, ctx?) {
- var results = [];
- var xpathResult = document.evaluate(
- xpath, // e.g. //*[text()='"+text+"']
- ctx || document,
- null,
- XPathResult.ORDERED_NODE_ITERATOR_TYPE,
- null
- );
- var node;
- while ((node = xpathResult.iterateNext()) != null) {
- results.push(node);
- }
- return results;
- }
- let num_expanded = 0
- const getLoadMoreLinks = () => [
- // find all the buttons/links to expand collapsed/hidden/lazy-loaded content
- ...document.querySelectorAll('faceplate-partial[loading=action]'), // new reddit
- ...document.querySelectorAll('a[onclick^="return morechildren"]'), // old reddit show more replies
- ...document.querySelectorAll('a[onclick^="return togglecomment"]'), // old reddit show hidden replies
- // ...document.querySelectorAll('a.js-show-link'), // stack overflow comments show more (TODO: make this only work on SO)
- // ...document.querySelectorAll('a.morelink'), // HackerNews profile show more (TODO: make this only work on HN)
- // ...getElementsByXPath("//*[text()~='View \d+ replies']"), // facebook comment expander
- ...getElementsByXPath("//*[text()='Show more replies']"), // twitter infiniscroll expander
- ...getElementsByXPath("//*[text()='Show replies']"), // twitter replies expander
- ]
- const wait = (ms) => new Promise(res => setTimeout(res, ms))
- let load_more_links = getLoadMoreLinks()
- while (load_more_links.length) {
- console.log('Expanding comments...', load_more_links.length)
- for (const link of load_more_links) {
- link.scrollIntoView({behavior: 'smooth'})
- if (link.slot == 'children') {
- continue
- // patch new reddit "More replies" links that would open in a new window to display inline instead
- // const comment_id = link.src.split('?')[0].split('/').at(-1)
- // link.slot = `children-${comment_id}-0`
- // link.__alwaysShowSlot = false
- }
- // click the "More replies" button
- link.click()
- num_expanded++
- await wait(delay)
- const time_elapsed = num_expanded * delay
- if ((num_expanded > limit) || (time_elapsed > timeout))
- return num_expanded
- }
- load_more_links = getLoadMoreLinks()
- }
- return num_expanded
- }, {timeout, limit, delay});
- page.off('hashchange')
- if (num_expanded) {
- console.log(`[🗃️] Expanded ${num_expanded} comments...`)
-
- // scroll to bottom, then back up to top
- const final_height = await page.evaluate('document.body.scrollHeight');
- await page.evaluate((top) => { window.scrollTo({ top, left: 0, behavior: 'smooth' }); }, final_height + 1000);
- await wait(delay);
- await page.evaluate(() => { window.scrollTo({ top: 0, left: 0, behavior: 'smooth' }); });
- await wait(delay);
- }
- }
- async function submitForm(page, _page_state, {timeout=5_000}={}) {
- try {
- await page.waitForSelector('form button[type=submit]', {timeout: 1_500});
- console.log('[☑️] Submitting form...')
- await page.click('form button[type=submit]')
- await page.waitForNavigation({timeout});
- await page.goBack();
- } catch (err) {
- // no form found
- }
- }
- // TODO: add an evasion to set navigator.connection.rtt = 365 (0 = detectable as headless)
- /******************************************************************************/
- /******************************************************************************/
- /**************** Extension-Based Archive Output Tasks ************************/
- async function saveSinglefile(page, {main_response, extensions}) {
- const extension = extensions.filter(({name}) => name === 'singlefile')[0]
- if (!extension.version) throw 'Could not find Singlefile extension ID, is it installed?'
- const url = await page.url() || main_response.url()
- if (URL_SCHEMES_IGNORED.includes(url.split(':')[0])) return null
- // get list of existing past files in downloads/* to ignore
- const files_before = new Set(
- (await fs.promises.readdir(CHROME_DOWNLOADS_DIR))
- .filter(fn => fn.endsWith('.html'))
- );
- const out_path = SINGLEFILE_PATH(page)
- console.log(`[🛠️] Saving Singlefile HTML using extension (${extension.id})...`.padEnd(82+1), prettyPath(CHROME_DOWNLOADS_DIR))
- await page.bringToFront() // action button acts on the foreground tab, so it has to be in front :(
- await extension.dispatchAction()
- let files_new = []
- const check_delay = 3_000
- for (const _try in [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]) {
- await wait(check_delay)
- const files_after = (await fs.promises.readdir(CHROME_DOWNLOADS_DIR)).filter(fn => fn.endsWith('.html'));
- files_new = files_after.filter(file => !files_before.has(file))
- if (files_new.length == 0) {
- // console.warn(` ...waiting for Singlefile to write HTML into ${CHROME_DOWNLOADS_DIR}...`)
- continue
- }
- // iterate through new downloads and find a matching .html containing our page's URL in the header
- for (const file of files_new) {
- const dl_path = path.join(CHROME_DOWNLOADS_DIR, file)
- const dl_text = await fs.promises.readFile(dl_path, 'utf-8')
- const dl_header = dl_text.split('meta charset')[0]
- if (dl_header.includes(`url: ${url}`)) {
- /// dont need this check anymore as now all output is versioned:
- // if (fs.existsSync(out_path)) {
- // const {size: existingSize} = await fs.promises.stat(out_path)
- // const {size: newFileSize} = await fs.promises.stat(dl_path)
- // if (newFileSize < existingSize) {
- // console.log(`[🗑️] Discarding singlefile output (${file}) as it's smaller than existing ${out_path}...`)
- // await fs.promises.rm(dl_path)
- // return out_path
- // }
- // }
- console.log(`[✍️] Moving Singlefile download from ${file}...`.padEnd(82), prettyPath(out_path))
- await fs.promises.rename(dl_path, out_path)
- return out_path
- }
- }
- }
- console.warn(`[❌] Couldn't find matching Singlefile HTML in ${CHROME_DOWNLOADS_DIR} after waiting ${(check_delay*10)/1000}s:`, files_new.join(', '))
- return null
- }
- async function saveArchiveWebPage(page, {extensions}, {timeout=30_000}={}) {
- // TODO: waiting on them to expose commands so we can generate .wacz easily
- // https://github.com/webrecorder/archiveweb.page/issues/207
- // ...
- const browser = await page.browser()
- const extension = extensions.filter(({name}) => name === 'archivewebpage')[0]
- await page.bringToFront()
- await extension.dispatchPopup()
- await extension.dispatchAction()
- const popup = await browser.waitForTarget(
- target => target.url().toString().startsWith(`chrome-extension://${extension.id}/popup.html`),
- {timeout: 5_000},
- )
- await page.bringToFront()
- // await puppeteer.Locator.race([
- // popup.locator('::-p-aria(Start With Autopilot)'),
- // popup.locator('wr-popup-viewer >>>> input'),
- // popup.locator(':scope >>> input')
- // ])
- // .setTimeout(timeout)
- // .click({
- // offset: {
- // x: 7.7265625,
- // y: 7.203125,
- // },
- // });
- // @ts-ignore
- await puppeteer.Locator.race([
- popup.locator('wr-popup-viewer >>>> div.status-row > p'),
- popup.locator(':scope >>> div.status-row > p'),
- popup.locator('::-p-text(Recording: \n)')
- ]).setTimeout(timeout).click({
- delay: 733.3000000007451,
- offset: {
- x: 293,
- y: 13.5,
- },
- })
- await wait(8_000)
- // @ts-ignore
- await puppeteer.Locator.race([
- popup.locator('wr-popup-viewer >>>> div:nth-of-type(2) > button > span:nth-of-type(2)'),
- popup.locator(':scope >>> div:nth-of-type(2) > button > span:nth-of-type(2)'),
- popup.locator('::-p-text(Stop)')
- ]).setTimeout(timeout).click({
- offset: {
- x: 7.859375,
- y: 23.203125,
- },
- });
- return null
- }
- async function savePocket(page, {extensions}) {
- const browser = await page.browser()
- const extension = extensions.filter(({name}) => name === 'pocket')[0]
- if (!extension.version) throw 'Could not find Pocket extension ID, is it installed?'
- console.log(`[🛠️] Saving URL to Pocket API using extension (${extension.id})...`, 'https://getpocket.com/saves')
- await page.bringToFront() // action button acts on the foreground tab, so it has to be in front
- await extension.dispatchAction()
- try {
- const login_window = await browser.waitForTarget(
- target => target.url().toString().startsWith('https://getpocket.com/'),
- {timeout: 3_000},
- )
- // login window will open if pocket is not signed-in
- if (login_window) return false
- } catch(e) {
- // no new window should open if it saves correctly
- return true
- }
- }
- /***************** Synchronous Archive Output Tasks ***************************/
- async function saveScreenrecording(page, page_state, {save_gif=true}={}) {
- if (page_state.recorder) {
- const duration = Date.now() - page_state.start_ts
- console.log(`[🎥] Saving screen-recording video (${duration/1000}s)...`.padEnd(82), prettyPath(SCREENRECORDING_PATH(page)))
- const recorder = page_state.recorder
- page_state.recorder = null
- await recorder.stop()
- // create symlink for legacy path
- const snap_dir = page_state.snapshot_dir
- const legacy_path = path.join(snap_dir, 'media', 'screenrecording.mp4')
- await overwriteSymlink(SCREENRECORDING_PATH(page), legacy_path, {relative: snap_dir, search_limit: snap_dir})
- // // remove duplicate frames (white frames at start while it loads + static image at end)
- // const video_path = SCREENRECORDING_PATH(page)
- // const short_path = video_path.replace('.mp4', '.short.mp4')
- // try {
- // await exec(
- // // create a shortened video starting from 0:02s to 0:01s with duplicate frames removed (can look jumpy sometimes)
- // `ffmpeg -ss 2 -sseof -1 -y -i ${video_path} -vf mpdecimate,setpts=N/FRAME_RATE/TB ${short_path}`
- // )
- // } catch(err) {
- // console.log('[❌] Failed to shorten screenrecording.mp4')
- // }
- // convert video to GIF
- if (save_gif) {
- try {
- const BIN_NAME = '/Volumes/NVME/Users/squash/bin/ffmpeg'
- const child = child_process.spawn(
- BIN_NAME,
- [
- '-hide_banner',
- '-loglevel', 'error',
- '-ss', '3',
- '-t', '10',
- '-y',
- '-i', SCREENRECORDING_PATH(page),
- '-vf', "fps=10,scale=1024:-1:flags=bicubic,split[s0][s1];[s0]palettegen[p];[s1][p]paletteuse",
- '-loop', '0',
- SCREENRECORDGIF_PATH(page),
- ],
- {
- cwd: path.dirname(SCREENRECORDING_PATH(page)),
- timeout: 60_000,
- // stdio: [null, 'pipe', 'pipe'],
- stdio: 'ignore',
- detached: true, // run in background, don't block on response
- },
- )
- await blockUntilExists(SCREENRECORDGIF_PATH(page), {min_bytes: 100, timeout: 40_000})
- console.log(`[🎥] Saved screen-recording GIF with ffmpeg pid=${child.pid} (${duration/1000}s)...`.padEnd(82), prettyPath(SCREENRECORDGIF_PATH(page)))
- const snap_dir = page_state.snapshot_dir
- const legacy_path = path.join(snap_dir, 'media', 'screenrecording.gif')
- await overwriteSymlink(SCREENRECORDGIF_PATH(page), legacy_path, {relative: snap_dir, search_limit: snap_dir})
- } catch(err) {
- console.log('[❌] Failed to convert video to GIF:', err)
- }
- }
- return SCREENRECORDING_PATH(page)
- }
- return null
- }
- async function saveScreenshot(page, _page_state, {aspect_ratio=SCREENSHOT_ASPECT_RATIO, width=null, height=null, jpg_width=1440, jpg_quality=90, timeout=30_000}={}) {
- try {await fs.promises.unlink(SCREENSHOT_PATH(page))} catch(err) {}
-
- // setup width and height
- width = width || DEFAULT_VIEWPORT.width
- assert((typeof width === 'number') && width > 200)
- height = height || Math.floor(width/aspect_ratio)
- assert((typeof height === 'number') && height > 200)
-
- console.log(`[📸] Saving full-page screenshot (${width}x${height}px)...`.padEnd(82), prettyPath(SCREENSHOT_PATH(page)))
- // set width, height, and deviceScale factor: https://github.com/puppeteer/puppeteer/issues/1576
- await page.setViewport({ ...DEFAULT_VIEWPORT, width, height, deviceScaleFactor: 2})
- await page.bringToFront()
- await wait(1_250) // page takes a sec settle after foregrounding and viewport update
- // take lossless fullpage screenshot of 1920x1440+px (4:3+) -> ./screenshot.png
- await page.screenshot({ path: SCREENSHOT_PATH(page), fullPage: true, type: 'png' })
-
- // wait for the screenshot to be created, then set the viewport to the next size
- await blockUntilExists(SCREENSHOT_PATH(page), {min_bytes: 100, timeout})
- await wait(6_000) // puppeteer takes a while to finish writing png data when fullPage: true
-
- const jpg_height = Math.floor(jpg_width/aspect_ratio)
- await page.setViewport({ ...DEFAULT_VIEWPORT, width: jpg_width, height: jpg_height, deviceScaleFactor: 2})
- await wait(1_250) // page takes a sec settle after foregrounding and viewport update
- // WARNING: make sure you never try to create two screenshots at the same time (especially not fullpage screenshots)
- // thats why there are all these delays here.
- // screenshot creation messes up the whole viewport while it's running,
- // and it writes bad/white empty screenshots if you try to make more than one concurrently
- // take compressed screenshot of jpg_width*jpg_height (4:3) -> ./screenshot.jpg
- await page.screenshot({
- path: SCREENSHOT_JPG_PATH(page),
- type: 'jpeg',
- quality: jpg_quality,
- clip: {
- x: 0,
- y: 0,
- width: jpg_width,
- height: jpg_height,
- },
- captureBeyondViewport: false,
- });
- await blockUntilExists(SCREENSHOT_JPG_PATH(page), {min_bytes: 100, timeout: timeout/2})
- console.log(`[📸] Saved screenshot as screenshot.jpg (${jpg_width}x${jpg_height}px)...`.padEnd(82), prettyPath(SCREENSHOT_JPG_PATH(page)))
- // reset viewport back to defaults
- await wait(1_250)
- await page.setViewport(DEFAULT_VIEWPORT)
- // ALTERNATIVE METHOD based on cropping fullpage png and converting to jpg manually:
- // import {PNG} from 'pngjs';
- // import jpeg from 'jpeg-js';
- // setTimeout(async () => {
- // try {
- // const screenshot_png = SCREENSHOT_PATH(page);
- // const screenshot_jpg = SCREENSHOT_JPG_PATH(page)
- // const jpg_max_height = height
- // const jpg_quality = quality; // Adjust the quality as needed (0-100)
- // fs.createReadStream(screenshot_png)
- // .pipe(new PNG())
- // .on('parsed', function () {
- // const width = this.width;
- // const height = this.height;
-
- // let cropped_height = height;
- // if (height > jpg_max_height) {
- // cropped_height = jpg_max_height;
- // }
-
- // const cropped_bytes = new Uint8Array(width * cropped_height * 4);
- // for (let y = 0; y < cropped_height; y++) {
- // for (let x = 0; x < width; x++) {
- // const idx = (width * y + x) << 2;
- // cropped_bytes[idx] = this.data[idx];
- // cropped_bytes[idx + 1] = this.data[idx + 1];
- // cropped_bytes[idx + 2] = this.data[idx + 2];
- // cropped_bytes[idx + 3] = this.data[idx + 3];
- // }
- // }
-
- // const jpeg_obj = {
- // data: cropped_bytes,
- // width: width,
- // height: cropped_height,
- // };
-
- // const jpeg_bytes = jpeg.encode(jpeg_obj, jpg_quality);
- // fs.writeFileSync(screenshot_jpg, jpeg_bytes.data);
- // console.log(`[📸] Saved screenshot as screenshot.jpg (${width}x${jpg_max_height}px)...`.padEnd(82), prettyPath(SCREENSHOT_JPG_PATH(page)))
- // });
- // } catch(err) {
- // console.error('[X] Error while generating JPG screenshot', SCREENSHOT_JPG_PATH(page), err)
- // }
- // }, DELAY_BEFORE_JPG_CONVERSION)
- // ALTERNATIVE METHOD TO WRITE SCREENSHOT JPG:
- // await wait(5_000) // puppeteer takes a while to finish writing png data when fullPage: true
- // if ((await page.evaluate('document.body.scrollHeight')) > max_height) {
- // // if page exceeds max_height, save additional cropped screenshot as screenshot.top.png
- // // (needed b.c. uncropped screenshot may have insane 1:20+ aspect ratio that is hard to use elsewhere)
- // await page.screenshot({ path: SCREENSHOT_JPG_PATH(page), type: 'jpg', quality: 100})
- // await wait(1_000) // page takes a sec settle after a screenshot
- // }
- return SCREENSHOT_PATH(page)
- }
- async function savePDF(page, _page_state, {timeout=30_000}={}) {
- const url = page.url() || 'about:blank'
- if (URL_SCHEMES_IGNORED.includes(url.split(':')[0])) return null
- const out_path = PDF_PATH(page)
- console.log(`[📓] Saving print-as-PDF export...`.padEnd(82), prettyPath(out_path))
- await page.bringToFront()
- try {await fs.promises.unlink(PDF_PATH(page))} catch(err) {}
- // await page.emulateMediaType('screen') // print as "@media(screen) instead of @media(print)"
- // page.createPDFStream lets us to save larger PDFs than page.pdf() before crashing
- // (streams to disk in chunks instead of all at once)
- const pdf_stream = await page.createPDFStream({
- timeout: timeout,
- printBackground: true,
- outline: true,
- tagged: true,
- format: 'A4',
- displayHeaderFooter: false,
- // margin: { top: '0.5cm', right: '1cm', bottom: '0.8cm', left: '1cm' },
- })
- const reader = pdf_stream.getReader()
- // iterate through reader and append chunks to out_path
- await fs.promises.rm(out_path, {force: true})
- let num_bytes = 0
- let error = '0 bytes written'
- try {
- while (true) {
- const {done, value} = await reader.read()
- if (done) break;
- await fs.promises.appendFile(out_path, value)
- num_bytes += value.length;
- }
- } catch(error) {
- num_bytes = 0
- }
- if (!num_bytes) {
- console.warn('[❌] Failed to save PDF', JSON.stringify(error, null, 4))
- await fs.promises.rm(out_path, {force: true})
- return null
- }
- return out_path
- }
- async function inlineShadowDOM(page, _page_state, {limit=100_000}={}) {
- console.log(`[😎] Replacing Shadow DOM elements with inline HTML...`)
- try {
- const num_replaced = await page.evaluate((limit) => {
- let num_replaced = 0
- // Returns HTML of given shadow DOM.
- const getShadowDomHtml = (shadowRoot) => {
- let shadowHTML = '';
- for (const el of shadowRoot.childNodes) {
- shadowHTML += el.nodeValue || el.outerHTML;
- }
- return shadowHTML;
- };
- // Recursively replaces shadow DOMs with their HTML.
- const replaceShadowDomsWithHtml = (rootElement) => {
- if (num_replaced > limit) return
- for (const el of rootElement.querySelectorAll('*')) {
- if (el.shadowRoot) {
- replaceShadowDomsWithHtml(el.shadowRoot);
- el.innerHTML += getShadowDomHtml(el.shadowRoot);
- }
- }
- num_replaced++
- };
- replaceShadowDomsWithHtml(document.body);
- return num_replaced
- }, limit)
- // console.log(' √ replaced', num_replaced, 'Shadow DOM trees')
- } catch(err) {
- console.log('[⚠️] Inlining Shadow DOM failed', err)
- }
- }
- async function saveAIQualityAssuranceResult(page, {original_url, version}) {
- console.log(`[🧠] Analyzing screenshot with GPT-4o for QA checks...`.padEnd(82), prettyPath(AIQA_PATH(page)))
-
- let screenshot_path = SCREENSHOT_PATH(page)
- const screenshot_cropped_path = SCREENSHOT_JPG_PATH(page)
- if (fs.existsSync(screenshot_cropped_path)) {
- // screenshot is too tall to pass to openai, send cropped version instead
- screenshot_path = screenshot_cropped_path
- }
- try {
- await blockUntilExists(screenshot_path, {min_bytes: 100, timeout: 7_500})
- } catch (err) {
- console.warn('[❌] Failed to send screenshot to GTP-4o for analysis, no screenshot.{png,jpg} exists', err)
- return null
- }
- var stdout = ''
- var stderr = ''
- let result = null
- const PYTHON_BIN = path.join(__dirname, '.venv/bin/python')
- const SCRIPT_PATH = path.join(__dirname, 'ai_qa.py')
- await blockUntilExists(PYTHON_BIN, {min_bytes: 1, timeout: 250})
- await blockUntilExists(SCRIPT_PATH, {min_bytes: 1, timeout: 250})
- try {
- var {stdout, stderr} = await exec(
- `${PYTHON_BIN} ${SCRIPT_PATH} --attach '${screenshot_path}'`
- )
- result = JSON.parse(stdout.toString())
- if (!result) throw 'Got empty result!'
- result = {
- TYPE: 'aiqa',
- VERSION: version,
- URL: original_url,
- ...result,
- }
- } catch(parse_err) {
- console.warn('[❌] Failed to get OpenAI analysis for screenshot.png', parse_err, stderr)
- }
- if (!(result || stdout)) {
- return null
- }
- await overwriteFile(
- AIQA_PATH(page),
- result || stdout.toString(),
- )
-
-
- return result
- }
- async function saveYTDLP(page, {original_url, version}, {max_size='750m'}={}) {
- console.log(`[🎥] Saving media with YT-DLP (<=${max_size})...`.padEnd(82), prettyPath(YTDLP_PATH(page)))
- await fs.promises.mkdir(YTDLP_PATH(page), {recursive: true})
- const cwd = YTDLP_PATH(page)
- const bin_name = 'yt-dlp'
- const timeout = 300_000 // 5min timeout
- const args = [
- '--restrict-filenames',
- '--trim-filenames', '128',
- '--write-description',
- '--write-info-json',
- '--write-annotations',
- '--write-thumbnail',
- '--no-call-home',
- '--write-sub',
- '--write-auto-subs',
- '--convert-subs=srt',
- '--yes-playlist',
- '--continue',
- '--no-abort-on-error',
- '--ignore-errors',
- '--geo-bypass',
- '--add-metadata',
- `--format=(bv*+ba/b)[filesize<=${max_size}][filesize_approx<=?${max_size}]/(bv*+ba/b)`,
- '--no-check-certificate',
- '--no-progress',
- // `--cookies=${COOKIES_TXT_PATH}`, // using logged in cookies actually makes it fail more often, not sure why
- original_url,
- ]
- const {getResult, ...exec_info} = await saveExecResult(bin_name, args, {original_url, version}, {cwd, timeout})
- return {getResult, ...exec_info}
- }
- async function saveGALLERYDL(page, {original_url, version}) {
- console.log(`[🎥] Saving photos with gallery-dl...`.padEnd(82), prettyPath(GALLERYDL_PATH(page)))
- await fs.promises.mkdir(GALLERYDL_PATH(page), {recursive: true})
- const cwd = GALLERYDL_PATH(page)
- const bin_name = 'gallery-dl'
- const timeout = 300_000 // 5min timeout
- const args = [
- '--verbose',
- '--write-metadata',
- '--write-infojson',
- '--write-tags',
- '--sleep=1.5-2.5',
- `--cookies=${COOKIES_TXT_PATH}`,
- // '--no-check-certificate',
- // `--directory=media`,
- original_url,
- ]
- const {getResult, ...exec_info} = await saveExecResult(bin_name, args, {original_url, version}, {cwd, timeout})
- return {getResult, ...exec_info}
- }
- // async function saveWget(page, {original_url, version}) {
- // console.log(`[⎒] Saving wget site clone...`.padEnd(82), prettyPath(WGET_PATH(page)))
- // const args = [
- // // ...
- // ]
-
- // spawn(
- // 'wget',
- // [
- // ...args,
- // original_url,
- // ],
- // {
- // cwd: WGET_PATH(page),
- // detached: true, // run in background, don't block on response
- // stdio: 'ignore',
- // timeout: 300_000, // 5min timeout
- // },
- // )
- // return {path: WGET_PATH(page)}
- // }
- /**************** Asynchronous Archive Output Tasks ***************************/
- type FaviconCandidate = {
- url: string,
- basename: string,
- extension: string,
- expected_mimetype: string,
- }
- const faviconFromDomain = (url) => {
- // https://auth:[email protected]:1234/a/bc123 -> https://auth:[email protected]:1234/favicon.ico
- const url_origin = (new URL(url)).origin
- return {
- url: url_origin ? `${url_origin}/favicon.ico` : null,
- basename: 'favicon',
- extension: undefined, // auto-detect extension at download time in case it redirects us to a png
- expected_mimetype: 'image/', // only accept image/* to avoid saving html/txt error reponses as icon
- } as FaviconCandidate
- }
- const faviconFromGoogle = (url, size=256) => {
- // https://auth:[email protected]:1234/a/bc123 -> https://www.google.com/s2.favicons?domain=t.co
- const domain = url && (new URL(url)).hostname
- return {
- url: domain?.includes('.') ? `https://www.google.com/s2/favicons?sz=${size},domain=${domain}` : null,
- basename: 'google_favicon',
- extension: 'png',
- expected_mimetype: 'image/png', // google always provides PNGs in response
- } as FaviconCandidate
- }
- const faviconFromHtml = async (page) => {
- // <link rel="icon" src="https://example.com/static/images/favicon.png"/> -> https://example.com/static/images/favicon.png
- let url
- try {
- url = await page.$eval('link[rel*="icon"]', (elem) => elem?.href)
- if (!url || !url.includes('://'))
- url = null
- } catch(err) {
- url = null
- // console.warn('Failed to find favicon tag in html', JSON.stringify(err, null, 4))
- }
- return {
- url,
- basename: 'favicon',
- extension: undefined, // auto-detect extension at download time
- expected_mimetype: 'image/', // accept any image/* mimetype at download time
- } as FaviconCandidate
- }
- type FaviconResult = {
- url: string,
- num_bytes: number,
- abspath?: string,
- dir?: string,
- filename?: string,
- mimeType?: string,
- }
- async function saveFavicon(page, {original_url, main_response, version}) {
- const dir = path.dirname(FAVICON_PATH(page))
- const response_url = main_response?.url()
- const favicon_downloads_to_try: {[key: string]: FaviconCandidate} = unique([
- await faviconFromHtml(page),
- faviconFromDomain(response_url),
- faviconFromDomain(original_url),
- faviconFromGoogle(response_url),
- faviconFromGoogle(original_url),
- ].filter(({url}) => url), 'url')
- const browser = await page.browser()
- // let logs = []
- // let errors = []
- let output_files: {[key: string]: FaviconResult} = {}
- for (const download_options of Object.values(favicon_downloads_to_try)) {
- let result: FaviconResult = {num_bytes: 0, url: download_options.url}
- // {url, num_bytes, abspath, dir, filename, basename, extension, mimeType}
- try {
- // try getting it with node-fetch first
- const response = await fetch(download_options.url) as Response
- const file_options = await detectFilename({...download_options, response, dir})
- if (response.headers.get("content-length")) {
- const favicon_stream = Readable.fromWeb(response.body as any)
- await overwriteFile(file_options.abspath, favicon_stream)
- result = {
- ...file_options,
- num_bytes: parseInt(response.headers.get("content-length") || '0'),
- mimeType: response.headers.get("content-type"),
- }
- } else {
- throw 'Failed to download favicon with fetch()'
- }
- } catch(err) {
- // console.warn('[!] Failed to get favicon with node-fetch', err)
- // fallback to getting it by opening a new browser tab
- result = await download({...download_options, browser, dir, page})
- }
- // logs.push(...(result.logs || []))
- // errors.push(...(result.errors || []))
- if (result.num_bytes) {
- console.log(`[🌠] Saving page favicon (${result.url.substring(0, 35)}... ${result.mimeType})...`.padEnd(82), prettyPath(result.abspath))
- output_files[result.filename] = result
- break // break here stops after the first successful download, comment out to keep going instead
- }
- }
- const output_file = Object.values(output_files).sort(file => file.num_bytes).at(-1)
- const favicon_info = {
- TYPE: 'favicon',
- VERSION: version,
- URL: original_url,
- succeeded: !!output_file,
- // stdout: JSON.stringify(logs),
- // stderr: JSON.stringify(errors),
- favicon_url: output_file?.url,
- favicon_urls: Object.keys(favicon_downloads_to_try),
- favicon_files: Object.keys(output_files).map(fname => fname.replace(dir, '.')),
- favicon_filename: output_file?.filename,
- favicon_num_bytes: output_file?.num_bytes,
- }
- await overwriteFile(FAVICON_PATH(page), favicon_info)
- return favicon_info
- }
- async function saveTitle(page, {original_url, version}) {
- const title_from_browser = (await page.title()) || null
- const title_from_js = await page.evaluate(() => document?.title || null)
- const title_from_html = await page.evaluate(() => document?.querySelector('title')?.innerText || null)
- const title_from_og = await page.evaluate(() => document?.querySelector('meta[property="og:title"]')?.getAttribute('content') || null)
- // best guess at best title = longest title
- const title = ([title_from_html, title_from_og, title_from_js, title_from_browser]
- .filter(title => title)
- .sort((a, b) => b.length - a.length)[0] || '')
- .replaceAll('\n', ' ')
- if (title?.length) {
- console.log(`[📗] Saving page title (${title.substring(0, 40)})...`.padEnd(82), prettyPath(TITLE_PATH(page)))
- await overwriteFile(TITLE_PATH(page), title)
- }
- const title_info = {
- TYPE: 'title',
- VERSION: version,
- URL: original_url,
- title,
- title_from_html,
- title_from_og,
- title_from_js,
- title_from_browser,
- }
- const title_json_path = TITLE_PATH(page).replace('.txt', '.json')
- await overwriteFile(title_json_path, title_info)
- return title_info
- }
- async function saveRaw(page, {main_response}) {
- const response = main_response
- if (!response) {
- console.warn('[⚠️] Failed to save page RAW bytes, main_response is null', response)
- }
- const dir = RAW_PATH(page)
- await fs.promises.mkdir(dir, {recursive: true})
- const {url, abspath, mimeType} = await detectFilename({page, response, dir})
- console.log(`[🔟] Saving raw response bytes (${mimeType})...`.padEnd(82), prettyPath(abspath))
- await download({page, response, abspath})
- return abspath
- }
- async function saveSourceMaps(page, {original_url, version}) {
- console.log(`[🐛] Saving source maps to ./responses/all/*.{js,css}.map...`)
- const response_index_path = path.join(RESPONSES_PATH(page), 'index.jsonl')
- const response_index = await fs.promises.readFile(response_index_path, 'utf-8')
- const urls_to_download = []
- for (const response of response_index.split('\n')) {
- try {
- const {url, extension} = JSON.parse(response)
- if (['css', 'js'].includes(extension?.toLowerCase())) {
- urls_to_download.push(url + '.map')
- }
- } catch(err) { continue }
- }
- // TODO: fix this, it needs to both after stopSavingMetadata and before stopSavingMetadata
- // fix is to use traffic_log to get response url list instead of waiting for index.jsonl to be created
- await page.evaluate(async (urls_to_download) => {
- const promises = []
- for (const sourcemap_url in urls_to_download) {
- promises.push(fetch(sourcemap_url))
- }
- return Promise.allSettled(promises)
- }, urls_to_download)
- return {
- TYPE: 'sourcemaps',
- URL: original_url,
- VERSION: version,
- sourcemaps: urls_to_download,
- }
- }
- async function saveRequests(page, {original_url, version, traffic_log}) {
- console.log(`[📼] Saving requests log (${Object.keys(traffic_log).length})...`.padEnd(82), prettyPath(REQUESTS_PATH(page)))
- const requests_info = {
- TYPE: 'requests',
- VERSION: version,
- URL: original_url,
- requests: traffic_log,
- }
- await overwriteFile(REQUESTS_PATH(page), requests_info)
- return requests_info
- }
- async function saveRedirects(page, {original_url, main_response, traffic_log, redirects, version}) {
- const main_request_id = Object.keys(traffic_log).filter(id => !id.includes('.'))[0]
- const main_response_traffic = traffic_log[main_request_id] || {}
- const url_from_browser = await page.url() || null
- const url_from_request = (
- main_response?.request()?.url()
- || main_response_traffic['Network.requestWillBeSent']?.request?.url
- || null)
- const url_from_response = (
- main_response?.url()
- || main_response_traffic['Network.responseReceived']?.main_response?.url
- || null)
- const http_redirects =
- Object.values(traffic_log)
- .filter(event => event['Network.requestWillBeSent']?.redirectResponse)
- .map(event => event['Network.requestWillBeSent'])
- .map(requestWillBeSent => ({
- url: requestWillBeSent.request.url,
- src: requestWillBeSent.redirectResponse.url,
- status: requestWillBeSent.redirectResponse.status,
- loaderId: requestWillBeSent.loaderId,
- requestId: requestWillBeSent.requestId,
- wallTime: requestWillBeSent.wallTime,
- initiator: requestWillBeSent.initiator,
- isMainFrame: (requestWillBeSent.loaderId == main_request_id),
- }))
-
- const url_parsed = new URL(url_from_response || url_from_request || url_from_browser)
- const redirects_info = {
- TYPE: 'redirects',
- VERSION: version,
- URL: original_url,
- url_parsed,
- url_from_request,
- url_from_response,
- url_from_browser,
- redirects_from_browser: redirects,
- redirects_from_http: http_redirects,
- }
- console.log(`[🔗] Saving page redirects log (${http_redirects.length})...`.padEnd(82), prettyPath(REDIRECTS_PATH(page)))
- await overwriteFile(REDIRECTS_PATH(page), redirects_info)
- return redirects_info
- }
- async function saveHeaders(page, {original_url, version, traffic_log}) {
- const main_request_id = Object.keys(traffic_log).filter(id => !id.includes('.'))[0]
- const main_response_traffic = traffic_log[main_request_id] || {}
- // combine base request with browser-added request headers
- const request = {...main_response_traffic['Network.requestWillBeSent']?.request}
- const request_extra_headers = main_response_traffic['Network.requestWillBeSentExtraInfo']?.headers || {}
- request.headers = {...request.headers, ...request_extra_headers}
- // combine base response with browser-added response headers
- const response = {...main_response_traffic['Network.responseReceived']?.response}
- const response_extra_headers = main_response_traffic['Network.responseReceivedExtraInfo']?.headers || {}
- response.headers = {...response.headers, ...response_extra_headers}
- const headers_info = {
- TYPE: 'headers',
- VERSION: version,
- URL: original_url,
- request,
- response,
- }
- const num_headers = Object.keys({...request.headers, ...response.headers}).length
- if (num_headers) {
- console.log(`[👾] Saving main request & response headers (${num_headers})...`.padEnd(82), prettyPath(HEADERS_PATH(page)))
- await overwriteFile(HEADERS_PATH(page), headers_info)
- }
- return headers_info
- }
-
- async function saveSSL(page, {original_url, version, traffic_log}) {
- const main_request_id = Object.keys(traffic_log).filter(id => !id.includes('.'))[0]
- const main_response_traffic = traffic_log[main_request_id] || {}
- const relevant_response_keys = [
- 'url',
- 'status',
- 'mimeType',
- 'connectionReused',
- 'remoteIPAddress',
- 'remotePort',
- 'fromServiceWorker',
- 'encodedDataLength',
- 'protocol',
- 'alternateProtocolUsage',
- 'securityState',
- 'securityDetails',
- ]
- let ssl_info = Object.entries(main_response_traffic['Network.responseReceived']?.response || {})
- .reduce((obj, [key, val]) => {
- if (relevant_response_keys.includes(key)) {
- obj[key] = val
- }
- return obj
- }, {}) as any
- // TODO: parse SSL certificate sha256 hash from chrome://system/#chrome_root_store
- // const ssl_certificate = await client.send('Network.getCertificate', {origin: original_url})
- // ssl_info.sslCertSha256 = '<unknown>'
- ssl_info = {
- TYPE: 'ssl',
- VERSION: version,
- URL: original_url,
- ...ssl_info,
- }
- if (Object.keys(ssl_info).length-3) {
- console.log(`[🔏] Saving page SSL details (${ssl_info?.securityDetails?.protocol})...`.padEnd(82), prettyPath(SSL_PATH(page)))
- await overwriteFile(SSL_PATH(page), ssl_info)
- }
- return ssl_info
- }
- async function saveDOM(page, {original_url, version}) {
- const html = await page.content();
- console.log(`[📖] Saving DOM dump (${html.length})...`.padEnd(82), prettyPath(DOM_PATH(page)))
- const html_with_header =
- `<!-- Saved by ArchiveBox TYPE=dom VERSION=${version} URL=${original_url} -->\n${html}`
- await overwriteFile(DOM_PATH(page), html_with_header)
- return DOM_PATH(page)
- }
- async function saveBodyText(page, _page_state) {
- const innerText = await page.evaluate(() => document?.body?.innerText);
- if (innerText?.length) {
- console.log(`[📃] Saving body text (${innerText.length})...`.padEnd(82), prettyPath(BODYTEXT_PATH(page)))
- await overwriteFile(BODYTEXT_PATH(page), innerText)
- }
- // // alternative method: emulate Ctrl+A, Ctrl+C (sometimes gets more than body.innerText)
- // const innerText = await page.$eval('*', (el) => {
- // const selection = window.getSelection();
- // const range = document.createRange();
- // range.selectNode(el);
- // selection.removeAllRanges();
- // selection.addRange(range);
- // return window.getSelection().toString();
- // });
- return innerText
- }
- async function savePandoc(page, { original_url, version }) {
- console.log(`[📒] Converting DOM HTML to markdown with Pandoc...`.padEnd(82), prettyPath(PANDOC_PATH(page)))
- let dom_paths = [DOM_PATH(page), SINGLEFILE_PATH(page)].filter(fs.existsSync)
- if (!dom_paths) return null
- const dom_path = dom_paths[0]
- var stdout: string = ''
- var stderr: string = ''
- let result: any = null
- const BIN_NAME = 'pandoc'
- // pandoc --from html --to markdown_github --citeproc --wrap=none --highlight-style=kate
- const args = [
- BIN_NAME,
- '--from=html',
- '--to=markdown_github',
- '--wrap=none',
- '--citeproc',
- '--highlight-style=kate',
- `--output='${PANDOC_PATH(page)}'`,
- dom_path,
- ]
- try {
- ;({ stdout, stderr } = await exec(args.join(' ')));
- stdout = stdout.toString().trim()
- if (!stdout) throw 'Got empty result!'
- result = {
- TYPE: 'pandoc',
- VERSION: version,
- URL: original_url,
- cmd: args,
- markdown_file: PANDOC_PATH(page),
- }
- } catch (parse_err) {
- console.warn('[❌] Failed to run Pandoc HTML to MD conversion', parse_err, stderr)
- }
- if (!stdout) {return null}
- await overwriteFile(
- PANDOC_PATH(page),
- stdout,
- )
- // pandoc --from markdown_github --to html --citeproc --wrap=none --highlight-style=kate
- const reverse_conversion_args = [
- BIN_NAME,
- '--from=markdown_github',
- '--to=html',
- '--wrap=none',
- '--citeproc',
- '--highlight-style=kate',
- `--output='${PANDOC_PATH(page).replace('.md', '.html')}'`,
- PANDOC_PATH(page),
- ]
- try {
- ; ({ stdout, stderr } = await exec(reverse_conversion_args.join(' ')));
- stdout = stdout.toString().trim()
- if (!stdout) throw 'Got empty result!'
- result = {
- ...result,
- html_file: PANDOC_PATH(page).replace('.md', '.html'),
- }
- } catch (parse_err) {
- console.warn('[❌] Failed to run Pandoc MD to HTML conversion', parse_err, stderr)
- }
- if (!result) { return null }
- await overwriteFile(
- PANDOC_PATH(page).replace('.md', '.html'),
- result,
- )
- return result
- }
- async function saveReadability(page, {original_url, version}) {
- const url = await page.url()
- let html = ''
- let article = null
- try {
- html = await page.content()
- if (html.length > 14_000_000) {
- console.warn('[⚠️] Truncating readability article text because html is too long...', html.length)
- html = html.substring(0, 13_900_000)
- }
- const virtualConsole = new VirtualConsole()
- const dom = new JSDOM(html, {url, virtualConsole})
- const reader = new Readability(dom.window.document);
- article = reader.parse()
- } catch(err) {
- console.warn(`[❌] Failed to get readability article text`)
- return null
- }
- if (article) {
- console.log(`[📜] Saving readability article text (${article.textContent?.length})...`.padEnd(82), prettyPath(READABILITY_PATH(page)))
- const {content, textContent, ...metadata} = article
- if (content.trim()) {
- await overwriteFile(READABILITY_PATH(page).replace('.json', '.html'), content);
- }
- if (textContent.trim()) {
- await overwriteFile(READABILITY_PATH(page).replace('.json', '.txt'), textContent);
- }
- const readability_info = {
- TYPE: 'readability',
- VERSION: version,
- URL: original_url,
- ...metadata,
- }
- await overwriteFile(READABILITY_PATH(page), readability_info)
- return readability_info
- }
- return null
- }
- async function saveAccessibility(page, {original_url, version}) {
- // get accessibility tree
- const accessibility_tree = await page.accessibility.snapshot({interestingOnly: true});
- // console.log(accessibility_tree);
- // get iframe tree
- const iframes = []
- function dumpFrameTree(frame, indent='>') {
- iframes.push(indent + frame.url());
- for (const child of frame.childFrames()) {
- dumpFrameTree(child, indent + '>');
- }
- }
- dumpFrameTree(page.mainFrame(), '');
- // console.log(iframes)
- // generate simple table-of-contents of all the key html elements (e.g. h1, h2, h3, article, main, etc.)
- const outline = await page.evaluate(() => {
- const headings = []
- for (const elem of [...document.querySelectorAll("h1, h2, h3, h4, h5, h6, a, header, footer, article, main, aside, nav, section, figure, summary, table, form, iframe")] as HTMLElement[]) {
-
- // skip a tags that aren't named anchors
- if (elem.tagName.toLowerCase() == 'a' && !(elem as HTMLAnchorElement).name) continue
- // e.g. article #main-article
- const elem_id = ((typeof elem.id === 'string' && elem.id) || (elem as HTMLAnchorElement).name || elem.ariaLabel || elem.role || '')
- const elem_classes = elem.className.trim().split(' ').slice(0, 3).join(' .') || ''
- const elem_action = (elem as any).action?.split('/')?.slice(-1)?.join('/')
- const summary = elem.innerText.length > 128
- ? `${elem.innerText?.slice(0, 128)}...`
- : elem.innerText
- let prefix = ''
- let title = (elem_id ? `#${elem_id}` : '')
- if (!title && elem_classes) title = `.${elem_classes}`
- if (elem_action) title = `${title} /${elem_action}`
- if (summary) title = `${title}: ${summary}`
- // if elem is a header, prepend a #### prefix based on its level
- const level = Number(elem.tagName.toLowerCase().replace('h', ''))
- if (!isNaN(level)) {
- prefix = '#'.repeat(level)
- title = elem.innerText || elem_id || elem_classes
- } else {
- // set prefix to element's breadcrumb path
- let node = elem
- const parents = [elem.tagName?.toLowerCase().trim()]
- while (node) {
- // add each parent element's name to the path
- // const elem_type = node.tagName?.toLowerCase().trim() || ''
- // if (elem_type && !['div', 'span', 'p', 'body', 'html'].includes(elem_type)) {
- // parents.unshift(elem_type);
- // }
- parents.unshift('') // add emptystring to abbreviate path as >>>> istead of main>article>header>div>...
- node = node.parentNode as HTMLElement
- }
- prefix = parents.join('>')
- }
- // strip all repeated whitespace and newlines
- title = title.replaceAll('\n', ' ').replace(/\s+/g, ' ').trim()
- if (prefix) {
- headings.push(`${prefix} ${title}`)
- }
- }
- // console.log(headings.join('\n'))
- return headings
- })
- console.log(`[🩼] Saving accessibility outline (${Object.keys(accessibility_tree).length})...`.padEnd(82), prettyPath(ACCESIBILITY_PATH(page)))
- // console.log(outline.filter(line => line.startsWith('#')).join('\n'))
- const accessibility_info = {
- TYPE: 'accessibility',
- VERSION: version,
- URL: original_url,
- iframes,
- headings: outline,
- tree: accessibility_tree,
- }
- await overwriteFile(
- ACCESIBILITY_PATH(page),
- accessibility_info,
- )
- return accessibility_info
- }
- async function saveSEO(page, {original_url, version}) {
- // collect all <meta name="title" property="og:title" content="Page Title for SEO | Somesite.com"> tags into dict
- const seo_vars = await page.evaluate(() =>
- [...document.querySelectorAll('meta')]
- .map(tag => ({key: tag.getAttribute('name') || tag.getAttribute('property') || '', value: tag.getAttribute('content') || ''}))
- .filter(obj => obj.key && obj.value)
- .sort((a, b) => a.value.length - b.value.length)
- .reduce((acc, node) => {acc[node.key] = node.value; return acc}, {})
- )
- const seo_info = {
- TYPE: 'seo',
- VERSION: version,
- URL: original_url,
- ...seo_vars,
- }
- const num_vars = Object.keys(seo_vars).length
- if (num_vars) {
- console.log(`[🔎] Saving page SEO metadata (${num_vars})...`.padEnd(82), prettyPath(SEO_PATH(page)))
- await overwriteFile(SEO_PATH(page), seo_info)
- }
- return seo_info
- }
- async function saveOutlinks(page, {original_url, version}) {
- // TODO: slow to iterate over all elements so many times, perhaps we can collapse everything down into one loop
- // Regular expression that matches syntax for a link (https://stackoverflow.com/a/3809435/117030):
- const LINK_REGEX = /https?:\/\/(www\.)?[-a-zA-Z0-9@:%._\+~#=]{1,256}\.[a-zA-Z0-9()]{1,6}\b([-a-zA-Z0-9()@:%_\+.~#?&//=]*)/gi;
- const filterW3Urls = (urls) =>
- urls.filter(url =>
- url && !url.startsWith('http://www.w3.org/'))
- const filterDataUrls = (urls) =>
- urls.filter(url =>
- url && !url.startsWith('data:'))
- const html = await page.content();
- const raw = html?.match(LINK_REGEX) || [];
- const hrefs = await page.$$eval(
- "pierce/a[href]",
- elems => elems
- .map(elem => elem.href)
- .filter(url => url),
- );
- const links = await page.$$eval(
- "pierce/link[href]",
- elems => elems
- .map(({rel, href}) => ({rel, href}))
- .filter(({rel, href}) => rel !== 'stylesheet')
- .reduce((collection, entry) => {
- const {rel, href} = entry
- const non_empty_rel = collection[href]?.rel || rel
- collection[href] = {rel: non_empty_rel, href}
- return collection
- }, {})
- );
- const iframes = await page.$$eval(
- "pierce/iframe[src]",
- elems => elems.map(iframe => iframe.src).filter(url => url)
- );
- const images = await page.$$eval(
- "pierce/img[src]",
- elems => elems.map(img => img.src).filter(url => url && !url.startsWith('data:'))
- );
- const css_images = await page.$$eval(
- "pierce/*",
- elems => elems
- .map(elem => {
- const css_url_ptn = /url\(\s*?['"]?\s*?(\S+?)\s*?["']?\s*?\)/i;
- const bg_img = window.getComputedStyle(elem, null).getPropertyValue('background-image')
- const bg_url = css_url_ptn.exec(bg_img)
- return bg_url ? bg_url[1] : null
- })
- )
- const css_stylesheets = await page.$$eval(
- "pierce/link[rel=stylesheet]",
- elems => elems.map(elem => elem.href).filter(url => url)
- );
- const js_scripts = await page.$$eval(
- "pierce/script[src]",
- elems => elems.map(elem => elem.src).filter(url => url)
- );
- const outlinks_info = {
- TYPE: 'outlinks',
- VERSION: version,
- URL: original_url,
- raw: [...new Set(filterDataUrls(filterW3Urls(raw)))],
- hrefs: [...new Set(filterDataUrls(hrefs))],
- links: [...Object.values(links)],
- iframes: [...new Set(iframes)],
- images: [...new Set(filterDataUrls(images))],
- css_images: [...new Set(filterDataUrls(css_images))],
- css_stylesheets: [...new Set(filterDataUrls(css_stylesheets))],
- js_scripts: [...new Set(filterDataUrls(js_scripts))],
- }
- if (raw?.length || hrefs?.length || links?.length || iframes?.length) {
- console.log(`[🖇️] Saving page outgoing links (${raw?.length || hrefs?.length})...`.padEnd(82+1), prettyPath(OUTLINKS_PATH(page)))
- await overwriteFile(OUTLINKS_PATH(page), outlinks_info)
- }
- return outlinks_info
- }
- async function saveAuthStorage(page, {client, version, original_url}) {
- const url = original_url || await page.url()
- if (URL_SCHEMES_IGNORED.includes(url.split(':')[0])) return null
- if (!SAVE_AUTH_STORAGE) return null
- // const cookies = JSON.stringify(await page.cookies()); // doesnt include httponly cookies
- const auth_from_browser = {
- cookies: (await client.send('Network.getAllCookies')).cookies,
- localStorage: {},
- sessionStorage: {},
- }
- // attempt to load localStorage and sessionStorage from browser (may fail in some cases https://github.com/puppeteer/puppeteer/issues/921)
- try {
- auth_from_browser.localStorage = (await page.evaluate(() =>
- JSON.parse(JSON.stringify({[window.location.origin]: window.localStorage}))))
- } catch(err) {
- throw `Failed to get page window.localStorage! ${err}`
- }
- try {
- auth_from_browser.sessionStorage = (await page.evaluate(() =>
- JSON.parse(JSON.stringify({[window.location.origin]: window.sessionStorage}))))
- } catch(err) {
- throw `Failed to get page window.sessionStorage! ${err}`
- }
- // WARNING: small TOCTTOU gap between this read-before-write and the write below
- // can possibly overwrite changes made by other processes in this gap
- const auth_on_disk = await loadAuthStorage(page, {client}, {apply: false})
- const cookies = dedupeCookies([...auth_on_disk.cookies, ...auth_from_browser.cookies])
- const auth_info = {
- TYPE: 'auth',
- VERSION: version,
- URL: original_url,
- cookies: cookies,
- sessionStorage: merge(auth_on_disk.sessionStorage, auth_from_browser.sessionStorage),
- localStorage: merge(auth_on_disk.localStorage, auth_from_browser.localStorage),
- }
- // console.log(`[⛙] Merged ${auth_on_disk.cookies.length} existing + ${auth_from_browser.cookies.length} new -> ${auth_info.cookies.length} cookies`)
-
- console.log(`[🍪] Saving cookies/localStorage/sessionStorage (${auth_info.cookies.length})...`.padEnd(82), prettyPath(AUTH_JSON_PATH));
- await overwriteFile(AUTH_JSON_PATH, auth_info);
-
- // Write to cookies.txt file using tough-cookie + @root/file-cookie-store
- await saveCookiesTxt(cookies)
- return auth_info
- }
- async function saveCookiesTxt(cookies) {
- const cookies_store = new FileCookieStore(COOKIES_TXT_PATH, {auto_sync: false, lockfile: false})
- const cookie_jar = new ToughCookie.CookieJar(cookies_store)
- cookie_jar.setCookieAsync = util.promisify(cookie_jar.setCookie)
- cookies_store.saveAsync = util.promisify(cookies_store.save)
- for (const cookie of cookies) {
- const cookie_for_tough = {
- domain: cookie.domain,
- path: cookie.path,
- key: cookie.name,
- value: cookie.value,
- expires: (new Date(cookie.expires * 1000)).toISOString(),
- hostOnly: cookie.domain.startsWith('.'),
- secure: cookie.secure,
- }
- // console.log('COOKIE_FOR_TOUGH_TXT', cookie_for_tough)
- const parsed_cookie = ToughCookie.Cookie.fromJSON(cookie_for_tough)
- // console.log('COOKIE_FOR_TOUGH_TXT_TO_DUMP', parsed_cookie)
- try {
- // assemble a fake URL just to satisfy ToughCookieJar's requirement of having a URL at set time
- let url = cookie.secure ? 'https://' : 'http://'
- if (cookie.domain.startsWith('.')) {
- url = url + cookie.domain.slice(1)
- } else {
- url = url + cookie.domain
- }
- if (cookie.sourcePort && ![80, 443].includes(cookie.sourcePort)) {
- url = `${url}:${cookie.sourcePort}`
- }
- url = `${url}${cookie.path || ''}`
- await cookie_jar.setCookieAsync(parsed_cookie, url, {ignoreError: true})
- } catch(err) {
- console.error('[❌] Failed to dump browser cookie for cookies.txt...', cookie_for_tough, '->', parsed_cookie, err)
- }
- }
- console.log(`[🍪] Saving cookies TXT (${cookies.length})...`.padEnd(82), prettyPath(COOKIES_TXT_PATH));
- await cookies_store.saveAsync()
- }
- async function saveMetrics(page, {original_url, version, start_time, start_ts, traffic_log, redirects}) {
- const end_time = (new Date()).toISOString()
- const end_ts = Date.now()
- const metrics_info = {
- TYPE: 'metrics',
- VERSION: version,
- URL: original_url,
- ...(await page.metrics()),
- start_time,
- start_ts,
- end_time,
- end_ts,
- duration: (end_ts - start_ts),
- num_requests: traffic_log.length,
- num_redirects: Object.keys(redirects).length -1,
- }
- console.log(`[🏎️] Saving final summary + timing metrics...`.padEnd(82+1), prettyPath(METRICS_PATH(page)))
- await overwriteFile(METRICS_PATH(page), metrics_info)
- return metrics_info
- }
- /******************************************************************************/
- /******************************************************************************/
- /**************************** Utility Helpers *********************************/
- function hashCode(str) {
- // get a simple integer hash for a given string (based on java String#hashCode)
- // useful only for throwaway nonces / easy deterministic random identifiers, not a replacement for sha256
- let hash = 0;
- for (let i=0; i<str.length; i++) {
- hash = str.charCodeAt(i) + ((hash << 5) - hash);
- }
- return Math.abs(hash)
- }
- function unique(iter, key: string | ((any, number) => string)='id') {
- // uniqueify an array of objects by a value within them, key can be name of attr or getter function
- // > iter = [{id: 1}, {id: 2}, {id: 1}]
- // > Object.entries(iter) = [
- // [ '0', { id: 1 } ],
- // [ '1', { id: 2 } ],
- // [ '2', { id: 1 } ] ]
- // > unique(iter, 'id') => {1: {id: 1}, 2: {id: 2}}
- // > iter = {a1: {id: 1}, b2: {id: 2}, a3: {id: 1}}
- // > Object.entries(iter) = [
- // [ 'a1', { id: 1 } ],
- // [ 'b2', { id: 2 } ],
- // [ 'a3', { id: 1 } ]
- // ]
- // > unique(iter, 'id') => {1: {id: 1}, 2: {id: 2}}
- const key_type = (typeof key)
- if (!['function', 'string'].includes(key_type))
- throw 'key must be either a string lookup key or a function (obj, idx) => return unique_id'
- const key_func = (key_type === 'string')
- ? (entry_obj, idx) => entry_obj[(key as string)]
- : (entry_obj, idx) => (key as Function)(entry_obj, idx) // otherwise key is a callback func
- const seen = {}
- for (const [idx, entry_obj] of Object.entries(iter)) {
- const unique_id = key_func(entry_obj, idx)
- if (seen[unique_id] === undefined) {
- seen[unique_id] = entry_obj
- }
- }
- return seen
- }
- const wait = (ms: number) => new Promise(res => {
- if (ms > 10_000) {
- console.debug(`[⏲️] Waiting ${Math.round(ms/1000)}s...`)
- }
- setTimeout(res, ms)
- })
- const TimeoutError = Symbol()
- const withTimeout = (promise, ms) => {
- // run a promise with a time limit, raises a TimeoutError if it fails
- let timer
- return Promise.race([
- promise,
- new Promise((_r, reject) =>
- timer = setTimeout(reject, ms, TimeoutError)
- ),
- ]).finally(() => clearTimeout(timer))
- }
- const MAX_VALID_DATE = new Date('2150-01-01T00:00:00.000Z')
- const MIN_VALID_DATE = new Date('2010-01-01T00:00:00.000Z')
- const UNIX_EPOCH_DATE = new Date(0)
- const validateDate = (date, {min=MIN_VALID_DATE, max=MAX_VALID_DATE, singleton=UNIX_EPOCH_DATE}={}) => {
- assert((date instanceof Date), `Got invalid type for Date: ${typeof date} ${date} (expected Date)`)
- assert(String(date) !== 'Invalid Date', `Got invalid value for Date: ${typeof date} ${date}`)
- if (Number(date) === Number(singleton)) return date // epoch singleton is always valid
- assert(date < max, `Got Date that was higher than MAX_VALID_DATE=${max}`)
- assert(date > min, `Got Date that was lower than MIN_VALID_DATE=${min}`)
- return date
- }
- const parseVersionDateStr = (yyyymmddtime) => {
- // YYYYMMDDhhmmssxxx or YYYYMMDDhhmmss or YYYYMMDDhhmm or YYYYMMDD -> Date
- const is_only_numbers = /^\d+$/.test(yyyymmddtime.replace('.', ''))
- assert(is_only_numbers, `Non-numeric characters in YYYYMMDD date are not allowed: ${yyyymmddtime} (while trying YYYYMMDDhhmmssxxx format)`)
- const num_digits = String(yyyymmddtime).split('.')[0].length
- assert([17, 14, 12, 8].includes(num_digits), `Got invalid number of digits (${num_digits}) in YYYYMMDD date: ${yyyymmddtime} (while trying YYYYMMDDhhmmssxxx format)`)
- const [_all, yyyy, mm, dd, hr, min, sec, ms] = /^(\d{4})(\d{2})(\d{2})(\d{2})?(\d{2})?(\d{2})?(\d{3})?$/.exec(yyyymmddtime)
- assert(yyyy && mm && dd, `Could not find YYYYMMDD`)
- const time_error_msg = `Detected YYYYMMDD[hhmm[ss[xxxx]]] but time segment is invalid ${hr}:${min || '__'}:${ms || '___'}`
- if (ms) assert(hr && min && sec, time_error_msg)
- if (sec) assert(hr && min, time_error_msg)
- if (min) assert(hr, time_error_msg)
- if (hr) assert (min, time_error_msg)
- const iso_str = `${yyyy}-${mm}-${dd}T${hr || '00'}:${min || '00'}:${sec || '00'}.${ms || '00'}Z`
- const parsed_date = new Date(iso_str)
- return validateDate(parsed_date) // 1970-01-01T00:00:00.000Z (ISO format)
- }
- const parseTimestampDateStr = (timestamp) => {
- // 1709724291000 or 1709724291000.000 or 1709724291 or 1709724291.000 -> Date
- timestamp = String(timestamp)
- const is_only_numbers = /^\d+$/.test(timestamp.replace('.', ''))
- assert(is_only_numbers, `Got invalid characters in timstamp: ${timestamp} (while trying xxxxxxxxxxxxx format)`)
- const num_digits = String(timestamp).split('.')[0].length
- assert([13, 10, 1].includes(num_digits), `Got invalid number of digits (${num_digits}) in timestamp: ${timestamp} (while trying xxxxxxxxxxxxx format)`)
- let parsed_date = null
- if (num_digits === 13) {
- parsed_date = new Date(Number(timestamp)) // 1709724291000 (unix timestamp w/ milliseconds)
- } else if (num_digits === 10) {
- parsed_date = new Date(Number(timestamp) * 1000) // 1709724291 (unix timestamp w/ seconds)
- } else if (num_digits === 1) {
- assert(String(timestamp) === '0', `Got invalid single-digit timestamp: ${timestamp} (while trying xxxxxxxxxxxxx format or 0 for UNIX epoch)`)
- parsed_date = UNIX_EPOCH_DATE
- }
- return validateDate(parsed_date)
- }
- const parseISODateStr = (iso_str) => {
- // 1970-01-01T00:00:00.000Z -> Date
- const num_digits = String(iso_str).length
- assert([24, 19, 16, 10].includes(num_digits), `Got invalid number of digits (${num_digits}) in ISO date: ${iso_str} (while trying 1970-01-01T00:00:00.000Z format)`)
- const parsed_date = new Date(iso_str)
- return validateDate(parsed_date)
- }
- const parseDate = (date) => {
- // date === undefined => use today/now
- // date === null => use unix epoch 0 aka 1970-01-01T00:00:00.000Z
- // date *= YYYYMMDDHHMMSS => use a version date string (e.g. 20010131235958)
- // date *= 1234567... => use a timestmap (e.g. 1709724291000)
- // date *= 1970-01-01T... => use iso datetime (e.g. 1970-01-01T00:00:00.000Z)
- // returns -> Date
- if (date === undefined) {
- return (new Date()) // today (2024-05-29T22:02:34.682Z) aka timestamp=1717020154682
- }
- if (date === null || date == 0) {
- return UNIX_EPOCH_DATE // unix epoch (1970-01-01T00:00:00.000Z) aka timestamp=0
- }
- if (date instanceof Date) {
- return validateDate(date) // JS date Date('1970-01-01T00:00:00.000Z')
- }
- if ((typeof date) === 'number') {
- date = String(date) // unix timestamp e.g. 1717020154682
- }
- assert((typeof date) === 'string', `Tried to parse date but got unsupported type ${(typeof date)}: ${date}`)
- const errors = [`Failed to parse Date from string: ${date}`]
- try {
- return parseVersionDateStr(date)
- } catch(err) { errors.push(err) }
- try {
- return parseTimestampDateStr(date)
- } catch(err) { errors.push(err) }
- try {
- return parseISODateStr(date)
- } catch(err) { errors.push(err) }
-
- throw errors.join('\n')
- }
- const versionStrFromDate = (date, {withDate=true, withTime=true, withSeconds=true, withMilliseconds=false}={}) => {
- // takes Date, returns YYYYMMDDHHMMSSXXX or YYYYMMDDHHMMSS or YYYYMMDDHHMM or YYYYMMDD
- const parsed_date = parseDate(date)
- const [date_iso, time_iso] = parsed_date.toISOString().split('T') // ['2001-01-31', '23:59:58.090Z']
- const components_to_use = []
- if (withDate) {
- components_to_use.push(date_iso.replaceAll('-', '')) // '20010131'
- }
- if (withTime) {
- const [hr, min, sec, ms] = time_iso.replace('Z', '').replace('.', ':').split(':') // ['23', '59', '58', '090']
- components_to_use.push(hr)
- components_to_use.push(min)
- if (withSeconds) {
- components_to_use.push(sec)
- if (withMilliseconds) {
- components_to_use.push(ms)
- }
- }
- }
- assert(components_to_use.length, 'At least one of {withDate, withTime} must be set.')
- const final_str = components_to_use.join('') // 20010131235958
- assert(parseVersionDateStr(final_str)) // sanity check to make sure it parses correctly
- return final_str
- }
- // test date functions:
- // console.log(parseDate('20120131'))
- // console.log(versionStrFromDate(parseDate('20120131')))
- // console.log(versionStrFromDate(parseDate('0')))
- // console.log(versionStrFromDate(parseDate(0)))
- // console.log(versionStrFromDate(parseDate(null)))
- // console.log(versionStrFromDate())
- // console.log(versionStrFromDate(parseDate('20120131235859090')))
- // console.log(versionStrFromDate(parseDate('1970-01-01T00:00:00.000Z')))
- // console.log(versionStrFromDate(parseDate('2024-12-01T00:00')))
- // console.log(versionStrFromDate(parseDate('2024-12-01'), {withTime: false}))
- const prettyPath = (path) => {
- // return a pretty-printable path where the abspath of the data dir is replaced with /data for brevity/privacy
- return path.replace(DATA_DIR, './data')
- }
- const pathIsHidden = (relpath) => {
- // check if a path or any of the directories above it are hidden (e.g. ./some/.dir/abc or ./.DS_Store)
-
- // make sure test path behaves like an abspath (avoids edge-cases messing up relpaths on '' or '.' or './')
- let test_path = relpath
- if (test_path.startsWith('./'))
- test_path = test_path.substring(2)
- if (!test_path.startsWith('/'))
- test_path = path.join('/', test_path)
- // iterate through parents, checking if any parent is hidden until we reach /
- while (test_path !== '/') {
- const basename = path.basename(test_path)
- if (basename.startsWith('.')) {
- // console.log('PATH IS HIDDEN', relpath)
- return true
- }
- // otherwise set test_path to parent dir and repeat
- test_path = path.dirname(test_path)
- }
- return false
- }
- const pathDepth = (child_path, relative_to='.') => {
- // get the number of directory hops deep a child path is relative to '.' (or a given parent)
-
- if (child_path.startsWith('/') && !relative_to.startsWith('/')) {
- // if child_path is absolute, then relative_to must be absolute as well otherwise depth will be depth all the way to the / root
- relative_to = fs.realpathSync(relative_to)
- }
- if (relative_to.startsWith('/') && !child_path.startsWith('/')) {
- // same deal, either both paths have to be relative, or both have to be absolute
- child_path = fs.realpathSync(child_path)
- }
- const relative_path_to_root = path.relative(relative_to, child_path)
- const num_hops_down = relative_path_to_root.split('/').length
- return num_hops_down
- }
- interface DirentWithExtras extends fs.Dirent {
- relpath: string,
- abspath: string,
- reldepth: number,
- }
- async function getDirEntries(dir_path, {pwd=null, recursive=true, includeHidden=false, includeFiles=true, includeDirs=true, includeLinks=false, filter=null, maxdepth=-1}={}) {
- // get the list of all sub-paths under a given path recursively
- // console.log('GETTING DIRECTORY ENTRIES', {dir_path, pwd, recursive, includeHidden, includeFiles, includeDirs, maxdepth})
- pwd = pwd || dir_path
- let dir_abspath = dir_path
- if (!dir_abspath.startsWith(pwd)) {
- dir_abspath = path.join(pwd, dir_abspath)
- }
- assert(fs.existsSync(dir_abspath), `Tried to get directory listing for dir that doesn't exist! ${prettyPath(dir_abspath)}`)
- return (await fs.promises.readdir(dir_abspath, { recursive, withFileTypes: true }))
- .map((dirent: DirentWithExtras) => {
- // filter combined with map because relpath is re-used in both operations
- const relpath = path.join(path.relative(pwd, dirent.parentPath), dirent.name)
- // console.log('CALCULATED RELATIVE PATH', relpath)
- const abspath = path.join(dir_abspath, relpath)
- const basename = path.basename(dirent.name)
- if (!includeLinks && dirent.isSymbolicLink()) return null
- if (!includeFiles && dirent.isFile()) return null
- if (!includeDirs && dirent.isDirectory()) return null
- if (!includeHidden && pathIsHidden(relpath)) return null
- dirent.relpath = relpath
- dirent.abspath = abspath
- dirent.reldepth = pathDepth(relpath)
- // console.log('RELATIVE DEPTH MEASURED', prettyPath(dir_abspath), prettyPath(relpath), dirent.reldepth)
- if (maxdepth >= 0) {
- if ((dirent.reldepth-1) > maxdepth) return null
- }
-
- if ((typeof filter) === 'function') {
- const should_keep = filter({abspath, relpath, basename, dirent})
- if (!should_keep) {
- // console.log('FILTER EXCLUDED RESULT', {abspath, relpath, basename, dirent})
- return null
- }
- }
- return relpath
- })
- .filter(Boolean)
- .sort() as string[]
- }
- async function getTotalSize(dir_or_file_path, {pwd=null, _cache=null, filter=null, subfiles=null}={}) {
- // get the total size in bytes of a file or directory (recursively adds up file sizes within directory)
- // check _cache first
- if (_cache && (dir_or_file_path in _cache))
- return _cache[dir_or_file_path]
- // make sure dir_or_file_path is under pwd
- pwd = pwd || path.dirname(dir_or_file_path)
- let abspath = dir_or_file_path
- if (!dir_or_file_path.startsWith(pwd)) {
- abspath = path.join(pwd, dir_or_file_path)
- }
- // if it's a file, stat it and return the size
- // console.log('CALCUALTED ABSPATH', {abspath, dir_or_file_path, pwd})
- const dirent = await fs.promises.stat(abspath)
- if (dirent.isFile()) {
- // console.log('CALCULATING FILE SIZE subfile=', prettyPath(abspath))
- return dirent.size
- }
- // if it's not a file and not a directory, give up, dont try to size special files like FIFO/socket/etc.
- if (!dirent.isDirectory()) return 0
- // if it's a directory, size is the sum of all the sizes of files within
- // console.log('CALCULATING SUBDIR SIZE subdir=', prettyPath(abspath))
- let total_bytes = 0
- const files_within = subfiles || await getDirEntries(dir_or_file_path, {
- pwd,
- recursive: true,
- includeDirs: false,
- includeFiles: true,
- filter,
- })
- for (const subpath of files_within) {
- total_bytes += await getTotalSize(subpath, {pwd, _cache, filter})
- }
- return total_bytes
- }
- async function getDirSizes(dir_path, {pwd=null, subfiles=null, withRoot=true, filter=null, maxdepth=-1}={}) {
- // get the size of a directory and all the files within (recursively) as a number of bytes
- // dir_path: path absolute or relative path of the directory you want size info for
- // pwd: path (optional) absolute path of the directory you want to interpret dir_path relative to
- // subfiles: dirent[] (optional) instead of reading disk, you can manually provide a getDirEntries results list to use
- // withRoot: bool include a summary entry for the root dir_path dir in the list as '.'
- // filter: function (optional) provide a filter func for dir entries ({abspath, relpath, basename, dirent}) => true/false
- // maxdepth: number (optional) does not affect actual calculations, but hides entries below a certain depth in the returned output for brevity
- assert((await fs.promises.stat(dir_path)).isDirectory(), `Tried to calculate directory sizes but path is not a directory! ${dir_path}`)
- pwd = pwd || dir_path
- // {'.': 246, 'example.json': 123, 'example2.txt': 123}
- const sizes = {}
- // first collect the list of all sub-files recursively and calculate their sizes individually
- const files_within = subfiles || await getDirEntries(dir_path, {
- pwd,
- recursive: true,
- includeDirs: false,
- includeFiles: true,
- // dont pass maxdepth here, we need the entire file listing to accurately calculate parent dir sizes
- // it never makes sense to ignore subfiles beyond a certain depth for size calculation
- filter, // filter is allowed though, useful to calculcate size of some subset of files that match a pattern
- })
- for (const subpath of files_within) {
- sizes[subpath] = await getTotalSize(subpath, {pwd, _cache: sizes, filter})
- }
-
- // then calculate the top-level directory total as the sum of all the file sizes under it
- const total_size = Object.values(sizes).reduce((a: number, b: number) => a + b, 0)
- // then calculate the subtotals of all the sub-directories
- const subdirs_within = await getDirEntries(dir_path, {pwd, recursive: true, includeDirs: true, includeFiles: false, filter, maxdepth})
- for (const subpath of subdirs_within) {
- sizes[subpath] = await getTotalSize(subpath, {pwd, _cache: sizes, filter}) // uses _cache to avoid re-computing
- }
- // if maxdepth is passed, filter results to only include paths shallower than max depth
- if (maxdepth >= 0) {
- for (const subpath of Object.keys(sizes)) {
- if (pathDepth(subpath) > maxdepth) {
- delete sizes[subpath]
- }
- }
- }
- // set total_size last so it appears at the bottom of the object in logs for convenience
- if (withRoot) {
- sizes['.'] = total_size
- }
- return sizes
- }
- async function getLargestPath(path_a, path_b) {
- // compare two files/directories and return the largest one of the two (calculating size recursively)
-
- path_a = await fs.promises.realpath(path_a)
- path_b = await fs.promises.realpath(path_b)
- const size_a = await getTotalSize(path_a)
- const size_b = await getTotalSize(path_b)
- // console.log('COMPARING', prettyPath(path_a), size_a, ' ', prettyPath(path_b), size_b)
- if (size_a > size_b) return path_a
- return path_b
- }
- async function findCommonAncestor(target_abspath, symlink_abspath, {relative=true, search_limit=DATA_DIR}: {relative?: boolean | string, search_limit?: string}={}) {
- // given a target path and a symlink path, find the common ancestor path they both share
- // (searches recursively through absolute path parent directories until a common dir is found, up to search_limit)
- search_limit = await fs.promises.realpath(search_limit)
- let relative_dir = search_limit
- if ((typeof relative) === 'boolean') {
- // if start dir is default, set it to symlinks directory path
- if (relative) {
- relative_dir = path.dirname(symlink_abspath)
- } else {
- relative_dir = search_limit
- }
- } else if ((typeof relative) === 'string') {
- // if start dir is a string, get its absolute path
- relative_dir = relative as string
- } else {
- throw `Got invalid type for relative path during common ancestor search: ${relative}`
- }
- if ((await fs.promises.stat(relative_dir)).isFile()) {
- // if start dir is a file, set it to its parent dir path
- relative_dir = path.dirname(relative_dir)
- }
- assert(
- (await fs.promises.stat(relative_dir)).isDirectory(),
- `Tried to find common ancestor starting from invalid search directory:\n 🔗 ${prettyPath(symlink_abspath)}\n -> ${prettyPath(target_abspath)}\n Error: search dir does not exist or is not a directory: ❌ ${prettyPath(relative_dir)}`,
- )
- const symlink_filename = path.basename(symlink_abspath)
- const target_filename = path.basename(target_abspath)
- const symlink_parent_abspath = await fs.promises.realpath(path.dirname(symlink_abspath))
- const target_parent_abspath = await fs.promises.realpath(path.dirname(target_abspath))
- const search_dir_abspath = await fs.promises.realpath(relative_dir)
- let closest_common_ancestor = search_dir_abspath
- const isAncestorCommon = (ancestor) => (
- target_parent_abspath.startsWith(ancestor)
- && symlink_parent_abspath.startsWith(ancestor))
- // check if both src and target start with the same ancestor path
- while (closest_common_ancestor !== search_limit) {
- if (isAncestorCommon(closest_common_ancestor)) break
- else {
- // otherwise go up one directory and try again
- // console.log(' ...going up a directory', prettyPath(closest_common_ancestor)+'/..')
- closest_common_ancestor = path.dirname(closest_common_ancestor)
- }
- }
- assert(
- isAncestorCommon(closest_common_ancestor),
- `Tried to create relative symlink but could not find common ancestor:\n 🔗 ${prettyPath(symlink_abspath)}\n -> ${prettyPath(target_abspath)}\n Error: target path and symlink path are not both under:\n ❌ ${prettyPath(closest_common_ancestor)}`,
- )
-
- const symlink_to_ancestor_relpath = path.relative(symlink_parent_abspath, closest_common_ancestor) // ../../..
- const target_from_ancestor_relpath = path.join(path.relative(closest_common_ancestor, target_parent_abspath), target_filename) // 'archive/19999999.23423523'
- const symlink_to_target_relpath = path.join(symlink_to_ancestor_relpath, target_from_ancestor_relpath) // '../../../archive/19999999.23423523'
- return {
- closest_common_ancestor,
- search_dir_abspath,
- target_abspath,
- target_filename,
- target_from_ancestor_relpath,
-
- symlink_abspath,
- symlink_filename,
- symlink_to_ancestor_relpath,
- symlink_to_target_relpath,
- }
- }
- interface StatsWithExtras extends fs.Stats {
- abspath: string
- relpath?: string
- reldepth?: number
- }
- async function blockUntilExists(file_path, {timeout=7_500, min_bytes=0}={}) {
- // wait up to timeout seconds until file we expect to exist appears on the filesystem
- // (used to handle eventual consistency in network filesystems where we need a delay after writing before reads show up)
- const interval = 250
- const max_tries = timeout / interval
- let tries = 0
-
- let abspath = null
- while (tries < max_tries) {
- try {
- const abspath = await fs.promises.realpath(file_path)
- assert(fs.existsSync(abspath))
-
- const dirent = await fs.promises.stat(abspath) as StatsWithExtras
- dirent.abspath = abspath
-
- if (min_bytes && (dirent.size < min_bytes)) {
- assert(dirent.size >= 1)
- // this is a valid warning but unfortunately its too common to bother showing:
- // console.warn(`[⚠️] Expected file to be >=${Math.round(min_bytes/1000)}kb but was only ${dirent.size/1000}kb:`, prettyPath(file_path))
- }
-
- return dirent
- } catch(err) {
- const waited = (tries * interval)
- if (waited === 5_000) {
- console.warn(`[⚠️] Waited >${waited/1000}s for file to appear (is filesystem or bg task running slow?):`, prettyPath(file_path))
- }
- await wait(interval)
- tries++
- }
- }
- throw `Expected file does not exist after ${timeout/1000}s: ${prettyPath(file_path)}`
- }
- async function overwriteSymlink(target_path, symlink_path, {relative=true, mkdirs=false, search_limit=DATA_DIR, timeout=5_000}: {relative?: boolean | string, mkdirs?: boolean, search_limit?: string, timeout?: number}={}) {
- // create a symlink from symlink_path -> target_path
- // relative: true => symlink is created as a relative link by default (it will auto-find the closest common ancestor dir, often DATA_DIR)
- // mkdirs: true => optionally creates symlink parent dirs automatically)
- // make sure target file actually exists first
- let target_dirent
- try {
- target_dirent = await blockUntilExists(target_path, {timeout})
- } catch(err) {
- throw `Tried to create symlink pointing to file that does not exist:\n 🔗 ${prettyPath(symlink_path)}\n -> ❌ ${prettyPath(target_path)}\n ${err}`
- }
- const target_abspath = target_dirent.abspath
- const target_filename = path.basename(target_abspath)
- const target_parent_abspath = path.dirname(target_abspath)
-
- // make sure target is a valid file or directory and not a special character/block device/other weird file
- const target_is_dir = target_dirent.isDirectory()
- const target_is_file = target_dirent.isFile()
- assert(target_is_dir || target_is_file, `Tried to create symlink to an unsupported file type:\n 🔗 ${prettyPath(symlink_path)}\n -> ❌ ${prettyPath(target_path)} (expected file or directory)`)
- // create symlink file parent directories if needed
- const symlink_filename = path.basename(symlink_path)
- const symlink_parent_dir = path.dirname(symlink_path)
- if (mkdirs) {
- await fs.promises.mkdir(symlink_parent_dir, {recursive: true})
- }
- try {
- assert((await fs.promises.stat(symlink_parent_dir)).isDirectory())
- } catch(err) {
- throw `Tried to create symlink in a directory that doesn't exist:\n 🔗 ${symlink_parent_dir}❌/${symlink_filename}\n -> ${target_path}\n ${err}`
- }
- const symlink_parent_abspath = await fs.promises.realpath(symlink_parent_dir)
- const symlink_abspath = path.join(symlink_parent_abspath, symlink_filename)
- // determine nearest common ancestor between symlink dir and target dir
- const {
- closest_common_ancestor,
- symlink_to_ancestor_relpath,
- target_from_ancestor_relpath,
- symlink_to_target_relpath,
- } = await findCommonAncestor(target_abspath, symlink_abspath, {relative, search_limit})
-
- // set final target path to abspath or relative path depending on {relative} options
- let target_path_final
- if (relative) {
- // make symlink into relative link (based on closest common ancestor dir between symlink_abspath and target_abspath)
- target_path_final = symlink_to_target_relpath
- // console.log(' 🔗', prettyPath(symlink_abspath), '->', prettyPath(target_abspath), `(as relative link: ${target_path_final})`)
- } else {
- // make symlink into an absolute path (verbatim passed target_path)
- target_path_final = target_path
- // console.log(' 🔗', prettyPath(symlink_abspath), '->', prettyPath(target_abspath), '(as absolute path)')
- }
- // remove any existing symlink at destination if there is already one there
- const random_nonce = crypto.randomBytes(16).toString('hex').substring(0, 8)
- const symlink_temp_path = `${symlink_abspath}.${random_nonce}.dup`
- try { await fs.promises.unlink(symlink_abspath) } catch(err) {}
- try { await fs.promises.unlink(symlink_temp_path) } catch(err) {}
- // create the symlink and check that it works after creation
- let created_symlink = null
- try {
- created_symlink = symlink_temp_path
- await fs.promises.symlink(target_path_final, symlink_temp_path)
- created_symlink = symlink_abspath
- await fs.promises.rename(symlink_temp_path, symlink_abspath)
- } catch(err) {
- if (String(err).includes('EISDIR')) {
- // console.warn('[⚠️] Tried to create symlink on top of existing directory', prettyPath(symlink_abspath))
- // no real recourse in this situation, and its too noisy to log every time this happens
- // it's also not always safe to move the dir out of the way, so better to just fail silently here, leaving:
- // ${symlink_abspath}.${random_nonce}.dup
- } else {
- console.warn('[⚠️] Failed to create symlink', prettyPath(created_symlink), err)
- }
- }
- let dirent
- try {
- dirent = await blockUntilExists(created_symlink, {timeout, min_bytes: 0})
- // best we can do here is just check that it exists ^, trying to check that it has the exact expected abspath that we set is bad, because its a race condition:
- // assert(dirent.abspath == target_abspath) // its often already overwritten by later activity, so final abspath may already be different
- } catch(err) {
- throw `Symlink created but does not seem to resolve to intended file:\n 🔗 ${symlink_path}\n -> ❌ ${target_path}\n actual=${dirent?.abspath}\n expected=${target_abspath}\n ${err}`
- }
- return {
- symlink_path,
- symlink_abspath: created_symlink,
- symlink_filename: path.basename(created_symlink),
- symlink_parent_abspath,
- symlink_to_ancestor_relpath,
- symlink_to_target_relpath,
-
- target_path,
- target_abspath,
- target_filename,
- target_parent_abspath,
- target_from_ancestor_relpath,
- target_path_final,
- target_is_dir,
- target_is_file,
- target_is_relative: Boolean(relative),
-
- closest_common_ancestor,
- }
- }
- // test symlink and common ancestor finding
- // console.log(await findCommonAncestor('/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/archive/1709724410.19269/seo.json', '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/archive/1709724410.19269/seo2.json'))
- // console.log(await findCommonAncestor('/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/archive/1709724410.19269', '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/index/snapshots_by_domain/twitter.com/1709724410.19269', {relative: true, search_limit: '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/'}))
- // console.log(await overwriteSymlink('/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/archive/1709724410.19269', '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/index/snapshots_by_domain/twitter.com/1709724410.19269'))
- // console.log(await overwriteSymlink('/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/index/snapshots_by_domain/twitter.com/1709724410.19269', '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/index/favorite_snapshots/1709724410.19269', {relative: false, mkdirs: true, search_limit: '/Volumes/NVME/Users/squash/Local/Code/archiveboxes/archivebox-spreadsheet-bot/data/'}))
- async function overwriteDir(path) {
- // delete any existing folder at the destination path (important otherwise we may create a folder inside an existing folder/symlink)
- try {
- await fs.promises.rm(path, { recursive: true, force: true });
- } catch(err) {}
- await fs.promises.mkdir(path, {recursive: true})
- return path
- }
- async function overwriteFile(path, contents, options={encoding: 'utf8', flag: 'w', flush: false, block: true}) {
- // write any JS value to a fresh file (e.g. String, Buffer, WritableStream, etc. anything JSON-serializable)
- const block_until_created = options.block || true
- delete options.block
- try {
- // delete any existing symlink/file present at the destination path
- // (important otherwise we may write into an existing symlink by accident)
- await fs.promises.unlink(path)
- } catch(err) {}
- try {
- let nonce = 1
- while ((await fs.promises.stat(path)).isDirectory()) {
- // if we try to write a file to a path that already has a directory in that location
- // (common when trying to write response JSON e.g. http://www.instagram.com/api/graphql returns json and www.instagram.com/api/graphql/abc returns json)
- path = path.replace(`.${nonce-1}`, '') + `.${nonce}`
- nonce++;
- if (nonce > 20) throw `Too many conflicting files while trying to write to ${prettyPath(path)}`
- }
- } catch(err) {
- if (!String(err).includes('no such file or directory')) {
- console.warn('[⚠️] Warning: Problem with conflicting directory at while trying to write file', err)
- }
- }
- // refuse writing undefined/null/function because its likely an error and not intended
- const content_is_null = (contents === null) || (contents === undefined)
- const content_is_func = (typeof contents === 'function')
- if (content_is_null || content_is_func) {
- throw `Cannot write ${typeof contents} ${contents} to file: ${path}`
- }
- // Numbers, BigInts, and Booleans can be cast to strings, then wrt
- const content_is_primitive = ['number', 'bigint', 'boolean'].includes(typeof contents)
- if (content_is_primitive) {
- contents = String(contents)
- await fs.promises.writeFile(path, contents, options as any)
- if (block_until_created) await blockUntilExists(path, {min_bytes: Buffer.byteLength(contents)})
- return path
- }
- // Strings and Buffers can be written directly to file
- const content_is_string = (typeof contents === 'string' || contents instanceof String)
- const content_is_buffer = Buffer.isBuffer(contents)
- if (content_is_string || content_is_buffer) {
- await fs.promises.writeFile(path, contents, options as any)
- if (block_until_created) await blockUntilExists(path, {min_bytes: Buffer.byteLength(contents)})
- return path
- }
-
- // WritableStream objects can be piped into file
- const content_is_stream = (contents?.pipe)
- if (content_is_stream) {
- const stream_byte_length = contents.writableLength
- const dest_file = fs.createWriteStream(path);
- await finished(contents.pipe(dest_file))
- if (block_until_created) await blockUntilExists(path, {min_bytes: stream_byte_length})
- return path
- }
- // Objects and Arrays can be JSON-stringified then written into file
- const content_is_obj = (Array.isArray(contents) || typeof contents === 'object')
- if (content_is_obj) {
- contents = JSON.stringify(contents, null, 4)
- await fs.promises.writeFile(path, contents, options as any)
- if (block_until_created) await blockUntilExists(path, {min_bytes: Buffer.byteLength(contents)})
- return path
- }
- throw `Cannot write contents of type ${typeof contents} to file: ${path} < ${contents}`
- }
- async function saveExecResult(bin, args=null, {original_url, version}, {cwd='.', timeout=60_000, ...spawn_options}={}) {
- assert(bin)
- assert(original_url && original_url.includes('://'))
- assert(version)
- const BIN_NAME = bin // 'yt-dlp'
- const ARGS = args || [] // ['--some-arg', '--some-other-arg']
- const CWD = cwd || process.cwd() // '.'
- const TIMEOUT = 300_000 // 5min timeout
- const PATH = process.env.PATH
- await fs.promises.mkdir(cwd, {recursive: true})
- // quick-n-dirty dump of cmd to bash script, but this might be better: https://github.com/nodejs/node/issues/34840#issuecomment-677402567
- const cmd_log_str = `#!/usr/bin/env bash
- TYPE="${BIN_NAME}"
- URL="${original_url}"
- VERSION="${version}"
- TIMEOUT=${TIMEOUT}
- CWD="${CWD}"
- PATH="${PATH}:$PATH"
- ${BIN_NAME} ${ARGS.map(arg => JSON.stringify(arg)).join(' ')}
- `
- const cmd_log = path.join(cwd, 'cmd.sh')
- await overwriteFile(cmd_log, cmd_log_str)
- const stdout_log = fs.createWriteStream(path.join(cwd, 'stdout.log'))
- const stderr_log = fs.createWriteStream(path.join(cwd, 'stderr.log'))
-
- const start_date = new Date()
- const start_ts = Number(start_date)
- const start_time = start_date.toISOString()
- const child = child_process.spawn(
- BIN_NAME,
- ARGS,
- {
- cwd: CWD,
- timeout: TIMEOUT, // 5min timeout
- stdio: [null, 'pipe', 'pipe'], // </dev/null >./stdout.log 2>./stderr.log
- // detached: true, // run in background, don't block on response
- ...(spawn_options || {}),
- },
- )
- child.stdout.setEncoding('utf8')
- child.stdout.pipe(stdout_log)
- child.stderr.setEncoding('utf8')
- child.stderr.pipe(stderr_log)
- const exec_info = {
- TYPE: BIN_NAME,
- URL: original_url,
- VERSION: version,
- bin_name: BIN_NAME,
- args: ARGS,
- timeout: TIMEOUT,
- hostname: os.hostname(),
- bin_paths: PATH,
- ppid: process.pid,
- pid: child.pid,
- start_ts,
- start_time,
- end_time: null,
- end_ts: null,
- duration: null,
- returncode: null,
- log_files: {},
- output_files: {},
- }
- // promise that resolves when the command is finished executing
- // TODO: refactor to use withTimeout
- const getResult = (timeout=TIMEOUT) =>
- new Promise((resolve, reject) => {
- const loop = setInterval(() => {
- if (exec_info.end_time) {
- clearInterval(loop)
- clearTimeout(timer)
- resolve(exec_info)
- }
- }, 100)
- const timer = setTimeout(() => {
- clearInterval(loop)
- if (!exec_info.end_time) {
- reject(new Error(`Process ${BIN_NAME} did not finish within TIMEOUT=${TIMEOUT}`))
- }
- }, timeout);
- })
- const logFilesFilter = ({relpath}) =>
- ['cmd.sh', 'stdout.log', 'stderr.log'].includes(relpath)
- const outputFilesFilter = ({relpath}) =>
- !['cmd.sh', 'stdout.log', 'stderr.log', 'index.json'].includes(relpath)
- const getOutputFiles = async (filter=outputFilesFilter) => {
- return await getDirInfo(CWD, {filter, withHelpers: false, withRoot: false, maxdepth: 6})
- }
- child.on('close', async (returncode) => {
- const end_date = new Date()
- exec_info.returncode = returncode
- exec_info.pid = child.pid
- exec_info.end_ts = Number(end_date)
- exec_info.end_time = end_date.toISOString()
- exec_info.duration = exec_info.end_ts - exec_info.start_ts
- exec_info.log_files = await getOutputFiles(logFilesFilter)
- exec_info.output_files = await getOutputFiles(outputFilesFilter)
- const end_metadata = `
- # END_TIME="${exec_info.end_time}"
- # DURATION=${exec_info.duration}
- # RETURNCODE=${exec_info.returncode }
- `
- await fs.promises.appendFile(cmd_log, end_metadata)
- // write exec_info json (which includes file list) to CWD/index.json
- await overwriteFile(path.join(CWD, 'index.json'), exec_info)
- })
- // child.unref() // dont wait for child process to close
-
- const start_metadata = `
- #################### LAST RUN LOG ####################
- # HOSTNAME="${exec_info.hostname}"
- # PPID=${exec_info.ppid}
- # PID=${exec_info.pid}
- # START_TIME="${exec_info.start_time}"
- `
- await fs.promises.appendFile(cmd_log, start_metadata)
- return {
- ...exec_info,
- getResult,
- }
- }
- const HASH_CACHE = {}
- async function sha256File(file_path: string, {pwd=null}: {pwd?: string}={}) {
- return new Promise((resolve, reject) => {
- pwd = pwd || path.dirname(file_path);
- if (!file_path.startsWith(pwd)) {
- file_path = path.join(pwd, file_path);
- }
- const dirent = fs.statSync(file_path);
- const abspath = fs.realpathSync(file_path);
- const cache_key = `${abspath}:${dirent.size}:${dirent.mtimeMs}`; // PATH:SIZE:LAST_MODIFIED_TIME
- if (cache_key in HASH_CACHE) {
- resolve(HASH_CACHE[cache_key]);
- }
- const hash = crypto.createHash('sha256');
- const rs = fs.createReadStream(abspath);
- rs.on('error', reject);
- rs.on('data', chunk => hash.update(chunk));
- rs.on('end', () => {
- const final_hash = hash.digest('hex');
- HASH_CACHE[cache_key] = final_hash;
- resolve(final_hash);
- });
- }) as Promise<string>
- }
- async function getDirSha256(dir_path, {pwd=null, withRoot=true, filter=null, maxdepth=-1, subfiles=null}={}) {
- // console.log('CALCULATING SHA256 OF FILES IN DIR', dir_path, {withRoot, filter, maxdepth})
- // dir_path: path absolute or relative path of the directory you want the merkle sha256 for
- // pwd: path (optional) absolute path of the directory you want to interpret dir_path relative to
- // withRoot: bool include a summary entry for the root dir_path dir in the list as '.'
- // filter: function (optional) provide a filter func for dir entries ({abspath, relpath, basename, dirent}) => true/false
- // maxdepth: number (optional) does not affect actual calculations, but hides entries below a certain depth in the returned output for brevity
- // subfiles: dirent[] (optional) instead of reading disk, you can manually provide a getDirEntries results list to use
- pwd = pwd || dir_path
- if (!dir_path.startsWith(pwd)) {
- dir_path = path.join(pwd, dir_path)
- }
- const dirent = await fs.promises.stat(dir_path)
- assert(dirent.isDirectory(), `Tried to compute sha256 of path but missing or not a directory! ${dir_path}`)
- assert((maxdepth >= -1), `maxdepth must be -1, 0, or 1, 2, 3, etc... (got ${maxdepth})`)
-
- // assert(!(filter && withRoot), `Cannot generate root hash (consistently) when a custom filter is provided!`)
- // get the sha256 of every file in a directory recursively (excluding hidden files and symlinks)
- // EQUIVALENT TO: find . -type f -not -path '*/.*' -print0 | sort -z | xargs -0 sha256sum
- const all_subfiles = (subfiles as string[]) || await getDirEntries(dir_path, {
- pwd,
- recursive: true,
- includeFiles: true,
- includeDirs: false,
- // ~~maxdepth,~~ // dont pass maxdepth here, we need the entire file listing to accurately calculate parent dir hashes.
- // it never makes sense to ignore subfiles beyond a certain depth for hash calculation. Hashes are
- // only useful IDs if they are consistent+repeatable, hashing to an arbitrary depth will produce
- // many different hashes for the same directory, which is not something we need/want polluting the hash space.
-
- filter, // we do however allow passing a manual filter funcs which does actually affect the hash
- // this is useful to allow quick checks to see whether a certain subset of files has changed or not
- })
- const hashes: {[key: string]: string} = {}
- let hashable_summary_str = ''
- for (const subfile of all_subfiles) {
- // {'versions/20240413144307/screen recording.mp4': '1df4d9c3aca8b36f1f73e327d56038f80a35db407a298edb16c72576d7dd894e', ...}
- hashes[subfile] = await sha256File(subfile, {pwd})
- const relpath = path.relative(await fs.promises.realpath(dir_path), await fs.promises.realpath(path.join(pwd, subfile)))
- hashable_summary_str += `${hashes[subfile]} ./${relpath}\n`
- }
- // console.log('CALCULATED HASHES FOR ALL SUBFILES IN DIR', dir_path, Object.keys(hashes).length)
- // get list of subdirectories and recursively hash every subdirectory
- // EQUIVALENT TO: find . -type d -not -path '*/.*' -maxdepth ${maxdepth} -print | sort
- const subdirs = await getDirEntries(dir_path, {pwd, recursive: true, includeHidden: false, includeDirs: true, includeFiles: false, filter, maxdepth})
- // for each subdirectory, get its hash recursively and store it in the hash list
- for (const subdir of subdirs) {
- // console.log('GETTING SUBDIR HASH', subdir)
- // a directory's hash is defined as the hash of all the *files* within (excluding dirs/symlinks/hidden)
- const subdir_hashes = await getDirSha256(
- subdir,
- {pwd, withRoot: true, filter, maxdepth: 0},
- )
- hashes[subdir] = subdir_hashes['.']
- }
- // console.log('CALCULATED HASHES FOR ALL SUBDIRS IN DIR', dir_path, subdirs.length)
- // filter results if maxdepth is provided
- if (maxdepth >= 0) {
- for (const subpath of Object.keys(hashes)) {
- if (pathDepth(subpath) > maxdepth) {
- delete hashes[subpath]
- }
- }
- }
- // console.log('LIMITED OUTPUT DUE TO MAXDEPTH', maxdepth, Object.keys(hashes).length)
- // calculate the hash of the root '.' folder by hashing all of hashes of its contents
- // EQUIVALENT TO: find . -type f -not -path '*/.*' -print0 | sort -z | xargs -0 sha256sum | sha256sum
- if (withRoot) {
- // pass the first command's output containing the file list + hashes into another sha256
- // to get the final hash of the whole directory combined
- // console.log('CALCULATING FINAL ROOT HASH for ', dir_path)
- // console.log(hashable_summary_str)
- hashes['.'] = crypto.createHash('sha256').update(hashable_summary_str).digest('hex') as string
- // console.log('--->', hashes['.'])
- }
- return hashes
- }
- async function getDirInfo(dir_path, {pwd=null, withRoot=true, withHelpers=true, filter=null, maxdepth=-1, subfiles=null}={}) {
- // get a detailed JSON/dumpable index of a directory's contents, w/ merkle sha256's, sizes, and mimeTypes
- // dir_path: path absolute or relative path of the directory you want size info for
- // pwd: path (optional) absolute path of the directory you want to interpret dir_path relative to
- // withRoot: bool include a summary entry for the root dir_path dir in the list as '.'
- // withHelpers: bool attach many extra helper attrs/funcs to results (beyond JSON-serializable core data)
- // filter: function (optional) provide a filter func for dir entries ({abspath, relpath, basename, dirent}) => true/false
- // maxdepth: number (optional) does not affect actual calculations, but hides entries below a certain depth in the returned output for brevity
- // subfiles: dirent[] (optional) instead of reading disk, you can manually provide a getDirEntries results list to use
- // {
- // ...
- // 'example.txt': { ... },
- // 'foobar/example.mp3': { ... },
- // '.': { // this is the fully agumented result when withHelpers=true
- // is_file: false,
- // is_dir: true,
- // filename: '.',
- // basename: '1709039915.378868',
- // mimeType: 'inode/directory'
- // extension: undefined,
- // num_bytes: 11540961,
- // num_subpaths: 15,
- // sha256: '9fc58b3ed887e7139338062ebd49bd6795373759e8acb73d2f7a40f1413789da',
- // reldepth: 1,
- // relpath: './',
- // cwd: '/opt/archivebox/data/archive/1709039915.378868/',
- // dirname: '/opt/archivebox/data/archive',
- // abspath: '/opt/archivebox/data/archive/1709039915.378868',
- // dirent: Stats {
- // dev: 16777240,
- // mode: 16895,
- // uid: 501,
- // ...
- // mtimeMs: 1717160622956.1357,
- // ctimeMs: 1717160622956.1357,
- // },
- // created: '2024-05-31T13:03:42.956Z',
- // modified: '2024-05-31T13:03:42.956Z',
- // summary: './data/archive/1709039915.378868 (inode/directory 11541kb 9fc58b3e)',
- // helptext: 'Verify these hashes by running:\n' +
- // ' cd /opt/archivebox/data/archive/1709039915.378868 \n' +
- // " find . -type f -not -path '*/.*' -print0 | sort -z | xargs -0 sha256sum | sha256sum",
- // },
- // }
- pwd = pwd || dir_path
- if (!dir_path.startsWith(pwd)) {
- dir_path = path.join(pwd, dir_path)
- }
- // calculate hashes and sizes recursively
- const hashes = await getDirSha256(dir_path, {pwd, withRoot, filter, maxdepth, subfiles})
- const sizes = await getDirSizes(dir_path, {pwd, withRoot, filter, maxdepth, subfiles})
- const num_total_subpaths = Object.keys(hashes).filter(name => name !== '.').length
- const details = {}
- for (const [filename, sha256] of Object.entries(hashes)) {
- if (filename === '.' && !withRoot) continue
- const abspath = await fs.promises.realpath(path.join(dir_path, filename))
- const dirent = await fs.promises.stat(abspath)
- const num_subpaths = Object.keys(hashes).filter(subpath => subpath.startsWith(filename + '/')).length
- const is_file = dirent.isFile()
- const is_dir = dirent.isDirectory()
- // bare-bones info suitable for JSON dumps/exports
- const basic_info = {
- sha256,
- num_bytes: sizes[filename],
- created: (new Date(dirent.ctimeMs)).toISOString(),
- mimeType: undefined,
- extension: undefined,
- num_subpaths: undefined,
- }
- if (is_dir) {
- basic_info.mimeType = 'inode/directory'
- basic_info.extension = undefined
- basic_info.num_subpaths = (filename === '.') ? num_total_subpaths : num_subpaths
- }
- if (is_file) {
- basic_info.mimeType = mime.lookup(abspath) || null
- basic_info.extension = path.extname(filename)
- basic_info.num_subpaths = undefined
- }
- // extra helpers suitable for usage in other areas of the codebase
- const info_with_helpers = {
- ...basic_info,
- filename,
- basename: path.basename(abspath),
- dirname: path.dirname(abspath),
- cwd: dir_path,
- relpath: is_dir ? (filename + '/') : filename,
- reldepth: pathDepth(filename),
- abspath,
- is_file,
- is_dir,
- dirent,
- modified: (new Date(dirent.mtimeMs)).toISOString(),
- summary: `${prettyPath(abspath)} (${basic_info.mimeType} ${Math.round(basic_info.num_bytes/1000)}kb ${sha256.substring(0, 8)})`,
- helptext: undefined,
- }
- if (filename === '.') {
- info_with_helpers.helptext = `Verify these hashes by running:\n cd ${prettyPath(abspath)} \n find . -type f -not -path '*/.*' -print0 | sort -z | xargs -0 sha256sum | sha256sum`
- }
- if ((typeof filter) === 'function') {
- if (!filter(info_with_helpers)) continue
- }
- details[filename] = withHelpers ? info_with_helpers : basic_info
- }
- return details
- }
- // console.log(await getDirSha256(
- // '/opt/archivebox/data/archive/1709039915.378868/',
- // {
- // withRoot: true,
- // maxdepth: -1,
- // filter: ({relpath}) => relpath.startsWith('versions'),
- // },
- // ))
- // console.log(await getDirSizes(
- // '/opt/archivebox/data/archive/1709039915.378868/',
- // {
- // withRoot: false,
- // maxdepth: 2,
- // filter: ({relpath}) => !relpath.startsWith('versions'),
- // },
- // ))
- // console.log(await getDirInfo(
- // '/opt/archivebox/data/archive/1709039915.378868/',
- // {
- // withRoot: true,
- // withHelpers: true,
- // maxdepth: 1,
- // // filter: ({relpath}) => relpath.startsWith('versions'),
- // },
- // ))
- type DetectFilenameOptions = {
- url?: string,
- response?: HTTPResponse | Response,
- page?: Page,
- dir?: string,
- abspath?: string,
- filename?: string,
- basename?: string,
- extension?: string,
- mimeType?: string,
- resourceType?: string,
- }
- async function detectFilename({ url, response, page, dir, abspath, filename, basename, extension, mimeType, resourceType }: DetectFilenameOptions) {
- // this function takes a url (and/or response/page), and detects the abspath,dir,filename,basename,extention,mimeType
- // from the URL (+ any enforced path components passed in via args)
- // example: detectFilename({url: 'https://example.com/favicon.png', extension: 'ico'}) outputs 'favicon.ico'
- //
- // it has some quirks that are specific to archiving and may not behave as you expect
- // e.g. if visiting the url https://example.com/error.zip returns a 500 text/html error page
- // this may still save it as a .zip with mimeType=application/x-zip and ignore the response mimeType the url ends in .zip
- // however, if the url has no extension, e.g. https://example.com/error it will
- // auto-detect the mimeType based on the response and append an extension, saving as error.html
- //
- // ⚠️ SECURITY WARNING: think carefully about the permissions, shell injection, and RCE implications of any changes made here ⚠️
- // this function writes untrusted web content to the filesystem using auto-detected mimetype to co-erce the extension,
- // which can be dangerous (e.g. what if one of these downloads is a malicious ransomware .exe, do we really want to give it .exe?
- // if we do, how do we make sure it never gets executed? (without damaging the integrity of the copy)
- if (!(response || page)) throw 'Either a page or a response must be provided in order to detect mimeType & URL'
- if (response && (typeof response.headers !== 'function')) {
- const node_fetch_response: Response = response as Response
- response = {
- url: () => node_fetch_response.url,
- headers: () => node_fetch_response.headers,
- } as unknown as HTTPResponse
- }
- response = response as HTTPResponse
- url = url || response?.url() || (await page.url())
- if (!url) throw 'URL was not provided and could not be detected from {response, page}'
- // Document, Stylesheet, Image, Media, Font, Script, TextTrack, XHR, Fetch, Prefetch, EventSource, WebSocket, Manifest, SignedExchange, Ping, CSPViolationReport, Preflight, Other
- try {
- resourceType = resourceType || response?.request()?.resourceType()
- } catch(err) {
- // ignore, sometimes response is null/not available
- }
- const resourceTypeToMimeType = {
- 'Stylesheet': 'text/css',
- 'Script': 'application/x-javascript',
- 'WebSocket': 'application/json',
- 'Website': 'text/html',
- }
-
- mimeType = mimeType || resourceTypeToMimeType[resourceType] // guess extension based on request resourceType
- extension = extension || (mimeType ? mime.extension(mimeType) : null)
- // handle special url cases (e.g. schemes in URL_SCHEMES_IGNORED)
- if (url.startsWith('about:blank')) {
- filename = 'about_blank'
- mimeType = 'text/html'
- }
- else if (url.startsWith('data:')) {
- filename = `data__${hashCode(url)}`
- }
- // console.log('detectFilename>', {url, dir, abspath, filename, basename, extension, mimeType, resourceType})
- if (abspath) {
- if (dir || filename || basename || extension)
- throw '{abspath} should not be passed with other options (e.g. dir, filename, basename, extension)'
- var {dir, base: filename, ext: extension, name: basename} = path.parse(abspath)
- // path.parse('/home/user/dir/file.txt') returns:
- // { root: '/',
- // dir: '/home/user/dir',
- // base: 'file.txt',
- // ext: '.txt',
- // name: 'file' }
- } else {
- dir = dir || path.resolve(process.cwd())
- filename = filename // https://example.com/a.1.zip?e.pdf=2#g.h=3 => a.1.zip
- || (new URL(url)).pathname.split('/').at(-1) // https://example.com/file124.rss => file124.rss prefers last component of path with no query/hash, falls back to domain name if no path
- || 'index' // https://example.com/abc/def/ => index.html
- //|| (new URL(url)).hostname.replaceAll('.', '_') // https://example.com => example_com (but if disabled, this would be index.html)
- }
- if (!filename) throw 'filename/abspath were not passed and could not be detected from url'
-
- const path_extname = path.extname(filename)
- const resp_mimetype = response && (
- (response as any).mimeType
- || response.headers()['content-type']?.split(';')[0]
- || resourceTypeToMimeType[resourceType]
- || 'application/octet-stream'
- )
- mimeType = mimeType // https://example.com/a.1.zip?e.pdf=2#g.h=3 => application/x-zip prefers mimetype based on extension in path, falls back to response mimeType
- || (path_extname && mime.lookup(path_extname)) // https://example.com/file124.rss => application/rss+xml
- || resp_mimetype // https://example.com/get?type=png => image/png
-
- extension = extension
- || (path_extname && path_extname.replace('.', '')) // https://example.com/a.1.zip?e.pdf=2#g.h=3 => zip prefers extension in path, falls back to response mimeType's suggested extension
- || (resp_mimetype && mime.extension(resp_mimetype)) // https://example.com => html
- || '' // https://example.com/websocket.1 =>
- if (extension.startsWith('.'))
- extension = extension.slice(1)
- basename = basename // https://example.com/a.1.zip?e.pdf=2#g.h=3 => a.1 prefers to filename in path (without extension), falls back to domain name
- || (path.parse(filename).name) // https://mp4dl.example.com => mp4dl_example_com
- basename = basename.slice(0, 120) // truncate at 120 characters (leaving 8 chars for .ext)
- basename = basename.replace(/[^a-zA-Z0-9%+?&=@;_ \.-]/g, '') // strip characters not allowed in filenames
- filename = basename + '.' + extension
- if (filename.endsWith('.'))
- filename = filename.slice(0, -1)
- abspath = abspath || path.join(dir, filename)
- // console.log('detectFilename<', {url, dir, abspath, filename, basename, extension, mimeType, resourceType})
- return {
- url,
- dir,
- abspath,
- filename,
- basename,
- extension,
- mimeType,
- resourceType,
- resp_mimetype,
- }
- }
- interface DowloadOptions extends DetectFilenameOptions {
- browser?: Browser
- expected_mimetype?: string
- timeout?: number
- }
- async function download({ url, browser, page, response, dir, abspath, filename, basename, extension, expected_mimetype, timeout }: DowloadOptions) {
- url = url || (response as HTTPResponse)?.url() || (await page?.url())
- ALREADY_ARCHIVED.add(url.slice(0, 4096)) // prevent running whole archive task on tabs we create for just for downloading
- browser = browser || (page && (await page.browser()))
- timeout = timeout || 120_000
- expected_mimetype = expected_mimetype || ''
- let newPage = null
- let errors = []
- let num_bytes = 0
- let bytesBuffer = null
- // if we need to fetch the url (i.e. it's not already been requested)
- if (!response) {
- if (!browser) throw 'No {browser} or {page} was provided to download with'
- newPage = await browser.newPage()
- if (page) await page.bringToFront() // if origin page is provided, make sure it stays in foreground
- response = await newPage.goto(url, {timeout: timeout, waitUntil: 'networkidle0'})
- if (page) await page.bringToFront() // if origin page is provided, make sure it stays in foreground
- }
- url = url || (response as HTTPResponse)?.url() || (await newPage?.url()) || (await page?.url());
- const response_mimetype = (response as HTTPResponse).headers()['content-type']?.split(';')[0] || 'text/html'
- // detect the filename we should write to based on provided url/response/page/filename/extension suggestions
- var {
- dir,
- abspath,
- filename,
- basename,
- extension,
- mimeType,
- } = await detectFilename({url, page, response, dir, abspath, filename, basename, extension, mimeType})
- // if mimeType is passed, make sure response matches expected mimetype, otherwise consider download a failure
- if (!response_mimetype.startsWith(expected_mimetype)) {
- errors.push(`Expected ${expected_mimetype} but got ${response_mimetype}`)
- } else {
- // download the file using puppeteer's response.buffer()
- try {
- // write the response bytes into the output file
- bytesBuffer = await (response as HTTPResponse).buffer()
- await overwriteFile(abspath, bytesBuffer)
- num_bytes = bytesBuffer.length
- } catch(err) {
- errors.push(err)
- }
- // security check to make sure downloaded file is not executable (random binaries downloaded off the internet = dangerous)
- fs.access(abspath, fs.constants.X_OK, (err) => {
- if (!err) console.warn(
- '[⚠️] SECURITY WARNING: Downloaded file appears to be executable:', prettyPath(abspath),
- '\n (be careful running untrusted programs downloaded from the internet!)'
- )
- })
- }
- // if we opened a dedicated page for downloading, close it now
- if (newPage) {
- newPage.close()
- }
- if (errors.length) {
- // console.warn(`[❌] Downloading ${url} (${mimeType}) to ${abspath} failed:`, JSON.stringify(errors, null, 4))
- } else {
- console.log(`[💾] Downloaded ${url.substring(0, 40)} (${num_bytes} ${mimeType})...`.padEnd(82), prettyPath(abspath))
- }
- return {
- url, response, errors,
- dir, abspath, filename, basename, extension, mimeType,
- bytesBuffer, num_bytes,
- }
- }
- /************************** Puppeteer Launching *******************************/
- async function startCluster(puppeteer, args=CHROME_ARGS_DEFAULT) {
- console.log(`[🎭] Launching ${CHROME_CLUSTER_WORKERS}x Chromium browsers with puppeteer-cluster:`.padEnd(82), prettyPath(CHROME_PROFILE_PATH))
- const cluster = await Cluster.launch({
- puppeteer,
- monitor: true,
- maxConcurrency: CHROME_CLUSTER_WORKERS,
- sameDomainDelay: 2550,
- workerCreationDelay: 250,
- timeout: 300_000, // total ms timeout for an entire task (1000ms * 60s * 5m)
- concurrency: Cluster.CONCURRENCY_PAGE, // share cookies between all tabs in a given browser
- puppeteerOptions: {
- args, // all the chrome launch CLI args
- ignoreDefaultArgs: true, // trust me, we have enough args already...
- // dumpio: true, // full debug log output, super noisy
- }
- })
- console.log('*************************************************************************')
- return cluster
- }
- async function remoteBrowser(puppeteer, {browserURL, browserWSEndpoint}) {
- console.log('[🎭] Connecting Puppeteer to existing Chromium browser via:', browserURL || browserWSEndpoint)
- let completed_initial_connection = false
- const browser = await puppeteer.connect({browserURL, browserWSEndpoint, defaultViewport: null, targetFilter: () => completed_initial_connection})
- completed_initial_connection = true
- console.log('*************************************************************************')
- return browser
- }
- async function startBrowser(puppeteer, args=CHROME_ARGS_DEFAULT) {
- console.log('[🎭] Launching Puppeteer Chromium browser...'.padEnd(82+1), prettyPath(CHROME_PROFILE_PATH))
- const browser = await puppeteer.launch({ignoreDefaultArgs: true, args, dumpio: true})
- globalThis.browser = browser
- console.log('*************************************************************************')
-
- // store all active tabs on global var by url for easier vscode interactive debugging
- const storeTabForDebugger = async (target) => {
- try {
- globalThis.tabs = globalThis.tabs || {}
- const url = target.url()
- const page = await target.page()
- if (!page || page?.isClosed()) {
- delete globalThis.tabs[url]
- } else {
- globalThis.tab = page
- globalThis.tabs[url] = page
- }
- } catch(err) {console.warn(err)}
- }
- browser.on('targetcreated', storeTabForDebugger)
- browser.on('targetchanged', storeTabForDebugger)
- browser.on('targetdestroyed', storeTabForDebugger)
- // wait for initial extension background.js/service worker targets to load
- await wait(3_000)
- // prime the extensions cache
- const extensions = await getChromeExtensionsFromCache({browser})
- globalThis.extensions = extensions // for easier debugging only
- // give the user 2min to check any issues with the initial startup pages (bot profile pages),
- // solve captchas, re-login, etc. then close them after that to save resources
- const startup_pages = (await browser.pages())
- const startup_page_close_delay = 120_000
- setTimeout(async () => {
- for (const page of startup_pages) {
- try { await page.close() } catch(err) { /* page may already be closed by now, which is fine */ }
- }
-
- }, startup_page_close_delay)
-
- // setup any extensions that need final runtime configuration using their options pages
- // await setup2CaptchaExtension({browser, extensions})
- // open a placeholder page so browser window stays open when there are no active archiving pages
- // (it's annoying to have the entire window open/close/open/close/etc every time an archive task runs)
- const empty_page = await browser.newPage()
- await wait(250)
- await empty_page.goto('chrome://version')
- await wait(500)
- console.log('*************************************************************************')
- return browser
- }
- async function startAPIServer(port=API_SERVER_PORT, host=API_SERVER_HOST, taskCallback=null) {
- // taskCallback should be an async function that takes ({url}) => and does something with it
- assert(taskCallback && (typeof taskCallback === 'function'))
- const server = createServer(async (req, res) => {
- if (req.method === 'POST') {
- console.log(`[API][POST] ${req.url}`)
- let body = '';
- req.on('data', (chunk) => {
- body += chunk;
- });
- req.on('end', () => {
- try {
- const jsonData = JSON.parse(body);
- // Process the JSON data
- console.log(jsonData);
-
- res.writeHead(200, { 'Content-Type': 'application/json' });
- res.end(JSON.stringify({ message: 'JSON data received' }));
- } catch (error) {
- res.writeHead(400, { 'Content-Type': 'application/json' });
- res.end(JSON.stringify({ error: 'Invalid JSON data' }));
- }
- });
- } else if (req.method === 'GET') {
- console.log(`[API][GET] ${req.url}`)
- const parsedUrl = new URL(`http://${host}:${port}${req.url}`)
- const query = new URLSearchParams(parsedUrl.search);
- const url = query.get('url');
- if (url && url.includes('://')) {
- res.writeHead(200, { 'Content-Type': 'text/plain' });
- try {
- await taskCallback({url})
- res.end(`${url}\n${TASK_PATH(url)}`);
- } catch(err) {
- res.end(`${url}\n${TASK_PATH(url)}\n${err}`);
- }
- } else {
- res.writeHead(500, { 'Content-Type': 'text/plain' });
- res.end(`Bad URL: ${url}\n\nExpected: /?url=https://example.com/url/to/archive`);
- }
- } else {
- res.writeHead(405, { 'Content-Type': 'application/json' });
- res.end(JSON.stringify({ error: 'Method not allowed' }));
- }
- })
- server.listen(port, host, () => {
- console.log(`[🎰] API Server listening for requests on http://${host}:${port}/?url=...`);
- })
- console.log('*************************************************************************')
- return server
- }
- async function main(urls, cluster=CHROME_CLUSTER) {
- process.chdir(DATA_DIR)
- const extensions = await getChromeExtensionsFromPersona({CHROME_EXTENSIONS, CHROME_EXTENSIONS_DIR})
- const args = getChromeArgs({...CHROME_LAUNCH_OPTIONS, CHROME_EXTENSIONS: extensions})
- const preferences = getChromePreferences({CHROME_PREFERENCES_DEFAULT, CHROME_PREFERENCES_EXTRA, CHROME_DOWNLOADS_DIR, CHROME_EXTENSIONS: extensions})
- const Puppeteer = applyChromePreferences(PupeteerExtra, CHROME_PREFERENCES_PATH, preferences)
-
- Puppeteer.use(StealthPlugin());
- // Puppeteer.use(ReplPlugin());
- // handled by uBlock Origin & ReCaptcha browser extensions, probably not needed here anymore:
- // Puppeteer.use(RecaptchaPlugin({
- // provider: {id: '2captcha', token: API_KEY_2CAPTCHA},
- // visualFeedback: true,
- // }))
- // const AdblockerPlugin = require('puppeteer-extra-plugin-adblocker')
- // puppeteer.use(AdblockerPlugin({ blockTrackers: true }))
- if (cluster) {
- // launch browser with multiple tabs w/ puppeteer
- const cluster = await startCluster(Puppeteer, args)
- const handleTask = async ({url}) => cluster.queue(url, botArchiveTask)
- const server = await startAPIServer(API_SERVER_PORT, API_SERVER_HOST, handleTask)
- console.log('[📋] Running tasks in parallel with puppeteer cluster...')
- for (const url of urls) {
- if (fs.existsSync(path.join(TASK_PATH(url), 'aiqa.json'))) {
- try {
- JSON.parse((await fs.promises.readFile(path.join(TASK_PATH(url), 'aiqa.json'))).toString())
- console.log(' skipping (already present):', TASK_PATH(url), url)
- continue
- } catch(err) {
- // pass
- }
- }
- cluster.queue(url, botArchiveTask)
- await wait(3_000)
- }
- await cluster.idle();
- await cluster.close();
- } else {
- // launch single new browser w/ puppeter / connect to remote CDP browser w/ puppeteer
- const browser = await startBrowser(Puppeteer, args)
- // const browser = await remoteBrowser(Puppeteer, {browserURL, browserWSEndpoint})
- // run speedtest in the background
- speedtest({browser})
- const handleTask = async ({url}) => await botArchiveTask({page: (await browser.newPage()), data: url})
- const server = await startAPIServer(API_SERVER_PORT, API_SERVER_HOST, handleTask)
- // wait for any pre-run setup tasks or server requests
- await wait(5_000)
- let num_succeeded = 0
- let num_failed = 0
- console.log(`[📋] Running ${urls.length} tasks sequentially with puppeteer browser...`)
- for (const url of urls) {
- const run_count = (num_succeeded + num_failed) || 1
- // check if task should be run or skipped based on existing snapshot data present in directory
- const metrics_path = path.join(TASK_PATH(url), 'metrics.json')
- const screenshot_path = path.join(TASK_PATH(url), 'screenrecording.gif')
- const aiqa_path = path.join(TASK_PATH(url), 'aiqa.json')
- const versions_path = path.join(TASK_PATH(url), 'versions')
- if (fs.existsSync(metrics_path) && fs.existsSync(screenshot_path) && fs.existsSync(aiqa_path) && fs.existsSync(versions_path)) {
- try {
- const ai_qa_result = JSON.parse(await fs.promises.readFile(aiqa_path, 'utf-8'))
- console.log(prettyPath(TASK_PATH(url)), `${ai_qa_result.pct_visible}%`, ai_qa_result.website_brand_name, url.substring(0, 80))
- assert(ai_qa_result.website_brand_name)
- continue
- } catch(err) {
- // pass
- }
- }
- let delay = 0
- // create a new browser page and run the archiving task
- const page = (await browser.newPage())
- try {
- console.log(ANSI.black + `◤==============================================================================[${String(run_count).padStart(3)}]/[${urls.length}]◥` + ANSI.reset)
- await botArchiveTask({page, data: url})
- delay = 1_000
- num_succeeded += 1
- } catch(err) {
- console.error('[❌] Archiving task failed!', url)
- console.error(err)
- num_failed += 1
- delay = 15_000 // extra delay if there are errors
- }
- console.log(ANSI.black + `◣==============================================================================[☑ ${num_succeeded}][🆇 ${num_failed}]◢` + ANSI.reset)
-
- // check for abnormally high failure rates and exit early if needed
- const failure_pct = Math.round((num_failed/run_count) * 100)
- if (failure_pct > 50) {
- if (run_count > 5) {
- console.warn(`[⚠️] ${failure_pct}% Task failure rate is very high! Will self-cancel after 10 URLs if >50% continue to fail...`)
- }
- if (run_count > 10) {
- throw `Too many tasks failed in a row! Quitting early after ${run_count}/${urls.length} tasks.`
- }
- }
-
- // increase the delay between tasks based on the ratio of how many are failing:succeeding
- delay = Math.pow(4, (num_failed/(num_succeeded + 3))) * delay
- // e.g. 0:1 failure ratio == 1 * delay == 1 ~ 15s
- // 1:1 failure ratio == 5 * delay == 5 ~ 1m ... 5^(failed:succeeded) exponential increase
- // 2:1 failure ratio == 25 * delay == 25s ~ 6m
- // 3:1 failure ratio == 125 * delay == 2m ~ 31m
- // etc...
- // up to 1hr+
- delay = Math.min(delay, 3_600_000) // 1hr maximum delay between tasks
- delay = Math.max(delay, 1_000) // 1s minimum delay between tasks
- if (delay > 2_500) {
- console.log('... waiting', Math.round(delay/1000), 'seconds (self rate-limit)...')
- }
- await wait(delay) // base ratelimit
- console.log()
- }
- if (PASSIVE_ARCHIVING) {
- // replace these as-needed:
- const browserURL = 'http://localhost:9222/'
- const browserWSEndpoint = 'ws://localhost:9222/devtools/browser'
- const driver_browser = browser || await remoteBrowser(Puppeteer, {browserURL, browserWSEndpoint})
- const archiver_browser = {} //await startBrowser(Puppeteer, args)
- const extensions = await getChromeExtensionsFromCache({browser: driver_browser})
- // close both browsers if either one is closed
- let browser_is_open = true
- driver_browser.on('disconnected', async () => {browser_is_open = false}) // await archiver_browser.close()
- // archiver_browser.on('disconnected', async () => {browser_is_open = false; await driver_browser.close()})
- // handle any tab navigation to a new URL in the driver browser
- const handleUserNavigation = async (target) => {
- const url = target.url()
- const page = await target.page()
- // const client = await target.createCDPSession()
- if (target.type() == 'page' && page && url) {
- console.log(ANSI.black + '==============================================================================' + ANSI.reset)
- console.warn('[➕] DRIVER BROWSER NAVIGATED:', ANSI.blue, url, ANSI.reset)
-
- try {
- await passiveArchiveTask({browser: driver_browser, page, url})
- await wait(3_000)
- } catch(err) {
- console.error('[❌] Archiving task failed!', url)
- console.error(err)
- await wait(10_000) // base ratelimit
- }
- console.log(ANSI.black + '==============================================================================' + ANSI.reset)
- // await client.send('Page.enable')
- // await client.send('Page.setWebLifecycleState', {state: 'active'})
- }
- // await client.send('Runtime.runIfWaitingForDebugger')
- }
- // setup handler to archive new page whenever one is opened
- driver_browser.on('targetcreated', handleUserNavigation)
- driver_browser.on('targetchanged', handleUserNavigation)
- console.log('------------------------------------------------------')
- console.log('[👀] Waiting for browser tabs to be opened by human...')
- while (browser_is_open) {
- await wait(2_000)
- }
- } else {
- while (true) {
- await wait(2_000)
- }
- }
- await browser.close()
- }
- console.log('[✅] Finished all tasks and stopped browsers.')
- process.exit(0);
- }
- /******************************************************************************/
- if (import.meta.main) {
- main(URLS).catch(console.error);
- }
- /******************************************************************************/
- // if we want to handle CLI args in the future, minimist is great:
- // var argv = require('minimist')(process.argv.slice(2));
- // console.log(argv); // --url=https://example.com --binpath=/browsers/chromium-1047/bin/chromium --datadir=/Chromium
- // const {url, binpath, datadir} = argv;
- // OLD CODE, may be useful in the future if we need audio in screenrecordings:
- // async function setupScreenrecordingWithAudio(page, wss) {
- // console.log('[🎬] Setting up screen-recording plugin...');
- // const stream_port = (await wss).options.port;
- // // streamPage = await (page.browser()).newPage()
- // await page.goto(`chrome-extension://jjndjgheafjngoipoacpjgeicjeomjli/options.html#${stream_port}`)
- //
- // // puppeteer-stream recording start
- // streamFile = fs.createWriteStream(SCREENRECORDING_PATH(page))
- // stream = await getStream(page, {
- // audio: true,
- // video: true,
- // bitsPerSecond: 8000000, // 1080p video
- // });
- // stream.pipe(streamFile);
- // return {stream, streamFile}
- //
- // // puppeteer-stream recording stop & cleanup
- // if (stream && streamFile) {
- // await stream?.destroy();
- // streamFile?.close();
- // // await streamPage.close();
- // }
- // }
|