1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349335033513352335333543355335633573358335933603361336233633364336533663367336833693370337133723373337433753376337733783379338033813382338333843385338633873388338933903391339233933394339533963397339833993400340134023403340434053406340734083409341034113412341334143415341634173418341934203421342234233424342534263427342834293430343134323433343434353436343734383439344034413442344334443445344634473448344934503451345234533454345534563457345834593460346134623463346434653466346734683469347034713472347334743475347634773478347934803481348234833484348534863487348834893490349134923493349434953496349734983499350035013502350335043505350635073508350935103511351235133514351535163517351835193520352135223523352435253526352735283529353035313532353335343535353635373538353935403541354235433544354535463547354835493550355135523553355435553556355735583559356035613562356335643565356635673568356935703571357235733574357535763577357835793580358135823583358435853586358735883589359035913592359335943595359635973598359936003601360236033604360536063607360836093610361136123613361436153616361736183619362036213622362336243625362636273628362936303631363236333634363536363637363836393640364136423643364436453646364736483649365036513652365336543655365636573658365936603661366236633664366536663667366836693670367136723673367436753676367736783679368036813682368336843685368636873688368936903691369236933694369536963697369836993700370137023703370437053706370737083709371037113712371337143715371637173718371937203721372237233724372537263727372837293730373137323733373437353736373737383739374037413742374337443745374637473748374937503751375237533754375537563757375837593760376137623763376437653766376737683769377037713772377337743775377637773778377937803781378237833784378537863787378837893790379137923793379437953796379737983799380038013802380338043805380638073808380938103811381238133814381538163817381838193820382138223823382438253826382738283829383038313832383338343835383638373838383938403841384238433844384538463847384838493850385138523853385438553856385738583859386038613862386338643865386638673868386938703871387238733874387538763877387838793880388138823883388438853886388738883889389038913892389338943895389638973898389939003901390239033904390539063907390839093910391139123913391439153916391739183919392039213922392339243925392639273928392939303931393239333934393539363937393839393940394139423943394439453946394739483949395039513952395339543955395639573958395939603961396239633964396539663967396839693970397139723973397439753976397739783979398039813982398339843985398639873988398939903991399239933994399539963997399839994000400140024003400440054006400740084009401040114012401340144015401640174018401940204021402240234024402540264027402840294030403140324033403440354036403740384039404040414042404340444045404640474048404940504051405240534054405540564057405840594060406140624063406440654066406740684069407040714072407340744075407640774078407940804081408240834084408540864087408840894090409140924093409440954096409740984099410041014102410341044105410641074108410941104111411241134114411541164117411841194120412141224123412441254126412741284129413041314132413341344135413641374138413941404141414241434144414541464147414841494150415141524153415441554156415741584159416041614162416341644165416641674168416941704171417241734174417541764177417841794180418141824183418441854186418741884189419041914192419341944195419641974198419942004201420242034204420542064207420842094210421142124213421442154216421742184219422042214222422342244225422642274228422942304231423242334234423542364237423842394240424142424243424442454246424742484249425042514252425342544255425642574258425942604261426242634264426542664267426842694270427142724273427442754276427742784279428042814282428342844285428642874288428942904291429242934294429542964297429842994300430143024303430443054306430743084309431043114312431343144315431643174318431943204321432243234324432543264327432843294330433143324333433443354336433743384339434043414342434343444345434643474348434943504351435243534354435543564357435843594360436143624363436443654366436743684369437043714372437343744375437643774378437943804381438243834384438543864387438843894390439143924393439443954396439743984399440044014402440344044405440644074408440944104411441244134414441544164417441844194420442144224423442444254426442744284429443044314432443344344435443644374438443944404441444244434444444544464447444844494450445144524453445444554456445744584459446044614462446344644465446644674468446944704471447244734474447544764477447844794480448144824483448444854486448744884489449044914492449344944495449644974498449945004501450245034504450545064507450845094510451145124513451445154516451745184519452045214522452345244525452645274528452945304531453245334534453545364537453845394540454145424543454445454546454745484549455045514552455345544555455645574558455945604561456245634564456545664567456845694570457145724573457445754576457745784579458045814582458345844585458645874588458945904591459245934594459545964597459845994600460146024603460446054606460746084609461046114612461346144615461646174618461946204621462246234624462546264627462846294630463146324633463446354636463746384639464046414642464346444645464646474648464946504651465246534654465546564657465846594660466146624663466446654666466746684669467046714672467346744675467646774678467946804681468246834684468546864687468846894690469146924693469446954696469746984699470047014702470347044705470647074708470947104711471247134714471547164717471847194720472147224723472447254726472747284729473047314732473347344735473647374738473947404741474247434744474547464747474847494750475147524753475447554756475747584759476047614762476347644765476647674768476947704771477247734774477547764777477847794780478147824783478447854786478747884789479047914792479347944795479647974798479948004801480248034804480548064807480848094810481148124813481448154816481748184819482048214822482348244825482648274828482948304831483248334834483548364837483848394840484148424843484448454846484748484849485048514852485348544855485648574858485948604861486248634864486548664867486848694870487148724873487448754876487748784879488048814882488348844885488648874888488948904891489248934894489548964897489848994900 |
- /*
- * $Id$
- *
- * Copyright (C) 2001-2003 FhG Fokus
- *
- * This file is part of ser, a free SIP server.
- *
- * ser is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation; either version 2 of the License, or
- * (at your option) any later version
- *
- * For a license to use the ser software under conditions
- * other than those described here, or to purchase support for this
- * software, please contact iptel.org by e-mail at the following addresses:
- * [email protected]
- *
- * ser is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
- /*
- * History:
- * --------
- * 2002-11-29 created by andrei
- * 2002-12-11 added tcp_send (andrei)
- * 2003-01-20 locking fixes, hashtables (andrei)
- * 2003-02-20 s/lock_t/gen_lock_t/ to avoid a conflict on solaris (andrei)
- * 2003-02-25 Nagle is disabled if -DDISABLE_NAGLE (andrei)
- * 2003-03-29 SO_REUSEADDR before calling bind to allow
- * server restart, Nagle set on the (hopefuly)
- * correct socket (jiri)
- * 2003-03-31 always try to find the corresponding tcp listen socket for
- * a temp. socket and store in in *->bind_address: added
- * find_tcp_si, modified tcpconn_connect (andrei)
- * 2003-04-14 set sockopts to TOS low delay (andrei)
- * 2003-06-30 moved tcp new connect checking & handling to
- * handle_new_connect (andrei)
- * 2003-07-09 tls_close called before closing the tcp connection (andrei)
- * 2003-10-24 converted to the new socket_info lists (andrei)
- * 2003-10-27 tcp port aliases support added (andrei)
- * 2003-11-04 always lock before manipulating refcnt; sendchild
- * does not inc refcnt by itself anymore (andrei)
- * 2003-11-07 different unix sockets are used for fd passing
- * to/from readers/writers (andrei)
- * 2003-11-17 handle_new_connect & tcp_connect will close the
- * new socket if tcpconn_new return 0 (e.g. out of mem) (andrei)
- * 2003-11-28 tcp_blocking_write & tcp_blocking_connect added (andrei)
- * 2004-11-08 dropped find_tcp_si and replaced with find_si (andrei)
- * 2005-06-07 new tcp optimized code, supports epoll (LT), sigio + real time
- * signals, poll & select (andrei)
- * 2005-06-26 *bsd kqueue support (andrei)
- * 2005-07-04 solaris /dev/poll support (andrei)
- * 2005-07-08 tcp_max_connections, tcp_connection_lifetime, don't accept
- * more connections if tcp_max_connections is exceeded (andrei)
- * 2005-10-21 cleanup all the open connections on exit
- * decrement the no. of open connections on timeout too (andrei) * 2006-01-30 queue send_fd request and execute them at the end of the
- * poll loop (#ifdef) (andrei)
- * process all children requests, before attempting to send
- * them new stuff (fixes some deadlocks) (andrei)
- * 2006-02-03 timers are run only once per s (andrei)
- * tcp children fds can be non-blocking; send fds are queued on
- * EAGAIN; lots of bug fixes (andrei)
- * 2006-02-06 better tcp_max_connections checks, tcp_connections_no moved to
- * shm (andrei)
- * 2006-04-12 tcp_send() changed to use struct dest_info (andrei)
- * 2006-11-02 switched to atomic ops for refcnt, locking improvements
- * (andrei)
- * 2006-11-04 switched to raw ticks (to fix conversion errors which could
- * result in inf. lifetime) (andrei)
- * 2007-07-25 tcpconn_connect can now bind the socket on a specified
- * source addr/port (andrei)
- * 2007-07-26 tcp_send() and tcpconn_get() can now use a specified source
- * addr./port (andrei)
- * 2007-08-23 getsockname() for INADDR_ANY(SI_IS_ANY) sockets (andrei)
- * 2007-08-27 split init_sock_opt into a lightweight init_sock_opt_accept()
- * used when accepting connections and init_sock_opt used for
- * connect/ new sockets (andrei)
- * 2007-11-22 always add the connection & clear the coresponding flags before
- * io_watch_add-ing its fd - it's safer this way (andrei)
- * 2007-11-26 improved tcp timers: switched to local_timer (andrei)
- * 2007-11-27 added send fd cache and reader fd reuse (andrei)
- * 2007-11-28 added support for TCP_DEFER_ACCEPT, KEEPALIVE, KEEPINTVL,
- * KEEPCNT, QUICKACK, SYNCNT, LINGER2 (andrei)
- * 2007-12-04 support for queueing write requests (andrei)
- * 2007-12-12 destroy connection asap on wbuf. timeout (andrei)
- * 2007-12-13 changed the refcnt and destroy scheme, now refcnt is 1 if
- * linked into the hash tables (was 0) (andrei)
- * 2007-12-21 support for pending connects (connections are added to the
- * hash immediately and writes on them are buffered) (andrei)
- * 2008-02-05 handle POLLRDHUP (if supported), POLLERR and
- * POLLHUP (andrei)
- * on write error check if there's still data in the socket
- * read buffer and process it first (andrei)
- * 2009-02-26 direct blacklist support (andrei)
- * 2009-03-20 s/wq_timeout/send_timeout ; send_timeout is now in ticks
- * (andrei)
- * 2009-04-09 tcp ev and tcp stats macros added (andrei)
- * 2009-09-15 support for force connection reuse and close after send
- * send flags (andrei)
- * 2010-03-23 tcp_send() split in 3 smaller functions (andrei)
- */
- /** tcp main/dispatcher and tcp send functions.
- * @file tcp_main.c
- * @ingroup core
- * Module: @ref core
- */
- #ifdef USE_TCP
- #ifndef SHM_MEM
- #error "shared memory support needed (add -DSHM_MEM to Makefile.defs)"
- #endif
- #define HANDLE_IO_INLINE
- #include "io_wait.h" /* include first to make sure the needed features are
- turned on (e.g. _GNU_SOURCE for POLLRDHUP) */
- #include <sys/time.h>
- #include <sys/types.h>
- #include <sys/select.h>
- #include <sys/socket.h>
- #ifdef HAVE_FILIO_H
- #include <sys/filio.h> /* needed on solaris 2.x for FIONREAD */
- #elif defined __OS_solaris
- #define BSD_COMP /* needed on older solaris for FIONREAD */
- #endif /* HAVE_FILIO_H / __OS_solaris */
- #include <sys/ioctl.h> /* ioctl() used on write error */
- #include <netinet/in.h>
- #include <netinet/in_systm.h>
- #include <netinet/ip.h>
- #include <netinet/tcp.h>
- #include <sys/uio.h> /* writev*/
- #include <netdb.h>
- #include <stdlib.h> /*exit() */
- #include <unistd.h>
- #include <errno.h>
- #include <string.h>
- #ifdef HAVE_SELECT
- #include <sys/select.h>
- #endif
- #include <poll.h>
- #include "ip_addr.h"
- #include "pass_fd.h"
- #include "tcp_conn.h"
- #include "globals.h"
- #include "pt.h"
- #include "locking.h"
- #include "mem/mem.h"
- #include "mem/shm_mem.h"
- #include "timer.h"
- #include "sr_module.h"
- #include "tcp_server.h"
- #include "tcp_init.h"
- #include "tcp_int_send.h"
- #include "tcp_stats.h"
- #include "tcp_ev.h"
- #include "tsend.h"
- #include "timer_ticks.h"
- #include "local_timer.h"
- #ifdef CORE_TLS
- #include "tls/tls_server.h"
- #define tls_loaded() 1
- #else
- #include "tls_hooks_init.h"
- #include "tls_hooks.h"
- #endif /* CORE_TLS*/
- #ifdef USE_DST_BLACKLIST
- #include "dst_blacklist.h"
- #endif /* USE_DST_BLACKLIST */
- #include "tcp_info.h"
- #include "tcp_options.h"
- #include "ut.h"
- #include "cfg/cfg_struct.h"
- #define local_malloc pkg_malloc
- #define local_free pkg_free
- #include <fcntl.h> /* must be included after io_wait.h if SIGIO_RT is used */
- #ifdef NO_MSG_DONTWAIT
- #ifndef MSG_DONTWAIT
- /* should work inside tcp_main */
- #define MSG_DONTWAIT 0
- #endif
- #endif /*NO_MSG_DONTWAIT */
- #define TCP_PASS_NEW_CONNECTION_ON_DATA /* don't pass a new connection
- immediately to a child, wait for
- some data on it first */
- #define TCP_LISTEN_BACKLOG 1024
- #define SEND_FD_QUEUE /* queue send fd requests on EAGAIN, instead of sending
- them immediately */
- #define TCP_CHILD_NON_BLOCKING
- #ifdef SEND_FD_QUEUE
- #ifndef TCP_CHILD_NON_BLOCKING
- #define TCP_CHILD_NON_BLOCKING
- #endif
- #define MAX_SEND_FD_QUEUE_SIZE tcp_main_max_fd_no
- #define SEND_FD_QUEUE_SIZE 128 /* initial size */
- #define SEND_FD_QUEUE_TIMEOUT MS_TO_TICKS(2000) /* 2 s */
- #endif
- /* minimum interval local_timer_run() is allowed to run, in ticks */
- #define TCPCONN_TIMEOUT_MIN_RUN 1 /* once per tick */
- #define TCPCONN_WAIT_TIMEOUT 1 /* 1 tick */
- #ifdef TCP_ASYNC
- static unsigned int* tcp_total_wq=0;
- #endif
- enum fd_types { F_NONE, F_SOCKINFO /* a tcp_listen fd */,
- F_TCPCONN, F_TCPCHILD, F_PROC };
- #ifdef TCP_FD_CACHE
- #define TCP_FD_CACHE_SIZE 8
- struct fd_cache_entry{
- struct tcp_connection* con;
- int id;
- int fd;
- };
- static struct fd_cache_entry fd_cache[TCP_FD_CACHE_SIZE];
- #endif /* TCP_FD_CACHE */
- static int is_tcp_main=0;
- enum poll_types tcp_poll_method=0; /* by default choose the best method */
- int tcp_main_max_fd_no=0;
- int tcp_max_connections=DEFAULT_TCP_MAX_CONNECTIONS;
- int tls_max_connections=DEFAULT_TLS_MAX_CONNECTIONS;
- static union sockaddr_union tcp_source_ipv4_addr; /* saved bind/srv v4 addr. */
- static union sockaddr_union* tcp_source_ipv4=0;
- static union sockaddr_union tcp_source_ipv6_addr; /* saved bind/src v6 addr. */
- static union sockaddr_union* tcp_source_ipv6=0;
- static int* tcp_connections_no=0; /* current tcp (+tls) open connections */
- static int* tls_connections_no=0; /* current tls open connections */
- /* connection hash table (after ip&port) , includes also aliases */
- struct tcp_conn_alias** tcpconn_aliases_hash=0;
- /* connection hash table (after connection id) */
- struct tcp_connection** tcpconn_id_hash=0;
- gen_lock_t* tcpconn_lock=0;
- struct tcp_child* tcp_children=0;
- static int* connection_id=0; /* unique for each connection, used for
- quickly finding the corresponding connection
- for a reply */
- int unix_tcp_sock;
- static int tcp_proto_no=-1; /* tcp protocol number as returned by
- getprotobyname */
- static io_wait_h io_h;
- static struct local_timer tcp_main_ltimer;
- static ticks_t tcp_main_prev_ticks;
- /* tell if there are tcp workers that should handle only specific socket
- * - used to optimize the search of least loaded worker for a tcp socket
- * - 0 - no workers per tcp sockets have been set
- * - 1 + generic_workers - when there are workers per tcp sockets
- */
- static int tcp_sockets_gworkers = 0;
- static ticks_t tcpconn_main_timeout(ticks_t , struct timer_ln* , void* );
- inline static int _tcpconn_add_alias_unsafe(struct tcp_connection* c, int port,
- struct ip_addr* l_ip, int l_port,
- int flags);
- /* sets source address used when opening new sockets and no source is specified
- * (by default the address is choosen by the kernel)
- * Should be used only on init.
- * returns -1 on error */
- int tcp_set_src_addr(struct ip_addr* ip)
- {
- switch (ip->af){
- case AF_INET:
- ip_addr2su(&tcp_source_ipv4_addr, ip, 0);
- tcp_source_ipv4=&tcp_source_ipv4_addr;
- break;
- case AF_INET6:
- ip_addr2su(&tcp_source_ipv6_addr, ip, 0);
- tcp_source_ipv6=&tcp_source_ipv6_addr;
- break;
- default:
- return -1;
- }
- return 0;
- }
- static inline int init_sock_keepalive(int s)
- {
- int optval;
-
- #ifdef HAVE_SO_KEEPALIVE
- if (cfg_get(tcp, tcp_cfg, keepalive)){
- optval=1;
- if (setsockopt(s, SOL_SOCKET, SO_KEEPALIVE, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to enable SO_KEEPALIVE: %s\n", strerror(errno));
- return -1;
- }
- }
- #endif
- #ifdef HAVE_TCP_KEEPINTVL
- if ((optval=cfg_get(tcp, tcp_cfg, keepintvl))){
- if (setsockopt(s, IPPROTO_TCP, TCP_KEEPINTVL, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set keepalive probes interval: %s\n", strerror(errno));
- }
- }
- #endif
- #ifdef HAVE_TCP_KEEPIDLE
- if ((optval=cfg_get(tcp, tcp_cfg, keepidle))){
- if (setsockopt(s, IPPROTO_TCP, TCP_KEEPIDLE, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set keepalive idle interval: %s\n", strerror(errno));
- }
- }
- #endif
- #ifdef HAVE_TCP_KEEPCNT
- if ((optval=cfg_get(tcp, tcp_cfg, keepcnt))){
- if (setsockopt(s, IPPROTO_TCP, TCP_KEEPCNT, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set maximum keepalive count: %s\n", strerror(errno));
- }
- }
- #endif
- return 0;
- }
- /* set all socket/fd options for new sockets (e.g. before connect):
- * disable nagle, tos lowdelay, reuseaddr, non-blocking
- *
- * return -1 on error */
- static int init_sock_opt(int s, int af)
- {
- int flags;
- int optval;
-
- #ifdef DISABLE_NAGLE
- flags=1;
- if ( (tcp_proto_no!=-1) && (setsockopt(s, tcp_proto_no , TCP_NODELAY,
- &flags, sizeof(flags))<0) ){
- LM_WARN("could not disable Nagle: %s\n", strerror(errno));
- }
- #endif
- /* tos*/
- optval = tos;
- if(af==AF_INET){
- if (setsockopt(s, IPPROTO_IP, IP_TOS, (void*)&optval,
- sizeof(optval)) ==-1){
- LM_WARN("setsockopt tos: %s\n", strerror(errno));
- /* continue since this is not critical */
- }
- } else if(af==AF_INET6){
- if (setsockopt(s, IPPROTO_IPV6, IPV6_TCLASS,
- (void*)&optval, sizeof(optval)) ==-1) {
- LM_WARN("setsockopt v6 tos: %s\n", strerror(errno));
- /* continue since this is not critical */
- }
- }
- #if !defined(TCP_DONT_REUSEADDR)
- optval=1;
- if (setsockopt(s, SOL_SOCKET, SO_REUSEADDR,
- (void*)&optval, sizeof(optval))==-1){
- LM_ERR("setsockopt SO_REUSEADDR %s\n", strerror(errno));
- /* continue, not critical */
- }
- #endif /* !TCP_DONT_REUSEADDR */
- #ifdef HAVE_TCP_SYNCNT
- if ((optval=cfg_get(tcp, tcp_cfg, syncnt))){
- if (setsockopt(s, IPPROTO_TCP, TCP_SYNCNT, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set maximum SYN retr. count: %s\n", strerror(errno));
- }
- }
- #endif
- #ifdef HAVE_TCP_LINGER2
- if ((optval=cfg_get(tcp, tcp_cfg, linger2))){
- if (setsockopt(s, IPPROTO_TCP, TCP_LINGER2, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set maximum LINGER2 timeout: %s\n", strerror(errno));
- }
- }
- #endif
- #ifdef HAVE_TCP_QUICKACK
- if (cfg_get(tcp, tcp_cfg, delayed_ack)){
- optval=0; /* reset quick ack => delayed ack */
- if (setsockopt(s, IPPROTO_TCP, TCP_QUICKACK, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to reset TCP_QUICKACK: %s\n", strerror(errno));
- }
- }
- #endif /* HAVE_TCP_QUICKACK */
- init_sock_keepalive(s);
-
- /* non-blocking */
- flags=fcntl(s, F_GETFL);
- if (flags==-1){
- LM_ERR("fnctl failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- if (fcntl(s, F_SETFL, flags|O_NONBLOCK)==-1){
- LM_ERR("fcntl: set non-blocking failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- return 0;
- error:
- return -1;
- }
- /* set all socket/fd options for "accepted" sockets
- * only nonblocking is set since the rest is inherited from the
- * "parent" (listening) socket
- * Note: setting O_NONBLOCK is required on linux but it's not needed on
- * BSD and possibly solaris (where the flag is inherited from the
- * parent socket). However since there is no standard document
- * requiring a specific behaviour in this case it's safer to always set
- * it (at least for now) --andrei
- * TODO: check on which OSes O_NONBLOCK is inherited and make this
- * function a nop.
- *
- * return -1 on error */
- static int init_sock_opt_accept(int s)
- {
- int flags;
-
- /* non-blocking */
- flags=fcntl(s, F_GETFL);
- if (flags==-1){
- LM_ERR("fnctl failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- if (fcntl(s, F_SETFL, flags|O_NONBLOCK)==-1){
- LM_ERR("fcntl: set non-blocking failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- return 0;
- error:
- return -1;
- }
- /** close a socket, handling errno.
- * On EINTR, repeat the close().
- * Filter expected errors (return success if close() failed because
- * EPIPE, ECONNRST a.s.o). Note that this happens on *BSDs (on linux close()
- * does not fail for socket level errors).
- * @param s - open valid socket.
- * @return - 0 on success, < 0 on error (whatever close() returns). On error
- * errno is set.
- */
- static int tcp_safe_close(int s)
- {
- int ret;
- retry:
- if (unlikely((ret = close(s)) < 0 )) {
- switch(errno) {
- case EINTR:
- goto retry;
- case EPIPE:
- case ENOTCONN:
- case ECONNRESET:
- case ECONNREFUSED:
- case ENETUNREACH:
- case EHOSTUNREACH:
- /* on *BSD we really get these errors at close() time
- => ignore them */
- ret = 0;
- break;
- default:
- break;
- }
- }
- return ret;
- }
- /* blocking connect on a non-blocking fd; it will timeout after
- * tcp_connect_timeout
- * if BLOCKING_USE_SELECT and HAVE_SELECT are defined it will internally
- * use select() instead of poll (bad if fd > FD_SET_SIZE, poll is preferred)
- */
- static int tcp_blocking_connect(int fd, int type, snd_flags_t* send_flags,
- const struct sockaddr *servaddr,
- socklen_t addrlen)
- {
- int n;
- #if defined(HAVE_SELECT) && defined(BLOCKING_USE_SELECT)
- fd_set sel_set;
- fd_set orig_set;
- struct timeval timeout;
- #else
- struct pollfd pf;
- #endif
- int elapsed;
- int to;
- int ticks;
- int err;
- unsigned int err_len;
- int poll_err;
-
- poll_err=0;
- to=cfg_get(tcp, tcp_cfg, connect_timeout_s);
- ticks=get_ticks();
- again:
- n=connect(fd, servaddr, addrlen);
- if (n==-1){
- if (errno==EINTR){
- elapsed=(get_ticks()-ticks)*TIMER_TICK;
- if (elapsed<to) goto again;
- else goto error_timeout;
- }
- if (errno!=EINPROGRESS && errno!=EALREADY){
- goto error_errno;
- }
- }else goto end;
-
- /* poll/select loop */
- #if defined(HAVE_SELECT) && defined(BLOCKING_USE_SELECT)
- FD_ZERO(&orig_set);
- FD_SET(fd, &orig_set);
- #else
- pf.fd=fd;
- pf.events=POLLOUT;
- #endif
- while(1){
- elapsed=(get_ticks()-ticks)*TIMER_TICK;
- if (elapsed>=to)
- goto error_timeout;
- #if defined(HAVE_SELECT) && defined(BLOCKING_USE_SELECT)
- sel_set=orig_set;
- timeout.tv_sec=to-elapsed;
- timeout.tv_usec=0;
- n=select(fd+1, 0, &sel_set, 0, &timeout);
- #else
- n=poll(&pf, 1, (to-elapsed)*1000);
- #endif
- if (n<0){
- if (errno==EINTR) continue;
- LM_ERR("%s: poll/select failed: (%d) %s\n",
- su2a((union sockaddr_union*)servaddr, addrlen),
- errno, strerror(errno));
- goto error;
- }else if (n==0) /* timeout */ continue;
- #if defined(HAVE_SELECT) && defined(BLOCKING_USE_SELECT)
- if (FD_ISSET(fd, &sel_set))
- #else
- if (pf.revents&(POLLERR|POLLHUP|POLLNVAL)){
- LM_ERR("%s: poll error: flags %x\n",
- su2a((union sockaddr_union*)servaddr, addrlen),
- pf.revents);
- poll_err=1;
- }
- #endif
- {
- err_len=sizeof(err);
- getsockopt(fd, SOL_SOCKET, SO_ERROR, &err, &err_len);
- if ((err==0) && (poll_err==0)) goto end;
- if (err!=EINPROGRESS && err!=EALREADY){
- LM_ERR("%s: SO_ERROR (%d) %s\n",
- su2a((union sockaddr_union*)servaddr, addrlen),
- err, strerror(err));
- errno=err;
- goto error_errno;
- }
- }
- }
- error_errno:
- switch(errno){
- case ENETUNREACH:
- case EHOSTUNREACH:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type,
- (union sockaddr_union*)servaddr, send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_UNREACHABLE(errno, 0, 0,
- (union sockaddr_union*)servaddr, type);
- break;
- case ETIMEDOUT:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type,
- (union sockaddr_union*)servaddr, send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(errno, 0, 0,
- (union sockaddr_union*)servaddr, type);
- break;
- case ECONNREFUSED:
- case ECONNRESET:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type,
- (union sockaddr_union*)servaddr, send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_RST(errno, 0, 0,
- (union sockaddr_union*)servaddr, type);
- break;
- case EAGAIN: /* not posix, but supported on linux and bsd */
- TCP_EV_CONNECT_NO_MORE_PORTS(errno, 0, 0,
- (union sockaddr_union*)servaddr, type);
- break;
- default:
- TCP_EV_CONNECT_ERR(errno, 0, 0,
- (union sockaddr_union*)servaddr, type);
- }
- LM_ERR("%s: (%d) %s\n",
- su2a((union sockaddr_union*)servaddr, addrlen),
- errno, strerror(errno));
- goto error;
- error_timeout:
- /* timeout */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type,
- (union sockaddr_union*)servaddr, send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(0, 0, 0, (union sockaddr_union*)servaddr, type);
- LM_ERR("%s: timeout %d s elapsed from %d s\n",
- su2a((union sockaddr_union*)servaddr, addrlen),
- elapsed, cfg_get(tcp, tcp_cfg, connect_timeout_s));
- error:
- TCP_STATS_CONNECT_FAILED();
- return -1;
- end:
- return 0;
- }
- #ifdef TCP_ASYNC
- /* unsafe version */
- #define _wbufq_empty(con) ((con)->wbuf_q.first==0)
- /* unsafe version */
- #define _wbufq_non_empty(con) ((con)->wbuf_q.first!=0)
- /* unsafe version, call while holding the connection write lock */
- inline static int _wbufq_add(struct tcp_connection* c, const char* data,
- unsigned int size)
- {
- struct tcp_wbuffer_queue* q;
- struct tcp_wbuffer* wb;
- unsigned int last_free;
- unsigned int wb_size;
- unsigned int crt_size;
- ticks_t t;
-
- q=&c->wbuf_q;
- t=get_ticks_raw();
- if (unlikely( ((q->queued+size)>cfg_get(tcp, tcp_cfg, tcpconn_wq_max)) ||
- ((*tcp_total_wq+size)>cfg_get(tcp, tcp_cfg, tcp_wq_max)) ||
- (q->first &&
- TICKS_LT(q->wr_timeout, t)) )){
- LM_ERR("(%d bytes): write queue full or timeout "
- " (%d, total %d, last write %d s ago)\n",
- size, q->queued, *tcp_total_wq,
- TICKS_TO_S(t-(q->wr_timeout-
- cfg_get(tcp, tcp_cfg, send_timeout))));
- if (q->first && TICKS_LT(q->wr_timeout, t)){
- if (unlikely(c->state==S_CONN_CONNECT)){
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su( BLST_ERR_CONNECT, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(0, TCP_LADDR(c), TCP_LPORT(c),
- TCP_PSU(c), TCP_PROTO(c));
- TCP_STATS_CONNECT_FAILED();
- }else{
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su( BLST_ERR_SEND, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_SEND_TIMEOUT(0, &c->rcv);
- TCP_STATS_SEND_TIMEOUT();
- }
- }else{
- /* if it's not a timeout => queue full */
- TCP_EV_SENDQ_FULL(0, &c->rcv);
- TCP_STATS_SENDQ_FULL();
- }
- goto error;
- }
-
- if (unlikely(q->last==0)){
- wb_size=MAX_unsigned(cfg_get(tcp, tcp_cfg, wq_blk_size), size);
- wb=shm_malloc(sizeof(*wb)+wb_size-1);
- if (unlikely(wb==0))
- goto error;
- wb->b_size=wb_size;
- wb->next=0;
- q->last=wb;
- q->first=wb;
- q->last_used=0;
- q->offset=0;
- q->wr_timeout=get_ticks_raw()+
- ((c->state==S_CONN_CONNECT)?
- S_TO_TICKS(cfg_get(tcp, tcp_cfg, connect_timeout_s)):
- cfg_get(tcp, tcp_cfg, send_timeout));
- }else{
- wb=q->last;
- }
-
- while(size){
- last_free=wb->b_size-q->last_used;
- if (last_free==0){
- wb_size=MAX_unsigned(cfg_get(tcp, tcp_cfg, wq_blk_size), size);
- wb=shm_malloc(sizeof(*wb)+wb_size-1);
- if (unlikely(wb==0))
- goto error;
- wb->b_size=wb_size;
- wb->next=0;
- q->last->next=wb;
- q->last=wb;
- q->last_used=0;
- last_free=wb->b_size;
- }
- crt_size=MIN_unsigned(last_free, size);
- memcpy(wb->buf+q->last_used, data, crt_size);
- q->last_used+=crt_size;
- size-=crt_size;
- data+=crt_size;
- q->queued+=crt_size;
- atomic_add_int((int*)tcp_total_wq, crt_size);
- }
- return 0;
- error:
- return -1;
- }
- /* unsafe version, call while holding the connection write lock
- * inserts data at the beginning, it ignores the max queue size checks and
- * the timeout (use sparingly)
- * Note: it should never be called on a write buffer after wbufq_run() */
- inline static int _wbufq_insert(struct tcp_connection* c, const char* data,
- unsigned int size)
- {
- struct tcp_wbuffer_queue* q;
- struct tcp_wbuffer* wb;
-
- q=&c->wbuf_q;
- if (likely(q->first==0)) /* if empty, use wbufq_add */
- return _wbufq_add(c, data, size);
-
- if (unlikely((*tcp_total_wq+size)>cfg_get(tcp, tcp_cfg, tcp_wq_max))){
- LM_ERR("(%d bytes): write queue full"
- " (%d, total %d, last write %d s ago)\n",
- size, q->queued, *tcp_total_wq,
- TICKS_TO_S(get_ticks_raw()-q->wr_timeout-
- cfg_get(tcp, tcp_cfg, send_timeout)));
- goto error;
- }
- if (unlikely(q->offset)){
- LM_CRIT("non-null offset %d (bad call, should"
- "never be called after the wbufq_run())\n", q->offset);
- goto error;
- }
- if ((q->first==q->last) && ((q->last->b_size-q->last_used)>=size)){
- /* one block with enough space in it for size bytes */
- memmove(q->first->buf+size, q->first->buf, q->last_used);
- memcpy(q->first->buf, data, size);
- q->last_used+=size;
- }else{
- /* create a size bytes block directly */
- wb=shm_malloc(sizeof(*wb)+size-1);
- if (unlikely(wb==0))
- goto error;
- wb->b_size=size;
- /* insert it */
- wb->next=q->first;
- q->first=wb;
- memcpy(wb->buf, data, size);
- }
-
- q->queued+=size;
- atomic_add_int((int*)tcp_total_wq, size);
- return 0;
- error:
- return -1;
- }
- /* unsafe version, call while holding the connection write lock */
- inline static void _wbufq_destroy( struct tcp_wbuffer_queue* q)
- {
- struct tcp_wbuffer* wb;
- struct tcp_wbuffer* next_wb;
- int unqueued;
-
- unqueued=0;
- if (likely(q->first)){
- wb=q->first;
- do{
- next_wb=wb->next;
- unqueued+=(wb==q->last)?q->last_used:wb->b_size;
- if (wb==q->first)
- unqueued-=q->offset;
- shm_free(wb);
- wb=next_wb;
- }while(wb);
- }
- memset(q, 0, sizeof(*q));
- atomic_add_int((int*)tcp_total_wq, -unqueued);
- }
- /* tries to empty the queue (safe version, c->write_lock must not be hold)
- * returns -1 on error, bytes written on success (>=0)
- * if the whole queue is emptied => sets *empty*/
- inline static int wbufq_run(int fd, struct tcp_connection* c, int* empty)
- {
- struct tcp_wbuffer_queue* q;
- struct tcp_wbuffer* wb;
- int n;
- int ret;
- int block_size;
- char* buf;
-
- *empty=0;
- ret=0;
- lock_get(&c->write_lock);
- q=&c->wbuf_q;
- while(q->first){
- block_size=((q->first==q->last)?q->last_used:q->first->b_size)-
- q->offset;
- buf=q->first->buf+q->offset;
- n=_tcpconn_write_nb(fd, c, buf, block_size);
- if (likely(n>0)){
- ret+=n;
- if (likely(n==block_size)){
- wb=q->first;
- q->first=q->first->next;
- shm_free(wb);
- q->offset=0;
- q->queued-=block_size;
- atomic_add_int((int*)tcp_total_wq, -block_size);
- }else{
- q->offset+=n;
- q->queued-=n;
- atomic_add_int((int*)tcp_total_wq, -n);
- break;
- }
- }else{
- if (n<0){
- /* EINTR is handled inside _tcpconn_write_nb */
- if (!(errno==EAGAIN || errno==EWOULDBLOCK)){
- if (unlikely(c->state==S_CONN_CONNECT)){
- switch(errno){
- case ENETUNREACH:
- case EHOSTUNREACH: /* not posix for send() */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT,
- c->rcv.proto,
- &c->rcv.src_su,
- &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_UNREACHABLE(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c),
- TCP_PROTO(c));
- break;
- case ECONNREFUSED:
- case ECONNRESET:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT,
- c->rcv.proto,
- &c->rcv.src_su,
- &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_RST(0, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c),
- TCP_PROTO(c));
- break;
- default:
- TCP_EV_CONNECT_ERR(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c),
- TCP_PROTO(c));
- }
- TCP_STATS_CONNECT_FAILED();
- }else{
- switch(errno){
- case ECONNREFUSED:
- case ECONNRESET:
- TCP_STATS_CON_RESET();
- /* no break */
- case ENETUNREACH:
- case EHOSTUNREACH: /* not posix for send() */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_SEND,
- c->rcv.proto,
- &c->rcv.src_su,
- &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- break;
- }
- }
- ret=-1;
- LM_ERR("%s [%d]\n", strerror(errno), errno);
- }
- }
- break;
- }
- }
- if (likely(q->first==0)){
- q->last=0;
- q->last_used=0;
- q->offset=0;
- *empty=1;
- }
- lock_release(&c->write_lock);
- if (likely(ret>0)){
- q->wr_timeout=get_ticks_raw()+cfg_get(tcp, tcp_cfg, send_timeout);
- if (unlikely(c->state==S_CONN_CONNECT || c->state==S_CONN_ACCEPT)){
- TCP_STATS_ESTABLISHED(c->state);
- c->state=S_CONN_OK;
- }
- }
- return ret;
- }
- #endif /* TCP_ASYNC */
- #if 0
- /* blocking write even on non-blocking sockets
- * if TCP_TIMEOUT will return with error */
- static int tcp_blocking_write(struct tcp_connection* c, int fd, char* buf,
- unsigned int len)
- {
- int n;
- fd_set sel_set;
- struct timeval timeout;
- int ticks;
- int initial_len;
-
- initial_len=len;
- again:
-
- n=send(fd, buf, len,
- #ifdef HAVE_MSG_NOSIGNAL
- MSG_NOSIGNAL
- #else
- 0
- #endif
- );
- if (n<0){
- if (errno==EINTR) goto again;
- else if (errno!=EAGAIN && errno!=EWOULDBLOCK){
- LM_ERR("failed to send: (%d) %s\n", errno, strerror(errno));
- TCP_EV_SEND_TIMEOUT(errno, &c->rcv);
- TCP_STATS_SEND_TIMEOUT();
- goto error;
- }
- }else if (n<len){
- /* partial write */
- buf+=n;
- len-=n;
- }else{
- /* success: full write */
- goto end;
- }
- while(1){
- FD_ZERO(&sel_set);
- FD_SET(fd, &sel_set);
- timeout.tv_sec=tcp_send_timeout;
- timeout.tv_usec=0;
- ticks=get_ticks();
- n=select(fd+1, 0, &sel_set, 0, &timeout);
- if (n<0){
- if (errno==EINTR) continue; /* signal, ignore */
- LM_ERR("select failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }else if (n==0){
- /* timeout */
- if (get_ticks()-ticks>=tcp_send_timeout){
- LM_ERR("send timeout (%d)\n", tcp_send_timeout);
- goto error;
- }
- continue;
- }
- if (FD_ISSET(fd, &sel_set)){
- /* we can write again */
- goto again;
- }
- }
- error:
- return -1;
- end:
- return initial_len;
- }
- #endif
- struct tcp_connection* tcpconn_new(int sock, union sockaddr_union* su,
- union sockaddr_union* local_addr,
- struct socket_info* ba, int type,
- int state)
- {
- struct tcp_connection *c;
- int rd_b_size;
-
- rd_b_size=cfg_get(tcp, tcp_cfg, rd_buf_size);
- c=shm_malloc(sizeof(struct tcp_connection) + rd_b_size);
- if (c==0){
- LM_ERR("mem. allocation failure\n");
- goto error;
- }
- memset(c, 0, sizeof(struct tcp_connection)); /* zero init (skip rd buf)*/
- c->s=sock;
- c->fd=-1; /* not initialized */
- if (lock_init(&c->write_lock)==0){
- LM_ERR("init lock failed\n");
- goto error;
- }
-
- c->rcv.src_su=*su;
-
- atomic_set(&c->refcnt, 0);
- local_timer_init(&c->timer, tcpconn_main_timeout, c, 0);
- su2ip_addr(&c->rcv.src_ip, su);
- c->rcv.src_port=su_getport(su);
- c->rcv.bind_address=ba;
- if (likely(local_addr)){
- su2ip_addr(&c->rcv.dst_ip, local_addr);
- c->rcv.dst_port=su_getport(local_addr);
- }else if (ba){
- c->rcv.dst_ip=ba->address;
- c->rcv.dst_port=ba->port_no;
- }
- print_ip("tcpconn_new: new tcp connection: ", &c->rcv.src_ip, "\n");
- DBG( "tcpconn_new: on port %d, type %d\n", c->rcv.src_port, type);
- init_tcp_req(&c->req, (char*)c+sizeof(struct tcp_connection), rd_b_size);
- c->id=(*connection_id)++;
- c->rcv.proto_reserved1=0; /* this will be filled before receive_message*/
- c->rcv.proto_reserved2=0;
- c->state=state;
- c->extra_data=0;
- #ifdef USE_TLS
- if (type==PROTO_TLS){
- if (tls_tcpconn_init(c, sock)==-1) goto error;
- }else
- #endif /* USE_TLS*/
- {
- c->type=PROTO_TCP;
- c->rcv.proto=PROTO_TCP;
- c->timeout=get_ticks_raw()+cfg_get(tcp, tcp_cfg, con_lifetime);
- }
-
- return c;
-
- error:
- if (c) shm_free(c);
- return 0;
- }
- /* do the actual connect, set sock. options a.s.o
- * returns socket on success, -1 on error
- * sets also *res_local_addr, res_si and state (S_CONN_CONNECT for an
- * unfinished connect and S_CONN_OK for a finished one)*/
- inline static int tcp_do_connect( union sockaddr_union* server,
- union sockaddr_union* from,
- int type,
- snd_flags_t* send_flags,
- union sockaddr_union* res_local_addr,
- struct socket_info** res_si,
- enum tcp_conn_states *state
- )
- {
- int s;
- union sockaddr_union my_name;
- socklen_t my_name_len;
- struct ip_addr ip;
- #ifdef TCP_ASYNC
- int n;
- #endif /* TCP_ASYNC */
- s=socket(AF2PF(server->s.sa_family), SOCK_STREAM, 0);
- if (unlikely(s==-1)){
- LM_ERR("%s: socket: (%d) %s\n",
- su2a(server, sizeof(*server)), errno, strerror(errno));
- goto error;
- }
- if (init_sock_opt(s, server->s.sa_family)<0){
- LM_ERR("%s: init_sock_opt failed\n",
- su2a(server, sizeof(*server)));
- goto error;
- }
-
- if (unlikely(from && bind(s, &from->s, sockaddru_len(*from)) != 0)){
- LM_WARN("binding to source address %s failed: %s [%d]\n",
- su2a(from, sizeof(*from)),
- strerror(errno), errno);
- }
- *state=S_CONN_OK;
- #ifdef TCP_ASYNC
- if (likely(cfg_get(tcp, tcp_cfg, async))){
- again:
- n=connect(s, &server->s, sockaddru_len(*server));
- if (likely(n==-1)){ /*non-blocking => most probable EINPROGRESS*/
- if (likely(errno==EINPROGRESS))
- *state=S_CONN_CONNECT;
- else if (errno==EINTR) goto again;
- else if (errno!=EALREADY){
- switch(errno){
- case ENETUNREACH:
- case EHOSTUNREACH:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type, server,
- send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_UNREACHABLE(errno, 0, 0, server, type);
- break;
- case ETIMEDOUT:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type, server,
- send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(errno, 0, 0, server, type);
- break;
- case ECONNREFUSED:
- case ECONNRESET:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, type, server,
- send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_RST(errno, 0, 0, server, type);
- break;
- case EAGAIN:/* not posix, but supported on linux and bsd */
- TCP_EV_CONNECT_NO_MORE_PORTS(errno, 0, 0, server,type);
- break;
- default:
- TCP_EV_CONNECT_ERR(errno, 0, 0, server, type);
- }
- TCP_STATS_CONNECT_FAILED();
- LM_ERR("connect %s: (%d) %s\n",
- su2a(server, sizeof(*server)),
- errno, strerror(errno));
- goto error;
- }
- }
- }else{
- #endif /* TCP_ASYNC */
- if (tcp_blocking_connect(s, type, send_flags, &server->s,
- sockaddru_len(*server))<0){
- LM_ERR("tcp_blocking_connect %s failed\n",
- su2a(server, sizeof(*server)));
- goto error;
- }
- #ifdef TCP_ASYNC
- }
- #endif /* TCP_ASYNC */
- if (from){
- su2ip_addr(&ip, from);
- if (!ip_addr_any(&ip))
- /* we already know the source ip, skip the sys. call */
- goto find_socket;
- }
- my_name_len=sizeof(my_name);
- if (unlikely(getsockname(s, &my_name.s, &my_name_len)!=0)){
- LM_ERR("getsockname failed: %s(%d)\n", strerror(errno), errno);
- *res_si=0;
- goto error;
- }
- from=&my_name; /* update from with the real "from" address */
- su2ip_addr(&ip, &my_name);
- find_socket:
- #ifdef USE_TLS
- if (unlikely(type==PROTO_TLS))
- *res_si=find_si(&ip, 0, PROTO_TLS);
- else
- #endif
- *res_si=find_si(&ip, 0, PROTO_TCP);
-
- if (unlikely(*res_si==0)){
- LM_WARN("%s: could not find corresponding"
- " listening socket for %s, using default...\n",
- su2a(server, sizeof(*server)), ip_addr2a(&ip));
- if (server->s.sa_family==AF_INET) *res_si=sendipv4_tcp;
- else *res_si=sendipv6_tcp;
- }
- *res_local_addr=*from;
- return s;
- error:
- if (s!=-1) tcp_safe_close(s);
- return -1;
- }
- struct tcp_connection* tcpconn_connect( union sockaddr_union* server,
- union sockaddr_union* from,
- int type, snd_flags_t* send_flags)
- {
- int s;
- struct socket_info* si;
- union sockaddr_union my_name;
- struct tcp_connection* con;
- enum tcp_conn_states state;
- s=-1;
-
- if (*tcp_connections_no >= cfg_get(tcp, tcp_cfg, max_connections)){
- LM_ERR("maximum number of connections exceeded (%d/%d)\n",
- *tcp_connections_no,
- cfg_get(tcp, tcp_cfg, max_connections));
- goto error;
- }
- if (unlikely(type==PROTO_TLS)) {
- if (*tls_connections_no >= cfg_get(tcp, tcp_cfg, max_tls_connections)){
- LM_ERR("maximum number of tls connections"
- " exceeded (%d/%d)\n",
- *tls_connections_no,
- cfg_get(tcp, tcp_cfg, max_tls_connections));
- goto error;
- }
- }
- s=tcp_do_connect(server, from, type, send_flags, &my_name, &si, &state);
- if (s==-1){
- LM_ERR("tcp_do_connect %s: failed (%d) %s\n",
- su2a(server, sizeof(*server)), errno, strerror(errno));
- goto error;
- }
- con=tcpconn_new(s, server, &my_name, si, type, state);
- if (con==0){
- LM_ERR("%s: tcpconn_new failed, closing the "
- " socket\n", su2a(server, sizeof(*server)));
- goto error;
- }
- tcpconn_set_send_flags(con, *send_flags);
- return con;
- error:
- if (s!=-1) tcp_safe_close(s); /* close the opened socket */
- return 0;
- }
- #ifdef TCP_CONNECT_WAIT
- int tcpconn_finish_connect( struct tcp_connection* c,
- union sockaddr_union* from)
- {
- int s;
- int r;
- union sockaddr_union local_addr;
- struct socket_info* si;
- enum tcp_conn_states state;
- struct tcp_conn_alias* a;
- int new_conn_alias_flags;
-
- s=tcp_do_connect(&c->rcv.src_su, from, c->type, &c->send_flags,
- &local_addr, &si, &state);
- if (unlikely(s==-1)){
- LM_ERR("%s: tcp_do_connect for %p failed\n",
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)), c);
- return -1;
- }
- c->rcv.bind_address=si;
- su2ip_addr(&c->rcv.dst_ip, &local_addr);
- c->rcv.dst_port=su_getport(&local_addr);
- /* update aliases if needed */
- if (likely(from==0)){
- new_conn_alias_flags=cfg_get(tcp, tcp_cfg, new_conn_alias_flags);
- /* add aliases */
- TCPCONN_LOCK;
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip, 0,
- new_conn_alias_flags);
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip,
- c->rcv.dst_port, new_conn_alias_flags);
- TCPCONN_UNLOCK;
- }else if (su_cmp(from, &local_addr)!=1){
- new_conn_alias_flags=cfg_get(tcp, tcp_cfg, new_conn_alias_flags);
- TCPCONN_LOCK;
- /* remove all the aliases except the first one and re-add them
- * (there shouldn't be more then the 3 default aliases at this
- * stage) */
- if (c->aliases > 1) {
- for (r=1; r<c->aliases; r++){
- a=&c->con_aliases[r];
- tcpconn_listrm(tcpconn_aliases_hash[a->hash],
- a, next, prev);
- }
- c->aliases=1;
- }
- /* add the local_ip:0 and local_ip:local_port aliases */
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip,
- 0, new_conn_alias_flags);
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip,
- c->rcv.dst_port, new_conn_alias_flags);
- TCPCONN_UNLOCK;
- }
-
- return s;
- }
- #endif /* TCP_CONNECT_WAIT */
- /* adds a tcp connection to the tcpconn hashes
- * Note: it's called _only_ from the tcp_main process */
- inline static struct tcp_connection* tcpconn_add(struct tcp_connection *c)
- {
- struct ip_addr zero_ip;
- int new_conn_alias_flags;
- if (likely(c)){
- ip_addr_mk_any(c->rcv.src_ip.af, &zero_ip);
- c->id_hash=tcp_id_hash(c->id);
- c->aliases=0;
- new_conn_alias_flags=cfg_get(tcp, tcp_cfg, new_conn_alias_flags);
- TCPCONN_LOCK;
- c->flags|=F_CONN_HASHED;
- /* add it at the begining of the list*/
- tcpconn_listadd(tcpconn_id_hash[c->id_hash], c, id_next, id_prev);
- /* set the aliases */
- /* first alias is for (peer_ip, peer_port, 0 ,0) -- for finding
- * any connection to peer_ip, peer_port
- * the second alias is for (peer_ip, peer_port, local_addr, 0) -- for
- * finding any conenction to peer_ip, peer_port from local_addr
- * the third alias is for (peer_ip, peer_port, local_addr, local_port)
- * -- for finding if a fully specified connection exists */
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &zero_ip, 0,
- new_conn_alias_flags);
- if (likely(c->rcv.dst_ip.af && ! ip_addr_any(&c->rcv.dst_ip))){
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip, 0,
- new_conn_alias_flags);
- _tcpconn_add_alias_unsafe(c, c->rcv.src_port, &c->rcv.dst_ip,
- c->rcv.dst_port, new_conn_alias_flags);
- }
- /* ignore add_alias errors, there are some valid cases when one
- * of the add_alias would fail (e.g. first add_alias for 2 connections
- * with the same destination but different src. ip*/
- TCPCONN_UNLOCK;
- DBG("tcpconn_add: hashes: %d:%d:%d, %d\n",
- c->con_aliases[0].hash,
- c->con_aliases[1].hash,
- c->con_aliases[2].hash,
- c->id_hash);
- return c;
- }else{
- LM_CRIT("null connection pointer\n");
- return 0;
- }
- }
- static inline void _tcpconn_detach(struct tcp_connection *c)
- {
- int r;
- tcpconn_listrm(tcpconn_id_hash[c->id_hash], c, id_next, id_prev);
- /* remove all the aliases */
- for (r=0; r<c->aliases; r++)
- tcpconn_listrm(tcpconn_aliases_hash[c->con_aliases[r].hash],
- &c->con_aliases[r], next, prev);
- c->aliases = 0;
- }
- static inline void _tcpconn_free(struct tcp_connection* c)
- {
- #ifdef TCP_ASYNC
- if (unlikely(_wbufq_non_empty(c)))
- _wbufq_destroy(&c->wbuf_q);
- #endif
- lock_destroy(&c->write_lock);
- #ifdef USE_TLS
- if (unlikely(c->type==PROTO_TLS || c->type==PROTO_WSS)) tls_tcpconn_clean(c);
- #endif
- shm_free(c);
- }
- /* unsafe tcpconn_rm version (nolocks) */
- void _tcpconn_rm(struct tcp_connection* c)
- {
- _tcpconn_detach(c);
- _tcpconn_free(c);
- }
- void tcpconn_rm(struct tcp_connection* c)
- {
- int r;
- TCPCONN_LOCK;
- tcpconn_listrm(tcpconn_id_hash[c->id_hash], c, id_next, id_prev);
- /* remove all the aliases */
- for (r=0; r<c->aliases; r++)
- tcpconn_listrm(tcpconn_aliases_hash[c->con_aliases[r].hash],
- &c->con_aliases[r], next, prev);
- c->aliases = 0;
- TCPCONN_UNLOCK;
- lock_destroy(&c->write_lock);
- #ifdef USE_TLS
- if ((c->type==PROTO_TLS || c->type==PROTO_WSS)&&(c->extra_data)) tls_tcpconn_clean(c);
- #endif
- shm_free(c);
- }
- /* finds a connection, if id=0 uses the ip addr, port, local_ip and local port
- * (host byte order) and tries to find the connection that matches all of
- * them. Wild cards can be used for local_ip and local_port (a 0 filled
- * ip address and/or a 0 local port).
- * WARNING: unprotected (locks) use tcpconn_get unless you really
- * know what you are doing */
- struct tcp_connection* _tcpconn_find(int id, struct ip_addr* ip, int port,
- struct ip_addr* l_ip, int l_port)
- {
- struct tcp_connection *c;
- struct tcp_conn_alias* a;
- unsigned hash;
- int is_local_ip_any;
-
- #ifdef EXTRA_DEBUG
- DBG("tcpconn_find: %d port %d\n",id, port);
- if (ip) print_ip("tcpconn_find: ip ", ip, "\n");
- #endif
- if (likely(id)){
- hash=tcp_id_hash(id);
- for (c=tcpconn_id_hash[hash]; c; c=c->id_next){
- #ifdef EXTRA_DEBUG
- DBG("c=%p, c->id=%d, port=%d\n",c, c->id, c->rcv.src_port);
- print_ip("ip=", &c->rcv.src_ip, "\n");
- #endif
- if ((id==c->id)&&(c->state!=S_CONN_BAD)) return c;
- }
- }else if (likely(ip)){
- hash=tcp_addr_hash(ip, port, l_ip, l_port);
- is_local_ip_any=ip_addr_any(l_ip);
- for (a=tcpconn_aliases_hash[hash]; a; a=a->next){
- #ifdef EXTRA_DEBUG
- DBG("a=%p, c=%p, c->id=%d, alias port= %d port=%d\n", a, a->parent,
- a->parent->id, a->port, a->parent->rcv.src_port);
- print_ip("ip=",&a->parent->rcv.src_ip,"\n");
- #endif
- if ( (a->parent->state!=S_CONN_BAD) && (port==a->port) &&
- ((l_port==0) || (l_port==a->parent->rcv.dst_port)) &&
- (ip_addr_cmp(ip, &a->parent->rcv.src_ip)) &&
- (is_local_ip_any ||
- ip_addr_cmp(l_ip, &a->parent->rcv.dst_ip))
- )
- return a->parent;
- }
- }
- return 0;
- }
- /* _tcpconn_find with locks and timeout
- * local_addr contains the desired local ip:port. If null any local address
- * will be used. IN*ADDR_ANY or 0 port are wild cards.
- */
- struct tcp_connection* tcpconn_get(int id, struct ip_addr* ip, int port,
- union sockaddr_union* local_addr,
- ticks_t timeout)
- {
- struct tcp_connection* c;
- struct ip_addr local_ip;
- int local_port;
-
- local_port=0;
- if (likely(ip)){
- if (unlikely(local_addr)){
- su2ip_addr(&local_ip, local_addr);
- local_port=su_getport(local_addr);
- }else{
- ip_addr_mk_any(ip->af, &local_ip);
- local_port=0;
- }
- }
- TCPCONN_LOCK;
- c=_tcpconn_find(id, ip, port, &local_ip, local_port);
- if (likely(c)){
- atomic_inc(&c->refcnt);
- /* update the timeout only if the connection is not handled
- * by a tcp reader _and_the timeout is non-zero (the tcp
- * reader process uses c->timeout for its own internal
- * timeout and c->timeout will be overwritten * anyway on
- * return to tcp_main) */
- if (likely(c->reader_pid==0 && timeout != 0))
- c->timeout=get_ticks_raw()+timeout;
- }
- TCPCONN_UNLOCK;
- return c;
- }
- /* add c->dst:port, local_addr as an alias for the "id" connection,
- * flags: TCP_ALIAS_FORCE_ADD - add an alias even if a previous one exists
- * TCP_ALIAS_REPLACE - if a prev. alias exists, replace it with the
- * new one
- * returns 0 on success, <0 on failure ( -1 - null c, -2 too many aliases,
- * -3 alias already present and pointing to another connection)
- * WARNING: must be called with TCPCONN_LOCK held */
- inline static int _tcpconn_add_alias_unsafe(struct tcp_connection* c, int port,
- struct ip_addr* l_ip, int l_port,
- int flags)
- {
- unsigned hash;
- struct tcp_conn_alias* a;
- struct tcp_conn_alias* nxt;
- struct tcp_connection* p;
- int is_local_ip_any;
- int i;
- int r;
-
- a=0;
- is_local_ip_any=ip_addr_any(l_ip);
- if (likely(c)){
- hash=tcp_addr_hash(&c->rcv.src_ip, port, l_ip, l_port);
- /* search the aliases for an already existing one */
- for (a=tcpconn_aliases_hash[hash], nxt=0; a; a=nxt){
- nxt=a->next;
- if ( (a->parent->state!=S_CONN_BAD) && (port==a->port) &&
- ( (l_port==0) || (l_port==a->parent->rcv.dst_port)) &&
- (ip_addr_cmp(&c->rcv.src_ip, &a->parent->rcv.src_ip)) &&
- ( is_local_ip_any ||
- ip_addr_cmp(&a->parent->rcv.dst_ip, l_ip))
- ){
- /* found */
- if (unlikely(a->parent!=c)){
- if (flags & TCP_ALIAS_FORCE_ADD)
- /* still have to walk the whole list to check if
- * the alias was not already added */
- continue;
- else if (flags & TCP_ALIAS_REPLACE){
- /* remove the alias =>
- * remove the current alias and all the following
- * ones from the corresponding connection, shift the
- * connection aliases array and re-add the other
- * aliases (!= current one) */
- p=a->parent;
- for (i=0; (i<p->aliases) && (&(p->con_aliases[i])!=a);
- i++);
- if (unlikely(i==p->aliases)){
- LM_CRIT("alias %p not found in con %p (id %d)\n",
- a, p, p->id);
- goto error_not_found;
- }
- for (r=i; r<p->aliases; r++){
- tcpconn_listrm(
- tcpconn_aliases_hash[p->con_aliases[r].hash],
- &p->con_aliases[r], next, prev);
- }
- if (likely((i+1)<p->aliases)){
- memmove(&p->con_aliases[i], &p->con_aliases[i+1],
- (p->aliases-i-1)*
- sizeof(p->con_aliases[0]));
- }
- p->aliases--;
- /* re-add the remaining aliases */
- for (r=i; r<p->aliases; r++){
- tcpconn_listadd(
- tcpconn_aliases_hash[p->con_aliases[r].hash],
- &p->con_aliases[r], next, prev);
- }
- }else
- goto error_sec;
- }else goto ok;
- }
- }
- if (unlikely(c->aliases>=TCP_CON_MAX_ALIASES)) goto error_aliases;
- c->con_aliases[c->aliases].parent=c;
- c->con_aliases[c->aliases].port=port;
- c->con_aliases[c->aliases].hash=hash;
- tcpconn_listadd(tcpconn_aliases_hash[hash],
- &c->con_aliases[c->aliases], next, prev);
- c->aliases++;
- }else goto error_not_found;
- ok:
- #ifdef EXTRA_DEBUG
- if (a) DBG("_tcpconn_add_alias_unsafe: alias already present\n");
- else DBG("_tcpconn_add_alias_unsafe: alias port %d for hash %d, id %d\n",
- port, hash, c->id);
- #endif
- return 0;
- error_aliases:
- /* too many aliases */
- return -2;
- error_not_found:
- /* null connection */
- return -1;
- error_sec:
- /* alias already present and pointing to a different connection
- * (hijack attempt?) */
- return -3;
- }
- /* add port as an alias for the "id" connection,
- * returns 0 on success,-1 on failure */
- int tcpconn_add_alias(int id, int port, int proto)
- {
- struct tcp_connection* c;
- int ret;
- struct ip_addr zero_ip;
- int r;
- int alias_flags;
-
- /* fix the port */
- port=port?port:((proto==PROTO_TLS)?SIPS_PORT:SIP_PORT);
- TCPCONN_LOCK;
- /* check if alias already exists */
- c=_tcpconn_find(id, 0, 0, 0, 0);
- if (likely(c)){
- ip_addr_mk_any(c->rcv.src_ip.af, &zero_ip);
- alias_flags=cfg_get(tcp, tcp_cfg, alias_flags);
- /* alias src_ip:port, 0, 0 */
- ret=_tcpconn_add_alias_unsafe(c, port, &zero_ip, 0,
- alias_flags);
- if (ret<0 && ret!=-3) goto error;
- /* alias src_ip:port, local_ip, 0 */
- ret=_tcpconn_add_alias_unsafe(c, port, &c->rcv.dst_ip, 0,
- alias_flags);
- if (ret<0 && ret!=-3) goto error;
- /* alias src_ip:port, local_ip, local_port */
- ret=_tcpconn_add_alias_unsafe(c, port, &c->rcv.dst_ip, c->rcv.dst_port,
- alias_flags);
- if (unlikely(ret<0)) goto error;
- }else goto error_not_found;
- TCPCONN_UNLOCK;
- return 0;
- error_not_found:
- TCPCONN_UNLOCK;
- LM_ERR("no connection found for id %d\n",id);
- return -1;
- error:
- TCPCONN_UNLOCK;
- switch(ret){
- case -2:
- LM_ERR("too many aliases (%d) for connection %p (id %d) %s:%d <- %d\n",
- c->aliases, c, c->id, ip_addr2a(&c->rcv.src_ip),
- c->rcv.src_port, port);
- for (r=0; r<c->aliases; r++){
- LM_ERR("alias %d: for %p (%d) %s:%d <-%d hash %x\n", r, c, c->id,
- ip_addr2a(&c->rcv.src_ip), c->rcv.src_port,
- c->con_aliases[r].port, c->con_aliases[r].hash);
- }
- break;
- case -3:
- LM_ERR("possible port hijack attempt\n");
- LM_ERR("alias for %d port %d already"
- " present and points to another connection \n",
- c->id, port);
- break;
- default:
- LM_ERR("unknown error %d\n", ret);
- }
- return -1;
- }
- #ifdef TCP_FD_CACHE
- static void tcp_fd_cache_init(void)
- {
- int r;
- for (r=0; r<TCP_FD_CACHE_SIZE; r++)
- fd_cache[r].fd=-1;
- }
- inline static struct fd_cache_entry* tcp_fd_cache_get(struct tcp_connection *c)
- {
- int h;
-
- h=c->id%TCP_FD_CACHE_SIZE;
- if ((fd_cache[h].fd>0) && (fd_cache[h].id==c->id) && (fd_cache[h].con==c))
- return &fd_cache[h];
- return 0;
- }
- inline static void tcp_fd_cache_rm(struct fd_cache_entry* e)
- {
- e->fd=-1;
- }
- inline static void tcp_fd_cache_add(struct tcp_connection *c, int fd)
- {
- int h;
-
- h=c->id%TCP_FD_CACHE_SIZE;
- if (likely(fd_cache[h].fd>0))
- tcp_safe_close(fd_cache[h].fd);
- fd_cache[h].fd=fd;
- fd_cache[h].id=c->id;
- fd_cache[h].con=c;
- }
- #endif /* TCP_FD_CACHE */
- inline static int tcpconn_chld_put(struct tcp_connection* tcpconn);
- static int tcpconn_send_put(struct tcp_connection* c, const char* buf,
- unsigned len, snd_flags_t send_flags);
- static int tcpconn_do_send(int fd, struct tcp_connection* c,
- const char* buf, unsigned len,
- snd_flags_t send_flags, long* resp, int locked);
- static int tcpconn_1st_send(int fd, struct tcp_connection* c,
- const char* buf, unsigned len,
- snd_flags_t send_flags, long* resp, int locked);
- /* finds a tcpconn & sends on it
- * uses the dst members to, proto (TCP|TLS) and id and tries to send
- * from the "from" address (if non null and id==0)
- * returns: number of bytes written (>=0) on success
- * <0 on error */
- int tcp_send(struct dest_info* dst, union sockaddr_union* from,
- const char* buf, unsigned len)
- {
- struct tcp_connection *c;
- struct ip_addr ip;
- int port;
- int fd;
- long response[2];
- int n;
- ticks_t con_lifetime;
- #ifdef USE_TLS
- const char* rest_buf;
- const char* t_buf;
- unsigned rest_len, t_len;
- long resp;
- snd_flags_t t_send_flags;
- #endif /* USE_TLS */
-
- port=su_getport(&dst->to);
- con_lifetime=cfg_get(tcp, tcp_cfg, con_lifetime);
- if (likely(port)){
- su2ip_addr(&ip, &dst->to);
- c=tcpconn_get(dst->id, &ip, port, from, con_lifetime);
- }else if (likely(dst->id)){
- c=tcpconn_get(dst->id, 0, 0, 0, con_lifetime);
- }else{
- LM_CRIT("null id & to\n");
- return -1;
- }
-
- if (likely(dst->id)){
- if (unlikely(c==0)) {
- if (likely(port)){
- /* try again w/o id */
- c=tcpconn_get(0, &ip, port, from, con_lifetime);
- }else{
- LM_ERR("id %d not found, dropping\n", dst->id);
- return -1;
- }
- }
- }
- /* connection not found or unusable => open a new one and send on it */
- if (unlikely((c==0) || tcpconn_close_after_send(c))){
- if (unlikely(c)){
- /* can't use c if it's marked as close-after-send =>
- release it and try opening new one */
- tcpconn_chld_put(c); /* release c (dec refcnt & free on 0) */
- c=0;
- }
- /* check if connect() is disabled */
- if (unlikely((dst->send_flags.f & SND_F_FORCE_CON_REUSE) ||
- cfg_get(tcp, tcp_cfg, no_connect)))
- return -1;
- DBG("tcp_send: no open tcp connection found, opening new one\n");
- /* create tcp connection */
- if (likely(from==0)){
- /* check to see if we have to use a specific source addr. */
- switch (dst->to.s.sa_family) {
- case AF_INET:
- from = tcp_source_ipv4;
- break;
- case AF_INET6:
- from = tcp_source_ipv6;
- break;
- default:
- /* error, bad af, ignore ... */
- break;
- }
- }
- #if defined(TCP_CONNECT_WAIT) && defined(TCP_ASYNC)
- if (likely(cfg_get(tcp, tcp_cfg, tcp_connect_wait) &&
- cfg_get(tcp, tcp_cfg, async) )){
- if (unlikely(*tcp_connections_no >=
- cfg_get(tcp, tcp_cfg, max_connections))){
- LM_ERR("%s: maximum number of connections exceeded (%d/%d)\n",
- su2a(&dst->to, sizeof(dst->to)),
- *tcp_connections_no,
- cfg_get(tcp, tcp_cfg, max_connections));
- return -1;
- }
- if (unlikely(dst->proto==PROTO_TLS)) {
- if (unlikely(*tls_connections_no >=
- cfg_get(tcp, tcp_cfg, max_tls_connections))){
- LM_ERR("%s: maximum number of tls connections exceeded (%d/%d)\n",
- su2a(&dst->to, sizeof(dst->to)),
- *tls_connections_no,
- cfg_get(tcp, tcp_cfg, max_tls_connections));
- return -1;
- }
- }
- c=tcpconn_new(-1, &dst->to, from, 0, dst->proto,
- S_CONN_CONNECT);
- if (unlikely(c==0)){
- LM_ERR("%s: could not create new connection\n",
- su2a(&dst->to, sizeof(dst->to)));
- return -1;
- }
- c->flags|=F_CONN_PENDING|F_CONN_FD_CLOSED;
- tcpconn_set_send_flags(c, dst->send_flags);
- atomic_set(&c->refcnt, 2); /* ref from here and from main hash
- table */
- /* add it to id hash and aliases */
- if (unlikely(tcpconn_add(c)==0)){
- LM_ERR("%s: could not add connection %p\n",
- su2a(&dst->to, sizeof(dst->to)), c);
- _tcpconn_free(c);
- n=-1;
- goto end_no_conn;
- }
- /* do connect and if src ip or port changed, update the
- * aliases */
- if (unlikely((fd=tcpconn_finish_connect(c, from))<0)){
- /* tcpconn_finish_connect will automatically blacklist
- on error => no need to do it here */
- LM_ERR("%s: tcpconn_finish_connect(%p) failed\n",
- su2a(&dst->to, sizeof(dst->to)), c);
- goto conn_wait_error;
- }
- /* ? TODO: it might be faster just to queue the write directly
- * and send to main CONN_NEW_PENDING_WRITE */
- /* delay sending the fd to main after the send */
-
- /* NOTE: no lock here, because the connection is marked as
- * pending and nobody else will try to write on it. However
- * this might produce out-of-order writes. If this is not
- * desired either lock before the write or use
- * _wbufq_insert(...)
- * NOTE2: _wbufq_insert() is used now (no out-of-order).
- */
- #ifdef USE_TLS
- if (unlikely(c->type==PROTO_TLS)) {
- /* for TLS the TLS processing and the send must happen
- atomically w/ respect to other sends on the same connection
- (otherwise reordering might occur which would break TLS) =>
- lock. However in this case this send will always be the first.
- We can have the send() outside the lock only if this is the
- first and only send (tls_encode is not called again), or
- this is the last send for a tls_encode() loop and all the
- previous ones did return CONN_NEW_COMPLETE or CONN_EOF.
- */
- response[1] = CONN_NOP;
- t_buf = buf;
- t_len = len;
- lock_get(&c->write_lock);
- redo_tls_encode:
- t_send_flags = dst->send_flags;
- n = tls_encode(c, &t_buf, &t_len, &rest_buf, &rest_len,
- &t_send_flags);
- /* There are 4 cases:
- 1. entire buffer consumed from the first try
- (rest_len == rest_buf == 0)
- 2. rest_buf & first call
- 3. rest_buf & not first call
- 3a. CONN_NEW_COMPLETE or CONN_EOF
- 3b. CONN_NEW_PENDING_WRITE
- 4. entire buffer consumed, but not first call
- 4a. CONN_NEW_COMPLETE or CONN_EOF
- 4b. CONN_NEW_PENDING_WRITE
- We misuse response[1] == CONN_NOP to test for the
- first call.
- */
- if (unlikely(n < 0)) {
- lock_release(&c->write_lock);
- goto conn_wait_error;
- }
- if (likely(rest_len == 0)) {
- /* 1 or 4*: CONN_NEW_COMPLETE, CONN_EOF, CONN_NOP
- or CONN_NEW_PENDING_WRITE (*rest_len == 0) */
- if (likely(response[1] != CONN_NEW_PENDING_WRITE)) {
- /* 1 or 4a => it's safe to do the send outside the
- lock (it will either send directly or
- wbufq_insert())
- */
- lock_release(&c->write_lock);
- if (likely(t_len != 0)) {
- n=tcpconn_1st_send(fd, c, t_buf, t_len,
- t_send_flags,
- &response[1], 0);
- } else { /* t_len == 0 */
- if (response[1] == CONN_NOP) {
- /* nothing to send (e.g parallel send
- tls_encode queues some data and then
- WANT_READ => this tls_encode will queue
- the cleartext too and will have nothing
- to send right now) and initial send =>
- behave as if the send was successful
- (but never return EOF here) */
- response[1] = CONN_NEW_COMPLETE;
- }
- }
- /* exit */
- } else {
- /* CONN_NEW_PENDING_WRITE: 4b: it was a
- repeated tls_encode() (or otherwise we would
- have here CONN_NOP) => add to the queue */
- if (unlikely(t_len &&
- _wbufq_add(c, t_buf, t_len) < 0)) {
- response[1] = CONN_ERROR;
- n = -1;
- }
- lock_release(&c->write_lock);
- /* exit (no send) */
- }
- } else { /* rest_len != 0 */
- /* 2 or 3*: if tls_encode hasn't finished, we have to
- call tcpconn_1st_send() under lock (otherwise if it
- returns CONN_NEW_PENDING_WRITE, there is no way
- to find the right place to add the new queued
- data from the 2nd tls_encode()) */
- if (likely((response[1] == CONN_NOP /*2*/ ||
- response[1] == CONN_NEW_COMPLETE /*3a*/ ||
- response[1] == CONN_EOF /*3a*/) && t_len))
- n = tcpconn_1st_send(fd, c, t_buf, t_len,
- t_send_flags,
- &response[1], 1);
- else if (unlikely(t_len &&
- _wbufq_add(c, t_buf, t_len) < 0)) {
- /*3b: CONN_NEW_PENDING_WRITE*/
- response[1] = CONN_ERROR;
- n = -1;
- }
- if (likely(n >= 0)) {
- /* if t_len == 0 => nothing was sent => previous
- response will be kept */
- t_buf = rest_buf;
- t_len = rest_len;
- goto redo_tls_encode;
- } else {
- lock_release(&c->write_lock);
- /* error exit */
- }
- }
- } else
- #endif /* USE_TLS */
- n=tcpconn_1st_send(fd, c, buf, len, dst->send_flags,
- &response[1], 0);
- if (unlikely(n<0)) /* this will catch CONN_ERROR too */
- goto conn_wait_error;
- if (unlikely(response[1]==CONN_EOF)){
- /* if close-after-send requested, don't bother
- sending the fd back to tcp_main, try closing it
- immediately (no other tcp_send should use it,
- because it is marked as close-after-send before
- being added to the hash) */
- goto conn_wait_close;
- }
- /* send to tcp_main */
- response[0]=(long)c;
- if (unlikely(send_fd(unix_tcp_sock, response,
- sizeof(response), fd) <= 0)){
- LM_ERR("%s: %ld for %p failed:" " %s (%d)\n",
- su2a(&dst->to, sizeof(dst->to)),
- response[1], c, strerror(errno), errno);
- goto conn_wait_error;
- }
- goto conn_wait_success;
- }
- #endif /* TCP_CONNECT_WAIT && TCP_ASYNC */
- if (unlikely((c=tcpconn_connect(&dst->to, from, dst->proto,
- &dst->send_flags))==0)){
- LM_ERR("%s: connect failed\n", su2a(&dst->to, sizeof(dst->to)));
- return -1;
- }
- tcpconn_set_send_flags(c, dst->send_flags);
- if (likely(c->state==S_CONN_OK))
- TCP_STATS_ESTABLISHED(S_CONN_CONNECT);
- atomic_set(&c->refcnt, 2); /* ref. from here and it will also
- be added in the tcp_main hash */
- fd=c->s;
- c->flags|=F_CONN_FD_CLOSED; /* not yet opened in main */
- /* ? TODO: it might be faster just to queue the write and
- * send to main a CONN_NEW_PENDING_WRITE */
-
- /* send the new tcpconn to "tcp main" */
- response[0]=(long)c;
- response[1]=CONN_NEW;
- n=send_fd(unix_tcp_sock, response, sizeof(response), c->s);
- if (unlikely(n<=0)){
- LM_ERR("%s: failed send_fd: %s (%d)\n",
- su2a(&dst->to, sizeof(dst->to)),
- strerror(errno), errno);
- /* we can safely delete it, it's not referenced by anybody */
- _tcpconn_free(c);
- n=-1;
- goto end_no_conn;
- }
- /* new connection => send on it directly */
- #ifdef USE_TLS
- if (unlikely(c->type==PROTO_TLS)) {
- /* for TLS the TLS processing and the send must happen
- atomically w/ respect to other sends on the same connection
- (otherwise reordering might occur which would break TLS) =>
- lock.
- */
- response[1] = CONN_NOP;
- t_buf = buf;
- t_len = len;
- lock_get(&c->write_lock);
- do {
- t_send_flags = dst->send_flags;
- n = tls_encode(c, &t_buf, &t_len, &rest_buf, &rest_len,
- &t_send_flags);
- if (likely(n > 0)) {
- n = tcpconn_do_send(fd, c, t_buf, t_len, t_send_flags,
- &resp, 1);
- if (likely(response[1] != CONN_QUEUED_WRITE ||
- resp == CONN_ERROR))
- /* don't overwrite a previous CONN_QUEUED_WRITE
- unless error */
- response[1] = resp;
- } else if (unlikely(n < 0)) {
- response[1] = CONN_ERROR;
- break;
- }
- /* else do nothing for n (t_len) == 0, keep
- the last reponse */
- t_buf = rest_buf;
- t_len = rest_len;
- } while(unlikely(rest_len && n > 0));
- lock_release(&c->write_lock);
- } else
- #endif /* USE_TLS */
- n = tcpconn_do_send(fd, c, buf, len, dst->send_flags,
- &response[1], 0);
- if (unlikely(response[1] != CONN_NOP)) {
- response[0]=(long)c;
- if (send_all(unix_tcp_sock, response, sizeof(response)) <= 0) {
- BUG("tcp_main command %ld sending failed (write):"
- "%s (%d)\n", response[1], strerror(errno), errno);
- /* all commands != CONN_NOP returned by tcpconn_do_send()
- (CONN_EOF, CONN_ERROR, CONN_QUEUED_WRITE) will auto-dec
- refcnt => if sending the command fails we have to
- dec. refcnt by hand */
- tcpconn_chld_put(c); /* deref. it manually */
- n=-1;
- }
- /* here refcnt for c is already decremented => c contents can
- no longer be used and refcnt _must_ _not_ be decremented
- again on exit */
- if (unlikely(n < 0 || response[1] == CONN_EOF)) {
- /* on error or eof, close fd */
- tcp_safe_close(fd);
- } else if (response[1] == CONN_QUEUED_WRITE) {
- #ifdef TCP_FD_CACHE
- if (cfg_get(tcp, tcp_cfg, fd_cache)) {
- tcp_fd_cache_add(c, fd);
- } else
- #endif /* TCP_FD_CACHE */
- tcp_safe_close(fd);
- } else {
- BUG("unexpected tcpconn_do_send() return & response:"
- " %d, %ld\n", n, response[1]);
- }
- goto end_no_deref;
- }
- #ifdef TCP_FD_CACHE
- if (cfg_get(tcp, tcp_cfg, fd_cache)) {
- tcp_fd_cache_add(c, fd);
- }else
- #endif /* TCP_FD_CACHE */
- tcp_safe_close(fd);
- /* here we can have only commands that _do_ _not_ dec refcnt.
- (CONN_EOF, CON_ERROR, CON_QUEUED_WRITE are all treated above) */
- goto release_c;
- } /* if (c==0 or unusable) new connection */
- /* existing connection, send on it */
- n = tcpconn_send_put(c, buf, len, dst->send_flags);
- /* no deref needed (automatically done inside tcpconn_send_put() */
- return n;
- #ifdef TCP_CONNECT_WAIT
- conn_wait_success:
- #ifdef TCP_FD_CACHE
- if (cfg_get(tcp, tcp_cfg, fd_cache)) {
- tcp_fd_cache_add(c, fd);
- } else
- #endif /* TCP_FD_CACHE */
- if (unlikely (tcp_safe_close(fd) < 0))
- LM_ERR("closing temporary send fd for %p: %s: "
- "close(%d) failed (flags 0x%x): %s (%d)\n", c,
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)),
- fd, c->flags, strerror(errno), errno);
- tcpconn_chld_put(c); /* release c (dec refcnt & free on 0) */
- return n;
- conn_wait_error:
- n=-1;
- conn_wait_close:
- /* connect or send failed or immediate close-after-send was requested on
- * newly created connection which was not yet sent to tcp_main (but was
- * already hashed) => don't send to main, unhash and destroy directly
- * (if refcnt>2 it will be destroyed when the last sender releases the
- * connection (tcpconn_chld_put(c))) or when tcp_main receives a
- * CONN_ERROR it*/
- c->state=S_CONN_BAD;
- /* we are here only if we opened a new fd (and not reused a cached or
- a reader one) => if the connect was successful close the fd */
- if (fd>=0) {
- if (unlikely(tcp_safe_close(fd) < 0 ))
- LM_ERR("closing temporary send fd for %p: %s: "
- "close(%d) failed (flags 0x%x): %s (%d)\n", c,
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)),
- fd, c->flags, strerror(errno), errno);
- }
- /* here the connection is for sure in the hash (tcp_main will not
- remove it because it's marked as PENDing) and the refcnt is at least
- 2
- */
- TCPCONN_LOCK;
- _tcpconn_detach(c);
- c->flags&=~F_CONN_HASHED;
- tcpconn_put(c);
- TCPCONN_UNLOCK;
- /* dec refcnt -> mark it for destruction */
- tcpconn_chld_put(c);
- return n;
- #endif /* TCP_CONNECT_WAIT */
- release_c:
- tcpconn_chld_put(c); /* release c (dec refcnt & free on 0) */
- end_no_deref:
- end_no_conn:
- return n;
- }
- /** sends on an existing tcpconn and auto-dec. con. ref counter.
- * As opposed to tcp_send(), this function requires an existing
- * tcp connection.
- * WARNING: the tcp_connection will be de-referenced.
- * @param c - existing tcp connection pointer.
- * @param buf - data to be sent.
- * @param len - data length,
- * @return >=0 on success, -1 on error.
- */
- static int tcpconn_send_put(struct tcp_connection* c, const char* buf,
- unsigned len, snd_flags_t send_flags)
- {
- struct tcp_connection *tmp;
- int fd;
- long response[2];
- int n;
- int do_close_fd;
- #ifdef USE_TLS
- const char* rest_buf;
- const char* t_buf;
- unsigned rest_len, t_len;
- long resp;
- snd_flags_t t_send_flags;
- #endif /* USE_TLS */
- #ifdef TCP_FD_CACHE
- struct fd_cache_entry* fd_cache_e;
- int use_fd_cache;
-
- use_fd_cache=cfg_get(tcp, tcp_cfg, fd_cache);
- fd_cache_e=0;
- #endif /* TCP_FD_CACHE */
- do_close_fd=1; /* close the fd on exit */
- response[1] = CONN_NOP;
- #ifdef TCP_ASYNC
- /* if data is already queued, we don't need the fd */
- #ifdef TCP_CONNECT_WAIT
- if (unlikely(cfg_get(tcp, tcp_cfg, async) &&
- (_wbufq_non_empty(c) || (c->flags&F_CONN_PENDING)) ))
- #else /* ! TCP_CONNECT_WAIT */
- if (unlikely(cfg_get(tcp, tcp_cfg, async) && (_wbufq_non_empty(c)) ))
- #endif /* TCP_CONNECT_WAIT */
- {
- lock_get(&c->write_lock);
- #ifdef TCP_CONNECT_WAIT
- if (likely(_wbufq_non_empty(c) || (c->flags&F_CONN_PENDING)))
- #else /* ! TCP_CONNECT_WAIT */
- if (likely(_wbufq_non_empty(c)))
- #endif /* TCP_CONNECT_WAIT */
- {
- do_close_fd=0;
- #ifdef USE_TLS
- if (unlikely(c->type==PROTO_TLS || c->type==PROTO_WSS)) {
- t_buf = buf;
- t_len = len;
- do {
- t_send_flags = send_flags;
- n = tls_encode(c, &t_buf, &t_len,
- &rest_buf, &rest_len,
- &t_send_flags);
- if (unlikely((n < 0) || (t_len &&
- (_wbufq_add(c, t_buf, t_len) < 0)))) {
- lock_release(&c->write_lock);
- n=-1;
- response[1] = CONN_ERROR;
- c->state=S_CONN_BAD;
- c->timeout=get_ticks_raw(); /* force timeout */
- goto error;
- }
- t_buf = rest_buf;
- t_len = rest_len;
- } while(unlikely(rest_len && n > 0));
- } else
- #endif /* USE_TLS */
- if (unlikely(len && (_wbufq_add(c, buf, len)<0))){
- lock_release(&c->write_lock);
- n=-1;
- response[1] = CONN_ERROR;
- c->state=S_CONN_BAD;
- c->timeout=get_ticks_raw(); /* force timeout */
- goto error;
- }
- n=len;
- lock_release(&c->write_lock);
- goto release_c;
- }
- lock_release(&c->write_lock);
- }
- #endif /* TCP_ASYNC */
- /* check if this is not the same reader process holding
- * c and if so send directly on c->fd */
- if (c->reader_pid==my_pid()){
- DBG("tcp_send: send from reader (%d (%d)), reusing fd\n",
- my_pid(), process_no);
- fd=c->fd;
- do_close_fd=0; /* don't close the fd on exit, it's in use */
- #ifdef TCP_FD_CACHE
- use_fd_cache=0; /* don't cache: problems would arise due to the
- close() on cache eviction (if the fd is still
- used). If it has to be cached then dup() _must_
- be used */
- }else if (likely(use_fd_cache &&
- ((fd_cache_e=tcp_fd_cache_get(c))!=0))){
- fd=fd_cache_e->fd;
- do_close_fd=0;
- DBG("tcp_send: found fd in cache ( %d, %p, %d)\n",
- fd, c, fd_cache_e->id);
- #endif /* TCP_FD_CACHE */
- }else{
- DBG("tcp_send: tcp connection found (%p), acquiring fd\n", c);
- /* get the fd */
- response[0]=(long)c;
- response[1]=CONN_GET_FD;
- n=send_all(unix_tcp_sock, response, sizeof(response));
- if (unlikely(n<=0)){
- LM_ERR("failed to get fd(write):%s (%d)\n", strerror(errno), errno);
- n=-1;
- goto release_c;
- }
- DBG("tcp_send, c= %p, n=%d\n", c, n);
- n=receive_fd(unix_tcp_sock, &tmp, sizeof(tmp), &fd, MSG_WAITALL);
- if (unlikely(n<=0)){
- LM_ERR("failed to get fd(receive_fd): %s (%d)\n",
- strerror(errno), errno);
- n=-1;
- do_close_fd=0;
- goto release_c;
- }
- /* handle fd closed or bad connection/error
- (it's possible that this happened in the time between
- we found the intial connection and the time when we get
- the fd)
- */
- if (unlikely(c!=tmp || fd==-1 || c->state==S_CONN_BAD)){
- if (unlikely(c!=tmp && tmp!=0))
- BUG("tcp_send: get_fd: got different connection:"
- " %p (id= %d, refcnt=%d state=%d) != "
- " %p (n=%d)\n",
- c, c->id, atomic_get(&c->refcnt), c->state,
- tmp, n
- );
- n=-1; /* fail */
- /* don't cache fd & close it */
- do_close_fd = (fd==-1)?0:1;
- #ifdef TCP_FD_CACHE
- use_fd_cache = 0;
- #endif /* TCP_FD_CACHE */
- goto end;
- }
- DBG("tcp_send: after receive_fd: c= %p n=%d fd=%d\n",c, n, fd);
- }
-
- #ifdef USE_TLS
- if (unlikely(c->type==PROTO_TLS || c->type==PROTO_WSS)) {
- /* for TLS the TLS processing and the send must happen
- atomically w/ respect to other sends on the same connection
- (otherwise reordering might occur which would break TLS) =>
- lock.
- */
- response[1] = CONN_NOP;
- t_buf = buf;
- t_len = len;
- lock_get(&c->write_lock);
- do {
- t_send_flags = send_flags;
- n = tls_encode(c, &t_buf, &t_len, &rest_buf, &rest_len,
- &t_send_flags);
- if (likely(n > 0)) {
- n = tcpconn_do_send(fd, c, t_buf, t_len, t_send_flags,
- &resp, 1);
- if (likely(response[1] != CONN_QUEUED_WRITE ||
- resp == CONN_ERROR))
- /* don't overwrite a previous CONN_QUEUED_WRITE
- unless error */
- response[1] = resp;
- } else if (unlikely(n < 0)) {
- response[1] = CONN_ERROR;
- break;
- }
- /* else do nothing for n (t_len) == 0, keep
- the last reponse */
- t_buf = rest_buf;
- t_len = rest_len;
- } while(unlikely(rest_len && n > 0));
- lock_release(&c->write_lock);
- } else
- #endif
- n = tcpconn_do_send(fd, c, buf, len, send_flags, &response[1], 0);
- if (unlikely(response[1] != CONN_NOP)) {
- error:
- response[0]=(long)c;
- if (send_all(unix_tcp_sock, response, sizeof(response)) <= 0) {
- BUG("tcp_main command %ld sending failed (write):%s (%d)\n",
- response[1], strerror(errno), errno);
- /* all commands != CONN_NOP returned by tcpconn_do_send()
- (CONN_EOF, CONN_ERROR, CONN_QUEUED_WRITE) will auto-dec refcnt
- => if sending the command fails we have to dec. refcnt by hand
- */
- tcpconn_chld_put(c); /* deref. it manually */
- n=-1;
- }
- /* here refcnt for c is already decremented => c contents can no
- longer be used and refcnt _must_ _not_ be decremented again
- on exit */
- if (unlikely(n < 0 || response[1] == CONN_EOF)) {
- /* on error or eof, remove from cache or close fd */
- #ifdef TCP_FD_CACHE
- if (unlikely(fd_cache_e)){
- tcp_fd_cache_rm(fd_cache_e);
- fd_cache_e = 0;
- tcp_safe_close(fd);
- }else
- #endif /* TCP_FD_CACHE */
- if (do_close_fd) tcp_safe_close(fd);
- } else if (response[1] == CONN_QUEUED_WRITE) {
- #ifdef TCP_FD_CACHE
- if (unlikely((fd_cache_e==0) && use_fd_cache)){
- tcp_fd_cache_add(c, fd);
- }else
- #endif /* TCP_FD_CACHE */
- if (do_close_fd) tcp_safe_close(fd);
- } else {
- BUG("unexpected tcpconn_do_send() return & response: %d, %ld\n",
- n, response[1]);
- }
- return n; /* no tcpconn_put */
- }
- end:
- #ifdef TCP_FD_CACHE
- if (unlikely((fd_cache_e==0) && use_fd_cache)){
- tcp_fd_cache_add(c, fd);
- }else
- #endif /* TCP_FD_CACHE */
- if (do_close_fd) {
- if (unlikely(tcp_safe_close(fd) < 0))
- LM_ERR("closing temporary send fd for %p: %s: "
- "close(%d) failed (flags 0x%x): %s (%d)\n", c,
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)),
- fd, c->flags, strerror(errno), errno);
- }
- /* here we can have only commands that _do_ _not_ dec refcnt.
- (CONN_EOF, CON_ERROR, CON_QUEUED_WRITE are all treated above) */
- release_c:
- tcpconn_chld_put(c); /* release c (dec refcnt & free on 0) */
- return n;
- }
- /* unsafe send on a known tcp connection.
- * Directly send on a known tcp connection with a given fd.
- * It is assumed that the connection locks are already held.
- * Side effects: if needed it will send state update commands to
- * tcp_main (e.g. CON_EOF, CON_ERROR, CON_QUEUED_WRITE).
- * @param fd - fd used for sending.
- * @param c - existing tcp connection pointer (state and flags might be
- * changed).
- * @param buf - data to be sent.
- * @param len - data length.
- * @param send_flags
- * @return <0 on error, number of bytes sent on success.
- */
- int tcpconn_send_unsafe(int fd, struct tcp_connection *c,
- const char* buf, unsigned len, snd_flags_t send_flags)
- {
- int n;
- long response[2];
-
- n = tcpconn_do_send(fd, c, buf, len, send_flags, &response[1], 1);
- if (unlikely(response[1] != CONN_NOP)) {
- /* all commands != CONN_NOP returned by tcpconn_do_send()
- (CONN_EOF, CONN_ERROR, CONN_QUEUED_WRITE) will auto-dec refcnt
- => increment it (we don't want the connection to be destroyed
- from under us)
- */
- atomic_inc(&c->refcnt);
- response[0]=(long)c;
- if (send_all(unix_tcp_sock, response, sizeof(response)) <= 0) {
- BUG("connection %p command %ld sending failed (write):%s (%d)\n",
- c, response[1], strerror(errno), errno);
- /* send failed => deref. it back by hand */
- tcpconn_chld_put(c);
- n=-1;
- }
- /* here refcnt for c is already decremented => c contents can no
- longer be used and refcnt _must_ _not_ be decremented again
- on exit */
- return n;
- }
- return n;
- }
- /** lower level send (connection and fd should be known).
- * It takes care of possible write-queueing, blacklisting a.s.o.
- * It expects a valid tcp connection. It doesn't touch the ref. cnts.
- * It will also set the connection flags from send_flags (it's better
- * to do it here, because it's guaranteed to be under lock).
- * @param fd - fd used for sending.
- * @param c - existing tcp connection pointer (state and flags might be
- * changed).
- * @param buf - data to be sent.
- * @param len - data length.
- * @param send_flags
- * @param resp - filled with a cmd. for tcp_main:
- * CONN_NOP - nothing needs to be done (do not send
- * anything to tcp_main).
- * CONN_ERROR - error, connection should be closed.
- * CONN_EOF - no error, but connection should be closed.
- * CONN_QUEUED_WRITE - new write queue (connection
- * should be watched for write and the wr.
- * queue flushed).
- * @param locked - if set assume the connection is already locked (call from
- * tls) and do not lock/unlock the connection.
- * @return >=0 on success, < 0 on error && *resp == CON_ERROR.
- *
- */
- static int tcpconn_do_send(int fd, struct tcp_connection* c,
- const char* buf, unsigned len,
- snd_flags_t send_flags, long* resp,
- int locked)
- {
- int n;
- #ifdef TCP_ASYNC
- int enable_write_watch;
- #endif /* TCP_ASYNC */
- DBG("tcp_send: sending...\n");
- *resp = CONN_NOP;
- if (likely(!locked)) lock_get(&c->write_lock);
- /* update connection send flags with the current ones */
- tcpconn_set_send_flags(c, send_flags);
- #ifdef TCP_ASYNC
- if (likely(cfg_get(tcp, tcp_cfg, async))){
- if (_wbufq_non_empty(c)
- #ifdef TCP_CONNECT_WAIT
- || (c->flags&F_CONN_PENDING)
- #endif /* TCP_CONNECT_WAIT */
- ){
- if (unlikely(_wbufq_add(c, buf, len)<0)){
- if (likely(!locked)) lock_release(&c->write_lock);
- n=-1;
- goto error;
- }
- if (likely(!locked)) lock_release(&c->write_lock);
- n=len;
- goto end;
- }
- n=_tcpconn_write_nb(fd, c, buf, len);
- }else{
- #endif /* TCP_ASYNC */
- /* n=tcp_blocking_write(c, fd, buf, len); */
- n=tsend_stream(fd, buf, len,
- TICKS_TO_S(cfg_get(tcp, tcp_cfg, send_timeout)) *
- 1000);
- #ifdef TCP_ASYNC
- }
- #else /* ! TCP_ASYNC */
- if (likely(!locked)) lock_release(&c->write_lock);
- #endif /* TCP_ASYNC */
-
- DBG("tcp_send: after real write: c= %p n=%d fd=%d\n",c, n, fd);
- DBG("tcp_send: buf=\n%.*s\n", (int)len, buf);
- if (unlikely(n<(int)len)){
- #ifdef TCP_ASYNC
- if (cfg_get(tcp, tcp_cfg, async) &&
- ((n>=0) || errno==EAGAIN || errno==EWOULDBLOCK)){
- enable_write_watch=_wbufq_empty(c);
- if (n<0) n=0;
- else if (unlikely(c->state==S_CONN_CONNECT ||
- c->state==S_CONN_ACCEPT)){
- TCP_STATS_ESTABLISHED(c->state);
- c->state=S_CONN_OK; /* something was written */
- }
- if (unlikely(_wbufq_add(c, buf+n, len-n)<0)){
- if (likely(!locked)) lock_release(&c->write_lock);
- n=-1;
- goto error;
- }
- if (likely(!locked)) lock_release(&c->write_lock);
- n=len;
- if (likely(enable_write_watch))
- *resp=CONN_QUEUED_WRITE;
- goto end;
- }else{
- if (likely(!locked)) lock_release(&c->write_lock);
- }
- #endif /* TCP_ASYNC */
- if (unlikely(c->state==S_CONN_CONNECT)){
- switch(errno){
- case ENETUNREACH:
- case EHOSTUNREACH: /* not posix for send() */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_UNREACHABLE(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c), TCP_PROTO(c));
- break;
- case ECONNREFUSED:
- case ECONNRESET:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_CONNECT, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_RST(errno, TCP_LADDR(c), TCP_LPORT(c),
- TCP_PSU(c), TCP_PROTO(c));
- break;
- default:
- TCP_EV_CONNECT_ERR(errno, TCP_LADDR(c), TCP_LPORT(c),
- TCP_PSU(c), TCP_PROTO(c));
- }
- TCP_STATS_CONNECT_FAILED();
- }else{
- switch(errno){
- case ECONNREFUSED:
- case ECONNRESET:
- TCP_STATS_CON_RESET();
- /* no break */
- case ENETUNREACH:
- /*case EHOSTUNREACH: -- not posix */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su(BLST_ERR_SEND, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- break;
- }
- }
- LM_ERR("failed to send on %p (%s:%d->%s): %s (%d)\n",
- c, ip_addr2a(&c->rcv.dst_ip), c->rcv.dst_port,
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)),
- strerror(errno), errno);
- n = -1;
- #ifdef TCP_ASYNC
- error:
- #endif /* TCP_ASYNC */
- /* error on the connection , mark it as bad and set 0 timeout */
- c->state=S_CONN_BAD;
- c->timeout=get_ticks_raw();
- /* tell "main" it should drop this (optional it will t/o anyway?)*/
- *resp=CONN_ERROR;
- return n; /* error return, no tcpconn_put */
- }
-
- #ifdef TCP_ASYNC
- if (likely(!locked)) lock_release(&c->write_lock);
- #endif /* TCP_ASYNC */
- /* in non-async mode here we're either in S_CONN_OK or S_CONN_ACCEPT*/
- if (unlikely(c->state==S_CONN_CONNECT || c->state==S_CONN_ACCEPT)){
- TCP_STATS_ESTABLISHED(c->state);
- c->state=S_CONN_OK;
- }
- if (unlikely(send_flags.f & SND_F_CON_CLOSE)){
- /* close after write => send EOF request to tcp_main */
- c->state=S_CONN_BAD;
- c->timeout=get_ticks_raw();
- /* tell "main" it should drop this*/
- *resp=CONN_EOF;
- return n;
- }
- end:
- return n;
- }
- /** low level 1st send on a new connection.
- * It takes care of possible write-queueing, blacklisting a.s.o.
- * It expects a valid just-opened tcp connection. It doesn't touch the
- * ref. counters. It's used only in the async first send case.
- * @param fd - fd used for sending.
- * @param c - existing tcp connection pointer (state and flags might be
- * changed). The connection must be new (no previous send on it).
- * @param buf - data to be sent.
- * @param len - data length.
- * @param send_flags
- * @param resp - filled with a fd sending cmd. for tcp_main on success. It
- * _must_ be one of the commands listed below:
- * CONN_NEW_PENDING_WRITE - new connection, first write
- * was partially successful (or EAGAIN) and
- * was queued (connection should be watched
- * for write and the write queue flushed).
- * The fd should be sent to tcp_main.
- * CONN_NEW_COMPLETE - new connection, first write
- * completed successfully and no data is
- * queued. The fd should be sent to tcp_main.
- * CONN_EOF - no error, but the connection should be
- * closed (e.g. SND_F_CON_CLOSE send flag).
- * CONN_ERROR - error, _must_ return < 0.
- * @param locked - if set assume the connection is already locked (call from
- * tls) and do not lock/unlock the connection.
- * @return >=0 on success, < 0 on error (on error *resp is undefined).
- *
- */
- static int tcpconn_1st_send(int fd, struct tcp_connection* c,
- const char* buf, unsigned len,
- snd_flags_t send_flags, long* resp,
- int locked)
- {
- int n;
-
- n=_tcpconn_write_nb(fd, c, buf, len);
- if (unlikely(n<(int)len)){
- /* on EAGAIN or ENOTCONN return success.
- ENOTCONN appears on newer FreeBSD versions (non-blocking socket,
- connect() & send immediately) */
- if ((n>=0) || errno==EAGAIN || errno==EWOULDBLOCK || errno==ENOTCONN){
- DBG("pending write on new connection %p "
- " (%d/%d bytes written)\n", c, n, len);
- if (unlikely(n<0)) n=0;
- else{
- if (likely(c->state == S_CONN_CONNECT))
- TCP_STATS_ESTABLISHED(S_CONN_CONNECT);
- c->state=S_CONN_OK; /* partial write => connect()
- ended */
- }
- /* add to the write queue */
- if (likely(!locked)) lock_get(&c->write_lock);
- if (unlikely(_wbufq_insert(c, buf+n, len-n)<0)){
- if (likely(!locked)) lock_release(&c->write_lock);
- n=-1;
- LM_ERR("%s: EAGAIN and write queue full or failed for %p\n",
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)), c);
- goto error;
- }
- if (likely(!locked)) lock_release(&c->write_lock);
- /* send to tcp_main */
- *resp=CONN_NEW_PENDING_WRITE;
- n=len;
- goto end;
- }
- /* n < 0 and not EAGAIN => write error */
- /* if first write failed it's most likely a
- connect error */
- switch(errno){
- case ENETUNREACH:
- case EHOSTUNREACH: /* not posix for send() */
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su( BLST_ERR_CONNECT, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_UNREACHABLE(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c), TCP_PROTO(c));
- break;
- case ECONNREFUSED:
- case ECONNRESET:
- #ifdef USE_DST_BLACKLIST
- dst_blacklist_su( BLST_ERR_CONNECT, c->rcv.proto,
- &c->rcv.src_su, &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_RST(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c), TCP_PROTO(c));
- break;
- default:
- TCP_EV_CONNECT_ERR(errno, TCP_LADDR(c),
- TCP_LPORT(c), TCP_PSU(c), TCP_PROTO(c));
- }
- /* error: destroy it directly */
- TCP_STATS_CONNECT_FAILED();
- LM_ERR("%s: connect & send for %p failed:" " %s (%d)\n",
- su2a(&c->rcv.src_su, sizeof(c->rcv.src_su)),
- c, strerror(errno), errno);
- goto error;
- }
- LM_INFO("quick connect for %p\n", c);
- if (likely(c->state == S_CONN_CONNECT))
- TCP_STATS_ESTABLISHED(S_CONN_CONNECT);
- if (unlikely(send_flags.f & SND_F_CON_CLOSE)){
- /* close after write => EOF => close immediately */
- c->state=S_CONN_BAD;
- /* tell our caller that it should drop this*/
- *resp=CONN_EOF;
- }else{
- c->state=S_CONN_OK;
- /* send to tcp_main */
- *resp=CONN_NEW_COMPLETE;
- }
- end:
- return n; /* >= 0 */
- error:
- *resp=CONN_ERROR;
- return -1;
- }
- int tcp_init(struct socket_info* sock_info)
- {
- union sockaddr_union* addr;
- int optval;
- #ifdef HAVE_TCP_ACCEPT_FILTER
- struct accept_filter_arg afa;
- #endif /* HAVE_TCP_ACCEPT_FILTER */
- #ifdef DISABLE_NAGLE
- int flag;
- struct protoent* pe;
- if (tcp_proto_no==-1){ /* if not already set */
- pe=getprotobyname("tcp");
- if (pe==0){
- LM_ERR("could not get TCP protocol number\n");
- tcp_proto_no=-1;
- }else{
- tcp_proto_no=pe->p_proto;
- }
- }
- #endif
-
- addr=&sock_info->su;
- /* sock_info->proto=PROTO_TCP; */
- if (init_su(addr, &sock_info->address, sock_info->port_no)<0){
- LM_ERR("could no init sockaddr_union\n");
- goto error;
- }
- DBG("tcp_init: added %s\n", su2a(addr, sizeof(*addr)));
- sock_info->socket=socket(AF2PF(addr->s.sa_family), SOCK_STREAM, 0);
- if (sock_info->socket==-1){
- LM_ERR("tcp_init: socket: %s\n", strerror(errno));
- goto error;
- }
- #ifdef DISABLE_NAGLE
- flag=1;
- if ( (tcp_proto_no!=-1) &&
- (setsockopt(sock_info->socket, tcp_proto_no , TCP_NODELAY,
- &flag, sizeof(flag))<0) ){
- LM_ERR("could not disable Nagle: %s\n", strerror(errno));
- }
- #endif
- #if !defined(TCP_DONT_REUSEADDR)
- /* Stevens, "Network Programming", Section 7.5, "Generic Socket
- * Options": "...server started,..a child continues..on existing
- * connection..listening server is restarted...call to bind fails
- * ... ALL TCP servers should specify the SO_REUSEADDRE option
- * to allow the server to be restarted in this situation
- *
- * Indeed, without this option, the server can't restart.
- * -jiri
- */
- optval=1;
- if (setsockopt(sock_info->socket, SOL_SOCKET, SO_REUSEADDR,
- (void*)&optval, sizeof(optval))==-1) {
- LM_ERR("setsockopt %s\n", strerror(errno));
- goto error;
- }
- #endif
- /* tos */
- optval = tos;
- if (setsockopt(sock_info->socket, IPPROTO_IP, IP_TOS, (void*)&optval,
- sizeof(optval)) ==-1){
- LM_WARN("setsockopt tos: %s\n", strerror(errno));
- /* continue since this is not critical */
- }
- #ifdef HAVE_TCP_DEFER_ACCEPT
- /* linux only */
- if ((optval=cfg_get(tcp, tcp_cfg, defer_accept))){
- if (setsockopt(sock_info->socket, IPPROTO_TCP, TCP_DEFER_ACCEPT,
- (void*)&optval, sizeof(optval)) ==-1){
- LM_WARN("setsockopt TCP_DEFER_ACCEPT %s\n", strerror(errno));
- /* continue since this is not critical */
- }
- }
- #endif /* HAVE_TCP_DEFFER_ACCEPT */
- #ifdef HAVE_TCP_SYNCNT
- if ((optval=cfg_get(tcp, tcp_cfg, syncnt))){
- if (setsockopt(sock_info->socket, IPPROTO_TCP, TCP_SYNCNT, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set maximum SYN retr. count: %s\n", strerror(errno));
- }
- }
- #endif
- #ifdef HAVE_TCP_LINGER2
- if ((optval=cfg_get(tcp, tcp_cfg, linger2))){
- if (setsockopt(sock_info->socket, IPPROTO_TCP, TCP_LINGER2, &optval,
- sizeof(optval))<0){
- LM_WARN("failed to set maximum LINGER2 timeout: %s\n", strerror(errno));
- }
- }
- #endif
- init_sock_keepalive(sock_info->socket);
- if (bind(sock_info->socket, &addr->s, sockaddru_len(*addr))==-1){
- LM_ERR("bind(%x, %p, %d) on %s:%d : %s\n",
- sock_info->socket, &addr->s,
- (unsigned)sockaddru_len(*addr),
- sock_info->address_str.s,
- sock_info->port_no,
- strerror(errno));
- goto error;
- }
- if (listen(sock_info->socket, TCP_LISTEN_BACKLOG)==-1){
- LM_ERR("listen(%x, %p, %d) on %s: %s\n",
- sock_info->socket, &addr->s,
- (unsigned)sockaddru_len(*addr),
- sock_info->address_str.s,
- strerror(errno));
- goto error;
- }
- #ifdef HAVE_TCP_ACCEPT_FILTER
- /* freebsd */
- if (cfg_get(tcp, tcp_cfg, defer_accept)){
- memset(&afa, 0, sizeof(afa));
- strcpy(afa.af_name, "dataready");
- if (setsockopt(sock_info->socket, SOL_SOCKET, SO_ACCEPTFILTER,
- (void*)&afa, sizeof(afa)) ==-1){
- LM_WARN("setsockopt SO_ACCEPTFILTER %s\n",
- strerror(errno));
- /* continue since this is not critical */
- }
- }
- #endif /* HAVE_TCP_ACCEPT_FILTER */
-
- return 0;
- error:
- if (sock_info->socket!=-1){
- tcp_safe_close(sock_info->socket);
- sock_info->socket=-1;
- }
- return -1;
- }
- /* close tcp_main's fd from a tcpconn
- * WARNING: call only in tcp_main context */
- inline static void tcpconn_close_main_fd(struct tcp_connection* tcpconn)
- {
- int fd;
-
-
- fd=tcpconn->s;
- #ifdef USE_TLS
- if (tcpconn->type==PROTO_TLS || tcpconn->type==PROTO_WSS)
- tls_close(tcpconn, fd);
- #endif
- #ifdef TCP_FD_CACHE
- if (likely(cfg_get(tcp, tcp_cfg, fd_cache))) shutdown(fd, SHUT_RDWR);
- #endif /* TCP_FD_CACHE */
- if (unlikely(tcp_safe_close(fd)<0))
- LM_ERR("(%p): %s close(%d) failed (flags 0x%x): %s (%d)\n", tcpconn,
- su2a(&tcpconn->rcv.src_su, sizeof(tcpconn->rcv.src_su)),
- fd, tcpconn->flags, strerror(errno), errno);
- tcpconn->s=-1;
- }
- /* dec refcnt & frees the connection if refcnt==0
- * returns 1 if the connection is freed, 0 otherwise
- *
- * WARNING: use only from child processes */
- inline static int tcpconn_chld_put(struct tcp_connection* tcpconn)
- {
- if (unlikely(atomic_dec_and_test(&tcpconn->refcnt))){
- DBG("tcpconn_chld_put: destroying connection %p (%d, %d) "
- "flags %04x\n", tcpconn, tcpconn->id,
- tcpconn->s, tcpconn->flags);
- /* sanity checks */
- membar_read_atomic_op(); /* make sure we see the current flags */
- if (unlikely(!(tcpconn->flags & F_CONN_FD_CLOSED) ||
- (tcpconn->flags &
- (F_CONN_HASHED|F_CONN_MAIN_TIMER|
- F_CONN_READ_W|F_CONN_WRITE_W)) )){
- LM_CRIT("%p bad flags = %0x\n", tcpconn, tcpconn->flags);
- abort();
- }
- _tcpconn_free(tcpconn); /* destroys also the wbuf_q if still present*/
- return 1;
- }
- return 0;
- }
- /* simple destroy function (the connection should be already removed
- * from the hashes. refcnt 0 and the fds should not be watched anymore for IO)
- */
- inline static void tcpconn_destroy(struct tcp_connection* tcpconn)
- {
- DBG("tcpconn_destroy: destroying connection %p (%d, %d) "
- "flags %04x\n", tcpconn, tcpconn->id,
- tcpconn->s, tcpconn->flags);
- if (unlikely(tcpconn->flags & F_CONN_HASHED)){
- LM_CRIT("called with hashed connection (%p)\n", tcpconn);
- /* try to continue */
- if (likely(tcpconn->flags & F_CONN_MAIN_TIMER))
- local_timer_del(&tcp_main_ltimer, &tcpconn->timer);
- TCPCONN_LOCK;
- _tcpconn_detach(tcpconn);
- tcpconn->flags &= ~(F_CONN_HASHED|F_CONN_MAIN_TIMER);
- TCPCONN_UNLOCK;
- }
- if (likely(!(tcpconn->flags & F_CONN_FD_CLOSED))){
- tcpconn_close_main_fd(tcpconn);
- tcpconn->flags|=F_CONN_FD_CLOSED;
- (*tcp_connections_no)--;
- if (unlikely(tcpconn->type==PROTO_TLS || tcpconn->type==PROTO_WSS))
- (*tls_connections_no)--;
- }
- _tcpconn_free(tcpconn); /* destroys also the wbuf_q if still present*/
- }
- /* tries to destroy the connection: dec. refcnt and if 0 destroys the
- * connection, else it will mark it as BAD and close the main fds
- *
- * returns 1 if the connection was destroyed, 0 otherwise
- *
- * WARNING: - the connection _has_ to be removed from the hash and timer
- * first (use tcpconn_try_unhash() for this )
- * - the fd should not be watched anymore (io_watch_del()...)
- * - must be called _only_ from the tcp_main process context
- * (or else the fd will remain open)
- */
- inline static int tcpconn_put_destroy(struct tcp_connection* tcpconn)
- {
- if (unlikely((tcpconn->flags &
- (F_CONN_WRITE_W|F_CONN_HASHED|F_CONN_MAIN_TIMER|F_CONN_READ_W)) )){
- /* sanity check */
- if (unlikely(tcpconn->flags & F_CONN_HASHED)){
- LM_CRIT("called with hashed and/or"
- "on timer connection (%p), flags = %0x\n",
- tcpconn, tcpconn->flags);
- /* try to continue */
- if (likely(tcpconn->flags & F_CONN_MAIN_TIMER))
- local_timer_del(&tcp_main_ltimer, &tcpconn->timer);
- TCPCONN_LOCK;
- _tcpconn_detach(tcpconn);
- tcpconn->flags &= ~(F_CONN_HASHED|F_CONN_MAIN_TIMER);
- TCPCONN_UNLOCK;
- }else{
- LM_CRIT("%p flags = %0x\n", tcpconn, tcpconn->flags);
- }
- }
- tcpconn->state=S_CONN_BAD;
- /* in case it's still in a reader timer */
- tcpconn->timeout=get_ticks_raw();
- /* fast close: close fds now */
- if (likely(!(tcpconn->flags & F_CONN_FD_CLOSED))){
- tcpconn_close_main_fd(tcpconn);
- tcpconn->flags|=F_CONN_FD_CLOSED;
- (*tcp_connections_no)--;
- if (unlikely(tcpconn->type==PROTO_TLS || tcpconn->type==PROTO_WSS))
- (*tls_connections_no)--;
- }
- /* all the flags / ops on the tcpconn must be done prior to decrementing
- * the refcnt. and at least a membar_write_atomic_op() mem. barrier or
- * a mb_atomic_* op must * be used to make sure all the changed flags are
- * written into memory prior to the new refcnt value */
- if (unlikely(mb_atomic_dec_and_test(&tcpconn->refcnt))){
- _tcpconn_free(tcpconn);
- return 1;
- }
- return 0;
- }
- /* try to remove a connection from the hashes and timer.
- * returns 1 if the connection was removed, 0 if not (connection not in
- * hash)
- *
- * WARNING: call it only in the tcp_main process context or else the
- * timer removal won't work.
- */
- inline static int tcpconn_try_unhash(struct tcp_connection* tcpconn)
- {
- if (likely(tcpconn->flags & F_CONN_HASHED)){
- tcpconn->state=S_CONN_BAD;
- if (likely(tcpconn->flags & F_CONN_MAIN_TIMER)){
- local_timer_del(&tcp_main_ltimer, &tcpconn->timer);
- tcpconn->flags&=~F_CONN_MAIN_TIMER;
- }else
- /* in case it's still in a reader timer */
- tcpconn->timeout=get_ticks_raw();
- TCPCONN_LOCK;
- if (tcpconn->flags & F_CONN_HASHED){
- tcpconn->flags&=~F_CONN_HASHED;
- _tcpconn_detach(tcpconn);
- TCPCONN_UNLOCK;
- }else{
- /* tcp_send was faster and did unhash it itself */
- TCPCONN_UNLOCK;
- return 0;
- }
- #ifdef TCP_ASYNC
- /* empty possible write buffers (optional) */
- if (unlikely(_wbufq_non_empty(tcpconn))){
- lock_get(&tcpconn->write_lock);
- /* check again, while holding the lock */
- if (likely(_wbufq_non_empty(tcpconn)))
- _wbufq_destroy(&tcpconn->wbuf_q);
- lock_release(&tcpconn->write_lock);
- }
- #endif /* TCP_ASYNC */
- return 1;
- }
- return 0;
- }
- #ifdef SEND_FD_QUEUE
- struct send_fd_info{
- struct tcp_connection* tcp_conn;
- ticks_t expire;
- int unix_sock;
- unsigned int retries; /* debugging */
- };
- struct tcp_send_fd_q{
- struct send_fd_info* data; /* buffer */
- struct send_fd_info* crt; /* pointer inside the buffer */
- struct send_fd_info* end; /* points after the last valid position */
- };
- static struct tcp_send_fd_q send2child_q;
- static int send_fd_queue_init(struct tcp_send_fd_q *q, unsigned int size)
- {
- q->data=pkg_malloc(size*sizeof(struct send_fd_info));
- if (q->data==0){
- LM_ERR("out of memory\n");
- return -1;
- }
- q->crt=&q->data[0];
- q->end=&q->data[size];
- return 0;
- }
- static void send_fd_queue_destroy(struct tcp_send_fd_q *q)
- {
- if (q->data){
- pkg_free(q->data);
- q->data=0;
- q->crt=q->end=0;
- }
- }
- static int init_send_fd_queues(void)
- {
- if (send_fd_queue_init(&send2child_q, SEND_FD_QUEUE_SIZE)!=0)
- goto error;
- return 0;
- error:
- LM_ERR("init failed\n");
- return -1;
- }
- static void destroy_send_fd_queues(void)
- {
- send_fd_queue_destroy(&send2child_q);
- }
- inline static int send_fd_queue_add( struct tcp_send_fd_q* q,
- int unix_sock,
- struct tcp_connection *t)
- {
- struct send_fd_info* tmp;
- unsigned long new_size;
-
- if (q->crt>=q->end){
- new_size=q->end-&q->data[0];
- if (new_size< MAX_SEND_FD_QUEUE_SIZE/2){
- new_size*=2;
- }else new_size=MAX_SEND_FD_QUEUE_SIZE;
- if (unlikely(q->crt>=&q->data[new_size])){
- LM_ERR("queue full: %ld/%ld\n",
- (long)(q->crt-&q->data[0]-1), new_size);
- goto error;
- }
- LM_CRIT("queue full: %ld, extending to %ld\n",
- (long)(q->end-&q->data[0]), new_size);
- tmp=pkg_realloc(q->data, new_size*sizeof(struct send_fd_info));
- if (unlikely(tmp==0)){
- LM_ERR("out of memory\n");
- goto error;
- }
- q->crt=(q->crt-&q->data[0])+tmp;
- q->data=tmp;
- q->end=&q->data[new_size];
- }
- q->crt->tcp_conn=t;
- q->crt->unix_sock=unix_sock;
- q->crt->expire=get_ticks_raw()+SEND_FD_QUEUE_TIMEOUT;
- q->crt->retries=0;
- q->crt++;
- return 0;
- error:
- return -1;
- }
- inline static void send_fd_queue_run(struct tcp_send_fd_q* q)
- {
- struct send_fd_info* p;
- struct send_fd_info* t;
-
- for (p=t=&q->data[0]; p<q->crt; p++){
- if (unlikely(p->tcp_conn->state == S_CONN_BAD ||
- p->tcp_conn->flags & F_CONN_FD_CLOSED ||
- p->tcp_conn->s ==-1)) {
- /* bad and/or already closed connection => remove */
- goto rm_con;
- }
- if (unlikely(send_fd(p->unix_sock, &(p->tcp_conn),
- sizeof(struct tcp_connection*), p->tcp_conn->s)<=0)){
- if ( ((errno==EAGAIN)||(errno==EWOULDBLOCK)) &&
- ((s_ticks_t)(p->expire-get_ticks_raw())>0)){
- /* leave in queue for a future try */
- *t=*p;
- t->retries++;
- t++;
- }else{
- LM_ERR("send_fd failed on socket %d , queue entry %ld, retries %d,"
- " connection %p, tcp socket %d, errno=%d (%s) \n",
- p->unix_sock, (long)(p-&q->data[0]), p->retries,
- p->tcp_conn, p->tcp_conn->s, errno,
- strerror(errno));
- rm_con:
- #ifdef TCP_ASYNC
- /* if a connection is on the send_fd queue it means it's
- not watched for read anymore => could be watched only for
- write */
- if (p->tcp_conn->flags & F_CONN_WRITE_W){
- io_watch_del(&io_h, p->tcp_conn->s, -1, IO_FD_CLOSING);
- p->tcp_conn->flags &=~F_CONN_WRITE_W;
- }
- #endif
- p->tcp_conn->flags &= ~F_CONN_READER;
- if (likely(tcpconn_try_unhash(p->tcp_conn)))
- tcpconn_put(p->tcp_conn);
- tcpconn_put_destroy(p->tcp_conn); /* dec refcnt & destroy */
- }
- }
- }
- q->crt=t;
- }
- #else
- #define send_fd_queue_run(q)
- #endif
- /* non blocking write() on a tcpconnection, unsafe version (should be called
- * while holding c->write_lock). The fd should be non-blocking.
- * returns number of bytes written on success, -1 on error (and sets errno)
- */
- int _tcpconn_write_nb(int fd, struct tcp_connection* c,
- const char* buf, int len)
- {
- int n;
-
- again:
- n=send(fd, buf, len,
- #ifdef HAVE_MSG_NOSIGNAL
- MSG_NOSIGNAL
- #else
- 0
- #endif /* HAVE_MSG_NOSIGNAL */
- );
- if (unlikely(n<0)){
- if (errno==EINTR) goto again;
- }
- return n;
- }
- /* handles io from a tcp child process
- * params: tcp_c - pointer in the tcp_children array, to the entry for
- * which an io event was detected
- * fd_i - fd index in the fd_array (usefull for optimizing
- * io_watch_deletes)
- * returns: handle_* return convention: -1 on error, 0 on EAGAIN (no more
- * io events queued), >0 on success. success/error refer only to
- * the reads from the fd.
- */
- inline static int handle_tcp_child(struct tcp_child* tcp_c, int fd_i)
- {
- struct tcp_connection* tcpconn;
- long response[2];
- int cmd;
- int bytes;
- int n;
- ticks_t t;
- ticks_t crt_timeout;
- ticks_t con_lifetime;
-
- if (unlikely(tcp_c->unix_sock<=0)){
- /* (we can't have a fd==0, 0 is never closed )*/
- LM_CRIT("fd %d for %d (pid %ld, ser no %d)\n", tcp_c->unix_sock,
- (int)(tcp_c-&tcp_children[0]), (long)tcp_c->pid,
- tcp_c->proc_no);
- goto error;
- }
- /* read until sizeof(response)
- * (this is a SOCK_STREAM so read is not atomic) */
- bytes=recv_all(tcp_c->unix_sock, response, sizeof(response), MSG_DONTWAIT);
- if (unlikely(bytes<(int)sizeof(response))){
- if (bytes==0){
- /* EOF -> bad, child has died */
- DBG("DBG: handle_tcp_child: dead tcp child %d (pid %ld, no %d)"
- " (shutting down?)\n", (int)(tcp_c-&tcp_children[0]),
- (long)tcp_c->pid, tcp_c->proc_no );
- /* don't listen on it any more */
- io_watch_del(&io_h, tcp_c->unix_sock, fd_i, 0);
- goto error; /* eof. so no more io here, it's ok to return error */
- }else if (bytes<0){
- /* EAGAIN is ok if we try to empty the buffer
- * e.g.: SIGIO_RT overflow mode or EPOLL ET */
- if ((errno!=EAGAIN) && (errno!=EWOULDBLOCK)){
- LM_CRIT("read from tcp child %ld (pid %ld, no %d) %s [%d]\n",
- (long)(tcp_c-&tcp_children[0]), (long)tcp_c->pid,
- tcp_c->proc_no, strerror(errno), errno );
- }else{
- bytes=0;
- }
- /* try to ignore ? */
- goto end;
- }else{
- /* should never happen */
- LM_CRIT("too few bytes received (%d)\n", bytes );
- bytes=0; /* something was read so there is no error; otoh if
- receive_fd returned less then requested => the receive
- buffer is empty => no more io queued on this fd */
- goto end;
- }
- }
-
- DBG("handle_tcp_child: reader response= %lx, %ld from %d \n",
- response[0], response[1], (int)(tcp_c-&tcp_children[0]));
- cmd=response[1];
- tcpconn=(struct tcp_connection*)response[0];
- if (unlikely(tcpconn==0)){
- /* should never happen */
- LM_CRIT("null tcpconn pointer received from tcp child %d (pid %ld): %lx, %lx\n",
- (int)(tcp_c-&tcp_children[0]), (long)tcp_c->pid,
- response[0], response[1]) ;
- goto end;
- }
- switch(cmd){
- case CONN_RELEASE:
- tcp_c->busy--;
- if (unlikely(tcpconn_put(tcpconn))){
- /* if refcnt was 1 => it was used only in the
- tcp reader => it's not hashed or watched for IO
- anymore => no need to io_watch_del() */
- tcpconn_destroy(tcpconn);
- break;
- }
- if (unlikely(tcpconn->state==S_CONN_BAD)){
- if (tcpconn_try_unhash(tcpconn)) {
- #ifdef TCP_ASYNC
- if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- tcpconn->flags &= ~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- tcpconn_put_destroy(tcpconn);
- }
- #ifdef TCP_ASYNC
- else if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- /* should never happen: if it's already unhashed, it
- should not be watched for IO */
- BUG("unhashed connection watched for write\n");
- io_watch_del(&io_h, tcpconn->s, -1, 0);
- tcpconn->flags &= ~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- break;
- }
- /* update the timeout*/
- t=get_ticks_raw();
- con_lifetime=cfg_get(tcp, tcp_cfg, con_lifetime);
- tcpconn->timeout=t+con_lifetime;
- crt_timeout=con_lifetime;
- #ifdef TCP_ASYNC
- if (unlikely(cfg_get(tcp, tcp_cfg, async) &&
- _wbufq_non_empty(tcpconn) )){
- if (unlikely(TICKS_GE(t, tcpconn->wbuf_q.wr_timeout))){
- DBG("handle_tcp_child: wr. timeout on CONN_RELEASE for %p "
- "refcnt= %d\n", tcpconn,
- atomic_get(&tcpconn->refcnt));
- /* timeout */
- if (unlikely(tcpconn->state==S_CONN_CONNECT)){
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su( BLST_ERR_CONNECT,
- tcpconn->rcv.proto,
- &tcpconn->rcv.src_su,
- &tcpconn->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(0, TCP_LADDR(tcpconn),
- TCP_LPORT(tcpconn), TCP_PSU(tcpconn),
- TCP_PROTO(tcpconn));
- TCP_STATS_CONNECT_FAILED();
- }else{
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su( BLST_ERR_SEND,
- tcpconn->rcv.proto,
- &tcpconn->rcv.src_su,
- &tcpconn->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_SEND_TIMEOUT(0, &tcpconn->rcv);
- TCP_STATS_SEND_TIMEOUT();
- }
- if (tcpconn_try_unhash(tcpconn)) {
- if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- tcpconn_put_destroy(tcpconn);
- } else if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- BUG("unhashed connection watched for write\n");
- io_watch_del(&io_h, tcpconn->s, -1, 0);
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- break;
- }else{
- crt_timeout=MIN_unsigned(con_lifetime,
- tcpconn->wbuf_q.wr_timeout-t);
- }
- }
- #endif /* TCP_ASYNC */
- /* re-activate the timer */
- tcpconn->timer.f=tcpconn_main_timeout;
- local_timer_reinit(&tcpconn->timer);
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer, crt_timeout, t);
- /* must be after the de-ref*/
- tcpconn->flags|=(F_CONN_MAIN_TIMER|F_CONN_READ_W|F_CONN_WANTS_RD);
- tcpconn->flags&=~(F_CONN_READER|F_CONN_OOB_DATA);
- #ifdef TCP_ASYNC
- if (unlikely(tcpconn->flags & F_CONN_WRITE_W))
- n=io_watch_chg(&io_h, tcpconn->s, POLLIN| POLLOUT, -1);
- else
- #endif /* TCP_ASYNC */
- n=io_watch_add(&io_h, tcpconn->s, POLLIN, F_TCPCONN, tcpconn);
- if (unlikely(n<0)){
- LM_CRIT("failed to add new socket to the fd list\n");
- tcpconn->flags&=~F_CONN_READ_W;
- if (tcpconn_try_unhash(tcpconn)) {
- #ifdef TCP_ASYNC
- if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- tcpconn_put_destroy(tcpconn);
- }
- #ifdef TCP_ASYNC
- else if (unlikely(tcpconn->flags & F_CONN_WRITE_W)) {
- BUG("unhashed connection watched for write\n");
- io_watch_del(&io_h, tcpconn->s, -1, 0);
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- break;
- }
- DBG("handle_tcp_child: CONN_RELEASE %p refcnt= %d\n",
- tcpconn, atomic_get(&tcpconn->refcnt));
- break;
- case CONN_ERROR:
- case CONN_DESTROY:
- case CONN_EOF:
- /* WARNING: this will auto-dec. refcnt! */
- tcp_c->busy--;
- /* main doesn't listen on it => we don't have to delete it
- if (tcpconn->s!=-1)
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- */
- #ifdef TCP_ASYNC
- if ((tcpconn->flags & F_CONN_WRITE_W) && (tcpconn->s!=-1)){
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- if (tcpconn_try_unhash(tcpconn))
- tcpconn_put(tcpconn);
- tcpconn_put_destroy(tcpconn); /* deref & delete if refcnt==0 */
- break;
- default:
- LM_CRIT("unknown cmd %d from tcp reader %d\n",
- cmd, (int)(tcp_c-&tcp_children[0]));
- }
- end:
- return bytes;
- error:
- return -1;
- }
- /* handles io from a "generic" ser process (get fd or new_fd from a tcp_send)
- *
- * params: p - pointer in the ser processes array (pt[]), to the entry for
- * which an io event was detected
- * fd_i - fd index in the fd_array (usefull for optimizing
- * io_watch_deletes)
- * returns: handle_* return convention:
- * -1 on error reading from the fd,
- * 0 on EAGAIN or when no more io events are queued
- * (receive buffer empty),
- * >0 on successfull reads from the fd (the receive buffer might
- * be non-empty).
- */
- inline static int handle_ser_child(struct process_table* p, int fd_i)
- {
- struct tcp_connection* tcpconn;
- struct tcp_connection* tmp;
- long response[2];
- int cmd;
- int bytes;
- int ret;
- int fd;
- int flags;
- ticks_t t;
- ticks_t con_lifetime;
- #ifdef TCP_ASYNC
- ticks_t nxt_timeout;
- #endif /* TCP_ASYNC */
-
- ret=-1;
- if (unlikely(p->unix_sock<=0)){
- /* (we can't have a fd==0, 0 is never closed )*/
- LM_CRIT("fd %d for %d (pid %d)\n", p->unix_sock, (int)(p-&pt[0]), p->pid);
- goto error;
- }
-
- /* get all bytes and the fd (if transmitted)
- * (this is a SOCK_STREAM so read is not atomic) */
- bytes=receive_fd(p->unix_sock, response, sizeof(response), &fd,
- MSG_DONTWAIT);
- if (unlikely(bytes<(int)sizeof(response))){
- /* too few bytes read */
- if (bytes==0){
- /* EOF -> bad, child has died */
- DBG("DBG: handle_ser_child: dead child %d, pid %d"
- " (shutting down?)\n", (int)(p-&pt[0]), p->pid);
- /* don't listen on it any more */
- io_watch_del(&io_h, p->unix_sock, fd_i, 0);
- goto error; /* child dead => no further io events from it */
- }else if (bytes<0){
- /* EAGAIN is ok if we try to empty the buffer
- * e.g: SIGIO_RT overflow mode or EPOLL ET */
- if ((errno!=EAGAIN) && (errno!=EWOULDBLOCK)){
- LM_CRIT("read from child %d (pid %d): %s [%d]\n",
- (int)(p-&pt[0]), p->pid,
- strerror(errno), errno);
- ret=-1;
- }else{
- ret=0;
- }
- /* try to ignore ? */
- goto end;
- }else{
- /* should never happen */
- LM_CRIT("too few bytes received (%d)\n", bytes );
- ret=0; /* something was read so there is no error; otoh if
- receive_fd returned less then requested => the receive
- buffer is empty => no more io queued on this fd */
- goto end;
- }
- }
- ret=1; /* something was received, there might be more queued */
- DBG("handle_ser_child: read response= %lx, %ld, fd %d from %d (%d)\n",
- response[0], response[1], fd, (int)(p-&pt[0]), p->pid);
- cmd=response[1];
- tcpconn=(struct tcp_connection*)response[0];
- if (unlikely(tcpconn==0)){
- LM_CRIT("null tcpconn pointer received from child %d (pid %d): %lx, %lx\n",
- (int)(p-&pt[0]), p->pid, response[0], response[1]) ;
- goto end;
- }
- switch(cmd){
- case CONN_ERROR:
- LM_ERR("received CON_ERROR for %p (id %d), refcnt %d, flags 0x%0x\n",
- tcpconn, tcpconn->id, atomic_get(&tcpconn->refcnt),
- tcpconn->flags);
- case CONN_EOF: /* forced EOF after full send, due to send flags */
- #ifdef TCP_CONNECT_WAIT
- /* if the connection is marked as pending => it might be on
- * the way of reaching tcp_main (e.g. CONN_NEW_COMPLETE or
- * CONN_NEW_PENDING_WRITE) => it cannot be destroyed here,
- * it will be destroyed on CONN_NEW_COMPLETE /
- * CONN_NEW_PENDING_WRITE or in the send error case by the
- * sender process */
- if (unlikely(tcpconn->flags & F_CONN_PENDING)) {
- if (tcpconn_put(tcpconn))
- tcpconn_destroy(tcpconn);
- /* no need for io_watch_del(), if PENDING it should not
- be watched for anything in tcp_main */
- break;
- }
- #endif /* TCP_CONNECT_WAIT */
- if ( tcpconn_try_unhash(tcpconn) )
- tcpconn_put(tcpconn);
- if ( ((tcpconn->flags & (F_CONN_WRITE_W|F_CONN_READ_W)) ) &&
- (tcpconn->s!=-1)){
- io_watch_del(&io_h, tcpconn->s, -1, IO_FD_CLOSING);
- tcpconn->flags&=~(F_CONN_WRITE_W|F_CONN_READ_W);
- }
- tcpconn_put_destroy(tcpconn); /* dec refcnt & destroy on 0 */
- break;
- case CONN_GET_FD:
- /* send the requested FD */
- /* WARNING: take care of setting refcnt properly to
- * avoid race conditions */
- if (unlikely(tcpconn->state == S_CONN_BAD ||
- (tcpconn->flags & F_CONN_FD_CLOSED) ||
- tcpconn->s ==-1)) {
- /* connection is already marked as bad and/or has no
- fd => don't try to send the fd (trying to send a
- closed fd _will_ fail) */
- tmp = 0;
- if (unlikely(send_all(p->unix_sock, &tmp, sizeof(tmp)) <= 0))
- BUG("handle_ser_child: CONN_GET_FD: send_all failed\n");
- /* no need to attempt to destroy the connection, it should
- be already in the process of being destroyed */
- } else if (unlikely(send_fd(p->unix_sock, &tcpconn,
- sizeof(tcpconn), tcpconn->s)<=0)){
- LM_ERR("CONN_GET_FD: send_fd failed\n");
- /* try sending error (better then not sending anything) */
- tmp = 0;
- if (unlikely(send_all(p->unix_sock, &tmp, sizeof(tmp)) <= 0))
- BUG("handle_ser_child: CONN_GET_FD:"
- " send_fd send_all fallback failed\n");
- }
- break;
- case CONN_NEW:
- /* update the fd in the requested tcpconn*/
- /* WARNING: take care of setting refcnt properly to
- * avoid race conditions */
- if (unlikely(fd==-1)){
- LM_CRIT("CONN_NEW: no fd received\n");
- tcpconn->flags|=F_CONN_FD_CLOSED;
- tcpconn_put_destroy(tcpconn);
- break;
- }
- (*tcp_connections_no)++;
- if (unlikely(tcpconn->type==PROTO_TLS))
- (*tls_connections_no)++;
- tcpconn->s=fd;
- /* add tcpconn to the list*/
- tcpconn_add(tcpconn);
- /* update the timeout*/
- t=get_ticks_raw();
- con_lifetime=cfg_get(tcp, tcp_cfg, con_lifetime);
- tcpconn->timeout=t+con_lifetime;
- /* activate the timer (already properly init. in tcpconn_new())
- * no need for reinit */
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer,
- con_lifetime, t);
- tcpconn->flags|=(F_CONN_MAIN_TIMER|F_CONN_READ_W|F_CONN_WANTS_RD)
- #ifdef TCP_ASYNC
- /* not used for now, the connection is sent to tcp_main
- * before knowing whether we can write on it or we should
- * wait */
- | (((int)!(tcpconn->flags & F_CONN_WANTS_WR)-1)&
- F_CONN_WRITE_W)
- #endif /* TCP_ASYNC */
- ;
- tcpconn->flags&=~F_CONN_FD_CLOSED;
- flags=POLLIN
- #ifdef TCP_ASYNC
- /* not used for now, the connection is sent to tcp_main
- * before knowing if we can write on it or we should
- * wait */
- | (((int)!(tcpconn->flags & F_CONN_WANTS_WR)-1) & POLLOUT)
- #endif /* TCP_ASYNC */
- ;
- if (unlikely(
- io_watch_add(&io_h, tcpconn->s, flags,
- F_TCPCONN, tcpconn)<0)){
- LM_CRIT("failed to add new socket to the fd list\n");
- tcpconn->flags&=~(F_CONN_WRITE_W|F_CONN_READ_W);
- tcpconn_try_unhash(tcpconn); /* unhash & dec refcnt */
- tcpconn_put_destroy(tcpconn);
- }
- break;
- #ifdef TCP_ASYNC
- case CONN_QUEUED_WRITE:
- /* received only if the wr. queue is empty and a write finishes
- * with EAGAIN (common after connect())
- * it should only enable write watching on the fd. The connection
- * should be already in the hash. The refcnt is automatically
- * decremented.
- */
- /* auto-dec refcnt */
- if (unlikely(tcpconn_put(tcpconn))){
- tcpconn_destroy(tcpconn);
- break;
- }
- if (unlikely((tcpconn->state==S_CONN_BAD) ||
- !(tcpconn->flags & F_CONN_HASHED) ))
- /* in the process of being destroyed => do nothing */
- break;
- if (!(tcpconn->flags & F_CONN_WANTS_WR)){
- tcpconn->flags|=F_CONN_WANTS_WR;
- t=get_ticks_raw();
- if (likely((tcpconn->flags & F_CONN_MAIN_TIMER) &&
- (TICKS_LT(tcpconn->wbuf_q.wr_timeout, tcpconn->timeout)) &&
- TICKS_LT(t, tcpconn->wbuf_q.wr_timeout) )){
- /* _wbufq_nonempty() is guaranteed here */
- /* update the timer */
- local_timer_del(&tcp_main_ltimer, &tcpconn->timer);
- local_timer_reinit(&tcpconn->timer);
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer,
- tcpconn->wbuf_q.wr_timeout-t, t);
- DBG("tcp_main: handle_ser_child: CONN_QUEUED_WRITE; %p "
- "timeout adjusted to %d s\n", tcpconn,
- TICKS_TO_S(tcpconn->wbuf_q.wr_timeout-t));
- }
- if (!(tcpconn->flags & F_CONN_WRITE_W)){
- tcpconn->flags|=F_CONN_WRITE_W;
- if (!(tcpconn->flags & F_CONN_READ_W)){
- if (unlikely(io_watch_add(&io_h, tcpconn->s, POLLOUT,
- F_TCPCONN, tcpconn)<0)){
- LM_CRIT("failed to enable write watch on socket\n");
- if (tcpconn_try_unhash(tcpconn))
- tcpconn_put_destroy(tcpconn);
- break;
- }
- }else{
- if (unlikely(io_watch_chg(&io_h, tcpconn->s,
- POLLIN|POLLOUT, -1)<0)){
- LM_CRIT("failed to change socket watch events\n");
- if (tcpconn_try_unhash(tcpconn)) {
- io_watch_del(&io_h, tcpconn->s, -1,
- IO_FD_CLOSING);
- tcpconn->flags&=~F_CONN_READ_W;
- tcpconn_put_destroy(tcpconn);
- } else {
- BUG("unhashed connection watched for IO\n");
- io_watch_del(&io_h, tcpconn->s, -1, 0);
- tcpconn->flags&=~F_CONN_READ_W;
- }
- break;
- }
- }
- }
- }else{
- LM_WARN("connection %p already watched for write\n", tcpconn);
- }
- break;
- #ifdef TCP_CONNECT_WAIT
- case CONN_NEW_COMPLETE:
- case CONN_NEW_PENDING_WRITE:
- /* received when a pending connect completes in the same
- * tcp_send() that initiated it
- * the connection is already in the hash with F_CONN_PENDING
- * flag (added by tcp_send()) and refcnt at least 1 (for the
- * hash)*/
- tcpconn->flags&=~(F_CONN_PENDING|F_CONN_FD_CLOSED);
- if (unlikely((tcpconn->state==S_CONN_BAD) || (fd==-1))){
- if (unlikely(fd==-1))
- LM_CRIT("CONN_NEW_COMPLETE: no fd received\n");
- else
- LM_WARN("CONN_NEW_COMPLETE: received connection with error\n");
- tcpconn->flags|=F_CONN_FD_CLOSED;
- tcpconn->state=S_CONN_BAD;
- tcpconn_try_unhash(tcpconn);
- tcpconn_put_destroy(tcpconn);
- break;
- }
- (*tcp_connections_no)++;
- if (unlikely(tcpconn->type==PROTO_TLS))
- (*tls_connections_no)++;
- tcpconn->s=fd;
- /* update the timeout*/
- t=get_ticks_raw();
- con_lifetime=cfg_get(tcp, tcp_cfg, con_lifetime);
- tcpconn->timeout=t+con_lifetime;
- nxt_timeout=con_lifetime;
- if (unlikely(cmd==CONN_NEW_COMPLETE)){
- /* check if needs to be watched for write */
- lock_get(&tcpconn->write_lock);
- /* if queue non empty watch it for write */
- flags=(_wbufq_empty(tcpconn)-1)&POLLOUT;
- lock_release(&tcpconn->write_lock);
- if (flags){
- if (TICKS_LT(tcpconn->wbuf_q.wr_timeout, tcpconn->timeout)
- && TICKS_LT(t, tcpconn->wbuf_q.wr_timeout))
- nxt_timeout=tcpconn->wbuf_q.wr_timeout-t;
- tcpconn->flags|=F_CONN_WRITE_W|F_CONN_WANTS_WR;
- }
- /* activate the timer (already properly init. in
- tcpconn_new()) no need for reinit */
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer, nxt_timeout,
- t);
- tcpconn->flags|=F_CONN_MAIN_TIMER|F_CONN_READ_W|
- F_CONN_WANTS_RD;
- }else{
- /* CONN_NEW_PENDING_WRITE */
- /* no need to check, we have something queued for write */
- flags=POLLOUT;
- if (TICKS_LT(tcpconn->wbuf_q.wr_timeout, tcpconn->timeout)
- && TICKS_LT(t, tcpconn->wbuf_q.wr_timeout))
- nxt_timeout=tcpconn->wbuf_q.wr_timeout-t;
- /* activate the timer (already properly init. in
- tcpconn_new()) no need for reinit */
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer, nxt_timeout,
- t);
- tcpconn->flags|=F_CONN_MAIN_TIMER|F_CONN_READ_W|
- F_CONN_WANTS_RD |
- F_CONN_WRITE_W|F_CONN_WANTS_WR;
- }
- flags|=POLLIN;
- if (unlikely(
- io_watch_add(&io_h, tcpconn->s, flags,
- F_TCPCONN, tcpconn)<0)){
- LM_CRIT("failed to add new socket to the fd list\n");
- tcpconn->flags&=~(F_CONN_WRITE_W|F_CONN_READ_W);
- tcpconn_try_unhash(tcpconn); /* unhash & dec refcnt */
- tcpconn_put_destroy(tcpconn);
- }
- break;
- #endif /* TCP_CONNECT_WAIT */
- #endif /* TCP_ASYNC */
- default:
- LM_CRIT("unknown cmd %d\n", cmd);
- }
- end:
- return ret;
- error:
- return -1;
- }
- /* sends a tcpconn + fd to a choosen child */
- inline static int send2child(struct tcp_connection* tcpconn)
- {
- int i;
- int min_busy;
- int idx;
- int wfirst;
- int wlast;
- static int crt=0; /* current child */
- int last;
-
- if(likely(tcp_sockets_gworkers==0)) {
- /* no child selection based on received socket
- * - use least loaded over all */
- min_busy=tcp_children[0].busy;
- idx=0;
- last=crt+tcp_children_no;
- for (; crt<last; crt++){
- i=crt%tcp_children_no;
- if (!tcp_children[i].busy){
- idx=i;
- min_busy=0;
- break;
- }else if (min_busy>tcp_children[i].busy){
- min_busy=tcp_children[i].busy;
- idx=i;
- }
- }
- crt=idx+1; /* next time we start with crt%tcp_children_no */
- } else {
- /* child selection based on received socket
- * - use least loaded per received socket, starting with the first
- * in its group */
- if(tcpconn->rcv.bind_address->workers>0) {
- wfirst = tcpconn->rcv.bind_address->workers_tcpidx;
- wlast = wfirst + tcpconn->rcv.bind_address->workers;
- LM_DBG("===== checking per-socket specific workers (%d/%d..%d/%d) [%s]\n",
- tcp_children[wfirst].pid, tcp_children[wfirst].proc_no,
- tcp_children[wlast-1].pid, tcp_children[wlast-1].proc_no,
- tcpconn->rcv.bind_address->sock_str.s);
- } else {
- wfirst = 0;
- wlast = tcp_sockets_gworkers - 1;
- LM_DBG("+++++ checking per-socket generic workers (%d/%d..%d/%d) [%s]\n",
- tcp_children[wfirst].pid, tcp_children[wfirst].proc_no,
- tcp_children[wlast-1].pid, tcp_children[wlast-1].proc_no,
- tcpconn->rcv.bind_address->sock_str.s);
- }
- idx = wfirst;
- min_busy = tcp_children[idx].busy;
- for(i=wfirst; i<wlast; i++) {
- if (!tcp_children[i].busy){
- idx=i;
- min_busy=0;
- break;
- } else {
- if (min_busy>tcp_children[i].busy) {
- min_busy=tcp_children[i].busy;
- idx=i;
- }
- }
- }
- }
-
- tcp_children[idx].busy++;
- tcp_children[idx].n_reqs++;
- if (unlikely(min_busy)){
- DBG("WARNING: send2child: no free tcp receiver, "
- " connection passed to the least busy one (%d)\n",
- min_busy);
- }
- LM_DBG("selected tcp worker %d %d(%ld) for activity on [%s], %p\n",
- idx, tcp_children[idx].proc_no, (long)tcp_children[idx].pid,
- tcpconn->rcv.bind_address->sock_str.s, tcpconn);
- /* first make sure this child doesn't have pending request for
- * tcp_main (to avoid a possible deadlock: e.g. child wants to
- * send a release command, but the master fills its socket buffer
- * with new connection commands => deadlock) */
- /* answer tcp_send requests first */
- while(unlikely((tcpconn->state != S_CONN_BAD) &&
- (handle_ser_child(&pt[tcp_children[idx].proc_no], -1)>0)));
- /* process tcp readers requests */
- while(unlikely((tcpconn->state != S_CONN_BAD &&
- (handle_tcp_child(&tcp_children[idx], -1)>0))));
-
- /* the above possible pending requests might have included a
- command to close this tcpconn (e.g. CONN_ERROR, CONN_EOF).
- In this case the fd is already closed here (and possible
- even replaced by another one with the same number) so it
- must not be sent to a reader anymore */
- if (unlikely(tcpconn->state == S_CONN_BAD ||
- (tcpconn->flags & F_CONN_FD_CLOSED)))
- return -1;
- #ifdef SEND_FD_QUEUE
- /* if queue full, try to queue the io */
- if (unlikely(send_fd(tcp_children[idx].unix_sock, &tcpconn,
- sizeof(tcpconn), tcpconn->s)<=0)){
- if ((errno==EAGAIN)||(errno==EWOULDBLOCK)){
- /* FIXME: remove after debugging */
- LM_CRIT("tcp child %d, socket %d: queue full, %d requests queued (total handled %d)\n",
- idx, tcp_children[idx].unix_sock, min_busy,
- tcp_children[idx].n_reqs-1);
- if (send_fd_queue_add(&send2child_q, tcp_children[idx].unix_sock,
- tcpconn)!=0){
- LM_ERR("queue send op. failed\n");
- return -1;
- }
- }else{
- LM_ERR("send_fd failed for %p (flags 0x%0x), fd %d\n",
- tcpconn, tcpconn->flags, tcpconn->s);
- return -1;
- }
- }
- #else
- if (unlikely(send_fd(tcp_children[idx].unix_sock, &tcpconn,
- sizeof(tcpconn), tcpconn->s)<=0)){
- LM_ERR("send_fd failed for %p (flags 0x%0x), fd %d\n",
- tcpconn, tcpconn->flags, tcpconn->s);
- return -1;
- }
- #endif
-
- return 0;
- }
- /* handles a new connection, called internally by tcp_main_loop/handle_io.
- * params: si - pointer to one of the tcp socket_info structures on which
- * an io event was detected (connection attempt)
- * returns: handle_* return convention: -1 on error, 0 on EAGAIN (no more
- * io events queued), >0 on success. success/error refer only to
- * the accept.
- */
- static inline int handle_new_connect(struct socket_info* si)
- {
- union sockaddr_union su;
- union sockaddr_union sock_name;
- unsigned sock_name_len;
- union sockaddr_union* dst_su;
- struct tcp_connection* tcpconn;
- socklen_t su_len;
- int new_sock;
-
- /* got a connection on r */
- su_len=sizeof(su);
- new_sock=accept(si->socket, &(su.s), &su_len);
- if (unlikely(new_sock==-1)){
- if ((errno==EAGAIN)||(errno==EWOULDBLOCK))
- return 0;
- LM_ERR("error while accepting connection(%d): %s\n", errno, strerror(errno));
- return -1;
- }
- if (unlikely(*tcp_connections_no>=cfg_get(tcp, tcp_cfg, max_connections))){
- LM_ERR("maximum number of connections exceeded: %d/%d\n",
- *tcp_connections_no,
- cfg_get(tcp, tcp_cfg, max_connections));
- tcp_safe_close(new_sock);
- TCP_STATS_LOCAL_REJECT();
- return 1; /* success, because the accept was succesfull */
- }
- if (unlikely(si->proto==PROTO_TLS)) {
- if (unlikely(*tls_connections_no>=cfg_get(tcp, tcp_cfg, max_tls_connections))){
- LM_ERR("maximum number of tls connections exceeded: %d/%d\n",
- *tls_connections_no,
- cfg_get(tcp, tcp_cfg, max_tls_connections));
- tcp_safe_close(new_sock);
- TCP_STATS_LOCAL_REJECT();
- return 1; /* success, because the accept was succesfull */
- }
- }
- if (unlikely(init_sock_opt_accept(new_sock)<0)){
- LM_ERR("init_sock_opt failed\n");
- tcp_safe_close(new_sock);
- return 1; /* success, because the accept was succesfull */
- }
- (*tcp_connections_no)++;
- if (unlikely(si->proto==PROTO_TLS))
- (*tls_connections_no)++;
- /* stats for established connections are incremented after
- the first received or sent packet.
- Alternatively they could be incremented here for accepted
- connections, but then the connection state must be changed to
- S_CONN_OK:
- TCP_STATS_ESTABLISHED(S_CONN_ACCEPT);
- ...
- tcpconn=tcpconn_new(new_sock, &su, dst_su, si, si->proto, S_CONN_OK);
- */
-
- dst_su=&si->su;
- if (unlikely(si->flags & SI_IS_ANY)){
- /* INADDR_ANY => get local dst */
- sock_name_len=sizeof(sock_name);
- if (getsockname(new_sock, &sock_name.s, &sock_name_len)!=0){
- LM_ERR("getsockname failed: %s(%d)\n",
- strerror(errno), errno);
- /* go on with the 0.0.0.0 dst from the sock_info */
- }else{
- dst_su=&sock_name;
- }
- }
- /* add socket to list */
- tcpconn=tcpconn_new(new_sock, &su, dst_su, si, si->proto, S_CONN_ACCEPT);
- if (likely(tcpconn)){
- tcpconn->flags|=F_CONN_PASSIVE;
- #ifdef TCP_PASS_NEW_CONNECTION_ON_DATA
- atomic_set(&tcpconn->refcnt, 1); /* safe, not yet available to the
- outside world */
- tcpconn_add(tcpconn);
- /* activate the timer */
- local_timer_add(&tcp_main_ltimer, &tcpconn->timer,
- cfg_get(tcp, tcp_cfg, con_lifetime),
- get_ticks_raw());
- tcpconn->flags|=(F_CONN_MAIN_TIMER|F_CONN_READ_W|F_CONN_WANTS_RD);
- if (unlikely(io_watch_add(&io_h, tcpconn->s, POLLIN,
- F_TCPCONN, tcpconn)<0)){
- LM_CRIT("failed to add new socket to the fd list\n");
- tcpconn->flags&=~F_CONN_READ_W;
- if (tcpconn_try_unhash(tcpconn))
- tcpconn_put_destroy(tcpconn);
- }
- #else
- atomic_set(&tcpconn->refcnt, 2); /* safe, not yet available to the
- outside world */
- /* prepare it for passing to a child */
- tcpconn->flags|=F_CONN_READER;
- tcpconn_add(tcpconn);
- DBG("handle_new_connect: new connection from %s: %p %d flags: %04x\n",
- su2a(&su, sizeof(su)), tcpconn, tcpconn->s, tcpconn->flags);
- if(unlikely(send2child(tcpconn)<0)){
- tcpconn->flags&=~F_CONN_READER;
- if (tcpconn_try_unhash(tcpconn))
- tcpconn_put(tcpconn);
- tcpconn_put_destroy(tcpconn);
- }
- #endif
- }else{ /*tcpconn==0 */
- LM_ERR("tcpconn_new failed, closing socket\n");
- tcp_safe_close(new_sock);
- (*tcp_connections_no)--;
- if (unlikely(si->proto==PROTO_TLS))
- (*tls_connections_no)--;
- }
- return 1; /* accept() was succesfull */
- }
- /* handles an io event on one of the watched tcp connections
- *
- * params: tcpconn - pointer to the tcp_connection for which we have an io ev.
- * fd_i - index in the fd_array table (needed for delete)
- * returns: handle_* return convention, but on success it always returns 0
- * (because it's one-shot, after a succesful execution the fd is
- * removed from tcp_main's watch fd list and passed to a child =>
- * tcp_main is not interested in further io events that might be
- * queued for this fd)
- */
- inline static int handle_tcpconn_ev(struct tcp_connection* tcpconn, short ev,
- int fd_i)
- {
- #ifdef TCP_ASYNC
- int empty_q;
- int bytes;
- #endif /* TCP_ASYNC */
- /* is refcnt!=0 really necessary?
- * No, in fact it's a bug: I can have the following situation: a send only
- * tcp connection used by n processes simultaneously => refcnt = n. In
- * the same time I can have a read event and this situation is perfectly
- * valid. -- andrei
- */
- #if 0
- if ((tcpconn->refcnt!=0)){
- /* FIXME: might be valid for sigio_rt iff fd flags are not cleared
- * (there is a short window in which it could generate a sig
- * that would be catched by tcp_main) */
- LM_CRIT("handle_tcpconn_ev: io event on referenced"
- " tcpconn (%p), refcnt=%d, fd=%d\n",
- tcpconn, tcpconn->refcnt, tcpconn->s);
- return -1;
- }
- #endif
- /* pass it to child, so remove it from the io watch list and the local
- * timer */
- #ifdef TCP_ASYNC
- empty_q=0; /* warning fix */
- if (unlikely((ev & (POLLOUT|POLLERR|POLLHUP)) &&
- (tcpconn->flags & F_CONN_WRITE_W))){
- if (unlikely((ev & (POLLERR|POLLHUP)) ||
- (wbufq_run(tcpconn->s, tcpconn, &empty_q)<0) ||
- (empty_q && tcpconn_close_after_send(tcpconn))
- )){
- if ((tcpconn->flags & F_CONN_READ_W) && (ev & POLLIN)){
- /* connection is watched for read and there is a read event
- * (unfortunately if we have POLLIN here we don't know if
- * there's really any data in the read buffer or the POLLIN
- * was generated by the error or EOF => to avoid loosing
- * data it's safer to either directly check the read buffer
- * or try a read)*/
- /* in most cases the read buffer will be empty, so in general
- * is cheaper to check it here and then send the
- * conn. to a a child only if needed (another syscall + at
- * least 2 * syscalls in the reader + ...) */
- if ((ioctl(tcpconn->s, FIONREAD, &bytes)>=0) && (bytes>0)){
- if (unlikely(io_watch_del(&io_h, tcpconn->s, fd_i, 0)<0)){
- LM_ERR("io_watch_del(1) failed: for %p, fd %d\n",
- tcpconn, tcpconn->s);
- }
- tcpconn->flags&=~(F_CONN_WRITE_W|F_CONN_READ_W|
- F_CONN_WANTS_RD|F_CONN_WANTS_WR);
- tcpconn->flags|=F_CONN_FORCE_EOF|F_CONN_WR_ERROR;
- goto send_to_child;
- }
- /* if bytes==0 or ioctl failed, destroy the connection now */
- }
- if (unlikely(io_watch_del(&io_h, tcpconn->s, fd_i,
- IO_FD_CLOSING) < 0)){
- LM_ERR("io_watch_del() failed: for %p, fd %d\n", tcpconn, tcpconn->s);
- }
- tcpconn->flags&=~(F_CONN_WRITE_W|F_CONN_READ_W|
- F_CONN_WANTS_RD|F_CONN_WANTS_WR);
- if (unlikely(ev & POLLERR)){
- if (unlikely(tcpconn->state==S_CONN_CONNECT)){
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su(BLST_ERR_CONNECT, tcpconn->rcv.proto,
- &tcpconn->rcv.src_su,
- &tcpconn->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_ERR(0, TCP_LADDR(tcpconn),
- TCP_LPORT(tcpconn), TCP_PSU(tcpconn),
- TCP_PROTO(tcpconn));
- TCP_STATS_CONNECT_FAILED();
- }else{
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su(BLST_ERR_SEND, tcpconn->rcv.proto,
- &tcpconn->rcv.src_su,
- &tcpconn->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_STATS_CON_RESET(); /* FIXME: it could != RST */
- }
- }
- if (unlikely(!tcpconn_try_unhash(tcpconn))){
- LM_CRIT("unhashed connection %p\n", tcpconn);
- }
- tcpconn_put_destroy(tcpconn);
- goto error;
- }
- if (empty_q){
- tcpconn->flags&=~F_CONN_WANTS_WR;
- if (!(tcpconn->flags & F_CONN_READ_W)){
- if (unlikely(io_watch_del(&io_h, tcpconn->s, fd_i, 0)==-1)){
- LM_ERR("io_watch_del(2) failed: for %p, fd %d\n",
- tcpconn, tcpconn->s);
- goto error;
- }
- }else{
- if (unlikely(io_watch_chg(&io_h, tcpconn->s,
- POLLIN, fd_i)==-1)){
- LM_ERR("io_watch_chg(1) failed: for %p, fd %d\n",
- tcpconn, tcpconn->s);
- goto error;
- }
- }
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- ev&=~POLLOUT; /* clear POLLOUT */
- }
- if (likely(ev && (tcpconn->flags & F_CONN_READ_W))){
- /* if still some other IO event (POLLIN|POLLHUP|POLLERR) and
- * connection is still watched in tcp_main for reads, send it to a
- * child and stop watching it for input (but continue watching for
- * writes if needed): */
- if (unlikely(tcpconn->flags & F_CONN_WRITE_W)){
- if (unlikely(io_watch_chg(&io_h, tcpconn->s, POLLOUT, fd_i)==-1)){
- LM_ERR("io_watch_chg(2) failed: for %p, fd %d\n",
- tcpconn, tcpconn->s);
- goto error;
- }
- }else
- #else
- {
- #endif /* TCP_ASYNC */
- if (unlikely(io_watch_del(&io_h, tcpconn->s, fd_i, 0)==-1)){
- LM_ERR("io_watch_del(3) failed: for %p, fd %d\n",
- tcpconn, tcpconn->s);
- goto error;
- }
- #ifdef TCP_ASYNC
- send_to_child:
- #endif
- DBG("tcp: DBG: sending to child, events %x\n", ev);
- #ifdef POLLRDHUP
- tcpconn->flags|=((int)!(ev & (POLLRDHUP|POLLHUP|POLLERR)) -1) &
- F_CONN_EOF_SEEN;
- #else /* POLLRDHUP */
- tcpconn->flags|=((int)!(ev & (POLLHUP|POLLERR)) -1) & F_CONN_EOF_SEEN;
- #endif /* POLLRDHUP */
- tcpconn->flags|= ((int)!(ev & POLLPRI) -1) & F_CONN_OOB_DATA;
- tcpconn->flags|=F_CONN_READER;
- local_timer_del(&tcp_main_ltimer, &tcpconn->timer);
- tcpconn->flags&=~(F_CONN_MAIN_TIMER|F_CONN_READ_W|F_CONN_WANTS_RD);
- tcpconn_ref(tcpconn); /* refcnt ++ */
- if (unlikely(send2child(tcpconn)<0)){
- tcpconn->flags&=~F_CONN_READER;
- #ifdef TCP_ASYNC
- if (tcpconn->flags & F_CONN_WRITE_W){
- if (unlikely(io_watch_del(&io_h, tcpconn->s, fd_i,
- IO_FD_CLOSING) < 0)){
- LM_ERR("io_watch_del(4) failed:" " for %p, fd %d\n",
- tcpconn, tcpconn->s);
- }
- tcpconn->flags&=~F_CONN_WRITE_W;
- }
- #endif /* TCP_ASYNC */
- if (tcpconn_try_unhash(tcpconn))
- tcpconn_put(tcpconn);
- tcpconn_put_destroy(tcpconn); /* because of the tcpconn_ref() */
- }
- }
- return 0; /* we are not interested in possibly queued io events,
- the fd was either passed to a child, closed, or for writes,
- everything possible was already written */
- error:
- return -1;
- }
- /* generic handle io routine, it will call the appropiate
- * handle_xxx() based on the fd_map type
- *
- * params: fm - pointer to a fd hash entry
- * idx - index in the fd_array (or -1 if not known)
- * return: -1 on error
- * 0 on EAGAIN or when by some other way it is known that no more
- * io events are queued on the fd (the receive buffer is empty).
- * Usefull to detect when there are no more io events queued for
- * sigio_rt, epoll_et, kqueue.
- * >0 on successfull read from the fd (when there might be more io
- * queued -- the receive buffer might still be non-empty)
- */
- inline static int handle_io(struct fd_map* fm, short ev, int idx)
- {
- int ret;
- /* update the local config */
- cfg_update();
-
- switch(fm->type){
- case F_SOCKINFO:
- ret=handle_new_connect((struct socket_info*)fm->data);
- break;
- case F_TCPCONN:
- ret=handle_tcpconn_ev((struct tcp_connection*)fm->data, ev, idx);
- break;
- case F_TCPCHILD:
- ret=handle_tcp_child((struct tcp_child*)fm->data, idx);
- break;
- case F_PROC:
- ret=handle_ser_child((struct process_table*)fm->data, idx);
- break;
- case F_NONE:
- LM_CRIT("empty fd map: %p {%d, %d, %p}, idx %d\n",
- fm, fm->fd, fm->type, fm->data, idx);
- goto error;
- default:
- LM_CRIT("unknown fd type %d\n", fm->type);
- goto error;
- }
- return ret;
- error:
- return -1;
- }
- /* timer handler for tcpconnection handled by tcp_main */
- static ticks_t tcpconn_main_timeout(ticks_t t, struct timer_ln* tl, void* data)
- {
- struct tcp_connection *c;
- int fd;
- int tcp_async;
-
- c=(struct tcp_connection*)data;
- /* or (struct tcp...*)(tl-offset(c->timer)) */
-
- #ifdef TCP_ASYNC
- DBG( "tcp_main: entering timer for %p (ticks=%d, timeout=%d (%d s), "
- "wr_timeout=%d (%d s)), write queue: %d bytes\n",
- c, t, c->timeout, TICKS_TO_S(c->timeout-t),
- c->wbuf_q.wr_timeout, TICKS_TO_S(c->wbuf_q.wr_timeout-t),
- c->wbuf_q.queued);
-
- tcp_async=cfg_get(tcp, tcp_cfg, async);
- if (likely(TICKS_LT(t, c->timeout) && ( !tcp_async || _wbufq_empty(c) ||
- TICKS_LT(t, c->wbuf_q.wr_timeout)) )){
- if (unlikely(tcp_async && _wbufq_non_empty(c)))
- return (ticks_t)MIN_unsigned(c->timeout-t, c->wbuf_q.wr_timeout-t);
- else
- return (ticks_t)(c->timeout - t);
- }
- /* if time out due to write, add it to the blacklist */
- if (tcp_async && _wbufq_non_empty(c) && TICKS_GE(t, c->wbuf_q.wr_timeout)){
- if (unlikely(c->state==S_CONN_CONNECT)){
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su(BLST_ERR_CONNECT, c->rcv.proto, &c->rcv.src_su,
- &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_CONNECT_TIMEOUT(0, TCP_LADDR(c), TCP_LPORT(c), TCP_PSU(c),
- TCP_PROTO(c));
- TCP_STATS_CONNECT_FAILED();
- }else{
- #ifdef USE_DST_BLACKLIST
- (void)dst_blacklist_su(BLST_ERR_SEND, c->rcv.proto, &c->rcv.src_su,
- &c->send_flags, 0);
- #endif /* USE_DST_BLACKLIST */
- TCP_EV_SEND_TIMEOUT(0, &c->rcv);
- TCP_STATS_SEND_TIMEOUT();
- }
- }else{
- /* idle timeout */
- TCP_EV_IDLE_CONN_CLOSED(0, &c->rcv);
- TCP_STATS_CON_TIMEOUT();
- }
- #else /* ! TCP_ASYNC */
- if (TICKS_LT(t, c->timeout)){
- /* timeout extended, exit */
- return (ticks_t)(c->timeout - t);
- }
- /* idle timeout */
- TCP_EV_IDLE_CONN_CLOSED(0, &c->rcv);
- TCP_STATS_CON_TIMEOUT();
- #endif /* TCP_ASYNC */
- DBG("tcp_main: timeout for %p\n", c);
- if (likely(c->flags & F_CONN_HASHED)){
- c->flags&=~(F_CONN_HASHED|F_CONN_MAIN_TIMER);
- c->state=S_CONN_BAD;
- TCPCONN_LOCK;
- _tcpconn_detach(c);
- TCPCONN_UNLOCK;
- }else{
- c->flags&=~F_CONN_MAIN_TIMER;
- LM_CRIT("timer: called with unhashed connection %p\n", c);
- tcpconn_ref(c); /* ugly hack to try to go on */
- }
- fd=c->s;
- if (likely(fd>0)){
- if (likely(c->flags & (F_CONN_READ_W|F_CONN_WRITE_W))){
- io_watch_del(&io_h, fd, -1, IO_FD_CLOSING);
- c->flags&=~(F_CONN_READ_W|F_CONN_WRITE_W);
- }
- }
- tcpconn_put_destroy(c);
- return 0;
- }
- static inline void tcp_timer_run(void)
- {
- ticks_t ticks;
-
- ticks=get_ticks_raw();
- if (unlikely((ticks-tcp_main_prev_ticks)<TCPCONN_TIMEOUT_MIN_RUN)) return;
- tcp_main_prev_ticks=ticks;
- local_timer_run(&tcp_main_ltimer, ticks);
- }
- /* keep in sync with tcpconn_destroy, the "delete" part should be
- * the same except for io_watch_del..
- * Note: this function is called only on shutdown by the main ser process via
- * cleanup(). However it's also safe to call it from the tcp_main process.
- * => with the ser shutdown exception, it cannot execute in parallel
- * with tcpconn_add() or tcpconn_destroy()*/
- static inline void tcpconn_destroy_all(void)
- {
- struct tcp_connection *c, *next;
- unsigned h;
- int fd;
-
-
- TCPCONN_LOCK;
- for(h=0; h<TCP_ID_HASH_SIZE; h++){
- c=tcpconn_id_hash[h];
- while(c){
- next=c->id_next;
- if (is_tcp_main){
- /* we cannot close or remove the fd if we are not in the
- * tcp main proc.*/
- if ((c->flags & F_CONN_MAIN_TIMER)){
- local_timer_del(&tcp_main_ltimer, &c->timer);
- c->flags&=~F_CONN_MAIN_TIMER;
- } /* else still in some reader */
- fd=c->s;
- if (fd>0 && (c->flags & (F_CONN_READ_W|F_CONN_WRITE_W))){
- io_watch_del(&io_h, fd, -1, IO_FD_CLOSING);
- c->flags&=~(F_CONN_READ_W|F_CONN_WRITE_W);
- }
- }else{
- fd=-1;
- }
- #ifdef USE_TLS
- if (fd>0 && (c->type==PROTO_TLS || c->type==PROTO_WSS))
- tls_close(c, fd);
- if (unlikely(c->type==PROTO_TLS || c->type==PROTO_WSS))
- (*tls_connections_no)--;
- #endif
- (*tcp_connections_no)--;
- c->flags &= ~F_CONN_HASHED;
- _tcpconn_rm(c);
- if (fd>0) {
- #ifdef TCP_FD_CACHE
- if (likely(cfg_get(tcp, tcp_cfg, fd_cache)))
- shutdown(fd, SHUT_RDWR);
- #endif /* TCP_FD_CACHE */
- tcp_safe_close(fd);
- }
- c=next;
- }
- }
- TCPCONN_UNLOCK;
- }
- /* tcp main loop */
- void tcp_main_loop()
- {
- struct socket_info* si;
- int r;
-
- is_tcp_main=1; /* mark this process as tcp main */
-
- tcp_main_max_fd_no=get_max_open_fds();
- /* init send fd queues (here because we want mem. alloc only in the tcp
- * process */
- #ifdef SEND_FD_QUEUE
- if (init_send_fd_queues()<0){
- LM_CRIT("could not init send fd queues\n");
- goto error;
- }
- #endif
- /* init io_wait (here because we want the memory allocated only in
- * the tcp_main process) */
- if (init_io_wait(&io_h, tcp_main_max_fd_no, tcp_poll_method)<0)
- goto error;
- /* init: start watching all the fds*/
-
- /* init local timer */
- tcp_main_prev_ticks=get_ticks_raw();
- if (init_local_timer(&tcp_main_ltimer, get_ticks_raw())!=0){
- LM_ERR("failed to init local timer\n");
- goto error;
- }
- #ifdef TCP_FD_CACHE
- if (cfg_get(tcp, tcp_cfg, fd_cache)) tcp_fd_cache_init();
- #endif /* TCP_FD_CACHE */
-
- /* add all the sockets we listen on for connections */
- for (si=tcp_listen; si; si=si->next){
- if ((si->proto==PROTO_TCP) &&(si->socket!=-1)){
- if (io_watch_add(&io_h, si->socket, POLLIN, F_SOCKINFO, si)<0){
- LM_CRIT("failed to add listen socket to the fd list\n");
- goto error;
- }
- }else{
- LM_CRIT("non tcp address in tcp_listen\n");
- }
- }
- #ifdef USE_TLS
- if (!tls_disable && tls_loaded()){
- for (si=tls_listen; si; si=si->next){
- if ((si->proto==PROTO_TLS) && (si->socket!=-1)){
- if (io_watch_add(&io_h, si->socket, POLLIN, F_SOCKINFO, si)<0){
- LM_CRIT("failed to add tls listen socket to the fd list\n");
- goto error;
- }
- }else{
- LM_CRIT("non tls address in tls_listen\n");
- }
- }
- }
- #endif
- /* add all the unix sockets used for communcation with other ser processes
- * (get fd, new connection a.s.o) */
- for (r=1; r<process_no; r++){
- if (pt[r].unix_sock>0) /* we can't have 0, we never close it!*/
- if (io_watch_add(&io_h, pt[r].unix_sock, POLLIN,F_PROC, &pt[r])<0){
- LM_CRIT("failed to add process %d unix socket to the fd list\n", r);
- goto error;
- }
- }
- /* add all the unix sokets used for communication with the tcp childs */
- for (r=0; r<tcp_children_no; r++){
- if (tcp_children[r].unix_sock>0)/*we can't have 0, we never close it!*/
- if (io_watch_add(&io_h, tcp_children[r].unix_sock, POLLIN,
- F_TCPCHILD, &tcp_children[r]) <0){
- LM_CRIT("failed to add tcp child %d unix socket to the fd list\n", r);
- goto error;
- }
- }
- /* initialize the cfg framework */
- if (cfg_child_init()) goto error;
- /* main loop */
- switch(io_h.poll_method){
- case POLL_POLL:
- while(1){
- /* wait and process IO */
- io_wait_loop_poll(&io_h, TCP_MAIN_SELECT_TIMEOUT, 0);
- send_fd_queue_run(&send2child_q); /* then new io */
- /* remove old connections */
- tcp_timer_run();
- }
- break;
- #ifdef HAVE_SELECT
- case POLL_SELECT:
- while(1){
- io_wait_loop_select(&io_h, TCP_MAIN_SELECT_TIMEOUT, 0);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- #endif
- #ifdef HAVE_SIGIO_RT
- case POLL_SIGIO_RT:
- while(1){
- io_wait_loop_sigio_rt(&io_h, TCP_MAIN_SELECT_TIMEOUT);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- #endif
- #ifdef HAVE_EPOLL
- case POLL_EPOLL_LT:
- while(1){
- io_wait_loop_epoll(&io_h, TCP_MAIN_SELECT_TIMEOUT, 0);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- case POLL_EPOLL_ET:
- while(1){
- io_wait_loop_epoll(&io_h, TCP_MAIN_SELECT_TIMEOUT, 1);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- #endif
- #ifdef HAVE_KQUEUE
- case POLL_KQUEUE:
- while(1){
- io_wait_loop_kqueue(&io_h, TCP_MAIN_SELECT_TIMEOUT, 0);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- #endif
- #ifdef HAVE_DEVPOLL
- case POLL_DEVPOLL:
- while(1){
- io_wait_loop_devpoll(&io_h, TCP_MAIN_SELECT_TIMEOUT, 0);
- send_fd_queue_run(&send2child_q); /* then new io */
- tcp_timer_run();
- }
- break;
- #endif
- default:
- LM_CRIT("no support for poll method %s (%d)\n",
- poll_method_name(io_h.poll_method), io_h.poll_method);
- goto error;
- }
- error:
- #ifdef SEND_FD_QUEUE
- destroy_send_fd_queues();
- #endif
- destroy_io_wait(&io_h);
- LM_CRIT("exiting...");
- exit(-1);
- }
- /* cleanup before exit */
- void destroy_tcp()
- {
- if (tcpconn_id_hash){
- if (tcpconn_lock)
- TCPCONN_UNLOCK; /* hack: force-unlock the tcp lock in case
- some process was terminated while holding
- it; this will allow an almost gracious
- shutdown */
- tcpconn_destroy_all();
- shm_free(tcpconn_id_hash);
- tcpconn_id_hash=0;
- }
- DESTROY_TCP_STATS();
- if (tcp_connections_no){
- shm_free(tcp_connections_no);
- tcp_connections_no=0;
- }
- if (tls_connections_no){
- shm_free(tls_connections_no);
- tls_connections_no=0;
- }
- #ifdef TCP_ASYNC
- if (tcp_total_wq){
- shm_free(tcp_total_wq);
- tcp_total_wq=0;
- }
- #endif /* TCP_ASYNC */
- if (connection_id){
- shm_free(connection_id);
- connection_id=0;
- }
- if (tcpconn_aliases_hash){
- shm_free(tcpconn_aliases_hash);
- tcpconn_aliases_hash=0;
- }
- if (tcpconn_lock){
- lock_destroy(tcpconn_lock);
- lock_dealloc((void*)tcpconn_lock);
- tcpconn_lock=0;
- }
- if (tcp_children){
- pkg_free(tcp_children);
- tcp_children=0;
- }
- destroy_local_timer(&tcp_main_ltimer);
- }
- int init_tcp()
- {
- char* poll_err;
-
- tcp_options_check();
- if (tcp_cfg==0){
- BUG("tcp_cfg not initialized\n");
- goto error;
- }
- /* init lock */
- tcpconn_lock=lock_alloc();
- if (tcpconn_lock==0){
- LM_CRIT("could not alloc lock\n");
- goto error;
- }
- if (lock_init(tcpconn_lock)==0){
- LM_CRIT("could not init lock\n");
- lock_dealloc((void*)tcpconn_lock);
- tcpconn_lock=0;
- goto error;
- }
- /* init globals */
- tcp_connections_no=shm_malloc(sizeof(int));
- if (tcp_connections_no==0){
- LM_CRIT("could not alloc globals\n");
- goto error;
- }
- *tcp_connections_no=0;
- tls_connections_no=shm_malloc(sizeof(int));
- if (tls_connections_no==0){
- LM_CRIT("could not alloc globals\n");
- goto error;
- }
- *tls_connections_no=0;
- if (INIT_TCP_STATS()!=0) goto error;
- connection_id=shm_malloc(sizeof(int));
- if (connection_id==0){
- LM_CRIT("could not alloc globals\n");
- goto error;
- }
- *connection_id=1;
- #ifdef TCP_ASYNC
- tcp_total_wq=shm_malloc(sizeof(*tcp_total_wq));
- if (tcp_total_wq==0){
- LM_CRIT("could not alloc globals\n");
- goto error;
- }
- #endif /* TCP_ASYNC */
- /* alloc hashtables*/
- tcpconn_aliases_hash=(struct tcp_conn_alias**)
- shm_malloc(TCP_ALIAS_HASH_SIZE* sizeof(struct tcp_conn_alias*));
- if (tcpconn_aliases_hash==0){
- LM_CRIT("could not alloc address hashtable\n");
- goto error;
- }
- tcpconn_id_hash=(struct tcp_connection**)shm_malloc(TCP_ID_HASH_SIZE*
- sizeof(struct tcp_connection*));
- if (tcpconn_id_hash==0){
- LM_CRIT("could not alloc id hashtable\n");
- goto error;
- }
- /* init hashtables*/
- memset((void*)tcpconn_aliases_hash, 0,
- TCP_ALIAS_HASH_SIZE * sizeof(struct tcp_conn_alias*));
- memset((void*)tcpconn_id_hash, 0,
- TCP_ID_HASH_SIZE * sizeof(struct tcp_connection*));
-
- /* fix config variables */
- poll_err=check_poll_method(tcp_poll_method);
-
- /* set an appropriate poll method */
- if (poll_err || (tcp_poll_method==0)){
- tcp_poll_method=choose_poll_method();
- if (poll_err){
- LM_ERR("%s, using %s instead\n",
- poll_err, poll_method_name(tcp_poll_method));
- }else{
- LM_INFO("using %s as the io watch method (auto detected)\n",
- poll_method_name(tcp_poll_method));
- }
- }else{
- LM_INFO("using %s io watch method (config)\n",
- poll_method_name(tcp_poll_method));
- }
-
- return 0;
- error:
- /* clean-up */
- destroy_tcp();
- return -1;
- }
- #ifdef TCP_CHILD_NON_BLOCKING
- /* returns -1 on error */
- static int set_non_blocking(int s)
- {
- int flags;
- /* non-blocking */
- flags=fcntl(s, F_GETFL);
- if (flags==-1){
- LM_ERR("fnctl failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- if (fcntl(s, F_SETFL, flags|O_NONBLOCK)==-1){
- LM_ERR("fcntl: set non-blocking failed: (%d) %s\n", errno, strerror(errno));
- goto error;
- }
- return 0;
- error:
- return -1;
- }
- #endif
- /* returns -1 on error, 0 on success */
- int tcp_fix_child_sockets(int* fd)
- {
- #ifdef TCP_CHILD_NON_BLOCKING
- if ((set_non_blocking(fd[0])<0) ||
- (set_non_blocking(fd[1])<0)){
- return -1;
- }
- #endif
- return 0;
- }
- /* starts the tcp processes */
- int tcp_init_children()
- {
- int r, i;
- int reader_fd_1; /* for comm. with the tcp children read */
- pid_t pid;
- char si_desc[MAX_PT_DESC];
- struct socket_info *si;
-
- /* estimate max fd. no:
- * 1 tcp send unix socket/all_proc,
- * + 1 udp sock/udp proc + 1 tcp_child sock/tcp child*
- * + no_listen_tcp */
- for(r=0, si=tcp_listen; si; si=si->next, r++);
- #ifdef USE_TLS
- if (! tls_disable)
- for (si=tls_listen; si; si=si->next, r++);
- #endif
-
- register_fds(r+tcp_max_connections+get_max_procs()-1 /* tcp main */);
- #if 0
- tcp_max_fd_no=get_max_procs()*2 +r-1 /* timer */ +3; /* stdin/out/err*/
- /* max connections can be temporarily exceeded with estimated_process_count
- * - tcp_main (tcpconn_connect called simultaneously in all all the
- * processes) */
- tcp_max_fd_no+=tcp_max_connections+get_max_procs()-1 /* tcp main */;
- #endif
- /* alloc the children array */
- tcp_children=pkg_malloc(sizeof(struct tcp_child)*tcp_children_no);
- if (tcp_children==0){
- LM_ERR("out of memory\n");
- goto error;
- }
- memset(tcp_children, 0, sizeof(struct tcp_child)*tcp_children_no);
- /* assign own socket for tcp workers, if it is the case
- * - add them from end to start of tcp children array
- * - thus, have generic tcp workers at beginning */
- i = tcp_children_no-1;
- for(si=tcp_listen; si; si=si->next) {
- if(si->workers>0) {
- si->workers_tcpidx = i - si->workers + 1;
- for(r=0; r<si->workers; r++) {
- tcp_children[i].mysocket = si;
- i--;
- }
- }
- }
- #ifdef USE_TLS
- for(si=tls_listen; si; si=si->next) {
- if(si->workers>0) {
- si->workers_tcpidx = i - si->workers + 1;
- for(r=0; r<si->workers; r++) {
- tcp_children[i].mysocket = si;
- i--;
- }
- }
- }
- #endif
- tcp_sockets_gworkers = (i != tcp_children_no-1)?(1 + i + 1):0;
- /* create the tcp sock_info structures */
- /* copy the sockets --moved to main_loop*/
-
- /* fork children & create the socket pairs*/
- for(r=0; r<tcp_children_no; r++){
- child_rank++;
- snprintf(si_desc, MAX_PT_DESC, "tcp receiver (%s)",
- (tcp_children[r].mysocket!=NULL)?
- tcp_children[r].mysocket->sock_str.s:"generic");
- pid=fork_tcp_process(child_rank, si_desc, r, &reader_fd_1);
- if (pid<0){
- LM_ERR("fork failed: %s\n", strerror(errno));
- goto error;
- }else if (pid>0){
- /* parent */
- }else{
- /* child */
- bind_address=0; /* force a SEGFAULT if someone uses a non-init.
- bind address on tcp */
- tcp_receive_loop(reader_fd_1);
- }
- }
- return 0;
- error:
- return -1;
- }
- void tcp_get_info(struct tcp_gen_info *ti)
- {
- ti->tcp_readers=tcp_children_no;
- ti->tcp_max_connections=tcp_max_connections;
- ti->tls_max_connections=tls_max_connections;
- ti->tcp_connections_no=*tcp_connections_no;
- ti->tls_connections_no=*tls_connections_no;
- #ifdef TCP_ASYNC
- ti->tcp_write_queued=*tcp_total_wq;
- #else
- ti->tcp_write_queued=0;
- #endif /* TCP_ASYNC */
- }
- #endif
|