123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930 |
- package nebula
- import (
- "context"
- "encoding/binary"
- "errors"
- "fmt"
- "net"
- "sync"
- "sync/atomic"
- "time"
- "unsafe"
- "github.com/rcrowley/go-metrics"
- "github.com/sirupsen/logrus"
- "github.com/slackhq/nebula/config"
- "github.com/slackhq/nebula/header"
- "github.com/slackhq/nebula/iputil"
- "github.com/slackhq/nebula/udp"
- "github.com/slackhq/nebula/util"
- )
- //TODO: if a lighthouse doesn't have an answer, clients AGGRESSIVELY REQUERY.. why? handshake manager and/or getOrHandshake?
- //TODO: nodes are roaming lighthouses, this is bad. How are they learning?
- var ErrHostNotKnown = errors.New("host not known")
- type netIpAndPort struct {
- ip net.IP
- port uint16
- }
- type LightHouse struct {
- //TODO: We need a timer wheel to kick out vpnIps that haven't reported in a long time
- sync.RWMutex //Because we concurrently read and write to our maps
- amLighthouse bool
- myVpnIp iputil.VpnIp
- myVpnZeros iputil.VpnIp
- myVpnNet *net.IPNet
- punchConn *udp.Conn
- punchy *Punchy
- // Local cache of answers from light houses
- // map of vpn Ip to answers
- addrMap map[iputil.VpnIp]*RemoteList
- // filters remote addresses allowed for each host
- // - When we are a lighthouse, this filters what addresses we store and
- // respond with.
- // - When we are not a lighthouse, this filters which addresses we accept
- // from lighthouses.
- atomicRemoteAllowList *RemoteAllowList
- // filters local addresses that we advertise to lighthouses
- atomicLocalAllowList *LocalAllowList
- // used to trigger the HandshakeManager when we receive HostQueryReply
- handshakeTrigger chan<- iputil.VpnIp
- // atomicStaticList exists to avoid having a bool in each addrMap entry
- // since static should be rare
- atomicStaticList map[iputil.VpnIp]struct{}
- atomicLighthouses map[iputil.VpnIp]struct{}
- atomicInterval int64
- updateCancel context.CancelFunc
- updateParentCtx context.Context
- updateUdp udp.EncWriter
- nebulaPort uint32 // 32 bits because protobuf does not have a uint16
- atomicAdvertiseAddrs []netIpAndPort
- // IP's of relays that can be used by peers to access me
- atomicRelaysForMe []iputil.VpnIp
- metrics *MessageMetrics
- metricHolepunchTx metrics.Counter
- l *logrus.Logger
- }
- // NewLightHouseFromConfig will build a Lighthouse struct from the values provided in the config object
- // addrMap should be nil unless this is during a config reload
- func NewLightHouseFromConfig(l *logrus.Logger, c *config.C, myVpnNet *net.IPNet, pc *udp.Conn, p *Punchy) (*LightHouse, error) {
- amLighthouse := c.GetBool("lighthouse.am_lighthouse", false)
- nebulaPort := uint32(c.GetInt("listen.port", 0))
- if amLighthouse && nebulaPort == 0 {
- return nil, util.NewContextualError("lighthouse.am_lighthouse enabled on node but no port number is set in config", nil, nil)
- }
- // If port is dynamic, discover it
- if nebulaPort == 0 && pc != nil {
- uPort, err := pc.LocalAddr()
- if err != nil {
- return nil, util.NewContextualError("Failed to get listening port", nil, err)
- }
- nebulaPort = uint32(uPort.Port)
- }
- ones, _ := myVpnNet.Mask.Size()
- h := LightHouse{
- amLighthouse: amLighthouse,
- myVpnIp: iputil.Ip2VpnIp(myVpnNet.IP),
- myVpnZeros: iputil.VpnIp(32 - ones),
- myVpnNet: myVpnNet,
- addrMap: make(map[iputil.VpnIp]*RemoteList),
- nebulaPort: nebulaPort,
- atomicLighthouses: make(map[iputil.VpnIp]struct{}),
- atomicStaticList: make(map[iputil.VpnIp]struct{}),
- punchConn: pc,
- punchy: p,
- l: l,
- }
- if c.GetBool("stats.lighthouse_metrics", false) {
- h.metrics = newLighthouseMetrics()
- h.metricHolepunchTx = metrics.GetOrRegisterCounter("messages.tx.holepunch", nil)
- } else {
- h.metricHolepunchTx = metrics.NilCounter{}
- }
- err := h.reload(c, true)
- if err != nil {
- return nil, err
- }
- c.RegisterReloadCallback(func(c *config.C) {
- err := h.reload(c, false)
- switch v := err.(type) {
- case util.ContextualError:
- v.Log(l)
- case error:
- l.WithError(err).Error("failed to reload lighthouse")
- }
- })
- return &h, nil
- }
- func (lh *LightHouse) GetStaticHostList() map[iputil.VpnIp]struct{} {
- return *(*map[iputil.VpnIp]struct{})(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicStaticList))))
- }
- func (lh *LightHouse) GetLighthouses() map[iputil.VpnIp]struct{} {
- return *(*map[iputil.VpnIp]struct{})(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLighthouses))))
- }
- func (lh *LightHouse) GetRemoteAllowList() *RemoteAllowList {
- return (*RemoteAllowList)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRemoteAllowList))))
- }
- func (lh *LightHouse) GetLocalAllowList() *LocalAllowList {
- return (*LocalAllowList)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLocalAllowList))))
- }
- func (lh *LightHouse) GetAdvertiseAddrs() []netIpAndPort {
- return *(*[]netIpAndPort)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicAdvertiseAddrs))))
- }
- func (lh *LightHouse) GetRelaysForMe() []iputil.VpnIp {
- return *(*[]iputil.VpnIp)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRelaysForMe))))
- }
- func (lh *LightHouse) GetUpdateInterval() int64 {
- return atomic.LoadInt64(&lh.atomicInterval)
- }
- func (lh *LightHouse) reload(c *config.C, initial bool) error {
- if initial || c.HasChanged("lighthouse.advertise_addrs") {
- rawAdvAddrs := c.GetStringSlice("lighthouse.advertise_addrs", []string{})
- advAddrs := make([]netIpAndPort, 0)
- for i, rawAddr := range rawAdvAddrs {
- fIp, fPort, err := udp.ParseIPAndPort(rawAddr)
- if err != nil {
- return util.NewContextualError("Unable to parse lighthouse.advertise_addrs entry", m{"addr": rawAddr, "entry": i + 1}, err)
- }
- if fPort == 0 {
- fPort = uint16(lh.nebulaPort)
- }
- if ip4 := fIp.To4(); ip4 != nil && lh.myVpnNet.Contains(fIp) {
- lh.l.WithField("addr", rawAddr).WithField("entry", i+1).
- Warn("Ignoring lighthouse.advertise_addrs report because it is within the nebula network range")
- continue
- }
- advAddrs = append(advAddrs, netIpAndPort{ip: fIp, port: fPort})
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicAdvertiseAddrs)), unsafe.Pointer(&advAddrs))
- if !initial {
- lh.l.Info("lighthouse.advertise_addrs has changed")
- }
- }
- if initial || c.HasChanged("lighthouse.interval") {
- atomic.StoreInt64(&lh.atomicInterval, int64(c.GetInt("lighthouse.interval", 10)))
- if !initial {
- lh.l.Infof("lighthouse.interval changed to %v", lh.atomicInterval)
- if lh.updateCancel != nil {
- // May not always have a running routine
- lh.updateCancel()
- }
- lh.LhUpdateWorker(lh.updateParentCtx, lh.updateUdp)
- }
- }
- if initial || c.HasChanged("lighthouse.remote_allow_list") || c.HasChanged("lighthouse.remote_allow_ranges") {
- ral, err := NewRemoteAllowListFromConfig(c, "lighthouse.remote_allow_list", "lighthouse.remote_allow_ranges")
- if err != nil {
- return util.NewContextualError("Invalid lighthouse.remote_allow_list", nil, err)
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRemoteAllowList)), unsafe.Pointer(ral))
- if !initial {
- //TODO: a diff will be annoyingly difficult
- lh.l.Info("lighthouse.remote_allow_list and/or lighthouse.remote_allow_ranges has changed")
- }
- }
- if initial || c.HasChanged("lighthouse.local_allow_list") {
- lal, err := NewLocalAllowListFromConfig(c, "lighthouse.local_allow_list")
- if err != nil {
- return util.NewContextualError("Invalid lighthouse.local_allow_list", nil, err)
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLocalAllowList)), unsafe.Pointer(lal))
- if !initial {
- //TODO: a diff will be annoyingly difficult
- lh.l.Info("lighthouse.local_allow_list has changed")
- }
- }
- //NOTE: many things will get much simpler when we combine static_host_map and lighthouse.hosts in config
- if initial || c.HasChanged("static_host_map") {
- staticList := make(map[iputil.VpnIp]struct{})
- err := lh.loadStaticMap(c, lh.myVpnNet, staticList)
- if err != nil {
- return err
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicStaticList)), unsafe.Pointer(&staticList))
- if !initial {
- //TODO: we should remove any remote list entries for static hosts that were removed/modified?
- lh.l.Info("static_host_map has changed")
- }
- }
- if initial || c.HasChanged("lighthouse.hosts") {
- lhMap := make(map[iputil.VpnIp]struct{})
- err := lh.parseLighthouses(c, lh.myVpnNet, lhMap)
- if err != nil {
- return err
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLighthouses)), unsafe.Pointer(&lhMap))
- if !initial {
- //NOTE: we are not tearing down existing lighthouse connections because they might be used for non lighthouse traffic
- lh.l.Info("lighthouse.hosts has changed")
- }
- }
- if initial || c.HasChanged("relay.relays") {
- switch c.GetBool("relay.am_relay", false) {
- case true:
- // Relays aren't allowed to specify other relays
- if len(c.GetStringSlice("relay.relays", nil)) > 0 {
- lh.l.Info("Ignoring relays from config because am_relay is true")
- }
- relaysForMe := []iputil.VpnIp{}
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRelaysForMe)), unsafe.Pointer(&relaysForMe))
- case false:
- relaysForMe := []iputil.VpnIp{}
- for _, v := range c.GetStringSlice("relay.relays", nil) {
- lh.l.WithField("RelayIP", v).Info("Read relay from config")
- configRIP := net.ParseIP(v)
- if configRIP != nil {
- relaysForMe = append(relaysForMe, iputil.Ip2VpnIp(configRIP))
- }
- }
- atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRelaysForMe)), unsafe.Pointer(&relaysForMe))
- }
- }
- return nil
- }
- func (lh *LightHouse) parseLighthouses(c *config.C, tunCidr *net.IPNet, lhMap map[iputil.VpnIp]struct{}) error {
- lhs := c.GetStringSlice("lighthouse.hosts", []string{})
- if lh.amLighthouse && len(lhs) != 0 {
- lh.l.Warn("lighthouse.am_lighthouse enabled on node but upstream lighthouses exist in config")
- }
- for i, host := range lhs {
- ip := net.ParseIP(host)
- if ip == nil {
- return util.NewContextualError("Unable to parse lighthouse host entry", m{"host": host, "entry": i + 1}, nil)
- }
- if !tunCidr.Contains(ip) {
- return util.NewContextualError("lighthouse host is not in our subnet, invalid", m{"vpnIp": ip, "network": tunCidr.String()}, nil)
- }
- lhMap[iputil.Ip2VpnIp(ip)] = struct{}{}
- }
- if !lh.amLighthouse && len(lhMap) == 0 {
- lh.l.Warn("No lighthouse.hosts configured, this host will only be able to initiate tunnels with static_host_map entries")
- }
- staticList := lh.GetStaticHostList()
- for lhIP, _ := range lhMap {
- if _, ok := staticList[lhIP]; !ok {
- return fmt.Errorf("lighthouse %s does not have a static_host_map entry", lhIP)
- }
- }
- return nil
- }
- func (lh *LightHouse) loadStaticMap(c *config.C, tunCidr *net.IPNet, staticList map[iputil.VpnIp]struct{}) error {
- shm := c.GetMap("static_host_map", map[interface{}]interface{}{})
- i := 0
- for k, v := range shm {
- rip := net.ParseIP(fmt.Sprintf("%v", k))
- if rip == nil {
- return util.NewContextualError("Unable to parse static_host_map entry", m{"host": k, "entry": i + 1}, nil)
- }
- if !tunCidr.Contains(rip) {
- return util.NewContextualError("static_host_map key is not in our subnet, invalid", m{"vpnIp": rip, "network": tunCidr.String(), "entry": i + 1}, nil)
- }
- vpnIp := iputil.Ip2VpnIp(rip)
- vals, ok := v.([]interface{})
- if ok {
- for _, v := range vals {
- ip, port, err := udp.ParseIPAndPort(fmt.Sprintf("%v", v))
- if err != nil {
- return util.NewContextualError("Static host address could not be parsed", m{"vpnIp": vpnIp, "entry": i + 1}, err)
- }
- lh.addStaticRemote(vpnIp, udp.NewAddr(ip, port), staticList)
- }
- } else {
- ip, port, err := udp.ParseIPAndPort(fmt.Sprintf("%v", v))
- if err != nil {
- return util.NewContextualError("Static host address could not be parsed", m{"vpnIp": vpnIp, "entry": i + 1}, err)
- }
- lh.addStaticRemote(vpnIp, udp.NewAddr(ip, port), staticList)
- }
- i++
- }
- return nil
- }
- func (lh *LightHouse) Query(ip iputil.VpnIp, f udp.EncWriter) *RemoteList {
- if !lh.IsLighthouseIP(ip) {
- lh.QueryServer(ip, f)
- }
- lh.RLock()
- if v, ok := lh.addrMap[ip]; ok {
- lh.RUnlock()
- return v
- }
- lh.RUnlock()
- return nil
- }
- // This is asynchronous so no reply should be expected
- func (lh *LightHouse) QueryServer(ip iputil.VpnIp, f udp.EncWriter) {
- if lh.amLighthouse {
- return
- }
- if lh.IsLighthouseIP(ip) {
- return
- }
- // Send a query to the lighthouses and hope for the best next time
- query, err := NewLhQueryByInt(ip).Marshal()
- if err != nil {
- lh.l.WithError(err).WithField("vpnIp", ip).Error("Failed to marshal lighthouse query payload")
- return
- }
- lighthouses := lh.GetLighthouses()
- lh.metricTx(NebulaMeta_HostQuery, int64(len(lighthouses)))
- nb := make([]byte, 12, 12)
- out := make([]byte, mtu)
- for n := range lighthouses {
- f.SendMessageToVpnIp(header.LightHouse, 0, n, query, nb, out)
- }
- }
- func (lh *LightHouse) QueryCache(ip iputil.VpnIp) *RemoteList {
- lh.RLock()
- if v, ok := lh.addrMap[ip]; ok {
- lh.RUnlock()
- return v
- }
- lh.RUnlock()
- lh.Lock()
- defer lh.Unlock()
- // Add an entry if we don't already have one
- return lh.unlockedGetRemoteList(ip)
- }
- // queryAndPrepMessage is a lock helper on RemoteList, assisting the caller to build a lighthouse message containing
- // details from the remote list. It looks for a hit in the addrMap and a hit in the RemoteList under the owner vpnIp
- // If one is found then f() is called with proper locking, f() must return result of n.MarshalTo()
- func (lh *LightHouse) queryAndPrepMessage(vpnIp iputil.VpnIp, f func(*cache) (int, error)) (bool, int, error) {
- lh.RLock()
- // Do we have an entry in the main cache?
- if v, ok := lh.addrMap[vpnIp]; ok {
- // Swap lh lock for remote list lock
- v.RLock()
- defer v.RUnlock()
- lh.RUnlock()
- // vpnIp should also be the owner here since we are a lighthouse.
- c := v.cache[vpnIp]
- // Make sure we have
- if c != nil {
- n, err := f(c)
- return true, n, err
- }
- return false, 0, nil
- }
- lh.RUnlock()
- return false, 0, nil
- }
- func (lh *LightHouse) DeleteVpnIp(vpnIp iputil.VpnIp) {
- // First we check the static mapping
- // and do nothing if it is there
- if _, ok := lh.GetStaticHostList()[vpnIp]; ok {
- return
- }
- lh.Lock()
- //l.Debugln(lh.addrMap)
- delete(lh.addrMap, vpnIp)
- if lh.l.Level >= logrus.DebugLevel {
- lh.l.Debugf("deleting %s from lighthouse.", vpnIp)
- }
- lh.Unlock()
- }
- // AddStaticRemote adds a static host entry for vpnIp as ourselves as the owner
- // We are the owner because we don't want a lighthouse server to advertise for static hosts it was configured with
- // And we don't want a lighthouse query reply to interfere with our learned cache if we are a client
- //NOTE: this function should not interact with any hot path objects, like lh.staticList, the caller should handle it
- func (lh *LightHouse) addStaticRemote(vpnIp iputil.VpnIp, toAddr *udp.Addr, staticList map[iputil.VpnIp]struct{}) {
- lh.Lock()
- am := lh.unlockedGetRemoteList(vpnIp)
- am.Lock()
- defer am.Unlock()
- lh.Unlock()
- if ipv4 := toAddr.IP.To4(); ipv4 != nil {
- to := NewIp4AndPort(ipv4, uint32(toAddr.Port))
- if !lh.unlockedShouldAddV4(vpnIp, to) {
- return
- }
- am.unlockedPrependV4(lh.myVpnIp, to)
- } else {
- to := NewIp6AndPort(toAddr.IP, uint32(toAddr.Port))
- if !lh.unlockedShouldAddV6(vpnIp, to) {
- return
- }
- am.unlockedPrependV6(lh.myVpnIp, to)
- }
- // Mark it as static in the caller provided map
- staticList[vpnIp] = struct{}{}
- }
- // unlockedGetRemoteList assumes you have the lh lock
- func (lh *LightHouse) unlockedGetRemoteList(vpnIp iputil.VpnIp) *RemoteList {
- am, ok := lh.addrMap[vpnIp]
- if !ok {
- am = NewRemoteList()
- lh.addrMap[vpnIp] = am
- }
- return am
- }
- // unlockedShouldAddV4 checks if to is allowed by our allow list
- func (lh *LightHouse) unlockedShouldAddV4(vpnIp iputil.VpnIp, to *Ip4AndPort) bool {
- allow := lh.GetRemoteAllowList().AllowIpV4(vpnIp, iputil.VpnIp(to.Ip))
- if lh.l.Level >= logrus.TraceLevel {
- lh.l.WithField("remoteIp", vpnIp).WithField("allow", allow).Trace("remoteAllowList.Allow")
- }
- if !allow || ipMaskContains(lh.myVpnIp, lh.myVpnZeros, iputil.VpnIp(to.Ip)) {
- return false
- }
- return true
- }
- // unlockedShouldAddV6 checks if to is allowed by our allow list
- func (lh *LightHouse) unlockedShouldAddV6(vpnIp iputil.VpnIp, to *Ip6AndPort) bool {
- allow := lh.GetRemoteAllowList().AllowIpV6(vpnIp, to.Hi, to.Lo)
- if lh.l.Level >= logrus.TraceLevel {
- lh.l.WithField("remoteIp", lhIp6ToIp(to)).WithField("allow", allow).Trace("remoteAllowList.Allow")
- }
- // We don't check our vpn network here because nebula does not support ipv6 on the inside
- if !allow {
- return false
- }
- return true
- }
- func lhIp6ToIp(v *Ip6AndPort) net.IP {
- ip := make(net.IP, 16)
- binary.BigEndian.PutUint64(ip[:8], v.Hi)
- binary.BigEndian.PutUint64(ip[8:], v.Lo)
- return ip
- }
- func (lh *LightHouse) IsLighthouseIP(vpnIp iputil.VpnIp) bool {
- if _, ok := lh.GetLighthouses()[vpnIp]; ok {
- return true
- }
- return false
- }
- func NewLhQueryByInt(VpnIp iputil.VpnIp) *NebulaMeta {
- return &NebulaMeta{
- Type: NebulaMeta_HostQuery,
- Details: &NebulaMetaDetails{
- VpnIp: uint32(VpnIp),
- },
- }
- }
- func NewIp4AndPort(ip net.IP, port uint32) *Ip4AndPort {
- ipp := Ip4AndPort{Port: port}
- ipp.Ip = uint32(iputil.Ip2VpnIp(ip))
- return &ipp
- }
- func NewIp6AndPort(ip net.IP, port uint32) *Ip6AndPort {
- return &Ip6AndPort{
- Hi: binary.BigEndian.Uint64(ip[:8]),
- Lo: binary.BigEndian.Uint64(ip[8:]),
- Port: port,
- }
- }
- func NewUDPAddrFromLH4(ipp *Ip4AndPort) *udp.Addr {
- ip := ipp.Ip
- return udp.NewAddr(
- net.IPv4(byte(ip&0xff000000>>24), byte(ip&0x00ff0000>>16), byte(ip&0x0000ff00>>8), byte(ip&0x000000ff)),
- uint16(ipp.Port),
- )
- }
- func NewUDPAddrFromLH6(ipp *Ip6AndPort) *udp.Addr {
- return udp.NewAddr(lhIp6ToIp(ipp), uint16(ipp.Port))
- }
- func (lh *LightHouse) LhUpdateWorker(ctx context.Context, f udp.EncWriter) {
- lh.updateParentCtx = ctx
- lh.updateUdp = f
- interval := lh.GetUpdateInterval()
- if lh.amLighthouse || interval == 0 {
- return
- }
- clockSource := time.NewTicker(time.Second * time.Duration(interval))
- updateCtx, cancel := context.WithCancel(ctx)
- lh.updateCancel = cancel
- defer clockSource.Stop()
- for {
- lh.SendUpdate(f)
- select {
- case <-updateCtx.Done():
- return
- case <-clockSource.C:
- continue
- }
- }
- }
- func (lh *LightHouse) SendUpdate(f udp.EncWriter) {
- var v4 []*Ip4AndPort
- var v6 []*Ip6AndPort
- for _, e := range lh.GetAdvertiseAddrs() {
- if ip := e.ip.To4(); ip != nil {
- v4 = append(v4, NewIp4AndPort(e.ip, uint32(e.port)))
- } else {
- v6 = append(v6, NewIp6AndPort(e.ip, uint32(e.port)))
- }
- }
- lal := lh.GetLocalAllowList()
- for _, e := range *localIps(lh.l, lal) {
- if ip4 := e.To4(); ip4 != nil && ipMaskContains(lh.myVpnIp, lh.myVpnZeros, iputil.Ip2VpnIp(ip4)) {
- continue
- }
- // Only add IPs that aren't my VPN/tun IP
- if ip := e.To4(); ip != nil {
- v4 = append(v4, NewIp4AndPort(e, lh.nebulaPort))
- } else {
- v6 = append(v6, NewIp6AndPort(e, lh.nebulaPort))
- }
- }
- var relays []uint32
- for _, r := range lh.GetRelaysForMe() {
- relays = append(relays, (uint32)(r))
- }
- m := &NebulaMeta{
- Type: NebulaMeta_HostUpdateNotification,
- Details: &NebulaMetaDetails{
- VpnIp: uint32(lh.myVpnIp),
- Ip4AndPorts: v4,
- Ip6AndPorts: v6,
- RelayVpnIp: relays,
- },
- }
- lighthouses := lh.GetLighthouses()
- lh.metricTx(NebulaMeta_HostUpdateNotification, int64(len(lighthouses)))
- nb := make([]byte, 12, 12)
- out := make([]byte, mtu)
- mm, err := m.Marshal()
- if err != nil {
- lh.l.WithError(err).Error("Error while marshaling for lighthouse update")
- return
- }
- for vpnIp := range lighthouses {
- f.SendMessageToVpnIp(header.LightHouse, 0, vpnIp, mm, nb, out)
- }
- }
- type LightHouseHandler struct {
- lh *LightHouse
- nb []byte
- out []byte
- pb []byte
- meta *NebulaMeta
- l *logrus.Logger
- }
- func (lh *LightHouse) NewRequestHandler() *LightHouseHandler {
- lhh := &LightHouseHandler{
- lh: lh,
- nb: make([]byte, 12, 12),
- out: make([]byte, mtu),
- l: lh.l,
- pb: make([]byte, mtu),
- meta: &NebulaMeta{
- Details: &NebulaMetaDetails{},
- },
- }
- return lhh
- }
- func (lh *LightHouse) metricRx(t NebulaMeta_MessageType, i int64) {
- lh.metrics.Rx(header.MessageType(t), 0, i)
- }
- func (lh *LightHouse) metricTx(t NebulaMeta_MessageType, i int64) {
- lh.metrics.Tx(header.MessageType(t), 0, i)
- }
- // This method is similar to Reset(), but it re-uses the pointer structs
- // so that we don't have to re-allocate them
- func (lhh *LightHouseHandler) resetMeta() *NebulaMeta {
- details := lhh.meta.Details
- lhh.meta.Reset()
- // Keep the array memory around
- details.Ip4AndPorts = details.Ip4AndPorts[:0]
- details.Ip6AndPorts = details.Ip6AndPorts[:0]
- details.RelayVpnIp = details.RelayVpnIp[:0]
- lhh.meta.Details = details
- return lhh.meta
- }
- func (lhh *LightHouseHandler) HandleRequest(rAddr *udp.Addr, vpnIp iputil.VpnIp, p []byte, w udp.EncWriter) {
- n := lhh.resetMeta()
- err := n.Unmarshal(p)
- if err != nil {
- lhh.l.WithError(err).WithField("vpnIp", vpnIp).WithField("udpAddr", rAddr).
- Error("Failed to unmarshal lighthouse packet")
- //TODO: send recv_error?
- return
- }
- if n.Details == nil {
- lhh.l.WithField("vpnIp", vpnIp).WithField("udpAddr", rAddr).
- Error("Invalid lighthouse update")
- //TODO: send recv_error?
- return
- }
- lhh.lh.metricRx(n.Type, 1)
- switch n.Type {
- case NebulaMeta_HostQuery:
- lhh.handleHostQuery(n, vpnIp, rAddr, w)
- case NebulaMeta_HostQueryReply:
- lhh.handleHostQueryReply(n, vpnIp)
- case NebulaMeta_HostUpdateNotification:
- lhh.handleHostUpdateNotification(n, vpnIp)
- case NebulaMeta_HostMovedNotification:
- case NebulaMeta_HostPunchNotification:
- lhh.handleHostPunchNotification(n, vpnIp, w)
- }
- }
- func (lhh *LightHouseHandler) handleHostQuery(n *NebulaMeta, vpnIp iputil.VpnIp, addr *udp.Addr, w udp.EncWriter) {
- // Exit if we don't answer queries
- if !lhh.lh.amLighthouse {
- if lhh.l.Level >= logrus.DebugLevel {
- lhh.l.Debugln("I don't answer queries, but received from: ", addr)
- }
- return
- }
- //TODO: we can DRY this further
- reqVpnIp := n.Details.VpnIp
- //TODO: Maybe instead of marshalling into n we marshal into a new `r` to not nuke our current request data
- found, ln, err := lhh.lh.queryAndPrepMessage(iputil.VpnIp(n.Details.VpnIp), func(c *cache) (int, error) {
- n = lhh.resetMeta()
- n.Type = NebulaMeta_HostQueryReply
- n.Details.VpnIp = reqVpnIp
- lhh.coalesceAnswers(c, n)
- return n.MarshalTo(lhh.pb)
- })
- if !found {
- return
- }
- if err != nil {
- lhh.l.WithError(err).WithField("vpnIp", vpnIp).Error("Failed to marshal lighthouse host query reply")
- return
- }
- lhh.lh.metricTx(NebulaMeta_HostQueryReply, 1)
- w.SendMessageToVpnIp(header.LightHouse, 0, vpnIp, lhh.pb[:ln], lhh.nb, lhh.out[:0])
- // This signals the other side to punch some zero byte udp packets
- found, ln, err = lhh.lh.queryAndPrepMessage(vpnIp, func(c *cache) (int, error) {
- n = lhh.resetMeta()
- n.Type = NebulaMeta_HostPunchNotification
- n.Details.VpnIp = uint32(vpnIp)
- lhh.coalesceAnswers(c, n)
- return n.MarshalTo(lhh.pb)
- })
- if !found {
- return
- }
- if err != nil {
- lhh.l.WithError(err).WithField("vpnIp", vpnIp).Error("Failed to marshal lighthouse host was queried for")
- return
- }
- lhh.lh.metricTx(NebulaMeta_HostPunchNotification, 1)
- w.SendMessageToVpnIp(header.LightHouse, 0, iputil.VpnIp(reqVpnIp), lhh.pb[:ln], lhh.nb, lhh.out[:0])
- }
- func (lhh *LightHouseHandler) coalesceAnswers(c *cache, n *NebulaMeta) {
- if c.v4 != nil {
- if c.v4.learned != nil {
- n.Details.Ip4AndPorts = append(n.Details.Ip4AndPorts, c.v4.learned)
- }
- if c.v4.reported != nil && len(c.v4.reported) > 0 {
- n.Details.Ip4AndPorts = append(n.Details.Ip4AndPorts, c.v4.reported...)
- }
- }
- if c.v6 != nil {
- if c.v6.learned != nil {
- n.Details.Ip6AndPorts = append(n.Details.Ip6AndPorts, c.v6.learned)
- }
- if c.v6.reported != nil && len(c.v6.reported) > 0 {
- n.Details.Ip6AndPorts = append(n.Details.Ip6AndPorts, c.v6.reported...)
- }
- }
- if c.relay != nil {
- n.Details.RelayVpnIp = append(n.Details.RelayVpnIp, c.relay.relay...)
- }
- }
- func (lhh *LightHouseHandler) handleHostQueryReply(n *NebulaMeta, vpnIp iputil.VpnIp) {
- if !lhh.lh.IsLighthouseIP(vpnIp) {
- return
- }
- lhh.lh.Lock()
- am := lhh.lh.unlockedGetRemoteList(iputil.VpnIp(n.Details.VpnIp))
- am.Lock()
- lhh.lh.Unlock()
- certVpnIp := iputil.VpnIp(n.Details.VpnIp)
- am.unlockedSetV4(vpnIp, certVpnIp, n.Details.Ip4AndPorts, lhh.lh.unlockedShouldAddV4)
- am.unlockedSetV6(vpnIp, certVpnIp, n.Details.Ip6AndPorts, lhh.lh.unlockedShouldAddV6)
- am.unlockedSetRelay(vpnIp, certVpnIp, n.Details.RelayVpnIp)
- am.Unlock()
- // Non-blocking attempt to trigger, skip if it would block
- select {
- case lhh.lh.handshakeTrigger <- iputil.VpnIp(n.Details.VpnIp):
- default:
- }
- }
- func (lhh *LightHouseHandler) handleHostUpdateNotification(n *NebulaMeta, vpnIp iputil.VpnIp) {
- if !lhh.lh.amLighthouse {
- if lhh.l.Level >= logrus.DebugLevel {
- lhh.l.Debugln("I am not a lighthouse, do not take host updates: ", vpnIp)
- }
- return
- }
- //Simple check that the host sent this not someone else
- if n.Details.VpnIp != uint32(vpnIp) {
- if lhh.l.Level >= logrus.DebugLevel {
- lhh.l.WithField("vpnIp", vpnIp).WithField("answer", iputil.VpnIp(n.Details.VpnIp)).Debugln("Host sent invalid update")
- }
- return
- }
- lhh.lh.Lock()
- am := lhh.lh.unlockedGetRemoteList(vpnIp)
- am.Lock()
- lhh.lh.Unlock()
- certVpnIp := iputil.VpnIp(n.Details.VpnIp)
- am.unlockedSetV4(vpnIp, certVpnIp, n.Details.Ip4AndPorts, lhh.lh.unlockedShouldAddV4)
- am.unlockedSetV6(vpnIp, certVpnIp, n.Details.Ip6AndPorts, lhh.lh.unlockedShouldAddV6)
- am.unlockedSetRelay(vpnIp, certVpnIp, n.Details.RelayVpnIp)
- am.Unlock()
- }
- func (lhh *LightHouseHandler) handleHostPunchNotification(n *NebulaMeta, vpnIp iputil.VpnIp, w udp.EncWriter) {
- if !lhh.lh.IsLighthouseIP(vpnIp) {
- return
- }
- empty := []byte{0}
- punch := func(vpnPeer *udp.Addr) {
- if vpnPeer == nil {
- return
- }
- go func() {
- time.Sleep(lhh.lh.punchy.GetDelay())
- lhh.lh.metricHolepunchTx.Inc(1)
- lhh.lh.punchConn.WriteTo(empty, vpnPeer)
- }()
- if lhh.l.Level >= logrus.DebugLevel {
- //TODO: lacking the ip we are actually punching on, old: l.Debugf("Punching %s on %d for %s", IntIp(a.Ip), a.Port, IntIp(n.Details.VpnIp))
- lhh.l.Debugf("Punching on %d for %s", vpnPeer.Port, iputil.VpnIp(n.Details.VpnIp))
- }
- }
- for _, a := range n.Details.Ip4AndPorts {
- punch(NewUDPAddrFromLH4(a))
- }
- for _, a := range n.Details.Ip6AndPorts {
- punch(NewUDPAddrFromLH6(a))
- }
- // This sends a nebula test packet to the host trying to contact us. In the case
- // of a double nat or other difficult scenario, this may help establish
- // a tunnel.
- if lhh.lh.punchy.GetRespond() {
- queryVpnIp := iputil.VpnIp(n.Details.VpnIp)
- go func() {
- time.Sleep(time.Second * 5)
- if lhh.l.Level >= logrus.DebugLevel {
- lhh.l.Debugf("Sending a nebula test packet to vpn ip %s", queryVpnIp)
- }
- //NOTE: we have to allocate a new output buffer here since we are spawning a new goroutine
- // for each punchBack packet. We should move this into a timerwheel or a single goroutine
- // managed by a channel.
- w.SendMessageToVpnIp(header.Test, header.TestRequest, queryVpnIp, []byte(""), make([]byte, 12, 12), make([]byte, mtu))
- }()
- }
- }
- // ipMaskContains checks if testIp is contained by ip after applying a cidr
- // zeros is 32 - bits from net.IPMask.Size()
- func ipMaskContains(ip iputil.VpnIp, zeros iputil.VpnIp, testIp iputil.VpnIp) bool {
- return (testIp^ip)>>zeros == 0
- }
|