|
@@ -7,14 +7,18 @@ import (
|
|
|
"fmt"
|
|
|
"net"
|
|
|
"sync"
|
|
|
+ "sync/atomic"
|
|
|
"time"
|
|
|
+ "unsafe"
|
|
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
|
"github.com/rcrowley/go-metrics"
|
|
|
"github.com/sirupsen/logrus"
|
|
|
+ "github.com/slackhq/nebula/config"
|
|
|
"github.com/slackhq/nebula/header"
|
|
|
"github.com/slackhq/nebula/iputil"
|
|
|
"github.com/slackhq/nebula/udp"
|
|
|
+ "github.com/slackhq/nebula/util"
|
|
|
)
|
|
|
|
|
|
//TODO: if a lighthouse doesn't have an answer, clients AGGRESSIVELY REQUERY.. why? handshake manager and/or getOrHandshake?
|
|
@@ -28,7 +32,9 @@ type LightHouse struct {
|
|
|
amLighthouse bool
|
|
|
myVpnIp iputil.VpnIp
|
|
|
myVpnZeros iputil.VpnIp
|
|
|
+ myVpnNet *net.IPNet
|
|
|
punchConn *udp.Conn
|
|
|
+ punchy *Punchy
|
|
|
|
|
|
// Local cache of answers from light houses
|
|
|
// map of vpn Ip to answers
|
|
@@ -39,80 +45,240 @@ type LightHouse struct {
|
|
|
// respond with.
|
|
|
// - When we are not a lighthouse, this filters which addresses we accept
|
|
|
// from lighthouses.
|
|
|
- remoteAllowList *RemoteAllowList
|
|
|
+ atomicRemoteAllowList *RemoteAllowList
|
|
|
|
|
|
// filters local addresses that we advertise to lighthouses
|
|
|
- localAllowList *LocalAllowList
|
|
|
+ atomicLocalAllowList *LocalAllowList
|
|
|
|
|
|
// used to trigger the HandshakeManager when we receive HostQueryReply
|
|
|
handshakeTrigger chan<- iputil.VpnIp
|
|
|
|
|
|
- // staticList exists to avoid having a bool in each addrMap entry
|
|
|
+ // atomicStaticList exists to avoid having a bool in each addrMap entry
|
|
|
// since static should be rare
|
|
|
- staticList map[iputil.VpnIp]struct{}
|
|
|
- lighthouses map[iputil.VpnIp]struct{}
|
|
|
- interval int
|
|
|
- nebulaPort uint32 // 32 bits because protobuf does not have a uint16
|
|
|
- punchBack bool
|
|
|
- punchDelay time.Duration
|
|
|
+ atomicStaticList map[iputil.VpnIp]struct{}
|
|
|
+ atomicLighthouses map[iputil.VpnIp]struct{}
|
|
|
+
|
|
|
+ atomicInterval int64
|
|
|
+ updateCancel context.CancelFunc
|
|
|
+ updateParentCtx context.Context
|
|
|
+ updateUdp udp.EncWriter
|
|
|
+ nebulaPort uint32 // 32 bits because protobuf does not have a uint16
|
|
|
|
|
|
metrics *MessageMetrics
|
|
|
metricHolepunchTx metrics.Counter
|
|
|
l *logrus.Logger
|
|
|
}
|
|
|
|
|
|
-func NewLightHouse(l *logrus.Logger, amLighthouse bool, myVpnIpNet *net.IPNet, ips []iputil.VpnIp, interval int, nebulaPort uint32, pc *udp.Conn, punchBack bool, punchDelay time.Duration, metricsEnabled bool) *LightHouse {
|
|
|
- ones, _ := myVpnIpNet.Mask.Size()
|
|
|
+// NewLightHouseFromConfig will build a Lighthouse struct from the values provided in the config object
|
|
|
+// addrMap should be nil unless this is during a config reload
|
|
|
+func NewLightHouseFromConfig(l *logrus.Logger, c *config.C, myVpnNet *net.IPNet, pc *udp.Conn, p *Punchy) (*LightHouse, error) {
|
|
|
+ amLighthouse := c.GetBool("lighthouse.am_lighthouse", false)
|
|
|
+ nebulaPort := uint32(c.GetInt("listen.port", 0))
|
|
|
+ if amLighthouse && nebulaPort == 0 {
|
|
|
+ return nil, util.NewContextualError("lighthouse.am_lighthouse enabled on node but no port number is set in config", nil, nil)
|
|
|
+ }
|
|
|
+
|
|
|
+ ones, _ := myVpnNet.Mask.Size()
|
|
|
h := LightHouse{
|
|
|
- amLighthouse: amLighthouse,
|
|
|
- myVpnIp: iputil.Ip2VpnIp(myVpnIpNet.IP),
|
|
|
- myVpnZeros: iputil.VpnIp(32 - ones),
|
|
|
- addrMap: make(map[iputil.VpnIp]*RemoteList),
|
|
|
- nebulaPort: nebulaPort,
|
|
|
- lighthouses: make(map[iputil.VpnIp]struct{}),
|
|
|
- staticList: make(map[iputil.VpnIp]struct{}),
|
|
|
- interval: interval,
|
|
|
- punchConn: pc,
|
|
|
- punchBack: punchBack,
|
|
|
- punchDelay: punchDelay,
|
|
|
- l: l,
|
|
|
- }
|
|
|
-
|
|
|
- if metricsEnabled {
|
|
|
+ amLighthouse: amLighthouse,
|
|
|
+ myVpnIp: iputil.Ip2VpnIp(myVpnNet.IP),
|
|
|
+ myVpnZeros: iputil.VpnIp(32 - ones),
|
|
|
+ myVpnNet: myVpnNet,
|
|
|
+ addrMap: make(map[iputil.VpnIp]*RemoteList),
|
|
|
+ nebulaPort: nebulaPort,
|
|
|
+ atomicLighthouses: make(map[iputil.VpnIp]struct{}),
|
|
|
+ atomicStaticList: make(map[iputil.VpnIp]struct{}),
|
|
|
+ punchConn: pc,
|
|
|
+ punchy: p,
|
|
|
+ l: l,
|
|
|
+ }
|
|
|
+
|
|
|
+ if c.GetBool("stats.lighthouse_metrics", false) {
|
|
|
h.metrics = newLighthouseMetrics()
|
|
|
-
|
|
|
h.metricHolepunchTx = metrics.GetOrRegisterCounter("messages.tx.holepunch", nil)
|
|
|
} else {
|
|
|
h.metricHolepunchTx = metrics.NilCounter{}
|
|
|
}
|
|
|
|
|
|
- for _, ip := range ips {
|
|
|
- h.lighthouses[ip] = struct{}{}
|
|
|
+ err := h.reload(c, true)
|
|
|
+ if err != nil {
|
|
|
+ return nil, err
|
|
|
}
|
|
|
|
|
|
- return &h
|
|
|
+ c.RegisterReloadCallback(func(c *config.C) {
|
|
|
+ err := h.reload(c, false)
|
|
|
+ switch v := err.(type) {
|
|
|
+ case util.ContextualError:
|
|
|
+ v.Log(l)
|
|
|
+ case error:
|
|
|
+ l.WithError(err).Error("failed to reload lighthouse")
|
|
|
+ }
|
|
|
+ })
|
|
|
+
|
|
|
+ return &h, nil
|
|
|
}
|
|
|
|
|
|
-func (lh *LightHouse) SetRemoteAllowList(allowList *RemoteAllowList) {
|
|
|
- lh.Lock()
|
|
|
- defer lh.Unlock()
|
|
|
+func (lh *LightHouse) GetStaticHostList() map[iputil.VpnIp]struct{} {
|
|
|
+ return *(*map[iputil.VpnIp]struct{})(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicStaticList))))
|
|
|
+}
|
|
|
|
|
|
- lh.remoteAllowList = allowList
|
|
|
+func (lh *LightHouse) GetLighthouses() map[iputil.VpnIp]struct{} {
|
|
|
+ return *(*map[iputil.VpnIp]struct{})(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLighthouses))))
|
|
|
}
|
|
|
|
|
|
-func (lh *LightHouse) SetLocalAllowList(allowList *LocalAllowList) {
|
|
|
- lh.Lock()
|
|
|
- defer lh.Unlock()
|
|
|
+func (lh *LightHouse) GetRemoteAllowList() *RemoteAllowList {
|
|
|
+ return (*RemoteAllowList)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRemoteAllowList))))
|
|
|
+}
|
|
|
+
|
|
|
+func (lh *LightHouse) GetLocalAllowList() *LocalAllowList {
|
|
|
+ return (*LocalAllowList)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLocalAllowList))))
|
|
|
+}
|
|
|
+
|
|
|
+func (lh *LightHouse) GetUpdateInterval() int64 {
|
|
|
+ return atomic.LoadInt64(&lh.atomicInterval)
|
|
|
+}
|
|
|
+
|
|
|
+func (lh *LightHouse) reload(c *config.C, initial bool) error {
|
|
|
+ if initial || c.HasChanged("lighthouse.interval") {
|
|
|
+ atomic.StoreInt64(&lh.atomicInterval, int64(c.GetInt("lighthouse.interval", 10)))
|
|
|
+
|
|
|
+ if !initial {
|
|
|
+ lh.l.Infof("lighthouse.interval changed to %v", lh.atomicInterval)
|
|
|
+
|
|
|
+ if lh.updateCancel != nil {
|
|
|
+ // May not always have a running routine
|
|
|
+ lh.updateCancel()
|
|
|
+ }
|
|
|
+
|
|
|
+ lh.LhUpdateWorker(lh.updateParentCtx, lh.updateUdp)
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if initial || c.HasChanged("lighthouse.remote_allow_list") || c.HasChanged("lighthouse.remote_allow_ranges") {
|
|
|
+ ral, err := NewRemoteAllowListFromConfig(c, "lighthouse.remote_allow_list", "lighthouse.remote_allow_ranges")
|
|
|
+ if err != nil {
|
|
|
+ return util.NewContextualError("Invalid lighthouse.remote_allow_list", nil, err)
|
|
|
+ }
|
|
|
+
|
|
|
+ atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicRemoteAllowList)), unsafe.Pointer(ral))
|
|
|
+ if !initial {
|
|
|
+ //TODO: a diff will be annoyingly difficult
|
|
|
+ lh.l.Info("lighthouse.remote_allow_list and/or lighthouse.remote_allow_ranges has changed")
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if initial || c.HasChanged("lighthouse.local_allow_list") {
|
|
|
+ lal, err := NewLocalAllowListFromConfig(c, "lighthouse.local_allow_list")
|
|
|
+ if err != nil {
|
|
|
+ return util.NewContextualError("Invalid lighthouse.local_allow_list", nil, err)
|
|
|
+ }
|
|
|
+
|
|
|
+ atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLocalAllowList)), unsafe.Pointer(lal))
|
|
|
+ if !initial {
|
|
|
+ //TODO: a diff will be annoyingly difficult
|
|
|
+ lh.l.Info("lighthouse.local_allow_list has changed")
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //NOTE: many things will get much simpler when we combine static_host_map and lighthouse.hosts in config
|
|
|
+ if initial || c.HasChanged("static_host_map") {
|
|
|
+ staticList := make(map[iputil.VpnIp]struct{})
|
|
|
+ err := lh.loadStaticMap(c, lh.myVpnNet, staticList)
|
|
|
+ if err != nil {
|
|
|
+ return err
|
|
|
+ }
|
|
|
+
|
|
|
+ atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicStaticList)), unsafe.Pointer(&staticList))
|
|
|
+ if !initial {
|
|
|
+ //TODO: we should remove any remote list entries for static hosts that were removed/modified?
|
|
|
+ lh.l.Info("static_host_map has changed")
|
|
|
+ }
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ if initial || c.HasChanged("lighthouse.hosts") {
|
|
|
+ lhMap := make(map[iputil.VpnIp]struct{})
|
|
|
+ err := lh.parseLighthouses(c, lh.myVpnNet, lhMap)
|
|
|
+ if err != nil {
|
|
|
+ return err
|
|
|
+ }
|
|
|
+
|
|
|
+ atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&lh.atomicLighthouses)), unsafe.Pointer(&lhMap))
|
|
|
+ if !initial {
|
|
|
+ //NOTE: we are not tearing down existing lighthouse connections because they might be used for non lighthouse traffic
|
|
|
+ lh.l.Info("lighthouse.hosts has changed")
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return nil
|
|
|
+}
|
|
|
+
|
|
|
+func (lh *LightHouse) parseLighthouses(c *config.C, tunCidr *net.IPNet, lhMap map[iputil.VpnIp]struct{}) error {
|
|
|
+ lhs := c.GetStringSlice("lighthouse.hosts", []string{})
|
|
|
+ if lh.amLighthouse && len(lhs) != 0 {
|
|
|
+ lh.l.Warn("lighthouse.am_lighthouse enabled on node but upstream lighthouses exist in config")
|
|
|
+ }
|
|
|
+
|
|
|
+ for i, host := range lhs {
|
|
|
+ ip := net.ParseIP(host)
|
|
|
+ if ip == nil {
|
|
|
+ return util.NewContextualError("Unable to parse lighthouse host entry", m{"host": host, "entry": i + 1}, nil)
|
|
|
+ }
|
|
|
+ if !tunCidr.Contains(ip) {
|
|
|
+ return util.NewContextualError("lighthouse host is not in our subnet, invalid", m{"vpnIp": ip, "network": tunCidr.String()}, nil)
|
|
|
+ }
|
|
|
+ lhMap[iputil.Ip2VpnIp(ip)] = struct{}{}
|
|
|
+ }
|
|
|
+
|
|
|
+ if !lh.amLighthouse && len(lhMap) == 0 {
|
|
|
+ lh.l.Warn("No lighthouse.hosts configured, this host will only be able to initiate tunnels with static_host_map entries")
|
|
|
+ }
|
|
|
|
|
|
- lh.localAllowList = allowList
|
|
|
+ staticList := lh.GetStaticHostList()
|
|
|
+ for lhIP, _ := range lhMap {
|
|
|
+ if _, ok := staticList[lhIP]; !ok {
|
|
|
+ return fmt.Errorf("lighthouse %s does not have a static_host_map entry", lhIP)
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return nil
|
|
|
}
|
|
|
|
|
|
-func (lh *LightHouse) ValidateLHStaticEntries() error {
|
|
|
- for lhIP, _ := range lh.lighthouses {
|
|
|
- if _, ok := lh.staticList[lhIP]; !ok {
|
|
|
- return fmt.Errorf("Lighthouse %s does not have a static_host_map entry", lhIP)
|
|
|
+func (lh *LightHouse) loadStaticMap(c *config.C, tunCidr *net.IPNet, staticList map[iputil.VpnIp]struct{}) error {
|
|
|
+ shm := c.GetMap("static_host_map", map[interface{}]interface{}{})
|
|
|
+ i := 0
|
|
|
+
|
|
|
+ for k, v := range shm {
|
|
|
+ rip := net.ParseIP(fmt.Sprintf("%v", k))
|
|
|
+ if rip == nil {
|
|
|
+ return util.NewContextualError("Unable to parse static_host_map entry", m{"host": k, "entry": i + 1}, nil)
|
|
|
+ }
|
|
|
+
|
|
|
+ if !tunCidr.Contains(rip) {
|
|
|
+ return util.NewContextualError("static_host_map key is not in our subnet, invalid", m{"vpnIp": rip, "network": tunCidr.String(), "entry": i + 1}, nil)
|
|
|
}
|
|
|
+
|
|
|
+ vpnIp := iputil.Ip2VpnIp(rip)
|
|
|
+ vals, ok := v.([]interface{})
|
|
|
+ if ok {
|
|
|
+ for _, v := range vals {
|
|
|
+ ip, port, err := udp.ParseIPAndPort(fmt.Sprintf("%v", v))
|
|
|
+ if err != nil {
|
|
|
+ return util.NewContextualError("Static host address could not be parsed", m{"vpnIp": vpnIp, "entry": i + 1}, err)
|
|
|
+ }
|
|
|
+ lh.addStaticRemote(vpnIp, udp.NewAddr(ip, port), staticList)
|
|
|
+ }
|
|
|
+
|
|
|
+ } else {
|
|
|
+ ip, port, err := udp.ParseIPAndPort(fmt.Sprintf("%v", v))
|
|
|
+ if err != nil {
|
|
|
+ return util.NewContextualError("Static host address could not be parsed", m{"vpnIp": vpnIp, "entry": i + 1}, err)
|
|
|
+ }
|
|
|
+ lh.addStaticRemote(vpnIp, udp.NewAddr(ip, port), staticList)
|
|
|
+ }
|
|
|
+ i++
|
|
|
}
|
|
|
+
|
|
|
return nil
|
|
|
}
|
|
|
|
|
@@ -146,10 +312,11 @@ func (lh *LightHouse) QueryServer(ip iputil.VpnIp, f udp.EncWriter) {
|
|
|
return
|
|
|
}
|
|
|
|
|
|
- lh.metricTx(NebulaMeta_HostQuery, int64(len(lh.lighthouses)))
|
|
|
+ lighthouses := lh.GetLighthouses()
|
|
|
+ lh.metricTx(NebulaMeta_HostQuery, int64(len(lighthouses)))
|
|
|
nb := make([]byte, 12, 12)
|
|
|
out := make([]byte, mtu)
|
|
|
- for n := range lh.lighthouses {
|
|
|
+ for n := range lighthouses {
|
|
|
f.SendMessageToVpnIp(header.LightHouse, 0, n, query, nb, out)
|
|
|
}
|
|
|
}
|
|
@@ -197,7 +364,7 @@ func (lh *LightHouse) queryAndPrepMessage(vpnIp iputil.VpnIp, f func(*cache) (in
|
|
|
func (lh *LightHouse) DeleteVpnIp(vpnIp iputil.VpnIp) {
|
|
|
// First we check the static mapping
|
|
|
// and do nothing if it is there
|
|
|
- if _, ok := lh.staticList[vpnIp]; ok {
|
|
|
+ if _, ok := lh.GetStaticHostList()[vpnIp]; ok {
|
|
|
return
|
|
|
}
|
|
|
lh.Lock()
|
|
@@ -211,10 +378,11 @@ func (lh *LightHouse) DeleteVpnIp(vpnIp iputil.VpnIp) {
|
|
|
lh.Unlock()
|
|
|
}
|
|
|
|
|
|
-// AddStaticRemote adds a static host entry for vpnIp as ourselves as the owner
|
|
|
+// addStaticRemote adds a static host entry for vpnIp as ourselves as the owner
|
|
|
// We are the owner because we don't want a lighthouse server to advertise for static hosts it was configured with
|
|
|
// And we don't want a lighthouse query reply to interfere with our learned cache if we are a client
|
|
|
-func (lh *LightHouse) AddStaticRemote(vpnIp iputil.VpnIp, toAddr *udp.Addr) {
|
|
|
+//NOTE: this function should not interact with any hot path objects, like lh.staticList, the caller should handle it
|
|
|
+func (lh *LightHouse) addStaticRemote(vpnIp iputil.VpnIp, toAddr *udp.Addr, staticList map[iputil.VpnIp]struct{}) {
|
|
|
lh.Lock()
|
|
|
am := lh.unlockedGetRemoteList(vpnIp)
|
|
|
am.Lock()
|
|
@@ -236,8 +404,8 @@ func (lh *LightHouse) AddStaticRemote(vpnIp iputil.VpnIp, toAddr *udp.Addr) {
|
|
|
am.unlockedPrependV6(lh.myVpnIp, to)
|
|
|
}
|
|
|
|
|
|
- // Mark it as static
|
|
|
- lh.staticList[vpnIp] = struct{}{}
|
|
|
+ // Mark it as static in the caller provided map
|
|
|
+ staticList[vpnIp] = struct{}{}
|
|
|
}
|
|
|
|
|
|
// unlockedGetRemoteList assumes you have the lh lock
|
|
@@ -252,7 +420,7 @@ func (lh *LightHouse) unlockedGetRemoteList(vpnIp iputil.VpnIp) *RemoteList {
|
|
|
|
|
|
// unlockedShouldAddV4 checks if to is allowed by our allow list
|
|
|
func (lh *LightHouse) unlockedShouldAddV4(vpnIp iputil.VpnIp, to *Ip4AndPort) bool {
|
|
|
- allow := lh.remoteAllowList.AllowIpV4(vpnIp, iputil.VpnIp(to.Ip))
|
|
|
+ allow := lh.GetRemoteAllowList().AllowIpV4(vpnIp, iputil.VpnIp(to.Ip))
|
|
|
if lh.l.Level >= logrus.TraceLevel {
|
|
|
lh.l.WithField("remoteIp", vpnIp).WithField("allow", allow).Trace("remoteAllowList.Allow")
|
|
|
}
|
|
@@ -266,7 +434,7 @@ func (lh *LightHouse) unlockedShouldAddV4(vpnIp iputil.VpnIp, to *Ip4AndPort) bo
|
|
|
|
|
|
// unlockedShouldAddV6 checks if to is allowed by our allow list
|
|
|
func (lh *LightHouse) unlockedShouldAddV6(vpnIp iputil.VpnIp, to *Ip6AndPort) bool {
|
|
|
- allow := lh.remoteAllowList.AllowIpV6(vpnIp, to.Hi, to.Lo)
|
|
|
+ allow := lh.GetRemoteAllowList().AllowIpV6(vpnIp, to.Hi, to.Lo)
|
|
|
if lh.l.Level >= logrus.TraceLevel {
|
|
|
lh.l.WithField("remoteIp", lhIp6ToIp(to)).WithField("allow", allow).Trace("remoteAllowList.Allow")
|
|
|
}
|
|
@@ -287,7 +455,7 @@ func lhIp6ToIp(v *Ip6AndPort) net.IP {
|
|
|
}
|
|
|
|
|
|
func (lh *LightHouse) IsLighthouseIP(vpnIp iputil.VpnIp) bool {
|
|
|
- if _, ok := lh.lighthouses[vpnIp]; ok {
|
|
|
+ if _, ok := lh.GetLighthouses()[vpnIp]; ok {
|
|
|
return true
|
|
|
}
|
|
|
return false
|
|
@@ -329,18 +497,24 @@ func NewUDPAddrFromLH6(ipp *Ip6AndPort) *udp.Addr {
|
|
|
}
|
|
|
|
|
|
func (lh *LightHouse) LhUpdateWorker(ctx context.Context, f udp.EncWriter) {
|
|
|
- if lh.amLighthouse || lh.interval == 0 {
|
|
|
+ lh.updateParentCtx = ctx
|
|
|
+ lh.updateUdp = f
|
|
|
+
|
|
|
+ interval := lh.GetUpdateInterval()
|
|
|
+ if lh.amLighthouse || interval == 0 {
|
|
|
return
|
|
|
}
|
|
|
|
|
|
- clockSource := time.NewTicker(time.Second * time.Duration(lh.interval))
|
|
|
+ clockSource := time.NewTicker(time.Second * time.Duration(interval))
|
|
|
+ updateCtx, cancel := context.WithCancel(ctx)
|
|
|
+ lh.updateCancel = cancel
|
|
|
defer clockSource.Stop()
|
|
|
|
|
|
for {
|
|
|
lh.SendUpdate(f)
|
|
|
|
|
|
select {
|
|
|
- case <-ctx.Done():
|
|
|
+ case <-updateCtx.Done():
|
|
|
return
|
|
|
case <-clockSource.C:
|
|
|
continue
|
|
@@ -352,7 +526,8 @@ func (lh *LightHouse) SendUpdate(f udp.EncWriter) {
|
|
|
var v4 []*Ip4AndPort
|
|
|
var v6 []*Ip6AndPort
|
|
|
|
|
|
- for _, e := range *localIps(lh.l, lh.localAllowList) {
|
|
|
+ lal := lh.GetLocalAllowList()
|
|
|
+ for _, e := range *localIps(lh.l, lal) {
|
|
|
if ip4 := e.To4(); ip4 != nil && ipMaskContains(lh.myVpnIp, lh.myVpnZeros, iputil.Ip2VpnIp(ip4)) {
|
|
|
continue
|
|
|
}
|
|
@@ -373,7 +548,8 @@ func (lh *LightHouse) SendUpdate(f udp.EncWriter) {
|
|
|
},
|
|
|
}
|
|
|
|
|
|
- lh.metricTx(NebulaMeta_HostUpdateNotification, int64(len(lh.lighthouses)))
|
|
|
+ lighthouses := lh.GetLighthouses()
|
|
|
+ lh.metricTx(NebulaMeta_HostUpdateNotification, int64(len(lighthouses)))
|
|
|
nb := make([]byte, 12, 12)
|
|
|
out := make([]byte, mtu)
|
|
|
|
|
@@ -383,7 +559,7 @@ func (lh *LightHouse) SendUpdate(f udp.EncWriter) {
|
|
|
return
|
|
|
}
|
|
|
|
|
|
- for vpnIp := range lh.lighthouses {
|
|
|
+ for vpnIp := range lighthouses {
|
|
|
f.SendMessageToVpnIp(header.LightHouse, 0, vpnIp, mm, nb, out)
|
|
|
}
|
|
|
}
|
|
@@ -609,7 +785,7 @@ func (lhh *LightHouseHandler) handleHostPunchNotification(n *NebulaMeta, vpnIp i
|
|
|
}
|
|
|
|
|
|
go func() {
|
|
|
- time.Sleep(lhh.lh.punchDelay)
|
|
|
+ time.Sleep(lhh.lh.punchy.GetDelay())
|
|
|
lhh.lh.metricHolepunchTx.Inc(1)
|
|
|
lhh.lh.punchConn.WriteTo(empty, vpnPeer)
|
|
|
}()
|
|
@@ -631,7 +807,7 @@ func (lhh *LightHouseHandler) handleHostPunchNotification(n *NebulaMeta, vpnIp i
|
|
|
// This sends a nebula test packet to the host trying to contact us. In the case
|
|
|
// of a double nat or other difficult scenario, this may help establish
|
|
|
// a tunnel.
|
|
|
- if lhh.lh.punchBack {
|
|
|
+ if lhh.lh.punchy.GetRespond() {
|
|
|
queryVpnIp := iputil.VpnIp(n.Details.VpnIp)
|
|
|
go func() {
|
|
|
time.Sleep(time.Second * 5)
|