mirror of
https://github.com/tailscale/tailscale.git
synced 2024-11-26 03:25:35 +00:00
2cff9016e4
I started to write a full DNS caching resolver and I realized it was overkill and wouldn't work on Windows even in Go 1.14 yet, so I'm doing this tiny one instead for now, just for all our netcheck STUN derp lookups, and connections to DERP servers. (This will be caching a exactly 8 DNS entries, all ours.) Fixes #145 (can be better later, of course)
255 lines
6.0 KiB
Go
255 lines
6.0 KiB
Go
// Copyright (c) 2020 Tailscale Inc & AUTHORS All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// Package netcheck checks the network conditions from the current host.
|
|
package netcheck
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"net"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"golang.org/x/sync/errgroup"
|
|
"tailscale.com/interfaces"
|
|
"tailscale.com/net/dnscache"
|
|
"tailscale.com/stun"
|
|
"tailscale.com/stunner"
|
|
"tailscale.com/types/logger"
|
|
"tailscale.com/types/opt"
|
|
)
|
|
|
|
type Report struct {
|
|
UDP bool // UDP works
|
|
IPv6 bool // IPv6 works
|
|
MappingVariesByDestIP opt.Bool // for IPv4
|
|
HairPinning opt.Bool // for IPv4
|
|
PreferredDERP int // or 0 for unknown
|
|
DERPLatency map[string]time.Duration // keyed by STUN host:port
|
|
|
|
// TODO: update Clone when adding new fields
|
|
}
|
|
|
|
func (r *Report) Clone() *Report {
|
|
if r == nil {
|
|
return nil
|
|
}
|
|
r2 := *r
|
|
if r2.DERPLatency != nil {
|
|
r2.DERPLatency = map[string]time.Duration{}
|
|
for k, v := range r.DERPLatency {
|
|
r2.DERPLatency[k] = v
|
|
}
|
|
}
|
|
return &r2
|
|
}
|
|
|
|
const derpNodes = 4 // [1,4] contiguous, at present
|
|
|
|
var derpLoc = map[int]string{
|
|
1: "New York",
|
|
2: "San Francsico",
|
|
3: "Singapore",
|
|
4: "Frankfurt",
|
|
}
|
|
|
|
func DERPNodeLocation(id int) string { return derpLoc[id] }
|
|
|
|
func GetReport(ctx context.Context, logf logger.Logf) (*Report, error) {
|
|
|
|
var stunServers []string
|
|
var stunServers6 []string
|
|
for i := 1; i <= derpNodes; i++ {
|
|
stunServers = append(stunServers, fmt.Sprintf("derp%v.tailscale.com:3478", i))
|
|
stunServers6 = append(stunServers6, fmt.Sprintf("derp%v-v6.tailscale.com:3478", i))
|
|
}
|
|
|
|
// Mask user context with ours that we guarantee to cancel so
|
|
// we can depend on it being closed in goroutines later.
|
|
// (User ctx might be context.Background, etc)
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
defer cancel()
|
|
|
|
closeOnCtx := func(c io.Closer) {
|
|
<-ctx.Done()
|
|
c.Close()
|
|
}
|
|
|
|
v6, err := interfaces.HaveIPv6GlobalAddress()
|
|
if err != nil {
|
|
logf("interfaces: %v", err)
|
|
}
|
|
var (
|
|
mu sync.Mutex
|
|
ret = &Report{
|
|
DERPLatency: map[string]time.Duration{},
|
|
}
|
|
gotIP = map[string]string{} // server -> IP
|
|
gotIPHair = map[string]string{} // server -> IP for second UDP4 for hairpinning
|
|
gotIP4 string
|
|
)
|
|
add := func(server, ip string, d time.Duration) {
|
|
logf("%s says we are %s (in %v)", server, ip, d)
|
|
|
|
mu.Lock()
|
|
defer mu.Unlock()
|
|
ret.UDP = true
|
|
ret.DERPLatency[server] = d
|
|
if strings.Contains(server, "-v6") {
|
|
ret.IPv6 = true
|
|
} else {
|
|
// IPv4
|
|
if gotIP4 == "" {
|
|
gotIP4 = ip
|
|
} else {
|
|
if gotIP4 != ip {
|
|
ret.MappingVariesByDestIP.Set(true)
|
|
} else if ret.MappingVariesByDestIP == "" {
|
|
ret.MappingVariesByDestIP.Set(false)
|
|
}
|
|
}
|
|
}
|
|
gotIP[server] = ip
|
|
|
|
if ret.PreferredDERP == 0 {
|
|
ret.PreferredDERP = derpIndexOfSTUNHostPort(server)
|
|
}
|
|
}
|
|
addHair := func(server, ip string, d time.Duration) {
|
|
mu.Lock()
|
|
defer mu.Unlock()
|
|
gotIPHair[server] = ip
|
|
}
|
|
|
|
var pc4, pc6 net.PacketConn
|
|
|
|
pc4, err = net.ListenPacket("udp4", ":0")
|
|
if err != nil {
|
|
logf("udp4: %v", err)
|
|
return nil, err
|
|
}
|
|
go closeOnCtx(pc4)
|
|
|
|
// And a second UDP4 socket to check hairpinning.
|
|
pc4Hair, err := net.ListenPacket("udp4", ":0")
|
|
if err != nil {
|
|
logf("udp4: %v", err)
|
|
return nil, err
|
|
}
|
|
go closeOnCtx(pc4Hair)
|
|
|
|
if v6 {
|
|
pc6, err = net.ListenPacket("udp6", ":0")
|
|
if err != nil {
|
|
logf("udp6: %v", err)
|
|
v6 = false
|
|
} else {
|
|
go closeOnCtx(pc6)
|
|
}
|
|
}
|
|
|
|
reader := func(s *stunner.Stunner, pc net.PacketConn, maxReads int) {
|
|
var buf [64 << 10]byte
|
|
for i := 0; i < maxReads; i++ {
|
|
n, addr, err := pc.ReadFrom(buf[:])
|
|
if err != nil {
|
|
if ctx.Err() != nil {
|
|
return
|
|
}
|
|
logf("ReadFrom: %v", err)
|
|
return
|
|
}
|
|
ua, ok := addr.(*net.UDPAddr)
|
|
if !ok {
|
|
logf("ReadFrom: unexpected addr %T", addr)
|
|
continue
|
|
}
|
|
s.Receive(buf[:n], ua)
|
|
}
|
|
|
|
}
|
|
|
|
var grp errgroup.Group
|
|
|
|
const unlimited = 9999 // effectively, closed on cancel anyway
|
|
s4 := &stunner.Stunner{
|
|
Send: pc4.WriteTo,
|
|
Endpoint: add,
|
|
Servers: stunServers,
|
|
Logf: logf,
|
|
DNSCache: dnscache.Get(),
|
|
}
|
|
grp.Go(func() error { return s4.Run(ctx) })
|
|
go reader(s4, pc4, unlimited)
|
|
|
|
s4Hair := &stunner.Stunner{
|
|
Send: pc4Hair.WriteTo,
|
|
Endpoint: addHair,
|
|
Servers: stunServers,
|
|
Logf: logf,
|
|
DNSCache: dnscache.Get(),
|
|
}
|
|
grp.Go(func() error { return s4Hair.Run(ctx) })
|
|
go reader(s4Hair, pc4Hair, 2)
|
|
|
|
if v6 {
|
|
s6 := &stunner.Stunner{
|
|
Endpoint: add,
|
|
Send: pc6.WriteTo,
|
|
Servers: stunServers6,
|
|
Logf: logf,
|
|
OnlyIPv6: true,
|
|
DNSCache: dnscache.Get(),
|
|
}
|
|
grp.Go(func() error { return s6.Run(ctx) })
|
|
go reader(s6, pc6, unlimited)
|
|
}
|
|
|
|
err = grp.Wait()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
mu.Lock()
|
|
defer mu.Unlock()
|
|
|
|
// Check hairpinning.
|
|
if ret.MappingVariesByDestIP == "false" {
|
|
hairIPStr, hairPortStr, _ := net.SplitHostPort(gotIPHair["derp1.tailscale.com:3478"])
|
|
hairIP := net.ParseIP(hairIPStr)
|
|
hairPort, _ := strconv.Atoi(hairPortStr)
|
|
if hairIP != nil && hairPort != 0 {
|
|
tx := stun.NewTxID() // random payload
|
|
pc4.WriteTo(tx[:], &net.UDPAddr{IP: hairIP, Port: hairPort})
|
|
var got stun.TxID
|
|
pc4Hair.SetReadDeadline(time.Now().Add(1 * time.Second))
|
|
_, _, err := pc4Hair.ReadFrom(got[:])
|
|
ret.HairPinning.Set(err == nil && got == tx)
|
|
}
|
|
}
|
|
|
|
// TODO: if UDP is blocked, try to measure TCP connect times
|
|
// to DERP nodes instead? So UDP-blocked users still get a
|
|
// decent DERP node, rather than being randomly assigned to
|
|
// the other side of the planet? Or try ICMP? (likely also
|
|
// blocked?)
|
|
|
|
return ret.Clone(), nil
|
|
}
|
|
|
|
// derpIndexOfSTUNHostPort extracts the derp indes from a STUN host:port like
|
|
// "derp1-v6.tailscale.com:3478" or "derp2.tailscale.com:3478".
|
|
// It returns 0 on unexpected input.
|
|
func derpIndexOfSTUNHostPort(hp string) int {
|
|
hp = strings.TrimSuffix(hp, ".tailscale.com:3478")
|
|
hp = strings.TrimSuffix(hp, "-v6")
|
|
hp = strings.TrimPrefix(hp, "derp")
|
|
n, _ := strconv.Atoi(hp)
|
|
return n // 0 on error is okay
|
|
}
|