tailscale/derp
Brad Fitzpatrick 3485e4bf5a derp: make RunConnectionLoop funcs take Messages, support PeerPresentFlags
PeerPresentFlags was added in 5ffb2668ef but wasn't plumbed through to
the RunConnectionLoop. Rather than add yet another parameter (as
IP:port was added earlier), pass in the raw PeerPresentMessage and
PeerGoneMessage struct values, which are the same things, plus two
fields: PeerGoneReasonType for gone and the PeerPresentFlags from
5ffb2668ef.

Updates tailscale/corp#17816

Change-Id: Ib19d9f95353651ada90656071fc3656cf58b7987
Signed-off-by: Brad Fitzpatrick <bradfitz@tailscale.com>
2024-06-25 09:47:25 -07:00
..
derphttp derp: make RunConnectionLoop funcs take Messages, support PeerPresentFlags 2024-06-25 09:47:25 -07:00
testdata derp: add debug traffic handler 2021-06-18 15:47:55 -07:00
xdp cmd/xdpderper,derp/xdp: implement mode that drops STUN packets (#12527) 2024-06-18 14:06:00 -07:00
derp_client.go derp: add PeerPresentFlags bitmask to Watch messages 2024-06-22 20:38:25 -07:00
derp_server_default.go derp: remove stats goroutine, use a timer 2024-05-14 11:05:11 -06:00
derp_server_linux.go derp: remove stats goroutine, use a timer 2024-05-14 11:05:11 -06:00
derp_server.go derp: redo, simplify how mesh update writes are queued/written 2024-06-24 21:42:14 -07:00
derp_test.go derp: add PeerPresentFlags bitmask to Watch messages 2024-06-22 20:38:25 -07:00
derp.go derp: make RunConnectionLoop funcs take Messages, support PeerPresentFlags 2024-06-25 09:47:25 -07:00
dropreason_string.go derp: add some guardrails for derpReason metrics getting out of sync 2024-05-31 10:06:42 -07:00
README.md derp: add a README.md with some docs 2023-05-02 13:42:25 -07:00

DERP

This directory (and subdirectories) contain the DERP code. The server itself is in ../cmd/derper.

DERP is a packet relay system (client and servers) where peers are addressed using WireGuard public keys instead of IP addresses.

It relays two types of packets:

  • "Disco" discovery messages (see ../disco) as the a side channel during NAT traversal.

  • Encrypted WireGuard packets as the fallback of last resort when UDP is blocked or NAT traversal fails.

DERP Map

Each client receives a "DERP Map" from the coordination server describing the DERP servers the client should try to use.

The client picks its home "DERP home" based on latency. This is done to keep costs low by avoid using cloud load balancers (pricey) or anycast, which would necessarily require server-side routing between DERP regions.

Clients pick their DERP home and report it to the coordination server which shares it to all the peers in the tailnet. When a peer wants to send a packet and it doesn't already have a WireGuard session open, it sends disco messages (some direct, and some over DERP), trying to do the NAT traversal. The client will make connections to multiple DERP regions as needed. Only the DERP home region connection needs to be alive forever.

DERP Regions

Tailscale runs 1 or more DERP nodes (instances of cmd/derper) in various geographic regions to make sure users have low latency to their DERP home.

Regions generally have multiple nodes per region "meshed" (routing to each other) together for redundancy: it allows for cloud failures or upgrades without kicking users out to a higher latency region. Instead, clients will reconnect to the next node in the region. Each node in the region is required to to be meshed with every other node in the region and forward packets to the other nodes in the region. Packets are forwarded only one hop within the region. There is no routing between regions. The assumption is that the mesh TCP connections are over a VPC that's very fast, low latency, and not charged per byte. The coordination server assigns the list of nodes in a region as a function of the tailnet, so all nodes within a tailnet should generally be on the same node and not require forwarding. Only after a failure do clients of a particular tailnet get split between nodes in a region and require inter-node forwarding. But over time it balances back out. There's also an admin-only DERP frame type to force close the TCP connection of a particular client to force them to reconnect to their primary if the operator wants to force things to balance out sooner. (Using the (*derphttp.Client).ClosePeer method, as used by Tailscale's internal rarely-used cmd/derpprune maintenance tool)

We generally run a minimum of three nodes in a region not for quorum reasons (there's no voting) but just because two is too uncomfortably few for cascading failure reasons: if you're running two nodes at 51% load (CPU, memory, etc) and then one fails, that makes the second one fail. With three or more nodes, you can run each node a bit hotter.