mirror of
https://github.com/juanfont/headscale.git
synced 2024-12-22 07:57:34 +00:00
Resolve merge conflict
This commit is contained in:
parent
5bfcf5c917
commit
9698abbfd5
@ -2,6 +2,7 @@
|
|||||||
// ignoring it let us speed up the integration test
|
// ignoring it let us speed up the integration test
|
||||||
// development
|
// development
|
||||||
integration_test.go
|
integration_test.go
|
||||||
|
integration_test/
|
||||||
|
|
||||||
Dockerfile*
|
Dockerfile*
|
||||||
docker-compose*
|
docker-compose*
|
||||||
|
125
api.go
125
api.go
@ -286,21 +286,6 @@ func (h *Headscale) PollNetMapHandler(c *gin.Context) {
|
|||||||
}
|
}
|
||||||
h.db.Save(&m)
|
h.db.Save(&m)
|
||||||
|
|
||||||
update := make(chan []byte, 1)
|
|
||||||
|
|
||||||
pollData := make(chan []byte, 1)
|
|
||||||
defer close(pollData)
|
|
||||||
|
|
||||||
cancelKeepAlive := make(chan []byte, 1)
|
|
||||||
defer close(cancelKeepAlive)
|
|
||||||
|
|
||||||
log.Trace().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("id", c.Param("id")).
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Msg("Storing update channel")
|
|
||||||
h.clientsPolling.Store(m.ID, update)
|
|
||||||
|
|
||||||
data, err := h.getMapResponse(mKey, req, m)
|
data, err := h.getMapResponse(mKey, req, m)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error().
|
log.Error().
|
||||||
@ -351,6 +336,23 @@ func (h *Headscale) PollNetMapHandler(c *gin.Context) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Only create update channel if it has not been created
|
||||||
|
var update chan []byte
|
||||||
|
log.Trace().
|
||||||
|
Str("handler", "PollNetMap").
|
||||||
|
Str("id", c.Param("id")).
|
||||||
|
Str("machine", m.Name).
|
||||||
|
Msg("Creating or loading update channel")
|
||||||
|
if result, ok := h.clientsPolling.LoadOrStore(m.ID, make(chan []byte, 1)); ok {
|
||||||
|
update = result.(chan []byte)
|
||||||
|
}
|
||||||
|
|
||||||
|
pollData := make(chan []byte, 1)
|
||||||
|
defer close(pollData)
|
||||||
|
|
||||||
|
cancelKeepAlive := make(chan []byte, 1)
|
||||||
|
defer close(cancelKeepAlive)
|
||||||
|
|
||||||
log.Info().
|
log.Info().
|
||||||
Str("handler", "PollNetMap").
|
Str("handler", "PollNetMap").
|
||||||
Str("machine", m.Name).
|
Str("machine", m.Name).
|
||||||
@ -365,87 +367,15 @@ func (h *Headscale) PollNetMapHandler(c *gin.Context) {
|
|||||||
Str("handler", "PollNetMap").
|
Str("handler", "PollNetMap").
|
||||||
Str("machine", m.Name).
|
Str("machine", m.Name).
|
||||||
Msg("Notifying peers")
|
Msg("Notifying peers")
|
||||||
peers, _ := h.getPeers(m)
|
// TODO: Why does this block?
|
||||||
for _, p := range *peers {
|
go h.notifyChangesToPeers(&m)
|
||||||
pUp, ok := h.clientsPolling.Load(uint64(p.ID))
|
|
||||||
if ok {
|
|
||||||
log.Info().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Str("peer", m.Name).
|
|
||||||
Str("address", p.Addresses[0].String()).
|
|
||||||
Msgf("Notifying peer %s (%s)", p.Name, p.Addresses[0])
|
|
||||||
pUp.(chan []byte) <- []byte{}
|
|
||||||
} else {
|
|
||||||
log.Info().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Str("peer", m.Name).
|
|
||||||
Msgf("Peer %s does not appear to be polling", p.Name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
go h.keepAlive(cancelKeepAlive, pollData, mKey, req, m)
|
h.PollNetMapStream(c, m, req, mKey, pollData, update, cancelKeepAlive)
|
||||||
|
log.Trace().
|
||||||
c.Stream(func(w io.Writer) bool {
|
Str("handler", "PollNetMap").
|
||||||
select {
|
Str("id", c.Param("id")).
|
||||||
case data := <-pollData:
|
Str("machine", m.Name).
|
||||||
log.Trace().
|
Msg("Finished stream, closing PollNetMap session")
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Int("bytes", len(data)).
|
|
||||||
Msg("Sending data")
|
|
||||||
_, err := w.Write(data)
|
|
||||||
if err != nil {
|
|
||||||
log.Error().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Err(err).
|
|
||||||
Msg("Cannot write data")
|
|
||||||
}
|
|
||||||
now := time.Now().UTC()
|
|
||||||
m.LastSeen = &now
|
|
||||||
h.db.Save(&m)
|
|
||||||
return true
|
|
||||||
|
|
||||||
case <-update:
|
|
||||||
log.Debug().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Msg("Received a request for update")
|
|
||||||
data, err := h.getMapResponse(mKey, req, m)
|
|
||||||
if err != nil {
|
|
||||||
log.Error().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Err(err).
|
|
||||||
Msg("Could not get the map update")
|
|
||||||
}
|
|
||||||
_, err = w.Write(*data)
|
|
||||||
if err != nil {
|
|
||||||
log.Error().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Err(err).
|
|
||||||
Msg("Could not write the map response")
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
|
|
||||||
case <-c.Request.Context().Done():
|
|
||||||
log.Info().
|
|
||||||
Str("handler", "PollNetMap").
|
|
||||||
Str("machine", m.Name).
|
|
||||||
Msg("The client has closed the connection")
|
|
||||||
now := time.Now().UTC()
|
|
||||||
m.LastSeen = &now
|
|
||||||
h.db.Save(&m)
|
|
||||||
cancelKeepAlive <- []byte{}
|
|
||||||
h.clientsPolling.Delete(m.ID)
|
|
||||||
close(update)
|
|
||||||
return false
|
|
||||||
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (h *Headscale) keepAlive(cancel chan []byte, pollData chan []byte, mKey wgkey.Key, req tailcfg.MapRequest, m Machine) {
|
func (h *Headscale) keepAlive(cancel chan []byte, pollData chan []byte, mKey wgkey.Key, req tailcfg.MapRequest, m Machine) {
|
||||||
@ -514,10 +444,15 @@ func (h *Headscale) getMapResponse(mKey wgkey.Key, req tailcfg.MapRequest, m Mac
|
|||||||
DERPMap: h.cfg.DerpMap,
|
DERPMap: h.cfg.DerpMap,
|
||||||
UserProfiles: []tailcfg.UserProfile{profile},
|
UserProfiles: []tailcfg.UserProfile{profile},
|
||||||
}
|
}
|
||||||
|
log.Trace().
|
||||||
|
Str("func", "getMapResponse").
|
||||||
|
Str("machine", req.Hostinfo.Hostname).
|
||||||
|
Msgf("Generated map response: %s", tailMapResponseToString(resp))
|
||||||
|
|
||||||
var respBody []byte
|
var respBody []byte
|
||||||
if req.Compress == "zstd" {
|
if req.Compress == "zstd" {
|
||||||
src, _ := json.Marshal(resp)
|
src, _ := json.Marshal(resp)
|
||||||
|
|
||||||
encoder, _ := zstd.NewWriter(nil)
|
encoder, _ := zstd.NewWriter(nil)
|
||||||
srcCompressed := encoder.EncodeAll(src, nil)
|
srcCompressed := encoder.EncodeAll(src, nil)
|
||||||
respBody, err = encodeMsg(srcCompressed, &mKey, h.privateKey)
|
respBody, err = encodeMsg(srcCompressed, &mKey, h.privateKey)
|
||||||
|
5
app.go
5
app.go
@ -135,10 +135,7 @@ func (h *Headscale) expireEphemeralNodesWorker() {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error().Err(err).Str("machine", m.Name).Msg("🤮 Cannot delete ephemeral machine from the database")
|
log.Error().Err(err).Str("machine", m.Name).Msg("🤮 Cannot delete ephemeral machine from the database")
|
||||||
}
|
}
|
||||||
err = h.notifyChangesToPeers(&m)
|
h.notifyChangesToPeers(&m)
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
1
go.mod
1
go.mod
@ -22,6 +22,7 @@ require (
|
|||||||
github.com/rs/zerolog v1.23.0 // indirect
|
github.com/rs/zerolog v1.23.0 // indirect
|
||||||
github.com/spf13/cobra v1.1.3
|
github.com/spf13/cobra v1.1.3
|
||||||
github.com/spf13/viper v1.8.1
|
github.com/spf13/viper v1.8.1
|
||||||
|
github.com/stretchr/testify v1.7.0 // indirect
|
||||||
github.com/tailscale/hujson v0.0.0-20200924210142-dde312d0d6a2
|
github.com/tailscale/hujson v0.0.0-20200924210142-dde312d0d6a2
|
||||||
github.com/xeipuuv/gojsonpointer v0.0.0-20190905194746-02993c407bfb // indirect
|
github.com/xeipuuv/gojsonpointer v0.0.0-20190905194746-02993c407bfb // indirect
|
||||||
golang.org/x/crypto v0.0.0-20210616213533-5ff15b29337e
|
golang.org/x/crypto v0.0.0-20210616213533-5ff15b29337e
|
||||||
|
1
go.sum
1
go.sum
@ -817,6 +817,7 @@ github.com/streadway/handy v0.0.0-20190108123426-d5acb3125c2a/go.mod h1:qNTQ5P5J
|
|||||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||||
github.com/stretchr/objx v0.1.1/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
github.com/stretchr/objx v0.1.1/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||||
github.com/stretchr/objx v0.2.0/go.mod h1:qt09Ya8vawLte6SNmTgCsAVtYtaKzEcn8ATUoHMkEqE=
|
github.com/stretchr/objx v0.2.0/go.mod h1:qt09Ya8vawLte6SNmTgCsAVtYtaKzEcn8ATUoHMkEqE=
|
||||||
|
github.com/stretchr/objx v0.3.0 h1:NGXK3lHquSN08v5vWalVI/L8XU9hdzE/G6xsrze47As=
|
||||||
github.com/stretchr/objx v0.3.0/go.mod h1:qt09Ya8vawLte6SNmTgCsAVtYtaKzEcn8ATUoHMkEqE=
|
github.com/stretchr/objx v0.3.0/go.mod h1:qt09Ya8vawLte6SNmTgCsAVtYtaKzEcn8ATUoHMkEqE=
|
||||||
github.com/stretchr/testify v1.2.1/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs=
|
github.com/stretchr/testify v1.2.1/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs=
|
||||||
github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs=
|
github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs=
|
||||||
|
@ -9,17 +9,24 @@ import (
|
|||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/ory/dockertest/v3"
|
"github.com/ory/dockertest/v3"
|
||||||
"github.com/ory/dockertest/v3/docker"
|
"github.com/ory/dockertest/v3/docker"
|
||||||
"inet.af/netaddr"
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/suite"
|
||||||
|
|
||||||
"gopkg.in/check.v1"
|
"inet.af/netaddr"
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = check.Suite(&IntegrationSuite{})
|
type IntegrationTestSuite struct {
|
||||||
|
suite.Suite
|
||||||
|
}
|
||||||
|
|
||||||
type IntegrationSuite struct{}
|
func TestIntegrationTestSuite(t *testing.T) {
|
||||||
|
suite.Run(t, new(IntegrationTestSuite))
|
||||||
|
}
|
||||||
|
|
||||||
var integrationTmpDir string
|
var integrationTmpDir string
|
||||||
var ih Headscale
|
var ih Headscale
|
||||||
@ -27,7 +34,7 @@ var ih Headscale
|
|||||||
var pool dockertest.Pool
|
var pool dockertest.Pool
|
||||||
var network dockertest.Network
|
var network dockertest.Network
|
||||||
var headscale dockertest.Resource
|
var headscale dockertest.Resource
|
||||||
var tailscaleCount int = 10
|
var tailscaleCount int = 5
|
||||||
var tailscales map[string]dockertest.Resource
|
var tailscales map[string]dockertest.Resource
|
||||||
|
|
||||||
func executeCommand(resource *dockertest.Resource, cmd []string) (string, error) {
|
func executeCommand(resource *dockertest.Resource, cmd []string) (string, error) {
|
||||||
@ -63,7 +70,7 @@ func dockerRestartPolicy(config *docker.HostConfig) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
func (s *IntegrationTestSuite) SetupSuite() {
|
||||||
var err error
|
var err error
|
||||||
h = Headscale{
|
h = Headscale{
|
||||||
dbType: "sqlite3",
|
dbType: "sqlite3",
|
||||||
@ -104,8 +111,7 @@ func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
|||||||
fmt.Sprintf("%s/derp.yaml:/etc/headscale/derp.yaml", currentPath),
|
fmt.Sprintf("%s/derp.yaml:/etc/headscale/derp.yaml", currentPath),
|
||||||
},
|
},
|
||||||
Networks: []*dockertest.Network{&network},
|
Networks: []*dockertest.Network{&network},
|
||||||
// Cmd: []string{"sleep", "3600"},
|
Cmd: []string{"headscale", "serve"},
|
||||||
Cmd: []string{"headscale", "serve"},
|
|
||||||
PortBindings: map[docker.Port][]docker.PortBinding{
|
PortBindings: map[docker.Port][]docker.PortBinding{
|
||||||
"8080/tcp": []docker.PortBinding{{HostPort: "8080"}},
|
"8080/tcp": []docker.PortBinding{{HostPort: "8080"}},
|
||||||
},
|
},
|
||||||
@ -127,10 +133,8 @@ func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
|||||||
tailscaleOptions := &dockertest.RunOptions{
|
tailscaleOptions := &dockertest.RunOptions{
|
||||||
Name: hostname,
|
Name: hostname,
|
||||||
Networks: []*dockertest.Network{&network},
|
Networks: []*dockertest.Network{&network},
|
||||||
// Make the container run until killed
|
Cmd: []string{"tailscaled", "--tun=userspace-networking", "--socks5-server=localhost:1055"},
|
||||||
// Cmd: []string{"sleep", "3600"},
|
Env: []string{},
|
||||||
Cmd: []string{"tailscaled", "--tun=userspace-networking", "--socks5-server=localhost:1055"},
|
|
||||||
Env: []string{},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if pts, err := pool.BuildAndRunWithBuildOptions(tailscaleBuildOptions, tailscaleOptions, dockerRestartPolicy); err == nil {
|
if pts, err := pool.BuildAndRunWithBuildOptions(tailscaleBuildOptions, tailscaleOptions, dockerRestartPolicy); err == nil {
|
||||||
@ -141,6 +145,7 @@ func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
|||||||
fmt.Printf("Created %s container\n", hostname)
|
fmt.Printf("Created %s container\n", hostname)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TODO: Replace this logic with something that can be detected on Github Actions
|
||||||
fmt.Println("Waiting for headscale to be ready")
|
fmt.Println("Waiting for headscale to be ready")
|
||||||
hostEndpoint := fmt.Sprintf("localhost:%s", headscale.GetPort("8080/tcp"))
|
hostEndpoint := fmt.Sprintf("localhost:%s", headscale.GetPort("8080/tcp"))
|
||||||
|
|
||||||
@ -164,14 +169,14 @@ func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
|||||||
&headscale,
|
&headscale,
|
||||||
[]string{"headscale", "namespaces", "create", "test"},
|
[]string{"headscale", "namespaces", "create", "test"},
|
||||||
)
|
)
|
||||||
c.Assert(err, check.IsNil)
|
assert.Nil(s.T(), err)
|
||||||
|
|
||||||
fmt.Println("Creating pre auth key")
|
fmt.Println("Creating pre auth key")
|
||||||
authKey, err := executeCommand(
|
authKey, err := executeCommand(
|
||||||
&headscale,
|
&headscale,
|
||||||
[]string{"headscale", "-n", "test", "preauthkeys", "create", "--reusable", "--expiration", "24h"},
|
[]string{"headscale", "-n", "test", "preauthkeys", "create", "--reusable", "--expiration", "24h"},
|
||||||
)
|
)
|
||||||
c.Assert(err, check.IsNil)
|
assert.Nil(s.T(), err)
|
||||||
|
|
||||||
headscaleEndpoint := fmt.Sprintf("http://headscale:%s", headscale.GetPort("8080/tcp"))
|
headscaleEndpoint := fmt.Sprintf("http://headscale:%s", headscale.GetPort("8080/tcp"))
|
||||||
|
|
||||||
@ -186,12 +191,16 @@ func (s *IntegrationSuite) SetUpSuite(c *check.C) {
|
|||||||
command,
|
command,
|
||||||
)
|
)
|
||||||
fmt.Println("tailscale result: ", result)
|
fmt.Println("tailscale result: ", result)
|
||||||
c.Assert(err, check.IsNil)
|
assert.Nil(s.T(), err)
|
||||||
fmt.Printf("%s joined\n", hostname)
|
fmt.Printf("%s joined\n", hostname)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// The nodes need a bit of time to get their updated maps from headscale
|
||||||
|
// TODO: See if we can have a more deterministic wait here.
|
||||||
|
time.Sleep(20 * time.Second)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *IntegrationSuite) TearDownSuite(c *check.C) {
|
func (s *IntegrationTestSuite) TearDownSuite() {
|
||||||
if err := pool.Purge(&headscale); err != nil {
|
if err := pool.Purge(&headscale); err != nil {
|
||||||
log.Printf("Could not purge resource: %s\n", err)
|
log.Printf("Could not purge resource: %s\n", err)
|
||||||
}
|
}
|
||||||
@ -207,21 +216,102 @@ func (s *IntegrationSuite) TearDownSuite(c *check.C) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *IntegrationSuite) TestListNodes(c *check.C) {
|
func (s *IntegrationTestSuite) TestListNodes() {
|
||||||
fmt.Println("Listing nodes")
|
fmt.Println("Listing nodes")
|
||||||
result, err := executeCommand(
|
result, err := executeCommand(
|
||||||
&headscale,
|
&headscale,
|
||||||
[]string{"headscale", "-n", "test", "nodes", "list"},
|
[]string{"headscale", "-n", "test", "nodes", "list"},
|
||||||
)
|
)
|
||||||
c.Assert(err, check.IsNil)
|
assert.Nil(s.T(), err)
|
||||||
|
|
||||||
|
fmt.Printf("List nodes: \n%s\n", result)
|
||||||
|
|
||||||
|
// Chck that the correct count of host is present in node list
|
||||||
|
lines := strings.Split(result, "\n")
|
||||||
|
assert.Equal(s.T(), len(tailscales), len(lines)-2)
|
||||||
|
|
||||||
for hostname, _ := range tailscales {
|
for hostname, _ := range tailscales {
|
||||||
c.Assert(strings.Contains(result, hostname), check.Equals, true)
|
assert.Contains(s.T(), result, hostname)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *IntegrationSuite) TestGetIpAddresses(c *check.C) {
|
func (s *IntegrationTestSuite) TestGetIpAddresses() {
|
||||||
ipPrefix := netaddr.MustParseIPPrefix("100.64.0.0/10")
|
ipPrefix := netaddr.MustParseIPPrefix("100.64.0.0/10")
|
||||||
|
ips, err := getIPs()
|
||||||
|
assert.Nil(s.T(), err)
|
||||||
|
|
||||||
|
for hostname, _ := range tailscales {
|
||||||
|
s.T().Run(hostname, func(t *testing.T) {
|
||||||
|
ip := ips[hostname]
|
||||||
|
|
||||||
|
fmt.Printf("IP for %s: %s\n", hostname, ip)
|
||||||
|
|
||||||
|
// c.Assert(ip.Valid(), check.IsTrue)
|
||||||
|
assert.True(t, ip.Is4())
|
||||||
|
assert.True(t, ipPrefix.Contains(ip))
|
||||||
|
|
||||||
|
ips[hostname] = ip
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *IntegrationTestSuite) TestStatus() {
|
||||||
|
ips, err := getIPs()
|
||||||
|
assert.Nil(s.T(), err)
|
||||||
|
|
||||||
|
for hostname, tailscale := range tailscales {
|
||||||
|
s.T().Run(hostname, func(t *testing.T) {
|
||||||
|
command := []string{"tailscale", "status"}
|
||||||
|
|
||||||
|
fmt.Printf("Getting status for %s\n", hostname)
|
||||||
|
result, err := executeCommand(
|
||||||
|
&tailscale,
|
||||||
|
command,
|
||||||
|
)
|
||||||
|
assert.Nil(t, err)
|
||||||
|
// fmt.Printf("Status for %s: %s", hostname, result)
|
||||||
|
|
||||||
|
// Check if we have as many nodes in status
|
||||||
|
// as we have IPs/tailscales
|
||||||
|
lines := strings.Split(result, "\n")
|
||||||
|
assert.Equal(t, len(ips), len(lines)-1)
|
||||||
|
assert.Equal(t, len(tailscales), len(lines)-1)
|
||||||
|
|
||||||
|
// Check that all hosts is present in all hosts status
|
||||||
|
for ipHostname, ip := range ips {
|
||||||
|
assert.Contains(t, result, ip.String())
|
||||||
|
assert.Contains(t, result, ipHostname)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// func (s *IntegrationTestSuite) TestPingAllPeers() {
|
||||||
|
// ips, err := getIPs()
|
||||||
|
// assert.Nil(s.T(), err)
|
||||||
|
//
|
||||||
|
// for hostname, tailscale := range tailscales {
|
||||||
|
// for peername, ip := range ips {
|
||||||
|
// s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
|
||||||
|
// // We currently cant ping ourselves, so skip that.
|
||||||
|
// if peername != hostname {
|
||||||
|
// command := []string{"tailscale", "ping", "--timeout=1s", "--c=1", ip.String()}
|
||||||
|
//
|
||||||
|
// fmt.Printf("Pinging from %s (%s) to %s (%s)\n", hostname, ips[hostname], peername, ip)
|
||||||
|
// result, err := executeCommand(
|
||||||
|
// &tailscale,
|
||||||
|
// command,
|
||||||
|
// )
|
||||||
|
// assert.Nil(t, err)
|
||||||
|
// fmt.Printf("Result for %s: %s\n", hostname, result)
|
||||||
|
// assert.Contains(t, result, "pong")
|
||||||
|
// }
|
||||||
|
// })
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
func getIPs() (map[string]netaddr.IP, error) {
|
||||||
ips := make(map[string]netaddr.IP)
|
ips := make(map[string]netaddr.IP)
|
||||||
for hostname, tailscale := range tailscales {
|
for hostname, tailscale := range tailscales {
|
||||||
command := []string{"tailscale", "ip"}
|
command := []string{"tailscale", "ip"}
|
||||||
@ -230,17 +320,16 @@ func (s *IntegrationSuite) TestGetIpAddresses(c *check.C) {
|
|||||||
&tailscale,
|
&tailscale,
|
||||||
command,
|
command,
|
||||||
)
|
)
|
||||||
c.Assert(err, check.IsNil)
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
ip, err := netaddr.ParseIP(strings.TrimSuffix(result, "\n"))
|
ip, err := netaddr.ParseIP(strings.TrimSuffix(result, "\n"))
|
||||||
c.Assert(err, check.IsNil)
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
fmt.Printf("IP for %s: %s", hostname, result)
|
}
|
||||||
|
|
||||||
// c.Assert(ip.Valid(), check.IsTrue)
|
|
||||||
c.Assert(ip.Is4(), check.Equals, true)
|
|
||||||
c.Assert(ipPrefix.Contains(ip), check.Equals, true)
|
|
||||||
|
|
||||||
ips[hostname] = ip
|
ips[hostname] = ip
|
||||||
}
|
}
|
||||||
|
return ips, nil
|
||||||
}
|
}
|
||||||
|
12
machine.go
12
machine.go
@ -159,6 +159,10 @@ func (m Machine) toNode() (*tailcfg.Node, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (h *Headscale) getPeers(m Machine) (*[]*tailcfg.Node, error) {
|
func (h *Headscale) getPeers(m Machine) (*[]*tailcfg.Node, error) {
|
||||||
|
log.Trace().
|
||||||
|
Str("func", "getPeers").
|
||||||
|
Str("machine", m.Name).
|
||||||
|
Msg("Finding peers")
|
||||||
machines := []Machine{}
|
machines := []Machine{}
|
||||||
if err := h.db.Where("namespace_id = ? AND machine_key <> ? AND registered",
|
if err := h.db.Where("namespace_id = ? AND machine_key <> ? AND registered",
|
||||||
m.NamespaceID, m.MachineKey).Find(&machines).Error; err != nil {
|
m.NamespaceID, m.MachineKey).Find(&machines).Error; err != nil {
|
||||||
@ -175,6 +179,11 @@ func (h *Headscale) getPeers(m Machine) (*[]*tailcfg.Node, error) {
|
|||||||
peers = append(peers, peer)
|
peers = append(peers, peer)
|
||||||
}
|
}
|
||||||
sort.Slice(peers, func(i, j int) bool { return peers[i].ID < peers[j].ID })
|
sort.Slice(peers, func(i, j int) bool { return peers[i].ID < peers[j].ID })
|
||||||
|
|
||||||
|
log.Trace().
|
||||||
|
Str("func", "getPeers").
|
||||||
|
Str("machine", m.Name).
|
||||||
|
Msgf("Found peers: %s", tailNodesToString(peers))
|
||||||
return &peers, nil
|
return &peers, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -239,7 +248,7 @@ func (m *Machine) GetHostInfo() (*tailcfg.Hostinfo, error) {
|
|||||||
return &hostinfo, nil
|
return &hostinfo, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (h *Headscale) notifyChangesToPeers(m *Machine) error {
|
func (h *Headscale) notifyChangesToPeers(m *Machine) {
|
||||||
peers, _ := h.getPeers(*m)
|
peers, _ := h.getPeers(*m)
|
||||||
for _, p := range *peers {
|
for _, p := range *peers {
|
||||||
pUp, ok := h.clientsPolling.Load(uint64(p.ID))
|
pUp, ok := h.clientsPolling.Load(uint64(p.ID))
|
||||||
@ -259,5 +268,4 @@ func (h *Headscale) notifyChangesToPeers(m *Machine) error {
|
|||||||
Msgf("Peer %s does not appear to be polling", p.Name)
|
Msgf("Peer %s does not appear to be polling", p.Name)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
@ -169,10 +169,7 @@ func (h *Headscale) checkForNamespacesPendingUpdates() {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
for _, m := range *machines {
|
for _, m := range *machines {
|
||||||
err = h.notifyChangesToPeers(&m)
|
h.notifyChangesToPeers(&m)
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
newV, err := h.getValue("namespaces_pending_updates")
|
newV, err := h.getValue("namespaces_pending_updates")
|
||||||
|
@ -49,12 +49,7 @@ func (h *Headscale) EnableNodeRoute(namespace string, nodeName string, routeStr
|
|||||||
// The peers map is stored in memory in the server process.
|
// The peers map is stored in memory in the server process.
|
||||||
// Definitely not accessible from the CLI tool.
|
// Definitely not accessible from the CLI tool.
|
||||||
// We need RPC to the server - or some kind of 'needsUpdate' field in the DB
|
// We need RPC to the server - or some kind of 'needsUpdate' field in the DB
|
||||||
peers, _ := h.getPeers(*m)
|
h.notifyChangesToPeers(m)
|
||||||
for _, p := range *peers {
|
|
||||||
if pUp, ok := h.clientsPolling.Load(uint64(p.ID)); ok {
|
|
||||||
pUp.(chan []byte) <- []byte{}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return &rIP, nil
|
return &rIP, nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
17
utils.go
17
utils.go
@ -10,9 +10,11 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
|
"strings"
|
||||||
|
|
||||||
"golang.org/x/crypto/nacl/box"
|
"golang.org/x/crypto/nacl/box"
|
||||||
"inet.af/netaddr"
|
"inet.af/netaddr"
|
||||||
|
"tailscale.com/tailcfg"
|
||||||
"tailscale.com/types/wgkey"
|
"tailscale.com/types/wgkey"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -58,6 +60,7 @@ func encode(v interface{}, pubKey *wgkey.Key, privKey *wgkey.Private) ([]byte, e
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return encodeMsg(b, pubKey, privKey)
|
return encodeMsg(b, pubKey, privKey)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -139,3 +142,17 @@ func containsIPs(ips []netaddr.IP, ip netaddr.IP) bool {
|
|||||||
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func tailNodesToString(nodes []*tailcfg.Node) string {
|
||||||
|
temp := make([]string, len(nodes))
|
||||||
|
|
||||||
|
for index, node := range nodes {
|
||||||
|
temp[index] = node.Name
|
||||||
|
}
|
||||||
|
|
||||||
|
return fmt.Sprintf("[ %s ](%d)", strings.Join(temp, ", "), len(temp))
|
||||||
|
}
|
||||||
|
|
||||||
|
func tailMapResponseToString(resp tailcfg.MapResponse) string {
|
||||||
|
return fmt.Sprintf("{ Node: %s, Peers: %s }", resp.Node.Name, tailNodesToString(resp.Peers))
|
||||||
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user