mirror of
https://github.com/juanfont/headscale.git
synced 2024-12-25 01:07:52 +00:00
Merge pull request #897 from kradalby/integration-remove-v1-genera
Remove v1 general integration tests
This commit is contained in:
commit
c1c22a4b51
35
.github/workflows/test-integration-general.yml
vendored
35
.github/workflows/test-integration-general.yml
vendored
@ -1,35 +0,0 @@
|
|||||||
name: CI
|
|
||||||
|
|
||||||
on: [pull_request]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
integration-test-general:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
fetch-depth: 2
|
|
||||||
|
|
||||||
- name: Set Swap Space
|
|
||||||
uses: pierotofy/set-swap-space@master
|
|
||||||
with:
|
|
||||||
swap-size-gb: 10
|
|
||||||
|
|
||||||
- name: Get changed files
|
|
||||||
id: changed-files
|
|
||||||
uses: tj-actions/changed-files@v14.1
|
|
||||||
with:
|
|
||||||
files: |
|
|
||||||
*.nix
|
|
||||||
go.*
|
|
||||||
**/*.go
|
|
||||||
integration_test/
|
|
||||||
config-example.yaml
|
|
||||||
|
|
||||||
- uses: cachix/install-nix-action@v16
|
|
||||||
if: steps.changed-files.outputs.any_changed == 'true'
|
|
||||||
|
|
||||||
- name: Run general integration tests
|
|
||||||
if: steps.changed-files.outputs.any_changed == 'true'
|
|
||||||
run: nix develop --command -- make test_integration_general
|
|
12
Makefile
12
Makefile
@ -24,7 +24,7 @@ dev: lint test build
|
|||||||
test:
|
test:
|
||||||
@go test -short -coverprofile=coverage.out ./...
|
@go test -short -coverprofile=coverage.out ./...
|
||||||
|
|
||||||
test_integration: test_integration_cli test_integration_derp test_integration_oidc test_integration_general
|
test_integration: test_integration_cli test_integration_derp test_integration_oidc test_integration_v2_general
|
||||||
|
|
||||||
test_integration_cli:
|
test_integration_cli:
|
||||||
docker network rm $$(docker network ls --filter name=headscale --quiet) || true
|
docker network rm $$(docker network ls --filter name=headscale --quiet) || true
|
||||||
@ -46,16 +46,6 @@ test_integration_derp:
|
|||||||
-v /var/run/docker.sock:/var/run/docker.sock golang:1 \
|
-v /var/run/docker.sock:/var/run/docker.sock golang:1 \
|
||||||
go test -failfast -timeout 30m -count=1 -run IntegrationDERP ./...
|
go test -failfast -timeout 30m -count=1 -run IntegrationDERP ./...
|
||||||
|
|
||||||
test_integration_general:
|
|
||||||
docker network rm $$(docker network ls --filter name=headscale --quiet) || true
|
|
||||||
docker network create headscale-test || true
|
|
||||||
docker run -t --rm \
|
|
||||||
--network headscale-test \
|
|
||||||
-v ~/.cache/hs-integration-go:/go \
|
|
||||||
-v $$PWD:$$PWD -w $$PWD \
|
|
||||||
-v /var/run/docker.sock:/var/run/docker.sock golang:1 \
|
|
||||||
go test -failfast -timeout 30m -count=1 -run IntegrationGeneral ./...
|
|
||||||
|
|
||||||
test_integration_oidc:
|
test_integration_oidc:
|
||||||
docker network rm $$(docker network ls --filter name=headscale --quiet) || true
|
docker network rm $$(docker network ls --filter name=headscale --quiet) || true
|
||||||
docker network create headscale-test || true
|
docker network create headscale-test || true
|
||||||
|
@ -1,807 +0,0 @@
|
|||||||
//nolint
|
|
||||||
package headscale
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"context"
|
|
||||||
"encoding/json"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"log"
|
|
||||||
"net/http"
|
|
||||||
"net/netip"
|
|
||||||
"os"
|
|
||||||
"path"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
|
||||||
"testing"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
v1 "github.com/juanfont/headscale/gen/go/headscale/v1"
|
|
||||||
"github.com/ory/dockertest/v3"
|
|
||||||
"github.com/ory/dockertest/v3/docker"
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
"github.com/stretchr/testify/suite"
|
|
||||||
"tailscale.com/client/tailscale/apitype"
|
|
||||||
"tailscale.com/ipn/ipnstate"
|
|
||||||
)
|
|
||||||
|
|
||||||
type IntegrationTestSuite struct {
|
|
||||||
suite.Suite
|
|
||||||
stats *suite.SuiteInformation
|
|
||||||
|
|
||||||
pool dockertest.Pool
|
|
||||||
network dockertest.Network
|
|
||||||
headscale dockertest.Resource
|
|
||||||
saveLogs bool
|
|
||||||
|
|
||||||
namespaces map[string]TestNamespace
|
|
||||||
|
|
||||||
joinWaitGroup sync.WaitGroup
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestIntegrationGeneralTestSuite(t *testing.T) {
|
|
||||||
if testing.Short() {
|
|
||||||
t.Skip("skipping integration tests due to short flag")
|
|
||||||
}
|
|
||||||
|
|
||||||
saveLogs, err := GetEnvBool("HEADSCALE_INTEGRATION_SAVE_LOG")
|
|
||||||
if err != nil {
|
|
||||||
saveLogs = false
|
|
||||||
}
|
|
||||||
|
|
||||||
s := new(IntegrationTestSuite)
|
|
||||||
|
|
||||||
s.namespaces = map[string]TestNamespace{
|
|
||||||
"thisspace": {
|
|
||||||
count: 10,
|
|
||||||
tailscales: make(map[string]dockertest.Resource),
|
|
||||||
},
|
|
||||||
"otherspace": {
|
|
||||||
count: 2,
|
|
||||||
tailscales: make(map[string]dockertest.Resource),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
s.saveLogs = saveLogs
|
|
||||||
|
|
||||||
suite.Run(t, s)
|
|
||||||
|
|
||||||
// HandleStats, which allows us to check if we passed and save logs
|
|
||||||
// is called after TearDown, so we cannot tear down containers before
|
|
||||||
// we have potentially saved the logs.
|
|
||||||
if s.saveLogs {
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
for _, tailscale := range scales.tailscales {
|
|
||||||
if err := s.pool.Purge(&tailscale); err != nil {
|
|
||||||
log.Printf("Could not purge resource: %s\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if !s.stats.Passed() {
|
|
||||||
err := s.saveLog(&s.headscale, "test_output")
|
|
||||||
if err != nil {
|
|
||||||
log.Printf("Could not save log: %s\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if err := s.pool.Purge(&s.headscale); err != nil {
|
|
||||||
log.Printf("Could not purge resource: %s\n", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := s.network.Close(); err != nil {
|
|
||||||
log.Printf("Could not close network: %s\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) saveLog(
|
|
||||||
resource *dockertest.Resource,
|
|
||||||
basePath string,
|
|
||||||
) error {
|
|
||||||
err := os.MkdirAll(basePath, os.ModePerm)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
var stdout bytes.Buffer
|
|
||||||
var stderr bytes.Buffer
|
|
||||||
|
|
||||||
err = s.pool.Client.Logs(
|
|
||||||
docker.LogsOptions{
|
|
||||||
Context: context.TODO(),
|
|
||||||
Container: resource.Container.ID,
|
|
||||||
OutputStream: &stdout,
|
|
||||||
ErrorStream: &stderr,
|
|
||||||
Tail: "all",
|
|
||||||
RawTerminal: false,
|
|
||||||
Stdout: true,
|
|
||||||
Stderr: true,
|
|
||||||
Follow: false,
|
|
||||||
Timestamps: false,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Printf("Saving logs for %s to %s\n", resource.Container.Name, basePath)
|
|
||||||
|
|
||||||
err = os.WriteFile(
|
|
||||||
path.Join(basePath, resource.Container.Name+".stdout.log"),
|
|
||||||
[]byte(stdout.String()),
|
|
||||||
0o644,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
err = os.WriteFile(
|
|
||||||
path.Join(basePath, resource.Container.Name+".stderr.log"),
|
|
||||||
[]byte(stdout.String()),
|
|
||||||
0o644,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) Join(
|
|
||||||
endpoint, key, hostname string,
|
|
||||||
tailscale dockertest.Resource,
|
|
||||||
) {
|
|
||||||
defer s.joinWaitGroup.Done()
|
|
||||||
|
|
||||||
command := []string{
|
|
||||||
"tailscale",
|
|
||||||
"up",
|
|
||||||
"-login-server",
|
|
||||||
endpoint,
|
|
||||||
"--authkey",
|
|
||||||
key,
|
|
||||||
"--hostname",
|
|
||||||
hostname,
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Println("Join command:", command)
|
|
||||||
log.Printf("Running join command for %s\n", hostname)
|
|
||||||
_, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
log.Printf("%s joined\n", hostname)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) tailscaleContainer(
|
|
||||||
namespace, identifier, version string,
|
|
||||||
) (string, *dockertest.Resource) {
|
|
||||||
tailscaleBuildOptions := getDockerBuildOptions(version)
|
|
||||||
|
|
||||||
hostname := fmt.Sprintf(
|
|
||||||
"%s-tailscale-%s-%s",
|
|
||||||
namespace,
|
|
||||||
strings.Replace(version, ".", "-", -1),
|
|
||||||
identifier,
|
|
||||||
)
|
|
||||||
tailscaleOptions := &dockertest.RunOptions{
|
|
||||||
Name: hostname,
|
|
||||||
Networks: []*dockertest.Network{&s.network},
|
|
||||||
Cmd: []string{
|
|
||||||
"tailscaled", "--tun=tsdev",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
err := s.pool.RemoveContainerByName(hostname)
|
|
||||||
if err != nil {
|
|
||||||
s.FailNow(
|
|
||||||
fmt.Sprintf(
|
|
||||||
"Could not remove existing container before building test: %s",
|
|
||||||
err,
|
|
||||||
),
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
pts, err := s.pool.BuildAndRunWithBuildOptions(
|
|
||||||
tailscaleBuildOptions,
|
|
||||||
tailscaleOptions,
|
|
||||||
DockerRestartPolicy,
|
|
||||||
DockerAllowLocalIPv6,
|
|
||||||
DockerAllowNetworkAdministration,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
log.Fatalf("Could not start tailscale container version %s: %s", version, err)
|
|
||||||
}
|
|
||||||
log.Printf("Created %s container\n", hostname)
|
|
||||||
|
|
||||||
return hostname, pts
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) SetupSuite() {
|
|
||||||
var err error
|
|
||||||
app = Headscale{
|
|
||||||
dbType: "sqlite3",
|
|
||||||
dbString: "integration_test_db.sqlite3",
|
|
||||||
}
|
|
||||||
|
|
||||||
if ppool, err := dockertest.NewPool(""); err == nil {
|
|
||||||
s.pool = *ppool
|
|
||||||
} else {
|
|
||||||
s.FailNow(fmt.Sprintf("Could not connect to docker: %s", err), "")
|
|
||||||
}
|
|
||||||
|
|
||||||
network, err := GetFirstOrCreateNetwork(&s.pool, headscaleNetwork)
|
|
||||||
if err != nil {
|
|
||||||
s.FailNow(fmt.Sprintf("Failed to create or get network: %s", err), "")
|
|
||||||
}
|
|
||||||
s.network = network
|
|
||||||
|
|
||||||
headscaleBuildOptions := &dockertest.BuildOptions{
|
|
||||||
Dockerfile: "Dockerfile",
|
|
||||||
ContextDir: ".",
|
|
||||||
}
|
|
||||||
|
|
||||||
currentPath, err := os.Getwd()
|
|
||||||
if err != nil {
|
|
||||||
s.FailNow(fmt.Sprintf("Could not determine current path: %s", err), "")
|
|
||||||
}
|
|
||||||
|
|
||||||
headscaleOptions := &dockertest.RunOptions{
|
|
||||||
Name: headscaleHostname,
|
|
||||||
Mounts: []string{
|
|
||||||
fmt.Sprintf("%s/integration_test/etc:/etc/headscale", currentPath),
|
|
||||||
},
|
|
||||||
ExposedPorts: []string{"8080/tcp"},
|
|
||||||
PortBindings: map[docker.Port][]docker.PortBinding{
|
|
||||||
"8080/tcp": {{HostPort: "8080"}},
|
|
||||||
},
|
|
||||||
Networks: []*dockertest.Network{&s.network},
|
|
||||||
Cmd: []string{"headscale", "serve"},
|
|
||||||
}
|
|
||||||
|
|
||||||
err = s.pool.RemoveContainerByName(headscaleHostname)
|
|
||||||
if err != nil {
|
|
||||||
s.FailNow(
|
|
||||||
fmt.Sprintf(
|
|
||||||
"Could not remove existing container before building test: %s",
|
|
||||||
err,
|
|
||||||
),
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Println("Creating headscale container for core integration tests")
|
|
||||||
if pheadscale, err := s.pool.BuildAndRunWithBuildOptions(headscaleBuildOptions, headscaleOptions, DockerRestartPolicy); err == nil {
|
|
||||||
s.headscale = *pheadscale
|
|
||||||
} else {
|
|
||||||
s.FailNow(fmt.Sprintf("Could not start headscale container for core integration tests: %s", err), "")
|
|
||||||
}
|
|
||||||
log.Println("Created headscale container for core integration tests")
|
|
||||||
|
|
||||||
log.Println("Creating tailscale containers for core integration tests")
|
|
||||||
for namespace, scales := range s.namespaces {
|
|
||||||
for i := 0; i < scales.count; i++ {
|
|
||||||
version := tailscaleVersions[i%len(tailscaleVersions)]
|
|
||||||
|
|
||||||
hostname, container := s.tailscaleContainer(
|
|
||||||
namespace,
|
|
||||||
fmt.Sprint(i),
|
|
||||||
version,
|
|
||||||
)
|
|
||||||
scales.tailscales[hostname] = *container
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Println("Waiting for headscale to be ready for core integration tests")
|
|
||||||
hostEndpoint := fmt.Sprintf("%s:%s",
|
|
||||||
s.headscale.GetIPInNetwork(&s.network),
|
|
||||||
s.headscale.GetPort("8080/tcp"))
|
|
||||||
|
|
||||||
if err := s.pool.Retry(func() error {
|
|
||||||
url := fmt.Sprintf("http://%s/health", hostEndpoint)
|
|
||||||
|
|
||||||
resp, err := http.Get(url)
|
|
||||||
if err != nil {
|
|
||||||
fmt.Printf("headscale for core integration test is not ready: %s\n", err)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
if resp.StatusCode != http.StatusOK {
|
|
||||||
return fmt.Errorf("status code not OK")
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}); err != nil {
|
|
||||||
// TODO(kradalby): If we cannot access headscale, or any other fatal error during
|
|
||||||
// test setup, we need to abort and tear down. However, testify does not seem to
|
|
||||||
// support that at the moment:
|
|
||||||
// https://github.com/stretchr/testify/issues/849
|
|
||||||
return // fmt.Errorf("Could not connect to headscale: %s", err)
|
|
||||||
}
|
|
||||||
log.Println("headscale container is ready for core integration tests")
|
|
||||||
|
|
||||||
for namespace, scales := range s.namespaces {
|
|
||||||
log.Printf("Creating headscale namespace: %s\n", namespace)
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&s.headscale,
|
|
||||||
[]string{"headscale", "namespaces", "create", namespace},
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
log.Println("headscale create namespace result: ", result)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
log.Printf("Creating pre auth key for %s\n", namespace)
|
|
||||||
preAuthResult, _, err := ExecuteCommand(
|
|
||||||
&s.headscale,
|
|
||||||
[]string{
|
|
||||||
"headscale",
|
|
||||||
"--namespace",
|
|
||||||
namespace,
|
|
||||||
"preauthkeys",
|
|
||||||
"create",
|
|
||||||
"--reusable",
|
|
||||||
"--expiration",
|
|
||||||
"24h",
|
|
||||||
"--output",
|
|
||||||
"json",
|
|
||||||
},
|
|
||||||
[]string{"LOG_LEVEL=error"},
|
|
||||||
)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
var preAuthKey v1.PreAuthKey
|
|
||||||
err = json.Unmarshal([]byte(preAuthResult), &preAuthKey)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
assert.True(s.T(), preAuthKey.Reusable)
|
|
||||||
|
|
||||||
headscaleEndpoint := "http://headscale:8080"
|
|
||||||
|
|
||||||
log.Printf(
|
|
||||||
"Joining tailscale containers to headscale at %s\n",
|
|
||||||
headscaleEndpoint,
|
|
||||||
)
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
s.joinWaitGroup.Add(1)
|
|
||||||
go s.Join(headscaleEndpoint, preAuthKey.Key, hostname, tailscale)
|
|
||||||
}
|
|
||||||
|
|
||||||
s.joinWaitGroup.Wait()
|
|
||||||
}
|
|
||||||
|
|
||||||
// The nodes need a bit of time to get their updated maps from headscale
|
|
||||||
// TODO: See if we can have a more deterministic wait here.
|
|
||||||
time.Sleep(60 * time.Second)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TearDownSuite() {
|
|
||||||
if !s.saveLogs {
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
for _, tailscale := range scales.tailscales {
|
|
||||||
if err := s.pool.Purge(&tailscale); err != nil {
|
|
||||||
log.Printf("Could not purge resource: %s\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := s.pool.Purge(&s.headscale); err != nil {
|
|
||||||
log.Printf("Could not purge resource: %s\n", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := s.network.Close(); err != nil {
|
|
||||||
log.Printf("Could not close network: %s\n", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) HandleStats(
|
|
||||||
suiteName string,
|
|
||||||
stats *suite.SuiteInformation,
|
|
||||||
) {
|
|
||||||
s.stats = stats
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TestListNodes() {
|
|
||||||
for namespace, scales := range s.namespaces {
|
|
||||||
log.Println("Listing nodes")
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&s.headscale,
|
|
||||||
[]string{"headscale", "--namespace", namespace, "nodes", "list"},
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
log.Printf("List nodes: \n%s\n", result)
|
|
||||||
|
|
||||||
// Chck that the correct count of host is present in node list
|
|
||||||
lines := strings.Split(result, "\n")
|
|
||||||
assert.Equal(s.T(), len(scales.tailscales), len(lines)-2)
|
|
||||||
|
|
||||||
for hostname := range scales.tailscales {
|
|
||||||
assert.Contains(s.T(), result, hostname)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TestGetIpAddresses() {
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
ips, err := getIPs(scales.tailscales)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
for hostname := range scales.tailscales {
|
|
||||||
ips := ips[hostname]
|
|
||||||
for _, ip := range ips {
|
|
||||||
s.T().Run(hostname, func(t *testing.T) {
|
|
||||||
assert.NotNil(t, ip)
|
|
||||||
|
|
||||||
log.Printf("IP for %s: %s\n", hostname, ip)
|
|
||||||
|
|
||||||
// c.Assert(ip.Valid(), check.IsTrue)
|
|
||||||
assert.True(t, ip.Is4() || ip.Is6())
|
|
||||||
switch {
|
|
||||||
case ip.Is4():
|
|
||||||
assert.True(t, IpPrefix4.Contains(ip))
|
|
||||||
case ip.Is6():
|
|
||||||
assert.True(t, IpPrefix6.Contains(ip))
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO(kradalby): fix this test
|
|
||||||
// We need some way to import ipnstate.Status from multiple go packages.
|
|
||||||
// Currently it will only work with 1.18.x since that is the last
|
|
||||||
// version we have in go.mod
|
|
||||||
// func (s *IntegrationTestSuite) TestStatus() {
|
|
||||||
// for _, scales := range s.namespaces {
|
|
||||||
// ips, err := getIPs(scales.tailscales)
|
|
||||||
// assert.Nil(s.T(), err)
|
|
||||||
//
|
|
||||||
// for hostname, tailscale := range scales.tailscales {
|
|
||||||
// s.T().Run(hostname, func(t *testing.T) {
|
|
||||||
// command := []string{"tailscale", "status", "--json"}
|
|
||||||
//
|
|
||||||
// log.Printf("Getting status for %s\n", hostname)
|
|
||||||
// result, err := ExecuteCommand(
|
|
||||||
// &tailscale,
|
|
||||||
// command,
|
|
||||||
// []string{},
|
|
||||||
// )
|
|
||||||
// assert.Nil(t, err)
|
|
||||||
//
|
|
||||||
// var status ipnstate.Status
|
|
||||||
// err = json.Unmarshal([]byte(result), &status)
|
|
||||||
// assert.Nil(s.T(), err)
|
|
||||||
//
|
|
||||||
// // TODO(kradalby): Replace this check with peer length of SAME namespace
|
|
||||||
// // Check if we have as many nodes in status
|
|
||||||
// // as we have IPs/tailscales
|
|
||||||
// // lines := strings.Split(result, "\n")
|
|
||||||
// // assert.Equal(t, len(ips), len(lines)-1)
|
|
||||||
// // assert.Equal(t, len(scales.tailscales), len(lines)-1)
|
|
||||||
//
|
|
||||||
// peerIps := getIPsfromIPNstate(status)
|
|
||||||
//
|
|
||||||
// // Check that all hosts is present in all hosts status
|
|
||||||
// for ipHostname, ip := range ips {
|
|
||||||
// if hostname != ipHostname {
|
|
||||||
// assert.Contains(t, peerIps, ip)
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// })
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
func getIPsfromIPNstate(status ipnstate.Status) []netip.Addr {
|
|
||||||
ips := make([]netip.Addr, 0)
|
|
||||||
|
|
||||||
for _, peer := range status.Peer {
|
|
||||||
ips = append(ips, peer.TailscaleIPs...)
|
|
||||||
}
|
|
||||||
|
|
||||||
return ips
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Adopt test for cross communication between namespaces.
|
|
||||||
func (s *IntegrationTestSuite) TestPingAllPeersByAddress() {
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
ips, err := getIPs(scales.tailscales)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
for peername, peerIPs := range ips {
|
|
||||||
for i, ip := range peerIPs {
|
|
||||||
// We currently cant ping ourselves, so skip that.
|
|
||||||
if peername == hostname {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
s.T().
|
|
||||||
Run(fmt.Sprintf("%s-%s-%d", hostname, peername, i), func(t *testing.T) {
|
|
||||||
// We are only interested in "direct ping" which means what we
|
|
||||||
// might need a couple of more attempts before reaching the node.
|
|
||||||
command := []string{
|
|
||||||
"tailscale", "ping",
|
|
||||||
"--timeout=1s",
|
|
||||||
"--c=10",
|
|
||||||
"--until-direct=true",
|
|
||||||
ip.String(),
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Printf(
|
|
||||||
"Pinging from %s to %s (%s)\n",
|
|
||||||
hostname,
|
|
||||||
peername,
|
|
||||||
ip,
|
|
||||||
)
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(t, err)
|
|
||||||
log.Printf("Result for %s: %s\n", hostname, result)
|
|
||||||
assert.Contains(t, result, "pong")
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TestTailDrop() {
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
ips, err := getIPs(scales.tailscales)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
retry := func(times int, sleepInverval time.Duration, doWork func() error) (err error) {
|
|
||||||
for attempts := 0; attempts < times; attempts++ {
|
|
||||||
err = doWork()
|
|
||||||
if err == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
time.Sleep(sleepInverval)
|
|
||||||
}
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
command := []string{"touch", fmt.Sprintf("/tmp/file_from_%s", hostname)}
|
|
||||||
_, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
for peername := range ips {
|
|
||||||
if peername == hostname {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
var ip4 netip.Addr
|
|
||||||
for _, ip := range ips[peername] {
|
|
||||||
if ip.Is4() {
|
|
||||||
ip4 = ip
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if ip4.IsUnspecified() {
|
|
||||||
panic("no ipv4 address found")
|
|
||||||
}
|
|
||||||
|
|
||||||
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
|
|
||||||
command := []string{
|
|
||||||
"tailscale", "file", "cp",
|
|
||||||
fmt.Sprintf("/tmp/file_from_%s", hostname),
|
|
||||||
fmt.Sprintf("%s:", ip4),
|
|
||||||
}
|
|
||||||
err := retry(10, 1*time.Second, func() error {
|
|
||||||
log.Printf(
|
|
||||||
"Sending file from %s to %s\n",
|
|
||||||
hostname,
|
|
||||||
peername,
|
|
||||||
)
|
|
||||||
_, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
ExecuteCommandTimeout(60*time.Second),
|
|
||||||
)
|
|
||||||
return err
|
|
||||||
})
|
|
||||||
|
|
||||||
assert.Nil(t, err)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
command := []string{
|
|
||||||
"tailscale", "file",
|
|
||||||
"get",
|
|
||||||
"/tmp/",
|
|
||||||
}
|
|
||||||
_, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
for peername, ip := range ips {
|
|
||||||
if peername == hostname {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
|
|
||||||
command := []string{
|
|
||||||
"ls",
|
|
||||||
fmt.Sprintf("/tmp/file_from_%s", peername),
|
|
||||||
}
|
|
||||||
log.Printf(
|
|
||||||
"Checking file in %s (%s) from %s (%s)\n",
|
|
||||||
hostname,
|
|
||||||
ips[hostname][1],
|
|
||||||
peername,
|
|
||||||
ip,
|
|
||||||
)
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(t, err)
|
|
||||||
log.Printf("Result for %s: %s\n", peername, result)
|
|
||||||
assert.Equal(
|
|
||||||
t,
|
|
||||||
fmt.Sprintf("/tmp/file_from_%s\n", peername),
|
|
||||||
result,
|
|
||||||
)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TestPingAllPeersByHostname() {
|
|
||||||
hostnames, err := getMagicFQDN(&s.headscale)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
log.Printf("Resolved hostnames: %#v", hostnames)
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
for _, peername := range hostnames {
|
|
||||||
if strings.Contains(peername, hostname) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
|
|
||||||
command := []string{
|
|
||||||
"tailscale", "ping",
|
|
||||||
"--timeout=10s",
|
|
||||||
"--c=20",
|
|
||||||
"--until-direct=true",
|
|
||||||
peername,
|
|
||||||
}
|
|
||||||
|
|
||||||
log.Printf(
|
|
||||||
"Pinging using hostname from %s to %s\n",
|
|
||||||
hostname,
|
|
||||||
peername,
|
|
||||||
)
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
assert.Nil(t, err)
|
|
||||||
log.Printf("Result for %s: %s\n", hostname, result)
|
|
||||||
assert.Contains(t, result, "pong")
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *IntegrationTestSuite) TestMagicDNS() {
|
|
||||||
hostnames, err := getMagicFQDN(&s.headscale)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
log.Printf("Resolved hostnames: %#v", hostnames)
|
|
||||||
|
|
||||||
for _, scales := range s.namespaces {
|
|
||||||
ips, err := getIPs(scales.tailscales)
|
|
||||||
assert.Nil(s.T(), err)
|
|
||||||
|
|
||||||
retry := func(times int, sleepInverval time.Duration, doWork func() (string, error)) (result string, err error) {
|
|
||||||
for attempts := 0; attempts < times; attempts++ {
|
|
||||||
result, err = doWork()
|
|
||||||
if err == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
time.Sleep(sleepInverval)
|
|
||||||
}
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
for hostname, tailscale := range scales.tailscales {
|
|
||||||
for _, peername := range hostnames {
|
|
||||||
if strings.Contains(peername, hostname) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
|
|
||||||
command := []string{
|
|
||||||
"tailscale", "ip", peername,
|
|
||||||
}
|
|
||||||
result, err := retry(10, 1*time.Second, func() (string, error) {
|
|
||||||
log.Printf(
|
|
||||||
"Resolving name %s from %s\n",
|
|
||||||
peername,
|
|
||||||
hostname,
|
|
||||||
)
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
return result, err
|
|
||||||
})
|
|
||||||
|
|
||||||
assert.Nil(t, err)
|
|
||||||
log.Printf("Result for %s: %s\n", hostname, result)
|
|
||||||
|
|
||||||
peerBaseName := peername[:len(peername)-MachineGivenNameHashLength-1]
|
|
||||||
expectedAddresses := ips[peerBaseName]
|
|
||||||
for _, ip := range expectedAddresses {
|
|
||||||
assert.Contains(t, result, ip.String())
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func getAPIURLs(
|
|
||||||
tailscales map[string]dockertest.Resource,
|
|
||||||
) (map[netip.Addr]string, error) {
|
|
||||||
fts := make(map[netip.Addr]string)
|
|
||||||
for _, tailscale := range tailscales {
|
|
||||||
command := []string{
|
|
||||||
"curl",
|
|
||||||
"--unix-socket",
|
|
||||||
"/run/tailscale/tailscaled.sock",
|
|
||||||
"http://localhost/localapi/v0/file-targets",
|
|
||||||
}
|
|
||||||
result, _, err := ExecuteCommand(
|
|
||||||
&tailscale,
|
|
||||||
command,
|
|
||||||
[]string{},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
var pft []apitype.FileTarget
|
|
||||||
if err := json.Unmarshal([]byte(result), &pft); err != nil {
|
|
||||||
return nil, fmt.Errorf("invalid JSON: %w", err)
|
|
||||||
}
|
|
||||||
for _, ft := range pft {
|
|
||||||
n := ft.Node
|
|
||||||
for _, a := range n.Addresses { // just add all the addresses
|
|
||||||
if _, ok := fts[a.Addr()]; !ok {
|
|
||||||
if ft.PeerAPIURL == "" {
|
|
||||||
return nil, errors.New("api url is empty")
|
|
||||||
}
|
|
||||||
fts[a.Addr()] = ft.PeerAPIURL
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return fts, nil
|
|
||||||
}
|
|
Loading…
x
Reference in New Issue
Block a user