headscale/integration_test.go

865 lines
20 KiB
Go
Raw Normal View History

//go:build integration
// +build integration
package headscale
import (
"bytes"
2021-08-20 09:50:55 -06:00
"context"
"encoding/json"
2021-10-04 07:14:12 -06:00
"errors"
"fmt"
2021-08-20 09:50:55 -06:00
"io/ioutil"
"log"
"net/http"
"os"
2021-08-20 09:50:55 -06:00
"path"
"strings"
"sync"
2021-08-13 03:33:19 -06:00
"testing"
"time"
v1 "github.com/juanfont/headscale/gen/go/headscale/v1"
"github.com/ory/dockertest/v3"
"github.com/ory/dockertest/v3/docker"
2021-08-13 03:33:19 -06:00
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/suite"
2021-11-14 10:35:49 -07:00
"inet.af/netaddr"
2021-09-26 06:22:11 -06:00
"tailscale.com/client/tailscale/apitype"
"tailscale.com/ipn/ipnstate"
)
2022-02-11 01:39:00 -07:00
var tailscaleVersions = []string{"1.20.4", "1.18.2", "1.16.2", "1.14.3", "1.12.3"}
type TestNamespace struct {
count int
tailscales map[string]dockertest.Resource
}
2021-08-13 03:33:19 -06:00
type IntegrationTestSuite struct {
suite.Suite
2021-08-20 09:50:55 -06:00
stats *suite.SuiteInformation
pool dockertest.Pool
network dockertest.Network
headscale dockertest.Resource
namespaces map[string]TestNamespace
joinWaitGroup sync.WaitGroup
2021-08-13 03:33:19 -06:00
}
2021-08-13 03:33:19 -06:00
func TestIntegrationTestSuite(t *testing.T) {
2021-08-20 09:50:55 -06:00
s := new(IntegrationTestSuite)
s.namespaces = map[string]TestNamespace{
"main": {
count: 20,
tailscales: make(map[string]dockertest.Resource),
},
"shared": {
count: 5,
tailscales: make(map[string]dockertest.Resource),
},
}
2021-08-20 09:50:55 -06:00
suite.Run(t, s)
// HandleStats, which allows us to check if we passed and save logs
// is called after TearDown, so we cannot tear down containers before
// we have potentially saved the logs.
for _, scales := range s.namespaces {
for _, tailscale := range scales.tailscales {
if err := s.pool.Purge(&tailscale); err != nil {
log.Printf("Could not purge resource: %s\n", err)
}
2021-08-20 09:50:55 -06:00
}
}
if !s.stats.Passed() {
err := s.saveLog(&s.headscale, "test_output")
2021-08-20 09:50:55 -06:00
if err != nil {
log.Printf("Could not save log: %s\n", err)
}
}
if err := s.pool.Purge(&s.headscale); err != nil {
2021-08-20 09:50:55 -06:00
log.Printf("Could not purge resource: %s\n", err)
}
if err := s.network.Close(); err != nil {
2021-08-20 09:50:55 -06:00
log.Printf("Could not close network: %s\n", err)
}
2021-08-13 03:33:19 -06:00
}
2021-11-13 01:36:45 -07:00
func (s *IntegrationTestSuite) saveLog(
resource *dockertest.Resource,
basePath string,
) error {
2021-08-20 09:50:55 -06:00
err := os.MkdirAll(basePath, os.ModePerm)
if err != nil {
return err
}
var stdout bytes.Buffer
var stderr bytes.Buffer
err = s.pool.Client.Logs(
2021-08-20 09:50:55 -06:00
docker.LogsOptions{
Context: context.TODO(),
Container: resource.Container.ID,
OutputStream: &stdout,
ErrorStream: &stderr,
Tail: "all",
RawTerminal: false,
Stdout: true,
Stderr: true,
Follow: false,
Timestamps: false,
},
)
if err != nil {
return err
}
log.Printf("Saving logs for %s to %s\n", resource.Container.Name, basePath)
2021-08-20 09:50:55 -06:00
2021-11-13 01:36:45 -07:00
err = ioutil.WriteFile(
path.Join(basePath, resource.Container.Name+".stdout.log"),
[]byte(stdout.String()),
0o644,
)
2021-08-20 09:50:55 -06:00
if err != nil {
return err
}
2021-11-13 01:36:45 -07:00
err = ioutil.WriteFile(
path.Join(basePath, resource.Container.Name+".stderr.log"),
[]byte(stdout.String()),
0o644,
)
2021-08-20 09:50:55 -06:00
if err != nil {
return err
}
return nil
}
func (s *IntegrationTestSuite) Join(
endpoint, key, hostname string,
tailscale dockertest.Resource,
) {
defer s.joinWaitGroup.Done()
command := []string{
"tailscale",
"up",
"-login-server",
endpoint,
"--authkey",
key,
"--hostname",
hostname,
}
log.Println("Join command:", command)
log.Printf("Running join command for %s\n", hostname)
_, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(s.T(), err)
log.Printf("%s joined\n", hostname)
}
func (s *IntegrationTestSuite) tailscaleContainer(
namespace, identifier, version string,
) (string, *dockertest.Resource) {
tailscaleBuildOptions := &dockertest.BuildOptions{
Dockerfile: "Dockerfile.tailscale",
ContextDir: ".",
BuildArgs: []docker.BuildArg{
{
Name: "TAILSCALE_VERSION",
Value: version,
},
},
}
2021-11-13 01:36:45 -07:00
hostname := fmt.Sprintf(
"%s-tailscale-%s-%s",
namespace,
strings.Replace(version, ".", "-", -1),
identifier,
)
tailscaleOptions := &dockertest.RunOptions{
Name: hostname,
Networks: []*dockertest.Network{&s.network},
2021-11-13 01:36:45 -07:00
Cmd: []string{
"tailscaled", "--tun=tsdev",
2021-11-13 01:36:45 -07:00
},
}
2021-11-13 01:36:45 -07:00
pts, err := s.pool.BuildAndRunWithBuildOptions(
tailscaleBuildOptions,
tailscaleOptions,
DockerRestartPolicy,
DockerAllowLocalIPv6,
DockerAllowNetworkAdministration,
2021-11-13 01:36:45 -07:00
)
if err != nil {
log.Fatalf("Could not start resource: %s", err)
}
log.Printf("Created %s container\n", hostname)
2021-11-14 10:35:49 -07:00
return hostname, pts
}
2021-08-13 03:33:19 -06:00
func (s *IntegrationTestSuite) SetupSuite() {
var err error
2021-11-15 09:16:04 -07:00
app = Headscale{
dbType: "sqlite3",
dbString: "integration_test_db.sqlite3",
}
if ppool, err := dockertest.NewPool(""); err == nil {
s.pool = *ppool
} else {
log.Fatalf("Could not connect to docker: %s", err)
}
if pnetwork, err := s.pool.CreateNetwork("headscale-test"); err == nil {
s.network = *pnetwork
} else {
log.Fatalf("Could not create network: %s", err)
}
headscaleBuildOptions := &dockertest.BuildOptions{
Dockerfile: "Dockerfile",
ContextDir: ".",
}
currentPath, err := os.Getwd()
if err != nil {
log.Fatalf("Could not determine current path: %s", err)
}
headscaleOptions := &dockertest.RunOptions{
Name: "headscale",
Mounts: []string{
fmt.Sprintf("%s/integration_test/etc:/etc/headscale", currentPath),
},
Networks: []*dockertest.Network{&s.network},
2021-08-13 03:33:19 -06:00
Cmd: []string{"headscale", "serve"},
}
log.Println("Creating headscale container")
if pheadscale, err := s.pool.BuildAndRunWithBuildOptions(headscaleBuildOptions, headscaleOptions, DockerRestartPolicy); err == nil {
s.headscale = *pheadscale
} else {
log.Fatalf("Could not start resource: %s", err)
}
log.Println("Created headscale container")
log.Println("Creating tailscale containers")
for namespace, scales := range s.namespaces {
for i := 0; i < scales.count; i++ {
version := tailscaleVersions[i%len(tailscaleVersions)]
2021-11-13 01:36:45 -07:00
hostname, container := s.tailscaleContainer(
namespace,
fmt.Sprint(i),
version,
)
scales.tailscales[hostname] = *container
}
}
log.Println("Waiting for headscale to be ready")
hostEndpoint := fmt.Sprintf("localhost:%s", s.headscale.GetPort("8080/tcp"))
if err := s.pool.Retry(func() error {
url := fmt.Sprintf("http://%s/health", hostEndpoint)
2021-11-14 10:35:49 -07:00
resp, err := http.Get(url)
if err != nil {
return err
}
2021-11-14 10:35:49 -07:00
if resp.StatusCode != http.StatusOK {
return fmt.Errorf("status code not OK")
}
2021-11-14 10:35:49 -07:00
return nil
}); err != nil {
// TODO(kradalby): If we cannot access headscale, or any other fatal error during
// test setup, we need to abort and tear down. However, testify does not seem to
// support that at the moment:
// https://github.com/stretchr/testify/issues/849
return // fmt.Errorf("Could not connect to headscale: %s", err)
}
log.Println("headscale container is ready")
for namespace, scales := range s.namespaces {
log.Printf("Creating headscale namespace: %s\n", namespace)
result, err := ExecuteCommand(
&s.headscale,
[]string{"headscale", "namespaces", "create", namespace},
2021-09-26 06:22:11 -06:00
[]string{},
)
log.Println("headscale create namespace result: ", result)
assert.Nil(s.T(), err)
log.Printf("Creating pre auth key for %s\n", namespace)
preAuthResult, err := ExecuteCommand(
&s.headscale,
[]string{
"headscale",
"--namespace",
namespace,
"preauthkeys",
"create",
"--reusable",
"--expiration",
"24h",
"--output",
"json",
},
[]string{"LOG_LEVEL=error"},
)
assert.Nil(s.T(), err)
var preAuthKey v1.PreAuthKey
err = json.Unmarshal([]byte(preAuthResult), &preAuthKey)
assert.Nil(s.T(), err)
assert.True(s.T(), preAuthKey.Reusable)
headscaleEndpoint := "http://headscale:8080"
log.Printf(
2021-11-13 01:36:45 -07:00
"Joining tailscale containers to headscale at %s\n",
headscaleEndpoint,
)
for hostname, tailscale := range scales.tailscales {
s.joinWaitGroup.Add(1)
go s.Join(headscaleEndpoint, preAuthKey.Key, hostname, tailscale)
}
s.joinWaitGroup.Wait()
}
2021-08-13 03:33:19 -06:00
// The nodes need a bit of time to get their updated maps from headscale
// TODO: See if we can have a more deterministic wait here.
2021-08-20 09:50:55 -06:00
time.Sleep(60 * time.Second)
}
2021-08-13 03:33:19 -06:00
func (s *IntegrationTestSuite) TearDownSuite() {
2021-08-20 09:50:55 -06:00
}
2021-11-13 01:36:45 -07:00
func (s *IntegrationTestSuite) HandleStats(
suiteName string,
stats *suite.SuiteInformation,
) {
2021-08-20 09:50:55 -06:00
s.stats = stats
}
2021-08-13 03:33:19 -06:00
func (s *IntegrationTestSuite) TestListNodes() {
for namespace, scales := range s.namespaces {
log.Println("Listing nodes")
result, err := ExecuteCommand(
&s.headscale,
[]string{"headscale", "--namespace", namespace, "nodes", "list"},
2021-09-26 06:22:11 -06:00
[]string{},
)
assert.Nil(s.T(), err)
2021-08-13 03:33:19 -06:00
log.Printf("List nodes: \n%s\n", result)
2021-08-13 03:33:19 -06:00
// Chck that the correct count of host is present in node list
lines := strings.Split(result, "\n")
assert.Equal(s.T(), len(scales.tailscales), len(lines)-2)
for hostname := range scales.tailscales {
assert.Contains(s.T(), result, hostname)
}
}
}
2021-08-13 03:33:19 -06:00
func (s *IntegrationTestSuite) TestGetIpAddresses() {
for _, scales := range s.namespaces {
ips, err := getIPs(scales.tailscales)
assert.Nil(s.T(), err)
2021-08-13 03:33:19 -06:00
2022-02-11 01:39:00 -07:00
for hostname := range scales.tailscales {
2022-01-16 06:16:59 -07:00
ips := ips[hostname]
for _, ip := range ips {
s.T().Run(hostname, func(t *testing.T) {
assert.NotNil(t, ip)
log.Printf("IP for %s: %s\n", hostname, ip)
2022-01-16 06:16:59 -07:00
// c.Assert(ip.Valid(), check.IsTrue)
assert.True(t, ip.Is4() || ip.Is6())
switch {
case ip.Is4():
assert.True(t, IpPrefix4.Contains(ip))
case ip.Is6():
assert.True(t, IpPrefix6.Contains(ip))
}
})
}
}
2021-08-13 03:33:19 -06:00
}
}
// TODO(kradalby): fix this test
2022-01-16 06:16:59 -07:00
// We need some way to import ipnstate.Status from multiple go packages.
// Currently it will only work with 1.18.x since that is the last
// version we have in go.mod
// func (s *IntegrationTestSuite) TestStatus() {
2022-01-16 06:16:59 -07:00
// for _, scales := range s.namespaces {
// ips, err := getIPs(scales.tailscales)
// assert.Nil(s.T(), err)
//
2022-01-16 06:16:59 -07:00
// for hostname, tailscale := range scales.tailscales {
// s.T().Run(hostname, func(t *testing.T) {
// command := []string{"tailscale", "status", "--json"}
//
// log.Printf("Getting status for %s\n", hostname)
2022-01-16 06:16:59 -07:00
// result, err := ExecuteCommand(
// &tailscale,
// command,
// []string{},
// )
// assert.Nil(t, err)
//
2022-01-16 06:16:59 -07:00
// var status ipnstate.Status
// err = json.Unmarshal([]byte(result), &status)
// assert.Nil(s.T(), err)
//
2022-01-16 06:16:59 -07:00
// // TODO(kradalby): Replace this check with peer length of SAME namespace
// // Check if we have as many nodes in status
// // as we have IPs/tailscales
// // lines := strings.Split(result, "\n")
// // assert.Equal(t, len(ips), len(lines)-1)
// // assert.Equal(t, len(scales.tailscales), len(lines)-1)
//
2022-01-16 06:16:59 -07:00
// peerIps := getIPsfromIPNstate(status)
//
2022-01-16 06:16:59 -07:00
// // Check that all hosts is present in all hosts status
// for ipHostname, ip := range ips {
// if hostname != ipHostname {
// assert.Contains(t, peerIps, ip)
// }
// }
// })
// }
// }
// }
2021-08-13 03:33:19 -06:00
func getIPsfromIPNstate(status ipnstate.Status) []netaddr.IP {
ips := make([]netaddr.IP, 0)
for _, peer := range status.Peer {
ips = append(ips, peer.TailscaleIPs...)
}
return ips
}
2022-01-16 06:16:59 -07:00
func (s *IntegrationTestSuite) TestPingAllPeersByAddress() {
for _, scales := range s.namespaces {
ips, err := getIPs(scales.tailscales)
assert.Nil(s.T(), err)
2021-08-13 04:01:23 -06:00
for hostname, tailscale := range scales.tailscales {
2022-01-16 06:16:59 -07:00
for peername, peerIPs := range ips {
for i, ip := range peerIPs {
// We currently cant ping ourselves, so skip that.
2022-01-16 06:16:59 -07:00
if peername == hostname {
continue
}
2022-02-11 01:39:00 -07:00
s.T().
Run(fmt.Sprintf("%s-%s-%d", hostname, peername, i), func(t *testing.T) {
// We are only interested in "direct ping" which means what we
// might need a couple of more attempts before reaching the node.
command := []string{
"tailscale", "ping",
"--timeout=1s",
"--c=10",
"--until-direct=true",
ip.String(),
}
log.Printf(
2022-02-11 01:39:00 -07:00
"Pinging from %s to %s (%s)\n",
hostname,
peername,
ip,
)
result, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(t, err)
log.Printf("Result for %s: %s\n", hostname, result)
2022-02-11 01:39:00 -07:00
assert.Contains(t, result, "pong")
})
2022-01-16 06:16:59 -07:00
}
}
2021-08-13 04:01:23 -06:00
}
}
}
2021-08-13 03:33:19 -06:00
func (s *IntegrationTestSuite) TestSharedNodes() {
main := s.namespaces["main"]
shared := s.namespaces["shared"]
result, err := ExecuteCommand(
&s.headscale,
2021-11-13 01:36:45 -07:00
[]string{
"headscale",
"nodes",
"list",
"--output",
"json",
"--namespace",
"shared",
},
2021-09-26 06:22:11 -06:00
[]string{},
)
assert.Nil(s.T(), err)
var machineList []v1.Machine
err = json.Unmarshal([]byte(result), &machineList)
assert.Nil(s.T(), err)
for _, machine := range machineList {
result, err := ExecuteCommand(
&s.headscale,
[]string{
"headscale",
"nodes",
"share",
"--identifier", fmt.Sprint(machine.Id),
"--namespace", "main",
},
2021-09-26 06:22:11 -06:00
[]string{},
)
assert.Nil(s.T(), err)
log.Println("Shared node with result: ", result)
}
result, err = ExecuteCommand(
&s.headscale,
[]string{"headscale", "nodes", "list", "--namespace", "main"},
2021-09-26 06:22:11 -06:00
[]string{},
)
assert.Nil(s.T(), err)
log.Println("Nodelist after sharing", result)
// Chck that the correct count of host is present in node list
lines := strings.Split(result, "\n")
assert.Equal(s.T(), len(main.tailscales)+len(shared.tailscales), len(lines)-2)
for hostname := range main.tailscales {
assert.Contains(s.T(), result, hostname)
}
for hostname := range shared.tailscales {
assert.Contains(s.T(), result, hostname)
}
2021-10-16 03:36:16 -06:00
// TODO(juanfont): We have to find out why do we need to wait
time.Sleep(100 * time.Second) // Wait for the nodes to receive updates
sharedIps, err := getIPs(shared.tailscales)
assert.Nil(s.T(), err)
for hostname, tailscale := range main.tailscales {
2022-01-16 06:16:59 -07:00
for peername, peerIPs := range sharedIps {
for i, ip := range peerIPs {
// We currently cant ping ourselves, so skip that.
2022-01-16 06:16:59 -07:00
if peername == hostname {
continue
}
2022-02-11 01:39:00 -07:00
s.T().
Run(fmt.Sprintf("%s-%s-%d", hostname, peername, i), func(t *testing.T) {
// We are only interested in "direct ping" which means what we
// might need a couple of more attempts before reaching the node.
command := []string{
"tailscale", "ping",
"--timeout=15s",
"--c=20",
"--until-direct=true",
ip.String(),
}
log.Printf(
2022-02-11 01:39:00 -07:00
"Pinging from %s to %s (%s)\n",
hostname,
peername,
ip,
)
result, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(t, err)
log.Printf("Result for %s: %s\n", hostname, result)
2022-02-11 01:39:00 -07:00
assert.Contains(t, result, "pong")
})
2022-01-16 06:16:59 -07:00
}
}
}
}
2021-09-25 05:12:44 -06:00
func (s *IntegrationTestSuite) TestTailDrop() {
for _, scales := range s.namespaces {
ips, err := getIPs(scales.tailscales)
assert.Nil(s.T(), err)
2021-09-26 04:22:59 -06:00
assert.Nil(s.T(), err)
2021-09-25 05:12:44 -06:00
retry := func(times int, sleepInverval time.Duration, doWork func() error) (err error) {
for attempts := 0; attempts < times; attempts++ {
err = doWork()
if err == nil {
return
}
time.Sleep(sleepInverval)
}
return
}
2021-09-25 05:12:44 -06:00
for hostname, tailscale := range scales.tailscales {
command := []string{"touch", fmt.Sprintf("/tmp/file_from_%s", hostname)}
_, err := ExecuteCommand(
2021-09-25 05:12:44 -06:00
&tailscale,
command,
[]string{},
2021-09-25 05:12:44 -06:00
)
assert.Nil(s.T(), err)
2022-02-11 01:39:00 -07:00
for peername := range ips {
if peername == hostname {
continue
}
2021-09-25 05:12:44 -06:00
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
command := []string{
"tailscale", "file", "cp",
fmt.Sprintf("/tmp/file_from_%s", hostname),
fmt.Sprintf("%s:", peername),
2021-09-25 05:12:44 -06:00
}
retry(10, 1*time.Second, func() error {
log.Printf(
"Sending file from %s to %s\n",
hostname,
peername,
)
_, err := ExecuteCommand(
&tailscale,
command,
[]string{},
ExecuteCommandTimeout(60*time.Second),
)
return err
})
assert.Nil(t, err)
2021-09-25 05:12:44 -06:00
})
}
}
for hostname, tailscale := range scales.tailscales {
command := []string{
"tailscale", "file",
"get",
2021-09-26 06:22:11 -06:00
"/tmp/",
2021-09-25 05:12:44 -06:00
}
_, err := ExecuteCommand(
2021-09-25 05:12:44 -06:00
&tailscale,
command,
2021-09-26 06:22:11 -06:00
[]string{},
2021-09-25 05:12:44 -06:00
)
assert.Nil(s.T(), err)
for peername, ip := range ips {
2022-01-16 06:16:59 -07:00
if peername == hostname {
continue
}
2021-09-25 05:12:44 -06:00
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
2022-01-16 06:16:59 -07:00
command := []string{
"ls",
fmt.Sprintf("/tmp/file_from_%s", peername),
2021-09-25 05:12:44 -06:00
}
log.Printf(
2022-01-16 06:16:59 -07:00
"Checking file in %s (%s) from %s (%s)\n",
hostname,
ips[hostname],
peername,
ip,
)
result, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(t, err)
log.Printf("Result for %s: %s\n", peername, result)
2022-01-16 06:16:59 -07:00
assert.Equal(
t,
fmt.Sprintf("/tmp/file_from_%s\n", peername),
result,
)
2021-09-25 05:12:44 -06:00
})
}
}
}
}
2022-01-16 06:16:59 -07:00
func (s *IntegrationTestSuite) TestPingAllPeersByHostname() {
2021-10-04 10:04:08 -06:00
for namespace, scales := range s.namespaces {
ips, err := getIPs(scales.tailscales)
assert.Nil(s.T(), err)
for hostname, tailscale := range scales.tailscales {
2022-02-11 01:39:00 -07:00
for peername := range ips {
2022-01-16 06:16:59 -07:00
if peername == hostname {
continue
}
2021-10-04 10:04:08 -06:00
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
2022-01-16 06:16:59 -07:00
command := []string{
"tailscale", "ping",
"--timeout=10s",
"--c=20",
"--until-direct=true",
fmt.Sprintf("%s.%s.headscale.net", peername, namespace),
}
log.Printf(
"Pinging using hostname from %s to %s\n",
2022-01-16 06:16:59 -07:00
hostname,
peername,
)
result, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(t, err)
log.Printf("Result for %s: %s\n", hostname, result)
2022-01-16 06:16:59 -07:00
assert.Contains(t, result, "pong")
})
}
}
}
}
func (s *IntegrationTestSuite) TestMagicDNS() {
2022-01-16 06:16:59 -07:00
for namespace, scales := range s.namespaces {
ips, err := getIPs(scales.tailscales)
assert.Nil(s.T(), err)
for hostname, tailscale := range scales.tailscales {
for peername, ips := range ips {
if peername == hostname {
continue
}
s.T().Run(fmt.Sprintf("%s-%s", hostname, peername), func(t *testing.T) {
2022-01-16 06:16:59 -07:00
command := []string{
"tailscale", "ip",
2022-01-16 06:16:59 -07:00
fmt.Sprintf("%s.%s.headscale.net", peername, namespace),
2021-10-04 10:04:08 -06:00
}
2022-01-16 06:16:59 -07:00
log.Printf(
"Resolving name %s from %s\n",
2022-01-16 06:16:59 -07:00
peername,
hostname,
)
result, err := ExecuteCommand(
&tailscale,
command,
[]string{},
)
assert.Nil(t, err)
log.Printf("Result for %s: %s\n", hostname, result)
2022-01-16 06:16:59 -07:00
for _, ip := range ips {
assert.Contains(t, result, ip.String())
2022-01-16 06:16:59 -07:00
}
2021-10-04 10:04:08 -06:00
})
}
}
}
}
2022-02-11 01:39:00 -07:00
func getIPs(
tailscales map[string]dockertest.Resource,
) (map[string][]netaddr.IP, error) {
2022-01-16 06:16:59 -07:00
ips := make(map[string][]netaddr.IP)
for hostname, tailscale := range tailscales {
command := []string{"tailscale", "ip"}
result, err := ExecuteCommand(
&tailscale,
command,
2021-09-26 06:22:11 -06:00
[]string{},
)
2021-08-13 03:33:19 -06:00
if err != nil {
return nil, err
}
2022-01-16 06:16:59 -07:00
for _, address := range strings.Split(result, "\n") {
address = strings.TrimSuffix(address, "\n")
if len(address) < 1 {
continue
}
ip, err := netaddr.ParseIP(address)
if err != nil {
return nil, err
}
ips[hostname] = append(ips[hostname], ip)
2021-08-13 03:33:19 -06:00
}
}
2021-11-14 10:35:49 -07:00
2021-08-13 03:33:19 -06:00
return ips, nil
}
2021-09-26 04:22:59 -06:00
2021-11-13 01:36:45 -07:00
func getAPIURLs(
tailscales map[string]dockertest.Resource,
) (map[netaddr.IP]string, error) {
2021-09-26 04:22:59 -06:00
fts := make(map[netaddr.IP]string)
2021-09-26 06:22:11 -06:00
for _, tailscale := range tailscales {
2021-09-26 07:17:27 -06:00
command := []string{
2021-09-26 04:22:59 -06:00
"curl",
"--unix-socket",
"/run/tailscale/tailscaled.sock",
"http://localhost/localapi/v0/file-targets",
}
result, err := ExecuteCommand(
2021-09-26 04:22:59 -06:00
&tailscale,
command,
2021-09-26 06:22:11 -06:00
[]string{},
2021-09-26 04:22:59 -06:00
)
if err != nil {
return nil, err
}
2021-09-26 06:33:01 -06:00
2021-09-26 04:22:59 -06:00
var pft []apitype.FileTarget
2021-09-26 06:22:11 -06:00
if err := json.Unmarshal([]byte(result), &pft); err != nil {
2021-09-26 04:22:59 -06:00
return nil, fmt.Errorf("invalid JSON: %w", err)
}
for _, ft := range pft {
n := ft.Node
2021-09-26 06:22:11 -06:00
for _, a := range n.Addresses { // just add all the addresses
if _, ok := fts[a.IP()]; !ok {
2021-10-04 07:14:12 -06:00
if ft.PeerAPIURL == "" {
return nil, errors.New("api url is empty")
}
2021-09-26 06:22:11 -06:00
fts[a.IP()] = ft.PeerAPIURL
2021-09-26 04:22:59 -06:00
}
}
}
}
2021-11-14 10:35:49 -07:00
2021-09-26 06:22:11 -06:00
return fts, nil
2021-09-26 04:22:59 -06:00
}