mirror of
				https://github.com/juanfont/headscale.git
				synced 2025-10-28 10:51:44 +01:00 
			
		
		
		
	* handle control protocol through websocket The necessary behaviour is already in place, but the wasm build only issued GETs, and the handler was not invoked. * get DERP-over-websocket working for wasm clients * Prepare for testing builtin websocket-over-DERP Still needs some way to assert that clients are connected through websockets, rather than the TCP hijacking version of DERP. * integration tests: properly differentiate between DERP transports * do not touch unrelated code * linter fixes * integration testing: unexport common implementation of derp server scenario * fixup! integration testing: unexport common implementation of derp server scenario * dockertestutil/logs: remove unhelpful comment * update changelog --------- Co-authored-by: Csaba Sarkadi <sarkadicsa@tutanota.de>
		
			
				
	
	
		
			431 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			431 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| package integration
 | |
| 
 | |
| import (
 | |
| 	"bufio"
 | |
| 	"bytes"
 | |
| 	"io"
 | |
| 	"os"
 | |
| 	"strings"
 | |
| 	"sync"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/juanfont/headscale/hscontrol/util"
 | |
| 	"github.com/juanfont/headscale/integration/tsic"
 | |
| 	"github.com/stretchr/testify/assert"
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	derpPingTimeout = 2 * time.Second
 | |
| 	derpPingCount   = 10
 | |
| )
 | |
| 
 | |
| func assertNoErr(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "unexpected error: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrf(t *testing.T, msg string, err error) {
 | |
| 	t.Helper()
 | |
| 	if err != nil {
 | |
| 		t.Fatalf(msg, err)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func assertNotNil(t *testing.T, thing interface{}) {
 | |
| 	t.Helper()
 | |
| 	if thing == nil {
 | |
| 		t.Fatal("got unexpected nil")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func assertNoErrHeadscaleEnv(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to create headscale environment: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrGetHeadscale(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to get headscale: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrListClients(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to list clients: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrListClientIPs(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to get client IPs: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrSync(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to have all clients sync up: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrListFQDN(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to list FQDNs: %s", err)
 | |
| }
 | |
| 
 | |
| func assertNoErrLogout(t *testing.T, err error) {
 | |
| 	t.Helper()
 | |
| 	assertNoErrf(t, "failed to log out tailscale nodes: %s", err)
 | |
| }
 | |
| 
 | |
| func assertContains(t *testing.T, str, subStr string) {
 | |
| 	t.Helper()
 | |
| 	if !strings.Contains(str, subStr) {
 | |
| 		t.Fatalf("%#v does not contain %#v", str, subStr)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func didClientUseWebsocketForDERP(t *testing.T, client TailscaleClient) bool {
 | |
| 	t.Helper()
 | |
| 
 | |
| 	buf := &bytes.Buffer{}
 | |
| 	err := client.WriteLogs(buf, buf)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("failed to fetch client logs: %s: %s", client.Hostname(), err)
 | |
| 	}
 | |
| 
 | |
| 	count, err := countMatchingLines(buf, func(line string) bool {
 | |
| 		return strings.Contains(line, "websocket: connected to ")
 | |
| 	})
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("failed to process client logs: %s: %s", client.Hostname(), err)
 | |
| 	}
 | |
| 
 | |
| 	return count > 0
 | |
| }
 | |
| 
 | |
| func pingAllHelper(t *testing.T, clients []TailscaleClient, addrs []string, opts ...tsic.PingOption) int {
 | |
| 	t.Helper()
 | |
| 	success := 0
 | |
| 
 | |
| 	for _, client := range clients {
 | |
| 		for _, addr := range addrs {
 | |
| 			err := client.Ping(addr, opts...)
 | |
| 			if err != nil {
 | |
| 				t.Errorf("failed to ping %s from %s: %s", addr, client.Hostname(), err)
 | |
| 			} else {
 | |
| 				success++
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return success
 | |
| }
 | |
| 
 | |
| func pingDerpAllHelper(t *testing.T, clients []TailscaleClient, addrs []string) int {
 | |
| 	t.Helper()
 | |
| 	success := 0
 | |
| 
 | |
| 	for _, client := range clients {
 | |
| 		for _, addr := range addrs {
 | |
| 			if isSelfClient(client, addr) {
 | |
| 				continue
 | |
| 			}
 | |
| 
 | |
| 			err := client.Ping(
 | |
| 				addr,
 | |
| 				tsic.WithPingTimeout(derpPingTimeout),
 | |
| 				tsic.WithPingCount(derpPingCount),
 | |
| 				tsic.WithPingUntilDirect(false),
 | |
| 			)
 | |
| 			if err != nil {
 | |
| 				t.Logf("failed to ping %s from %s: %s", addr, client.Hostname(), err)
 | |
| 			} else {
 | |
| 				success++
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return success
 | |
| }
 | |
| 
 | |
| // assertClientsState validates the status and netmap of a list of
 | |
| // clients for the general case of all to all connectivity.
 | |
| func assertClientsState(t *testing.T, clients []TailscaleClient) {
 | |
| 	t.Helper()
 | |
| 
 | |
| 	var wg sync.WaitGroup
 | |
| 
 | |
| 	for _, client := range clients {
 | |
| 		wg.Add(1)
 | |
| 		c := client // Avoid loop pointer
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			assertValidStatus(t, c)
 | |
| 			assertValidNetcheck(t, c)
 | |
| 			assertValidNetmap(t, c)
 | |
| 		}()
 | |
| 	}
 | |
| 
 | |
| 	t.Logf("waiting for client state checks to finish")
 | |
| 	wg.Wait()
 | |
| }
 | |
| 
 | |
| // assertValidNetmap asserts that the netmap of a client has all
 | |
| // the minimum required fields set to a known working config for
 | |
| // the general case. Fields are checked on self, then all peers.
 | |
| // This test is not suitable for ACL/partial connection tests.
 | |
| // This test can only be run on clients from 1.56.1. It will
 | |
| // automatically pass all clients below that and is safe to call
 | |
| // for all versions.
 | |
| func assertValidNetmap(t *testing.T, client TailscaleClient) {
 | |
| 	t.Helper()
 | |
| 
 | |
| 	if !util.TailscaleVersionNewerOrEqual("1.56", client.Version()) {
 | |
| 		t.Logf("%q has version %q, skipping netmap check...", client.Hostname(), client.Version())
 | |
| 
 | |
| 		return
 | |
| 	}
 | |
| 
 | |
| 	t.Logf("Checking netmap of %q", client.Hostname())
 | |
| 
 | |
| 	netmap, err := client.Netmap()
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("getting netmap for %q: %s", client.Hostname(), err)
 | |
| 	}
 | |
| 
 | |
| 	assert.Truef(t, netmap.SelfNode.Hostinfo().Valid(), "%q does not have Hostinfo", client.Hostname())
 | |
| 	if hi := netmap.SelfNode.Hostinfo(); hi.Valid() {
 | |
| 		assert.LessOrEqual(t, 1, netmap.SelfNode.Hostinfo().Services().Len(), "%q does not have enough services, got: %v", client.Hostname(), netmap.SelfNode.Hostinfo().Services())
 | |
| 	}
 | |
| 
 | |
| 	assert.NotEmptyf(t, netmap.SelfNode.AllowedIPs(), "%q does not have any allowed IPs", client.Hostname())
 | |
| 	assert.NotEmptyf(t, netmap.SelfNode.Addresses(), "%q does not have any addresses", client.Hostname())
 | |
| 
 | |
| 	if netmap.SelfNode.Online() != nil {
 | |
| 		assert.Truef(t, *netmap.SelfNode.Online(), "%q is not online", client.Hostname())
 | |
| 	} else {
 | |
| 		t.Errorf("Online should not be nil for %s", client.Hostname())
 | |
| 	}
 | |
| 
 | |
| 	assert.Falsef(t, netmap.SelfNode.Key().IsZero(), "%q does not have a valid NodeKey", client.Hostname())
 | |
| 	assert.Falsef(t, netmap.SelfNode.Machine().IsZero(), "%q does not have a valid MachineKey", client.Hostname())
 | |
| 	assert.Falsef(t, netmap.SelfNode.DiscoKey().IsZero(), "%q does not have a valid DiscoKey", client.Hostname())
 | |
| 
 | |
| 	for _, peer := range netmap.Peers {
 | |
| 		assert.NotEqualf(t, "127.3.3.40:0", peer.DERP(), "peer (%s) has no home DERP in %q's netmap, got: %s", peer.ComputedName(), client.Hostname(), peer.DERP())
 | |
| 
 | |
| 		assert.Truef(t, peer.Hostinfo().Valid(), "peer (%s) of %q does not have Hostinfo", peer.ComputedName(), client.Hostname())
 | |
| 		if hi := peer.Hostinfo(); hi.Valid() {
 | |
| 			assert.LessOrEqualf(t, 3, peer.Hostinfo().Services().Len(), "peer (%s) of %q does not have enough services, got: %v", peer.ComputedName(), client.Hostname(), peer.Hostinfo().Services())
 | |
| 
 | |
| 			// Netinfo is not always set
 | |
| 			// assert.Truef(t, hi.NetInfo().Valid(), "peer (%s) of %q does not have NetInfo", peer.ComputedName(), client.Hostname())
 | |
| 			if ni := hi.NetInfo(); ni.Valid() {
 | |
| 				assert.NotEqualf(t, 0, ni.PreferredDERP(), "peer (%s) has no home DERP in %q's netmap, got: %s", peer.ComputedName(), client.Hostname(), peer.Hostinfo().NetInfo().PreferredDERP())
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		assert.NotEmptyf(t, peer.Endpoints(), "peer (%s) of %q does not have any endpoints", peer.ComputedName(), client.Hostname())
 | |
| 		assert.NotEmptyf(t, peer.AllowedIPs(), "peer (%s) of %q does not have any allowed IPs", peer.ComputedName(), client.Hostname())
 | |
| 		assert.NotEmptyf(t, peer.Addresses(), "peer (%s) of %q does not have any addresses", peer.ComputedName(), client.Hostname())
 | |
| 
 | |
| 		assert.Truef(t, *peer.Online(), "peer (%s) of %q is not online", peer.ComputedName(), client.Hostname())
 | |
| 
 | |
| 		assert.Falsef(t, peer.Key().IsZero(), "peer (%s) of %q does not have a valid NodeKey", peer.ComputedName(), client.Hostname())
 | |
| 		assert.Falsef(t, peer.Machine().IsZero(), "peer (%s) of %q does not have a valid MachineKey", peer.ComputedName(), client.Hostname())
 | |
| 		assert.Falsef(t, peer.DiscoKey().IsZero(), "peer (%s) of %q does not have a valid DiscoKey", peer.ComputedName(), client.Hostname())
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // assertValidStatus asserts that the status of a client has all
 | |
| // the minimum required fields set to a known working config for
 | |
| // the general case. Fields are checked on self, then all peers.
 | |
| // This test is not suitable for ACL/partial connection tests.
 | |
| func assertValidStatus(t *testing.T, client TailscaleClient) {
 | |
| 	t.Helper()
 | |
| 	status, err := client.Status(true)
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("getting status for %q: %s", client.Hostname(), err)
 | |
| 	}
 | |
| 
 | |
| 	assert.NotEmptyf(t, status.Self.HostName, "%q does not have HostName set, likely missing Hostinfo", client.Hostname())
 | |
| 	assert.NotEmptyf(t, status.Self.OS, "%q does not have OS set, likely missing Hostinfo", client.Hostname())
 | |
| 	assert.NotEmptyf(t, status.Self.Relay, "%q does not have a relay, likely missing Hostinfo/Netinfo", client.Hostname())
 | |
| 
 | |
| 	assert.NotEmptyf(t, status.Self.TailscaleIPs, "%q does not have Tailscale IPs", client.Hostname())
 | |
| 
 | |
| 	// This seem to not appear until version 1.56
 | |
| 	if status.Self.AllowedIPs != nil {
 | |
| 		assert.NotEmptyf(t, status.Self.AllowedIPs, "%q does not have any allowed IPs", client.Hostname())
 | |
| 	}
 | |
| 
 | |
| 	assert.NotEmptyf(t, status.Self.Addrs, "%q does not have any endpoints", client.Hostname())
 | |
| 
 | |
| 	assert.Truef(t, status.Self.Online, "%q is not online", client.Hostname())
 | |
| 
 | |
| 	assert.Truef(t, status.Self.InNetworkMap, "%q is not in network map", client.Hostname())
 | |
| 
 | |
| 	// This isnt really relevant for Self as it wont be in its own socket/wireguard.
 | |
| 	// assert.Truef(t, status.Self.InMagicSock, "%q is not tracked by magicsock", client.Hostname())
 | |
| 	// assert.Truef(t, status.Self.InEngine, "%q is not in in wireguard engine", client.Hostname())
 | |
| 
 | |
| 	for _, peer := range status.Peer {
 | |
| 		assert.NotEmptyf(t, peer.HostName, "peer (%s) of %q does not have HostName set, likely missing Hostinfo", peer.DNSName, client.Hostname())
 | |
| 		assert.NotEmptyf(t, peer.OS, "peer (%s) of %q does not have OS set, likely missing Hostinfo", peer.DNSName, client.Hostname())
 | |
| 		assert.NotEmptyf(t, peer.Relay, "peer (%s) of %q does not have a relay, likely missing Hostinfo/Netinfo", peer.DNSName, client.Hostname())
 | |
| 
 | |
| 		assert.NotEmptyf(t, peer.TailscaleIPs, "peer (%s) of %q does not have Tailscale IPs", peer.DNSName, client.Hostname())
 | |
| 
 | |
| 		// This seem to not appear until version 1.56
 | |
| 		if peer.AllowedIPs != nil {
 | |
| 			assert.NotEmptyf(t, peer.AllowedIPs, "peer (%s) of %q does not have any allowed IPs", peer.DNSName, client.Hostname())
 | |
| 		}
 | |
| 
 | |
| 		// Addrs does not seem to appear in the status from peers.
 | |
| 		// assert.NotEmptyf(t, peer.Addrs, "peer (%s) of %q does not have any endpoints", peer.DNSName, client.Hostname())
 | |
| 
 | |
| 		assert.Truef(t, peer.Online, "peer (%s) of %q is not online", peer.DNSName, client.Hostname())
 | |
| 
 | |
| 		assert.Truef(t, peer.InNetworkMap, "peer (%s) of %q is not in network map", peer.DNSName, client.Hostname())
 | |
| 		assert.Truef(t, peer.InMagicSock, "peer (%s) of %q is not tracked by magicsock", peer.DNSName, client.Hostname())
 | |
| 
 | |
| 		// TODO(kradalby): InEngine is only true when a proper tunnel is set up,
 | |
| 		// there might be some interesting stuff to test here in the future.
 | |
| 		// assert.Truef(t, peer.InEngine, "peer (%s) of %q is not in wireguard engine", peer.DNSName, client.Hostname())
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func assertValidNetcheck(t *testing.T, client TailscaleClient) {
 | |
| 	t.Helper()
 | |
| 	report, err := client.Netcheck()
 | |
| 	if err != nil {
 | |
| 		t.Fatalf("getting status for %q: %s", client.Hostname(), err)
 | |
| 	}
 | |
| 
 | |
| 	assert.NotEqualf(t, 0, report.PreferredDERP, "%q does not have a DERP relay", client.Hostname())
 | |
| }
 | |
| 
 | |
| func isSelfClient(client TailscaleClient, addr string) bool {
 | |
| 	if addr == client.Hostname() {
 | |
| 		return true
 | |
| 	}
 | |
| 
 | |
| 	ips, err := client.IPs()
 | |
| 	if err != nil {
 | |
| 		return false
 | |
| 	}
 | |
| 
 | |
| 	for _, ip := range ips {
 | |
| 		if ip.String() == addr {
 | |
| 			return true
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func isCI() bool {
 | |
| 	if _, ok := os.LookupEnv("CI"); ok {
 | |
| 		return true
 | |
| 	}
 | |
| 
 | |
| 	if _, ok := os.LookupEnv("GITHUB_RUN_ID"); ok {
 | |
| 		return true
 | |
| 	}
 | |
| 
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func dockertestMaxWait() time.Duration {
 | |
| 	wait := 120 * time.Second //nolint
 | |
| 
 | |
| 	if isCI() {
 | |
| 		wait = 300 * time.Second //nolint
 | |
| 	}
 | |
| 
 | |
| 	return wait
 | |
| }
 | |
| 
 | |
| func countMatchingLines(in io.Reader, predicate func(string) bool) (int, error) {
 | |
| 	count := 0
 | |
| 	scanner := bufio.NewScanner(in)
 | |
| 	{
 | |
| 		const logBufferInitialSize = 1024 << 10 // preallocate 1 MiB
 | |
| 		buff := make([]byte, logBufferInitialSize)
 | |
| 		scanner.Buffer(buff, len(buff))
 | |
| 		scanner.Split(bufio.ScanLines)
 | |
| 	}
 | |
| 
 | |
| 	for scanner.Scan() {
 | |
| 		if predicate(scanner.Text()) {
 | |
| 			count += 1
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return count, scanner.Err()
 | |
| }
 | |
| 
 | |
| // func dockertestCommandTimeout() time.Duration {
 | |
| // 	timeout := 10 * time.Second //nolint
 | |
| //
 | |
| // 	if isCI() {
 | |
| // 		timeout = 60 * time.Second //nolint
 | |
| // 	}
 | |
| //
 | |
| // 	return timeout
 | |
| // }
 | |
| 
 | |
| // pingAllNegativeHelper is intended to have 1 or more nodes timing out from the ping,
 | |
| // it counts failures instead of successes.
 | |
| // func pingAllNegativeHelper(t *testing.T, clients []TailscaleClient, addrs []string) int {
 | |
| // 	t.Helper()
 | |
| // 	failures := 0
 | |
| //
 | |
| // 	timeout := 100
 | |
| // 	count := 3
 | |
| //
 | |
| // 	for _, client := range clients {
 | |
| // 		for _, addr := range addrs {
 | |
| // 			err := client.Ping(
 | |
| // 				addr,
 | |
| // 				tsic.WithPingTimeout(time.Duration(timeout)*time.Millisecond),
 | |
| // 				tsic.WithPingCount(count),
 | |
| // 			)
 | |
| // 			if err != nil {
 | |
| // 				failures++
 | |
| // 			}
 | |
| // 		}
 | |
| // 	}
 | |
| //
 | |
| // 	return failures
 | |
| // }
 | |
| 
 | |
| // // findPeerByIP takes an IP and a map of peers from status.Peer, and returns a *ipnstate.PeerStatus
 | |
| // // if there is a peer with the given IP. If no peer is found, nil is returned.
 | |
| // func findPeerByIP(
 | |
| // 	ip netip.Addr,
 | |
| // 	peers map[key.NodePublic]*ipnstate.PeerStatus,
 | |
| // ) *ipnstate.PeerStatus {
 | |
| // 	for _, peer := range peers {
 | |
| // 		for _, peerIP := range peer.TailscaleIPs {
 | |
| // 			if ip == peerIP {
 | |
| // 				return peer
 | |
| // 			}
 | |
| // 		}
 | |
| // 	}
 | |
| //
 | |
| // 	return nil
 | |
| // }
 | |
| //
 | |
| // // findPeerByHostname takes a hostname and a map of peers from status.Peer, and returns a *ipnstate.PeerStatus
 | |
| // // if there is a peer with the given hostname. If no peer is found, nil is returned.
 | |
| // func findPeerByHostname(
 | |
| // 	hostname string,
 | |
| // 	peers map[key.NodePublic]*ipnstate.PeerStatus,
 | |
| // ) *ipnstate.PeerStatus {
 | |
| // 	for _, peer := range peers {
 | |
| // 		if hostname == peer.HostName {
 | |
| // 			return peer
 | |
| // 		}
 | |
| // 	}
 | |
| //
 | |
| // 	return nil
 | |
| // }
 |