// Copyright (c) HashiCorp, Inc. // SPDX-License-Identifier: BUSL-1.1 package cluster import ( "context" "encoding/json" "errors" "fmt" "os" "path/filepath" "strconv" "strings" "testing" "time" "github.com/hashicorp/consul/api" "github.com/hashicorp/consul/sdk/testutil/retry" "github.com/hashicorp/consul/test/integration/consul-container/libs/utils" "github.com/hashicorp/serf/serf" goretry "github.com/avast/retry-go" "github.com/stretchr/testify/require" "github.com/teris-io/shortid" "github.com/testcontainers/testcontainers-go" ) // Cluster provides an interface for creating and controlling a Consul cluster // in integration tests, with agents running in containers. // These fields are public in the event someone might want to surgically // craft a test case. type Cluster struct { Agents []Agent // BuildContext *BuildContext // TODO CACert string CAKey string ID string Index int Network testcontainers.Network NetworkName string ScratchDir string ACLEnabled bool TokenBootstrap string } type TestingT interface { Logf(format string, args ...any) Cleanup(f func()) } func NewN(t TestingT, conf Config, count int) (*Cluster, error) { var configs []Config for i := 0; i < count; i++ { configs = append(configs, conf) } return New(t, configs) } // New creates a Consul cluster. An agent will be started for each of the given // configs and joined to the cluster. // // A cluster has its own docker network for DNS connectivity, but is also // joined // // The provided TestingT is used to register a cleanup function to terminate // the cluster. func New(t TestingT, configs []Config, ports ...int) (*Cluster, error) { id, err := shortid.Generate() if err != nil { return nil, fmt.Errorf("could not generate cluster id: %w", err) } name := fmt.Sprintf("consul-int-cluster-%s", id) network, err := createNetwork(t, name) if err != nil { return nil, fmt.Errorf("could not create cluster container network: %w", err) } // Rig up one scratch dir for the cluster with auto-cleanup on test exit. scratchDir, err := os.MkdirTemp("", name) if err != nil { return nil, err } t.Cleanup(func() { _ = os.RemoveAll(scratchDir) }) err = os.Chmod(scratchDir, 0777) if err != nil { return nil, err } cluster := &Cluster{ ID: id, Network: network, NetworkName: name, ScratchDir: scratchDir, ACLEnabled: configs[0].ACLEnabled, } t.Cleanup(func() { _ = cluster.Terminate() }) if err := cluster.Add(configs, true, ports...); err != nil { return nil, fmt.Errorf("could not start or join all agents: %w", err) } return cluster, nil } func (c *Cluster) AddN(conf Config, count int, join bool) error { var configs []Config for i := 0; i < count; i++ { configs = append(configs, conf) } return c.Add(configs, join) } // Add starts agents with the given configurations and joins them to the existing cluster func (c *Cluster) Add(configs []Config, serfJoin bool, ports ...int) (xe error) { if c.Index == 0 && !serfJoin { return fmt.Errorf("the first call to Cluster.Add must have serfJoin=true") } var agents []Agent for idx, conf := range configs { // Each agent gets it's own area in the cluster scratch. conf.ScratchDir = filepath.Join(c.ScratchDir, strconv.Itoa(c.Index)) if err := os.MkdirAll(conf.ScratchDir, 0777); err != nil { return fmt.Errorf("container %d making scratchDir: %w", idx, err) } if err := os.Chmod(conf.ScratchDir, 0777); err != nil { return fmt.Errorf("container %d perms on scratchDir: %w", idx, err) } var n Agent // retry creating client every ten seconds. with local development, we've found // that this "port not found" error occurs when runs happen too close together if err := goretry.Do( func() (err error) { n, err = NewConsulContainer( context.Background(), conf, c, ports..., ) return err }, goretry.Delay(10*time.Second), goretry.RetryIf(func(err error) bool { return strings.Contains(err.Error(), "port not found") }), ); err != nil { return fmt.Errorf("container %d creating: %s", idx, err) } agents = append(agents, n) c.Index++ } if serfJoin { if err := c.Join(agents); err != nil { return fmt.Errorf("could not join agents to cluster: %w", err) } } else { if err := c.JoinExternally(agents); err != nil { return fmt.Errorf("could not join agents to cluster: %w", err) } } if utils.Debug { c.PrintDebugInfo(agents) } return nil } // Join joins the given agent to the cluster. func (c *Cluster) Join(agents []Agent) error { return c.join(agents, false) } func (c *Cluster) JoinExternally(agents []Agent) error { return c.join(agents, true) } func (c *Cluster) join(agents []Agent, skipSerfJoin bool) error { if len(agents) == 0 { return nil // no change } if len(c.Agents) == 0 { // if acl enabled, generate the bootstrap tokens at the first agent if c.ACLEnabled { var ( output string err error ) // retry since agent needs to start the ACL system err = goretry.Do( func() error { output, err = agents[0].Exec(context.Background(), []string{"consul", "acl", "bootstrap"}) if err != nil { return err } return nil }, goretry.Delay(time.Second*1), ) if err != nil { return fmt.Errorf("error generating the bootstrap token, %s", err) } c.TokenBootstrap, err = extractSecretIDFrom(output) if err != nil { return err } fmt.Println("Cluster bootstrap token:", c.TokenBootstrap) // The first node's default client needs to be updated after bootstrap token // is created _, err = agents[0].NewClient(c.TokenBootstrap, true) if err != nil { return fmt.Errorf("error updating the first node's client, %s", err) } } // Join the rest to the first. c.Agents = append(c.Agents, agents[0]) return c.join(agents[1:], skipSerfJoin) } // Always join to the original server. joinAddr := c.Agents[0].GetIP() for _, n := range agents { if !skipSerfJoin { // retry in case the agent token is updated at the agent err := goretry.Do( func() error { err := n.GetClient().Agent().Join(joinAddr, false) if err != nil { return fmt.Errorf("could not join agent %s to %s: %w", n.GetName(), joinAddr, err) } return nil }, ) if err != nil { return err } } c.Agents = append(c.Agents, n) } return nil } func (c *Cluster) CreateAgentToken(datacenter string, agentName string) (string, error) { output, err := c.Agents[0].Exec(context.Background(), []string{"consul", "acl", "token", "create", "-description", "\"agent token\"", "-token", c.TokenBootstrap, "-node-identity", fmt.Sprintf("%s:%s", agentName, datacenter)}) if err != nil { return "", fmt.Errorf("after retry, error generating agent token, %s", err) } secretID, err := extractSecretIDFrom(output) return secretID, err } // Remove instructs the agent to leave the cluster then removes it // from the cluster Agent list. func (c *Cluster) Remove(n Agent) error { err := n.GetClient().Agent().Leave() if err != nil { return fmt.Errorf("could not remove agent %s: %w", n.GetName(), err) } foundIdx := -1 for idx, this := range c.Agents { if this == n { foundIdx = idx break } } if foundIdx == -1 { return errors.New("could not find agent in cluster") } c.Agents = append(c.Agents[:foundIdx], c.Agents[foundIdx+1:]...) return nil } // StandardUpgrade upgrades a running consul cluster following the steps from // // https://developer.hashicorp.com/consul/docs/upgrading#standard-upgrades // // - takes a snapshot (which is discarded) // - terminate and rejoin the pod of a new version of consul // // NOTE: we pass in a *testing.T but this method also returns an error. JUST // within this method when in doubt return an error. A testing assertion should // be saved only for using t.Cleanup or in a few of the retry-until-working // helpers below. // // This lets us have tests that assert that an upgrade will fail. func (c *Cluster) StandardUpgrade(t *testing.T, ctx context.Context, targetImage string, targetVersion string) error { var err error // We take a snapshot, but note that we currently do nothing with it. if c.ACLEnabled { _, err = c.Agents[0].Exec(context.Background(), []string{"consul", "snapshot", "save", "-token", c.TokenBootstrap, "backup.snap"}) } else { _, err = c.Agents[0].Exec(context.Background(), []string{"consul", "snapshot", "save", "backup.snap"}) } if err != nil { return fmt.Errorf("error taking the snapshot: %s", err) } // Upgrade individual agent to the target version in the following order // 1. followers // 2. leader // 3. clients (TODO) // Grab a client connected to the leader, which we will upgrade last so our // connection remains ok. leader, err := c.Leader() if err != nil { return err } t.Logf("Leader name: %s", leader.GetName()) followers, err := c.Followers() if err != nil { return err } t.Logf("The number of followers = %d", len(followers)) // NOTE: we only assert the number of agents in default partition // TODO: add partition to the cluster struct to assert partition size clusterSize := 0 for _, agent := range c.Agents { if agent.GetPartition() == "" || agent.GetPartition() == "default" { clusterSize++ } } t.Logf("The number of agents in default partition = %d", clusterSize) upgradeFn := func(agent Agent, clientFactory func() (*api.Client, error)) error { config := agent.GetConfig() config.Image = targetImage config.Version = targetVersion if agent.IsServer() { // You only ever need bootstrap settings the FIRST time, so we do not need // them again. config.ConfigBuilder.Unset("bootstrap") } else { // If we upgrade the clients fast enough // membership might not be gossiped to all of // the clients to persist into their serf // snapshot, so force them to rejoin the // normal way on restart. config.ConfigBuilder.Set("retry_join", []string{"agent-0"}) } newJSON, err := json.MarshalIndent(config.ConfigBuilder, "", " ") if err != nil { return fmt.Errorf("could not re-generate json config: %w", err) } config.JSON = string(newJSON) t.Logf("Upgraded cluster config for %q:\n%s", agent.GetName(), config.JSON) err = agent.Upgrade(context.Background(), config) if err != nil { return err } client, err := clientFactory() if err != nil { return err } // wait until the agent rejoin and leader is elected; skip non-default agent if agent.GetPartition() == "" || agent.GetPartition() == "default" { WaitForMembers(t, client, clusterSize) } WaitForLeader(t, c, client) return nil } for _, agent := range followers { t.Logf("Upgrade follower: %s", agent.GetName()) err := upgradeFn(agent, func() (*api.Client, error) { return leader.GetClient(), nil }) if err != nil { return fmt.Errorf("error upgrading follower %q: %w", agent.GetName(), err) } } t.Logf("Upgrade leader: %s", leader.GetAgentName()) err = upgradeFn(leader, func() (*api.Client, error) { if len(followers) > 0 { return followers[0].GetClient(), nil } return leader.GetClient(), nil }) if err != nil { return fmt.Errorf("error upgrading leader %q: %w", leader.GetName(), err) } clientAgents := c.Clients() for _, agent := range clientAgents { t.Logf("Upgrade client agent: %s", agent.GetName()) err = upgradeFn(agent, func() (*api.Client, error) { leader, err = c.Leader() if err != nil { return nil, err } return leader.GetClient(), nil }) if err != nil { return fmt.Errorf("error upgrading client agent %q: %w", agent.GetName(), err) } } t.Log("Update completed\n") return nil } // Terminate will attempt to terminate all agents in the cluster and its network. If any agent // termination fails, Terminate will abort and return an error. func (c *Cluster) Terminate() error { for _, n := range c.Agents { err := n.Terminate() if err != nil { return err } } // Testcontainers seems to clean this the network. // Trigger it now will throw an error while the containers are still shutting down // if err := c.Network.Remove(context.Background()); err != nil { // return fmt.Errorf("could not terminate cluster network %s: %w", c.ID, err) // } return nil } // Leader returns the cluster leader agent, or an error if no leader is // available. func (c *Cluster) Leader() (Agent, error) { if len(c.Agents) < 1 { return nil, fmt.Errorf("no agent available") } n0 := c.Agents[0] leaderAdd, err := getLeader(n0.GetClient()) if err != nil { return nil, err } for _, n := range c.Agents { addr := n.GetIP() if strings.Contains(leaderAdd, addr) { return n, nil } } return nil, fmt.Errorf("leader not found") } func getLeader(client *api.Client) (string, error) { leaderAdd, err := client.Status().Leader() if err != nil { return "", fmt.Errorf("could not query leader: %w", err) } if leaderAdd == "" { return "", errors.New("no leader available") } return leaderAdd, nil } // Followers returns the cluster following servers. func (c *Cluster) Followers() ([]Agent, error) { var followers []Agent leader, err := c.Leader() if err != nil { return nil, fmt.Errorf("could not determine leader: %w", err) } for _, n := range c.Agents { if n != leader && n.IsServer() { followers = append(followers, n) } } return followers, nil } // Servers returns the handle to server agents func (c *Cluster) Servers() []Agent { var servers []Agent for _, n := range c.Agents { if n.IsServer() { servers = append(servers, n) } } return servers } // Clients returns the handle to client agents in provided partition func (c *Cluster) ClientsInPartition(partition string) []Agent { var clients []Agent for _, n := range c.Agents { if n.IsServer() { continue } if n.GetPartition() == partition { clients = append(clients, n) } } return clients } // Clients returns the handle to client agents in all partitions func (c *Cluster) Clients() []Agent { var clients []Agent for _, n := range c.Agents { if !n.IsServer() { clients = append(clients, n) } } return clients } func (c *Cluster) APIClient(index int) *api.Client { nodes := c.Clients() if len(nodes) == 0 { nodes = c.Servers() if len(nodes) == 0 { return nil } } return nodes[0].GetClient() } // GetClient returns a consul API client to the node if node is provided. // Otherwise, GetClient returns the API client to the first node of either // server or client agent. // // TODO: see about switching to just APIClient() calls instead? func (c *Cluster) GetClient(node Agent, isServer bool) (*api.Client, error) { if node != nil { return node.GetClient(), nil } nodes := c.Clients() if isServer { nodes = c.Servers() } if len(nodes) <= 0 { return nil, errors.New("no nodes") } return nodes[0].GetClient(), nil } // PeerWithCluster establishes peering with the acceptor cluster func (c *Cluster) PeerWithCluster(acceptingClient *api.Client, acceptingPeerName string, dialingPeerName string) error { dialingClient := c.APIClient(0) generateReq := api.PeeringGenerateTokenRequest{ PeerName: acceptingPeerName, } generateRes, _, err := acceptingClient.Peerings().GenerateToken(context.Background(), generateReq, &api.WriteOptions{}) if err != nil { return fmt.Errorf("error generate token: %w", err) } establishReq := api.PeeringEstablishRequest{ PeerName: dialingPeerName, PeeringToken: generateRes.PeeringToken, } _, _, err = dialingClient.Peerings().Establish(context.Background(), establishReq, &api.WriteOptions{}) if err != nil { return fmt.Errorf("error establish peering: %w", err) } return nil } const retryTimeout = 90 * time.Second const retryFrequency = 500 * time.Millisecond func LongFailer() *retry.Timer { return &retry.Timer{Timeout: retryTimeout, Wait: retryFrequency} } func WaitForLeader(t *testing.T, cluster *Cluster, client *api.Client) { retry.RunWith(LongFailer(), t, func(r *retry.R) { leader, err := cluster.Leader() require.NoError(r, err) require.NotEmpty(r, leader) }) if client != nil { waitForLeaderFromClient(t, client) } } func waitForLeaderFromClient(t *testing.T, client *api.Client) { retry.RunWith(LongFailer(), t, func(r *retry.R) { leader, err := getLeader(client) require.NoError(r, err) require.NotEmpty(r, leader) }) } func WaitForMembers(t *testing.T, client *api.Client, expectN int) { retry.RunWith(LongFailer(), t, func(r *retry.R) { members, err := client.Agent().Members(false) var activeMembers int for _, member := range members { if serf.MemberStatus(member.Status) == serf.StatusAlive { activeMembers++ } } require.NoError(r, err) require.Equal(r, expectN, activeMembers) }) } func (c *Cluster) ConfigEntryWrite(entry api.ConfigEntry) error { client, _ := c.GetClient(nil, true) entries := client.ConfigEntries() written := false written, _, err := entries.Set(entry, nil) if err != nil { return fmt.Errorf("error set config entry: %v", err) } if !written { return fmt.Errorf("config entry not updated: %s/%s", entry.GetKind(), entry.GetName()) } return err } func (c *Cluster) ConfigEntryDelete(entry api.ConfigEntry) error { client, err := c.GetClient(nil, true) if err != nil { return err } entries := client.ConfigEntries() _, err = entries.Delete(entry.GetKind(), entry.GetName(), nil) if err != nil { return fmt.Errorf("error deleting config entry: %v", err) } return err } func (c *Cluster) PrintDebugInfo(agents []Agent) { for _, a := range agents { uri := a.GetInfo().DebugURI n := a.GetAgentName() s := a.IsServer() l := "NA" if s { leader, err := c.Leader() if err == nil { if leader == a { l = "true" } else { l = "false" } } } fmt.Printf("\ndebug info:: n=%s,s=%t,l=%s,uri=%s\n\n", n, s, l, uri) } } func extractSecretIDFrom(tokenOutput string) (string, error) { lines := strings.Split(tokenOutput, "\n") for _, line := range lines { if strings.Contains(line, "SecretID") { secretIDtoken := strings.Split(line, ":") return strings.TrimSpace(secretIDtoken[1]), nil } } return "", fmt.Errorf("can't found secretID in token") }