mirror of
https://github.com/status-im/consul.git
synced 2025-01-25 13:10:32 +00:00
5fb9df1640
* Adding explicit MPL license for sub-package This directory and its subdirectories (packages) contain files licensed with the MPLv2 `LICENSE` file in this directory and are intentionally licensed separately from the BSL `LICENSE` file at the root of this repository. * Adding explicit MPL license for sub-package This directory and its subdirectories (packages) contain files licensed with the MPLv2 `LICENSE` file in this directory and are intentionally licensed separately from the BSL `LICENSE` file at the root of this repository. * Updating the license from MPL to Business Source License Going forward, this project will be licensed under the Business Source License v1.1. Please see our blog post for more details at <Blog URL>, FAQ at www.hashicorp.com/licensing-faq, and details of the license at www.hashicorp.com/bsl. * add missing license headers * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 --------- Co-authored-by: hashicorp-copywrite[bot] <110428419+hashicorp-copywrite[bot]@users.noreply.github.com>
700 lines
18 KiB
Go
700 lines
18 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: BUSL-1.1
|
|
|
|
package cluster
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/hashicorp/consul/api"
|
|
"github.com/hashicorp/consul/sdk/testutil/retry"
|
|
"github.com/hashicorp/consul/test/integration/consul-container/libs/utils"
|
|
"github.com/hashicorp/serf/serf"
|
|
|
|
goretry "github.com/avast/retry-go"
|
|
"github.com/stretchr/testify/require"
|
|
"github.com/teris-io/shortid"
|
|
"github.com/testcontainers/testcontainers-go"
|
|
)
|
|
|
|
// Cluster provides an interface for creating and controlling a Consul cluster
|
|
// in integration tests, with agents running in containers.
|
|
// These fields are public in the event someone might want to surgically
|
|
// craft a test case.
|
|
type Cluster struct {
|
|
Agents []Agent
|
|
// BuildContext *BuildContext // TODO
|
|
CACert string
|
|
CAKey string
|
|
ID string
|
|
Index int
|
|
Network testcontainers.Network
|
|
NetworkName string
|
|
ScratchDir string
|
|
|
|
ACLEnabled bool
|
|
TokenBootstrap string
|
|
}
|
|
|
|
type TestingT interface {
|
|
Logf(format string, args ...any)
|
|
Cleanup(f func())
|
|
}
|
|
|
|
func NewN(t TestingT, conf Config, count int) (*Cluster, error) {
|
|
var configs []Config
|
|
for i := 0; i < count; i++ {
|
|
configs = append(configs, conf)
|
|
}
|
|
|
|
return New(t, configs)
|
|
}
|
|
|
|
// New creates a Consul cluster. An agent will be started for each of the given
|
|
// configs and joined to the cluster.
|
|
//
|
|
// A cluster has its own docker network for DNS connectivity, but is also
|
|
// joined
|
|
//
|
|
// The provided TestingT is used to register a cleanup function to terminate
|
|
// the cluster.
|
|
func New(t TestingT, configs []Config, ports ...int) (*Cluster, error) {
|
|
id, err := shortid.Generate()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("could not generate cluster id: %w", err)
|
|
}
|
|
|
|
name := fmt.Sprintf("consul-int-cluster-%s", id)
|
|
network, err := createNetwork(t, name)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("could not create cluster container network: %w", err)
|
|
}
|
|
|
|
// Rig up one scratch dir for the cluster with auto-cleanup on test exit.
|
|
scratchDir, err := os.MkdirTemp("", name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
t.Cleanup(func() {
|
|
_ = os.RemoveAll(scratchDir)
|
|
})
|
|
err = os.Chmod(scratchDir, 0777)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
cluster := &Cluster{
|
|
ID: id,
|
|
Network: network,
|
|
NetworkName: name,
|
|
ScratchDir: scratchDir,
|
|
ACLEnabled: configs[0].ACLEnabled,
|
|
}
|
|
t.Cleanup(func() {
|
|
_ = cluster.Terminate()
|
|
})
|
|
|
|
if err := cluster.Add(configs, true, ports...); err != nil {
|
|
return nil, fmt.Errorf("could not start or join all agents: %w", err)
|
|
}
|
|
|
|
return cluster, nil
|
|
}
|
|
|
|
func (c *Cluster) AddN(conf Config, count int, join bool) error {
|
|
var configs []Config
|
|
for i := 0; i < count; i++ {
|
|
configs = append(configs, conf)
|
|
}
|
|
return c.Add(configs, join)
|
|
}
|
|
|
|
// Add starts agents with the given configurations and joins them to the existing cluster
|
|
func (c *Cluster) Add(configs []Config, serfJoin bool, ports ...int) (xe error) {
|
|
if c.Index == 0 && !serfJoin {
|
|
return fmt.Errorf("the first call to Cluster.Add must have serfJoin=true")
|
|
}
|
|
|
|
var agents []Agent
|
|
for idx, conf := range configs {
|
|
// Each agent gets it's own area in the cluster scratch.
|
|
conf.ScratchDir = filepath.Join(c.ScratchDir, strconv.Itoa(c.Index))
|
|
if err := os.MkdirAll(conf.ScratchDir, 0777); err != nil {
|
|
return fmt.Errorf("container %d making scratchDir: %w", idx, err)
|
|
}
|
|
if err := os.Chmod(conf.ScratchDir, 0777); err != nil {
|
|
return fmt.Errorf("container %d perms on scratchDir: %w", idx, err)
|
|
}
|
|
|
|
var n Agent
|
|
|
|
// retry creating client every ten seconds. with local development, we've found
|
|
// that this "port not found" error occurs when runs happen too close together
|
|
if err := goretry.Do(
|
|
func() (err error) {
|
|
n, err = NewConsulContainer(
|
|
context.Background(),
|
|
conf,
|
|
c,
|
|
ports...,
|
|
)
|
|
return err
|
|
},
|
|
goretry.Delay(10*time.Second),
|
|
goretry.RetryIf(func(err error) bool {
|
|
return strings.Contains(err.Error(), "port not found")
|
|
}),
|
|
); err != nil {
|
|
return fmt.Errorf("container %d creating: %s", idx, err)
|
|
}
|
|
|
|
agents = append(agents, n)
|
|
c.Index++
|
|
}
|
|
|
|
if serfJoin {
|
|
if err := c.Join(agents); err != nil {
|
|
return fmt.Errorf("could not join agents to cluster: %w", err)
|
|
}
|
|
} else {
|
|
if err := c.JoinExternally(agents); err != nil {
|
|
return fmt.Errorf("could not join agents to cluster: %w", err)
|
|
}
|
|
}
|
|
|
|
if utils.Debug {
|
|
c.PrintDebugInfo(agents)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Join joins the given agent to the cluster.
|
|
func (c *Cluster) Join(agents []Agent) error {
|
|
return c.join(agents, false)
|
|
}
|
|
|
|
func (c *Cluster) JoinExternally(agents []Agent) error {
|
|
return c.join(agents, true)
|
|
}
|
|
|
|
func (c *Cluster) join(agents []Agent, skipSerfJoin bool) error {
|
|
if len(agents) == 0 {
|
|
return nil // no change
|
|
}
|
|
|
|
if len(c.Agents) == 0 {
|
|
// if acl enabled, generate the bootstrap tokens at the first agent
|
|
if c.ACLEnabled {
|
|
var (
|
|
output string
|
|
err error
|
|
)
|
|
// retry since agent needs to start the ACL system
|
|
err = goretry.Do(
|
|
func() error {
|
|
output, err = agents[0].Exec(context.Background(), []string{"consul", "acl", "bootstrap"})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
},
|
|
goretry.Delay(time.Second*1),
|
|
)
|
|
if err != nil {
|
|
return fmt.Errorf("error generating the bootstrap token, %s", err)
|
|
}
|
|
c.TokenBootstrap, err = extractSecretIDFrom(output)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
fmt.Println("Cluster bootstrap token:", c.TokenBootstrap)
|
|
|
|
// The first node's default client needs to be updated after bootstrap token
|
|
// is created
|
|
_, err = agents[0].NewClient(c.TokenBootstrap, true)
|
|
if err != nil {
|
|
return fmt.Errorf("error updating the first node's client, %s", err)
|
|
}
|
|
}
|
|
|
|
// Join the rest to the first.
|
|
c.Agents = append(c.Agents, agents[0])
|
|
return c.join(agents[1:], skipSerfJoin)
|
|
}
|
|
|
|
// Always join to the original server.
|
|
joinAddr := c.Agents[0].GetIP()
|
|
|
|
for _, n := range agents {
|
|
if !skipSerfJoin {
|
|
// retry in case the agent token is updated at the agent
|
|
err := goretry.Do(
|
|
func() error {
|
|
err := n.GetClient().Agent().Join(joinAddr, false)
|
|
if err != nil {
|
|
return fmt.Errorf("could not join agent %s to %s: %w", n.GetName(), joinAddr, err)
|
|
}
|
|
return nil
|
|
},
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
c.Agents = append(c.Agents, n)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (c *Cluster) CreateAgentToken(datacenter string, agentName string) (string, error) {
|
|
output, err := c.Agents[0].Exec(context.Background(), []string{"consul", "acl", "token", "create", "-description", "\"agent token\"",
|
|
"-token", c.TokenBootstrap,
|
|
"-node-identity", fmt.Sprintf("%s:%s", agentName, datacenter)})
|
|
if err != nil {
|
|
return "", fmt.Errorf("after retry, error generating agent token, %s", err)
|
|
}
|
|
secretID, err := extractSecretIDFrom(output)
|
|
return secretID, err
|
|
}
|
|
|
|
// Remove instructs the agent to leave the cluster then removes it
|
|
// from the cluster Agent list.
|
|
func (c *Cluster) Remove(n Agent) error {
|
|
err := n.GetClient().Agent().Leave()
|
|
if err != nil {
|
|
return fmt.Errorf("could not remove agent %s: %w", n.GetName(), err)
|
|
}
|
|
|
|
foundIdx := -1
|
|
for idx, this := range c.Agents {
|
|
if this == n {
|
|
foundIdx = idx
|
|
break
|
|
}
|
|
}
|
|
|
|
if foundIdx == -1 {
|
|
return errors.New("could not find agent in cluster")
|
|
}
|
|
|
|
c.Agents = append(c.Agents[:foundIdx], c.Agents[foundIdx+1:]...)
|
|
return nil
|
|
}
|
|
|
|
// StandardUpgrade upgrades a running consul cluster following the steps from
|
|
//
|
|
// https://developer.hashicorp.com/consul/docs/upgrading#standard-upgrades
|
|
//
|
|
// - takes a snapshot (which is discarded)
|
|
// - terminate and rejoin the pod of a new version of consul
|
|
//
|
|
// NOTE: we pass in a *testing.T but this method also returns an error. JUST
|
|
// within this method when in doubt return an error. A testing assertion should
|
|
// be saved only for using t.Cleanup or in a few of the retry-until-working
|
|
// helpers below.
|
|
//
|
|
// This lets us have tests that assert that an upgrade will fail.
|
|
func (c *Cluster) StandardUpgrade(t *testing.T, ctx context.Context, targetImage string, targetVersion string) error {
|
|
var err error
|
|
// We take a snapshot, but note that we currently do nothing with it.
|
|
if c.ACLEnabled {
|
|
_, err = c.Agents[0].Exec(context.Background(), []string{"consul", "snapshot", "save",
|
|
"-token", c.TokenBootstrap, "backup.snap"})
|
|
} else {
|
|
_, err = c.Agents[0].Exec(context.Background(), []string{"consul", "snapshot", "save", "backup.snap"})
|
|
}
|
|
if err != nil {
|
|
return fmt.Errorf("error taking the snapshot: %s", err)
|
|
}
|
|
|
|
// Upgrade individual agent to the target version in the following order
|
|
// 1. followers
|
|
// 2. leader
|
|
// 3. clients (TODO)
|
|
|
|
// Grab a client connected to the leader, which we will upgrade last so our
|
|
// connection remains ok.
|
|
leader, err := c.Leader()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
t.Logf("Leader name: %s", leader.GetName())
|
|
|
|
followers, err := c.Followers()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
t.Logf("The number of followers = %d", len(followers))
|
|
|
|
// NOTE: we only assert the number of agents in default partition
|
|
// TODO: add partition to the cluster struct to assert partition size
|
|
clusterSize := 0
|
|
for _, agent := range c.Agents {
|
|
if agent.GetPartition() == "" || agent.GetPartition() == "default" {
|
|
clusterSize++
|
|
}
|
|
}
|
|
t.Logf("The number of agents in default partition = %d", clusterSize)
|
|
|
|
upgradeFn := func(agent Agent, clientFactory func() (*api.Client, error)) error {
|
|
config := agent.GetConfig()
|
|
config.Image = targetImage
|
|
config.Version = targetVersion
|
|
|
|
if agent.IsServer() {
|
|
// You only ever need bootstrap settings the FIRST time, so we do not need
|
|
// them again.
|
|
config.ConfigBuilder.Unset("bootstrap")
|
|
} else {
|
|
// If we upgrade the clients fast enough
|
|
// membership might not be gossiped to all of
|
|
// the clients to persist into their serf
|
|
// snapshot, so force them to rejoin the
|
|
// normal way on restart.
|
|
config.ConfigBuilder.Set("retry_join", []string{"agent-0"})
|
|
}
|
|
|
|
newJSON, err := json.MarshalIndent(config.ConfigBuilder, "", " ")
|
|
if err != nil {
|
|
return fmt.Errorf("could not re-generate json config: %w", err)
|
|
}
|
|
config.JSON = string(newJSON)
|
|
t.Logf("Upgraded cluster config for %q:\n%s", agent.GetName(), config.JSON)
|
|
|
|
err = agent.Upgrade(context.Background(), config)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
client, err := clientFactory()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// wait until the agent rejoin and leader is elected; skip non-default agent
|
|
if agent.GetPartition() == "" || agent.GetPartition() == "default" {
|
|
WaitForMembers(t, client, clusterSize)
|
|
}
|
|
WaitForLeader(t, c, client)
|
|
|
|
return nil
|
|
}
|
|
|
|
for _, agent := range followers {
|
|
t.Logf("Upgrade follower: %s", agent.GetName())
|
|
|
|
err := upgradeFn(agent, func() (*api.Client, error) {
|
|
return leader.GetClient(), nil
|
|
})
|
|
if err != nil {
|
|
return fmt.Errorf("error upgrading follower %q: %w", agent.GetName(), err)
|
|
}
|
|
}
|
|
|
|
t.Logf("Upgrade leader: %s", leader.GetAgentName())
|
|
err = upgradeFn(leader, func() (*api.Client, error) {
|
|
if len(followers) > 0 {
|
|
return followers[0].GetClient(), nil
|
|
}
|
|
return leader.GetClient(), nil
|
|
})
|
|
if err != nil {
|
|
return fmt.Errorf("error upgrading leader %q: %w", leader.GetName(), err)
|
|
}
|
|
|
|
clientAgents := c.Clients()
|
|
for _, agent := range clientAgents {
|
|
t.Logf("Upgrade client agent: %s", agent.GetName())
|
|
|
|
err = upgradeFn(agent, func() (*api.Client, error) {
|
|
leader, err = c.Leader()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return leader.GetClient(), nil
|
|
})
|
|
if err != nil {
|
|
return fmt.Errorf("error upgrading client agent %q: %w", agent.GetName(), err)
|
|
}
|
|
}
|
|
|
|
t.Log("Update completed\n")
|
|
return nil
|
|
}
|
|
|
|
// Terminate will attempt to terminate all agents in the cluster and its network. If any agent
|
|
// termination fails, Terminate will abort and return an error.
|
|
func (c *Cluster) Terminate() error {
|
|
for _, n := range c.Agents {
|
|
err := n.Terminate()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Testcontainers seems to clean this the network.
|
|
// Trigger it now will throw an error while the containers are still shutting down
|
|
// if err := c.Network.Remove(context.Background()); err != nil {
|
|
// return fmt.Errorf("could not terminate cluster network %s: %w", c.ID, err)
|
|
// }
|
|
return nil
|
|
}
|
|
|
|
// Leader returns the cluster leader agent, or an error if no leader is
|
|
// available.
|
|
func (c *Cluster) Leader() (Agent, error) {
|
|
if len(c.Agents) < 1 {
|
|
return nil, fmt.Errorf("no agent available")
|
|
}
|
|
n0 := c.Agents[0]
|
|
|
|
leaderAdd, err := getLeader(n0.GetClient())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, n := range c.Agents {
|
|
addr := n.GetIP()
|
|
if strings.Contains(leaderAdd, addr) {
|
|
return n, nil
|
|
}
|
|
}
|
|
return nil, fmt.Errorf("leader not found")
|
|
}
|
|
|
|
func getLeader(client *api.Client) (string, error) {
|
|
leaderAdd, err := client.Status().Leader()
|
|
if err != nil {
|
|
return "", fmt.Errorf("could not query leader: %w", err)
|
|
}
|
|
if leaderAdd == "" {
|
|
return "", errors.New("no leader available")
|
|
}
|
|
return leaderAdd, nil
|
|
}
|
|
|
|
// Followers returns the cluster following servers.
|
|
func (c *Cluster) Followers() ([]Agent, error) {
|
|
var followers []Agent
|
|
|
|
leader, err := c.Leader()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("could not determine leader: %w", err)
|
|
}
|
|
|
|
for _, n := range c.Agents {
|
|
if n != leader && n.IsServer() {
|
|
followers = append(followers, n)
|
|
}
|
|
}
|
|
return followers, nil
|
|
}
|
|
|
|
// Servers returns the handle to server agents
|
|
func (c *Cluster) Servers() []Agent {
|
|
var servers []Agent
|
|
|
|
for _, n := range c.Agents {
|
|
if n.IsServer() {
|
|
servers = append(servers, n)
|
|
}
|
|
}
|
|
return servers
|
|
}
|
|
|
|
// Clients returns the handle to client agents in provided partition
|
|
func (c *Cluster) ClientsInPartition(partition string) []Agent {
|
|
var clients []Agent
|
|
|
|
for _, n := range c.Agents {
|
|
if n.IsServer() {
|
|
continue
|
|
}
|
|
|
|
if n.GetPartition() == partition {
|
|
clients = append(clients, n)
|
|
}
|
|
}
|
|
return clients
|
|
}
|
|
|
|
// Clients returns the handle to client agents in all partitions
|
|
func (c *Cluster) Clients() []Agent {
|
|
var clients []Agent
|
|
|
|
for _, n := range c.Agents {
|
|
if !n.IsServer() {
|
|
clients = append(clients, n)
|
|
}
|
|
}
|
|
return clients
|
|
}
|
|
|
|
func (c *Cluster) APIClient(index int) *api.Client {
|
|
nodes := c.Clients()
|
|
if len(nodes) == 0 {
|
|
nodes = c.Servers()
|
|
if len(nodes) == 0 {
|
|
return nil
|
|
}
|
|
}
|
|
return nodes[0].GetClient()
|
|
}
|
|
|
|
// GetClient returns a consul API client to the node if node is provided.
|
|
// Otherwise, GetClient returns the API client to the first node of either
|
|
// server or client agent.
|
|
//
|
|
// TODO: see about switching to just APIClient() calls instead?
|
|
func (c *Cluster) GetClient(node Agent, isServer bool) (*api.Client, error) {
|
|
if node != nil {
|
|
return node.GetClient(), nil
|
|
}
|
|
|
|
nodes := c.Clients()
|
|
if isServer {
|
|
nodes = c.Servers()
|
|
}
|
|
|
|
if len(nodes) <= 0 {
|
|
return nil, errors.New("no nodes")
|
|
}
|
|
|
|
return nodes[0].GetClient(), nil
|
|
}
|
|
|
|
// PeerWithCluster establishes peering with the acceptor cluster
|
|
func (c *Cluster) PeerWithCluster(acceptingClient *api.Client, acceptingPeerName string, dialingPeerName string) error {
|
|
dialingClient := c.APIClient(0)
|
|
|
|
generateReq := api.PeeringGenerateTokenRequest{
|
|
PeerName: acceptingPeerName,
|
|
}
|
|
generateRes, _, err := acceptingClient.Peerings().GenerateToken(context.Background(), generateReq, &api.WriteOptions{})
|
|
if err != nil {
|
|
return fmt.Errorf("error generate token: %w", err)
|
|
}
|
|
|
|
establishReq := api.PeeringEstablishRequest{
|
|
PeerName: dialingPeerName,
|
|
PeeringToken: generateRes.PeeringToken,
|
|
}
|
|
_, _, err = dialingClient.Peerings().Establish(context.Background(), establishReq, &api.WriteOptions{})
|
|
if err != nil {
|
|
return fmt.Errorf("error establish peering: %w", err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
const retryTimeout = 90 * time.Second
|
|
const retryFrequency = 500 * time.Millisecond
|
|
|
|
func LongFailer() *retry.Timer {
|
|
return &retry.Timer{Timeout: retryTimeout, Wait: retryFrequency}
|
|
}
|
|
|
|
func WaitForLeader(t *testing.T, cluster *Cluster, client *api.Client) {
|
|
retry.RunWith(LongFailer(), t, func(r *retry.R) {
|
|
leader, err := cluster.Leader()
|
|
require.NoError(r, err)
|
|
require.NotEmpty(r, leader)
|
|
})
|
|
|
|
if client != nil {
|
|
waitForLeaderFromClient(t, client)
|
|
}
|
|
}
|
|
|
|
func waitForLeaderFromClient(t *testing.T, client *api.Client) {
|
|
retry.RunWith(LongFailer(), t, func(r *retry.R) {
|
|
leader, err := getLeader(client)
|
|
require.NoError(r, err)
|
|
require.NotEmpty(r, leader)
|
|
})
|
|
}
|
|
|
|
func WaitForMembers(t *testing.T, client *api.Client, expectN int) {
|
|
retry.RunWith(LongFailer(), t, func(r *retry.R) {
|
|
members, err := client.Agent().Members(false)
|
|
var activeMembers int
|
|
for _, member := range members {
|
|
if serf.MemberStatus(member.Status) == serf.StatusAlive {
|
|
activeMembers++
|
|
}
|
|
}
|
|
require.NoError(r, err)
|
|
require.Equal(r, expectN, activeMembers)
|
|
})
|
|
}
|
|
|
|
func (c *Cluster) ConfigEntryWrite(entry api.ConfigEntry) error {
|
|
client, _ := c.GetClient(nil, true)
|
|
|
|
entries := client.ConfigEntries()
|
|
written := false
|
|
written, _, err := entries.Set(entry, nil)
|
|
if err != nil {
|
|
return fmt.Errorf("error set config entry: %v", err)
|
|
}
|
|
if !written {
|
|
return fmt.Errorf("config entry not updated: %s/%s", entry.GetKind(), entry.GetName())
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (c *Cluster) ConfigEntryDelete(entry api.ConfigEntry) error {
|
|
client, err := c.GetClient(nil, true)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
entries := client.ConfigEntries()
|
|
_, err = entries.Delete(entry.GetKind(), entry.GetName(), nil)
|
|
if err != nil {
|
|
return fmt.Errorf("error deleting config entry: %v", err)
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (c *Cluster) PrintDebugInfo(agents []Agent) {
|
|
for _, a := range agents {
|
|
uri := a.GetInfo().DebugURI
|
|
n := a.GetAgentName()
|
|
s := a.IsServer()
|
|
l := "NA"
|
|
if s {
|
|
leader, err := c.Leader()
|
|
if err == nil {
|
|
if leader == a {
|
|
l = "true"
|
|
} else {
|
|
l = "false"
|
|
}
|
|
}
|
|
}
|
|
fmt.Printf("\ndebug info:: n=%s,s=%t,l=%s,uri=%s\n\n", n, s, l, uri)
|
|
}
|
|
}
|
|
|
|
func extractSecretIDFrom(tokenOutput string) (string, error) {
|
|
lines := strings.Split(tokenOutput, "\n")
|
|
for _, line := range lines {
|
|
if strings.Contains(line, "SecretID") {
|
|
secretIDtoken := strings.Split(line, ":")
|
|
return strings.TrimSpace(secretIDtoken[1]), nil
|
|
}
|
|
}
|
|
return "", fmt.Errorf("can't found secretID in token")
|
|
}
|