mirror of https://github.com/status-im/consul.git
upgrade test: use retry with ModifyIndex and remove ent test file (#16553)
This commit is contained in:
parent
a5b8256111
commit
b649a5e8e4
|
@ -6,6 +6,7 @@ import (
|
|||
"testing"
|
||||
"time"
|
||||
|
||||
goretry "github.com/avast/retry-go"
|
||||
"github.com/stretchr/testify/require"
|
||||
|
||||
"github.com/hashicorp/consul/api"
|
||||
|
@ -75,22 +76,33 @@ func TestStandardUpgradeToTarget_fromLatest(t *testing.T) {
|
|||
const serviceName = "api"
|
||||
index := libservice.ServiceCreate(t, client, serviceName)
|
||||
|
||||
ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index)
|
||||
require.NoError(t, client.Agent().ServiceRegister(
|
||||
&api.AgentServiceRegistration{Name: serviceName, Port: 9998},
|
||||
))
|
||||
|
||||
timer := time.NewTimer(3 * time.Second)
|
||||
err = goretry.Do(
|
||||
func() error {
|
||||
ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index)
|
||||
select {
|
||||
case err := <-errCh:
|
||||
require.NoError(t, err)
|
||||
case service := <-ch:
|
||||
require.Len(t, service, 1)
|
||||
require.Equal(t, serviceName, service[0].Service.Service)
|
||||
require.Equal(t, 9998, service[0].Service.Port)
|
||||
case <-timer.C:
|
||||
t.Fatalf("test timeout")
|
||||
index = service[0].Service.ModifyIndex
|
||||
if len(service) != 1 {
|
||||
return fmt.Errorf("service is %d, want 1", len(service))
|
||||
}
|
||||
if serviceName != service[0].Service.Service {
|
||||
return fmt.Errorf("service name is %s, want %s", service[0].Service.Service, serviceName)
|
||||
}
|
||||
if service[0].Service.Port != 9998 {
|
||||
return fmt.Errorf("service is %d, want 9998", service[0].Service.Port)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
},
|
||||
goretry.Attempts(5),
|
||||
goretry.Delay(time.Second),
|
||||
)
|
||||
require.NoError(t, err)
|
||||
|
||||
// upgrade the cluster to the Target version
|
||||
t.Logf("initiating standard upgrade to version=%q", tc.targetVersion)
|
||||
|
|
|
@ -1,271 +0,0 @@
|
|||
//go:build consulent
|
||||
// +build consulent
|
||||
|
||||
package upgrade
|
||||
|
||||
import (
|
||||
"context"
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io"
|
||||
"testing"
|
||||
|
||||
"github.com/hashicorp/consul/api"
|
||||
"github.com/hashicorp/consul/sdk/testutil"
|
||||
"github.com/hashicorp/consul/sdk/testutil/retry"
|
||||
"github.com/stretchr/testify/require"
|
||||
|
||||
libcluster "github.com/hashicorp/consul/test/integration/consul-container/libs/cluster"
|
||||
libservice "github.com/hashicorp/consul/test/integration/consul-container/libs/service"
|
||||
"github.com/hashicorp/consul/test/integration/consul-container/libs/utils"
|
||||
)
|
||||
|
||||
// Test partition crud using Current Clients and Latest GA Servers
|
||||
func TestLatestGAServersWithCurrentClients_PartitionCRUD(t *testing.T) {
|
||||
testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Partitions",
|
||||
func(t *testing.T, client *api.Client) {
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
|
||||
// CRUD partitions
|
||||
partition, _, err := client.Partitions().Read(ctx, "default", nil)
|
||||
require.NoError(t, err)
|
||||
fmt.Printf("%+v\n", partition)
|
||||
require.NotNil(t, partition)
|
||||
require.Equal(t, "default", partition.Name)
|
||||
|
||||
fooPartReq := api.Partition{Name: "foo-part"}
|
||||
fooPart, _, err := client.Partitions().Create(ctx, &api.Partition{Name: "foo-part"}, nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, fooPart)
|
||||
require.Equal(t, "foo-part", fooPart.Name)
|
||||
|
||||
partition, _, err = client.Partitions().Read(ctx, "foo-part", nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, partition)
|
||||
require.Equal(t, "foo-part", partition.Name)
|
||||
|
||||
fooPartReq.Description = "foo-part part"
|
||||
partition, _, err = client.Partitions().Update(ctx, &fooPartReq, nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, partition)
|
||||
require.Equal(t, "foo-part", partition.Name)
|
||||
require.Equal(t, "foo-part part", partition.Description)
|
||||
},
|
||||
func(t *testing.T, client *api.Client) {
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
|
||||
//Read partition again
|
||||
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
|
||||
partition, _, err := client.Partitions().Read(ctx, "default", nil)
|
||||
require.NoError(r, err)
|
||||
require.NotNil(r, partition)
|
||||
require.Equal(r, "default", partition.Name)
|
||||
})
|
||||
|
||||
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
|
||||
partition, _, err := client.Partitions().Read(ctx, "foo-part", nil)
|
||||
require.NoError(r, err)
|
||||
require.NotNil(r, partition)
|
||||
require.Equal(r, "foo-part", partition.Name)
|
||||
require.Equal(r, "foo-part part", partition.Description)
|
||||
})
|
||||
},
|
||||
)
|
||||
}
|
||||
|
||||
// Test namespace crud using Current Clients and Latest GA Servers
|
||||
func TestLatestGAServersWithCurrentClients_NamespaceCRUD(t *testing.T) {
|
||||
testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Namespaces",
|
||||
func(t *testing.T, client *api.Client) {
|
||||
// CRUD namespaces
|
||||
namespace, _, err := client.Namespaces().Read("default", nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, namespace, "default namespace does not exist yet")
|
||||
require.Equal(t, "default", namespace.Name)
|
||||
|
||||
fooNsReq := api.Namespace{Name: "foo-ns"}
|
||||
fooNs, _, err := client.Namespaces().Create(&api.Namespace{Name: "foo-ns"}, nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, fooNs)
|
||||
require.Equal(t, "foo-ns", fooNs.Name)
|
||||
|
||||
namespace, _, err = client.Namespaces().Read("foo-ns", nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, namespace)
|
||||
require.Equal(t, "foo-ns", namespace.Name)
|
||||
|
||||
fooNsReq.Description = "foo-ns ns"
|
||||
namespace, _, err = client.Namespaces().Update(&fooNsReq, nil)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, namespace)
|
||||
require.Equal(t, "foo-ns", namespace.Name)
|
||||
require.Equal(t, "foo-ns ns", namespace.Description)
|
||||
},
|
||||
func(t *testing.T, client *api.Client) {
|
||||
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
|
||||
namespace, _, err := client.Namespaces().Read("default", nil)
|
||||
require.NoError(r, err)
|
||||
require.NotNil(r, namespace)
|
||||
require.Equal(r, "default", namespace.Name)
|
||||
})
|
||||
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
|
||||
namespace, _, err := client.Namespaces().Read("foo-ns", nil)
|
||||
require.NoError(r, err)
|
||||
require.NotNil(r, namespace)
|
||||
require.Equal(r, "foo-ns", namespace.Name)
|
||||
require.Equal(r, "foo-ns ns", namespace.Description)
|
||||
})
|
||||
},
|
||||
)
|
||||
}
|
||||
|
||||
func testLatestGAServersWithCurrentClients_TenancyCRUD(
|
||||
t *testing.T,
|
||||
tenancyName string,
|
||||
createFn func(t *testing.T, client *api.Client),
|
||||
readFn func(t *testing.T, client *api.Client),
|
||||
) {
|
||||
const (
|
||||
numServers = 3
|
||||
numClients = 2
|
||||
)
|
||||
|
||||
// Create initial cluster
|
||||
cluster, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{
|
||||
NumServers: numServers,
|
||||
NumClients: numClients,
|
||||
BuildOpts: &libcluster.BuildOptions{
|
||||
Datacenter: "dc1",
|
||||
ConsulImageName: utils.LatestImageName,
|
||||
ConsulVersion: utils.LatestVersion,
|
||||
},
|
||||
ApplyDefaultProxySettings: true,
|
||||
})
|
||||
|
||||
client := cluster.APIClient(0)
|
||||
libcluster.WaitForLeader(t, cluster, client)
|
||||
libcluster.WaitForMembers(t, client, 5)
|
||||
|
||||
testutil.RunStep(t, "Create "+tenancyName, func(t *testing.T) {
|
||||
createFn(t, client)
|
||||
})
|
||||
|
||||
ctx := context.Background()
|
||||
|
||||
var snapshot io.ReadCloser
|
||||
testutil.RunStep(t, "Save snapshot", func(t *testing.T) {
|
||||
var err error
|
||||
snapshot, _, err = client.Snapshot().Save(nil)
|
||||
require.NoError(t, err)
|
||||
})
|
||||
|
||||
testutil.RunStep(t, "Check "+tenancyName+" after upgrade", func(t *testing.T) {
|
||||
// Upgrade nodes
|
||||
leader, err := cluster.Leader()
|
||||
require.NoError(t, err)
|
||||
|
||||
// upgrade things in the following order:
|
||||
//
|
||||
// 1. follower servers
|
||||
// 2. leader server
|
||||
// 3. clients
|
||||
var upgradeOrder []libcluster.Agent
|
||||
|
||||
followers, err := cluster.Followers()
|
||||
require.NoError(t, err)
|
||||
upgradeOrder = append(upgradeOrder, followers...)
|
||||
upgradeOrder = append(upgradeOrder, leader)
|
||||
upgradeOrder = append(upgradeOrder, cluster.Clients()...)
|
||||
|
||||
for _, n := range upgradeOrder {
|
||||
conf := n.GetConfig()
|
||||
|
||||
// TODO: ensure this makes sense again, it was doing an apples/orange version!=image comparison
|
||||
if conf.Version == utils.TargetVersion {
|
||||
return
|
||||
}
|
||||
|
||||
conf.Version = utils.TargetVersion
|
||||
|
||||
if n.IsServer() {
|
||||
// You only ever need bootstrap settings the FIRST time, so we do not need
|
||||
// them again.
|
||||
conf.ConfigBuilder.Unset("bootstrap")
|
||||
} else {
|
||||
// If we upgrade the clients fast enough
|
||||
// membership might not be gossiped to all of
|
||||
// the clients to persist into their serf
|
||||
// snapshot, so force them to rejoin the
|
||||
// normal way on restart.
|
||||
conf.ConfigBuilder.Set("retry_join", []string{"agent-0"})
|
||||
}
|
||||
|
||||
newJSON, err := json.MarshalIndent(conf.ConfigBuilder, "", " ")
|
||||
require.NoError(t, err)
|
||||
conf.JSON = string(newJSON)
|
||||
t.Logf("Upgraded cluster config for %q:\n%s", n.GetName(), conf.JSON)
|
||||
|
||||
selfBefore, err := n.GetClient().Agent().Self()
|
||||
require.NoError(t, err)
|
||||
|
||||
require.NoError(t, n.Upgrade(ctx, conf))
|
||||
|
||||
selfAfter, err := n.GetClient().Agent().Self()
|
||||
require.NoError(t, err)
|
||||
require.Truef(t,
|
||||
(selfBefore["Config"]["Version"] != selfAfter["Config"]["Version"]) || (selfBefore["Config"]["Revision"] != selfAfter["Config"]["Revision"]),
|
||||
fmt.Sprintf("upgraded version must be different (%s, %s), (%s, %s)", selfBefore["Config"]["Version"], selfBefore["Config"]["Revision"], selfAfter["Config"]["Version"], selfAfter["Config"]["Revision"]),
|
||||
)
|
||||
|
||||
client := n.GetClient()
|
||||
|
||||
libcluster.WaitForLeader(t, cluster, nil)
|
||||
libcluster.WaitForMembers(t, client, 5)
|
||||
}
|
||||
|
||||
//get the client again as it changed after upgrade.
|
||||
client := cluster.APIClient(0)
|
||||
libcluster.WaitForLeader(t, cluster, client)
|
||||
|
||||
// Read data again
|
||||
readFn(t, client)
|
||||
})
|
||||
|
||||
// Terminate the cluster for the snapshot test
|
||||
testutil.RunStep(t, "Terminate the cluster", func(t *testing.T) {
|
||||
require.NoError(t, cluster.Terminate())
|
||||
})
|
||||
|
||||
{ // Clear these so they super break if you tried to use them.
|
||||
cluster = nil
|
||||
client = nil
|
||||
}
|
||||
|
||||
// Create a fresh cluster from scratch
|
||||
cluster2, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{
|
||||
NumServers: numServers,
|
||||
NumClients: numClients,
|
||||
BuildOpts: &libcluster.BuildOptions{
|
||||
Datacenter: "dc1",
|
||||
ConsulImageName: utils.LatestImageName,
|
||||
ConsulVersion: utils.LatestVersion,
|
||||
},
|
||||
ApplyDefaultProxySettings: true,
|
||||
})
|
||||
client2 := cluster2.APIClient(0)
|
||||
|
||||
testutil.RunStep(t, "Restore saved snapshot", func(t *testing.T) {
|
||||
libcluster.WaitForLeader(t, cluster2, client2)
|
||||
libcluster.WaitForMembers(t, client2, 5)
|
||||
|
||||
// Restore the saved snapshot
|
||||
require.NoError(t, client2.Snapshot().Restore(nil, snapshot))
|
||||
|
||||
libcluster.WaitForLeader(t, cluster2, client2)
|
||||
|
||||
// make sure we still have the right data
|
||||
readFn(t, client2)
|
||||
})
|
||||
}
|
Loading…
Reference in New Issue