upgrade test: use retry with ModifyIndex and remove ent test file (#16553)

This commit is contained in:
cskh 2023-03-07 13:27:47 -05:00 committed by GitHub
parent a5b8256111
commit b649a5e8e4
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 25 additions and 284 deletions

View File

@ -6,6 +6,7 @@ import (
"testing" "testing"
"time" "time"
goretry "github.com/avast/retry-go"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
"github.com/hashicorp/consul/api" "github.com/hashicorp/consul/api"
@ -75,22 +76,33 @@ func TestStandardUpgradeToTarget_fromLatest(t *testing.T) {
const serviceName = "api" const serviceName = "api"
index := libservice.ServiceCreate(t, client, serviceName) index := libservice.ServiceCreate(t, client, serviceName)
ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index)
require.NoError(t, client.Agent().ServiceRegister( require.NoError(t, client.Agent().ServiceRegister(
&api.AgentServiceRegistration{Name: serviceName, Port: 9998}, &api.AgentServiceRegistration{Name: serviceName, Port: 9998},
)) ))
err = goretry.Do(
timer := time.NewTimer(3 * time.Second) func() error {
ch, errCh := libservice.ServiceHealthBlockingQuery(client, serviceName, index)
select { select {
case err := <-errCh: case err := <-errCh:
require.NoError(t, err) require.NoError(t, err)
case service := <-ch: case service := <-ch:
require.Len(t, service, 1) index = service[0].Service.ModifyIndex
require.Equal(t, serviceName, service[0].Service.Service) if len(service) != 1 {
require.Equal(t, 9998, service[0].Service.Port) return fmt.Errorf("service is %d, want 1", len(service))
case <-timer.C:
t.Fatalf("test timeout")
} }
if serviceName != service[0].Service.Service {
return fmt.Errorf("service name is %s, want %s", service[0].Service.Service, serviceName)
}
if service[0].Service.Port != 9998 {
return fmt.Errorf("service is %d, want 9998", service[0].Service.Port)
}
}
return nil
},
goretry.Attempts(5),
goretry.Delay(time.Second),
)
require.NoError(t, err)
// upgrade the cluster to the Target version // upgrade the cluster to the Target version
t.Logf("initiating standard upgrade to version=%q", tc.targetVersion) t.Logf("initiating standard upgrade to version=%q", tc.targetVersion)

View File

@ -1,271 +0,0 @@
//go:build consulent
// +build consulent
package upgrade
import (
"context"
"encoding/json"
"fmt"
"io"
"testing"
"github.com/hashicorp/consul/api"
"github.com/hashicorp/consul/sdk/testutil"
"github.com/hashicorp/consul/sdk/testutil/retry"
"github.com/stretchr/testify/require"
libcluster "github.com/hashicorp/consul/test/integration/consul-container/libs/cluster"
libservice "github.com/hashicorp/consul/test/integration/consul-container/libs/service"
"github.com/hashicorp/consul/test/integration/consul-container/libs/utils"
)
// Test partition crud using Current Clients and Latest GA Servers
func TestLatestGAServersWithCurrentClients_PartitionCRUD(t *testing.T) {
testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Partitions",
func(t *testing.T, client *api.Client) {
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
// CRUD partitions
partition, _, err := client.Partitions().Read(ctx, "default", nil)
require.NoError(t, err)
fmt.Printf("%+v\n", partition)
require.NotNil(t, partition)
require.Equal(t, "default", partition.Name)
fooPartReq := api.Partition{Name: "foo-part"}
fooPart, _, err := client.Partitions().Create(ctx, &api.Partition{Name: "foo-part"}, nil)
require.NoError(t, err)
require.NotNil(t, fooPart)
require.Equal(t, "foo-part", fooPart.Name)
partition, _, err = client.Partitions().Read(ctx, "foo-part", nil)
require.NoError(t, err)
require.NotNil(t, partition)
require.Equal(t, "foo-part", partition.Name)
fooPartReq.Description = "foo-part part"
partition, _, err = client.Partitions().Update(ctx, &fooPartReq, nil)
require.NoError(t, err)
require.NotNil(t, partition)
require.Equal(t, "foo-part", partition.Name)
require.Equal(t, "foo-part part", partition.Description)
},
func(t *testing.T, client *api.Client) {
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
//Read partition again
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
partition, _, err := client.Partitions().Read(ctx, "default", nil)
require.NoError(r, err)
require.NotNil(r, partition)
require.Equal(r, "default", partition.Name)
})
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
partition, _, err := client.Partitions().Read(ctx, "foo-part", nil)
require.NoError(r, err)
require.NotNil(r, partition)
require.Equal(r, "foo-part", partition.Name)
require.Equal(r, "foo-part part", partition.Description)
})
},
)
}
// Test namespace crud using Current Clients and Latest GA Servers
func TestLatestGAServersWithCurrentClients_NamespaceCRUD(t *testing.T) {
testLatestGAServersWithCurrentClients_TenancyCRUD(t, "Namespaces",
func(t *testing.T, client *api.Client) {
// CRUD namespaces
namespace, _, err := client.Namespaces().Read("default", nil)
require.NoError(t, err)
require.NotNil(t, namespace, "default namespace does not exist yet")
require.Equal(t, "default", namespace.Name)
fooNsReq := api.Namespace{Name: "foo-ns"}
fooNs, _, err := client.Namespaces().Create(&api.Namespace{Name: "foo-ns"}, nil)
require.NoError(t, err)
require.NotNil(t, fooNs)
require.Equal(t, "foo-ns", fooNs.Name)
namespace, _, err = client.Namespaces().Read("foo-ns", nil)
require.NoError(t, err)
require.NotNil(t, namespace)
require.Equal(t, "foo-ns", namespace.Name)
fooNsReq.Description = "foo-ns ns"
namespace, _, err = client.Namespaces().Update(&fooNsReq, nil)
require.NoError(t, err)
require.NotNil(t, namespace)
require.Equal(t, "foo-ns", namespace.Name)
require.Equal(t, "foo-ns ns", namespace.Description)
},
func(t *testing.T, client *api.Client) {
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
namespace, _, err := client.Namespaces().Read("default", nil)
require.NoError(r, err)
require.NotNil(r, namespace)
require.Equal(r, "default", namespace.Name)
})
retry.RunWith(libcluster.LongFailer(), t, func(r *retry.R) {
namespace, _, err := client.Namespaces().Read("foo-ns", nil)
require.NoError(r, err)
require.NotNil(r, namespace)
require.Equal(r, "foo-ns", namespace.Name)
require.Equal(r, "foo-ns ns", namespace.Description)
})
},
)
}
func testLatestGAServersWithCurrentClients_TenancyCRUD(
t *testing.T,
tenancyName string,
createFn func(t *testing.T, client *api.Client),
readFn func(t *testing.T, client *api.Client),
) {
const (
numServers = 3
numClients = 2
)
// Create initial cluster
cluster, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{
NumServers: numServers,
NumClients: numClients,
BuildOpts: &libcluster.BuildOptions{
Datacenter: "dc1",
ConsulImageName: utils.LatestImageName,
ConsulVersion: utils.LatestVersion,
},
ApplyDefaultProxySettings: true,
})
client := cluster.APIClient(0)
libcluster.WaitForLeader(t, cluster, client)
libcluster.WaitForMembers(t, client, 5)
testutil.RunStep(t, "Create "+tenancyName, func(t *testing.T) {
createFn(t, client)
})
ctx := context.Background()
var snapshot io.ReadCloser
testutil.RunStep(t, "Save snapshot", func(t *testing.T) {
var err error
snapshot, _, err = client.Snapshot().Save(nil)
require.NoError(t, err)
})
testutil.RunStep(t, "Check "+tenancyName+" after upgrade", func(t *testing.T) {
// Upgrade nodes
leader, err := cluster.Leader()
require.NoError(t, err)
// upgrade things in the following order:
//
// 1. follower servers
// 2. leader server
// 3. clients
var upgradeOrder []libcluster.Agent
followers, err := cluster.Followers()
require.NoError(t, err)
upgradeOrder = append(upgradeOrder, followers...)
upgradeOrder = append(upgradeOrder, leader)
upgradeOrder = append(upgradeOrder, cluster.Clients()...)
for _, n := range upgradeOrder {
conf := n.GetConfig()
// TODO: ensure this makes sense again, it was doing an apples/orange version!=image comparison
if conf.Version == utils.TargetVersion {
return
}
conf.Version = utils.TargetVersion
if n.IsServer() {
// You only ever need bootstrap settings the FIRST time, so we do not need
// them again.
conf.ConfigBuilder.Unset("bootstrap")
} else {
// If we upgrade the clients fast enough
// membership might not be gossiped to all of
// the clients to persist into their serf
// snapshot, so force them to rejoin the
// normal way on restart.
conf.ConfigBuilder.Set("retry_join", []string{"agent-0"})
}
newJSON, err := json.MarshalIndent(conf.ConfigBuilder, "", " ")
require.NoError(t, err)
conf.JSON = string(newJSON)
t.Logf("Upgraded cluster config for %q:\n%s", n.GetName(), conf.JSON)
selfBefore, err := n.GetClient().Agent().Self()
require.NoError(t, err)
require.NoError(t, n.Upgrade(ctx, conf))
selfAfter, err := n.GetClient().Agent().Self()
require.NoError(t, err)
require.Truef(t,
(selfBefore["Config"]["Version"] != selfAfter["Config"]["Version"]) || (selfBefore["Config"]["Revision"] != selfAfter["Config"]["Revision"]),
fmt.Sprintf("upgraded version must be different (%s, %s), (%s, %s)", selfBefore["Config"]["Version"], selfBefore["Config"]["Revision"], selfAfter["Config"]["Version"], selfAfter["Config"]["Revision"]),
)
client := n.GetClient()
libcluster.WaitForLeader(t, cluster, nil)
libcluster.WaitForMembers(t, client, 5)
}
//get the client again as it changed after upgrade.
client := cluster.APIClient(0)
libcluster.WaitForLeader(t, cluster, client)
// Read data again
readFn(t, client)
})
// Terminate the cluster for the snapshot test
testutil.RunStep(t, "Terminate the cluster", func(t *testing.T) {
require.NoError(t, cluster.Terminate())
})
{ // Clear these so they super break if you tried to use them.
cluster = nil
client = nil
}
// Create a fresh cluster from scratch
cluster2, _, _ := libtopology.NewCluster(t, &libtopology.ClusterConfig{
NumServers: numServers,
NumClients: numClients,
BuildOpts: &libcluster.BuildOptions{
Datacenter: "dc1",
ConsulImageName: utils.LatestImageName,
ConsulVersion: utils.LatestVersion,
},
ApplyDefaultProxySettings: true,
})
client2 := cluster2.APIClient(0)
testutil.RunStep(t, "Restore saved snapshot", func(t *testing.T) {
libcluster.WaitForLeader(t, cluster2, client2)
libcluster.WaitForMembers(t, client2, 5)
// Restore the saved snapshot
require.NoError(t, client2.Snapshot().Restore(nil, snapshot))
libcluster.WaitForLeader(t, cluster2, client2)
// make sure we still have the right data
readFn(t, client2)
})
}