mirror of https://github.com/status-im/consul.git
counter part of ent pr (#17618)
This commit is contained in:
parent
1864874726
commit
b782f2e39a
|
@ -194,6 +194,7 @@ dev-docker: linux dev-build
|
|||
--label version=$(CONSUL_VERSION) \
|
||||
--load \
|
||||
-f $(CURDIR)/build-support/docker/Consul-Dev-Multiarch.dockerfile $(CURDIR)/pkg/bin/
|
||||
docker tag 'consul:local' '$(CONSUL_COMPAT_TEST_IMAGE):local'
|
||||
|
||||
check-remote-dev-image-env:
|
||||
ifndef REMOTE_DEV_IMAGE
|
||||
|
|
|
@ -32,8 +32,6 @@ const (
|
|||
// - logs for exceeding
|
||||
|
||||
func TestServerRequestRateLimit(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
type action struct {
|
||||
function func(client *api.Client) error
|
||||
rateLimitOperation string
|
||||
|
@ -52,6 +50,7 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
mode string
|
||||
}
|
||||
|
||||
// getKV and putKV are net/RPC calls
|
||||
getKV := action{
|
||||
function: func(client *api.Client) error {
|
||||
_, _, err := client.KV().Get("foo", &api.QueryOptions{})
|
||||
|
@ -99,13 +98,13 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
action: putKV,
|
||||
expectedErrorMsg: "",
|
||||
expectExceededLog: true,
|
||||
expectMetric: false,
|
||||
expectMetric: true,
|
||||
},
|
||||
{
|
||||
action: getKV,
|
||||
expectedErrorMsg: "",
|
||||
expectExceededLog: true,
|
||||
expectMetric: false,
|
||||
expectMetric: true,
|
||||
},
|
||||
},
|
||||
},
|
||||
|
@ -127,10 +126,13 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
expectMetric: true,
|
||||
},
|
||||
},
|
||||
}}
|
||||
},
|
||||
}
|
||||
|
||||
for _, tc := range testCases {
|
||||
tc := tc
|
||||
t.Run(tc.description, func(t *testing.T) {
|
||||
t.Parallel()
|
||||
clusterConfig := &libtopology.ClusterConfig{
|
||||
NumServers: 1,
|
||||
NumClients: 0,
|
||||
|
@ -144,12 +146,9 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
ApplyDefaultProxySettings: false,
|
||||
}
|
||||
|
||||
cluster, _, _ := libtopology.NewCluster(t, clusterConfig)
|
||||
cluster, client := setupClusterAndClient(t, clusterConfig, true)
|
||||
defer terminate(t, cluster)
|
||||
|
||||
client, err := cluster.GetClient(nil, true)
|
||||
require.NoError(t, err)
|
||||
|
||||
// perform actions and validate returned errors to client
|
||||
for _, op := range tc.operations {
|
||||
err := op.action.function(client)
|
||||
|
@ -165,22 +164,14 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
// doing this in a separate loop so we can perform actions, allow metrics
|
||||
// and logs to collect and then assert on each.
|
||||
for _, op := range tc.operations {
|
||||
timer := &retry.Timer{Timeout: 10 * time.Second, Wait: 500 * time.Millisecond}
|
||||
timer := &retry.Timer{Timeout: 15 * time.Second, Wait: 500 * time.Millisecond}
|
||||
retry.RunWith(timer, t, func(r *retry.R) {
|
||||
// validate metrics
|
||||
metricsInfo, err := client.Agent().Metrics()
|
||||
// TODO(NET-1978): currently returns NaN error
|
||||
// require.NoError(t, err)
|
||||
if metricsInfo != nil && err == nil {
|
||||
if op.expectMetric {
|
||||
checkForMetric(r, metricsInfo, op.action.rateLimitOperation, op.action.rateLimitType, tc.mode)
|
||||
}
|
||||
}
|
||||
checkForMetric(t, cluster, op.action.rateLimitOperation, op.action.rateLimitType, tc.mode, op.expectMetric)
|
||||
|
||||
// validate logs
|
||||
// putting this last as there are cases where logs
|
||||
// were not present in consumer when assertion was made.
|
||||
checkLogsForMessage(r, clusterConfig.LogConsumer.Msgs,
|
||||
checkLogsForMessage(t, clusterConfig.LogConsumer.Msgs,
|
||||
fmt.Sprintf("[DEBUG] agent.server.rpc-rate-limit: RPC exceeded allowed rate limit: rpc=%s", op.action.rateLimitOperation),
|
||||
op.action.rateLimitOperation, "exceeded", op.expectExceededLog)
|
||||
|
||||
|
@ -190,7 +181,24 @@ func TestServerRequestRateLimit(t *testing.T) {
|
|||
}
|
||||
}
|
||||
|
||||
func checkForMetric(t *retry.R, metricsInfo *api.MetricsInfo, operationName string, expectedLimitType string, expectedMode string) {
|
||||
func setupClusterAndClient(t *testing.T, config *libtopology.ClusterConfig, isServer bool) (*libcluster.Cluster, *api.Client) {
|
||||
cluster, _, _ := libtopology.NewCluster(t, config)
|
||||
|
||||
client, err := cluster.GetClient(nil, isServer)
|
||||
require.NoError(t, err)
|
||||
|
||||
return cluster, client
|
||||
}
|
||||
|
||||
func checkForMetric(t *testing.T, cluster *libcluster.Cluster, operationName string, expectedLimitType string, expectedMode string, expectMetric bool) {
|
||||
// validate metrics
|
||||
server, err := cluster.GetClient(nil, true)
|
||||
require.NoError(t, err)
|
||||
metricsInfo, err := server.Agent().Metrics()
|
||||
// TODO(NET-1978): currently returns NaN error
|
||||
// require.NoError(t, err)
|
||||
if metricsInfo != nil && err == nil {
|
||||
if expectMetric {
|
||||
const counterName = "consul.rpc.rate_limit.exceeded"
|
||||
|
||||
var counter api.SampledValue
|
||||
|
@ -217,8 +225,11 @@ func checkForMetric(t *retry.R, metricsInfo *api.MetricsInfo, operationName stri
|
|||
require.Equal(t, expectedMode, mode)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func checkLogsForMessage(t *retry.R, logs []string, msg string, operationName string, logType string, logShouldExist bool) {
|
||||
func checkLogsForMessage(t *testing.T, logs []string, msg string, operationName string, logType string, logShouldExist bool) {
|
||||
if logShouldExist {
|
||||
found := false
|
||||
for _, log := range logs {
|
||||
if strings.Contains(log, msg) {
|
||||
|
@ -226,7 +237,9 @@ func checkLogsForMessage(t *retry.R, logs []string, msg string, operationName st
|
|||
break
|
||||
}
|
||||
}
|
||||
require.Equal(t, logShouldExist, found, fmt.Sprintf("%s log check failed for: %s. Log expected: %t", logType, operationName, logShouldExist))
|
||||
expectedLog := fmt.Sprintf("%s log check failed for: %s. Log expected: %t", logType, operationName, logShouldExist)
|
||||
require.Equal(t, logShouldExist, found, expectedLog)
|
||||
}
|
||||
}
|
||||
|
||||
func terminate(t *testing.T, cluster *libcluster.Cluster) {
|
||||
|
|
Loading…
Reference in New Issue