mirror of
https://github.com/status-im/consul.git
synced 2025-01-09 05:23:04 +00:00
5fb9df1640
* Adding explicit MPL license for sub-package This directory and its subdirectories (packages) contain files licensed with the MPLv2 `LICENSE` file in this directory and are intentionally licensed separately from the BSL `LICENSE` file at the root of this repository. * Adding explicit MPL license for sub-package This directory and its subdirectories (packages) contain files licensed with the MPLv2 `LICENSE` file in this directory and are intentionally licensed separately from the BSL `LICENSE` file at the root of this repository. * Updating the license from MPL to Business Source License Going forward, this project will be licensed under the Business Source License v1.1. Please see our blog post for more details at <Blog URL>, FAQ at www.hashicorp.com/licensing-faq, and details of the license at www.hashicorp.com/bsl. * add missing license headers * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 * Update copyright file headers to BUSL-1.1 --------- Co-authored-by: hashicorp-copywrite[bot] <110428419+hashicorp-copywrite[bot]@users.noreply.github.com>
312 lines
8.3 KiB
Go
312 lines
8.3 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: BUSL-1.1
|
|
|
|
package peerstream
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"math/rand"
|
|
"sort"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/hashicorp/go-hclog"
|
|
"github.com/stretchr/testify/require"
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"github.com/hashicorp/consul/agent/cache"
|
|
"github.com/hashicorp/consul/agent/consul/state"
|
|
"github.com/hashicorp/consul/agent/consul/stream"
|
|
"github.com/hashicorp/consul/agent/structs"
|
|
"github.com/hashicorp/consul/agent/submatview"
|
|
"github.com/hashicorp/consul/proto/private/pbservice"
|
|
"github.com/hashicorp/consul/proto/private/pbsubscribe"
|
|
)
|
|
|
|
// testInitialIndex is the first index that will be used in simulated updates.
|
|
//
|
|
// This is set to something arbitrarily high so that we can ignore the initial
|
|
// snapshot which may or may not be empty depending on timing.
|
|
const testInitialIndex uint64 = 9000
|
|
|
|
// TestExportedServiceSubscription tests the exported services view and the backing submatview.LocalMaterializer.
|
|
func TestExportedServiceSubscription(t *testing.T) {
|
|
s := &stateMap{
|
|
states: make(map[string]*serviceState),
|
|
}
|
|
|
|
sh := snapshotHandler{stateMap: s}
|
|
pub := stream.NewEventPublisher(10 * time.Millisecond)
|
|
pub.RegisterHandler(pbsubscribe.Topic_ServiceHealth, sh.Snapshot, false)
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
go pub.Run(ctx)
|
|
|
|
apiSN := structs.NewServiceName("api", nil)
|
|
webSN := structs.NewServiceName("web", nil)
|
|
|
|
newRegisterHealthEvent := func(id, service string) stream.Event {
|
|
return stream.Event{
|
|
Topic: pbsubscribe.Topic_ServiceHealth,
|
|
Payload: state.EventPayloadCheckServiceNode{
|
|
Op: pbsubscribe.CatalogOp_Register,
|
|
Value: &structs.CheckServiceNode{
|
|
Node: &structs.Node{},
|
|
Service: &structs.NodeService{
|
|
ID: id,
|
|
Service: service,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
// List of updates to the state store:
|
|
// - api: {register api-1, register api-2, register api-3}
|
|
// - web: {register web-1, deregister web-1, register web-2}1
|
|
events := []map[string]stream.Event{
|
|
{
|
|
apiSN.String(): newRegisterHealthEvent("api-1", "api"),
|
|
webSN.String(): newRegisterHealthEvent("web-1", "web"),
|
|
},
|
|
{
|
|
apiSN.String(): newRegisterHealthEvent("api-2", "api"),
|
|
webSN.String(): newRegisterHealthEvent("web-1", "web"),
|
|
},
|
|
{
|
|
apiSN.String(): newRegisterHealthEvent("api-3", "api"),
|
|
webSN.String(): newRegisterHealthEvent("web-2", "web"),
|
|
},
|
|
}
|
|
|
|
// store represents Consul's memdb state store.
|
|
// A stream of event updates
|
|
store := store{stateMap: s, pub: pub}
|
|
|
|
// This errgroup is used to issue simulate async updates to the state store,
|
|
// and also consume that fixed number of updates.
|
|
group, gctx := errgroup.WithContext(ctx)
|
|
group.Go(func() error {
|
|
store.simulateUpdates(gctx, events)
|
|
return nil
|
|
})
|
|
|
|
// viewStore is the store shared by the two service consumer's materializers.
|
|
// It is intentionally not run in the errgroup because it will block until the context is canceled.
|
|
viewStore := submatview.NewStore(hclog.New(nil))
|
|
go viewStore.Run(ctx)
|
|
|
|
// Each consumer represents a subscriber to exported service updates, and will consume
|
|
// stream events for the service name it is interested in.
|
|
consumers := make(map[string]*consumer)
|
|
for _, svc := range []structs.ServiceName{apiSN, webSN} {
|
|
c := &consumer{
|
|
viewStore: viewStore,
|
|
publisher: pub,
|
|
seenByIndex: make(map[uint64][]string),
|
|
}
|
|
service := svc
|
|
group.Go(func() error {
|
|
return c.consume(gctx, service.Name, len(events))
|
|
})
|
|
consumers[service.String()] = c
|
|
}
|
|
|
|
// Wait until all the events have been simulated and consumed.
|
|
done := make(chan struct{})
|
|
go func() {
|
|
defer close(done)
|
|
_ = group.Wait()
|
|
}()
|
|
|
|
select {
|
|
case <-done:
|
|
// finished
|
|
case <-time.After(500 * time.Millisecond):
|
|
// timed out, the Wait context will be cancelled by
|
|
t.Fatalf("timed out waiting for producers and consumers")
|
|
}
|
|
|
|
for svc, c := range consumers {
|
|
require.NotEmpty(t, c.seenByIndex)
|
|
|
|
// Note that store.states[svc].idsByIndex does not assert against a slice of expectations because
|
|
// the index that the different events will arrive in the simulation is not deterministic.
|
|
require.Equal(t, store.states[svc].idsByIndex, c.seenByIndex)
|
|
}
|
|
}
|
|
|
|
// stateMap is a map keyed by service to the state of the store at different indexes
|
|
type stateMap struct {
|
|
mu sync.Mutex
|
|
states map[string]*serviceState
|
|
}
|
|
|
|
type store struct {
|
|
*stateMap
|
|
|
|
pub *stream.EventPublisher
|
|
}
|
|
|
|
// simulateUpdates will publish events and also store the state at each index for later assertions.
|
|
func (s *store) simulateUpdates(ctx context.Context, events []map[string]stream.Event) {
|
|
idx := testInitialIndex
|
|
|
|
for _, m := range events {
|
|
if ctx.Err() != nil {
|
|
return
|
|
}
|
|
|
|
for svc, event := range m {
|
|
idx++
|
|
event.Index = idx
|
|
s.pub.Publish([]stream.Event{event})
|
|
|
|
s.stateMap.mu.Lock()
|
|
svcState, ok := s.states[svc]
|
|
if !ok {
|
|
svcState = &serviceState{
|
|
current: make(map[string]*structs.CheckServiceNode),
|
|
idsByIndex: make(map[uint64][]string),
|
|
}
|
|
s.states[svc] = svcState
|
|
}
|
|
s.stateMap.mu.Unlock()
|
|
|
|
svcState.mu.Lock()
|
|
svcState.idx = idx
|
|
|
|
// Updating the svcState.current map allows us to capture snapshots from a stream of add/delete events.
|
|
payload := event.Payload.(state.EventPayloadCheckServiceNode)
|
|
switch payload.Op {
|
|
case pbsubscribe.CatalogOp_Register:
|
|
svcState.current[payload.Value.Service.ID] = payload.Value
|
|
case pbsubscribe.CatalogOp_Deregister:
|
|
delete(svcState.current, payload.Value.Service.ID)
|
|
default:
|
|
panic(fmt.Sprintf("unable to handle op type %v", payload.Op))
|
|
}
|
|
|
|
svcState.idsByIndex[idx] = serviceIDsFromMap(svcState.current)
|
|
svcState.mu.Unlock()
|
|
|
|
delay := time.Duration(rand.Intn(25)) * time.Millisecond
|
|
time.Sleep(5*time.Millisecond + delay)
|
|
}
|
|
}
|
|
}
|
|
|
|
func serviceIDsFromMap(m map[string]*structs.CheckServiceNode) []string {
|
|
var result []string
|
|
for id := range m {
|
|
result = append(result, id)
|
|
}
|
|
sort.Strings(result)
|
|
return result
|
|
}
|
|
|
|
type snapshotHandler struct {
|
|
*stateMap
|
|
}
|
|
|
|
type serviceState struct {
|
|
mu sync.Mutex
|
|
idx uint64
|
|
|
|
// The current snapshot of data, given the observed events.
|
|
current map[string]*structs.CheckServiceNode
|
|
|
|
// The list of service IDs seen at each index that an update was received for the given service name.
|
|
idsByIndex map[uint64][]string
|
|
}
|
|
|
|
// Snapshot dumps the currently registered service instances.
|
|
//
|
|
// Snapshot implements stream.SnapshotFunc.
|
|
func (s *snapshotHandler) Snapshot(req stream.SubscribeRequest, buf stream.SnapshotAppender) (index uint64, err error) {
|
|
s.stateMap.mu.Lock()
|
|
svcState, ok := s.states[req.Subject.String()]
|
|
if !ok {
|
|
svcState = &serviceState{
|
|
current: make(map[string]*structs.CheckServiceNode),
|
|
idsByIndex: make(map[uint64][]string),
|
|
}
|
|
s.states[req.Subject.String()] = svcState
|
|
}
|
|
s.stateMap.mu.Unlock()
|
|
|
|
svcState.mu.Lock()
|
|
defer svcState.mu.Unlock()
|
|
|
|
for _, node := range svcState.current {
|
|
event := stream.Event{
|
|
Topic: pbsubscribe.Topic_ServiceHealth,
|
|
Index: svcState.idx,
|
|
Payload: state.EventPayloadCheckServiceNode{
|
|
Op: pbsubscribe.CatalogOp_Register,
|
|
Value: node,
|
|
},
|
|
}
|
|
buf.Append([]stream.Event{event})
|
|
}
|
|
return svcState.idx, nil
|
|
}
|
|
|
|
type consumer struct {
|
|
viewStore *submatview.Store
|
|
publisher *stream.EventPublisher
|
|
seenByIndex map[uint64][]string
|
|
}
|
|
|
|
func (c *consumer) consume(ctx context.Context, service string, countExpected int) error {
|
|
group, gctx := errgroup.WithContext(ctx)
|
|
updateCh := make(chan cache.UpdateEvent, 10)
|
|
|
|
group.Go(func() error {
|
|
sr := newExportedStandardServiceRequest(
|
|
hclog.New(nil),
|
|
structs.NewServiceName(service, nil),
|
|
c.publisher,
|
|
)
|
|
return c.viewStore.Notify(gctx, sr, "", updateCh)
|
|
})
|
|
group.Go(func() error {
|
|
var n int
|
|
for {
|
|
if n >= countExpected {
|
|
return nil
|
|
}
|
|
select {
|
|
case u := <-updateCh:
|
|
idx := u.Meta.Index
|
|
|
|
// This is the initial/empty state. Skip over it and wait for the first
|
|
// real event.
|
|
if idx < testInitialIndex {
|
|
continue
|
|
}
|
|
|
|
// Each update contains the current snapshot of registered services.
|
|
c.seenByIndex[idx] = serviceIDsFromUpdates(u)
|
|
n++
|
|
|
|
case <-gctx.Done():
|
|
return nil
|
|
}
|
|
}
|
|
})
|
|
return group.Wait()
|
|
}
|
|
|
|
func serviceIDsFromUpdates(u cache.UpdateEvent) []string {
|
|
var result []string
|
|
for _, node := range u.Result.(*pbservice.IndexedCheckServiceNodes).Nodes {
|
|
result = append(result, node.Service.ID)
|
|
}
|
|
sort.Strings(result)
|
|
return result
|
|
}
|