Merge branch 'master' into release/1-6

This commit is contained in:
Alvin Huang 2019-07-17 15:43:30 -04:00
commit 6f1953d96d
8 changed files with 166 additions and 111 deletions

View File

@ -587,9 +587,8 @@ workflows:
- dev-build - dev-build
filters: filters:
branches: branches:
only: ignore:
- /^bug\/flaky-test-.*$/ # only run go tests on bug/flaky-test-* for now since we are fixing tests - /^pull\/.*$/ # only run go tests on non forks
- release/1-6
- go-test-api: *go-test - go-test-api: *go-test
- dev-upload-s3: - dev-upload-s3:
requires: requires:

View File

@ -35,6 +35,7 @@ IMPROVEMENTS:
BUG FIXES: BUG FIXES:
* autopilot: update to also remove failed nodes from WAN gossip pool [[GH-6028](https://github.com/hashicorp/consul/pull/6028)] * autopilot: update to also remove failed nodes from WAN gossip pool [[GH-6028](https://github.com/hashicorp/consul/pull/6028)]
* agent: avoid reverting any check updates that occur while a service is being added or the config is reloaded [[GH-6144](https://github.com/hashicorp/consul/issues/6144)]
## 1.5.2 (June 27, 2019) ## 1.5.2 (June 27, 2019)

View File

@ -457,7 +457,7 @@ func (a *Agent) Start() error {
if err := a.loadProxies(c); err != nil { if err := a.loadProxies(c); err != nil {
return err return err
} }
if err := a.loadChecks(c); err != nil { if err := a.loadChecks(c, nil); err != nil {
return err return err
} }
if err := a.loadMetadata(c); err != nil { if err := a.loadMetadata(c); err != nil {
@ -2109,10 +2109,10 @@ func (a *Agent) addServiceInternal(service *structs.NodeService, chkTypes []*str
a.PauseSync() a.PauseSync()
defer a.ResumeSync() defer a.ResumeSync()
// Take a snapshot of the current state of checks (if any), and // Take a snapshot of the current state of checks (if any), and when adding
// restore them before resuming anti-entropy. // a check that already existed carry over the state before resuming
// anti-entropy.
snap := a.snapshotCheckState() snap := a.snapshotCheckState()
defer a.restoreCheckState(snap)
var checks []*structs.HealthCheck var checks []*structs.HealthCheck
@ -2143,6 +2143,13 @@ func (a *Agent) addServiceInternal(service *structs.NodeService, chkTypes []*str
check.Status = chkType.Status check.Status = chkType.Status
} }
// Restore the fields from the snapshot.
prev, ok := snap[check.CheckID]
if ok {
check.Output = prev.Output
check.Status = prev.Status
}
checks = append(checks, check) checks = append(checks, check)
} }
@ -3346,10 +3353,17 @@ func (a *Agent) unloadServices() error {
// loadChecks loads check definitions and/or persisted check definitions from // loadChecks loads check definitions and/or persisted check definitions from
// disk and re-registers them with the local agent. // disk and re-registers them with the local agent.
func (a *Agent) loadChecks(conf *config.RuntimeConfig) error { func (a *Agent) loadChecks(conf *config.RuntimeConfig, snap map[types.CheckID]*structs.HealthCheck) error {
// Register the checks from config // Register the checks from config
for _, check := range conf.Checks { for _, check := range conf.Checks {
health := check.HealthCheck(conf.NodeName) health := check.HealthCheck(conf.NodeName)
// Restore the fields from the snapshot.
if prev, ok := snap[health.CheckID]; ok {
health.Output = prev.Output
health.Status = prev.Status
}
chkType := check.CheckType() chkType := check.CheckType()
if err := a.addCheckLocked(health, chkType, false, check.Token, ConfigSourceLocal); err != nil { if err := a.addCheckLocked(health, chkType, false, check.Token, ConfigSourceLocal); err != nil {
return fmt.Errorf("Failed to register check '%s': %v %v", check.Name, err, check) return fmt.Errorf("Failed to register check '%s': %v %v", check.Name, err, check)
@ -3406,6 +3420,12 @@ func (a *Agent) loadChecks(conf *config.RuntimeConfig) error {
// services into the active pool // services into the active pool
p.Check.Status = api.HealthCritical p.Check.Status = api.HealthCritical
// Restore the fields from the snapshot.
if prev, ok := snap[p.Check.CheckID]; ok {
p.Check.Output = prev.Output
p.Check.Status = prev.Status
}
if err := a.addCheckLocked(p.Check, p.ChkType, false, p.Token, ConfigSourceLocal); err != nil { if err := a.addCheckLocked(p.Check, p.ChkType, false, p.Token, ConfigSourceLocal); err != nil {
// Purge the check if it is unable to be restored. // Purge the check if it is unable to be restored.
a.logger.Printf("[WARN] agent: Failed to restore check %q: %s", a.logger.Printf("[WARN] agent: Failed to restore check %q: %s",
@ -3634,15 +3654,6 @@ func (a *Agent) snapshotCheckState() map[types.CheckID]*structs.HealthCheck {
return a.State.Checks() return a.State.Checks()
} }
// restoreCheckState is used to reset the health state based on a snapshot.
// This is done after we finish the reload to avoid any unnecessary flaps
// in health state and potential session invalidations.
func (a *Agent) restoreCheckState(snap map[types.CheckID]*structs.HealthCheck) {
for id, check := range snap {
a.State.UpdateCheck(id, check.Status, check.Output)
}
}
// loadMetadata loads node metadata fields from the agent config and // loadMetadata loads node metadata fields from the agent config and
// updates them on the local agent. // updates them on the local agent.
func (a *Agent) loadMetadata(conf *config.RuntimeConfig) error { func (a *Agent) loadMetadata(conf *config.RuntimeConfig) error {
@ -3765,9 +3776,9 @@ func (a *Agent) ReloadConfig(newCfg *config.RuntimeConfig) error {
a.stateLock.Lock() a.stateLock.Lock()
defer a.stateLock.Unlock() defer a.stateLock.Unlock()
// Snapshot the current state, and restore it afterwards // Snapshot the current state, and use that to initialize the checks when
// they are recreated.
snap := a.snapshotCheckState() snap := a.snapshotCheckState()
defer a.restoreCheckState(snap)
// First unload all checks, services, and metadata. This lets us begin the reload // First unload all checks, services, and metadata. This lets us begin the reload
// with a clean slate. // with a clean slate.
@ -3798,7 +3809,7 @@ func (a *Agent) ReloadConfig(newCfg *config.RuntimeConfig) error {
if err := a.loadProxies(newCfg); err != nil { if err := a.loadProxies(newCfg); err != nil {
return fmt.Errorf("Failed reloading proxies: %s", err) return fmt.Errorf("Failed reloading proxies: %s", err)
} }
if err := a.loadChecks(newCfg); err != nil { if err := a.loadChecks(newCfg, snap); err != nil {
return fmt.Errorf("Failed reloading checks: %s", err) return fmt.Errorf("Failed reloading checks: %s", err)
} }
if err := a.loadMetadata(newCfg); err != nil { if err := a.loadMetadata(newCfg); err != nil {

View File

@ -500,6 +500,62 @@ func TestAgent_AddService(t *testing.T) {
} }
} }
func TestAgent_AddServices_AliasUpdateCheckNotReverted(t *testing.T) {
t.Parallel()
a := NewTestAgent(t, t.Name(), `
node_name = "node1"
`)
defer a.Shutdown()
// It's tricky to get an UpdateCheck call to be timed properly so it lands
// right in the middle of an addServiceInternal call so we cheat a bit and
// rely upon alias checks to do that work for us. We add enough services
// that probabilistically one of them is going to end up properly in the
// critical section.
//
// The first number I picked here (10) surprisingly failed every time prior
// to PR #6144 solving the underlying problem.
const numServices = 10
services := make([]*structs.ServiceDefinition, numServices)
checkIDs := make([]types.CheckID, numServices)
for i := 0; i < numServices; i++ {
name := fmt.Sprintf("web-%d", i)
services[i] = &structs.ServiceDefinition{
ID: name,
Name: name,
Port: 8080 + i,
Checks: []*structs.CheckType{
&structs.CheckType{
Name: "alias-for-fake-service",
AliasService: "fake",
},
},
}
checkIDs[i] = types.CheckID("service:" + name)
}
// Add all of the services quickly as you might do from config file snippets.
for _, service := range services {
ns := service.NodeService()
chkTypes, err := service.CheckTypes()
require.NoError(t, err)
require.NoError(t, a.AddService(ns, chkTypes, false, service.Token, ConfigSourceLocal))
}
retry.Run(t, func(r *retry.R) {
gotChecks := a.State.Checks()
for id, check := range gotChecks {
require.Equal(r, "passing", check.Status, "check %q is wrong", id)
require.Equal(r, "No checks found.", check.Output, "check %q is wrong", id)
}
})
}
func TestAgent_AddServiceNoExec(t *testing.T) { func TestAgent_AddServiceNoExec(t *testing.T) {
t.Parallel() t.Parallel()
a := NewTestAgent(t, t.Name(), ` a := NewTestAgent(t, t.Name(), `
@ -2965,14 +3021,11 @@ func TestAgent_checkStateSnapshot(t *testing.T) {
t.Fatalf("err: %s", err) t.Fatalf("err: %s", err)
} }
// Reload the checks // Reload the checks and restore the snapshot.
if err := a.loadChecks(a.Config); err != nil { if err := a.loadChecks(a.Config, snap); err != nil {
t.Fatalf("err: %s", err) t.Fatalf("err: %s", err)
} }
// Restore the state
a.restoreCheckState(snap)
// Search for the check // Search for the check
out, ok := a.State.Checks()[check1.CheckID] out, ok := a.State.Checks()[check1.CheckID]
if !ok { if !ok {
@ -3010,7 +3063,7 @@ func TestAgent_loadChecks_checkFails(t *testing.T) {
} }
// Try loading the checks from the persisted files // Try loading the checks from the persisted files
if err := a.loadChecks(a.Config); err != nil { if err := a.loadChecks(a.Config, nil); err != nil {
t.Fatalf("err: %s", err) t.Fatalf("err: %s", err)
} }

View File

@ -68,13 +68,15 @@ func TestLeader_RegisterMember(t *testing.T) {
} }
// Server should be registered // Server should be registered
retry.Run(t, func(r *retry.R) {
_, node, err := state.GetNode(s1.config.NodeName) _, node, err := state.GetNode(s1.config.NodeName)
if err != nil { if err != nil {
t.Fatalf("err: %v", err) r.Fatalf("err: %v", err)
} }
if node == nil { if node == nil {
t.Fatalf("server not registered") r.Fatalf("server not registered")
} }
})
// Service should be registered // Service should be registered
_, services, err := state.NodeServices(nil, s1.config.NodeName) _, services, err := state.NodeServices(nil, s1.config.NodeName)

View File

@ -742,53 +742,41 @@ func TestServer_AvoidReBootstrap(t *testing.T) {
func TestServer_Expect_NonVoters(t *testing.T) { func TestServer_Expect_NonVoters(t *testing.T) {
t.Parallel() t.Parallel()
dir1, s1 := testServerDCExpectNonVoter(t, "dc1", 3) dir1, s1 := testServerDCExpectNonVoter(t, "dc1", 2)
defer os.RemoveAll(dir1) defer os.RemoveAll(dir1)
defer s1.Shutdown() defer s1.Shutdown()
dir2, s2 := testServerDCExpectNonVoter(t, "dc1", 3) dir2, s2 := testServerDCExpect(t, "dc1", 2)
defer os.RemoveAll(dir2) defer os.RemoveAll(dir2)
defer s2.Shutdown() defer s2.Shutdown()
dir3, s3 := testServerDCExpect(t, "dc1", 3) dir3, s3 := testServerDCExpect(t, "dc1", 2)
defer os.RemoveAll(dir3) defer os.RemoveAll(dir3)
defer s3.Shutdown() defer s3.Shutdown()
dir4, s4 := testServerDCExpect(t, "dc1", 3) // Join the first two servers.
defer os.RemoveAll(dir4)
defer s4.Shutdown()
dir5, s5 := testServerDCExpect(t, "dc1", 3)
defer os.RemoveAll(dir5)
defer s5.Shutdown()
// Join the first three servers.
joinLAN(t, s2, s1) joinLAN(t, s2, s1)
joinLAN(t, s3, s1)
// Should have no peers yet since the bootstrap didn't occur. // Should have no peers yet since the bootstrap didn't occur.
retry.Run(t, func(r *retry.R) { retry.Run(t, func(r *retry.R) {
r.Check(wantPeers(s1, 0)) r.Check(wantPeers(s1, 0))
r.Check(wantPeers(s2, 0)) r.Check(wantPeers(s2, 0))
r.Check(wantPeers(s3, 0))
}) })
// Join the fourth node. // Join the third node.
joinLAN(t, s4, s1) joinLAN(t, s3, s1)
joinLAN(t, s5, s1)
// Now we have three servers so we should bootstrap. // Now we have three servers so we should bootstrap.
retry.Run(t, func(r *retry.R) { retry.Run(t, func(r *retry.R) {
r.Check(wantPeers(s1, 3)) r.Check(wantPeers(s1, 2))
r.Check(wantPeers(s2, 3)) r.Check(wantPeers(s2, 2))
r.Check(wantPeers(s3, 3)) r.Check(wantPeers(s3, 2))
r.Check(wantPeers(s4, 3))
}) })
// Make sure a leader is elected // Make sure a leader is elected
testrpc.WaitForLeader(t, s1.RPC, "dc1") testrpc.WaitForLeader(t, s1.RPC, "dc1")
retry.Run(t, func(r *retry.R) { retry.Run(t, func(r *retry.R) {
r.Check(wantRaft([]*Server{s1, s2, s3, s4, s5})) r.Check(wantRaft([]*Server{s1, s2, s3}))
}) })
} }

View File

@ -1,19 +1,18 @@
package api package api
import ( import (
"github.com/hashicorp/consul/sdk/testutil/retry"
"strings" "strings"
"testing" "testing"
"time" "time"
"github.com/hashicorp/go-uuid" "github.com/hashicorp/go-uuid"
"github.com/pascaldekloe/goe/verify"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
) )
func TestAPI_ClientTxn(t *testing.T) { func TestAPI_ClientTxn(t *testing.T) {
t.Parallel() t.Parallel()
require := require.New(t)
c, s := makeClient(t) c, s := makeClient(t)
defer s.Stop() defer s.Stop()
@ -24,7 +23,7 @@ func TestAPI_ClientTxn(t *testing.T) {
// Set up a test service and health check. // Set up a test service and health check.
nodeID, err := uuid.GenerateUUID() nodeID, err := uuid.GenerateUUID()
require.NoError(err) require.NoError(t, err)
catalog := c.Catalog() catalog := c.Catalog()
reg := &CatalogRegistration{ reg := &CatalogRegistration{
@ -59,11 +58,11 @@ func TestAPI_ClientTxn(t *testing.T) {
}, },
} }
_, err = catalog.Register(reg, nil) _, err = catalog.Register(reg, nil)
require.NoError(err) require.NoError(t, err)
node, _, err := catalog.Node("foo", nil) node, _, err := catalog.Node("foo", nil)
require.NoError(err) require.NoError(t, err)
require.Equal(nodeID, node.Node.ID) require.Equal(t, nodeID, node.Node.ID)
// Make a session. // Make a session.
id, _, err := session.CreateNoChecks(nil, nil) id, _, err := session.CreateNoChecks(nil, nil)
@ -218,8 +217,9 @@ func TestAPI_ClientTxn(t *testing.T) {
}, },
}, },
} }
verify.Values(t, "", ret.Results, expected) require.Equal(t, ret.Results, expected)
retry.Run(t, func(r *retry.R) {
// Run a read-only transaction. // Run a read-only transaction.
ops = TxnOps{ ops = TxnOps{
&TxnOp{ &TxnOp{
@ -237,9 +237,9 @@ func TestAPI_ClientTxn(t *testing.T) {
} }
ok, ret, _, err = txn.Txn(ops, nil) ok, ret, _, err = txn.Txn(ops, nil)
if err != nil { if err != nil {
t.Fatalf("err: %v", err) r.Fatalf("err: %v", err)
} else if !ok { } else if !ok {
t.Fatalf("transaction failure") r.Fatalf("transaction failure")
} }
expected = TxnResults{ expected = TxnResults{
@ -269,7 +269,8 @@ func TestAPI_ClientTxn(t *testing.T) {
}, },
}, },
} }
verify.Values(t, "", ret.Results, expected) require.Equal(r, ret.Results, expected)
})
// Sanity check using the regular GET API. // Sanity check using the regular GET API.
kv := c.KV() kv := c.KV()

View File

@ -715,7 +715,7 @@ func TestConnectRootsWatch(t *testing.T) {
return // ignore return // ignore
} }
v, ok := raw.(*api.CARootList) v, ok := raw.(*api.CARootList)
if !ok || v == nil { if !ok || v == nil || len(v.Roots) == 0 {
return // ignore return // ignore
} }
// Only 1 CA is the bootstrapped state (i.e. first response). Ignore this // Only 1 CA is the bootstrapped state (i.e. first response). Ignore this