status-go/peers/peerpool.go

411 lines
11 KiB
Go
Raw Normal View History

2018-04-10 06:44:09 +00:00
package peers
import (
"errors"
"sync"
"time"
"github.com/ethereum/go-ethereum/common"
2018-04-10 06:44:09 +00:00
"github.com/ethereum/go-ethereum/common/mclock"
"github.com/ethereum/go-ethereum/event"
"github.com/ethereum/go-ethereum/log"
"github.com/ethereum/go-ethereum/p2p"
2018-04-12 13:08:49 +00:00
"github.com/ethereum/go-ethereum/p2p/discover"
2018-04-10 06:44:09 +00:00
"github.com/ethereum/go-ethereum/p2p/discv5"
"github.com/status-im/status-go/contracts"
"github.com/status-im/status-go/discovery"
"github.com/status-im/status-go/mailserver/registry"
"github.com/status-im/status-go/params"
"github.com/status-im/status-go/peers/verifier"
2018-05-03 07:35:58 +00:00
"github.com/status-im/status-go/signal"
2018-04-10 06:44:09 +00:00
)
var (
// ErrDiscv5NotRunning returned when pool is started but discover v5 is not running or not enabled.
ErrDiscv5NotRunning = errors.New("Discovery v5 is not running")
)
// PoolEvent is a type used to for peer pool events.
type PoolEvent string
2018-04-10 06:44:09 +00:00
const (
// expirationPeriod is an amount of time while peer is considered as a connectable
expirationPeriod = 60 * time.Minute
2018-04-12 13:08:49 +00:00
// discoveryRestartTimeout defines how often loop will try to start discovery server
discoveryRestartTimeout = 2 * time.Second
2018-04-10 06:44:09 +00:00
// DefaultFastSync is a recommended value for aggressive peers search.
DefaultFastSync = 3 * time.Second
// DefaultSlowSync is a recommended value for slow (background) peers search.
DefaultSlowSync = 30 * time.Second
// DefaultDiscV5Timeout is a timeout after which Discv5 is stopped.
DefaultDiscV5Timeout = 3 * time.Minute
// DefaultTopicFastModeTimeout is a timeout after which sync mode is switched to slow mode.
DefaultTopicFastModeTimeout = 30 * time.Second
// DefaultTopicStopSearchDelay is the default delay when stopping a topic search.
DefaultTopicStopSearchDelay = 10 * time.Second
2018-04-10 06:44:09 +00:00
)
// Options is a struct with PeerPool configuration.
type Options struct {
FastSync time.Duration
SlowSync time.Duration
// After this time, Discovery is stopped even if max peers is not reached.
DiscServerTimeout time.Duration
// AllowStop allows stopping Discovery when reaching max peers or after timeout.
AllowStop bool
// TopicStopSearchDelay time stopSearch will be waiting for max cached peers to be
// filled before really stopping the search.
TopicStopSearchDelay time.Duration
// TrustedMailServers is a list of trusted nodes.
TrustedMailServers []discover.NodeID
// MailServerRegistryAddress is the MailServerRegistry contract address
MailServerRegistryAddress string
}
// NewDefaultOptions returns a struct with default Options.
func NewDefaultOptions() *Options {
return &Options{
FastSync: DefaultFastSync,
SlowSync: DefaultSlowSync,
DiscServerTimeout: DefaultDiscV5Timeout,
AllowStop: false,
TopicStopSearchDelay: DefaultTopicStopSearchDelay,
2018-04-10 06:44:09 +00:00
}
}
type peerInfo struct {
// discoveredTime last time when node was found by v5
discoveredTime mclock.AbsTime
// dismissed is true when our node requested a disconnect
dismissed bool
2018-04-10 06:44:09 +00:00
node *discv5.Node
}
// PeerPool manages discovered peers and connects them to p2p server
type PeerPool struct {
opts *Options
discovery discovery.Discovery
2018-07-03 11:27:04 +00:00
2018-04-10 06:44:09 +00:00
// config can be set only once per pool life cycle
config map[discv5.Topic]params.Limits
cache *Cache
2018-04-10 06:44:09 +00:00
mu sync.RWMutex
topics []TopicPoolInterface
2018-04-10 06:44:09 +00:00
serverSubscription event.Subscription
events chan *p2p.PeerEvent
2018-04-10 06:44:09 +00:00
quit chan struct{}
wg sync.WaitGroup
timeout <-chan time.Time
updateTopic chan *updateTopicRequest
2018-04-10 06:44:09 +00:00
}
// NewPeerPool creates instance of PeerPool
func NewPeerPool(discovery discovery.Discovery, config map[discv5.Topic]params.Limits, cache *Cache, options *Options) *PeerPool {
return &PeerPool{
2018-07-03 11:27:04 +00:00
opts: options,
discovery: discovery,
config: config,
cache: cache,
}
}
func (p *PeerPool) setDiscoveryTimeout() {
if p.opts.AllowStop && p.opts.DiscServerTimeout > 0 {
p.timeout = time.After(p.opts.DiscServerTimeout)
}
}
2018-04-10 06:44:09 +00:00
// Start creates topic pool for each topic in config and subscribes to server events.
func (p *PeerPool) Start(server *p2p.Server, rpcClient contracts.RPCClient) error {
2018-07-03 11:27:04 +00:00
if !p.discovery.Running() {
2018-04-10 06:44:09 +00:00
return ErrDiscv5NotRunning
}
2018-04-10 06:44:09 +00:00
p.mu.Lock()
defer p.mu.Unlock()
// init channels
2018-04-10 06:44:09 +00:00
p.quit = make(chan struct{})
p.updateTopic = make(chan *updateTopicRequest)
p.setDiscoveryTimeout()
// subscribe to peer events
p.events = make(chan *p2p.PeerEvent, 20)
p.serverSubscription = server.SubscribeEvents(p.events)
p.wg.Add(1)
go func() {
p.handleServerPeers(server, p.events)
p.wg.Done()
}()
// collect topics and start searching for nodes
p.topics = make([]TopicPoolInterface, 0, len(p.config))
2018-04-10 06:44:09 +00:00
for topic, limits := range p.config {
var topicPool TopicPoolInterface
t := newTopicPool(p.discovery, topic, limits, p.opts.SlowSync, p.opts.FastSync, p.cache)
if topic == MailServerDiscoveryTopic {
v, err := p.initVerifier(rpcClient)
if err != nil {
return err
}
topicPool = newCacheOnlyTopicPool(t, v)
} else {
topicPool = t
}
2018-04-10 06:44:09 +00:00
if err := topicPool.StartSearch(server); err != nil {
return err
}
p.topics = append(p.topics, topicPool)
}
// discovery must be already started when pool is started
signal.SendDiscoveryStarted()
2018-04-10 06:44:09 +00:00
return nil
}
func (p *PeerPool) initVerifier(rpcClient contracts.RPCClient) (v Verifier, err error) {
if addr := p.opts.MailServerRegistryAddress; addr != "" {
caller := contracts.NewContractCaller(rpcClient)
addrBytes := common.FromHex(addr)
return registry.NewVerifier(caller, common.BytesToAddress(addrBytes))
}
return verifier.NewLocalVerifier(p.opts.TrustedMailServers), nil
}
2018-07-03 11:27:04 +00:00
func (p *PeerPool) startDiscovery() error {
if p.discovery.Running() {
return nil
}
2018-07-03 11:27:04 +00:00
if err := p.discovery.Start(); err != nil {
return err
}
p.mu.Lock()
p.setDiscoveryTimeout()
p.mu.Unlock()
2018-05-03 07:35:58 +00:00
signal.SendDiscoveryStarted()
return nil
}
func (p *PeerPool) stopDiscovery(server *p2p.Server) {
2018-07-03 11:27:04 +00:00
if !p.discovery.Running() {
return
}
if err := p.discovery.Stop(); err != nil {
log.Error("discovery errored when stopping", "err", err)
}
for _, t := range p.topics {
t.StopSearch(server)
}
p.mu.Lock()
p.timeout = nil
p.mu.Unlock()
2018-05-03 07:35:58 +00:00
signal.SendDiscoveryStopped()
}
2018-04-12 13:08:49 +00:00
// restartDiscovery and search for topics that have peer count below min
func (p *PeerPool) restartDiscovery(server *p2p.Server) error {
2018-07-03 11:27:04 +00:00
if !p.discovery.Running() {
if err := p.startDiscovery(); err != nil {
2018-04-12 13:08:49 +00:00
return err
}
log.Debug("restarted discovery from peer pool")
}
for _, t := range p.topics {
if !t.BelowMin() || t.SearchRunning() {
continue
}
err := t.StartSearch(server)
if err != nil {
log.Error("search failed to start", "error", err)
}
}
return nil
}
2018-04-10 06:44:09 +00:00
// handleServerPeers watches server peer events, notifies topic pools about changes
// in the peer set and stops the discv5 if all topic pools collected enough peers.
//
// @TODO(adam): split it into peers and discovery management loops. This should
// simplify the whole logic and allow to remove `timeout` field from `PeerPool`.
2018-04-10 06:44:09 +00:00
func (p *PeerPool) handleServerPeers(server *p2p.Server, events <-chan *p2p.PeerEvent) {
2018-04-12 13:08:49 +00:00
var retryDiscv5 <-chan time.Time
var stopDiscv5 <-chan time.Time
2018-04-12 13:08:49 +00:00
2018-04-10 06:44:09 +00:00
for {
p.mu.RLock()
timeout := p.timeout
p.mu.RUnlock()
2018-04-10 06:44:09 +00:00
select {
case <-p.quit:
2018-07-03 11:27:04 +00:00
log.Debug("stopping DiscV5 because of quit")
p.stopDiscovery(server)
2018-04-10 06:44:09 +00:00
return
case <-timeout:
2018-07-03 11:27:04 +00:00
log.Info("DiscV5 timed out")
p.stopDiscovery(server)
2018-04-12 13:08:49 +00:00
case <-retryDiscv5:
if err := p.restartDiscovery(server); err != nil {
retryDiscv5 = time.After(discoveryRestartTimeout)
log.Error("starting discv5 failed", "error", err, "retry", discoveryRestartTimeout)
2018-04-12 13:08:49 +00:00
}
case <-stopDiscv5:
p.handleStopTopics(server)
case req := <-p.updateTopic:
if p.updateTopicLimits(server, req) == nil {
if !p.discovery.Running() {
retryDiscv5 = time.After(0)
}
}
2018-04-10 06:44:09 +00:00
case event := <-events:
switch event.Type {
case p2p.PeerEventTypeDrop:
2018-04-12 13:08:49 +00:00
log.Debug("confirm peer dropped", "ID", event.Peer)
if p.handleDroppedPeer(server, event.Peer) {
2018-04-12 13:08:49 +00:00
retryDiscv5 = time.After(0)
2018-04-10 06:44:09 +00:00
}
case p2p.PeerEventTypeAdd: // skip other events
log.Debug("confirm peer added", "ID", event.Peer)
p.handleAddedPeer(server, event.Peer)
stopDiscv5 = time.After(p.opts.TopicStopSearchDelay)
default:
continue
2018-04-10 06:44:09 +00:00
}
SendDiscoverySummary(server.PeersInfo())
2018-04-10 06:44:09 +00:00
}
}
}
// handleAddedPeer notifies all topics about added peer.
func (p *PeerPool) handleAddedPeer(server *p2p.Server, nodeID discover.NodeID) {
2018-04-12 13:08:49 +00:00
p.mu.Lock()
defer p.mu.Unlock()
for _, t := range p.topics {
t.ConfirmAdded(server, nodeID)
if p.opts.AllowStop && t.MaxReached() {
t.setStopSearchTimeout(p.opts.TopicStopSearchDelay)
}
}
}
// handleStopTopics stops the search on any topics having reached its max cached
// limit or its delay stop is expired, additionally will stop discovery if all
// peers are stopped.
func (p *PeerPool) handleStopTopics(server *p2p.Server) {
if !p.opts.AllowStop {
return
}
for _, t := range p.topics {
if t.readyToStopSearch() {
t.StopSearch(server)
}
}
if p.allTopicsStopped() {
2018-07-03 11:27:04 +00:00
log.Debug("closing discv5 connection because all topics reached max limit")
p.stopDiscovery(server)
}
}
// allTopicsStopped returns true if all topics are stopped.
func (p *PeerPool) allTopicsStopped() (all bool) {
if !p.opts.AllowStop {
return false
}
all = true
for _, t := range p.topics {
if !t.isStopped() {
2018-04-12 13:08:49 +00:00
all = false
}
}
return all
}
// handleDroppedPeer notifies every topic about dropped peer and returns true if any peer have connections
// below min limit
func (p *PeerPool) handleDroppedPeer(server *p2p.Server, nodeID discover.NodeID) (any bool) {
p.mu.Lock()
defer p.mu.Unlock()
for _, t := range p.topics {
confirmed := t.ConfirmDropped(server, nodeID)
if confirmed {
newPeer := t.AddPeerFromTable(server)
if newPeer != nil {
log.Debug("added peer from local table", "ID", newPeer.ID)
}
}
log.Debug("search", "topic", t.Topic(), "below min", t.BelowMin())
2018-04-12 13:08:49 +00:00
if t.BelowMin() && !t.SearchRunning() {
any = true
}
}
return any
}
2018-04-10 06:44:09 +00:00
// Stop closes pool quit channel and all channels that are watched by search queries
// and waits till all goroutines will exit.
func (p *PeerPool) Stop() {
// pool wasn't started
if p.quit == nil {
return
}
select {
case <-p.quit:
return
default:
log.Debug("started closing peer pool")
close(p.quit)
}
p.serverSubscription.Unsubscribe()
p.wg.Wait()
2018-04-10 06:44:09 +00:00
}
type updateTopicRequest struct {
Topic string
Limits params.Limits
}
// UpdateTopic updates the pre-existing TopicPool limits.
func (p *PeerPool) UpdateTopic(topic string, limits params.Limits) error {
if _, err := p.getTopic(topic); err != nil {
return err
}
p.updateTopic <- &updateTopicRequest{
Topic: topic,
Limits: limits,
}
return nil
}
func (p *PeerPool) updateTopicLimits(server *p2p.Server, req *updateTopicRequest) error {
t, err := p.getTopic(req.Topic)
if err != nil {
return err
}
t.SetLimits(req.Limits)
return nil
}
func (p *PeerPool) getTopic(topic string) (TopicPoolInterface, error) {
for _, t := range p.topics {
if t.Topic() == discv5.Topic(topic) {
return t, nil
}
}
return nil, errors.New("topic not found")
}