2018-07-04 10:51:47 +00:00
|
|
|
package swarm
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
2019-06-09 07:24:20 +00:00
|
|
|
"github.com/libp2p/go-libp2p-core/network"
|
|
|
|
"github.com/libp2p/go-libp2p-core/peer"
|
|
|
|
"github.com/libp2p/go-libp2p-core/transport"
|
|
|
|
lgbl "github.com/libp2p/go-libp2p-loggables"
|
|
|
|
|
2018-07-04 10:51:47 +00:00
|
|
|
logging "github.com/ipfs/go-log"
|
|
|
|
addrutil "github.com/libp2p/go-addr-util"
|
|
|
|
ma "github.com/multiformats/go-multiaddr"
|
|
|
|
)
|
|
|
|
|
|
|
|
// Diagram of dial sync:
|
|
|
|
//
|
|
|
|
// many callers of Dial() synched w. dials many addrs results to callers
|
|
|
|
// ----------------------\ dialsync use earliest /--------------
|
|
|
|
// -----------------------\ |----------\ /----------------
|
|
|
|
// ------------------------>------------<------- >---------<-----------------
|
|
|
|
// -----------------------| \----x \----------------
|
|
|
|
// ----------------------| \-----x \---------------
|
|
|
|
// any may fail if no addr at end
|
|
|
|
// retry dialAttempt x
|
|
|
|
|
|
|
|
var (
|
|
|
|
// ErrDialBackoff is returned by the backoff code when a given peer has
|
|
|
|
// been dialed too frequently
|
|
|
|
ErrDialBackoff = errors.New("dial backoff")
|
|
|
|
|
|
|
|
// ErrDialToSelf is returned if we attempt to dial our own peer
|
|
|
|
ErrDialToSelf = errors.New("dial to self attempted")
|
|
|
|
|
|
|
|
// ErrNoTransport is returned when we don't know a transport for the
|
|
|
|
// given multiaddr.
|
|
|
|
ErrNoTransport = errors.New("no transport for protocol")
|
2019-06-09 07:24:20 +00:00
|
|
|
|
|
|
|
// ErrAllDialsFailed is returned when connecting to a peer has ultimately failed
|
|
|
|
ErrAllDialsFailed = errors.New("all dials failed")
|
|
|
|
|
|
|
|
// ErrNoAddresses is returned when we fail to find any addresses for a
|
|
|
|
// peer we're trying to dial.
|
|
|
|
ErrNoAddresses = errors.New("no addresses")
|
|
|
|
|
|
|
|
// ErrNoAddresses is returned when we find addresses for a peer but
|
|
|
|
// can't use any of them.
|
|
|
|
ErrNoGoodAddresses = errors.New("no good addresses")
|
2018-07-04 10:51:47 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// DialAttempts governs how many times a goroutine will try to dial a given peer.
|
|
|
|
// Note: this is down to one, as we have _too many dials_ atm. To add back in,
|
|
|
|
// add loop back in Dial(.)
|
|
|
|
const DialAttempts = 1
|
|
|
|
|
|
|
|
// ConcurrentFdDials is the number of concurrent outbound dials over transports
|
|
|
|
// that consume file descriptors
|
|
|
|
const ConcurrentFdDials = 160
|
|
|
|
|
|
|
|
// DefaultPerPeerRateLimit is the number of concurrent outbound dials to make
|
|
|
|
// per peer
|
|
|
|
const DefaultPerPeerRateLimit = 8
|
|
|
|
|
|
|
|
// dialbackoff is a struct used to avoid over-dialing the same, dead peers.
|
|
|
|
// Whenever we totally time out on a peer (all three attempts), we add them
|
|
|
|
// to dialbackoff. Then, whenevers goroutines would _wait_ (dialsync), they
|
|
|
|
// check dialbackoff. If it's there, they don't wait and exit promptly with
|
|
|
|
// an error. (the single goroutine that is actually dialing continues to
|
|
|
|
// dial). If a dial is successful, the peer is removed from backoff.
|
|
|
|
// Example:
|
|
|
|
//
|
|
|
|
// for {
|
|
|
|
// if ok, wait := dialsync.Lock(p); !ok {
|
|
|
|
// if backoff.Backoff(p) {
|
|
|
|
// return errDialFailed
|
|
|
|
// }
|
|
|
|
// <-wait
|
|
|
|
// continue
|
|
|
|
// }
|
|
|
|
// defer dialsync.Unlock(p)
|
|
|
|
// c, err := actuallyDial(p)
|
|
|
|
// if err != nil {
|
|
|
|
// dialbackoff.AddBackoff(p)
|
|
|
|
// continue
|
|
|
|
// }
|
|
|
|
// dialbackoff.Clear(p)
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
|
|
|
|
// DialBackoff is a type for tracking peer dial backoffs.
|
|
|
|
//
|
2019-06-09 07:24:20 +00:00
|
|
|
// * It's safe to use its zero value.
|
2018-07-04 10:51:47 +00:00
|
|
|
// * It's thread-safe.
|
|
|
|
// * It's *not* safe to move this type after using.
|
|
|
|
type DialBackoff struct {
|
|
|
|
entries map[peer.ID]*backoffPeer
|
|
|
|
lock sync.RWMutex
|
|
|
|
}
|
|
|
|
|
|
|
|
type backoffPeer struct {
|
|
|
|
tries int
|
|
|
|
until time.Time
|
|
|
|
}
|
|
|
|
|
|
|
|
func (db *DialBackoff) init() {
|
|
|
|
if db.entries == nil {
|
|
|
|
db.entries = make(map[peer.ID]*backoffPeer)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Backoff returns whether the client should backoff from dialing
|
|
|
|
// peer p
|
|
|
|
func (db *DialBackoff) Backoff(p peer.ID) (backoff bool) {
|
|
|
|
db.lock.Lock()
|
|
|
|
defer db.lock.Unlock()
|
|
|
|
db.init()
|
|
|
|
bp, found := db.entries[p]
|
|
|
|
if found && time.Now().Before(bp.until) {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// BackoffBase is the base amount of time to backoff (default: 5s).
|
|
|
|
var BackoffBase = time.Second * 5
|
|
|
|
|
|
|
|
// BackoffCoef is the backoff coefficient (default: 1s).
|
|
|
|
var BackoffCoef = time.Second
|
|
|
|
|
|
|
|
// BackoffMax is the maximum backoff time (default: 5m).
|
|
|
|
var BackoffMax = time.Minute * 5
|
|
|
|
|
|
|
|
// AddBackoff lets other nodes know that we've entered backoff with
|
|
|
|
// peer p, so dialers should not wait unnecessarily. We still will
|
|
|
|
// attempt to dial with one goroutine, in case we get through.
|
|
|
|
//
|
|
|
|
// Backoff is not exponential, it's quadratic and computed according to the
|
|
|
|
// following formula:
|
|
|
|
//
|
|
|
|
// BackoffBase + BakoffCoef * PriorBackoffs^2
|
|
|
|
//
|
|
|
|
// Where PriorBackoffs is the number of previous backoffs.
|
|
|
|
func (db *DialBackoff) AddBackoff(p peer.ID) {
|
|
|
|
db.lock.Lock()
|
|
|
|
defer db.lock.Unlock()
|
|
|
|
db.init()
|
|
|
|
bp, ok := db.entries[p]
|
|
|
|
if !ok {
|
|
|
|
db.entries[p] = &backoffPeer{
|
|
|
|
tries: 1,
|
|
|
|
until: time.Now().Add(BackoffBase),
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
backoffTime := BackoffBase + BackoffCoef*time.Duration(bp.tries*bp.tries)
|
|
|
|
if backoffTime > BackoffMax {
|
|
|
|
backoffTime = BackoffMax
|
|
|
|
}
|
|
|
|
bp.until = time.Now().Add(backoffTime)
|
|
|
|
bp.tries++
|
|
|
|
}
|
|
|
|
|
|
|
|
// Clear removes a backoff record. Clients should call this after a
|
|
|
|
// successful Dial.
|
|
|
|
func (db *DialBackoff) Clear(p peer.ID) {
|
|
|
|
db.lock.Lock()
|
|
|
|
defer db.lock.Unlock()
|
|
|
|
db.init()
|
|
|
|
delete(db.entries, p)
|
|
|
|
}
|
|
|
|
|
|
|
|
// DialPeer connects to a peer.
|
|
|
|
//
|
|
|
|
// The idea is that the client of Swarm does not need to know what network
|
|
|
|
// the connection will happen over. Swarm can use whichever it choses.
|
|
|
|
// This allows us to use various transport protocols, do NAT traversal/relay,
|
|
|
|
// etc. to achieve connection.
|
2019-06-09 07:24:20 +00:00
|
|
|
func (s *Swarm) DialPeer(ctx context.Context, p peer.ID) (network.Conn, error) {
|
2018-07-04 10:51:47 +00:00
|
|
|
return s.dialPeer(ctx, p)
|
|
|
|
}
|
|
|
|
|
|
|
|
// internal dial method that returns an unwrapped conn
|
|
|
|
//
|
|
|
|
// It is gated by the swarm's dial synchronization systems: dialsync and
|
|
|
|
// dialbackoff.
|
|
|
|
func (s *Swarm) dialPeer(ctx context.Context, p peer.ID) (*Conn, error) {
|
|
|
|
log.Debugf("[%s] swarm dialing peer [%s]", s.local, p)
|
|
|
|
var logdial = lgbl.Dial("swarm", s.LocalPeer(), p, nil, nil)
|
2019-06-09 07:24:20 +00:00
|
|
|
err := p.Validate()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2018-07-04 10:51:47 +00:00
|
|
|
if p == s.local {
|
|
|
|
log.Event(ctx, "swarmDialSelf", logdial)
|
|
|
|
return nil, ErrDialToSelf
|
|
|
|
}
|
|
|
|
|
|
|
|
defer log.EventBegin(ctx, "swarmDialAttemptSync", p).Done()
|
|
|
|
|
|
|
|
// check if we already have an open connection first
|
|
|
|
conn := s.bestConnToPeer(p)
|
|
|
|
if conn != nil {
|
|
|
|
return conn, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// if this peer has been backed off, lets get out of here
|
|
|
|
if s.backf.Backoff(p) {
|
|
|
|
log.Event(ctx, "swarmDialBackoff", p)
|
|
|
|
return nil, ErrDialBackoff
|
|
|
|
}
|
|
|
|
|
|
|
|
// apply the DialPeer timeout
|
2019-06-09 07:24:20 +00:00
|
|
|
ctx, cancel := context.WithTimeout(ctx, network.GetDialPeerTimeout(ctx))
|
2018-07-04 10:51:47 +00:00
|
|
|
defer cancel()
|
|
|
|
|
2019-06-09 07:24:20 +00:00
|
|
|
conn, err = s.dsync.DialLock(ctx, p)
|
2018-07-04 10:51:47 +00:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Debugf("network for %s finished dialing %s", s.local, p)
|
|
|
|
return conn, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// doDial is an ugly shim method to retain all the logging and backoff logic
|
|
|
|
// of the old dialsync code
|
|
|
|
func (s *Swarm) doDial(ctx context.Context, p peer.ID) (*Conn, error) {
|
|
|
|
// Short circuit.
|
|
|
|
// By the time we take the dial lock, we may already *have* a connection
|
|
|
|
// to the peer.
|
|
|
|
c := s.bestConnToPeer(p)
|
|
|
|
if c != nil {
|
|
|
|
return c, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
logdial := lgbl.Dial("swarm", s.LocalPeer(), p, nil, nil)
|
|
|
|
|
|
|
|
// ok, we have been charged to dial! let's do it.
|
|
|
|
// if it succeeds, dial will add the conn to the swarm itself.
|
|
|
|
defer log.EventBegin(ctx, "swarmDialAttemptStart", logdial).Done()
|
|
|
|
|
|
|
|
conn, err := s.dial(ctx, p)
|
|
|
|
if err != nil {
|
|
|
|
conn = s.bestConnToPeer(p)
|
|
|
|
if conn != nil {
|
|
|
|
// Hm? What error?
|
|
|
|
// Could have canceled the dial because we received a
|
|
|
|
// connection or some other random reason.
|
|
|
|
// Just ignore the error and return the connection.
|
|
|
|
log.Debugf("ignoring dial error because we have a connection: %s", err)
|
|
|
|
return conn, nil
|
|
|
|
}
|
|
|
|
if err != context.Canceled {
|
|
|
|
log.Event(ctx, "swarmDialBackoffAdd", logdial)
|
|
|
|
s.backf.AddBackoff(p) // let others know to backoff
|
|
|
|
}
|
|
|
|
|
|
|
|
// ok, we failed.
|
2019-06-09 07:24:20 +00:00
|
|
|
return nil, err
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
|
|
|
return conn, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Swarm) canDial(addr ma.Multiaddr) bool {
|
|
|
|
t := s.TransportForDialing(addr)
|
|
|
|
return t != nil && t.CanDial(addr)
|
|
|
|
}
|
|
|
|
|
|
|
|
// dial is the actual swarm's dial logic, gated by Dial.
|
|
|
|
func (s *Swarm) dial(ctx context.Context, p peer.ID) (*Conn, error) {
|
|
|
|
var logdial = lgbl.Dial("swarm", s.LocalPeer(), p, nil, nil)
|
|
|
|
if p == s.local {
|
|
|
|
log.Event(ctx, "swarmDialDoDialSelf", logdial)
|
|
|
|
return nil, ErrDialToSelf
|
|
|
|
}
|
|
|
|
defer log.EventBegin(ctx, "swarmDialDo", logdial).Done()
|
|
|
|
logdial["dial"] = "failure" // start off with failure. set to "success" at the end.
|
|
|
|
|
|
|
|
sk := s.peers.PrivKey(s.local)
|
2019-06-09 07:24:20 +00:00
|
|
|
logdial["encrypted"] = sk != nil // log whether this will be an encrypted dial or not.
|
2018-07-04 10:51:47 +00:00
|
|
|
if sk == nil {
|
|
|
|
// fine for sk to be nil, just log.
|
|
|
|
log.Debug("Dial not given PrivateKey, so WILL NOT SECURE conn.")
|
|
|
|
}
|
|
|
|
|
|
|
|
//////
|
|
|
|
/*
|
2019-06-09 07:24:20 +00:00
|
|
|
This slice-to-chan code is temporary, the peerstore can currently provide
|
2018-07-04 10:51:47 +00:00
|
|
|
a channel as an interface for receiving addresses, but more thought
|
|
|
|
needs to be put into the execution. For now, this allows us to use
|
|
|
|
the improved rate limiter, while maintaining the outward behaviour
|
|
|
|
that we previously had (halting a dial when we run out of addrs)
|
|
|
|
*/
|
2019-06-09 07:24:20 +00:00
|
|
|
peerAddrs := s.peers.Addrs(p)
|
|
|
|
if len(peerAddrs) == 0 {
|
|
|
|
return nil, &DialError{Peer: p, Cause: ErrNoAddresses}
|
|
|
|
}
|
|
|
|
goodAddrs := s.filterKnownUndialables(peerAddrs)
|
|
|
|
if len(goodAddrs) == 0 {
|
|
|
|
return nil, &DialError{Peer: p, Cause: ErrNoGoodAddresses}
|
|
|
|
}
|
|
|
|
goodAddrsChan := make(chan ma.Multiaddr, len(goodAddrs))
|
2018-07-04 10:51:47 +00:00
|
|
|
for _, a := range goodAddrs {
|
2019-06-09 07:24:20 +00:00
|
|
|
goodAddrsChan <- a
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
2019-06-09 07:24:20 +00:00
|
|
|
close(goodAddrsChan)
|
2018-07-04 10:51:47 +00:00
|
|
|
/////////
|
|
|
|
|
|
|
|
// try to get a connection to any addr
|
2019-06-09 07:24:20 +00:00
|
|
|
connC, dialErr := s.dialAddrs(ctx, p, goodAddrsChan)
|
|
|
|
if dialErr != nil {
|
|
|
|
logdial["error"] = dialErr.Cause.Error()
|
|
|
|
if dialErr.Cause == context.Canceled {
|
|
|
|
// always prefer the "context canceled" error.
|
|
|
|
// we rely on behing able to check `err == context.Canceled`
|
|
|
|
//
|
|
|
|
// Removing this will BREAK backoff (causing us to
|
|
|
|
// backoff when canceling dials).
|
|
|
|
return nil, context.Canceled
|
|
|
|
}
|
|
|
|
return nil, dialErr
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
|
|
|
logdial["conn"] = logging.Metadata{
|
|
|
|
"localAddr": connC.LocalMultiaddr(),
|
|
|
|
"remoteAddr": connC.RemoteMultiaddr(),
|
|
|
|
}
|
2019-06-09 07:24:20 +00:00
|
|
|
swarmC, err := s.addConn(connC, network.DirOutbound)
|
2018-07-04 10:51:47 +00:00
|
|
|
if err != nil {
|
|
|
|
logdial["error"] = err.Error()
|
|
|
|
connC.Close() // close the connection. didn't work out :(
|
2019-06-09 07:24:20 +00:00
|
|
|
return nil, &DialError{Peer: p, Cause: err}
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
logdial["dial"] = "success"
|
|
|
|
return swarmC, nil
|
|
|
|
}
|
|
|
|
|
2019-06-09 07:24:20 +00:00
|
|
|
// filterKnownUndialables takes a list of multiaddrs, and removes those
|
|
|
|
// that we definitely don't want to dial: addresses configured to be blocked,
|
|
|
|
// IPv6 link-local addresses, addresses without a dial-capable transport,
|
|
|
|
// and addresses that we know to be our own.
|
|
|
|
// This is an optimization to avoid wasting time on dials that we know are going to fail.
|
|
|
|
func (s *Swarm) filterKnownUndialables(addrs []ma.Multiaddr) []ma.Multiaddr {
|
|
|
|
lisAddrs, _ := s.InterfaceListenAddresses()
|
|
|
|
var ourAddrs []ma.Multiaddr
|
|
|
|
for _, addr := range lisAddrs {
|
|
|
|
protos := addr.Protocols()
|
|
|
|
// we're only sure about filtering out /ip4 and /ip6 addresses, so far
|
|
|
|
if len(protos) == 2 && (protos[0].Code == ma.P_IP4 || protos[0].Code == ma.P_IP6) {
|
|
|
|
ourAddrs = append(ourAddrs, addr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return addrutil.FilterAddrs(addrs,
|
|
|
|
addrutil.SubtractFilter(ourAddrs...),
|
|
|
|
s.canDial,
|
|
|
|
// TODO: Consider allowing link-local addresses
|
|
|
|
addrutil.AddrOverNonLocalIP,
|
|
|
|
addrutil.FilterNeg(s.Filters.AddrBlocked),
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Swarm) dialAddrs(ctx context.Context, p peer.ID, remoteAddrs <-chan ma.Multiaddr) (transport.CapableConn, *DialError) {
|
2018-07-04 10:51:47 +00:00
|
|
|
log.Debugf("%s swarm dialing %s", s.local, p)
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
|
|
defer cancel() // cancel work when we exit func
|
|
|
|
|
|
|
|
// use a single response type instead of errs and conns, reduces complexity *a ton*
|
|
|
|
respch := make(chan dialResult)
|
2019-06-09 07:24:20 +00:00
|
|
|
err := new(DialError)
|
2018-07-04 10:51:47 +00:00
|
|
|
|
|
|
|
defer s.limiter.clearAllPeerDials(p)
|
|
|
|
|
|
|
|
var active int
|
2019-06-09 07:24:20 +00:00
|
|
|
dialLoop:
|
|
|
|
for remoteAddrs != nil || active > 0 {
|
|
|
|
// Check for context cancellations and/or responses first.
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
break dialLoop
|
|
|
|
case resp := <-respch:
|
|
|
|
active--
|
|
|
|
if resp.Err != nil {
|
|
|
|
// Errors are normal, lots of dials will fail
|
|
|
|
log.Infof("got error on dial: %s", resp.Err)
|
|
|
|
err.recordErr(resp.Addr, resp.Err)
|
|
|
|
} else if resp.Conn != nil {
|
|
|
|
return resp.Conn, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// We got a result, try again from the top.
|
|
|
|
continue
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now, attempt to dial.
|
2018-07-04 10:51:47 +00:00
|
|
|
select {
|
|
|
|
case addr, ok := <-remoteAddrs:
|
|
|
|
if !ok {
|
|
|
|
remoteAddrs = nil
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
s.limitedDial(ctx, p, addr, respch)
|
|
|
|
active++
|
|
|
|
case <-ctx.Done():
|
2019-06-09 07:24:20 +00:00
|
|
|
break dialLoop
|
2018-07-04 10:51:47 +00:00
|
|
|
case resp := <-respch:
|
|
|
|
active--
|
|
|
|
if resp.Err != nil {
|
|
|
|
// Errors are normal, lots of dials will fail
|
2019-06-09 07:24:20 +00:00
|
|
|
log.Infof("got error on dial: %s", resp.Err)
|
|
|
|
err.recordErr(resp.Addr, resp.Err)
|
2018-07-04 10:51:47 +00:00
|
|
|
} else if resp.Conn != nil {
|
|
|
|
return resp.Conn, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2019-06-09 07:24:20 +00:00
|
|
|
|
|
|
|
if ctxErr := ctx.Err(); ctxErr != nil {
|
|
|
|
err.Cause = ctxErr
|
|
|
|
} else if len(err.DialErrors) == 0 {
|
|
|
|
err.Cause = network.ErrNoRemoteAddrs
|
|
|
|
} else {
|
|
|
|
err.Cause = ErrAllDialsFailed
|
|
|
|
}
|
|
|
|
return nil, err
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// limitedDial will start a dial to the given peer when
|
|
|
|
// it is able, respecting the various different types of rate
|
|
|
|
// limiting that occur without using extra goroutines per addr
|
|
|
|
func (s *Swarm) limitedDial(ctx context.Context, p peer.ID, a ma.Multiaddr, resp chan dialResult) {
|
|
|
|
s.limiter.AddDialJob(&dialJob{
|
|
|
|
addr: a,
|
|
|
|
peer: p,
|
|
|
|
resp: resp,
|
|
|
|
ctx: ctx,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2019-06-09 07:24:20 +00:00
|
|
|
func (s *Swarm) dialAddr(ctx context.Context, p peer.ID, addr ma.Multiaddr) (transport.CapableConn, error) {
|
2018-07-04 10:51:47 +00:00
|
|
|
// Just to double check. Costs nothing.
|
|
|
|
if s.local == p {
|
|
|
|
return nil, ErrDialToSelf
|
|
|
|
}
|
|
|
|
log.Debugf("%s swarm dialing %s %s", s.local, p, addr)
|
|
|
|
|
|
|
|
tpt := s.TransportForDialing(addr)
|
|
|
|
if tpt == nil {
|
|
|
|
return nil, ErrNoTransport
|
|
|
|
}
|
|
|
|
|
|
|
|
connC, err := tpt.Dial(ctx, addr, p)
|
|
|
|
if err != nil {
|
2019-06-09 07:24:20 +00:00
|
|
|
return nil, err
|
2018-07-04 10:51:47 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Trust the transport? Yeah... right.
|
|
|
|
if connC.RemotePeer() != p {
|
|
|
|
connC.Close()
|
|
|
|
err = fmt.Errorf("BUG in transport %T: tried to dial %s, dialed %s", p, connC.RemotePeer(), tpt)
|
|
|
|
log.Error(err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// success! we got one!
|
|
|
|
return connC, nil
|
|
|
|
}
|