Do peer requests separately for each peer

This commit is contained in:
Matt Joiner 2021-09-18 20:34:14 +10:00
parent d90c41c534
commit 21358ba458
5 changed files with 83 additions and 36 deletions

View File

@ -305,7 +305,7 @@ func NewClient(cfg *ClientConfig) (cl *Client, err error) {
}, },
} }
go cl.requester() //go cl.requester()
return return
} }

View File

@ -365,7 +365,7 @@ func (cn *Peer) writeStatus(w io.Writer, t *Torrent) {
&cn._stats.ChunksReadUseful, &cn._stats.ChunksReadUseful,
&cn._stats.ChunksRead, &cn._stats.ChunksRead,
&cn._stats.ChunksWritten, &cn._stats.ChunksWritten,
cn.numLocalRequests(), len(cn.actualRequestState.Requests),
cn.nominalMaxRequests(), cn.nominalMaxRequests(),
cn.PeerMaxRequests, cn.PeerMaxRequests,
len(cn.peerRequests), len(cn.peerRequests),
@ -463,7 +463,7 @@ func (cn *PeerConn) requestedMetadataPiece(index int) bool {
// The actual value to use as the maximum outbound requests. // The actual value to use as the maximum outbound requests.
func (cn *Peer) nominalMaxRequests() (ret maxRequests) { func (cn *Peer) nominalMaxRequests() (ret maxRequests) {
return int(clamp(1, 2*int64(cn.maxPiecesReceivedBetweenRequestUpdates), int64(cn.PeerMaxRequests))) return maxRequests(clamp(1, int64(cn.PeerMaxRequests), 128))
} }
func (cn *Peer) totalExpectingTime() (ret time.Duration) { func (cn *Peer) totalExpectingTime() (ret time.Duration) {
@ -576,7 +576,7 @@ func (cn *Peer) request(r Request) (more bool, err error) {
if _, ok := cn.actualRequestState.Requests[r]; ok { if _, ok := cn.actualRequestState.Requests[r]; ok {
return true, nil return true, nil
} }
if cn.numLocalRequests() >= cn.nominalMaxRequests() { if len(cn.actualRequestState.Requests) >= cn.nominalMaxRequests() {
return true, errors.New("too many outstanding requests") return true, errors.New("too many outstanding requests")
} }
if cn.actualRequestState.Requests == nil { if cn.actualRequestState.Requests == nil {
@ -653,7 +653,7 @@ func (cn *PeerConn) postBitfield() {
} }
func (cn *PeerConn) updateRequests() { func (cn *PeerConn) updateRequests() {
cn.t.cl.tickleRequester() cn.tickleWriter()
} }
// Emits the indices in the Bitmaps bms in order, never repeating any index. // Emits the indices in the Bitmaps bms in order, never repeating any index.
@ -1453,10 +1453,6 @@ func (c *Peer) peerHasWantedPieces() bool {
return !c._pieceRequestOrder.IsEmpty() return !c._pieceRequestOrder.IsEmpty()
} }
func (c *Peer) numLocalRequests() int {
return len(c.actualRequestState.Requests)
}
func (c *Peer) deleteRequest(r Request) bool { func (c *Peer) deleteRequest(r Request) bool {
delete(c.nextRequestState.Requests, r) delete(c.nextRequestState.Requests, r)
if _, ok := c.actualRequestState.Requests[r]; !ok { if _, ok := c.actualRequestState.Requests[r]; !ok {

View File

@ -1,11 +1,13 @@
package request_strategy package request_strategy
import ( import (
"bytes"
"fmt" "fmt"
"sort" "sort"
"sync" "sync"
"github.com/anacrolix/multiless" "github.com/anacrolix/multiless"
"github.com/anacrolix/torrent/metainfo"
"github.com/anacrolix/torrent/storage" "github.com/anacrolix/torrent/storage"
pp "github.com/anacrolix/torrent/peer_protocol" pp "github.com/anacrolix/torrent/peer_protocol"
@ -41,9 +43,12 @@ func sortFilterPieces(pieces []filterPiece) {
i.Availability, j.Availability, i.Availability, j.Availability,
).Int( ).Int(
i.index, j.index, i.index, j.index,
).Uintptr( ).Lazy(func() multiless.Computation {
i.t.StableId, j.t.StableId, return multiless.New().Cmp(bytes.Compare(
).MustLess() i.t.InfoHash[:],
j.t.InfoHash[:],
))
}).MustLess()
}) })
} }
@ -170,7 +175,7 @@ func Run(input Input) map[PeerId]PeerNextRequestState {
}) })
}) })
torrents := input.Torrents torrents := input.Torrents
allPeers := make(map[uintptr][]*requestsPeer, len(torrents)) allPeers := make(map[metainfo.Hash][]*requestsPeer, len(torrents))
for _, t := range torrents { for _, t := range torrents {
peers := make([]*requestsPeer, 0, len(t.Peers)) peers := make([]*requestsPeer, 0, len(t.Peers))
for _, p := range t.Peers { for _, p := range t.Peers {
@ -181,17 +186,17 @@ func Run(input Input) map[PeerId]PeerNextRequestState {
}, },
}) })
} }
allPeers[t.StableId] = peers allPeers[t.InfoHash] = peers
} }
for _, piece := range requestPieces { for _, piece := range requestPieces {
for _, peer := range allPeers[piece.t.StableId] { for _, peer := range allPeers[piece.t.InfoHash] {
if peer.canRequestPiece(piece.index) { if peer.canRequestPiece(piece.index) {
peer.requestablePiecesRemaining++ peer.requestablePiecesRemaining++
} }
} }
} }
for _, piece := range requestPieces { for _, piece := range requestPieces {
allocatePendingChunks(piece, allPeers[piece.t.StableId]) allocatePendingChunks(piece, allPeers[piece.t.InfoHash])
} }
ret := make(map[PeerId]PeerNextRequestState) ret := make(map[PeerId]PeerNextRequestState)
for _, peers := range allPeers { for _, peers := range allPeers {

View File

@ -1,6 +1,7 @@
package request_strategy package request_strategy
import ( import (
"github.com/anacrolix/torrent/metainfo"
"github.com/anacrolix/torrent/storage" "github.com/anacrolix/torrent/storage"
) )
@ -10,7 +11,7 @@ type Torrent struct {
// Unclosed Peers. Not necessary for getting requestable piece ordering. // Unclosed Peers. Not necessary for getting requestable piece ordering.
Peers []Peer Peers []Peer
// Some value that's unique and stable between runs. Could even use the infohash? // Some value that's unique and stable between runs. Could even use the infohash?
StableId uintptr InfoHash metainfo.Hash
MaxUnverifiedBytes int64 MaxUnverifiedBytes int64
} }

View File

@ -5,6 +5,7 @@ import (
"unsafe" "unsafe"
"github.com/anacrolix/missinggo/v2/bitmap" "github.com/anacrolix/missinggo/v2/bitmap"
pp "github.com/anacrolix/torrent/peer_protocol"
"github.com/anacrolix/chansync" "github.com/anacrolix/chansync"
request_strategy "github.com/anacrolix/torrent/request-strategy" request_strategy "github.com/anacrolix/torrent/request-strategy"
@ -43,7 +44,7 @@ func (cl *Client) getRequestStrategyInput() request_strategy.Input {
ts := make([]request_strategy.Torrent, 0, len(cl.torrents)) ts := make([]request_strategy.Torrent, 0, len(cl.torrents))
for _, t := range cl.torrents { for _, t := range cl.torrents {
rst := request_strategy.Torrent{ rst := request_strategy.Torrent{
StableId: uintptr(unsafe.Pointer(t)), InfoHash: t.infoHash,
} }
if t.storage != nil { if t.storage != nil {
rst.Capacity = t.storage.Capacity rst.Capacity = t.storage.Capacity
@ -119,28 +120,72 @@ func setPeerNextRequestState(_p request_strategy.PeerId, rp request_strategy.Pee
} }
func (p *Peer) applyNextRequestState() bool { func (p *Peer) applyNextRequestState() bool {
next := p.nextRequestState if len(p.actualRequestState.Requests) > p.nominalMaxRequests()/2 {
current := p.actualRequestState return true
if !p.setInterested(next.Interested) {
return false
} }
for req := range current.Requests { type piece struct {
if _, ok := next.Requests[req]; !ok { index int
if !p.cancel(req) { endGame bool
return false }
var pieceOrder []piece
request_strategy.GetRequestablePieces(
p.t.cl.getRequestStrategyInput(),
func(t *request_strategy.Torrent, rsp *request_strategy.Piece, pieceIndex int) {
if t.InfoHash != p.t.infoHash {
return
}
if !p.peerHasPiece(pieceIndex) {
return
}
pieceOrder = append(pieceOrder, piece{
index: pieceIndex,
endGame: rsp.Priority == PiecePriorityNow,
})
},
)
more := true
interested := false
for _, endGameIter := range []bool{false, true} {
for _, piece := range pieceOrder {
tp := p.t.piece(piece.index)
tp.iterUndirtiedChunks(func(cs ChunkSpec) {
req := Request{pp.Integer(piece.index), cs}
if !piece.endGame && !endGameIter && p.t.pendingRequests[req] > 0 {
return
}
interested = true
more = p.setInterested(true)
if !more {
return
}
if len(p.actualRequestState.Requests) >= p.nominalMaxRequests() {
return
}
if p.peerChoking && !p.peerAllowedFast.Contains(bitmap.BitIndex(req.Index)) {
return
}
var err error
more, err = p.request(req)
if err != nil {
panic(err)
}
})
if interested && len(p.actualRequestState.Requests) >= p.nominalMaxRequests() {
break
}
if !more {
break
} }
} }
}
for req := range next.Requests {
more, err := p.request(req)
if err != nil {
panic(err)
} /* else {
log.Print(req)
} */
if !more { if !more {
return false break
} }
} }
return true if !more {
return false
}
if !interested {
p.setInterested(false)
}
return more
} }