90 lines
2.6 KiB
Go
90 lines
2.6 KiB
Go
|
package request_strategy
|
||
|
|
||
|
import (
|
||
|
"bytes"
|
||
|
"expvar"
|
||
|
|
||
|
"github.com/anacrolix/multiless"
|
||
|
"github.com/anacrolix/torrent/metainfo"
|
||
|
"github.com/google/btree"
|
||
|
|
||
|
"github.com/anacrolix/torrent/types"
|
||
|
)
|
||
|
|
||
|
type (
|
||
|
RequestIndex = uint32
|
||
|
ChunkIndex = uint32
|
||
|
Request = types.Request
|
||
|
pieceIndex = types.PieceIndex
|
||
|
piecePriority = types.PiecePriority
|
||
|
// This can be made into a type-param later, will be great for testing.
|
||
|
ChunkSpec = types.ChunkSpec
|
||
|
)
|
||
|
|
||
|
func pieceOrderLess(i, j *pieceRequestOrderItem) multiless.Computation {
|
||
|
return multiless.New().Int(
|
||
|
int(j.state.Priority), int(i.state.Priority),
|
||
|
// TODO: Should we match on complete here to prevent churn when availability changes?
|
||
|
).Bool(
|
||
|
j.state.Partial, i.state.Partial,
|
||
|
).Int(
|
||
|
// If this is done with relative availability, do we lose some determinism? If completeness
|
||
|
// is used, would that push this far enough down?
|
||
|
i.state.Availability, j.state.Availability,
|
||
|
).Int(
|
||
|
i.key.Index, j.key.Index,
|
||
|
).Lazy(func() multiless.Computation {
|
||
|
return multiless.New().Cmp(bytes.Compare(
|
||
|
i.key.InfoHash[:],
|
||
|
j.key.InfoHash[:],
|
||
|
))
|
||
|
})
|
||
|
}
|
||
|
|
||
|
var packageExpvarMap = expvar.NewMap("request-strategy")
|
||
|
|
||
|
// Calls f with requestable pieces in order.
|
||
|
func GetRequestablePieces(input Input, pro *PieceRequestOrder, f func(ih metainfo.Hash, pieceIndex int)) {
|
||
|
// Storage capacity left for this run, keyed by the storage capacity pointer on the storage
|
||
|
// TorrentImpl. A nil value means no capacity limit.
|
||
|
var storageLeft *int64
|
||
|
if cap, ok := input.Capacity(); ok {
|
||
|
storageLeft = &cap
|
||
|
}
|
||
|
var allTorrentsUnverifiedBytes int64
|
||
|
pro.tree.Ascend(func(i btree.Item) bool {
|
||
|
_i := i.(*pieceRequestOrderItem)
|
||
|
ih := _i.key.InfoHash
|
||
|
var t Torrent = input.Torrent(ih)
|
||
|
var piece Piece = t.Piece(_i.key.Index)
|
||
|
pieceLength := t.PieceLength()
|
||
|
if storageLeft != nil {
|
||
|
if *storageLeft < pieceLength {
|
||
|
return false
|
||
|
}
|
||
|
*storageLeft -= pieceLength
|
||
|
}
|
||
|
if !piece.Request() || piece.NumPendingChunks() == 0 {
|
||
|
// TODO: Clarify exactly what is verified. Stuff that's being hashed should be
|
||
|
// considered unverified and hold up further requests.
|
||
|
return true
|
||
|
}
|
||
|
if input.MaxUnverifiedBytes() != 0 && allTorrentsUnverifiedBytes+pieceLength > input.MaxUnverifiedBytes() {
|
||
|
return true
|
||
|
}
|
||
|
allTorrentsUnverifiedBytes += pieceLength
|
||
|
f(ih, _i.key.Index)
|
||
|
return true
|
||
|
})
|
||
|
return
|
||
|
}
|
||
|
|
||
|
type Input interface {
|
||
|
Torrent(metainfo.Hash) Torrent
|
||
|
// Storage capacity, shared among all Torrents with the same storage.TorrentCapacity pointer in
|
||
|
// their storage.Torrent references.
|
||
|
Capacity() (cap int64, capped bool)
|
||
|
// Across all the Torrents. This might be partitioned by storage capacity key now.
|
||
|
MaxUnverifiedBytes() int64
|
||
|
}
|