]> Sergey Matveev's repositories - btrtrc.git/blobdiff - torrent.go
Configurable hashers amount per torrent (#867)
[btrtrc.git] / torrent.go
index a4ae4b48c4bb340a310edefa1a67a304038e34e6..8b4745218864d81bb19ffda48140aa245f4cc044 100644 (file)
@@ -9,35 +9,44 @@ import (
        "fmt"
        "io"
        "math/rand"
-       "net/http"
+       "net/netip"
        "net/url"
        "sort"
-       "sync"
+       "strings"
        "text/tabwriter"
        "time"
        "unsafe"
 
-       "github.com/anacrolix/torrent/common"
-       "github.com/anacrolix/torrent/segments"
-       "github.com/anacrolix/torrent/webseed"
-       "github.com/davecgh/go-spew/spew"
-       "github.com/pion/datachannel"
-
+       "github.com/RoaringBitmap/roaring"
+       "github.com/anacrolix/chansync"
+       "github.com/anacrolix/chansync/events"
        "github.com/anacrolix/dht/v2"
+       . "github.com/anacrolix/generics"
+       g "github.com/anacrolix/generics"
        "github.com/anacrolix/log"
-       "github.com/anacrolix/missinggo"
-       "github.com/anacrolix/missinggo/iter"
        "github.com/anacrolix/missinggo/perf"
-       "github.com/anacrolix/missinggo/pubsub"
        "github.com/anacrolix/missinggo/slices"
+       "github.com/anacrolix/missinggo/v2"
        "github.com/anacrolix/missinggo/v2/bitmap"
-       "github.com/anacrolix/missinggo/v2/prioritybitmap"
+       "github.com/anacrolix/missinggo/v2/pubsub"
+       "github.com/anacrolix/multiless"
+       "github.com/anacrolix/sync"
+       "github.com/pion/datachannel"
+       "golang.org/x/exp/maps"
 
        "github.com/anacrolix/torrent/bencode"
+       "github.com/anacrolix/torrent/common"
+       "github.com/anacrolix/torrent/internal/check"
+       "github.com/anacrolix/torrent/internal/nestedmaps"
        "github.com/anacrolix/torrent/metainfo"
        pp "github.com/anacrolix/torrent/peer_protocol"
+       utHolepunch "github.com/anacrolix/torrent/peer_protocol/ut-holepunch"
+       request_strategy "github.com/anacrolix/torrent/request-strategy"
+       "github.com/anacrolix/torrent/segments"
        "github.com/anacrolix/torrent/storage"
        "github.com/anacrolix/torrent/tracker"
+       typedRoaring "github.com/anacrolix/torrent/typed-roaring"
+       "github.com/anacrolix/torrent/webseed"
        "github.com/anacrolix/torrent/webtorrent"
 )
 
@@ -50,26 +59,27 @@ type Torrent struct {
        cl     *Client
        logger log.Logger
 
-       networkingEnabled      bool
-       dataDownloadDisallowed bool
+       networkingEnabled      chansync.Flag
+       dataDownloadDisallowed chansync.Flag
        dataUploadDisallowed   bool
        userOnWriteChunkErr    func(error)
 
-       // Determines what chunks to request from peers.
-       requestStrategy requestStrategy
-
-       closed   missinggo.Event
+       closed   chansync.SetOnce
+       onClose  []func()
        infoHash metainfo.Hash
        pieces   []Piece
+
+       // The order pieces are requested if there's no stronger reason like availability or priority.
+       pieceRequestOrder []int
        // Values are the piece indices that changed.
-       pieceStateChanges *pubsub.PubSub
+       pieceStateChanges pubsub.PubSub[PieceStateChange]
        // The size of chunks to request from peers over the wire. This is
        // normally 16KiB by convention these days.
        chunkSize pp.Integer
-       chunkPool *sync.Pool
+       chunkPool sync.Pool
        // Total length of the torrent in bytes. Stored because it's not O(1) to
        // get this from the info dict.
-       length *int64
+       _length Option[int64]
 
        // The storage to open when the info dict becomes available.
        storageOpener *storage.Client
@@ -86,23 +96,23 @@ type Torrent struct {
        fileIndex segments.Index
        files     *[]*File
 
-       webSeeds map[string]*Peer
+       _chunksPerRegularPiece chunkIndexType
 
+       webSeeds map[string]*Peer
        // Active peer connections, running message stream loops. TODO: Make this
        // open (not-closed) connections only.
        conns               map[*PeerConn]struct{}
        maxEstablishedConns int
        // Set of addrs to which we're attempting to connect. Connections are
        // half-open until all handshakes are completed.
-       halfOpen    map[string]PeerInfo
-       fastestPeer *Peer
+       halfOpen map[string]map[outgoingConnAttemptKey]*PeerInfo
 
        // Reserve of peers to connect to. A peer can be both here and in the
        // active connections if were told about the peer after connecting with
        // them. That encourages us to reconnect to peers that are well known in
        // the swarm.
        peers prioritizedPeers
-       // Whether we want to know to know more peers.
+       // Whether we want to know more peers.
        wantPeersEvent missinggo.Event
        // An announcer for each tracker URL.
        trackerAnnouncers map[string]torrentTrackerAnnouncer
@@ -123,8 +133,8 @@ type Torrent struct {
        metadataCompletedChunks []bool
        metadataChanged         sync.Cond
 
-       // Set when .Info is obtained.
-       gotMetainfo missinggo.Event
+       // Closed when .Info is obtained.
+       gotMetainfoC chan struct{}
 
        readers                map[*reader]struct{}
        _readerNowPieces       bitmap.Bitmap
@@ -132,62 +142,91 @@ type Torrent struct {
 
        // A cache of pieces we need to get. Calculated from various piece and
        // file priorities and completion states elsewhere.
-       _pendingPieces prioritybitmap.PriorityBitmap
+       _pendingPieces roaring.Bitmap
        // A cache of completed piece indices.
-       _completedPieces bitmap.Bitmap
+       _completedPieces roaring.Bitmap
        // Pieces that need to be hashed.
-       piecesQueuedForHash bitmap.Bitmap
-       activePieceHashes   int
+       piecesQueuedForHash       bitmap.Bitmap
+       activePieceHashes         int
+       initialPieceCheckDisabled bool
 
-       // A pool of piece priorities []int for assignment to new connections.
-       // These "inclinations" are used to give connections preference for
-       // different pieces.
-       connPieceInclinationPool sync.Pool
+       connsWithAllPieces map[*Peer]struct{}
 
-       // Count of each request across active connections.
-       pendingRequests map[Request]int
+       requestState map[RequestIndex]requestState
+       // Chunks we've written to since the corresponding piece was last checked.
+       dirtyChunks typedRoaring.Bitmap[RequestIndex]
 
        pex pexState
-}
 
-func (t *Torrent) numConns() int {
-       return len(t.conns)
+       // Is On when all pieces are complete.
+       Complete chansync.Flag
+
+       // Torrent sources in use keyed by the source string.
+       activeSources sync.Map
+       sourcesLogger log.Logger
+
+       smartBanCache smartBanCache
+
+       // Large allocations reused between request state updates.
+       requestPieceStates []request_strategy.PieceRequestOrderState
+       requestIndexes     []RequestIndex
 }
 
-func (t *Torrent) numReaders() int {
-       return len(t.readers)
+type outgoingConnAttemptKey = *PeerInfo
+
+func (t *Torrent) length() int64 {
+       return t._length.Value
 }
 
-func (t *Torrent) readerNowPieces() bitmap.Bitmap {
-       return t._readerNowPieces
+func (t *Torrent) selectivePieceAvailabilityFromPeers(i pieceIndex) (count int) {
+       // This could be done with roaring.BitSliceIndexing.
+       t.iterPeers(func(peer *Peer) {
+               if _, ok := t.connsWithAllPieces[peer]; ok {
+                       return
+               }
+               if peer.peerHasPiece(i) {
+                       count++
+               }
+       })
+       return
 }
 
-func (t *Torrent) readerReadaheadPieces() bitmap.Bitmap {
-       return t._readerReadaheadPieces
+func (t *Torrent) decPieceAvailability(i pieceIndex) {
+       if !t.haveInfo() {
+               return
+       }
+       p := t.piece(i)
+       if p.relativeAvailability <= 0 {
+               panic(p.relativeAvailability)
+       }
+       p.relativeAvailability--
+       t.updatePieceRequestOrder(i)
 }
 
-func (t *Torrent) ignorePieces() bitmap.Bitmap {
-       ret := t._completedPieces.Copy()
-       ret.Union(t.piecesQueuedForHash)
-       for i := 0; i < t.numPieces(); i++ {
-               if t.piece(i).hashing {
-                       ret.Set(i, true)
-               }
+func (t *Torrent) incPieceAvailability(i pieceIndex) {
+       // If we don't the info, this should be reconciled when we do.
+       if t.haveInfo() {
+               p := t.piece(i)
+               p.relativeAvailability++
+               t.updatePieceRequestOrder(i)
        }
-       return ret
 }
 
-func (t *Torrent) pendingPieces() *prioritybitmap.PriorityBitmap {
-       return &t._pendingPieces
+func (t *Torrent) readerNowPieces() bitmap.Bitmap {
+       return t._readerNowPieces
 }
 
-func (t *Torrent) tickleReaders() {
-       t.cl.event.Broadcast()
+func (t *Torrent) readerReadaheadPieces() bitmap.Bitmap {
+       return t._readerReadaheadPieces
+}
+
+func (t *Torrent) ignorePieceForRequests(i pieceIndex) bool {
+       return !t.wantPieceIndex(i)
 }
 
 // Returns a channel that is closed when the Torrent is closed.
-func (t *Torrent) Closed() <-chan struct{} {
-       return t.closed.LockedChan(t.cl.locker())
+func (t *Torrent) Closed() events.Done {
+       return t.closed.Done()
 }
 
 // KnownSwarm returns the known subset of the peers in the Torrent's swarm, including active,
@@ -199,13 +238,14 @@ func (t *Torrent) KnownSwarm() (ks []PeerInfo) {
        })
 
        // Add half-open peers to the list
-       for _, peer := range t.halfOpen {
-               ks = append(ks, peer)
+       for _, attempts := range t.halfOpen {
+               for _, peer := range attempts {
+                       ks = append(ks, *peer)
+               }
        }
 
        // Add active peers to the list
        for conn := range t.conns {
-
                ks = append(ks, PeerInfo{
                        Id:     conn.PeerID,
                        Addr:   conn.RemoteAddr,
@@ -225,7 +265,7 @@ func (t *Torrent) KnownSwarm() (ks []PeerInfo) {
 
 func (t *Torrent) setChunkSize(size pp.Integer) {
        t.chunkSize = size
-       t.chunkPool = &sync.Pool{
+       t.chunkPool = sync.Pool{
                New: func() interface{} {
                        b := make([]byte, size)
                        return &b
@@ -234,10 +274,13 @@ func (t *Torrent) setChunkSize(size pp.Integer) {
 }
 
 func (t *Torrent) pieceComplete(piece pieceIndex) bool {
-       return t._completedPieces.Get(bitmap.BitIndex(piece))
+       return t._completedPieces.Contains(bitmap.BitIndex(piece))
 }
 
 func (t *Torrent) pieceCompleteUncached(piece pieceIndex) storage.Completion {
+       if t.storage == nil {
+               return storage.Completion{Complete: false, Ok: true}
+       }
        return t.pieces[piece].Storage().Completion()
 }
 
@@ -255,14 +298,19 @@ func (t *Torrent) addrActive(addr string) bool {
        return false
 }
 
-func (t *Torrent) unclosedConnsAsSlice() (ret []*PeerConn) {
-       ret = make([]*PeerConn, 0, len(t.conns))
+func (t *Torrent) appendUnclosedConns(ret []*PeerConn) []*PeerConn {
+       return t.appendConns(ret, func(conn *PeerConn) bool {
+               return !conn.closed.IsSet()
+       })
+}
+
+func (t *Torrent) appendConns(ret []*PeerConn, f func(*PeerConn) bool) []*PeerConn {
        for c := range t.conns {
-               if !c.closed.IsSet() {
+               if f(c) {
                        ret = append(ret, c)
                }
        }
-       return
+       return ret
 }
 
 func (t *Torrent) addPeer(p PeerInfo) (added bool) {
@@ -298,10 +346,11 @@ func (t *Torrent) addPeer(p PeerInfo) (added bool) {
 }
 
 func (t *Torrent) invalidateMetadata() {
-       for i := range t.metadataCompletedChunks {
+       for i := 0; i < len(t.metadataCompletedChunks); i++ {
                t.metadataCompletedChunks[i] = false
        }
        t.nameMu.Lock()
+       t.gotMetainfoC = make(chan struct{})
        t.info = nil
        t.nameMu.Unlock()
 }
@@ -384,9 +433,11 @@ func (t *Torrent) cacheLength() {
        for _, f := range t.info.UpvertedFiles() {
                l += f.Length
        }
-       t.length = &l
+       t._length = Some(l)
 }
 
+// TODO: This shouldn't fail for storage reasons. Instead we should handle storage failure
+// separately.
 func (t *Torrent) setInfo(info *metainfo.Info) error {
        if err := validateInfo(info); err != nil {
                return fmt.Errorf("bad info: %s", err)
@@ -401,6 +452,8 @@ func (t *Torrent) setInfo(info *metainfo.Info) error {
        t.nameMu.Lock()
        t.info = info
        t.nameMu.Unlock()
+       t._chunksPerRegularPiece = chunkIndexType((pp.Integer(t.usualPieceSize()) + t.chunkSize - 1) / t.chunkSize)
+       t.updateComplete()
        t.fileIndex = segments.NewIndex(common.LengthIterFromUpvertedFiles(info.UpvertedFiles()))
        t.displayName = "" // Save a few bytes lol.
        t.initFiles()
@@ -409,28 +462,46 @@ func (t *Torrent) setInfo(info *metainfo.Info) error {
        return nil
 }
 
+func (t *Torrent) pieceRequestOrderKey(i int) request_strategy.PieceRequestOrderKey {
+       return request_strategy.PieceRequestOrderKey{
+               InfoHash: t.infoHash,
+               Index:    i,
+       }
+}
+
 // This seems to be all the follow-up tasks after info is set, that can't fail.
 func (t *Torrent) onSetInfo() {
-       t.iterPeers(func(p *Peer) {
-               p.onGotInfo(t.info)
-       })
+       t.pieceRequestOrder = rand.Perm(t.numPieces())
+       t.initPieceRequestOrder()
+       MakeSliceWithLength(&t.requestPieceStates, t.numPieces())
        for i := range t.pieces {
-               t.updatePieceCompletion(pieceIndex(i))
                p := &t.pieces[i]
-               if !p.storageCompletionOk {
+               // Need to add relativeAvailability before updating piece completion, as that may result in conns
+               // being dropped.
+               if p.relativeAvailability != 0 {
+                       panic(p.relativeAvailability)
+               }
+               p.relativeAvailability = t.selectivePieceAvailabilityFromPeers(i)
+               t.addRequestOrderPiece(i)
+               t.updatePieceCompletion(i)
+               if !t.initialPieceCheckDisabled && !p.storageCompletionOk {
                        // t.logger.Printf("piece %s completion unknown, queueing check", p)
-                       t.queuePieceCheck(pieceIndex(i))
+                       t.queuePieceCheck(i)
                }
        }
        t.cl.event.Broadcast()
-       t.gotMetainfo.Set()
+       close(t.gotMetainfoC)
        t.updateWantPeersEvent()
-       t.pendingRequests = make(map[Request]int)
+       t.requestState = make(map[RequestIndex]requestState)
        t.tryCreateMorePieceHashers()
+       t.iterPeers(func(p *Peer) {
+               p.onGotInfo(t.info)
+               p.updateRequests("onSetInfo")
+       })
 }
 
 // Called when metadata for a torrent becomes available.
-func (t *Torrent) setInfoBytes(b []byte) error {
+func (t *Torrent) setInfoBytesLocked(b []byte) error {
        if metainfo.HashBytes(b) != t.infoHash {
                return errors.New("info bytes have wrong hash")
        }
@@ -466,19 +537,19 @@ func (t *Torrent) haveAllMetadataPieces() bool {
 }
 
 // TODO: Propagate errors to disconnect peer.
-func (t *Torrent) setMetadataSize(bytes int) (err error) {
+func (t *Torrent) setMetadataSize(size int) (err error) {
        if t.haveInfo() {
                // We already know the correct metadata size.
                return
        }
-       if bytes <= 0 || bytes > 10000000 { // 10MB, pulled from my ass.
-               return errors.New("bad size")
+       if uint32(size) > maxMetadataSize {
+               return log.WithLevel(log.Warning, errors.New("bad size"))
        }
-       if t.metadataBytes != nil && len(t.metadataBytes) == int(bytes) {
+       if len(t.metadataBytes) == size {
                return
        }
-       t.metadataBytes = make([]byte, bytes)
-       t.metadataCompletedChunks = make([]bool, (bytes+(1<<14)-1)/(1<<14))
+       t.metadataBytes = make([]byte, size)
+       t.metadataCompletedChunks = make([]bool, (size+(1<<14)-1)/(1<<14))
        t.metadataChanged.Broadcast()
        for c := range t.conns {
                c.requestPendingMetadata()
@@ -492,9 +563,12 @@ func (t *Torrent) name() string {
        t.nameMu.RLock()
        defer t.nameMu.RUnlock()
        if t.haveInfo() {
-               return t.info.Name
+               return t.info.BestName()
        }
-       return t.displayName
+       if t.displayName != "" {
+               return t.displayName
+       }
+       return "infohash:" + t.infoHash.HexString()
 }
 
 func (t *Torrent) pieceState(index pieceIndex) (ret PieceState) {
@@ -515,20 +589,44 @@ func (t *Torrent) metadataPieceSize(piece int) int {
        return metadataPieceSize(len(t.metadataBytes), piece)
 }
 
-func (t *Torrent) newMetadataExtensionMessage(c *PeerConn, msgType int, piece int, data []byte) pp.Message {
-       d := map[string]int{
-               "msg_type": msgType,
-               "piece":    piece,
+func (t *Torrent) newMetadataExtensionMessage(c *PeerConn, msgType pp.ExtendedMetadataRequestMsgType, piece int, data []byte) pp.Message {
+       return pp.Message{
+               Type:       pp.Extended,
+               ExtendedID: c.PeerExtensionIDs[pp.ExtensionNameMetadata],
+               ExtendedPayload: append(bencode.MustMarshal(pp.ExtendedMetadataRequestMsg{
+                       Piece:     piece,
+                       TotalSize: len(t.metadataBytes),
+                       Type:      msgType,
+               }), data...),
        }
-       if data != nil {
-               d["total_size"] = len(t.metadataBytes)
+}
+
+type pieceAvailabilityRun struct {
+       Count        pieceIndex
+       Availability int
+}
+
+func (me pieceAvailabilityRun) String() string {
+       return fmt.Sprintf("%v(%v)", me.Count, me.Availability)
+}
+
+func (t *Torrent) pieceAvailabilityRuns() (ret []pieceAvailabilityRun) {
+       rle := missinggo.NewRunLengthEncoder(func(el interface{}, count uint64) {
+               ret = append(ret, pieceAvailabilityRun{Availability: el.(int), Count: int(count)})
+       })
+       for i := range t.pieces {
+               rle.Append(t.pieces[i].availability(), 1)
        }
-       p := bencode.MustMarshal(d)
-       return pp.Message{
-               Type:            pp.Extended,
-               ExtendedID:      c.PeerExtensionIDs[pp.ExtensionNameMetadata],
-               ExtendedPayload: append(p, data...),
+       rle.Flush()
+       return
+}
+
+func (t *Torrent) pieceAvailabilityFrequencies() (freqs []int) {
+       freqs = make([]int, t.numActivePeers()+1)
+       for i := range t.pieces {
+               freqs[t.piece(i).availability()]++
        }
+       return
 }
 
 func (t *Torrent) pieceStateRuns() (ret PieceStateRuns) {
@@ -601,17 +699,41 @@ func (t *Torrent) writeStatus(w io.Writer) {
                }
                fmt.Fprintln(w)
        }
-       fmt.Fprintf(w, "Piece length: %s\n", func() string {
-               if t.haveInfo() {
-                       return fmt.Sprint(t.usualPieceSize())
-               } else {
-                       return "?"
-               }
-       }())
+       fmt.Fprintf(w, "Piece length: %s\n",
+               func() string {
+                       if t.haveInfo() {
+                               return fmt.Sprintf("%v (%v chunks)",
+                                       t.usualPieceSize(),
+                                       float64(t.usualPieceSize())/float64(t.chunkSize))
+                       } else {
+                               return "no info"
+                       }
+               }(),
+       )
        if t.info != nil {
                fmt.Fprintf(w, "Num Pieces: %d (%d completed)\n", t.numPieces(), t.numPiecesCompleted())
-               fmt.Fprintf(w, "Piece States: %s", t.pieceStateRuns())
-               fmt.Fprintln(w)
+               fmt.Fprintf(w, "Piece States: %s\n", t.pieceStateRuns())
+               // Generates a huge, unhelpful listing when piece availability is very scattered. Prefer
+               // availability frequencies instead.
+               if false {
+                       fmt.Fprintf(w, "Piece availability: %v\n", strings.Join(func() (ret []string) {
+                               for _, run := range t.pieceAvailabilityRuns() {
+                                       ret = append(ret, run.String())
+                               }
+                               return
+                       }(), " "))
+               }
+               fmt.Fprintf(w, "Piece availability frequency: %v\n", strings.Join(
+                       func() (ret []string) {
+                               for avail, freq := range t.pieceAvailabilityFrequencies() {
+                                       if freq == 0 {
+                                               continue
+                                       }
+                                       ret = append(ret, fmt.Sprintf("%v: %v", avail, freq))
+                               }
+                               return
+                       }(),
+                       ", "))
        }
        fmt.Fprintf(w, "Reader Pieces:")
        t.forReaderOffsetPieces(func(begin, end pieceIndex) (again bool) {
@@ -642,16 +764,40 @@ func (t *Torrent) writeStatus(w io.Writer) {
 
        fmt.Fprintf(w, "DHT Announces: %d\n", t.numDHTAnnounces)
 
-       spew.NewDefaultConfig()
-       spew.Fdump(w, t.statsLocked())
-
-       peers := t.peersAsSlice()
-       sort.Slice(peers, func(i, j int) bool {
-               return worseConn(peers[i], peers[j])
+       dumpStats(w, t.statsLocked())
+
+       fmt.Fprintf(w, "webseeds:\n")
+       t.writePeerStatuses(w, maps.Values(t.webSeeds))
+
+       peerConns := maps.Keys(t.conns)
+       // Peers without priorities first, then those with. I'm undecided about how to order peers
+       // without priorities.
+       sort.Slice(peerConns, func(li, ri int) bool {
+               l := peerConns[li]
+               r := peerConns[ri]
+               ml := multiless.New()
+               lpp := g.ResultFromTuple(l.peerPriority()).ToOption()
+               rpp := g.ResultFromTuple(r.peerPriority()).ToOption()
+               ml = ml.Bool(lpp.Ok, rpp.Ok)
+               ml = ml.Uint32(rpp.Value, lpp.Value)
+               return ml.Less()
        })
-       for i, c := range peers {
-               fmt.Fprintf(w, "%2d. ", i+1)
-               c.writeStatus(w, t)
+
+       fmt.Fprintf(w, "%v peer conns:\n", len(peerConns))
+       t.writePeerStatuses(w, g.SliceMap(peerConns, func(pc *PeerConn) *Peer {
+               return &pc.Peer
+       }))
+}
+
+func (t *Torrent) writePeerStatuses(w io.Writer, peers []*Peer) {
+       var buf bytes.Buffer
+       for _, c := range peers {
+               fmt.Fprintf(w, "- ")
+               buf.Reset()
+               c.writeStatus(&buf)
+               w.Write(bytes.TrimRight(
+                       bytes.ReplaceAll(buf.Bytes(), []byte("\n"), []byte("\n  ")),
+                       " "))
        }
 }
 
@@ -674,22 +820,38 @@ func (t *Torrent) newMetaInfo() metainfo.MetaInfo {
                                return nil
                        }
                }(),
+               UrlList: func() []string {
+                       ret := make([]string, 0, len(t.webSeeds))
+                       for url := range t.webSeeds {
+                               ret = append(ret, url)
+                       }
+                       return ret
+               }(),
        }
 }
 
-func (t *Torrent) BytesMissing() int64 {
+// Returns a count of bytes that are not complete in storage, and not pending being written to
+// storage. This value is from the perspective of the download manager, and may not agree with the
+// actual state in storage. If you want read data synchronously you should use a Reader. See
+// https://github.com/anacrolix/torrent/issues/828.
+func (t *Torrent) BytesMissing() (n int64) {
        t.cl.rLock()
-       defer t.cl.rUnlock()
-       return t.bytesMissingLocked()
+       n = t.bytesMissingLocked()
+       t.cl.rUnlock()
+       return
 }
 
 func (t *Torrent) bytesMissingLocked() int64 {
        return t.bytesLeft()
 }
 
+func iterFlipped(b *roaring.Bitmap, end uint64, cb func(uint32) bool) {
+       roaring.Flip(b, 0, end).Iterate(cb)
+}
+
 func (t *Torrent) bytesLeft() (left int64) {
-       bitmap.Flip(t._completedPieces, 0, bitmap.BitIndex(t.numPieces())).IterTyped(func(piece int) bool {
-               p := &t.pieces[piece]
+       iterFlipped(&t._completedPieces, uint64(t.numPieces()), func(x uint32) bool {
+               p := t.piece(pieceIndex(x))
                left += int64(p.length() - p.numDirtyBytes())
                return true
        })
@@ -720,24 +882,42 @@ func (t *Torrent) usualPieceSize() int {
 }
 
 func (t *Torrent) numPieces() pieceIndex {
-       return pieceIndex(t.info.NumPieces())
+       return t.info.NumPieces()
 }
 
-func (t *Torrent) numPiecesCompleted() (num int) {
-       return t._completedPieces.Len()
+func (t *Torrent) numPiecesCompleted() (num pieceIndex) {
+       return pieceIndex(t._completedPieces.GetCardinality())
 }
 
-func (t *Torrent) close() (err error) {
-       t.closed.Set()
-       t.tickleReaders()
+func (t *Torrent) close(wg *sync.WaitGroup) (err error) {
+       if !t.closed.Set() {
+               err = errors.New("already closed")
+               return
+       }
+       for _, f := range t.onClose {
+               f()
+       }
        if t.storage != nil {
-               t.storageLock.Lock()
-               t.storage.Close()
-               t.storageLock.Unlock()
+               wg.Add(1)
+               go func() {
+                       defer wg.Done()
+                       t.storageLock.Lock()
+                       defer t.storageLock.Unlock()
+                       if f := t.storage.Close; f != nil {
+                               err1 := f()
+                               if err1 != nil {
+                                       t.logger.WithDefaultLevel(log.Warning).Printf("error closing storage: %v", err1)
+                               }
+                       }
+               }()
        }
-       for conn := range t.conns {
-               conn.close()
+       t.iterPeers(func(p *Peer) {
+               p.close()
+       })
+       if t.storage != nil {
+               t.deletePieceRequestOrder()
        }
+       t.assertAllPiecesRelativeAvailabilityZero()
        t.pex.Reset()
        t.cl.event.Broadcast()
        t.pieceStateChanges.Close()
@@ -745,14 +925,23 @@ func (t *Torrent) close() (err error) {
        return
 }
 
+func (t *Torrent) assertAllPiecesRelativeAvailabilityZero() {
+       for i := range t.pieces {
+               p := t.piece(i)
+               if p.relativeAvailability != 0 {
+                       panic(fmt.Sprintf("piece %v has relative availability %v", i, p.relativeAvailability))
+               }
+       }
+}
+
 func (t *Torrent) requestOffset(r Request) int64 {
-       return torrentRequestOffset(*t.length, int64(t.usualPieceSize()), r)
+       return torrentRequestOffset(t.length(), int64(t.usualPieceSize()), r)
 }
 
 // Return the request that would include the given offset into the torrent data. Returns !ok if
 // there is no such request.
 func (t *Torrent) offsetRequest(off int64) (req Request, ok bool) {
-       return torrentOffsetRequest(*t.length, t.info.PieceLength, int64(t.chunkSize), off)
+       return torrentOffsetRequest(t.length(), t.info.PieceLength, int64(t.chunkSize), off)
 }
 
 func (t *Torrent) writeChunk(piece int, begin int64, data []byte) (err error) {
@@ -766,19 +955,32 @@ func (t *Torrent) writeChunk(piece int, begin int64, data []byte) (err error) {
 
 func (t *Torrent) bitfield() (bf []bool) {
        bf = make([]bool, t.numPieces())
-       t._completedPieces.IterTyped(func(piece int) (again bool) {
+       t._completedPieces.Iterate(func(piece uint32) (again bool) {
                bf[piece] = true
                return true
        })
        return
 }
 
-func (t *Torrent) pieceNumChunks(piece pieceIndex) pp.Integer {
-       return (t.pieceLength(piece) + t.chunkSize - 1) / t.chunkSize
+func (t *Torrent) pieceNumChunks(piece pieceIndex) chunkIndexType {
+       return chunkIndexType((t.pieceLength(piece) + t.chunkSize - 1) / t.chunkSize)
+}
+
+func (t *Torrent) chunksPerRegularPiece() chunkIndexType {
+       return t._chunksPerRegularPiece
+}
+
+func (t *Torrent) numChunks() RequestIndex {
+       if t.numPieces() == 0 {
+               return 0
+       }
+       return RequestIndex(t.numPieces()-1)*t.chunksPerRegularPiece() + t.pieceNumChunks(t.numPieces()-1)
 }
 
 func (t *Torrent) pendAllChunkSpecs(pieceIndex pieceIndex) {
-       t.pieces[pieceIndex]._dirtyChunks.Clear()
+       t.dirtyChunks.RemoveRange(
+               uint64(t.pieceRequestIndexOffset(pieceIndex)),
+               uint64(t.pieceRequestIndexOffset(pieceIndex+1)))
 }
 
 func (t *Torrent) pieceLength(piece pieceIndex) pp.Integer {
@@ -787,7 +989,7 @@ func (t *Torrent) pieceLength(piece pieceIndex) pp.Integer {
                return 0
        }
        if piece == t.numPieces()-1 {
-               ret := pp.Integer(*t.length % t.info.PieceLength)
+               ret := pp.Integer(t.length() % t.info.PieceLength)
                if ret != 0 {
                        return ret
                }
@@ -795,32 +997,60 @@ func (t *Torrent) pieceLength(piece pieceIndex) pp.Integer {
        return pp.Integer(t.info.PieceLength)
 }
 
-func (t *Torrent) hashPiece(piece pieceIndex) (ret metainfo.Hash, err error) {
-       hash := pieceHash.New()
+func (t *Torrent) smartBanBlockCheckingWriter(piece pieceIndex) *blockCheckingWriter {
+       return &blockCheckingWriter{
+               cache:        &t.smartBanCache,
+               requestIndex: t.pieceRequestIndexOffset(piece),
+               chunkSize:    t.chunkSize.Int(),
+       }
+}
+
+func (t *Torrent) hashPiece(piece pieceIndex) (
+       ret metainfo.Hash,
+       // These are peers that sent us blocks that differ from what we hash here.
+       differingPeers map[bannableAddr]struct{},
+       err error,
+) {
        p := t.piece(piece)
        p.waitNoPendingWrites()
        storagePiece := t.pieces[piece].Storage()
+
+       // Does the backend want to do its own hashing?
+       if i, ok := storagePiece.PieceImpl.(storage.SelfHashing); ok {
+               var sum metainfo.Hash
+               // log.Printf("A piece decided to self-hash: %d", piece)
+               sum, err = i.SelfHash()
+               missinggo.CopyExact(&ret, sum)
+               return
+       }
+
+       hash := pieceHash.New()
        const logPieceContents = false
+       smartBanWriter := t.smartBanBlockCheckingWriter(piece)
+       writers := []io.Writer{hash, smartBanWriter}
+       var examineBuf bytes.Buffer
        if logPieceContents {
-               var examineBuf bytes.Buffer
-               _, err = storagePiece.WriteTo(io.MultiWriter(hash, &examineBuf))
-               log.Printf("hashed %q with copy err %v", examineBuf.Bytes(), err)
-       } else {
-               _, err = storagePiece.WriteTo(hash)
+               writers = append(writers, &examineBuf)
+       }
+       _, err = storagePiece.WriteTo(io.MultiWriter(writers...))
+       if logPieceContents {
+               t.logger.WithDefaultLevel(log.Debug).Printf("hashed %q with copy err %v", examineBuf.Bytes(), err)
        }
+       smartBanWriter.Flush()
+       differingPeers = smartBanWriter.badPeers
        missinggo.CopyExact(&ret, hash.Sum(nil))
        return
 }
 
 func (t *Torrent) haveAnyPieces() bool {
-       return t._completedPieces.Len() != 0
+       return !t._completedPieces.IsEmpty()
 }
 
 func (t *Torrent) haveAllPieces() bool {
        if !t.haveInfo() {
                return false
        }
-       return t._completedPieces.Len() == bitmap.BitIndex(t.numPieces())
+       return t._completedPieces.GetCardinality() == bitmap.BitRange(t.numPieces())
 }
 
 func (t *Torrent) havePiece(index pieceIndex) bool {
@@ -828,8 +1058,9 @@ func (t *Torrent) havePiece(index pieceIndex) bool {
 }
 
 func (t *Torrent) maybeDropMutuallyCompletePeer(
-       // I'm not sure about taking peer here, not all peer implementations actually drop. Maybe that's okay?
-       p *Peer,
+       // I'm not sure about taking peer here, not all peer implementations actually drop. Maybe that's
+       // okay?
+       p *PeerConn,
 ) {
        if !t.cl.config.DropMutuallyCompletePeers {
                return
@@ -843,7 +1074,7 @@ func (t *Torrent) maybeDropMutuallyCompletePeer(
        if p.useful() {
                return
        }
-       t.logger.WithDefaultLevel(log.Debug).Printf("dropping %v, which is mutually complete", p)
+       p.logger.Levelf(log.Debug, "is mutually complete; dropping")
        p.drop()
 }
 
@@ -861,44 +1092,47 @@ func (t *Torrent) haveChunk(r Request) (ret bool) {
        return !p.pendingChunk(r.ChunkSpec, t.chunkSize)
 }
 
-func chunkIndex(cs ChunkSpec, chunkSize pp.Integer) int {
-       return int(cs.Begin / chunkSize)
+func chunkIndexFromChunkSpec(cs ChunkSpec, chunkSize pp.Integer) chunkIndexType {
+       return chunkIndexType(cs.Begin / chunkSize)
 }
 
 func (t *Torrent) wantPieceIndex(index pieceIndex) bool {
-       if !t.haveInfo() {
-               return false
-       }
-       if index < 0 || index >= t.numPieces() {
-               return false
-       }
-       p := &t.pieces[index]
-       if p.queuedForHash() {
-               return false
-       }
-       if p.hashing {
-               return false
-       }
-       if t.pieceComplete(index) {
-               return false
-       }
-       if t._pendingPieces.Contains(bitmap.BitIndex(index)) {
-               return true
+       return t._pendingPieces.Contains(uint32(index))
+}
+
+// A pool of []*PeerConn, to reduce allocations in functions that need to index or sort Torrent
+// conns (which is a map).
+var peerConnSlices sync.Pool
+
+func getPeerConnSlice(cap int) []*PeerConn {
+       getInterface := peerConnSlices.Get()
+       if getInterface == nil {
+               return make([]*PeerConn, 0, cap)
+       } else {
+               return getInterface.([]*PeerConn)[:0]
        }
-       // t.logger.Printf("piece %d not pending", index)
-       return !t.forReaderOffsetPieces(func(begin, end pieceIndex) bool {
-               return index < begin || index >= end
-       })
 }
 
-// The worst connection is one that hasn't been sent, or sent anything useful for the longest. A bad
-// connection is one that usually sends us unwanted pieces, or has been in worser half of the
-// established connections for more than a minute.
-func (t *Torrent) worstBadConn() *PeerConn {
-       wcs := worseConnSlice{t.unclosedConnsAsSlice()}
+// Calls the given function with a slice of unclosed conns. It uses a pool to reduce allocations as
+// this is a frequent occurrence.
+func (t *Torrent) withUnclosedConns(f func([]*PeerConn)) {
+       sl := t.appendUnclosedConns(getPeerConnSlice(len(t.conns)))
+       f(sl)
+       peerConnSlices.Put(sl)
+}
+
+func (t *Torrent) worstBadConnFromSlice(opts worseConnLensOpts, sl []*PeerConn) *PeerConn {
+       wcs := worseConnSlice{conns: sl}
+       wcs.initKeys(opts)
        heap.Init(&wcs)
        for wcs.Len() != 0 {
                c := heap.Pop(&wcs).(*PeerConn)
+               if opts.incomingIsBad && !c.outgoing {
+                       return c
+               }
+               if opts.outgoingIsBad && c.outgoing {
+                       return c
+               }
                if c._stats.ChunksReadWasted.Int64() >= 6 && c._stats.ChunksReadWasted.Int64() > c._stats.ChunksReadUseful.Int64() {
                        return c
                }
@@ -914,6 +1148,17 @@ func (t *Torrent) worstBadConn() *PeerConn {
        return nil
 }
 
+// The worst connection is one that hasn't been sent, or sent anything useful for the longest. A bad
+// connection is one that usually sends us unwanted pieces, or has been in the worse half of the
+// established connections for more than a minute. This is O(n log n). If there was a way to not
+// consider the position of a conn relative to the total number, it could be reduced to O(n).
+func (t *Torrent) worstBadConn(opts worseConnLensOpts) (ret *PeerConn) {
+       t.withUnclosedConns(func(ucs []*PeerConn) {
+               ret = t.worstBadConnFromSlice(opts, ucs)
+       })
+       return
+}
+
 type PieceStateChange struct {
        Index int
        PieceState
@@ -937,16 +1182,16 @@ func (t *Torrent) pieceNumPendingChunks(piece pieceIndex) pp.Integer {
        if t.pieceComplete(piece) {
                return 0
        }
-       return t.pieceNumChunks(piece) - t.pieces[piece].numDirtyChunks()
+       return pp.Integer(t.pieceNumChunks(piece) - t.pieces[piece].numDirtyChunks())
 }
 
 func (t *Torrent) pieceAllDirty(piece pieceIndex) bool {
-       return t.pieces[piece]._dirtyChunks.Len() == int(t.pieceNumChunks(piece))
+       return t.pieces[piece].allChunksDirty()
 }
 
 func (t *Torrent) readersChanged() {
        t.updateReaderPieces()
-       t.updateAllPiecePriorities()
+       t.updateAllPiecePriorities("Torrent.readersChanged")
 }
 
 func (t *Torrent) updateReaderPieces() {
@@ -965,15 +1210,15 @@ func (t *Torrent) readerPosChanged(from, to pieceRange) {
        }
        if l.end < h.begin {
                // Two distinct ranges.
-               t.updatePiecePriorities(l.begin, l.end)
-               t.updatePiecePriorities(h.begin, h.end)
+               t.updatePiecePriorities(l.begin, l.end, "Torrent.readerPosChanged")
+               t.updatePiecePriorities(h.begin, h.end, "Torrent.readerPosChanged")
        } else {
                // Ranges overlap.
                end := l.end
                if h.end > end {
                        end = h.end
                }
-               t.updatePiecePriorities(l.begin, end)
+               t.updatePiecePriorities(l.begin, end, "Torrent.readerPosChanged")
        }
 }
 
@@ -983,49 +1228,64 @@ func (t *Torrent) maybeNewConns() {
        t.openNewConns()
 }
 
-func (t *Torrent) piecePriorityChanged(piece pieceIndex) {
-       // t.logger.Printf("piece %d priority changed", piece)
-       t.iterPeers(func(c *Peer) {
-               if c.updatePiecePriority(piece) {
-                       // log.Print("conn piece priority changed")
-                       c.updateRequests()
-               }
-       })
+func (t *Torrent) piecePriorityChanged(piece pieceIndex, reason string) {
+       if t._pendingPieces.Contains(uint32(piece)) {
+               t.iterPeers(func(c *Peer) {
+                       // if c.requestState.Interested {
+                       //      return
+                       // }
+                       if !c.isLowOnRequests() {
+                               return
+                       }
+                       if !c.peerHasPiece(piece) {
+                               return
+                       }
+                       if c.requestState.Interested && c.peerChoking && !c.peerAllowedFast.Contains(piece) {
+                               return
+                       }
+                       c.updateRequests(reason)
+               })
+       }
        t.maybeNewConns()
        t.publishPieceChange(piece)
 }
 
-func (t *Torrent) updatePiecePriority(piece pieceIndex) {
+func (t *Torrent) updatePiecePriority(piece pieceIndex, reason string) {
+       if !t.closed.IsSet() {
+               // It would be possible to filter on pure-priority changes here to avoid churning the piece
+               // request order.
+               t.updatePieceRequestOrder(piece)
+       }
        p := &t.pieces[piece]
        newPrio := p.uncachedPriority()
        // t.logger.Printf("torrent %p: piece %d: uncached priority: %v", t, piece, newPrio)
        if newPrio == PiecePriorityNone {
-               if !t._pendingPieces.Remove(bitmap.BitIndex(piece)) {
+               if !t._pendingPieces.CheckedRemove(uint32(piece)) {
                        return
                }
        } else {
-               if !t._pendingPieces.Set(bitmap.BitIndex(piece), newPrio.BitmapPriority()) {
+               if !t._pendingPieces.CheckedAdd(uint32(piece)) {
                        return
                }
        }
-       t.piecePriorityChanged(piece)
+       t.piecePriorityChanged(piece, reason)
 }
 
-func (t *Torrent) updateAllPiecePriorities() {
-       t.updatePiecePriorities(0, t.numPieces())
+func (t *Torrent) updateAllPiecePriorities(reason string) {
+       t.updatePiecePriorities(0, t.numPieces(), reason)
 }
 
 // Update all piece priorities in one hit. This function should have the same
 // output as updatePiecePriority, but across all pieces.
-func (t *Torrent) updatePiecePriorities(begin, end pieceIndex) {
+func (t *Torrent) updatePiecePriorities(begin, end pieceIndex, reason string) {
        for i := begin; i < end; i++ {
-               t.updatePiecePriority(i)
+               t.updatePiecePriority(i, reason)
        }
 }
 
 // Returns the range of pieces [begin, end) that contains the extent of bytes.
 func (t *Torrent) byteRegionPieces(off, size int64) (begin, end pieceIndex) {
-       if off >= *t.length {
+       if off >= t.length() {
                return
        }
        if off < 0 {
@@ -1043,9 +1303,9 @@ func (t *Torrent) byteRegionPieces(off, size int64) (begin, end pieceIndex) {
        return
 }
 
-// Returns true if all iterations complete without breaking. Returns the read
-// regions for all readers. The reader regions should not be merged as some
-// callers depend on this method to enumerate readers.
+// Returns true if all iterations complete without breaking. Returns the read regions for all
+// readers. The reader regions should not be merged as some callers depend on this method to
+// enumerate readers.
 func (t *Torrent) forReaderOffsetPieces(f func(begin, end pieceIndex) (more bool)) (all bool) {
        for r := range t.readers {
                p := r.pieces
@@ -1060,34 +1320,21 @@ func (t *Torrent) forReaderOffsetPieces(f func(begin, end pieceIndex) (more bool
 }
 
 func (t *Torrent) piecePriority(piece pieceIndex) piecePriority {
-       prio, ok := t._pendingPieces.GetPriority(bitmap.BitIndex(piece))
-       if !ok {
-               return PiecePriorityNone
-       }
-       if prio > 0 {
-               panic(prio)
-       }
-       ret := piecePriority(-prio)
-       if ret == PiecePriorityNone {
-               panic(piece)
-       }
-       return ret
+       return t.piece(piece).uncachedPriority()
 }
 
-func (t *Torrent) pendRequest(req Request) {
-       ci := chunkIndex(req.ChunkSpec, t.chunkSize)
-       t.pieces[req.Index].pendChunkIndex(ci)
+func (t *Torrent) pendRequest(req RequestIndex) {
+       t.piece(t.pieceIndexOfRequestIndex(req)).pendChunkIndex(req % t.chunksPerRegularPiece())
 }
 
-func (t *Torrent) pieceCompletionChanged(piece pieceIndex) {
-       t.tickleReaders()
+func (t *Torrent) pieceCompletionChanged(piece pieceIndex, reason string) {
        t.cl.event.Broadcast()
        if t.pieceComplete(piece) {
                t.onPieceCompleted(piece)
        } else {
                t.onIncompletePiece(piece)
        }
-       t.updatePiecePriority(piece)
+       t.updatePiecePriority(piece, reason)
 }
 
 func (t *Torrent) numReceivedConns() (ret int) {
@@ -1099,6 +1346,15 @@ func (t *Torrent) numReceivedConns() (ret int) {
        return
 }
 
+func (t *Torrent) numOutgoingConns() (ret int) {
+       for c := range t.conns {
+               if c.outgoing {
+                       ret++
+               }
+       }
+       return
+}
+
 func (t *Torrent) maxHalfOpen() int {
        // Note that if we somehow exceed the maximum established conns, we want
        // the negative value to have an effect.
@@ -1106,13 +1362,16 @@ func (t *Torrent) maxHalfOpen() int {
        extraIncoming := int64(t.numReceivedConns() - t.maxEstablishedConns/2)
        // We want to allow some experimentation with new peers, and to try to
        // upset an oversupply of received connections.
-       return int(min(max(5, extraIncoming)+establishedHeadroom, int64(t.cl.config.HalfOpenConnsPerTorrent)))
+       return int(min(
+               max(5, extraIncoming)+establishedHeadroom,
+               int64(t.cl.config.HalfOpenConnsPerTorrent),
+       ))
 }
 
 func (t *Torrent) openNewConns() (initiated int) {
        defer t.updateWantPeersEvent()
        for t.peers.Len() != 0 {
-               if !t.wantConns() {
+               if !t.wantOutgoingConns() {
                        return
                }
                if len(t.halfOpen) >= t.maxHalfOpen() {
@@ -1125,27 +1384,20 @@ func (t *Torrent) openNewConns() (initiated int) {
                        return
                }
                p := t.peers.PopMax()
-               t.initiateConn(p)
+               opts := outgoingConnOpts{
+                       peerInfo:                 p,
+                       t:                        t,
+                       requireRendezvous:        false,
+                       skipHolepunchRendezvous:  false,
+                       receivedHolepunchConnect: false,
+                       HeaderObfuscationPolicy:  t.cl.config.HeaderObfuscationPolicy,
+               }
+               initiateConn(opts, false)
                initiated++
        }
        return
 }
 
-func (t *Torrent) getConnPieceInclination() []int {
-       _ret := t.connPieceInclinationPool.Get()
-       if _ret == nil {
-               pieceInclinationsNew.Add(1)
-               return rand.Perm(int(t.numPieces()))
-       }
-       pieceInclinationsReused.Add(1)
-       return *_ret.(*[]int)
-}
-
-func (t *Torrent) putPieceInclination(pi []int) {
-       t.connPieceInclinationPool.Put(&pi)
-       pieceInclinationsPut.Add(1)
-}
-
 func (t *Torrent) updatePieceCompletion(piece pieceIndex) bool {
        p := t.piece(piece)
        uncached := t.pieceCompleteUncached(piece)
@@ -1153,13 +1405,21 @@ func (t *Torrent) updatePieceCompletion(piece pieceIndex) bool {
        changed := cached != uncached
        complete := uncached.Complete
        p.storageCompletionOk = uncached.Ok
-       t._completedPieces.Set(bitmap.BitIndex(piece), complete)
+       x := uint32(piece)
+       if complete {
+               t._completedPieces.Add(x)
+               t.openNewConns()
+       } else {
+               t._completedPieces.Remove(x)
+       }
+       p.t.updatePieceRequestOrder(piece)
+       t.updateComplete()
        if complete && len(p.dirtiers) != 0 {
                t.logger.Printf("marked piece %v complete but still has dirtiers", piece)
        }
        if changed {
-               log.Fstr("piece %d completion changed: %+v -> %+v", piece, cached, uncached).SetLevel(log.Debug).Log(t.logger)
-               t.pieceCompletionChanged(piece)
+               log.Fstr("piece %d completion changed: %+v -> %+v", piece, cached, uncached).LogLevel(log.Debug, t.logger)
+               t.pieceCompletionChanged(piece, "Torrent.updatePieceCompletion")
        }
        return changed
 }
@@ -1181,8 +1441,9 @@ func (t *Torrent) readAt(b []byte, off int64) (n int, err error) {
        return
 }
 
-// Returns an error if the metadata was completed, but couldn't be set for
-// some reason. Blame it on the last peer to contribute.
+// Returns an error if the metadata was completed, but couldn't be set for some reason. Blame it on
+// the last peer to contribute. TODO: Actually we shouldn't blame peers for failure to open storage
+// etc. Also we should probably cached metadata pieces per-Peer, to isolate failure appropriately.
 func (t *Torrent) maybeCompleteMetadata() error {
        if t.haveInfo() {
                // Nothing to do.
@@ -1192,7 +1453,7 @@ func (t *Torrent) maybeCompleteMetadata() error {
                // Don't have enough metadata pieces.
                return nil
        }
-       err := t.setInfoBytes(t.metadataBytes)
+       err := t.setInfoBytesLocked(t.metadataBytes)
        if err != nil {
                t.invalidateMetadata()
                return fmt.Errorf("error setting info bytes: %s", err)
@@ -1207,7 +1468,7 @@ func (t *Torrent) readerPiecePriorities() (now, readahead bitmap.Bitmap) {
        t.forReaderOffsetPieces(func(begin, end pieceIndex) bool {
                if end > begin {
                        now.Add(bitmap.BitIndex(begin))
-                       readahead.AddRange(bitmap.BitIndex(begin)+1, bitmap.BitIndex(end))
+                       readahead.AddRange(bitmap.BitRange(begin)+1, bitmap.BitRange(end))
                }
                return true
        })
@@ -1221,7 +1482,7 @@ func (t *Torrent) needData() bool {
        if !t.haveInfo() {
                return true
        }
-       return t._pendingPieces.Len() != 0
+       return !t._pendingPieces.IsEmpty()
 }
 
 func appendMissingStrings(old, new []string) (ret []string) {
@@ -1261,17 +1522,17 @@ func (t *Torrent) bytesCompleted() int64 {
        if !t.haveInfo() {
                return 0
        }
-       return t.info.TotalLength() - t.bytesLeft()
+       return t.length() - t.bytesLeft()
 }
 
 func (t *Torrent) SetInfoBytes(b []byte) (err error) {
        t.cl.lock()
        defer t.cl.unlock()
-       return t.setInfoBytes(b)
+       return t.setInfoBytesLocked(b)
 }
 
 // Returns true if connection is removed from torrent.Conns.
-func (t *Torrent) deleteConnection(c *PeerConn) (ret bool) {
+func (t *Torrent) deletePeerConn(c *PeerConn) (ret bool) {
        if !c.closed.IsSet() {
                panic("connection is not closed")
                // There are behaviours prevented by the closed state that will fail
@@ -1287,37 +1548,56 @@ func (t *Torrent) deleteConnection(c *PeerConn) (ret bool) {
                }
        }
        torrent.Add("deleted connections", 1)
-       c.deleteAllRequests()
-       if t.numActivePeers() == 0 {
-               t.assertNoPendingRequests()
+       c.deleteAllRequests("Torrent.deletePeerConn")
+       t.assertPendingRequests()
+       if t.numActivePeers() == 0 && len(t.connsWithAllPieces) != 0 {
+               panic(t.connsWithAllPieces)
        }
        return
 }
 
-func (t *Torrent) numActivePeers() (num int) {
-       t.iterPeers(func(*Peer) {
-               num++
+func (t *Torrent) decPeerPieceAvailability(p *Peer) {
+       if t.deleteConnWithAllPieces(p) {
+               return
+       }
+       if !t.haveInfo() {
+               return
+       }
+       p.peerPieces().Iterate(func(i uint32) bool {
+               p.t.decPieceAvailability(pieceIndex(i))
+               return true
        })
-       return
 }
 
-func (t *Torrent) assertNoPendingRequests() {
-       if len(t.pendingRequests) != 0 {
-               panic(t.pendingRequests)
+func (t *Torrent) assertPendingRequests() {
+       if !check.Enabled {
+               return
        }
-       //if len(t.lastRequested) != 0 {
-       //      panic(t.lastRequested)
-       //}
+       // var actual pendingRequests
+       // if t.haveInfo() {
+       //      actual.m = make([]int, t.numChunks())
+       // }
+       // t.iterPeers(func(p *Peer) {
+       //      p.requestState.Requests.Iterate(func(x uint32) bool {
+       //              actual.Inc(x)
+       //              return true
+       //      })
+       // })
+       // diff := cmp.Diff(actual.m, t.pendingRequests.m)
+       // if diff != "" {
+       //      panic(diff)
+       // }
 }
 
 func (t *Torrent) dropConnection(c *PeerConn) {
        t.cl.event.Broadcast()
        c.close()
-       if t.deleteConnection(c) {
+       if t.deletePeerConn(c) {
                t.openNewConns()
        }
 }
 
+// Peers as in contact information for dialing out.
 func (t *Torrent) wantPeers() bool {
        if t.closed.IsSet() {
                return false
@@ -1325,7 +1605,7 @@ func (t *Torrent) wantPeers() bool {
        if t.peers.Len() > t.cl.config.TorrentPeersLowWater {
                return false
        }
-       return t.needData() || t.seeding()
+       return t.wantOutgoingConns()
 }
 
 func (t *Torrent) updateWantPeersEvent() {
@@ -1362,15 +1642,28 @@ func (t *Torrent) onWebRtcConn(
        dcc webtorrent.DataChannelContext,
 ) {
        defer c.Close()
+       netConn := webrtcNetConn{
+               ReadWriteCloser:    c,
+               DataChannelContext: dcc,
+       }
+       peerRemoteAddr := netConn.RemoteAddr()
+       //t.logger.Levelf(log.Critical, "onWebRtcConn remote addr: %v", peerRemoteAddr)
+       if t.cl.badPeerAddr(peerRemoteAddr) {
+               return
+       }
+       localAddrIpPort := missinggo.IpPortFromNetAddr(netConn.LocalAddr())
        pc, err := t.cl.initiateProtocolHandshakes(
                context.Background(),
-               webrtcNetConn{c, dcc},
+               netConn,
                t,
-               dcc.LocalOffered,
                false,
-               webrtcNetAddr{dcc.Remote},
-               webrtcNetwork,
-               fmt.Sprintf("webrtc offer_id %x", dcc.OfferId),
+               newConnectionOpts{
+                       outgoing:        dcc.LocalOffered,
+                       remoteAddr:      peerRemoteAddr,
+                       localPublicAddr: localAddrIpPort,
+                       network:         webrtcNetwork,
+                       connString:      fmt.Sprintf("webrtc offer_id %x: %v", dcc.OfferId, regularNetConnPeerConnConnString(netConn)),
+               },
        )
        if err != nil {
                t.logger.WithDefaultLevel(log.Error).Printf("error in handshaking webrtc connection: %v", err)
@@ -1381,18 +1674,19 @@ func (t *Torrent) onWebRtcConn(
        } else {
                pc.Discovery = PeerSourceIncoming
        }
+       pc.conn.SetWriteDeadline(time.Time{})
        t.cl.lock()
        defer t.cl.unlock()
        err = t.cl.runHandshookConn(pc, t)
        if err != nil {
-               t.logger.WithDefaultLevel(log.Critical).Printf("error running handshook webrtc conn: %v", err)
+               t.logger.WithDefaultLevel(log.Debug).Printf("error running handshook webrtc conn: %v", err)
        }
 }
 
 func (t *Torrent) logRunHandshookConn(pc *PeerConn, logAll bool, level log.Level) {
        err := t.cl.runHandshookConn(pc, t)
        if err != nil || logAll {
-               t.logger.WithDefaultLevel(level).Printf("error running handshook conn: %v", err)
+               t.logger.WithDefaultLevel(level).Levelf(log.ErrorLevel(err), "error running handshook conn: %v", err)
        }
 }
 
@@ -1401,11 +1695,10 @@ func (t *Torrent) runHandshookConnLoggingErr(pc *PeerConn) {
 }
 
 func (t *Torrent) startWebsocketAnnouncer(u url.URL) torrentTrackerAnnouncer {
-       wtc, release := t.cl.websocketTrackers.Get(u.String())
-       go func() {
-               <-t.closed.LockedChan(t.cl.locker())
-               release()
-       }()
+       wtc, release := t.cl.websocketTrackers.Get(u.String(), t.infoHash)
+       // This needs to run before the Torrent is dropped from the Client, to prevent a new webtorrent.TrackerClient for
+       // the same info hash before the old one is cleaned up.
+       t.onClose = append(t.onClose, release)
        wst := websocketTrackerStatus{u, wtc}
        go func() {
                err := wtc.Announce(tracker.Started, t.infoHash)
@@ -1417,7 +1710,6 @@ func (t *Torrent) startWebsocketAnnouncer(u url.URL) torrentTrackerAnnouncer {
                }
        }()
        return wst
-
 }
 
 func (t *Torrent) startScrapingTracker(_url string) {
@@ -1426,10 +1718,9 @@ func (t *Torrent) startScrapingTracker(_url string) {
        }
        u, err := url.Parse(_url)
        if err != nil {
-               // URLs with a leading '*' appear to be a uTorrent convention to
-               // disable trackers.
+               // URLs with a leading '*' appear to be a uTorrent convention to disable trackers.
                if _url[0] != '*' {
-                       log.Str("error parsing tracker url").AddValues("url", _url).Log(t.logger)
+                       t.logger.Levelf(log.Warning, "error parsing tracker url: %v", err)
                }
                return
        }
@@ -1460,8 +1751,9 @@ func (t *Torrent) startScrapingTracker(_url string) {
                        }
                }
                newAnnouncer := &trackerScraper{
-                       u: *u,
-                       t: t,
+                       u:               *u,
+                       t:               t,
+                       lookupTrackerIp: t.cl.config.LookupTrackerIp,
                }
                go newAnnouncer.Run()
                return newAnnouncer
@@ -1498,7 +1790,7 @@ func (t *Torrent) announceRequest(event tracker.AnnounceEvent) tracker.AnnounceR
                Event: event,
                NumWant: func() int32 {
                        if t.wantPeers() && len(t.cl.dialers) > 0 {
-                               return -1
+                               return 200 // Win has UDP packet limit. See: https://github.com/anacrolix/torrent/issues/764
                        } else {
                                return 0
                        }
@@ -1523,31 +1815,53 @@ func (t *Torrent) consumeDhtAnnouncePeers(pvs <-chan dht.PeersValues) {
        cl := t.cl
        for v := range pvs {
                cl.lock()
+               added := 0
                for _, cp := range v.Peers {
                        if cp.Port == 0 {
                                // Can't do anything with this.
                                continue
                        }
-                       t.addPeer(PeerInfo{
+                       if t.addPeer(PeerInfo{
                                Addr:   ipPortAddr{cp.IP, cp.Port},
                                Source: PeerSourceDhtGetPeers,
-                       })
+                       }) {
+                               added++
+                       }
                }
                cl.unlock()
+               // if added != 0 {
+               //      log.Printf("added %v peers from dht for %v", added, t.InfoHash().HexString())
+               // }
        }
 }
 
-func (t *Torrent) announceToDht(impliedPort bool, s DhtServer) error {
-       ps, err := s.Announce(t.infoHash, t.cl.incomingPeerPort(), impliedPort)
+// Announce using the provided DHT server. Peers are consumed automatically. done is closed when the
+// announce ends. stop will force the announce to end.
+func (t *Torrent) AnnounceToDht(s DhtServer) (done <-chan struct{}, stop func(), err error) {
+       ps, err := s.Announce(t.infoHash, t.cl.incomingPeerPort(), true)
+       if err != nil {
+               return
+       }
+       _done := make(chan struct{})
+       done = _done
+       stop = ps.Close
+       go func() {
+               t.consumeDhtAnnouncePeers(ps.Peers())
+               close(_done)
+       }()
+       return
+}
+
+func (t *Torrent) timeboxedAnnounceToDht(s DhtServer) error {
+       _, stop, err := t.AnnounceToDht(s)
        if err != nil {
                return err
        }
-       go t.consumeDhtAnnouncePeers(ps.Peers())
        select {
-       case <-t.closed.LockedChan(t.cl.locker()):
+       case <-t.closed.Done():
        case <-time.After(5 * time.Minute):
        }
-       ps.Close()
+       stop()
        return nil
 }
 
@@ -1560,7 +1874,10 @@ func (t *Torrent) dhtAnnouncer(s DhtServer) {
                        if t.closed.IsSet() {
                                return
                        }
-                       if !t.wantPeers() {
+                       // We're also announcing ourselves as a listener, so we don't just want peer addresses.
+                       // TODO: We can include the announce_peer step depending on whether we can receive
+                       // inbound connections. We should probably only announce once every 15 mins too.
+                       if !t.wantAnyConns() {
                                goto wait
                        }
                        // TODO: Determine if there's a listener on the port we're announcing.
@@ -1575,7 +1892,7 @@ func (t *Torrent) dhtAnnouncer(s DhtServer) {
                        t.numDHTAnnounces++
                        cl.unlock()
                        defer cl.lock()
-                       err := t.announceToDht(true, s)
+                       err := t.timeboxedAnnounceToDht(s)
                        if err != nil {
                                t.logger.WithDefaultLevel(log.Warning).Printf("error announcing %q to DHT: %s", t, err)
                        }
@@ -1612,6 +1929,7 @@ func (t *Torrent) statsLocked() (ret TorrentStats) {
                }
        }
        ret.ConnStats = t.stats.Copy()
+       ret.PiecesComplete = t.numPiecesCompleted()
        return
 }
 
@@ -1653,7 +1971,7 @@ func (t *Torrent) reconcileHandshakeStats(c *PeerConn) {
 }
 
 // Returns true if the connection is added.
-func (t *Torrent) addConnection(c *PeerConn) (err error) {
+func (t *Torrent) addPeerConn(c *PeerConn) (err error) {
        defer func() {
                if err == nil {
                        torrent.Add("added connections", 1)
@@ -1669,45 +1987,92 @@ func (t *Torrent) addConnection(c *PeerConn) (err error) {
                if !t.cl.config.DropDuplicatePeerIds {
                        continue
                }
-               if left, ok := c.hasPreferredNetworkOver(c0); ok && left {
+               if c.hasPreferredNetworkOver(c0) {
                        c0.close()
-                       t.deleteConnection(c0)
+                       t.deletePeerConn(c0)
                } else {
                        return errors.New("existing connection preferred")
                }
        }
        if len(t.conns) >= t.maxEstablishedConns {
-               c := t.worstBadConn()
+               numOutgoing := t.numOutgoingConns()
+               numIncoming := len(t.conns) - numOutgoing
+               c := t.worstBadConn(worseConnLensOpts{
+                       // We've already established that we have too many connections at this point, so we just
+                       // need to match what kind we have too many of vs. what we're trying to add now.
+                       incomingIsBad: (numIncoming-numOutgoing > 1) && c.outgoing,
+                       outgoingIsBad: (numOutgoing-numIncoming > 1) && !c.outgoing,
+               })
                if c == nil {
-                       return errors.New("don't want conns")
+                       return errors.New("don't want conn")
                }
                c.close()
-               t.deleteConnection(c)
+               t.deletePeerConn(c)
        }
        if len(t.conns) >= t.maxEstablishedConns {
                panic(len(t.conns))
        }
        t.conns[c] = struct{}{}
+       t.cl.event.Broadcast()
+       // We'll never receive the "p" extended handshake parameter.
        if !t.cl.config.DisablePEX && !c.PeerExtensionBytes.SupportsExtended() {
-               t.pex.Add(c) // as no further extended handshake expected
+               t.pex.Add(c)
        }
        return nil
 }
 
-func (t *Torrent) wantConns() bool {
-       if !t.networkingEnabled {
+func (t *Torrent) newConnsAllowed() bool {
+       if !t.networkingEnabled.Bool() {
+               return false
+       }
+       if t.closed.IsSet() {
+               return false
+       }
+       if !t.needData() && (!t.seeding() || !t.haveAnyPieces()) {
+               return false
+       }
+       return true
+}
+
+func (t *Torrent) wantAnyConns() bool {
+       if !t.networkingEnabled.Bool() {
                return false
        }
        if t.closed.IsSet() {
                return false
        }
-       if !t.seeding() && !t.needData() {
+       if !t.needData() && (!t.seeding() || !t.haveAnyPieces()) {
+               return false
+       }
+       return len(t.conns) < t.maxEstablishedConns
+}
+
+func (t *Torrent) wantOutgoingConns() bool {
+       if !t.newConnsAllowed() {
+               return false
+       }
+       if len(t.conns) < t.maxEstablishedConns {
+               return true
+       }
+       numIncomingConns := len(t.conns) - t.numOutgoingConns()
+       return t.worstBadConn(worseConnLensOpts{
+               incomingIsBad: numIncomingConns-t.numOutgoingConns() > 1,
+               outgoingIsBad: false,
+       }) != nil
+}
+
+func (t *Torrent) wantIncomingConns() bool {
+       if !t.newConnsAllowed() {
                return false
        }
        if len(t.conns) < t.maxEstablishedConns {
                return true
        }
-       return t.worstBadConn() != nil
+       numIncomingConns := len(t.conns) - t.numOutgoingConns()
+       return t.worstBadConn(worseConnLensOpts{
+               incomingIsBad: false,
+               outgoingIsBad: t.numOutgoingConns()-numIncomingConns > 1,
+       }) != nil
 }
 
 func (t *Torrent) SetMaxEstablishedConns(max int) (oldMax int) {
@@ -1715,18 +2080,24 @@ func (t *Torrent) SetMaxEstablishedConns(max int) (oldMax int) {
        defer t.cl.unlock()
        oldMax = t.maxEstablishedConns
        t.maxEstablishedConns = max
-       wcs := slices.HeapInterface(slices.FromMapKeys(t.conns), func(l, r *PeerConn) bool {
-               return worseConn(&l.Peer, &r.Peer)
-       })
+       wcs := worseConnSlice{
+               conns: t.appendConns(nil, func(*PeerConn) bool {
+                       return true
+               }),
+       }
+       wcs.initKeys(worseConnLensOpts{})
+       heap.Init(&wcs)
        for len(t.conns) > t.maxEstablishedConns && wcs.Len() > 0 {
-               t.dropConnection(wcs.Pop().(*PeerConn))
+               t.dropConnection(heap.Pop(&wcs).(*PeerConn))
        }
        t.openNewConns()
        return oldMax
 }
 
 func (t *Torrent) pieceHashed(piece pieceIndex, passed bool, hashIoErr error) {
-       t.logger.Log(log.Fstr("hashed piece %d (passed=%t)", piece, passed).SetLevel(log.Debug))
+       t.logger.LazyLog(log.Debug, func() log.Msg {
+               return log.Fstr("hashed piece %d (passed=%t)", piece, passed)
+       })
        p := t.piece(piece)
        p.numVerifies++
        t.cl.event.Broadcast()
@@ -1739,7 +2110,12 @@ func (t *Torrent) pieceHashed(piece pieceIndex, passed bool, hashIoErr error) {
                if passed {
                        pieceHashedCorrect.Add(1)
                } else {
-                       log.Fmsg("piece %d failed hash: %d connections contributed", piece, len(p.dirtiers)).AddValues(t, p).Log(t.logger)
+                       log.Fmsg(
+                               "piece %d failed hash: %d connections contributed", piece, len(p.dirtiers),
+                       ).AddValues(t, p).LogLevel(
+
+                               log.Debug, t.logger)
+
                        pieceHashedNotCorrect.Add(1)
                }
        }
@@ -1760,7 +2136,11 @@ func (t *Torrent) pieceHashed(piece pieceIndex, passed bool, hashIoErr error) {
                        c._stats.incrementPiecesDirtiedGood()
                }
                t.clearPieceTouchers(piece)
+               hasDirty := p.hasDirtyChunks()
                t.cl.unlock()
+               if hasDirty {
+                       p.Flush() // You can be synchronous here!
+               }
                err := p.Storage().MarkComplete()
                if err != nil {
                        t.logger.Printf("%T: error marking piece complete %d: %s", t.storage, piece, err)
@@ -1807,8 +2187,16 @@ func (t *Torrent) pieceHashed(piece pieceIndex, passed bool, hashIoErr error) {
 
                        if len(bannableTouchers) >= 1 {
                                c := bannableTouchers[0]
-                               t.cl.banPeerIP(c.remoteIp())
-                               c.drop()
+                               if len(bannableTouchers) != 1 {
+                                       t.logger.Levelf(log.Debug, "would have banned %v for touching piece %v after failed piece check", c.remoteIp(), piece)
+                               } else {
+                                       // Turns out it's still useful to ban peers like this because if there's only a
+                                       // single peer for a piece, and we never progress that piece to completion, we
+                                       // will never smart-ban them. Discovered in
+                                       // https://github.com/anacrolix/torrent/issues/715.
+                                       t.logger.Levelf(log.Warning, "banning %v for being sole dirtier of piece %v after failed piece check", c, piece)
+                                       c.ban()
+                               }
                        }
                }
                t.onIncompletePiece(piece)
@@ -1818,18 +2206,20 @@ func (t *Torrent) pieceHashed(piece pieceIndex, passed bool, hashIoErr error) {
 }
 
 func (t *Torrent) cancelRequestsForPiece(piece pieceIndex) {
-       // TODO: Make faster
-       for cn := range t.conns {
-               cn.tickleWriter()
+       start := t.pieceRequestIndexOffset(piece)
+       end := start + t.pieceNumChunks(piece)
+       for ri := start; ri < end; ri++ {
+               t.cancelRequest(ri)
        }
 }
 
 func (t *Torrent) onPieceCompleted(piece pieceIndex) {
        t.pendAllChunkSpecs(piece)
        t.cancelRequestsForPiece(piece)
+       t.piece(piece).readerCond.Broadcast()
        for conn := range t.conns {
                conn.have(piece)
-               t.maybeDropMutuallyCompletePeer(&conn.Peer)
+               t.maybeDropMutuallyCompletePeer(conn)
        }
 }
 
@@ -1855,13 +2245,13 @@ func (t *Torrent) onIncompletePiece(piece pieceIndex) {
        // }
        t.iterPeers(func(conn *Peer) {
                if conn.peerHasPiece(piece) {
-                       conn.updateRequests()
+                       conn.updateRequests("piece incomplete")
                }
        })
 }
 
 func (t *Torrent) tryCreateMorePieceHashers() {
-       for !t.closed.IsSet() && t.activePieceHashes < 2 && t.tryCreatePieceHasher() {
+       for !t.closed.IsSet() && t.activePieceHashes < t.cl.config.PieceHashersPerTorrent && t.tryCreatePieceHasher() {
        }
 }
 
@@ -1874,10 +2264,10 @@ func (t *Torrent) tryCreatePieceHasher() bool {
                return false
        }
        p := t.piece(pi)
-       t.piecesQueuedForHash.Remove(pi)
+       t.piecesQueuedForHash.Remove(bitmap.BitIndex(pi))
        p.hashing = true
        t.publishPieceChange(pi)
-       t.updatePiecePriority(pi)
+       t.updatePiecePriority(pi, "Torrent.tryCreatePieceHasher")
        t.storageLock.RLock()
        t.activePieceHashes++
        go t.pieceHasher(pi)
@@ -1896,9 +2286,32 @@ func (t *Torrent) getPieceToHash() (ret pieceIndex, ok bool) {
        return
 }
 
+func (t *Torrent) dropBannedPeers() {
+       t.iterPeers(func(p *Peer) {
+               remoteIp := p.remoteIp()
+               if remoteIp == nil {
+                       if p.bannableAddr.Ok {
+                               t.logger.WithDefaultLevel(log.Debug).Printf("can't get remote ip for peer %v", p)
+                       }
+                       return
+               }
+               netipAddr := netip.MustParseAddr(remoteIp.String())
+               if Some(netipAddr) != p.bannableAddr {
+                       t.logger.WithDefaultLevel(log.Debug).Printf(
+                               "peer remote ip does not match its bannable addr [peer=%v, remote ip=%v, bannable addr=%v]",
+                               p, remoteIp, p.bannableAddr)
+               }
+               if _, ok := t.cl.badPeerIPs[netipAddr]; ok {
+                       // Should this be a close?
+                       p.drop()
+                       t.logger.WithDefaultLevel(log.Debug).Printf("dropped %v for banned remote IP %v", p, netipAddr)
+               }
+       })
+}
+
 func (t *Torrent) pieceHasher(index pieceIndex) {
        p := t.piece(index)
-       sum, copyErr := t.hashPiece(index)
+       sum, failedPeers, copyErr := t.hashPiece(index)
        correct := sum == *p.hash
        switch copyErr {
        case nil, io.EOF:
@@ -1908,10 +2321,19 @@ func (t *Torrent) pieceHasher(index pieceIndex) {
        t.storageLock.RUnlock()
        t.cl.lock()
        defer t.cl.unlock()
+       if correct {
+               for peer := range failedPeers {
+                       t.cl.banPeerIP(peer.AsSlice())
+                       t.logger.WithDefaultLevel(log.Debug).Printf("smart banned %v for piece %v", peer, index)
+               }
+               t.dropBannedPeers()
+               for ri := t.pieceRequestIndexOffset(index); ri < t.pieceRequestIndexOffset(index+1); ri++ {
+                       t.smartBanCache.ForgetBlock(ri)
+               }
+       }
        p.hashing = false
-       t.updatePiecePriority(index)
        t.pieceHashed(index, correct, copyErr)
-       t.publishPieceChange(index)
+       t.updatePiecePriority(index, "Torrent.pieceHasher")
        t.activePieceHashes--
        t.tryCreateMorePieceHashers()
 }
@@ -1939,7 +2361,7 @@ func (t *Torrent) queuePieceCheck(pieceIndex pieceIndex) {
        }
        t.piecesQueuedForHash.Add(bitmap.BitIndex(pieceIndex))
        t.publishPieceChange(pieceIndex)
-       t.updatePiecePriority(pieceIndex)
+       t.updatePiecePriority(pieceIndex, "Torrent.queuePieceCheck")
        t.tryCreateMorePieceHashers()
 }
 
@@ -1951,8 +2373,45 @@ func (t *Torrent) VerifyData() {
        }
 }
 
-// Start the process of connecting to the given peer for the given torrent if appropriate.
-func (t *Torrent) initiateConn(peer PeerInfo) {
+func (t *Torrent) connectingToPeerAddr(addrStr string) bool {
+       return len(t.halfOpen[addrStr]) != 0
+}
+
+func (t *Torrent) hasPeerConnForAddr(x PeerRemoteAddr) bool {
+       addrStr := x.String()
+       for c := range t.conns {
+               ra := c.RemoteAddr
+               if ra.String() == addrStr {
+                       return true
+               }
+       }
+       return false
+}
+
+func (t *Torrent) getHalfOpenPath(
+       addrStr string,
+       attemptKey outgoingConnAttemptKey,
+) nestedmaps.Path[*PeerInfo] {
+       return nestedmaps.Next(nestedmaps.Next(nestedmaps.Begin(&t.halfOpen), addrStr), attemptKey)
+}
+
+func (t *Torrent) addHalfOpen(addrStr string, attemptKey *PeerInfo) {
+       path := t.getHalfOpenPath(addrStr, attemptKey)
+       if path.Exists() {
+               panic("should be unique")
+       }
+       path.Set(attemptKey)
+       t.cl.numHalfOpen++
+}
+
+// Start the process of connecting to the given peer for the given torrent if appropriate. I'm not
+// sure all the PeerInfo fields are being used.
+func initiateConn(
+       opts outgoingConnOpts,
+       ignoreLimits bool,
+) {
+       t := opts.t
+       peer := opts.peerInfo
        if peer.Id == t.cl.peerID {
                return
        }
@@ -1960,12 +2419,21 @@ func (t *Torrent) initiateConn(peer PeerInfo) {
                return
        }
        addr := peer.Addr
-       if t.addrActive(addr.String()) {
+       addrStr := addr.String()
+       if !ignoreLimits {
+               if t.connectingToPeerAddr(addrStr) {
+                       return
+               }
+       }
+       if t.hasPeerConnForAddr(addr) {
                return
        }
-       t.cl.numHalfOpen++
-       t.halfOpen[addr.String()] = peer
-       go t.cl.outgoingConnection(t, addr, peer.Source, peer.Trusted)
+       attemptKey := &peer
+       t.addHalfOpen(addrStr, attemptKey)
+       go t.cl.outgoingConnection(
+               opts,
+               attemptKey,
+       )
 }
 
 // Adds a trusted, pending peer for each of the given Client's addresses. Typically used in tests to
@@ -1986,7 +2454,7 @@ func (t *Torrent) AddClientPeer(cl *Client) int {
 // connection.
 func (t *Torrent) allStats(f func(*ConnStats)) {
        f(&t.stats)
-       f(&t.cl.stats)
+       f(&t.cl.connStats)
 }
 
 func (t *Torrent) hashingPiece(i pieceIndex) bool {
@@ -2005,30 +2473,6 @@ func (t *Torrent) piece(i int) *Piece {
        return &t.pieces[i]
 }
 
-func (t *Torrent) requestStrategyTorrent() requestStrategyTorrent {
-       return t
-}
-
-type torrentRequestStrategyCallbacks struct {
-       t *Torrent
-}
-
-func (cb torrentRequestStrategyCallbacks) requestTimedOut(r Request) {
-       torrent.Add("Request timeouts", 1)
-       cb.t.cl.lock()
-       defer cb.t.cl.unlock()
-       cb.t.iterPeers(func(cn *Peer) {
-               if cn.peerHasPiece(pieceIndex(r.Index)) {
-                       cn.updateRequests()
-               }
-       })
-
-}
-
-func (t *Torrent) requestStrategyCallbacks() requestStrategyCallbacks {
-       return torrentRequestStrategyCallbacks{t}
-}
-
 func (t *Torrent) onWriteChunkErr(err error) {
        if t.userOnWriteChunkErr != nil {
                go t.userOnWriteChunkErr(err)
@@ -2039,27 +2483,15 @@ func (t *Torrent) onWriteChunkErr(err error) {
 }
 
 func (t *Torrent) DisallowDataDownload() {
-       t.cl.lock()
-       defer t.cl.unlock()
        t.disallowDataDownloadLocked()
 }
 
 func (t *Torrent) disallowDataDownloadLocked() {
-       t.dataDownloadDisallowed = true
-       t.iterPeers(func(c *Peer) {
-               c.updateRequests()
-       })
-       t.tickleReaders()
+       t.dataDownloadDisallowed.Set()
 }
 
 func (t *Torrent) AllowDataDownload() {
-       t.cl.lock()
-       defer t.cl.unlock()
-       t.dataDownloadDisallowed = false
-       t.tickleReaders()
-       t.iterPeers(func(c *Peer) {
-               c.updateRequests()
-       })
+       t.dataDownloadDisallowed.Clear()
 }
 
 // Enables uploading data, if it was disabled.
@@ -2068,7 +2500,7 @@ func (t *Torrent) AllowDataUpload() {
        defer t.cl.unlock()
        t.dataUploadDisallowed = false
        for c := range t.conns {
-               c.updateRequests()
+               c.updateRequests("allow data upload")
        }
 }
 
@@ -2078,7 +2510,8 @@ func (t *Torrent) DisallowDataUpload() {
        defer t.cl.unlock()
        t.dataUploadDisallowed = true
        for c := range t.conns {
-               c.updateRequests()
+               // TODO: This doesn't look right. Shouldn't we tickle writers to choke peers or something instead?
+               c.updateRequests("disallow data upload")
        }
 }
 
@@ -2090,7 +2523,7 @@ func (t *Torrent) SetOnWriteChunkError(f func(error)) {
        t.userOnWriteChunkErr = f
 }
 
-func (t *Torrent) iterPeers(f func(*Peer)) {
+func (t *Torrent) iterPeers(f func(*Peer)) {
        for pc := range t.conns {
                f(&pc.Peer)
        }
@@ -2103,42 +2536,71 @@ func (t *Torrent) callbacks() *Callbacks {
        return &t.cl.config.Callbacks
 }
 
-var WebseedHttpClient = &http.Client{
-       Transport: &http.Transport{
-               MaxConnsPerHost: 10,
-       },
+type AddWebSeedsOpt func(*webseed.Client)
+
+// Sets the WebSeed trailing path escaper for a webseed.Client.
+func WebSeedPathEscaper(custom webseed.PathEscaper) AddWebSeedsOpt {
+       return func(c *webseed.Client) {
+               c.PathEscaper = custom
+       }
+}
+
+func (t *Torrent) AddWebSeeds(urls []string, opts ...AddWebSeedsOpt) {
+       t.cl.lock()
+       defer t.cl.unlock()
+       for _, u := range urls {
+               t.addWebSeed(u, opts...)
+       }
 }
 
-func (t *Torrent) addWebSeed(url string) {
+func (t *Torrent) addWebSeed(url string, opts ...AddWebSeedsOpt) {
        if t.cl.config.DisableWebseeds {
                return
        }
        if _, ok := t.webSeeds[url]; ok {
                return
        }
-       const maxRequests = 10
+       // I don't think Go http supports pipelining requests. However, we can have more ready to go
+       // right away. This value should be some multiple of the number of connections to a host. I
+       // would expect that double maxRequests plus a bit would be appropriate. This value is based on
+       // downloading Sintel (08ada5a7a6183aae1e09d831df6748d566095a10) from
+       // "https://webtorrent.io/torrents/".
+       const maxRequests = 16
        ws := webseedPeer{
                peer: Peer{
                        t:                        t,
                        outgoing:                 true,
                        Network:                  "http",
                        reconciledHandshakeStats: true,
-                       peerSentHaveAll:          true,
-                       // TODO: Raise this limit, and instead limit concurrent fetches.
-                       PeerMaxRequests: 32,
-                       RemoteAddr:      remoteAddrFromUrl(url),
-                       callbacks:       t.callbacks(),
+                       // This should affect how often we have to recompute requests for this peer. Note that
+                       // because we can request more than 1 thing at a time over HTTP, we will hit the low
+                       // requests mark more often, so recomputation is probably sooner than with regular peer
+                       // conns. ~4x maxRequests would be about right.
+                       PeerMaxRequests: 128,
+                       // TODO: Set ban prefix?
+                       RemoteAddr: remoteAddrFromUrl(url),
+                       callbacks:  t.callbacks(),
                },
                client: webseed.Client{
-                       // Consider a MaxConnsPerHost in the transport for this, possibly in a global Client.
-                       HttpClient: WebseedHttpClient,
+                       HttpClient: t.cl.httpClient,
                        Url:        url,
+                       ResponseBodyWrapper: func(r io.Reader) io.Reader {
+                               return &rateLimitedReader{
+                                       l: t.cl.config.DownloadRateLimiter,
+                                       r: r,
+                               }
+                       },
                },
                activeRequests: make(map[Request]webseed.Request, maxRequests),
        }
+       ws.peer.initRequestState()
+       for _, opt := range opts {
+               opt(&ws.client)
+       }
+       ws.peer.initUpdateRequestsTimer()
        ws.requesterCond.L = t.cl.locker()
-       for range iter.N(maxRequests) {
-               go ws.requester()
+       for i := 0; i < maxRequests; i += 1 {
+               go ws.requester(i)
        }
        for _, f := range t.callbacks().NewPeer {
                f(&ws.peer)
@@ -2159,3 +2621,285 @@ func (t *Torrent) peerIsActive(p *Peer) (active bool) {
        })
        return
 }
+
+func (t *Torrent) requestIndexToRequest(ri RequestIndex) Request {
+       index := t.pieceIndexOfRequestIndex(ri)
+       return Request{
+               pp.Integer(index),
+               t.piece(index).chunkIndexSpec(ri % t.chunksPerRegularPiece()),
+       }
+}
+
+func (t *Torrent) requestIndexFromRequest(r Request) RequestIndex {
+       return t.pieceRequestIndexOffset(pieceIndex(r.Index)) + RequestIndex(r.Begin/t.chunkSize)
+}
+
+func (t *Torrent) pieceRequestIndexOffset(piece pieceIndex) RequestIndex {
+       return RequestIndex(piece) * t.chunksPerRegularPiece()
+}
+
+func (t *Torrent) updateComplete() {
+       t.Complete.SetBool(t.haveAllPieces())
+}
+
+func (t *Torrent) cancelRequest(r RequestIndex) *Peer {
+       p := t.requestingPeer(r)
+       if p != nil {
+               p.cancel(r)
+       }
+       // TODO: This is a check that an old invariant holds. It can be removed after some testing.
+       //delete(t.pendingRequests, r)
+       if _, ok := t.requestState[r]; ok {
+               panic("expected request state to be gone")
+       }
+       return p
+}
+
+func (t *Torrent) requestingPeer(r RequestIndex) *Peer {
+       return t.requestState[r].peer
+}
+
+func (t *Torrent) addConnWithAllPieces(p *Peer) {
+       if t.connsWithAllPieces == nil {
+               t.connsWithAllPieces = make(map[*Peer]struct{}, t.maxEstablishedConns)
+       }
+       t.connsWithAllPieces[p] = struct{}{}
+}
+
+func (t *Torrent) deleteConnWithAllPieces(p *Peer) bool {
+       _, ok := t.connsWithAllPieces[p]
+       delete(t.connsWithAllPieces, p)
+       return ok
+}
+
+func (t *Torrent) numActivePeers() int {
+       return len(t.conns) + len(t.webSeeds)
+}
+
+func (t *Torrent) hasStorageCap() bool {
+       f := t.storage.Capacity
+       if f == nil {
+               return false
+       }
+       _, ok := (*f)()
+       return ok
+}
+
+func (t *Torrent) pieceIndexOfRequestIndex(ri RequestIndex) pieceIndex {
+       return pieceIndex(ri / t.chunksPerRegularPiece())
+}
+
+func (t *Torrent) iterUndirtiedRequestIndexesInPiece(
+       reuseIter *typedRoaring.Iterator[RequestIndex],
+       piece pieceIndex,
+       f func(RequestIndex),
+) {
+       reuseIter.Initialize(&t.dirtyChunks)
+       pieceRequestIndexOffset := t.pieceRequestIndexOffset(piece)
+       iterBitmapUnsetInRange(
+               reuseIter,
+               pieceRequestIndexOffset, pieceRequestIndexOffset+t.pieceNumChunks(piece),
+               f,
+       )
+}
+
+type requestState struct {
+       peer *Peer
+       when time.Time
+}
+
+// Returns an error if a received chunk is out of bounds in someway.
+func (t *Torrent) checkValidReceiveChunk(r Request) error {
+       if !t.haveInfo() {
+               return errors.New("torrent missing info")
+       }
+       if int(r.Index) >= t.numPieces() {
+               return fmt.Errorf("chunk index %v, torrent num pieces %v", r.Index, t.numPieces())
+       }
+       pieceLength := t.pieceLength(pieceIndex(r.Index))
+       if r.Begin >= pieceLength {
+               return fmt.Errorf("chunk begins beyond end of piece (%v >= %v)", r.Begin, pieceLength)
+       }
+       // We could check chunk lengths here, but chunk request size is not changed often, and tricky
+       // for peers to manipulate as they need to send potentially large buffers to begin with. There
+       // should be considerable checks elsewhere for this case due to the network overhead. We should
+       // catch most of the overflow manipulation stuff by checking index and begin above.
+       return nil
+}
+
+func (t *Torrent) peerConnsWithDialAddrPort(target netip.AddrPort) (ret []*PeerConn) {
+       for pc := range t.conns {
+               dialAddr, err := pc.remoteDialAddrPort()
+               if err != nil {
+                       continue
+               }
+               if dialAddr != target {
+                       continue
+               }
+               ret = append(ret, pc)
+       }
+       return
+}
+
+func wrapUtHolepunchMsgForPeerConn(
+       recipient *PeerConn,
+       msg utHolepunch.Msg,
+) pp.Message {
+       extendedPayload, err := msg.MarshalBinary()
+       if err != nil {
+               panic(err)
+       }
+       return pp.Message{
+               Type:            pp.Extended,
+               ExtendedID:      MapMustGet(recipient.PeerExtensionIDs, utHolepunch.ExtensionName),
+               ExtendedPayload: extendedPayload,
+       }
+}
+
+func sendUtHolepunchMsg(
+       pc *PeerConn,
+       msgType utHolepunch.MsgType,
+       addrPort netip.AddrPort,
+       errCode utHolepunch.ErrCode,
+) {
+       holepunchMsg := utHolepunch.Msg{
+               MsgType:  msgType,
+               AddrPort: addrPort,
+               ErrCode:  errCode,
+       }
+       incHolepunchMessagesSent(holepunchMsg)
+       ppMsg := wrapUtHolepunchMsgForPeerConn(pc, holepunchMsg)
+       pc.write(ppMsg)
+}
+
+func incHolepunchMessages(msg utHolepunch.Msg, verb string) {
+       torrent.Add(
+               fmt.Sprintf(
+                       "holepunch %v %v messages %v",
+                       msg.MsgType,
+                       addrPortProtocolStr(msg.AddrPort),
+                       verb,
+               ),
+               1,
+       )
+}
+
+func incHolepunchMessagesReceived(msg utHolepunch.Msg) {
+       incHolepunchMessages(msg, "received")
+}
+
+func incHolepunchMessagesSent(msg utHolepunch.Msg) {
+       incHolepunchMessages(msg, "sent")
+}
+
+func (t *Torrent) handleReceivedUtHolepunchMsg(msg utHolepunch.Msg, sender *PeerConn) error {
+       incHolepunchMessagesReceived(msg)
+       switch msg.MsgType {
+       case utHolepunch.Rendezvous:
+               t.logger.Printf("got holepunch rendezvous request for %v from %p", msg.AddrPort, sender)
+               sendMsg := sendUtHolepunchMsg
+               senderAddrPort, err := sender.remoteDialAddrPort()
+               if err != nil {
+                       sender.logger.Levelf(
+                               log.Warning,
+                               "error getting ut_holepunch rendezvous sender's dial address: %v",
+                               err,
+                       )
+                       // There's no better error code. The sender's address itself is invalid. I don't see
+                       // this error message being appropriate anywhere else anyway.
+                       sendMsg(sender, utHolepunch.Error, msg.AddrPort, utHolepunch.NoSuchPeer)
+               }
+               targets := t.peerConnsWithDialAddrPort(msg.AddrPort)
+               if len(targets) == 0 {
+                       sendMsg(sender, utHolepunch.Error, msg.AddrPort, utHolepunch.NotConnected)
+                       return nil
+               }
+               for _, pc := range targets {
+                       if !pc.supportsExtension(utHolepunch.ExtensionName) {
+                               sendMsg(sender, utHolepunch.Error, msg.AddrPort, utHolepunch.NoSupport)
+                               continue
+                       }
+                       sendMsg(sender, utHolepunch.Connect, msg.AddrPort, 0)
+                       sendMsg(pc, utHolepunch.Connect, senderAddrPort, 0)
+               }
+               return nil
+       case utHolepunch.Connect:
+               holepunchAddr := msg.AddrPort
+               t.logger.Printf("got holepunch connect request for %v from %p", holepunchAddr, sender)
+               if g.MapContains(t.cl.undialableWithoutHolepunch, holepunchAddr) {
+                       setAdd(&t.cl.undialableWithoutHolepunchDialedAfterHolepunchConnect, holepunchAddr)
+                       if g.MapContains(t.cl.accepted, holepunchAddr) {
+                               setAdd(&t.cl.probablyOnlyConnectedDueToHolepunch, holepunchAddr)
+                       }
+               }
+               opts := outgoingConnOpts{
+                       peerInfo: PeerInfo{
+                               Addr:         msg.AddrPort,
+                               Source:       PeerSourceUtHolepunch,
+                               PexPeerFlags: sender.pex.remoteLiveConns[msg.AddrPort].UnwrapOrZeroValue(),
+                       },
+                       t: t,
+                       // Don't attempt to start our own rendezvous if we fail to connect.
+                       skipHolepunchRendezvous:  true,
+                       receivedHolepunchConnect: true,
+                       // Assume that the other end initiated the rendezvous, and will use our preferred
+                       // encryption. So we will act normally.
+                       HeaderObfuscationPolicy: t.cl.config.HeaderObfuscationPolicy,
+               }
+               initiateConn(opts, true)
+               return nil
+       case utHolepunch.Error:
+               torrent.Add("holepunch error messages received", 1)
+               t.logger.Levelf(log.Debug, "received ut_holepunch error message from %v: %v", sender, msg.ErrCode)
+               return nil
+       default:
+               return fmt.Errorf("unhandled msg type %v", msg.MsgType)
+       }
+}
+
+func addrPortProtocolStr(addrPort netip.AddrPort) string {
+       addr := addrPort.Addr()
+       switch {
+       case addr.Is4():
+               return "ipv4"
+       case addr.Is6():
+               return "ipv6"
+       default:
+               panic(addrPort)
+       }
+}
+
+func (t *Torrent) trySendHolepunchRendezvous(addrPort netip.AddrPort) error {
+       rzsSent := 0
+       for pc := range t.conns {
+               if !pc.supportsExtension(utHolepunch.ExtensionName) {
+                       continue
+               }
+               if pc.supportsExtension(pp.ExtensionNamePex) {
+                       if !g.MapContains(pc.pex.remoteLiveConns, addrPort) {
+                               continue
+                       }
+               }
+               t.logger.Levelf(log.Debug, "sent ut_holepunch rendezvous message to %v for %v", pc, addrPort)
+               sendUtHolepunchMsg(pc, utHolepunch.Rendezvous, addrPort, 0)
+               rzsSent++
+       }
+       if rzsSent == 0 {
+               return errors.New("no eligible relays")
+       }
+       return nil
+}
+
+func (t *Torrent) numHalfOpenAttempts() (num int) {
+       for _, attempts := range t.halfOpen {
+               num += len(attempts)
+       }
+       return
+}
+
+func (t *Torrent) getDialTimeoutUnlocked() time.Duration {
+       cl := t.cl
+       cl.rLock()
+       defer cl.rUnlock()
+       return t.dialTimeout()
+}