"context"
"errors"
"fmt"
+ "math/rand"
"sync"
+ "time"
"github.com/RoaringBitmap/roaring"
"github.com/anacrolix/log"
+
"github.com/anacrolix/torrent/metainfo"
pp "github.com/anacrolix/torrent/peer_protocol"
"github.com/anacrolix/torrent/webseed"
)
type webseedPeer struct {
+ // First field for stats alignment.
+ peer Peer
client webseed.Client
activeRequests map[Request]webseed.Request
requesterCond sync.Cond
- peer Peer
// Number of requester routines.
maxRequests int
}
}
func (ws *webseedPeer) _cancel(r RequestIndex) bool {
- active, ok := ws.activeRequests[ws.peer.t.requestIndexToRequest(r)]
- if ok {
+ if active, ok := ws.activeRequests[ws.peer.t.requestIndexToRequest(r)]; ok {
active.Cancel()
+ // The requester is running and will handle the result.
+ return true
}
- if !ws.peer.deleteRequest(r) {
- panic("cancelled webseed request should exist")
- }
- if ws.peer.isLowOnRequests() {
- ws.peer.updateRequests("webseedPeer._cancel")
- }
- return true
+ // There should be no requester handling this, so no further events will occur.
+ return false
}
func (ws *webseedPeer) intoSpec(r Request) webseed.RequestSpec {
return true
}
-func (ws *webseedPeer) doRequest(r Request) {
+func (ws *webseedPeer) doRequest(r Request) error {
webseedRequest := ws.client.NewRequest(ws.intoSpec(r))
ws.activeRequests[r] = webseedRequest
- func() {
+ err := func() error {
ws.requesterCond.L.Unlock()
defer ws.requesterCond.L.Lock()
- ws.requestResultHandler(r, webseedRequest)
+ return ws.requestResultHandler(r, webseedRequest)
}()
delete(ws.activeRequests, r)
+ return err
}
-func (ws *webseedPeer) requester() {
+func (ws *webseedPeer) requester(i int) {
ws.requesterCond.L.Lock()
defer ws.requesterCond.L.Unlock()
start:
for !ws.peer.closed.IsSet() {
restart := false
- ws.peer.actualRequestState.Requests.Iterate(func(x uint32) bool {
+ ws.peer.requestState.Requests.Iterate(func(x RequestIndex) bool {
r := ws.peer.t.requestIndexToRequest(x)
if _, ok := ws.activeRequests[r]; ok {
return true
}
- ws.doRequest(r)
+ err := ws.doRequest(r)
+ ws.requesterCond.L.Unlock()
+ if err != nil && !errors.Is(err, context.Canceled) {
+ log.Printf("requester %v: error doing webseed request %v: %v", i, r, err)
+ }
restart = true
+ if errors.Is(err, webseed.ErrTooFast) {
+ time.Sleep(time.Duration(rand.Int63n(int64(10 * time.Second))))
+ }
+ ws.requesterCond.L.Lock()
return false
})
if restart {
return "WS"
}
-// TODO: This is called when banning peers. Perhaps we want to be able to ban webseeds too. We could
-// return bool if this is even possible, and if it isn't, skip to the next drop candidate.
+// Maybe this should drop all existing connections, or something like that.
func (ws *webseedPeer) drop() {}
+func (cn *webseedPeer) ban() {
+ cn.peer.close()
+}
+
func (ws *webseedPeer) handleUpdateRequests() {
// Because this is synchronous, webseed peers seem to get first dibs on newly prioritized
// pieces.
- ws.peer.maybeUpdateActualRequestState()
+ go func() {
+ ws.peer.t.cl.lock()
+ defer ws.peer.t.cl.unlock()
+ ws.peer.maybeUpdateActualRequestState()
+ }()
}
func (ws *webseedPeer) onClose() {
- ws.peer.logger.WithLevel(log.Debug).Print("closing")
- ws.peer.deleteAllRequests()
- for _, r := range ws.activeRequests {
- r.Cancel()
- }
+ ws.peer.logger.Levelf(log.Debug, "closing")
+ // Just deleting them means we would have to manually cancel active requests.
+ ws.peer.cancelAllRequests()
+ ws.peer.t.iterPeers(func(p *Peer) {
+ if p.isLowOnRequests() {
+ p.updateRequests("webseedPeer.onClose")
+ }
+ })
ws.requesterCond.Broadcast()
}
-func (ws *webseedPeer) requestResultHandler(r Request, webseedRequest webseed.Request) {
+func (ws *webseedPeer) requestResultHandler(r Request, webseedRequest webseed.Request) error {
result := <-webseedRequest.Result
close(webseedRequest.Result) // one-shot
// We do this here rather than inside receiveChunk, since we want to count errors too. I'm not
ws.peer.t.cl.lock()
defer ws.peer.t.cl.unlock()
if ws.peer.t.closed.IsSet() {
- return
+ return nil
}
- if result.Err != nil {
- if !errors.Is(result.Err, context.Canceled) && !ws.peer.closed.IsSet() {
+ err := result.Err
+ if err != nil {
+ switch {
+ case errors.Is(err, context.Canceled):
+ case errors.Is(err, webseed.ErrTooFast):
+ case ws.peer.closed.IsSet():
+ default:
ws.peer.logger.Printf("Request %v rejected: %v", r, result.Err)
// // Here lies my attempt to extract something concrete from Go's error system. RIP.
// cfg := spew.NewDefaultConfig()
log.Printf("closing %v", ws)
ws.peer.close()
}
- ws.peer.remoteRejectedRequest(ws.peer.t.requestIndexFromRequest(r))
- } else {
- err := ws.peer.receiveChunk(&pp.Message{
- Type: pp.Piece,
- Index: r.Index,
- Begin: r.Begin,
- Piece: result.Bytes,
- })
- if err != nil {
- panic(err)
+ if !ws.peer.remoteRejectedRequest(ws.peer.t.requestIndexFromRequest(r)) {
+ panic("invalid reject")
}
+ return err
}
-}
-
-func (me *webseedPeer) isLowOnRequests() bool {
- return me.peer.actualRequestState.Requests.GetCardinality() < uint64(me.maxRequests)
+ err = ws.peer.receiveChunk(&pp.Message{
+ Type: pp.Piece,
+ Index: r.Index,
+ Begin: r.Begin,
+ Piece: result.Bytes,
+ })
+ if err != nil {
+ panic(err)
+ }
+ return err
}
func (me *webseedPeer) peerPieces() *roaring.Bitmap {