10 "github.com/anacrolix/missinggo/pubsub"
11 "github.com/frankban/quicktest"
12 "github.com/stretchr/testify/require"
14 "github.com/anacrolix/torrent/metainfo"
15 pp "github.com/anacrolix/torrent/peer_protocol"
16 "github.com/anacrolix/torrent/storage"
19 // Ensure that no race exists between sending a bitfield, and a subsequent
20 // Have that would potentially alter it.
21 func TestSendBitfieldThenHave(t *testing.T) {
23 cl.init(TestingConfig(t))
25 c := cl.newConnection(nil, false, nil, "io.Pipe", "")
26 c.setTorrent(cl.newTorrent(metainfo.Hash{}, nil))
27 if err := c.t.setInfo(&metainfo.Info{Pieces: make([]byte, metainfo.HashSize*3)}); err != nil {
35 c.t._completedPieces.Add(1)
36 c.postBitfield( /*[]bool{false, true, false}*/ )
42 n, err := io.ReadFull(r, b)
44 // This will cause connection.writer to terminate.
47 require.NoError(t, err)
48 require.EqualValues(t, 15, n)
49 // Here we see that the bitfield doesn't have piece 2 set, as that should
50 // arrive in the following Have message.
51 require.EqualValues(t, "\x00\x00\x00\x02\x05@\x00\x00\x00\x05\x04\x00\x00\x00\x02", string(b))
54 type torrentStorage struct {
58 func (me *torrentStorage) Close() error { return nil }
60 func (me *torrentStorage) Piece(mp metainfo.Piece) storage.PieceImpl {
64 func (me *torrentStorage) Completion() storage.Completion {
65 return storage.Completion{}
68 func (me *torrentStorage) MarkComplete() error {
72 func (me *torrentStorage) MarkNotComplete() error {
76 func (me *torrentStorage) ReadAt([]byte, int64) (int, error) {
77 panic("shouldn't be called")
80 func (me *torrentStorage) WriteAt(b []byte, _ int64) (int, error) {
81 if len(b) != defaultChunkSize {
88 func BenchmarkConnectionMainReadLoop(b *testing.B) {
91 cl.init(&ClientConfig{
92 DownloadRateLimiter: unlimited,
95 ts := &torrentStorage{}
98 storage: &storage.Torrent{TorrentImpl: storage.TorrentImpl{Piece: ts.Piece, Close: ts.Close}},
99 pieceStateChanges: pubsub.NewPubSub(),
101 t.setChunkSize(defaultChunkSize)
102 require.NoError(b, t.setInfo(&metainfo.Info{
103 Pieces: make([]byte, 20),
105 PieceLength: 1 << 20,
107 t._pendingPieces.Add(0)
109 cn := cl.newConnection(r, true, r.RemoteAddr(), r.RemoteAddr().Network(), regularNetConnPeerConnConnString(r))
111 mrlErrChan := make(chan error)
114 Piece: make([]byte, defaultChunkSize),
118 err := cn.mainReadLoop()
124 wb := msg.MustMarshalBinary()
125 b.SetBytes(int64(len(msg.Piece)))
128 for i := 0; i < b.N; i += 1 {
130 // The chunk must be written to storage everytime, to ensure the
131 // writeSem is unlocked.
132 t.pendAllChunkSpecs(0)
133 cn.validReceiveChunks = map[RequestIndex]int{
134 t.requestIndexFromRequest(newRequestFromMessage(&msg)): 1,
137 n, err := w.Write(wb)
138 require.NoError(b, err)
139 require.EqualValues(b, len(wb), n)
142 if err := w.Close(); err != nil {
146 mrlErr := <-mrlErrChan
147 if mrlErr != nil && !errors.Is(mrlErr, io.EOF) {
150 c.Assert(cn._stats.ChunksReadUseful.Int64(), quicktest.Equals, int64(b.N))
153 func TestConnPexPeerFlags(t *testing.T) {
155 tcpAddr = &net.TCPAddr{IP: net.IPv6loopback, Port: 4848}
156 udpAddr = &net.UDPAddr{IP: net.IPv6loopback, Port: 4848}
158 testcases := []struct {
162 {&PeerConn{Peer: Peer{outgoing: false, PeerPrefersEncryption: false}}, 0},
163 {&PeerConn{Peer: Peer{outgoing: false, PeerPrefersEncryption: true}}, pp.PexPrefersEncryption},
164 {&PeerConn{Peer: Peer{outgoing: true, PeerPrefersEncryption: false}}, pp.PexOutgoingConn},
165 {&PeerConn{Peer: Peer{outgoing: true, PeerPrefersEncryption: true}}, pp.PexOutgoingConn | pp.PexPrefersEncryption},
166 {&PeerConn{Peer: Peer{RemoteAddr: udpAddr, Network: udpAddr.Network()}}, pp.PexSupportsUtp},
167 {&PeerConn{Peer: Peer{RemoteAddr: udpAddr, Network: udpAddr.Network(), outgoing: true}}, pp.PexOutgoingConn | pp.PexSupportsUtp},
168 {&PeerConn{Peer: Peer{RemoteAddr: tcpAddr, Network: tcpAddr.Network(), outgoing: true}}, pp.PexOutgoingConn},
169 {&PeerConn{Peer: Peer{RemoteAddr: tcpAddr, Network: tcpAddr.Network()}}, 0},
171 for i, tc := range testcases {
172 f := tc.conn.pexPeerFlags()
173 require.EqualValues(t, tc.f, f, i)
177 func TestConnPexEvent(t *testing.T) {
179 udpAddr = &net.UDPAddr{IP: net.IPv6loopback, Port: 4848}
180 tcpAddr = &net.TCPAddr{IP: net.IPv6loopback, Port: 4848}
181 dialTcpAddr = &net.TCPAddr{IP: net.IPv6loopback, Port: 4747}
182 dialUdpAddr = &net.UDPAddr{IP: net.IPv6loopback, Port: 4747}
184 testcases := []struct {
191 &PeerConn{Peer: Peer{RemoteAddr: udpAddr, Network: udpAddr.Network()}},
192 pexEvent{pexAdd, udpAddr, pp.PexSupportsUtp, nil},
196 &PeerConn{Peer: Peer{RemoteAddr: tcpAddr, Network: tcpAddr.Network(), outgoing: true, PeerListenPort: dialTcpAddr.Port}},
197 pexEvent{pexDrop, tcpAddr, pp.PexOutgoingConn, nil},
201 &PeerConn{Peer: Peer{RemoteAddr: tcpAddr, Network: tcpAddr.Network(), PeerListenPort: dialTcpAddr.Port}},
202 pexEvent{pexAdd, dialTcpAddr, 0, nil},
206 &PeerConn{Peer: Peer{RemoteAddr: udpAddr, Network: udpAddr.Network(), PeerListenPort: dialUdpAddr.Port}},
207 pexEvent{pexDrop, dialUdpAddr, pp.PexSupportsUtp, nil},
210 for i, tc := range testcases {
211 e := tc.c.pexEvent(tc.t)
212 require.EqualValues(t, tc.e, e, i)