2014-07-06 21:18:00 +02:00
|
|
|
/*
|
|
|
|
|
2016-01-18 00:57:46 +01:00
|
|
|
Copyright (c) 2003-2016, Arvid Norberg
|
2014-07-06 21:18:00 +02:00
|
|
|
All rights reserved.
|
|
|
|
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
|
|
modification, are permitted provided that the following conditions
|
|
|
|
are met:
|
|
|
|
|
|
|
|
* Redistributions of source code must retain the above copyright
|
|
|
|
notice, this list of conditions and the following disclaimer.
|
|
|
|
* Redistributions in binary form must reproduce the above copyright
|
|
|
|
notice, this list of conditions and the following disclaimer in
|
|
|
|
the documentation and/or other materials provided with the distribution.
|
|
|
|
* Neither the name of the author nor the names of its
|
|
|
|
contributors may be used to endorse or promote products derived
|
|
|
|
from this software without specific prior written permission.
|
|
|
|
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "libtorrent/bitfield.hpp"
|
|
|
|
#include "libtorrent/peer_connection.hpp"
|
|
|
|
#include "libtorrent/torrent.hpp"
|
|
|
|
#include "libtorrent/socket_type.hpp"
|
|
|
|
#include "libtorrent/peer_info.hpp" // for peer_info flags
|
|
|
|
#include "libtorrent/performance_counters.hpp" // for counters
|
2015-04-21 06:30:34 +02:00
|
|
|
#include "libtorrent/request_blocks.hpp"
|
2015-11-29 06:58:46 +01:00
|
|
|
#include "libtorrent/alert_manager.hpp"
|
2016-09-24 17:46:56 +02:00
|
|
|
#include "libtorrent/aux_/has_block.hpp"
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
#include <vector>
|
|
|
|
|
|
|
|
namespace libtorrent
|
|
|
|
{
|
|
|
|
int source_rank(int source_bitmask)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
if (source_bitmask & peer_info::tracker) ret |= 1 << 5;
|
|
|
|
if (source_bitmask & peer_info::lsd) ret |= 1 << 4;
|
|
|
|
if (source_bitmask & peer_info::dht) ret |= 1 << 3;
|
|
|
|
if (source_bitmask & peer_info::pex) ret |= 1 << 2;
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
// the case where ignore_peer is motivated is if two peers
|
|
|
|
// have only one piece that we don't have, and it's the
|
|
|
|
// same piece for both peers. Then they might get into an
|
|
|
|
// infinite loop, fighting to request the same blocks.
|
|
|
|
// returns false if the function is aborted by an early-exit
|
|
|
|
// condition.
|
|
|
|
bool request_a_block(torrent& t, peer_connection& c)
|
|
|
|
{
|
|
|
|
if (t.is_seed()) return false;
|
|
|
|
if (c.no_download()) return false;
|
|
|
|
if (t.upload_mode()) return false;
|
|
|
|
if (c.is_disconnecting()) return false;
|
|
|
|
|
|
|
|
// don't request pieces before we have the metadata
|
|
|
|
if (!t.valid_metadata()) return false;
|
|
|
|
|
|
|
|
// don't request pieces before the peer is properly
|
|
|
|
// initialized after we have the metadata
|
|
|
|
if (!t.are_files_checked()) return false;
|
|
|
|
|
2015-12-04 07:08:01 +01:00
|
|
|
// we don't want to request more blocks while trying to gracefully pause
|
|
|
|
if (t.graceful_pause()) return false;
|
|
|
|
|
2016-07-09 22:26:26 +02:00
|
|
|
TORRENT_ASSERT(c.peer_info_struct() != nullptr || c.type() != peer_connection::bittorrent_connection);
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
bool time_critical_mode = t.num_time_critical_pieces() > 0;
|
|
|
|
|
|
|
|
// in time critical mode, only have 1 outstanding request at a time
|
|
|
|
// via normal requests
|
2016-06-27 07:02:00 +02:00
|
|
|
int const desired_queue_size = time_critical_mode
|
|
|
|
? 1 : c.desired_queue_size();
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
int num_requests = desired_queue_size
|
2015-08-16 18:17:23 +02:00
|
|
|
- int(c.download_queue().size())
|
|
|
|
- int(c.request_queue().size());
|
2014-07-06 21:18:00 +02:00
|
|
|
|
2015-04-17 03:15:33 +02:00
|
|
|
#ifndef TORRENT_DISABLE_LOGGING
|
2016-09-14 04:46:07 +02:00
|
|
|
if (c.should_log(peer_log_alert::info))
|
|
|
|
{
|
|
|
|
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
|
|
|
|
, "dlq: %d rqq: %d target: %d req: %d engame: %d"
|
|
|
|
, int(c.download_queue().size()), int(c.request_queue().size())
|
|
|
|
, desired_queue_size, num_requests, c.endgame());
|
|
|
|
}
|
2014-07-06 21:18:00 +02:00
|
|
|
#endif
|
2016-06-27 07:02:00 +02:00
|
|
|
TORRENT_ASSERT(desired_queue_size > 0);
|
2014-07-06 21:18:00 +02:00
|
|
|
// if our request queue is already full, we
|
|
|
|
// don't have to make any new requests yet
|
|
|
|
if (num_requests <= 0) return false;
|
|
|
|
|
|
|
|
t.need_picker();
|
|
|
|
|
|
|
|
piece_picker& p = t.picker();
|
|
|
|
std::vector<piece_block> interesting_pieces;
|
|
|
|
interesting_pieces.reserve(100);
|
|
|
|
|
2015-01-20 03:34:55 +01:00
|
|
|
int prefer_contiguous_blocks = c.prefer_contiguous_blocks();
|
2014-07-06 21:18:00 +02:00
|
|
|
|
2015-01-20 03:34:55 +01:00
|
|
|
if (prefer_contiguous_blocks == 0 && !time_critical_mode)
|
2014-07-06 21:18:00 +02:00
|
|
|
{
|
2015-01-20 03:34:55 +01:00
|
|
|
int blocks_per_piece = t.torrent_file().piece_length() / t.block_size();
|
|
|
|
prefer_contiguous_blocks = c.statistics().download_payload_rate()
|
2014-07-06 21:18:00 +02:00
|
|
|
* t.settings().get_int(settings_pack::whole_pieces_threshold)
|
2015-01-20 03:34:55 +01:00
|
|
|
> t.torrent_file().piece_length() ? blocks_per_piece : 0;
|
2014-07-06 21:18:00 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// if we prefer whole pieces, the piece picker will pick at least
|
|
|
|
// the number of blocks we want, but it will try to make the picked
|
|
|
|
// blocks be from whole pieces, possibly by returning more blocks
|
|
|
|
// than we requested.
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2014-07-06 21:18:00 +02:00
|
|
|
error_code ec;
|
|
|
|
TORRENT_ASSERT(c.remote() == c.get_socket()->remote_endpoint(ec) || ec);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
aux::session_interface& ses = t.session();
|
|
|
|
|
|
|
|
std::vector<pending_block> const& dq = c.download_queue();
|
|
|
|
std::vector<pending_block> const& rq = c.request_queue();
|
|
|
|
|
|
|
|
std::vector<int> const& suggested = c.suggested_pieces();
|
|
|
|
bitfield const* bits = &c.get_bitfield();
|
|
|
|
bitfield fast_mask;
|
2015-11-29 06:58:46 +01:00
|
|
|
|
2014-07-06 21:18:00 +02:00
|
|
|
if (c.has_peer_choked())
|
|
|
|
{
|
|
|
|
// if we are choked we can only pick pieces from the
|
|
|
|
// allowed fast set. The allowed fast set is sorted
|
|
|
|
// in ascending priority order
|
|
|
|
std::vector<int> const& allowed_fast = c.allowed_fast();
|
|
|
|
|
|
|
|
// build a bitmask with only the allowed pieces in it
|
|
|
|
fast_mask.resize(c.get_bitfield().size(), false);
|
|
|
|
for (std::vector<int>::const_iterator i = allowed_fast.begin()
|
|
|
|
, end(allowed_fast.end()); i != end; ++i)
|
|
|
|
if ((*bits)[*i]) fast_mask.set_bit(*i);
|
|
|
|
bits = &fast_mask;
|
|
|
|
}
|
|
|
|
|
|
|
|
// picks the interesting pieces from this peer
|
|
|
|
// the integer is the number of pieces that
|
|
|
|
// should be guaranteed to be available for download
|
|
|
|
// (if num_requests is too big, too many pieces are
|
|
|
|
// picked and cpu-time is wasted)
|
|
|
|
// the last argument is if we should prefer whole pieces
|
|
|
|
// for this peer. If we're downloading one piece in 20 seconds
|
|
|
|
// then use this mode.
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t flags = p.pick_pieces(*bits, interesting_pieces
|
2015-01-20 03:34:55 +01:00
|
|
|
, num_requests, prefer_contiguous_blocks, c.peer_info_struct()
|
2015-02-08 22:17:00 +01:00
|
|
|
, c.picker_options(), suggested, t.num_peers()
|
2014-07-06 21:18:00 +02:00
|
|
|
, ses.stats_counters());
|
|
|
|
|
2015-04-17 03:15:33 +02:00
|
|
|
#ifndef TORRENT_DISABLE_LOGGING
|
2015-11-29 06:58:46 +01:00
|
|
|
if (t.alerts().should_post<picker_log_alert>()
|
|
|
|
&& !interesting_pieces.empty())
|
|
|
|
{
|
|
|
|
t.alerts().emplace_alert<picker_log_alert>(t.get_handle(), c.remote()
|
|
|
|
, c.pid(), flags, &interesting_pieces[0], int(interesting_pieces.size()));
|
|
|
|
}
|
2015-05-03 04:53:54 +02:00
|
|
|
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
|
|
|
|
, "prefer_contiguous: %d picked: %d"
|
2015-01-20 03:34:55 +01:00
|
|
|
, prefer_contiguous_blocks, int(interesting_pieces.size()));
|
2015-11-29 06:58:46 +01:00
|
|
|
#else
|
|
|
|
TORRENT_UNUSED(flags);
|
2014-07-06 21:18:00 +02:00
|
|
|
#endif
|
|
|
|
|
|
|
|
// if the number of pieces we have + the number of pieces
|
|
|
|
// we're requesting from is less than the number of pieces
|
|
|
|
// in the torrent, there are still some unrequested pieces
|
|
|
|
// and we're not strictly speaking in end-game mode yet
|
|
|
|
// also, if we already have at least one outstanding
|
|
|
|
// request, we shouldn't pick any busy pieces either
|
|
|
|
// in time critical mode, it's OK to request busy blocks
|
|
|
|
bool dont_pick_busy_blocks
|
|
|
|
= ((ses.settings().get_bool(settings_pack::strict_end_game_mode)
|
|
|
|
&& p.get_download_queue_size() < p.num_want_left())
|
|
|
|
|| dq.size() + rq.size() > 0)
|
|
|
|
&& !time_critical_mode;
|
|
|
|
|
|
|
|
// this is filled with an interesting piece
|
|
|
|
// that some other peer is currently downloading
|
|
|
|
piece_block busy_block = piece_block::invalid;
|
|
|
|
|
|
|
|
for (std::vector<piece_block>::iterator i = interesting_pieces.begin();
|
|
|
|
i != interesting_pieces.end(); ++i)
|
|
|
|
{
|
2015-01-20 03:34:55 +01:00
|
|
|
if (prefer_contiguous_blocks == 0 && num_requests <= 0) break;
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
if (time_critical_mode && p.piece_priority(i->piece_index) != 7)
|
|
|
|
{
|
|
|
|
// assume the subsequent pieces are not prio 7 and
|
|
|
|
// be done
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
int num_block_requests = p.num_peers(*i);
|
|
|
|
if (num_block_requests > 0)
|
|
|
|
{
|
|
|
|
// have we picked enough pieces?
|
|
|
|
if (num_requests <= 0) break;
|
|
|
|
|
|
|
|
// this block is busy. This means all the following blocks
|
|
|
|
// in the interesting_pieces list are busy as well, we might
|
|
|
|
// as well just exit the loop
|
|
|
|
if (dont_pick_busy_blocks) break;
|
|
|
|
|
|
|
|
TORRENT_ASSERT(p.num_peers(*i) > 0);
|
|
|
|
busy_block = *i;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
TORRENT_ASSERT(p.num_peers(*i) == 0);
|
|
|
|
|
|
|
|
// don't request pieces we already have in our request queue
|
|
|
|
// This happens when pieces time out or the peer sends us
|
|
|
|
// pieces we didn't request. Those aren't marked in the
|
|
|
|
// piece picker, but we still keep track of them in the
|
|
|
|
// download queue
|
2016-09-24 17:46:56 +02:00
|
|
|
if (std::find_if(dq.begin(), dq.end(), aux::has_block(*i)) != dq.end()
|
|
|
|
|| std::find_if(rq.begin(), rq.end(), aux::has_block(*i)) != rq.end())
|
2014-07-06 21:18:00 +02:00
|
|
|
{
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2014-07-06 21:18:00 +02:00
|
|
|
std::vector<pending_block>::const_iterator j
|
2016-09-24 17:46:56 +02:00
|
|
|
= std::find_if(dq.begin(), dq.end(), aux::has_block(*i));
|
2014-07-06 21:18:00 +02:00
|
|
|
if (j != dq.end()) TORRENT_ASSERT(j->timed_out || j->not_wanted);
|
|
|
|
#endif
|
2015-04-17 03:15:33 +02:00
|
|
|
#ifndef TORRENT_DISABLE_LOGGING
|
2015-05-03 04:53:54 +02:00
|
|
|
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
|
|
|
|
, "not_picking: %d,%d already in queue"
|
2014-07-06 21:18:00 +02:00
|
|
|
, i->piece_index, i->block_index);
|
|
|
|
#endif
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// ok, we found a piece that's not being downloaded
|
|
|
|
// by somebody else. request it from this peer
|
|
|
|
// and return
|
|
|
|
if (!c.add_request(*i, 0)) continue;
|
|
|
|
TORRENT_ASSERT(p.num_peers(*i) == 1);
|
|
|
|
TORRENT_ASSERT(p.is_requested(*i));
|
|
|
|
num_requests--;
|
|
|
|
}
|
|
|
|
|
|
|
|
// we have picked as many blocks as we should
|
|
|
|
// we're done!
|
|
|
|
if (num_requests <= 0)
|
|
|
|
{
|
|
|
|
// since we could pick as many blocks as we
|
|
|
|
// requested without having to resort to picking
|
|
|
|
// busy ones, we're not in end-game mode
|
|
|
|
c.set_endgame(false);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// we did not pick as many pieces as we wanted, because
|
|
|
|
// there aren't enough. This means we're in end-game mode
|
|
|
|
// as long as we have at least one request outstanding,
|
|
|
|
// we shouldn't pick another piece
|
|
|
|
// if we are attempting to download 'allowed' pieces
|
|
|
|
// and can't find any, that doesn't count as end-game
|
|
|
|
if (!c.has_peer_choked())
|
|
|
|
c.set_endgame(true);
|
2016-07-02 01:46:59 +02:00
|
|
|
|
2014-07-06 21:18:00 +02:00
|
|
|
// if we don't have any potential busy blocks to request
|
|
|
|
// or if we already have outstanding requests, don't
|
|
|
|
// pick a busy piece
|
|
|
|
if (busy_block == piece_block::invalid
|
|
|
|
|| dq.size() + rq.size() > 0)
|
|
|
|
{
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2014-07-06 21:18:00 +02:00
|
|
|
piece_picker::downloading_piece st;
|
|
|
|
p.piece_info(busy_block.piece_index, st);
|
|
|
|
TORRENT_ASSERT(st.requested + st.finished + st.writing
|
|
|
|
== p.blocks_in_piece(busy_block.piece_index));
|
|
|
|
#endif
|
|
|
|
TORRENT_ASSERT(p.is_requested(busy_block));
|
|
|
|
TORRENT_ASSERT(!p.is_downloaded(busy_block));
|
|
|
|
TORRENT_ASSERT(!p.is_finished(busy_block));
|
|
|
|
TORRENT_ASSERT(p.num_peers(busy_block) > 0);
|
|
|
|
|
|
|
|
c.add_request(busy_block, peer_connection::req_busy);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|