premiere-libtorrent/src/request_blocks.cpp

307 lines
11 KiB
C++
Raw Normal View History

2014-07-06 21:18:00 +02:00
/*
Copyright (c) 2003-2016, Arvid Norberg
2014-07-06 21:18:00 +02:00
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the distribution.
* Neither the name of the author nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
POSSIBILITY OF SUCH DAMAGE.
*/
#include "libtorrent/bitfield.hpp"
#include "libtorrent/peer_connection.hpp"
#include "libtorrent/torrent.hpp"
#include "libtorrent/socket_type.hpp"
#include "libtorrent/peer_info.hpp" // for peer_info flags
#include "libtorrent/performance_counters.hpp" // for counters
2015-04-21 06:30:34 +02:00
#include "libtorrent/request_blocks.hpp"
#include "libtorrent/alert_manager.hpp"
#include "libtorrent/aux_/has_block.hpp"
2014-07-06 21:18:00 +02:00
#include <vector>
namespace libtorrent {
2014-07-06 21:18:00 +02:00
int source_rank(int source_bitmask)
{
int ret = 0;
if (source_bitmask & peer_info::tracker) ret |= 1 << 5;
if (source_bitmask & peer_info::lsd) ret |= 1 << 4;
if (source_bitmask & peer_info::dht) ret |= 1 << 3;
if (source_bitmask & peer_info::pex) ret |= 1 << 2;
return ret;
}
// the case where ignore_peer is motivated is if two peers
// have only one piece that we don't have, and it's the
// same piece for both peers. Then they might get into an
// infinite loop, fighting to request the same blocks.
// returns false if the function is aborted by an early-exit
// condition.
bool request_a_block(torrent& t, peer_connection& c)
{
if (t.is_seed()) return false;
if (c.no_download()) return false;
if (t.upload_mode()) return false;
if (c.is_disconnecting()) return false;
// don't request pieces before we have the metadata
if (!t.valid_metadata()) return false;
// don't request pieces before the peer is properly
// initialized after we have the metadata
if (!t.are_files_checked()) return false;
// we don't want to request more blocks while trying to gracefully pause
if (t.graceful_pause()) return false;
TORRENT_ASSERT(c.peer_info_struct() != nullptr
|| c.type() != connection_type::bittorrent);
2014-07-06 21:18:00 +02:00
bool const time_critical_mode = t.num_time_critical_pieces() > 0;
2014-07-06 21:18:00 +02:00
// in time critical mode, only have 1 outstanding request at a time
// via normal requests
int const desired_queue_size = time_critical_mode
? 1 : c.desired_queue_size();
2014-07-06 21:18:00 +02:00
int num_requests = desired_queue_size
2015-08-16 18:17:23 +02:00
- int(c.download_queue().size())
- int(c.request_queue().size());
2014-07-06 21:18:00 +02:00
#ifndef TORRENT_DISABLE_LOGGING
if (c.should_log(peer_log_alert::info))
{
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
, "dlq: %d rqq: %d target: %d req: %d engame: %d"
, int(c.download_queue().size()), int(c.request_queue().size())
, desired_queue_size, num_requests, c.endgame());
}
2014-07-06 21:18:00 +02:00
#endif
TORRENT_ASSERT(desired_queue_size > 0);
2014-07-06 21:18:00 +02:00
// if our request queue is already full, we
// don't have to make any new requests yet
if (num_requests <= 0) return false;
t.need_picker();
piece_picker& p = t.picker();
std::vector<piece_block> interesting_pieces;
interesting_pieces.reserve(100);
int prefer_contiguous_blocks = c.prefer_contiguous_blocks();
2014-07-06 21:18:00 +02:00
if (prefer_contiguous_blocks == 0 && !time_critical_mode)
2014-07-06 21:18:00 +02:00
{
int blocks_per_piece = t.torrent_file().piece_length() / t.block_size();
prefer_contiguous_blocks = c.statistics().download_payload_rate()
2014-07-06 21:18:00 +02:00
* t.settings().get_int(settings_pack::whole_pieces_threshold)
> t.torrent_file().piece_length() ? blocks_per_piece : 0;
2014-07-06 21:18:00 +02:00
}
// if we prefer whole pieces, the piece picker will pick at least
// the number of blocks we want, but it will try to make the picked
// blocks be from whole pieces, possibly by returning more blocks
// than we requested.
#if TORRENT_USE_ASSERTS
2014-07-06 21:18:00 +02:00
error_code ec;
TORRENT_ASSERT(c.remote() == c.get_socket()->remote_endpoint(ec) || ec);
#endif
aux::session_interface& ses = t.session();
std::vector<pending_block> const& dq = c.download_queue();
std::vector<pending_block> const& rq = c.request_queue();
std::vector<piece_index_t> const& suggested = c.suggested_pieces();
auto const* bits = &c.get_bitfield();
typed_bitfield<piece_index_t> fast_mask;
2014-07-06 21:18:00 +02:00
if (c.has_peer_choked())
{
// if we are choked we can only pick pieces from the
// allowed fast set. The allowed fast set is sorted
// in ascending priority order
// build a bitmask with only the allowed pieces in it
fast_mask.resize(c.get_bitfield().size(), false);
for (auto const& i : c.allowed_fast())
{
if ((*bits)[i]) fast_mask.set_bit(i);
}
2014-07-06 21:18:00 +02:00
bits = &fast_mask;
}
// picks the interesting pieces from this peer
// the integer is the number of pieces that
// should be guaranteed to be available for download
// (if num_requests is too big, too many pieces are
// picked and cpu-time is wasted)
// the last argument is if we should prefer whole pieces
// for this peer. If we're downloading one piece in 20 seconds
// then use this mode.
std::uint32_t const flags = p.pick_pieces(*bits, interesting_pieces
, num_requests, prefer_contiguous_blocks, c.peer_info_struct()
, c.picker_options(), suggested, t.num_peers()
2014-07-06 21:18:00 +02:00
, ses.stats_counters());
#ifndef TORRENT_DISABLE_LOGGING
if (t.alerts().should_post<picker_log_alert>()
&& !interesting_pieces.empty())
{
t.alerts().emplace_alert<picker_log_alert>(t.get_handle(), c.remote()
, c.pid(), flags, &interesting_pieces[0], int(interesting_pieces.size()));
}
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
, "prefer_contiguous: %d picked: %d"
, prefer_contiguous_blocks, int(interesting_pieces.size()));
#else
TORRENT_UNUSED(flags);
2014-07-06 21:18:00 +02:00
#endif
// if the number of pieces we have + the number of pieces
// we're requesting from is less than the number of pieces
// in the torrent, there are still some unrequested pieces
// and we're not strictly speaking in end-game mode yet
// also, if we already have at least one outstanding
// request, we shouldn't pick any busy pieces either
// in time critical mode, it's OK to request busy blocks
bool const dont_pick_busy_blocks
2014-07-06 21:18:00 +02:00
= ((ses.settings().get_bool(settings_pack::strict_end_game_mode)
&& p.get_download_queue_size() < p.num_want_left())
|| dq.size() + rq.size() > 0)
&& !time_critical_mode;
// this is filled with an interesting piece
// that some other peer is currently downloading
piece_block busy_block = piece_block::invalid;
for (std::vector<piece_block>::iterator i = interesting_pieces.begin();
i != interesting_pieces.end(); ++i)
{
if (prefer_contiguous_blocks == 0 && num_requests <= 0) break;
2014-07-06 21:18:00 +02:00
if (time_critical_mode && p.piece_priority(i->piece_index) != 7)
{
// assume the subsequent pieces are not prio 7 and
// be done
break;
}
int num_block_requests = p.num_peers(*i);
if (num_block_requests > 0)
{
// have we picked enough pieces?
if (num_requests <= 0) break;
// this block is busy. This means all the following blocks
// in the interesting_pieces list are busy as well, we might
// as well just exit the loop
if (dont_pick_busy_blocks) break;
TORRENT_ASSERT(p.num_peers(*i) > 0);
busy_block = *i;
continue;
}
TORRENT_ASSERT(p.num_peers(*i) == 0);
// don't request pieces we already have in our request queue
// This happens when pieces time out or the peer sends us
// pieces we didn't request. Those aren't marked in the
// piece picker, but we still keep track of them in the
// download queue
if (std::find_if(dq.begin(), dq.end(), aux::has_block(*i)) != dq.end()
|| std::find_if(rq.begin(), rq.end(), aux::has_block(*i)) != rq.end())
2014-07-06 21:18:00 +02:00
{
#if TORRENT_USE_ASSERTS
2014-07-06 21:18:00 +02:00
std::vector<pending_block>::const_iterator j
= std::find_if(dq.begin(), dq.end(), aux::has_block(*i));
2014-07-06 21:18:00 +02:00
if (j != dq.end()) TORRENT_ASSERT(j->timed_out || j->not_wanted);
#endif
#ifndef TORRENT_DISABLE_LOGGING
c.peer_log(peer_log_alert::info, "PIECE_PICKER"
, "not_picking: %d,%d already in queue"
, static_cast<int>(i->piece_index), i->block_index);
2014-07-06 21:18:00 +02:00
#endif
continue;
}
// ok, we found a piece that's not being downloaded
// by somebody else. request it from this peer
// and return
if (!c.add_request(*i, 0)) continue;
TORRENT_ASSERT(p.num_peers(*i) == 1);
TORRENT_ASSERT(p.is_requested(*i));
num_requests--;
}
// we have picked as many blocks as we should
// we're done!
if (num_requests <= 0)
{
// since we could pick as many blocks as we
// requested without having to resort to picking
// busy ones, we're not in end-game mode
c.set_endgame(false);
return true;
}
// we did not pick as many pieces as we wanted, because
// there aren't enough. This means we're in end-game mode
// as long as we have at least one request outstanding,
// we shouldn't pick another piece
// if we are attempting to download 'allowed' pieces
// and can't find any, that doesn't count as end-game
if (!c.has_peer_choked())
c.set_endgame(true);
2014-07-06 21:18:00 +02:00
// if we don't have any potential busy blocks to request
// or if we already have outstanding requests, don't
// pick a busy piece
if (busy_block == piece_block::invalid
|| dq.size() + rq.size() > 0)
{
return true;
}
#if TORRENT_USE_ASSERTS
2014-07-06 21:18:00 +02:00
piece_picker::downloading_piece st;
p.piece_info(busy_block.piece_index, st);
TORRENT_ASSERT(st.requested + st.finished + st.writing
== p.blocks_in_piece(busy_block.piece_index));
#endif
TORRENT_ASSERT(p.is_requested(busy_block));
TORRENT_ASSERT(!p.is_downloaded(busy_block));
TORRENT_ASSERT(!p.is_finished(busy_block));
TORRENT_ASSERT(p.num_peers(busy_block) > 0);
c.add_request(busy_block, peer_connection::req_busy);
return true;
}
}