premiere-libtorrent/src/web_peer_connection.cpp

1170 lines
36 KiB
C++
Raw Normal View History

/*
Copyright (c) 2003-2016, Arvid Norberg
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
2014-07-06 21:18:00 +02:00
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the distribution.
* Neither the name of the author nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
POSSIBILITY OF SUCH DAMAGE.
*/
2015-08-20 01:33:20 +02:00
#include "libtorrent/config.hpp"
#include <functional>
#include <vector>
#include <cstdlib>
#include <cstdio> // for snprintf
#include <cinttypes> // for PRId64 et.al.
#include "libtorrent/web_peer_connection.hpp"
#include "libtorrent/session.hpp"
#include "libtorrent/identify_client.hpp"
#include "libtorrent/entry.hpp"
#include "libtorrent/bencode.hpp"
#include "libtorrent/alert_types.hpp"
#include "libtorrent/invariant_check.hpp"
#include "libtorrent/io.hpp"
#include "libtorrent/version.hpp"
#include "libtorrent/parse_url.hpp"
2009-11-26 06:45:43 +01:00
#include "libtorrent/peer_info.hpp"
2014-07-06 21:18:00 +02:00
#include "libtorrent/aux_/session_interface.hpp"
#include "libtorrent/alert_manager.hpp" // for alert_manageralert_manager
2015-04-21 02:23:00 +02:00
#include "libtorrent/aux_/escape_string.hpp" // for escape_path
#include "libtorrent/hex.hpp" // for is_hex
namespace libtorrent
{
2016-09-06 04:25:20 +02:00
constexpr int request_size_overhead = 5000;
2014-07-06 21:18:00 +02:00
struct disk_interface;
web_peer_connection::web_peer_connection(peer_connection_args const& pack
, web_seed_t& web)
: web_connection_base(pack, web)
2014-07-06 21:18:00 +02:00
, m_url(web.url)
, m_web(&web)
2014-07-06 21:18:00 +02:00
, m_received_body(0)
, m_chunk_pos(0)
, m_partial_chunk_header(0)
, m_num_responses(0)
{
INVARIANT_CHECK;
2014-07-06 21:18:00 +02:00
if (!m_settings.get_bool(settings_pack::report_web_seed_downloads))
ignore_stats(true);
std::shared_ptr<torrent> tor = pack.tor.lock();
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(tor);
2014-07-06 21:18:00 +02:00
// we always prefer downloading 1 MiB chunks
// from web seeds, or whole pieces if pieces
// are larger than a MiB
int preferred_size = 1024 * 1024;
2014-07-06 21:18:00 +02:00
// if the web server is known not to support keep-alive.
// request even larger blocks at a time
if (!web.supports_keepalive) preferred_size *= 4;
prefer_contiguous_blocks((std::max)(preferred_size / tor->block_size(), 1));
std::shared_ptr<torrent> t = associated_torrent().lock();
bool const single_file_request = t->torrent_file().num_files() == 1;
if (!single_file_request)
{
// handle incorrect .torrent files which are multi-file
// but have web seeds not ending with a slash
2016-12-11 02:51:55 +01:00
if (m_path.empty() || m_path[m_path.size() - 1] != '/') m_path += '/';
if (m_url.empty() || m_url[m_url.size() - 1] != '/') m_url += '/';
}
else
{
// handle .torrent files that don't include the filename in the url
if (m_path.empty()) m_path += '/';
2016-12-11 02:51:55 +01:00
if (m_path[m_path.size() - 1] == '/')
{
m_path += escape_string(t->torrent_file().name());
}
if (!m_url.empty() && m_url[m_url.size() - 1] == '/')
{
std::string tmp = t->torrent_file().files().file_path(0);
#ifdef TORRENT_WINDOWS
convert_path_to_posix(tmp);
#endif
m_url += escape_path(tmp);
}
}
2014-07-06 21:18:00 +02:00
// we want large blocks as well, so
// we can request more bytes at once
// this setting will merge adjacent requests
// into single larger ones
request_large_blocks(true);
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::info, "URL", "web_peer_connection %s", m_url.c_str());
#endif
2014-07-06 21:18:00 +02:00
}
2014-07-06 21:18:00 +02:00
void web_peer_connection::on_connected()
{
if (m_web->have_files.empty())
{
incoming_have_all();
}
else
{
std::shared_ptr<torrent> t = associated_torrent().lock();
// only advertise pieces that are contained within the files we have as
// indicated by m_web->have_files AND padfiles!
// it's important to include pieces that may overlap many files, as long
// as we have all those files, so instead of starting with a clear bitfield
// and setting the pieces corresponding to files we have, we do it the
// other way around. Start with assuming we have all files, and clear
// pieces overlapping with files we *don't* have.
bitfield have;
file_storage const& fs = t->torrent_file().files();
have.resize(fs.num_pieces(), true);
int const num_files = m_web->have_files.size();
for (int i = 0; i < num_files; ++i)
{
// if we have the file, no need to do anything
if (m_web->have_files.get_bit(i) || fs.pad_file_at(i)) continue;
std::tuple<int, int> const range = aux::file_piece_range_inclusive(fs, i);
for (int k = std::get<0>(range); k < std::get<1>(range); ++k)
have.clear_bit(k);
}
incoming_bitfield(have);
}
if (m_web->restart_request.piece != -1)
{
2014-07-06 21:18:00 +02:00
// increase the chances of requesting the block
// we have partial data for already, to finish it
incoming_suggest(m_web->restart_request.piece);
}
2014-07-06 21:18:00 +02:00
web_connection_base::on_connected();
}
void web_peer_connection::disconnect(error_code const& ec
, operation_t op, int error)
2014-07-06 21:18:00 +02:00
{
if (is_disconnecting()) return;
if (op == op_sock_write && ec == boost::system::errc::broken_pipe)
{
#ifndef TORRENT_DISABLE_LOGGING
// a write operation failed with broken-pipe. This typically happens
// with HTTP 1.0 servers that close their incoming channel of the TCP
// stream whenever they're done reading one full request. Instead of
// us bailing out and failing the entire request just because our
// write-end was closed, ignore it and keep reading until the read-end
// also is closed.
peer_log(peer_log_alert::info, "WRITE_DIRECTION", "CLOSED");
#endif
// prevent the peer from trying to send anything more
m_send_buffer.clear();
// when the web server closed our write-end of the socket (i.e. its
// read-end), if it's an HTTP 1.0 server. we will stop sending more
// requests. We'll close the connection once we receive the last bytes,
// and our read end is closed as well.
incoming_choke();
return;
}
if (op == op_connect && m_web && !m_web->endpoints.empty())
{
// we failed to connect to this IP. remove it so that the next attempt
// uses the next IP in the list.
m_web->endpoints.erase(m_web->endpoints.begin());
}
std::shared_ptr<torrent> t = associated_torrent().lock();
2014-07-06 21:18:00 +02:00
if (!m_requests.empty() && !m_file_requests.empty()
&& !m_piece.empty() && m_web)
2014-07-06 21:18:00 +02:00
{
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "SAVE_RESTART_DATA"
, "data: %d req: %d off: %d"
, int(m_piece.size()), int(m_requests.front().piece)
, int(m_requests.front().start));
}
#endif
m_web->restart_request = m_requests.front();
if (!m_web->restart_piece.empty())
{
2014-07-06 21:18:00 +02:00
// we're about to replace a different restart piece
// buffer. So it was wasted download
if (t) t->add_redundant_bytes(int(m_web->restart_piece.size())
2016-09-02 14:27:38 +02:00
, waste_reason::piece_closing);
}
m_web->restart_piece.swap(m_piece);
2014-07-06 21:18:00 +02:00
// we have to do this to not count this data as redundant. The
// upper layer will call downloading_piece_progress and assume
// it's all wasted download. Since we're saving it here, it isn't.
m_requests.clear();
}
if (m_web && !m_web->supports_keepalive && error == 0)
{
2014-07-06 21:18:00 +02:00
// if the web server doesn't support keepalive and we were
// disconnected as a graceful EOF, reconnect right away
if (t) get_io_service().post(
std::bind(&torrent::maybe_connect_web_seeds, t));
2014-07-06 21:18:00 +02:00
}
peer_connection::disconnect(ec, op, error);
if (t) t->disconnect_web_seed(this);
}
piece_block_progress web_peer_connection::downloading_piece_progress() const
2014-07-06 21:18:00 +02:00
{
if (m_requests.empty())
return piece_block_progress();
2014-07-06 21:18:00 +02:00
std::shared_ptr<torrent> t = associated_torrent().lock();
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(t);
piece_block_progress ret;
ret.piece_index = m_requests.front().piece;
ret.bytes_downloaded = int(m_piece.size());
2014-07-06 21:18:00 +02:00
// this is used to make sure that the block_index stays within
// bounds. If the entire piece is downloaded, the block_index
// would otherwise point to one past the end
int correction = m_piece.size() ? -1 : 0;
ret.block_index = int((m_requests.front().start + m_piece.size() + correction) / t->block_size());
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(ret.block_index < int(piece_block::invalid.block_index));
TORRENT_ASSERT(ret.piece_index < int(piece_block::invalid.piece_index));
ret.full_block_bytes = t->block_size();
const int last_piece = t->torrent_file().num_pieces() - 1;
if (ret.piece_index == last_piece && ret.block_index
== t->torrent_file().piece_size(last_piece) / t->block_size())
ret.full_block_bytes = t->torrent_file().piece_size(last_piece) % t->block_size();
return ret;
}
void web_peer_connection::write_request(peer_request const& r)
{
INVARIANT_CHECK;
std::shared_ptr<torrent> t = associated_torrent().lock();
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(t);
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(t->valid_metadata());
2014-07-06 21:18:00 +02:00
torrent_info const& info = t->torrent_file();
peer_request req = r;
2014-07-06 21:18:00 +02:00
std::string request;
request.reserve(400);
int size = r.length;
const int block_size = t->block_size();
const int piece_size = t->torrent_file().piece_length();
peer_request pr;
while (size > 0)
{
int request_offset = r.start + r.length - size;
pr.start = request_offset % piece_size;
pr.length = (std::min)(block_size, size);
pr.piece = r.piece + request_offset / piece_size;
m_requests.push_back(pr);
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::outgoing_message, "REQUESTING", "piece: %d start: %d len: %d"
, pr.piece, pr.start, pr.length);
#endif
if (m_web->restart_request == m_requests.front())
{
m_piece.swap(m_web->restart_piece);
peer_request const& front = m_requests.front();
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(front.length > int(m_piece.size()));
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "RESTART_DATA", "data: %d req: (%d, %d) size: %d"
, int(m_piece.size()), front.piece, front.start
, front.start + front.length - 1);
}
#else
TORRENT_UNUSED(front);
#endif
req.start += int(m_piece.size());
req.length -= int(m_piece.size());
2014-07-06 21:18:00 +02:00
// just to keep the accounting straight for the upper layer.
// it doesn't know we just re-wrote the request
incoming_piece_fragment(int(m_piece.size()));
m_web->restart_request.piece = -1;
2014-07-06 21:18:00 +02:00
}
#if 0
std::cerr << this << " REQ: p: " << pr.piece << " " << pr.start << std::endl;
#endif
2014-07-06 21:18:00 +02:00
size -= pr.length;
}
bool const single_file_request = t->torrent_file().num_files() == 1;
int const proxy_type = m_settings.get_int(settings_pack::proxy_type);
bool const using_proxy = (proxy_type == settings_pack::http
2014-07-06 21:18:00 +02:00
|| proxy_type == settings_pack::http_pw) && !m_ssl;
// the number of pad files that have been "requested". In case we _only_
// request padfiles, we can't rely on handling them in the on_receive()
// callback (because we won't receive anything), instead we have to post a
// pretend read callback where we can deliver the zeroes for the partfile
int num_pad_files = 0;
// TODO: 3 do we really need a special case here? wouldn't the multi-file
// case handle single file torrents correctly too?
2014-07-06 21:18:00 +02:00
if (single_file_request)
{
file_request_t file_req;
file_req.file_index = 0;
file_req.start = std::int64_t(req.piece) * info.piece_length()
+ req.start;
file_req.length = req.length;
2014-07-06 21:18:00 +02:00
request += "GET ";
// do not encode single file paths, they are
2014-07-06 21:18:00 +02:00
// assumed to be encoded in the torrent file
request += using_proxy ? m_url : m_path;
request += " HTTP/1.1\r\n";
add_headers(request, m_settings, using_proxy);
2014-07-06 21:18:00 +02:00
request += "\r\nRange: bytes=";
2016-05-01 05:10:47 +02:00
request += to_string(file_req.start).data();
2014-07-06 21:18:00 +02:00
request += "-";
2016-05-01 05:10:47 +02:00
request += to_string(file_req.start + file_req.length - 1).data();
2014-07-06 21:18:00 +02:00
request += "\r\n\r\n";
m_first_request = false;
m_file_requests.push_back(file_req);
2014-07-06 21:18:00 +02:00
}
else
{
std::vector<file_slice> files = info.orig_files().map_block(req.piece, req.start
, req.length);
2014-07-06 21:18:00 +02:00
for (std::vector<file_slice>::iterator i = files.begin();
i != files.end(); ++i)
{
2014-07-06 21:18:00 +02:00
file_slice const& f = *i;
file_request_t file_req;
file_req.file_index = f.file_index;
file_req.start = f.offset;
file_req.length = int(f.size);
2014-07-06 21:18:00 +02:00
if (info.orig_files().pad_file_at(f.file_index))
{
m_file_requests.push_back(file_req);
++num_pad_files;
2014-07-06 21:18:00 +02:00
continue;
}
TORRENT_ASSERT(m_web->have_files.empty()
|| m_web->have_files.get_bit(f.file_index));
2014-07-06 21:18:00 +02:00
request += "GET ";
if (using_proxy)
{
2014-07-06 21:18:00 +02:00
// m_url is already a properly escaped URL
// with the correct slashes. Don't encode it again
request += m_url;
}
auto redirection = m_web->redirects.find(f.file_index);
if (redirection != m_web->redirects.end())
{
request += redirection->second;
2014-07-06 21:18:00 +02:00
}
else
{
if (!using_proxy)
{
// m_path is already a properly escaped URL
// with the correct slashes. Don't encode it again
request += m_path;
}
2013-08-16 18:38:09 +02:00
2014-07-06 21:18:00 +02:00
std::string path = info.orig_files().file_path(f.file_index);
2009-10-30 04:42:29 +01:00
#ifdef TORRENT_WINDOWS
2014-07-06 21:18:00 +02:00
convert_path_to_posix(path);
2009-10-30 04:42:29 +01:00
#endif
request += escape_path(path);
2014-07-06 21:18:00 +02:00
}
request += " HTTP/1.1\r\n";
add_headers(request, m_settings, using_proxy);
2014-07-06 21:18:00 +02:00
request += "\r\nRange: bytes=";
2016-05-01 05:10:47 +02:00
request += to_string(f.offset).data();
2014-07-06 21:18:00 +02:00
request += "-";
2016-05-01 05:10:47 +02:00
request += to_string(f.offset + f.size - 1).data();
2014-07-06 21:18:00 +02:00
request += "\r\n\r\n";
m_first_request = false;
#if 0
2014-07-06 21:18:00 +02:00
std::cerr << this << " SEND-REQUEST: f: " << f.file_index
<< " s: " << f.offset
<< " e: " << (f.offset + f.size - 1) << std::endl;
#endif
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(f.file_index >= 0);
m_file_requests.push_back(file_req);
}
2014-07-06 21:18:00 +02:00
}
if (num_pad_files == int(m_file_requests.size()))
{
get_io_service().post(std::bind(
&web_peer_connection::on_receive_padfile,
std::static_pointer_cast<web_peer_connection>(self())));
return;
}
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::outgoing_message, "REQUEST", "%s", request.c_str());
#endif
send_buffer(request.c_str(), int(request.size()), message_type_request);
2014-07-06 21:18:00 +02:00
}
namespace {
std::string get_peer_name(http_parser const& p, std::string const& host)
{
std::string ret = "URL seed @ ";
ret += host;
std::string const& server_version = p.header("server");
if (!server_version.empty())
{
ret += " (";
ret += server_version;
ret += ")";
}
return ret;
}
2014-07-06 21:18:00 +02:00
std::tuple<std::int64_t, std::int64_t> get_range(
http_parser const& parser, error_code& ec)
{
std::int64_t range_start;
std::int64_t range_end;
if (parser.status_code() == 206)
{
std::tie(range_start, range_end) = parser.content_range();
if (range_start < 0 || range_end < range_start)
{
ec = errors::invalid_range;
range_start = 0;
range_end = 0;
}
else
{
// the http range is inclusive
range_end++;
}
}
else
{
range_start = 0;
range_end = parser.content_length();
if (range_end < 0)
{
range_end = 0;
ec = errors::no_content_length;
}
}
return std::tuple<std::int64_t, std::int64_t>(range_start, range_end);
}
2014-07-06 21:18:00 +02:00
}
// --------------------------
// RECEIVE DATA
// --------------------------
2014-07-06 21:18:00 +02:00
bool web_peer_connection::received_invalid_data(int index, bool single_peer)
{
if (!single_peer) return peer_connection::received_invalid_data(index, single_peer);
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
// when a web seed fails a hash check, do the following:
// 1. if the whole piece only overlaps a single file, mark that file as not
// have for this peer
// 2. if the piece overlaps more than one file, mark the piece as not have
// for this peer
// 3. if it's a single file torrent, just ban it right away
// this handles the case where web seeds may have some files updated but not other
2012-09-29 19:46:41 +02:00
std::shared_ptr<torrent> t = associated_torrent().lock();
2014-07-06 21:18:00 +02:00
file_storage const& fs = t->torrent_file().files();
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
// single file torrent
if (fs.num_files() == 1) return peer_connection::received_invalid_data(index, single_peer);
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
std::vector<file_slice> files = fs.map_block(index, 0, fs.piece_size(index));
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
if (files.size() == 1)
{
// assume the web seed has a different copy of this specific file
// than what we expect, and pretend not to have it.
int const fi = files[0].file_index;
int const first_piece = int(fs.file_offset(fi) / fs.piece_length());
2014-07-06 21:18:00 +02:00
// one past last piece
int const end_piece = int((fs.file_offset(fi) + fs.file_size(fi) + 1) / fs.piece_length());
2014-07-06 21:18:00 +02:00
for (int i = first_piece; i < end_piece; ++i)
incoming_dont_have(i);
}
else
{
incoming_dont_have(index);
}
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
peer_connection::received_invalid_data(index, single_peer);
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
// if we don't think we have any of the files, allow banning the web seed
if (num_have_pieces() == 0) return true;
2012-09-29 19:46:41 +02:00
2014-07-06 21:18:00 +02:00
// don't disconnect, we won't request anything from this file again
return false;
}
2012-09-29 19:46:41 +02:00
void web_peer_connection::on_receive_padfile()
{
handle_padfile();
}
void web_peer_connection::handle_error(int const bytes_left)
{
std::shared_ptr<torrent> t = associated_torrent().lock();
TORRENT_ASSERT(t);
// TODO: 2 just make this peer not have the pieces
// associated with the file we just requested. Only
// when it doesn't have any of the file do the following
// pad files will make it complicated
int retry_time = atoi(m_parser.header("retry-after").c_str());
if (retry_time <= 0) retry_time = m_settings.get_int(settings_pack::urlseed_wait_retry);
// temporarily unavailable, retry later
t->retry_web_seed(this, retry_time);
2016-05-01 05:10:47 +02:00
std::string error_msg = to_string(m_parser.status_code()).data()
+ (" " + m_parser.message());
if (t->alerts().should_post<url_seed_alert>())
{
t->alerts().emplace_alert<url_seed_alert>(t->get_handle(), m_url
, error_msg);
}
received_bytes(0, bytes_left);
disconnect(error_code(m_parser.status_code(), http_category()), op_bittorrent, 1);
return;
}
void web_peer_connection::handle_redirect(int const bytes_left)
{
// this means we got a redirection request
// look for the location header
std::string location = m_parser.header("location");
received_bytes(0, bytes_left);
std::shared_ptr<torrent> t = associated_torrent().lock();
TORRENT_ASSERT(t);
if (location.empty())
{
// we should not try this server again.
t->remove_web_seed_conn(this, errors::missing_location, op_bittorrent, 2);
m_web = nullptr;
TORRENT_ASSERT(is_disconnecting());
return;
}
bool const single_file_request = !m_path.empty()
&& m_path[m_path.size() - 1] != '/';
// add the redirected url and remove the current one
if (!single_file_request)
{
TORRENT_ASSERT(!m_file_requests.empty());
int const file_index = m_file_requests.front().file_index;
location = resolve_redirect_location(m_url, location);
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::info, "LOCATION", "%s", location.c_str());
#endif
// TODO: 3 this could be made more efficient for the case when we use an
// HTTP proxy. Then we wouldn't need to add new web seeds to the torrent,
// we could just make the redirect table contain full URLs.
std::string redirect_base;
std::string redirect_path;
error_code ec;
std::tie(redirect_base, redirect_path) = split_url(location, ec);
if (ec)
{
// we should not try this server again.
disconnect(errors::missing_location, op_bittorrent, 1);
return;
}
// add_web_seed won't add duplicates. If we have already added an entry
// with this URL, we'll get back the existing entry
web_seed_t* web = t->add_web_seed(redirect_base, web_seed_entry::url_seed, m_external_auth, m_extra_headers);
web->have_files.resize(t->torrent_file().num_files(), false);
// the new web seed we're adding only has this file for now
// we may add more files later
web->redirects[file_index] = redirect_path;
if (web->have_files.get_bit(file_index) == false)
{
web->have_files.set_bit(file_index);
if (web->peer_info.connection != nullptr)
{
peer_connection* pc = static_cast<peer_connection*>(web->peer_info.connection);
// we just learned that this host has this file, and we're currently
// connected to it. Make it advertise that it has this file to the
// bittorrent engine
file_storage const& fs = t->torrent_file().files();
std::tuple<int, int> const range = aux::file_piece_range_exclusive(fs, file_index);
for (int i = std::get<0>(range); i < std::get<1>(range); ++i)
pc->incoming_have(i);
}
}
// we don't have this file on this server. Don't ask for it again
m_web->have_files.resize(t->torrent_file().num_files(), true);
if (m_web->have_files.get_bit(file_index) == true)
{
m_web->have_files.clear_bit(file_index);
disconnect(errors::redirecting, op_bittorrent, 2);
}
}
else
{
location = resolve_redirect_location(m_url, location);
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::info, "LOCATION", "%s", location.c_str());
#endif
t->add_web_seed(location, web_seed_entry::url_seed, m_external_auth, m_extra_headers);
// this web seed doesn't have any files. Don't try to request from it
// again this session
m_web->have_files.resize(t->torrent_file().num_files(), false);
disconnect(errors::redirecting, op_bittorrent, 2);
m_web = nullptr;
TORRENT_ASSERT(is_disconnecting());
}
return;
}
2014-07-06 21:18:00 +02:00
void web_peer_connection::on_receive(error_code const& error
, std::size_t bytes_transferred)
{
INVARIANT_CHECK;
2014-07-06 21:18:00 +02:00
if (error)
{
received_bytes(0, int(bytes_transferred));
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "ERROR"
, "web_peer_connection error: %s", error.message().c_str());
}
#endif
2014-07-06 21:18:00 +02:00
return;
}
std::shared_ptr<torrent> t = associated_torrent().lock();
2014-07-06 21:18:00 +02:00
TORRENT_ASSERT(t);
// in case the first file on this series of requests is a padfile
// we need to handle it right now
span<char const> recv_buffer = m_recv_buffer.get();
handle_padfile();
if (associated_torrent().expired()) return;
2014-07-06 21:18:00 +02:00
for (;;)
{
int payload;
int protocol;
bool header_finished = m_parser.header_finished();
if (!header_finished)
{
bool failed = false;
std::tie(payload, protocol) = m_parser.incoming(recv_buffer, failed);
2014-07-06 21:18:00 +02:00
received_bytes(0, protocol);
TORRENT_ASSERT(int(recv_buffer.size()) >= protocol);
2014-07-06 21:18:00 +02:00
if (failed)
{
received_bytes(0, int(recv_buffer.size()));
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "RECEIVE_BYTES"
, "%s", std::string(recv_buffer.data(), recv_buffer.size()).c_str());
}
#endif
2014-07-06 21:18:00 +02:00
disconnect(errors::http_parse_error, op_bittorrent, 2);
return;
}
2007-12-29 19:24:50 +01:00
TORRENT_ASSERT(recv_buffer.empty() || recv_buffer[0] == 'H');
TORRENT_ASSERT(int(recv_buffer.size()) <= m_recv_buffer.packet_size());
2015-05-19 05:13:49 +02:00
2014-07-06 21:18:00 +02:00
// this means the entire status line hasn't been received yet
if (m_parser.status_code() == -1)
{
TORRENT_ASSERT(payload == 0);
break;
}
2014-07-06 21:18:00 +02:00
if (!m_parser.header_finished())
{
TORRENT_ASSERT(payload == 0);
break;
}
2014-07-06 21:18:00 +02:00
m_body_start = m_parser.body_start();
m_received_body = 0;
}
// we just completed reading the header
if (!header_finished)
{
++m_num_responses;
if (m_parser.connection_close())
{
2014-07-06 21:18:00 +02:00
incoming_choke();
if (m_num_responses == 1)
m_web->supports_keepalive = false;
2014-07-06 21:18:00 +02:00
}
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "STATUS"
, "%d %s", m_parser.status_code(), m_parser.message().c_str());
std::multimap<std::string, std::string> const& headers = m_parser.headers();
for (std::multimap<std::string, std::string>::const_iterator i = headers.begin()
, end(headers.end()); i != end; ++i)
peer_log(peer_log_alert::info, "STATUS", " %s: %s", i->first.c_str(), i->second.c_str());
}
2014-07-06 21:18:00 +02:00
#endif
2014-07-06 21:18:00 +02:00
// if the status code is not one of the accepted ones, abort
if (!is_ok_status(m_parser.status_code()))
{
file_request_t const& file_req = m_file_requests.front();
m_web->have_files.resize(t->torrent_file().num_files(), true);
m_web->have_files.clear_bit(file_req.file_index);
handle_error(int(recv_buffer.size()));
2014-07-06 21:18:00 +02:00
return;
}
2014-07-06 21:18:00 +02:00
if (is_redirect(m_parser.status_code()))
{
handle_redirect(int(recv_buffer.size()));
return;
}
m_server_string = get_peer_name(m_parser, m_host);
2015-05-28 16:46:12 +02:00
recv_buffer = recv_buffer.subspan(m_body_start);
m_body_start = m_parser.body_start();
m_received_body = 0;
}
// we only received the header, no data
if (recv_buffer.size() == 0) break;
// ===================================
// ======= RESPONSE BYTE RANGE =======
// ===================================
// despite the HTTP range being inclusive, range_start and range_end are
// exclusive to fit better into C++. i.e. range_end points one byte past
// the end of the payload
std::int64_t range_start;
std::int64_t range_end;
error_code ec;
std::tie(range_start, range_end) = get_range(m_parser, ec);
if (ec)
{
received_bytes(0, int(recv_buffer.size()));
// we should not try this server again.
t->remove_web_seed_conn(this, ec, op_bittorrent, 2);
m_web = nullptr;
TORRENT_ASSERT(is_disconnecting());
return;
}
TORRENT_ASSERT(!m_file_requests.empty());
file_request_t const& file_req = m_file_requests.front();
if (range_start != file_req.start
|| range_end != file_req.start + file_req.length)
{
// the byte range in the http response is different what we expected
received_bytes(0, int(recv_buffer.size()));
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::incoming))
{
peer_log(peer_log_alert::incoming, "INVALID HTTP RESPONSE"
, "in=(%d, %" PRId64 "-%" PRId64 ") expected=(%d, %" PRId64 "-%" PRId64 ") ]"
, file_req.file_index, range_start, range_end
, file_req.file_index, file_req.start, file_req.start + file_req.length - 1);
}
#endif
disconnect(errors::invalid_range, op_bittorrent, 2);
return;
}
if (m_parser.chunked_encoding())
{
// =========================
// === CHUNKED ENCODING ===
// =========================
while (m_chunk_pos >= 0 && recv_buffer.size() > 0)
{
// first deliver any payload we have in the buffer so far, ahead of
// the next chunk header.
if (m_chunk_pos > 0)
{
int const copy_size = (std::min)(m_chunk_pos, int(recv_buffer.size()));
TORRENT_ASSERT(copy_size > 0);
if (m_received_body + copy_size > file_req.length)
{
// the byte range in the http response is different what we expected
received_bytes(0, int(recv_buffer.size()));
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming, "INVALID HTTP RESPONSE"
, "received body: %d request size: %d"
, m_received_body, file_req.length);
#endif
disconnect(errors::invalid_range, op_bittorrent, 2);
return;
}
incoming_payload(recv_buffer.data(), copy_size);
recv_buffer = recv_buffer.subspan(copy_size);
m_chunk_pos -= copy_size;
if (recv_buffer.size() == 0) goto done;
}
TORRENT_ASSERT(m_chunk_pos == 0);
int header_size = 0;
std::int64_t chunk_size = 0;
span<char const> chunk_start = recv_buffer.subspan(m_chunk_pos);
TORRENT_ASSERT(chunk_start[0] == '\r'
|| aux::is_hex({chunk_start.data(), 1}));
bool const ret = m_parser.parse_chunk_header(chunk_start, &chunk_size, &header_size);
if (!ret)
{
received_bytes(0, int(chunk_start.size() - m_partial_chunk_header));
m_partial_chunk_header = int(chunk_start.size());
goto done;
}
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::info, "CHUNKED_ENCODING"
, "parsed chunk: %" PRId64 " header_size: %d"
, chunk_size, header_size);
#endif
received_bytes(0, header_size - m_partial_chunk_header);
m_partial_chunk_header = 0;
TORRENT_ASSERT(chunk_size != 0
|| int(chunk_start.size()) <= header_size || chunk_start[header_size] == 'H');
TORRENT_ASSERT(m_body_start + m_chunk_pos < INT_MAX);
m_chunk_pos += chunk_size;
recv_buffer = recv_buffer.subspan(header_size);
// a chunk size of zero means the request is complete. Make sure the
// number of payload bytes we've received matches the number we
// requested. If that's not the case, we got an invalid response.
if (chunk_size == 0)
{
TORRENT_ASSERT_VAL(m_chunk_pos == 0, m_chunk_pos);
#if TORRENT_USE_ASSERTS
span<char const> chunk = recv_buffer.subspan(m_chunk_pos);
TORRENT_ASSERT(chunk.size() == 0 || chunk[0] == 'H');
#endif
m_chunk_pos = -1;
TORRENT_ASSERT(m_received_body <= file_req.length);
if (m_received_body != file_req.length)
{
// the byte range in the http response is different what we expected
received_bytes(0, int(recv_buffer.size()));
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming, "INVALID HTTP RESPONSE"
, "received body: %d request size: %d"
, m_received_body, file_req.length);
#endif
disconnect(errors::invalid_range, op_bittorrent, 2);
return;
}
// we just completed an HTTP file request. pop it from m_file_requests
m_file_requests.pop_front();
m_parser.reset();
m_body_start = 0;
m_received_body = 0;
m_chunk_pos = 0;
m_partial_chunk_header = 0;
// in between each file request, there may be an implicit
// pad-file request
handle_padfile();
break;
}
// if all of the receive buffer was just consumed as chunk
// header, we're done
if (recv_buffer.size() == 0) goto done;
}
}
else
{
// this is the simple case, where we don't have chunked encoding
TORRENT_ASSERT(m_received_body <= file_req.length);
int const copy_size = (std::min)(file_req.length - m_received_body
, int(recv_buffer.size()));
incoming_payload(recv_buffer.data(), copy_size);
recv_buffer = recv_buffer.subspan(copy_size);
TORRENT_ASSERT(m_received_body <= file_req.length);
if (m_received_body == file_req.length)
{
// we just completed an HTTP file request. pop it from m_file_requests
m_file_requests.pop_front();
m_parser.reset();
m_body_start = 0;
m_received_body = 0;
m_chunk_pos = 0;
m_partial_chunk_header = 0;
// in between each file request, there may be an implicit
// pad-file request
handle_padfile();
}
}
if (recv_buffer.size() == 0) break;
}
done:
2009-04-11 22:45:14 +02:00
// now, remove all the bytes we've processed from the receive buffer
m_recv_buffer.cut(int(recv_buffer.data() - m_recv_buffer.get().begin())
, t->block_size() + request_size_overhead);
}
2009-04-11 22:45:14 +02:00
void web_peer_connection::incoming_payload(char const* buf, int len)
{
received_bytes(len, 0);
m_received_body += len;
2009-04-11 22:45:14 +02:00
if (is_disconnecting()) return;
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming_message, "INCOMING_PAYLOAD", "%d bytes", len);
#endif
// deliver all complete bittorrent requests to the bittorrent engine
while (len > 0)
{
if (m_requests.empty()) return;
TORRENT_ASSERT(!m_requests.empty());
peer_request const& front_request = m_requests.front();
int const piece_size = int(m_piece.size());
int const copy_size = (std::min)(front_request.length - piece_size, len);
// m_piece may not hold more than the response to the next BT request
TORRENT_ASSERT(front_request.length > piece_size);
// copy_size is the number of bytes we need to add to the end of m_piece
// to not exceed the size of the next bittorrent request to be delivered.
// m_piece can only hold the response for a single BT request at a time
m_piece.resize(piece_size + copy_size);
std::memcpy(m_piece.data() + piece_size, buf, copy_size);
len -= copy_size;
buf += copy_size;
// keep peer stats up-to-date
incoming_piece_fragment(copy_size);
TORRENT_ASSERT(front_request.length >= piece_size);
if (int(m_piece.size()) == front_request.length)
{
std::shared_ptr<torrent> t = associated_torrent().lock();
TORRENT_ASSERT(t);
2009-04-11 22:45:14 +02:00
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming_message, "POP_REQUEST"
, "piece: %d start: %d len: %d"
, front_request.piece, front_request.start, front_request.length);
#endif
// Make a copy of the request and pop it off the queue before calling
// incoming_piece because that may lead to a call to disconnect()
// which will clear the request queue and invalidate any references
// to the request
peer_request const front_request_copy = front_request;
m_requests.pop_front();
incoming_piece(front_request_copy, m_piece.data());
m_piece.clear();
}
}
}
void web_peer_connection::incoming_zeroes(int len)
{
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming_message, "INCOMING_ZEROES", "%d bytes", len);
#endif
// deliver all complete bittorrent requests to the bittorrent engine
while (len > 0)
{
TORRENT_ASSERT(!m_requests.empty());
peer_request const& front_request = m_requests.front();
int const piece_size = int(m_piece.size());
int const copy_size = (std::min)(front_request.length - piece_size, len);
// m_piece may not hold more than the response to the next BT request
TORRENT_ASSERT(front_request.length > piece_size);
// copy_size is the number of bytes we need to add to the end of m_piece
// to not exceed the size of the next bittorrent request to be delivered.
// m_piece can only hold the response for a single BT request at a time
m_piece.resize(piece_size + copy_size, 0);
len -= copy_size;
// keep peer stats up-to-date
incoming_piece_fragment(copy_size);
maybe_harvest_piece();
}
}
void web_peer_connection::maybe_harvest_piece()
{
peer_request const& front_request = m_requests.front();
TORRENT_ASSERT(front_request.length >= int(m_piece.size()));
if (int(m_piece.size()) != front_request.length) return;
std::shared_ptr<torrent> t = associated_torrent().lock();
TORRENT_ASSERT(t);
#ifndef TORRENT_DISABLE_LOGGING
peer_log(peer_log_alert::incoming_message, "POP_REQUEST"
, "piece: %d start: %d len: %d"
, front_request.piece, front_request.start, front_request.length);
#endif
m_requests.pop_front();
incoming_piece(front_request, m_piece.data());
m_piece.clear();
}
void web_peer_connection::get_specific_peer_info(peer_info& p) const
{
web_connection_base::get_specific_peer_info(p);
p.flags |= peer_info::local_connection;
p.connection_type = peer_info::web_seed;
}
2014-07-06 21:18:00 +02:00
void web_peer_connection::handle_padfile()
{
if (m_file_requests.empty()) return;
if (m_requests.empty()) return;
std::shared_ptr<torrent> t = associated_torrent().lock();
TORRENT_ASSERT(t);
torrent_info const& info = t->torrent_file();
while (!m_file_requests.empty()
&& info.orig_files().pad_file_at(m_file_requests.front().file_index))
{
// the next file is a pad file. We didn't actually send
// a request for this since it most likely doesn't exist on
// the web server anyway. Just pretend that we received a
// bunch of zeroes here and pop it again
std::int64_t file_size = m_file_requests.front().length;
// in theory the pad file can span multiple bocks, hence the loop
while (file_size > 0)
{
peer_request const front_request = m_requests.front();
TORRENT_ASSERT(int(m_piece.size()) < front_request.length);
int pad_size = int(std::min(file_size
, std::int64_t(front_request.length - m_piece.size())));
TORRENT_ASSERT(pad_size > 0);
file_size -= pad_size;
incoming_zeroes(pad_size);
#ifndef TORRENT_DISABLE_LOGGING
if (should_log(peer_log_alert::info))
{
peer_log(peer_log_alert::info, "HANDLE_PADFILE"
, "file: %d start: %" PRId64 " len: %d"
, m_file_requests.front().file_index
, m_file_requests.front().start
, m_file_requests.front().length);
}
#endif
}
m_file_requests.pop_front();
}
}
} // libtorrent namespace