2010-11-29 02:33:05 +01:00
|
|
|
/*
|
|
|
|
|
2018-04-09 09:04:33 +02:00
|
|
|
Copyright (c) 2009-2018, Arvid Norberg
|
2010-11-29 02:33:05 +01:00
|
|
|
All rights reserved.
|
|
|
|
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
|
|
modification, are permitted provided that the following conditions
|
|
|
|
are met:
|
|
|
|
|
|
|
|
* Redistributions of source code must retain the above copyright
|
|
|
|
notice, this list of conditions and the following disclaimer.
|
|
|
|
* Redistributions in binary form must reproduce the above copyright
|
|
|
|
notice, this list of conditions and the following disclaimer in
|
|
|
|
the documentation and/or other materials provided with the distribution.
|
|
|
|
* Neither the name of the author nor the names of its
|
|
|
|
contributors may be used to endorse or promote products derived
|
|
|
|
from this software without specific prior written permission.
|
|
|
|
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
|
|
*/
|
|
|
|
|
2012-06-24 21:50:49 +02:00
|
|
|
#include "libtorrent/config.hpp"
|
2010-11-29 02:33:05 +01:00
|
|
|
#include "libtorrent/utp_stream.hpp"
|
|
|
|
#include "libtorrent/sliding_average.hpp"
|
|
|
|
#include "libtorrent/utp_socket_manager.hpp"
|
2017-01-27 18:43:34 +01:00
|
|
|
#include "libtorrent/aux_/alloca.hpp"
|
2010-11-29 02:33:05 +01:00
|
|
|
#include "libtorrent/timestamp_history.hpp"
|
|
|
|
#include "libtorrent/error.hpp"
|
2011-02-26 08:55:51 +01:00
|
|
|
#include "libtorrent/random.hpp"
|
2012-06-26 05:42:01 +02:00
|
|
|
#include "libtorrent/invariant_check.hpp"
|
2014-07-06 21:18:00 +02:00
|
|
|
#include "libtorrent/performance_counters.hpp"
|
2015-06-06 07:22:53 +02:00
|
|
|
#include "libtorrent/io_service.hpp"
|
2016-06-18 20:01:38 +02:00
|
|
|
#include <cstdint>
|
2014-12-01 11:43:34 +01:00
|
|
|
#include <limits>
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-08-08 02:15:37 +02:00
|
|
|
// the behavior of the sequence numbers as implemented by uTorrent is not
|
|
|
|
// particularly regular. This switch indicates the odd parts.
|
2010-11-29 02:33:05 +01:00
|
|
|
#define TORRENT_UT_SEQ 1
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
2016-07-10 05:17:55 +02:00
|
|
|
#include <cstdarg>
|
2016-05-19 19:25:39 +02:00
|
|
|
#include <cinttypes> // for PRId64 et.al.
|
2010-11-29 02:33:05 +01:00
|
|
|
#include "libtorrent/socket_io.hpp"
|
|
|
|
#endif
|
|
|
|
|
|
|
|
namespace libtorrent {
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
|
2016-08-03 06:35:40 +02:00
|
|
|
static char const* packet_type_names[] = { "ST_DATA", "ST_FIN", "ST_STATE", "ST_RESET", "ST_SYN" };
|
|
|
|
static char const* socket_state_names[] = { "NONE", "SYN_SENT", "CONNECTED", "FIN_SENT", "ERROR", "DELETE" };
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
static struct utp_logger
|
|
|
|
{
|
|
|
|
FILE* utp_log_file;
|
2016-05-01 00:54:23 +02:00
|
|
|
std::mutex utp_log_mutex;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-06-20 17:32:06 +02:00
|
|
|
utp_logger() : utp_log_file(nullptr) {}
|
2010-11-29 02:33:05 +01:00
|
|
|
~utp_logger()
|
|
|
|
{
|
|
|
|
if (utp_log_file) fclose(utp_log_file);
|
|
|
|
}
|
|
|
|
} log_file_holder;
|
|
|
|
|
2015-09-02 07:30:40 +02:00
|
|
|
TORRENT_FORMAT(1, 2)
|
2010-11-29 02:33:05 +01:00
|
|
|
void utp_log(char const* fmt, ...)
|
|
|
|
{
|
2016-06-20 17:32:06 +02:00
|
|
|
if (log_file_holder.utp_log_file == nullptr) return;
|
2015-08-13 02:55:07 +02:00
|
|
|
|
2016-05-01 00:54:23 +02:00
|
|
|
std::lock_guard<std::mutex> lock(log_file_holder.utp_log_mutex);
|
2015-03-12 05:34:54 +01:00
|
|
|
static time_point start = clock_type::now();
|
2016-05-17 15:24:06 +02:00
|
|
|
std::fprintf(log_file_holder.utp_log_file, "[%012" PRId64 "] ", total_microseconds(clock_type::now() - start));
|
2010-11-29 02:33:05 +01:00
|
|
|
va_list l;
|
|
|
|
va_start(l, fmt);
|
|
|
|
vfprintf(log_file_holder.utp_log_file, fmt, l);
|
|
|
|
va_end(l);
|
|
|
|
}
|
|
|
|
|
2015-08-13 03:36:20 +02:00
|
|
|
bool is_utp_stream_logging() {
|
2016-06-20 17:32:06 +02:00
|
|
|
return log_file_holder.utp_log_file != nullptr;
|
2015-08-13 03:36:20 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void set_utp_stream_logging(bool enable) {
|
|
|
|
if (enable)
|
|
|
|
{
|
2016-06-20 17:32:06 +02:00
|
|
|
if (log_file_holder.utp_log_file == nullptr)
|
2015-08-13 03:36:20 +02:00
|
|
|
{
|
|
|
|
log_file_holder.utp_log_file = fopen("utp.log", "w+");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2016-06-20 17:32:06 +02:00
|
|
|
if (log_file_holder.utp_log_file != nullptr)
|
2015-08-13 03:36:20 +02:00
|
|
|
{
|
|
|
|
FILE* f = log_file_holder.utp_log_file;
|
2016-06-20 17:32:06 +02:00
|
|
|
log_file_holder.utp_log_file = nullptr;
|
2015-08-13 03:36:20 +02:00
|
|
|
fclose(f);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
#define UTP_LOG utp_log
|
|
|
|
#if TORRENT_VERBOSE_UTP_LOG
|
|
|
|
#define UTP_LOGV utp_log
|
|
|
|
#else
|
2015-04-27 04:21:12 +02:00
|
|
|
#define UTP_LOGV TORRENT_WHILE_0 printf
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2015-09-02 07:30:40 +02:00
|
|
|
#if __cplusplus >= 201103L || defined __clang__
|
2015-08-02 21:55:05 +02:00
|
|
|
|
|
|
|
#define UTP_LOG(...) do {} while(false)
|
|
|
|
#define UTP_LOGV(...) do {} while(false)
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2015-04-27 04:21:12 +02:00
|
|
|
#define UTP_LOG TORRENT_WHILE_0 printf
|
|
|
|
#define UTP_LOGV TORRENT_WHILE_0 printf
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-08-02 21:55:05 +02:00
|
|
|
#endif // cplusplus
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
enum
|
|
|
|
{
|
|
|
|
ACK_MASK = 0xffff,
|
|
|
|
|
|
|
|
// if a packet receives more than this number of
|
|
|
|
// duplicate acks, we'll trigger a fast re-send
|
|
|
|
dup_ack_limit = 3,
|
|
|
|
|
|
|
|
// the max number of packets to fast-resend per
|
|
|
|
// selective ack message
|
2013-09-12 13:28:54 +02:00
|
|
|
// only re-sending a single packet per sack
|
|
|
|
// appears to improve performance by making it
|
|
|
|
// less likely to loose the re-sent packet. Because
|
|
|
|
// when that happens, we must time-out in order
|
|
|
|
// to continue, which takes a long time.
|
2014-07-05 01:40:31 +02:00
|
|
|
sack_resend_limit = 1
|
2010-11-29 02:33:05 +01:00
|
|
|
};
|
|
|
|
|
|
|
|
// compare if lhs is less than rhs, taking wrapping
|
|
|
|
// into account. if lhs is close to UINT_MAX and rhs
|
|
|
|
// is close to 0, lhs is assumed to have wrapped and
|
|
|
|
// considered smaller
|
2017-04-04 22:42:37 +02:00
|
|
|
bool compare_less_wrap(std::uint32_t lhs
|
2016-06-18 20:01:38 +02:00
|
|
|
, std::uint32_t rhs, std::uint32_t mask)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
// distance walking from lhs to rhs, downwards
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t dist_down = (lhs - rhs) & mask;
|
2010-11-29 02:33:05 +01:00
|
|
|
// distance walking from lhs to rhs, upwards
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t dist_up = (rhs - lhs) & mask;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// if the distance walking up is shorter, lhs
|
|
|
|
// is less than rhs. If the distance walking down
|
|
|
|
// is shorter, then rhs is less than lhs
|
|
|
|
return dist_up < dist_down;
|
|
|
|
}
|
|
|
|
|
|
|
|
// since the uTP socket state may be needed after the
|
|
|
|
// utp_stream is closed, it's kept in a separate struct
|
|
|
|
// whose lifetime is not tied to the lifetime of utp_stream
|
|
|
|
|
|
|
|
// the utp socket is closely modelled after the asio async
|
|
|
|
// operations and handler model. For writing to the socket,
|
|
|
|
// the client provides a list of buffers (for gather/writev
|
|
|
|
// style of I/O) and whenever the socket can write another
|
|
|
|
// packet to the stream, it picks up data from these buffers.
|
|
|
|
// When all of the data has been written, or enough time has
|
|
|
|
// passed since we first started writing, the write handler
|
|
|
|
// is called and the write buffer is reset. This means that
|
|
|
|
// we're not writing anything at all while waiting for the
|
|
|
|
// client to re-issue a write request.
|
|
|
|
|
|
|
|
// reading is a little bit more complicated, since we must
|
|
|
|
// be able to receive data even when the user doesn't have
|
|
|
|
// an outstanding read operation on the socket. When the user
|
|
|
|
// does however, we want to receive data directly into the
|
|
|
|
// user's buffer instead of first copying it into our receive
|
|
|
|
// buffer. This is why the receive case is more complicated.
|
|
|
|
// There are two receive buffers. One provided by the user,
|
|
|
|
// which when present is always used. The other one is used
|
|
|
|
// when the user doesn't have an outstanding read request,
|
|
|
|
// and hence hasn't provided any buffer space to receive into.
|
|
|
|
|
|
|
|
// the user provided read buffer is called "m_read_buffer" and
|
|
|
|
// its size is "m_read_buffer_size". The buffer we spill over
|
|
|
|
// into when the user provided buffer is full or when there
|
|
|
|
// is none, is "m_receive_buffer" and "m_receive_buffer_size"
|
|
|
|
// respectively.
|
|
|
|
|
|
|
|
// in order to know when to trigger the read and write handlers
|
|
|
|
// there are two counters, m_read and m_written, which count
|
|
|
|
// the number of bytes we've stuffed into the user provided
|
|
|
|
// read buffer or written to the stream from the write buffer.
|
|
|
|
// These are used to trigger the handlers if we're written a
|
|
|
|
// large number of bytes. It's also triggered if we're filled
|
|
|
|
// the whole read buffer, or written the entire write buffer.
|
|
|
|
// The last way the handlers can be triggered is if we're read
|
|
|
|
// or written some, and enough time has elapsed since then.
|
|
|
|
|
|
|
|
// when we receive data into m_receive_buffer (i.e. the buffer
|
|
|
|
// used when there's no user provided one) is stored as a
|
|
|
|
// number of heap allocated packets. This is just because it's
|
|
|
|
// simple to reuse the data structured and it provides all the
|
|
|
|
// functionality needed for this buffer.
|
|
|
|
|
|
|
|
struct utp_socket_impl
|
|
|
|
{
|
2016-06-18 20:01:38 +02:00
|
|
|
utp_socket_impl(std::uint16_t recv_id, std::uint16_t send_id
|
2017-02-18 21:40:19 +01:00
|
|
|
, void* userdata, utp_socket_manager& sm)
|
2010-11-29 02:33:05 +01:00
|
|
|
: m_sm(sm)
|
|
|
|
, m_userdata(userdata)
|
2017-02-18 21:40:19 +01:00
|
|
|
, m_timeout(clock_type::now() + milliseconds(m_sm.connect_timeout()))
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_send_id(send_id)
|
|
|
|
, m_recv_id(recv_id)
|
|
|
|
, m_delay_sample_idx(0)
|
|
|
|
, m_state(UTP_STATE_NONE)
|
|
|
|
, m_eof(false)
|
|
|
|
, m_attached(true)
|
|
|
|
, m_nagle(true)
|
2012-06-21 18:41:05 +02:00
|
|
|
, m_slow_start(true)
|
2011-11-05 10:30:38 +01:00
|
|
|
, m_cwnd_full(false)
|
2014-07-06 21:18:00 +02:00
|
|
|
, m_null_buffers(false)
|
2012-06-21 17:05:57 +02:00
|
|
|
, m_deferred_ack(false)
|
2013-02-06 05:38:30 +01:00
|
|
|
, m_subscribe_drained(false)
|
2012-07-01 20:44:46 +02:00
|
|
|
, m_stalled(false)
|
2015-07-19 03:55:26 +02:00
|
|
|
, m_confirmed(false)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2018-02-05 01:29:14 +01:00
|
|
|
TORRENT_ASSERT((m_recv_id == ((m_send_id + 1) & 0xffff))
|
|
|
|
|| (m_send_id == ((m_recv_id + 1) & 0xffff)));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::num_utp_idle);
|
2011-08-24 07:13:25 +02:00
|
|
|
TORRENT_ASSERT(m_userdata);
|
2016-10-08 18:12:33 +02:00
|
|
|
m_delay_sample_hist.fill(std::numeric_limits<std::uint32_t>::max());
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
~utp_socket_impl();
|
|
|
|
|
2015-03-12 05:34:54 +01:00
|
|
|
void tick(time_point now);
|
2010-11-29 02:33:05 +01:00
|
|
|
void init_mtu(int link_mtu, int utp_mtu);
|
2016-07-22 18:31:42 +02:00
|
|
|
bool incoming_packet(span<std::uint8_t const> buf
|
2015-03-12 05:34:54 +01:00
|
|
|
, udp::endpoint const& ep, time_point receive_time);
|
2012-07-01 20:44:46 +02:00
|
|
|
void writable();
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
bool should_delete() const;
|
|
|
|
tcp::endpoint remote_endpoint(error_code& ec) const
|
|
|
|
{
|
|
|
|
if (m_state == UTP_STATE_NONE)
|
2015-06-06 07:22:53 +02:00
|
|
|
ec = boost::asio::error::not_connected;
|
2010-11-29 02:33:05 +01:00
|
|
|
else
|
|
|
|
TORRENT_ASSERT(m_remote_address != address_v4::any());
|
|
|
|
return tcp::endpoint(m_remote_address, m_port);
|
|
|
|
}
|
|
|
|
std::size_t available() const;
|
2011-07-14 05:46:03 +02:00
|
|
|
// returns true if there were handlers cancelled
|
|
|
|
// if it returns false, we can detach immediately
|
|
|
|
bool destroy();
|
2017-02-04 02:27:31 +01:00
|
|
|
void set_close_reason(close_reason_t code);
|
2010-11-29 02:33:05 +01:00
|
|
|
void detach();
|
|
|
|
void send_syn();
|
|
|
|
void send_fin();
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
void subscribe_drained();
|
2012-06-21 17:05:57 +02:00
|
|
|
void defer_ack();
|
2012-06-25 08:27:37 +02:00
|
|
|
void remove_sack_header(packet* p);
|
2012-06-28 05:53:52 +02:00
|
|
|
|
|
|
|
enum packet_flags_t { pkt_ack = 1, pkt_fin = 2 };
|
|
|
|
bool send_pkt(int flags = 0);
|
2010-11-29 02:33:05 +01:00
|
|
|
bool resend_packet(packet* p, bool fast_resend = false);
|
2015-08-19 01:39:01 +02:00
|
|
|
void send_reset(utp_header const* ph);
|
2017-02-18 19:16:55 +01:00
|
|
|
std::pair<std::uint32_t, int> parse_sack(std::uint16_t packet_ack, std::uint8_t const* ptr
|
2017-08-13 03:24:15 +02:00
|
|
|
, int size, time_point now);
|
2016-06-18 20:01:38 +02:00
|
|
|
void parse_close_reason(std::uint8_t const* ptr, int size);
|
|
|
|
void write_payload(std::uint8_t* ptr, int size);
|
2010-12-05 04:03:56 +01:00
|
|
|
void maybe_inc_acked_seq_nr();
|
2017-02-18 19:16:55 +01:00
|
|
|
std::uint32_t ack_packet(packet_ptr p, time_point const& receive_time
|
|
|
|
, std::uint16_t seq_nr);
|
2016-06-18 20:01:38 +02:00
|
|
|
void write_sack(std::uint8_t* buf, int size) const;
|
2017-02-11 20:21:48 +01:00
|
|
|
void incoming(std::uint8_t const* buf, int size, packet_ptr p, time_point now);
|
2014-07-07 08:28:48 +02:00
|
|
|
void do_ledbat(int acked_bytes, int delay, int in_flight);
|
2010-11-29 02:33:05 +01:00
|
|
|
int packet_timeout() const;
|
|
|
|
bool test_socket_state();
|
2013-02-06 05:38:30 +01:00
|
|
|
void maybe_trigger_receive_callback();
|
|
|
|
void maybe_trigger_send_callback();
|
2017-09-17 13:12:54 +02:00
|
|
|
bool cancel_handlers(error_code const& ec, bool shutdown);
|
2010-11-29 02:33:05 +01:00
|
|
|
bool consume_incoming_data(
|
2016-06-18 20:01:38 +02:00
|
|
|
utp_header const* ph, std::uint8_t const* ptr, int payload_size, time_point now);
|
2010-11-29 02:33:05 +01:00
|
|
|
void update_mtu_limits();
|
2017-02-26 18:08:35 +01:00
|
|
|
void experienced_loss(std::uint32_t seq_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-01-04 22:31:02 +01:00
|
|
|
void set_state(int s);
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
packet_ptr acquire_packet(int const allocate) { return m_sm.acquire_packet(allocate); }
|
|
|
|
void release_packet(packet_ptr p) { m_sm.release_packet(std::move(p)); }
|
2017-02-10 14:25:03 +01:00
|
|
|
|
2015-01-04 22:31:02 +01:00
|
|
|
private:
|
|
|
|
|
|
|
|
// non-copyable
|
2017-08-13 03:24:15 +02:00
|
|
|
utp_socket_impl(utp_socket_impl const&) = delete;
|
|
|
|
utp_socket_impl const& operator=(utp_socket_impl const&) = delete;
|
2015-01-04 22:31:02 +01:00
|
|
|
|
|
|
|
// TODO: 2 it would be nice if not everything would have to be public here
|
|
|
|
public:
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
void check_receive_buffers() const;
|
|
|
|
|
2014-01-21 20:26:09 +01:00
|
|
|
#if TORRENT_USE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
void check_invariant() const;
|
|
|
|
#endif
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
utp_socket_manager& m_sm;
|
2017-04-26 05:00:41 +02:00
|
|
|
std::weak_ptr<utp_socket_interface> m_sock;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// userdata pointer passed along
|
|
|
|
// with any callback. This is initialized to 0
|
|
|
|
// then set to point to the utp_stream when
|
|
|
|
// hooked up, and then reset to 0 once the utp_stream
|
|
|
|
// detaches. This is used to know whether or not
|
|
|
|
// the socket impl is still attached to a utp_stream
|
|
|
|
// object. When it isn't, we'll never be able to
|
|
|
|
// signal anything back to the client, and in case
|
|
|
|
// of errors, we just have to delete ourselves
|
|
|
|
// i.e. transition to the UTP_STATE_DELETED state
|
|
|
|
void* m_userdata;
|
|
|
|
|
|
|
|
// This is a platform-independent replacement
|
|
|
|
// for the regular iovec type in posix. Since
|
|
|
|
// it's not used in any system call, we might as
|
|
|
|
// well define our own type instead of wrapping
|
|
|
|
// the system's type.
|
|
|
|
struct iovec_t
|
|
|
|
{
|
2017-08-13 03:24:15 +02:00
|
|
|
iovec_t(void* b, std::size_t l): buf(b), len(l) {}
|
2010-11-29 02:33:05 +01:00
|
|
|
void* buf;
|
2017-02-18 12:52:46 +01:00
|
|
|
std::size_t len;
|
2010-11-29 02:33:05 +01:00
|
|
|
};
|
|
|
|
|
|
|
|
// if there's currently an async read or write
|
|
|
|
// operation in progress, these buffers are initialized
|
|
|
|
// and used, otherwise any bytes received are stuck in
|
|
|
|
// m_receive_buffer until another read is made
|
|
|
|
// as we flush from the write buffer, individual iovecs
|
|
|
|
// are updated to only refer to unflushed portions of the
|
|
|
|
// buffers. Buffers that empty are erased from the vector.
|
|
|
|
std::vector<iovec_t> m_write_buffer;
|
|
|
|
|
2016-06-20 17:32:06 +02:00
|
|
|
// if this is non nullptr, it's a packet. This packet was held off because
|
2012-06-25 08:27:37 +02:00
|
|
|
// of NAGLE. We couldn't send it immediately. It's left
|
|
|
|
// here to accrue more bytes before we send it.
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr m_nagle_packet;
|
2012-06-25 08:27:37 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the user provided read buffer. If this has a size greater
|
|
|
|
// than 0, we'll always prefer using it over putting received
|
|
|
|
// data in the m_receive_buffer. As data is stored in the
|
|
|
|
// read buffer, the iovec_t elements are adjusted to only
|
|
|
|
// refer to the unwritten portions of the buffers, and the
|
|
|
|
// ones that fill up are erased from the vector
|
|
|
|
std::vector<iovec_t> m_read_buffer;
|
2015-07-19 03:55:26 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// packets we've received without a read operation
|
|
|
|
// active. Store them here until the client triggers
|
|
|
|
// an async_read_some
|
2017-02-11 20:21:48 +01:00
|
|
|
std::vector<packet_ptr> m_receive_buffer;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the error on this socket. If m_state is
|
|
|
|
// set to UTP_STATE_ERROR_WAIT, this error should be
|
|
|
|
// forwarded to the client as soon as we have a new
|
|
|
|
// async operation initiated
|
|
|
|
error_code m_error;
|
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
// these indicate whether or not there is an outstanding read/write or
|
|
|
|
// connect operation. i.e. is there upper layer subscribed to these events.
|
2016-10-08 18:12:33 +02:00
|
|
|
bool m_read_handler = false;
|
|
|
|
bool m_write_handler = false;
|
|
|
|
bool m_connect_handler = false;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the address of the remote endpoint
|
|
|
|
address m_remote_address;
|
|
|
|
|
|
|
|
// the send and receive buffers
|
|
|
|
// maps packet sequence numbers
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_buffer m_inbuf;
|
|
|
|
packet_buffer m_outbuf;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the time when the last packet we sent times out. Including re-sends.
|
|
|
|
// if we ever end up not having sent anything in one second (
|
|
|
|
// or one mean rtt + 2 average deviations, whichever is greater)
|
|
|
|
// we set our cwnd to 1 MSS. This condition can happen either because
|
|
|
|
// a packet has timed out and needs to be resent or because our
|
|
|
|
// cwnd is set to less than one MSS during congestion control.
|
2016-04-26 06:37:47 +02:00
|
|
|
// it can also happen if the other end sends an advertised window
|
2010-11-29 02:33:05 +01:00
|
|
|
// size less than one MSS.
|
2015-03-12 05:34:54 +01:00
|
|
|
time_point m_timeout;
|
2015-07-20 17:20:17 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the last time we stepped the timestamp history
|
2016-10-08 18:12:33 +02:00
|
|
|
time_point m_last_history_step = clock_type::now();
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the max number of bytes in-flight. This is a fixed point
|
|
|
|
// value, to get the true number of bytes, shift right 16 bits
|
|
|
|
// the value is always >= 0, but the calculations performed on
|
2011-11-05 10:30:38 +01:00
|
|
|
// it in do_ledbat() are signed.
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int64_t m_cwnd = TORRENT_ETHERNET_MTU << 16;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
timestamp_history m_delay_hist;
|
|
|
|
timestamp_history m_their_delay_hist;
|
|
|
|
|
2014-08-20 09:02:03 +02:00
|
|
|
// the slow-start threshold. This is the congestion window size (m_cwnd)
|
|
|
|
// in bytes the last time we left slow-start mode. This is used as a
|
|
|
|
// threshold to leave slow-start earlier next time, to avoid packet-loss
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_ssthres = 0;
|
2014-08-20 09:02:03 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the number of bytes we have buffered in m_inbuf
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_buffered_incoming_bytes = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the timestamp diff in the last packet received
|
|
|
|
// this is what we'll send back
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint32_t m_reply_micro = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-04-04 04:30:56 +02:00
|
|
|
// this is the advertised receive window the other end sent
|
2010-11-29 02:33:05 +01:00
|
|
|
// we'll never have more un-acked bytes in flight
|
|
|
|
// if this ever gets set to zero, we'll try one packet every
|
|
|
|
// second until the window opens up again
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint32_t m_adv_wnd = TORRENT_ETHERNET_MTU;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the number of un-acked bytes we have sent
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_bytes_in_flight = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the number of bytes read into the user provided
|
|
|
|
// buffer. If this grows too big, we'll trigger the
|
|
|
|
// read handler.
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_read = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the sum of the lengths of all iovec in m_write_buffer
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_write_buffer_size = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the number of bytes already written to packets
|
|
|
|
// from m_write_buffer
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_written = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the sum of all packets stored in m_receive_buffer
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_receive_buffer_size = 0;
|
2015-07-19 03:55:26 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the sum of all buffers in m_read_buffer
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_read_buffer_size = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// max number of bytes to allocate for receive buffer
|
2018-04-12 16:21:20 +02:00
|
|
|
std::int32_t m_receive_buffer_capacity = 1024 * 1024;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this holds the 3 last delay measurements,
|
|
|
|
// these are the actual corrected delay measurements.
|
|
|
|
// the lowest of the 3 last ones is used in the congestion
|
|
|
|
// controller. This is to not completely close the cwnd
|
|
|
|
// by a single outlier.
|
2016-10-08 18:12:33 +02:00
|
|
|
std::array<std::uint32_t, 3> m_delay_sample_hist;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// counters
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint32_t m_in_packets = 0;
|
|
|
|
std::uint32_t m_out_packets = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2011-06-21 09:44:13 +02:00
|
|
|
// the last send delay sample
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_send_delay = 0;
|
2011-06-21 09:44:13 +02:00
|
|
|
// the last receive delay sample
|
2016-10-08 18:12:33 +02:00
|
|
|
std::int32_t m_recv_delay = 0;
|
2011-06-21 09:44:13 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// average RTT
|
|
|
|
sliding_average<16> m_rtt;
|
|
|
|
|
2015-02-28 20:51:15 +01:00
|
|
|
// if this is != 0, it means the upper layer provided a reason for why
|
|
|
|
// the connection is being closed. The reason is indicated by this
|
|
|
|
// non-zero value which is included in a packet header extension
|
2017-02-04 02:27:31 +01:00
|
|
|
close_reason_t m_close_reason = close_reason_t::none;
|
2015-02-28 20:51:15 +01:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// port of destination endpoint
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_port = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint16_t m_send_id;
|
|
|
|
std::uint16_t m_recv_id;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the ack we're sending back. We have
|
|
|
|
// received all packets up to this sequence number
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_ack_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the sequence number of the next packet
|
|
|
|
// we'll send
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_seq_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the sequence number of the packet that
|
|
|
|
// everything has been ACKed up to. Everything we've
|
|
|
|
// sent up to this point has been received by the other
|
|
|
|
// end.
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_acked_seq_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// each packet gets one chance of "fast resend". i.e.
|
|
|
|
// if we have multiple duplicate acks, we may send a
|
|
|
|
// packet immediately, if m_fast_resend_seq_nr is set
|
|
|
|
// to that packet's sequence number
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_fast_resend_seq_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the sequence number of the FIN packet
|
|
|
|
// we've received. This sequence number is only
|
|
|
|
// valid if m_eof is true. We should not accept
|
|
|
|
// any packets beyond this sequence number from the
|
|
|
|
// other end
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_eof_seq_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the lowest sequence number that, when lost,
|
|
|
|
// will cause the window size to be cut in half
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_loss_seq_nr = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the max number of bytes we can send in a packet
|
|
|
|
// including the header
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_mtu = TORRENT_ETHERNET_MTU - TORRENT_IPV4_HEADER - TORRENT_UDP_HEADER - 8 - 24 - 36;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the floor is the largest packet that we have
|
|
|
|
// been able to get through without fragmentation
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_mtu_floor = TORRENT_INET_MIN_MTU - TORRENT_IPV4_HEADER - TORRENT_UDP_HEADER;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the ceiling is the largest packet that we might
|
|
|
|
// be able to get through without fragmentation.
|
|
|
|
// i.e. ceiling +1 is very likely to not get through
|
|
|
|
// or we have in fact experienced a drop or ICMP
|
|
|
|
// message indicating that it is
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_mtu_ceiling = TORRENT_ETHERNET_MTU - TORRENT_IPV4_HEADER - TORRENT_UDP_HEADER;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the sequence number of the probe in-flight
|
|
|
|
// this is 0 if there is no probe in flight
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint16_t m_mtu_seq = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is a counter of how many times the current m_acked_seq_nr
|
|
|
|
// has been ACKed. If it's ACKed more than 3 times, we assume the
|
|
|
|
// packet with the next sequence number has been lost, and we trigger
|
2016-04-04 04:30:56 +02:00
|
|
|
// a re-send. Obviously an ACK only counts as a duplicate as long as
|
2010-11-29 02:33:05 +01:00
|
|
|
// we have outstanding packets following it.
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint8_t m_duplicate_acks = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the number of packet timeouts we've seen in a row
|
|
|
|
// this affects the packet timeout time
|
2016-10-08 18:12:33 +02:00
|
|
|
std::uint8_t m_num_timeouts = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-01-04 22:31:02 +01:00
|
|
|
// it's important that these match the enums in performance_counters for
|
|
|
|
// num_utp_idle etc.
|
2010-11-29 02:33:05 +01:00
|
|
|
enum state_t {
|
|
|
|
// not yet connected
|
|
|
|
UTP_STATE_NONE,
|
|
|
|
// sent a syn packet, not received any acks
|
|
|
|
UTP_STATE_SYN_SENT,
|
|
|
|
// syn-ack received and in normal operation
|
|
|
|
// of sending and receiving data
|
|
|
|
UTP_STATE_CONNECTED,
|
|
|
|
// fin sent, but all packets up to the fin packet
|
|
|
|
// have not yet been acked. We might still be waiting
|
|
|
|
// for a FIN from the other end
|
|
|
|
UTP_STATE_FIN_SENT,
|
|
|
|
|
|
|
|
// ====== states beyond this point =====
|
|
|
|
// === are considered closing states ===
|
|
|
|
// === and will cause the socket to ====
|
|
|
|
// ============ be deleted =============
|
|
|
|
|
|
|
|
// the socket has been gracefully disconnected
|
|
|
|
// and is waiting for the client to make a
|
|
|
|
// socket call so that we can communicate this
|
|
|
|
// fact and actually delete all the state, or
|
|
|
|
// there is an error on this socket and we're
|
|
|
|
// waiting to communicate this to the client in
|
|
|
|
// a callback. The error in either case is stored
|
|
|
|
// in m_error. If the socket has gracefully shut
|
|
|
|
// down, the error is error::eof.
|
|
|
|
UTP_STATE_ERROR_WAIT,
|
|
|
|
|
|
|
|
// there are no more references to this socket
|
|
|
|
// and we can delete it
|
|
|
|
UTP_STATE_DELETE
|
|
|
|
};
|
|
|
|
|
|
|
|
// this is the cursor into m_delay_sample_hist
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t m_delay_sample_idx:2;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the state the socket is in
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t m_state:3;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is set to true when we receive a fin
|
|
|
|
bool m_eof:1;
|
|
|
|
|
|
|
|
// is this socket state attached to a user space socket?
|
|
|
|
bool m_attached:1;
|
|
|
|
|
|
|
|
// this is true if nagle is enabled (which it is by default)
|
|
|
|
bool m_nagle:1;
|
2011-11-05 10:30:38 +01:00
|
|
|
|
|
|
|
// this is true while the socket is in slow start mode. It's
|
2012-06-21 18:41:05 +02:00
|
|
|
// only in slow-start during the start-up phase. Slow start
|
|
|
|
// (contrary to what its name suggest) means that we're growing
|
2016-04-04 04:30:56 +02:00
|
|
|
// the congestion window (cwnd) exponentially rather than linearly.
|
2012-06-21 18:41:05 +02:00
|
|
|
// this is done at startup of a socket in order to find its
|
|
|
|
// link capacity faster. This behaves similar to TCP slow start
|
2011-11-05 10:30:38 +01:00
|
|
|
bool m_slow_start:1;
|
2015-07-20 17:20:17 +02:00
|
|
|
|
2011-11-05 10:30:38 +01:00
|
|
|
// this is true as long as we have as many packets in
|
|
|
|
// flight as allowed by the congestion window (cwnd)
|
|
|
|
bool m_cwnd_full:1;
|
2012-06-21 17:05:57 +02:00
|
|
|
|
2014-07-06 21:18:00 +02:00
|
|
|
// this is set to one if the current read operation
|
2016-06-20 17:32:06 +02:00
|
|
|
// has a null_buffer. i.e. we're not reading into a user-provided
|
2014-07-06 21:18:00 +02:00
|
|
|
// buffer, we're just signalling when there's something
|
|
|
|
// to read from our internal receive buffer
|
|
|
|
bool m_null_buffers:1;
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
// this is set to true when this socket has added itself to
|
|
|
|
// the utp socket manager's list of deferred acks. Once the
|
|
|
|
// burst of incoming UDP packets is all drained, the utp socket
|
|
|
|
// manager will send acks for all sockets on this list.
|
|
|
|
bool m_deferred_ack:1;
|
2012-07-01 20:44:46 +02:00
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
// this is true if this socket has subscribed to be notified
|
|
|
|
// when this receive round is done
|
|
|
|
bool m_subscribe_drained:1;
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
// if this socket tries to send a packet via the utp socket
|
|
|
|
// manager, and it fails with EWOULDBLOCK, the socket
|
|
|
|
// is stalled and this is set. It's also added to a list
|
|
|
|
// of sockets in the utp_socket_manager to be notified of
|
|
|
|
// the socket being writable again
|
|
|
|
bool m_stalled:1;
|
2015-07-19 03:55:26 +02:00
|
|
|
|
|
|
|
// this is false by default and set to true once we've received a non-SYN
|
|
|
|
// packet for this connection with a correct ack_nr, confirming that the
|
|
|
|
// other end is not spoofing its source IP
|
|
|
|
bool m_confirmed:1;
|
2010-11-29 02:33:05 +01:00
|
|
|
};
|
|
|
|
|
2016-06-18 20:01:38 +02:00
|
|
|
utp_socket_impl* construct_utp_impl(std::uint16_t recv_id
|
|
|
|
, std::uint16_t send_id, void* userdata
|
2017-02-18 21:40:19 +01:00
|
|
|
, utp_socket_manager& sm)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
return new utp_socket_impl(recv_id, send_id, userdata, sm);
|
|
|
|
}
|
|
|
|
|
|
|
|
void detach_utp_impl(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
s->detach();
|
|
|
|
}
|
|
|
|
|
|
|
|
void delete_utp_impl(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
delete s;
|
|
|
|
}
|
|
|
|
|
2017-04-26 05:00:41 +02:00
|
|
|
void utp_abort(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
s->m_error = boost::asio::error::connection_aborted;
|
|
|
|
s->set_state(utp_socket_impl::UTP_STATE_ERROR_WAIT);
|
|
|
|
s->test_socket_state();
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
bool should_delete(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
return s->should_delete();
|
|
|
|
}
|
|
|
|
|
2017-04-26 05:00:41 +02:00
|
|
|
bool bound_to_udp_socket(utp_socket_impl* s, std::weak_ptr<utp_socket_interface> sock)
|
|
|
|
{
|
|
|
|
return s->m_sock.lock() == sock.lock();
|
|
|
|
}
|
|
|
|
|
2015-03-12 05:34:54 +01:00
|
|
|
void tick_utp_impl(utp_socket_impl* s, time_point now)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
s->tick(now);
|
|
|
|
}
|
|
|
|
|
|
|
|
void utp_init_mtu(utp_socket_impl* s, int link_mtu, int utp_mtu)
|
|
|
|
{
|
|
|
|
s->init_mtu(link_mtu, utp_mtu);
|
|
|
|
}
|
|
|
|
|
2017-04-26 05:00:41 +02:00
|
|
|
void utp_init_socket(utp_socket_impl* s, std::weak_ptr<utp_socket_interface> sock)
|
|
|
|
{
|
2017-08-13 03:24:15 +02:00
|
|
|
s->m_sock = std::move(sock);
|
2017-04-26 05:00:41 +02:00
|
|
|
}
|
|
|
|
|
2016-04-28 14:20:10 +02:00
|
|
|
bool utp_incoming_packet(utp_socket_impl* s
|
2016-07-22 18:31:42 +02:00
|
|
|
, span<char const> p
|
2017-02-18 12:52:46 +01:00
|
|
|
, udp::endpoint const& ep, time_point const receive_time)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2016-12-27 07:45:48 +01:00
|
|
|
return s->incoming_packet({reinterpret_cast<std::uint8_t const*>(p.data()), p.size()}
|
2015-08-19 01:39:01 +02:00
|
|
|
, ep, receive_time);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
bool utp_match(utp_socket_impl* s, udp::endpoint const& ep, std::uint16_t const id)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2016-10-21 02:49:59 +02:00
|
|
|
return s->m_recv_id == id
|
2010-11-29 02:33:05 +01:00
|
|
|
&& s->m_port == ep.port()
|
2016-10-21 02:49:59 +02:00
|
|
|
&& s->m_remote_address == ep.address();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
udp::endpoint utp_remote_endpoint(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
return udp::endpoint(s->m_remote_address, s->m_port);
|
|
|
|
}
|
|
|
|
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint16_t utp_receive_id(utp_socket_impl* s)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
return s->m_recv_id;
|
|
|
|
}
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
void utp_writable(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(s->m_stalled);
|
|
|
|
s->m_stalled = false;
|
|
|
|
s->writable();
|
|
|
|
}
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
void utp_send_ack(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(s->m_deferred_ack);
|
|
|
|
s->m_deferred_ack = false;
|
2012-06-28 05:53:52 +02:00
|
|
|
s->send_pkt(utp_socket_impl::pkt_ack);
|
2012-06-21 17:05:57 +02:00
|
|
|
}
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
void utp_socket_drained(utp_socket_impl* s)
|
|
|
|
{
|
|
|
|
s->m_subscribe_drained = false;
|
|
|
|
|
|
|
|
// at this point, we know we won't receive any
|
|
|
|
// more packets this round. So, we may want to
|
|
|
|
// call the receive callback function to
|
|
|
|
// let the user consume it
|
|
|
|
|
|
|
|
s->maybe_trigger_receive_callback();
|
|
|
|
s->maybe_trigger_send_callback();
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
void utp_socket_impl::update_mtu_limits()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
if (m_mtu_floor > m_mtu_ceiling) m_mtu_floor = m_mtu_ceiling;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
m_mtu = (m_mtu_floor + m_mtu_ceiling) / 2;
|
|
|
|
|
2016-10-08 20:17:51 +02:00
|
|
|
if ((m_cwnd >> 16) < m_mtu) m_cwnd = std::int64_t(m_mtu) * (1 << 16);
|
2012-08-13 18:49:28 +02:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: updating MTU to: %d [%d, %d]\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_mtu, m_mtu_floor, m_mtu_ceiling);
|
2012-08-13 18:49:28 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// clear the mtu probe sequence number since
|
|
|
|
// it was either dropped or acked
|
|
|
|
m_mtu_seq = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int utp_socket_state(utp_socket_impl const* s)
|
|
|
|
{
|
|
|
|
return s->m_state;
|
|
|
|
}
|
|
|
|
|
2011-06-21 09:44:13 +02:00
|
|
|
int utp_stream::send_delay() const
|
|
|
|
{
|
|
|
|
return m_impl ? m_impl->m_send_delay : 0;
|
|
|
|
}
|
|
|
|
|
2011-09-28 02:03:12 +02:00
|
|
|
int utp_stream::recv_delay() const
|
|
|
|
{
|
|
|
|
return m_impl ? m_impl->m_recv_delay : 0;
|
|
|
|
}
|
|
|
|
|
2015-06-06 07:22:53 +02:00
|
|
|
utp_stream::utp_stream(io_service& io_service)
|
2010-11-29 02:33:05 +01:00
|
|
|
: m_io_service(io_service)
|
2016-07-09 22:26:26 +02:00
|
|
|
, m_impl(nullptr)
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_open(false)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
utp_socket_impl* utp_stream::get_impl()
|
|
|
|
{
|
|
|
|
return m_impl;
|
|
|
|
}
|
|
|
|
|
2017-02-04 02:27:31 +01:00
|
|
|
void utp_stream::set_close_reason(close_reason_t code)
|
2015-02-28 20:51:15 +01:00
|
|
|
{
|
|
|
|
if (!m_impl) return;
|
|
|
|
m_impl->set_close_reason(code);
|
|
|
|
}
|
|
|
|
|
2017-02-04 02:27:31 +01:00
|
|
|
close_reason_t utp_stream::get_close_reason()
|
2015-02-28 20:51:15 +01:00
|
|
|
{
|
|
|
|
return m_incoming_close_reason;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
void utp_stream::close()
|
|
|
|
{
|
|
|
|
if (!m_impl) return;
|
2011-07-14 05:46:03 +02:00
|
|
|
if (!m_impl->destroy())
|
|
|
|
{
|
2011-08-10 07:23:59 +02:00
|
|
|
if (!m_impl) return;
|
2011-07-14 05:46:03 +02:00
|
|
|
detach_utp_impl(m_impl);
|
2016-07-09 22:26:26 +02:00
|
|
|
m_impl = nullptr;
|
2011-07-14 05:46:03 +02:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
std::size_t utp_stream::available() const
|
|
|
|
{
|
2014-07-06 21:18:00 +02:00
|
|
|
return m_impl ? m_impl->available() : 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
utp_stream::endpoint_type utp_stream::remote_endpoint(error_code& ec) const
|
|
|
|
{
|
|
|
|
if (!m_impl)
|
|
|
|
{
|
2015-06-06 07:22:53 +02:00
|
|
|
ec = boost::asio::error::not_connected;
|
2010-11-29 02:33:05 +01:00
|
|
|
return endpoint_type();
|
|
|
|
}
|
|
|
|
return m_impl->remote_endpoint(ec);
|
|
|
|
}
|
|
|
|
|
|
|
|
utp_stream::endpoint_type utp_stream::local_endpoint(error_code& ec) const
|
|
|
|
{
|
2017-02-18 21:40:19 +01:00
|
|
|
if (m_impl == nullptr)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-06-06 07:22:53 +02:00
|
|
|
ec = boost::asio::error::not_connected;
|
2017-04-26 05:00:41 +02:00
|
|
|
return endpoint_type();
|
|
|
|
}
|
|
|
|
|
|
|
|
auto s = m_impl->m_sock.lock();
|
|
|
|
if (!s)
|
|
|
|
{
|
|
|
|
ec = boost::asio::error::not_connected;
|
|
|
|
return endpoint_type();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2017-04-26 05:00:41 +02:00
|
|
|
|
|
|
|
udp::endpoint ep = s->local_endpoint();
|
|
|
|
return endpoint_type(ep.address(), ep.port());
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
utp_stream::~utp_stream()
|
|
|
|
{
|
|
|
|
if (m_impl)
|
|
|
|
{
|
2015-09-06 23:29:01 +02:00
|
|
|
UTP_LOGV("%8p: utp_stream destructed\n", static_cast<void*>(m_impl));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_impl->destroy();
|
|
|
|
detach_utp_impl(m_impl);
|
|
|
|
}
|
|
|
|
|
2016-07-09 22:26:26 +02:00
|
|
|
m_impl = nullptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
void utp_stream::set_impl(utp_socket_impl* impl)
|
|
|
|
{
|
2016-07-09 22:26:26 +02:00
|
|
|
TORRENT_ASSERT(m_impl == nullptr);
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(!m_open);
|
|
|
|
m_impl = impl;
|
|
|
|
m_open = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
int utp_stream::read_buffer_size() const
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(m_impl);
|
|
|
|
return m_impl->m_receive_buffer_size;
|
|
|
|
}
|
|
|
|
|
2017-02-04 02:27:31 +01:00
|
|
|
void utp_stream::on_close_reason(void* self, close_reason_t reason)
|
2015-02-28 20:51:15 +01:00
|
|
|
{
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* s = static_cast<utp_stream*>(self);
|
2016-03-03 04:30:02 +01:00
|
|
|
|
|
|
|
// it's possible the socket has been unlinked already, in which case m_impl
|
2016-06-20 17:32:06 +02:00
|
|
|
// will be nullptr
|
2016-03-03 04:30:02 +01:00
|
|
|
if (s->m_impl)
|
2017-02-04 02:27:31 +01:00
|
|
|
s->m_incoming_close_reason = reason;
|
2015-02-28 20:51:15 +01:00
|
|
|
}
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
void utp_stream::on_read(void* self, std::size_t const bytes_transferred
|
2017-09-17 13:12:54 +02:00
|
|
|
, error_code const& ec, bool const shutdown)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* s = static_cast<utp_stream*>(self);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
UTP_LOGV("%8p: calling read handler read:%d ec:%s shutdown:%d\n", static_cast<void*>(s->m_impl)
|
|
|
|
, int(bytes_transferred), ec.message().c_str(), shutdown);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(s->m_read_handler);
|
2014-07-06 21:18:00 +02:00
|
|
|
TORRENT_ASSERT(bytes_transferred > 0 || ec || s->m_impl->m_null_buffers);
|
2017-01-21 01:03:20 +01:00
|
|
|
s->m_io_service.post(std::bind<void>(std::move(s->m_read_handler), ec, bytes_transferred));
|
2016-08-13 03:31:55 +02:00
|
|
|
s->m_read_handler = nullptr;
|
2017-09-17 13:12:54 +02:00
|
|
|
if (shutdown && s->m_impl)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-02-15 01:25:56 +01:00
|
|
|
TORRENT_ASSERT(ec);
|
2010-11-29 02:33:05 +01:00
|
|
|
detach_utp_impl(s->m_impl);
|
2016-07-09 22:26:26 +02:00
|
|
|
s->m_impl = nullptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
void utp_stream::on_write(void* self, std::size_t const bytes_transferred
|
2017-09-17 13:12:54 +02:00
|
|
|
, error_code const& ec, bool const shutdown)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* s = static_cast<utp_stream*>(self);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
UTP_LOGV("%8p: calling write handler written:%d ec:%s shutdown:%d\n"
|
2015-09-06 23:29:01 +02:00
|
|
|
, static_cast<void*>(s->m_impl)
|
2017-09-17 13:12:54 +02:00
|
|
|
, int(bytes_transferred), ec.message().c_str(), shutdown);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(s->m_write_handler);
|
|
|
|
TORRENT_ASSERT(bytes_transferred > 0 || ec);
|
2017-01-21 01:03:20 +01:00
|
|
|
s->m_io_service.post(std::bind<void>(std::move(s->m_write_handler), ec, bytes_transferred));
|
2016-08-13 03:31:55 +02:00
|
|
|
s->m_write_handler = nullptr;
|
2017-09-17 13:12:54 +02:00
|
|
|
if (shutdown && s->m_impl)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-02-15 01:25:56 +01:00
|
|
|
TORRENT_ASSERT(ec);
|
2010-11-29 02:33:05 +01:00
|
|
|
detach_utp_impl(s->m_impl);
|
2016-07-09 22:26:26 +02:00
|
|
|
s->m_impl = nullptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
void utp_stream::on_connect(void* self, error_code const& ec, bool shutdown)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* s = static_cast<utp_stream*>(self);
|
2011-08-24 07:13:25 +02:00
|
|
|
TORRENT_ASSERT(s);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
UTP_LOGV("%8p: calling connect handler ec:%s shutdown:%d\n"
|
|
|
|
, static_cast<void*>(s->m_impl), ec.message().c_str(), shutdown);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(s->m_connect_handler);
|
2017-01-21 01:03:20 +01:00
|
|
|
s->m_io_service.post(std::bind<void>(std::move(s->m_connect_handler), ec));
|
2016-08-13 03:31:55 +02:00
|
|
|
s->m_connect_handler = nullptr;
|
2017-09-17 13:12:54 +02:00
|
|
|
if (shutdown && s->m_impl)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2014-07-06 21:18:00 +02:00
|
|
|
TORRENT_ASSERT(ec);
|
2010-11-29 02:33:05 +01:00
|
|
|
detach_utp_impl(s->m_impl);
|
2016-07-09 22:26:26 +02:00
|
|
|
s->m_impl = nullptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-13 03:24:15 +02:00
|
|
|
void utp_stream::add_read_buffer(void* buf, std::size_t const len)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
TORRENT_ASSERT(m_impl);
|
|
|
|
TORRENT_ASSERT(len < INT_MAX);
|
|
|
|
TORRENT_ASSERT(len > 0);
|
|
|
|
TORRENT_ASSERT(buf);
|
2017-08-13 03:24:15 +02:00
|
|
|
m_impl->m_read_buffer.emplace_back(buf, len);
|
2016-04-25 23:22:09 +02:00
|
|
|
m_impl->m_read_buffer_size += int(len);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-08-13 03:24:15 +02:00
|
|
|
UTP_LOGV("%8p: add_read_buffer %d bytes\n", static_cast<void*>(m_impl), int(len));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// this is the wrapper to add a user provided write buffer to the
|
|
|
|
// utp_socket_impl. It makes sure the m_write_buffer_size is kept
|
|
|
|
// up to date
|
2017-08-13 03:24:15 +02:00
|
|
|
void utp_stream::add_write_buffer(void const* buf, std::size_t const len)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
TORRENT_ASSERT(m_impl);
|
|
|
|
TORRENT_ASSERT(len < INT_MAX);
|
|
|
|
TORRENT_ASSERT(len > 0);
|
|
|
|
TORRENT_ASSERT(buf);
|
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2010-11-29 02:33:05 +01:00
|
|
|
int write_buffer_size = 0;
|
2017-03-12 16:34:42 +01:00
|
|
|
for (auto const& i : m_impl->m_write_buffer)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2017-03-12 16:34:42 +01:00
|
|
|
TORRENT_ASSERT(std::numeric_limits<int>::max() - int(i.len) > write_buffer_size);
|
|
|
|
write_buffer_size += int(i.len);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
TORRENT_ASSERT(m_impl->m_write_buffer_size == write_buffer_size);
|
|
|
|
#endif
|
|
|
|
|
2017-08-13 03:24:15 +02:00
|
|
|
m_impl->m_write_buffer.emplace_back(const_cast<void*>(buf), len);
|
2016-04-25 23:22:09 +02:00
|
|
|
m_impl->m_write_buffer_size += int(len);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2010-11-29 02:33:05 +01:00
|
|
|
write_buffer_size = 0;
|
2017-03-12 16:34:42 +01:00
|
|
|
for (auto const& i : m_impl->m_write_buffer)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2017-03-12 16:34:42 +01:00
|
|
|
TORRENT_ASSERT(std::numeric_limits<int>::max() - int(i.len) > write_buffer_size);
|
|
|
|
write_buffer_size += int(i.len);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
TORRENT_ASSERT(m_impl->m_write_buffer_size == write_buffer_size);
|
|
|
|
#endif
|
|
|
|
|
2015-09-06 23:29:01 +02:00
|
|
|
UTP_LOGV("%8p: add_write_buffer %d bytes\n", static_cast<void*>(m_impl), int(len));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// this is called when all user provided read buffers have been added
|
|
|
|
// and it's time to execute the async operation. The first thing we
|
|
|
|
// do is to copy any data stored in m_receive_buffer into the user
|
|
|
|
// provided buffer. This might be enough to in turn trigger the read
|
|
|
|
// handler immediately.
|
2015-02-15 01:25:56 +01:00
|
|
|
void utp_stream::issue_read()
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2011-08-24 07:13:25 +02:00
|
|
|
TORRENT_ASSERT(m_impl->m_userdata);
|
2015-02-15 01:25:56 +01:00
|
|
|
TORRENT_ASSERT(!m_impl->m_read_handler);
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
m_impl->m_null_buffers = m_impl->m_read_buffer_size == 0;
|
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
m_impl->m_read_handler = true;
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_impl->test_socket_state()) return;
|
|
|
|
|
|
|
|
UTP_LOGV("%8p: new read handler. %d bytes in buffer\n"
|
2015-09-06 23:29:01 +02:00
|
|
|
, static_cast<void*>(m_impl), m_impl->m_receive_buffer_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// so, the client wants to read. If we already
|
|
|
|
// have some data in the read buffer, move it into the
|
|
|
|
// client's buffer right away
|
|
|
|
|
2016-04-25 23:22:09 +02:00
|
|
|
m_impl->m_read += int(read_some(false));
|
2013-02-06 05:38:30 +01:00
|
|
|
m_impl->maybe_trigger_receive_callback();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-08-13 03:24:15 +02:00
|
|
|
std::size_t utp_stream::read_some(bool const clear_buffers)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
if (m_impl->m_receive_buffer_size == 0)
|
|
|
|
{
|
|
|
|
if (clear_buffers)
|
|
|
|
{
|
|
|
|
m_impl->m_read_buffer_size = 0;
|
|
|
|
m_impl->m_read_buffer.clear();
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-03-12 16:34:42 +01:00
|
|
|
auto target = m_impl->m_read_buffer.begin();
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
std::size_t ret = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
int pop_packets = 0;
|
2017-02-11 20:21:48 +01:00
|
|
|
for (auto i = m_impl->m_receive_buffer.begin()
|
2010-11-29 02:33:05 +01:00
|
|
|
, end(m_impl->m_receive_buffer.end()); i != end;)
|
|
|
|
{
|
2015-08-17 15:01:43 +02:00
|
|
|
if (target == m_impl->m_read_buffer.end())
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
UTP_LOGV(" No more target buffers: %d bytes left in buffer\n"
|
|
|
|
, m_impl->m_receive_buffer_size);
|
|
|
|
TORRENT_ASSERT(m_impl->m_read_buffer.empty());
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
m_impl->check_receive_buffers();
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
packet* p = i->get();
|
2017-02-18 12:52:46 +01:00
|
|
|
int to_copy = std::min(p->size - p->header_size, aux::numeric_cast<int>(target->len));
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(to_copy >= 0);
|
2017-02-18 12:52:46 +01:00
|
|
|
std::memcpy(target->buf, p->buf + p->header_size, std::size_t(to_copy));
|
|
|
|
ret += std::size_t(to_copy);
|
2015-08-17 15:01:43 +02:00
|
|
|
target->buf = static_cast<char*>(target->buf) + to_copy;
|
2017-02-18 12:52:46 +01:00
|
|
|
TORRENT_ASSERT(target->len >= std::size_t(to_copy));
|
|
|
|
target->len -= std::size_t(to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
m_impl->m_receive_buffer_size -= to_copy;
|
|
|
|
TORRENT_ASSERT(m_impl->m_read_buffer_size >= to_copy);
|
|
|
|
m_impl->m_read_buffer_size -= to_copy;
|
2016-12-13 16:30:36 +01:00
|
|
|
p->header_size += std::uint16_t(to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (target->len == 0) target = m_impl->m_read_buffer.erase(target);
|
|
|
|
|
|
|
|
m_impl->check_receive_buffers();
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_impl->m_receive_buffer_size >= 0);
|
|
|
|
|
|
|
|
// Consumed entire packet
|
|
|
|
if (p->header_size == p->size)
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
m_impl->release_packet(std::move(*i));
|
|
|
|
i->reset();
|
2010-11-29 02:33:05 +01:00
|
|
|
++pop_packets;
|
|
|
|
++i;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (m_impl->m_receive_buffer_size == 0)
|
|
|
|
{
|
|
|
|
UTP_LOGV(" Didn't fill entire target: %d bytes left in buffer\n"
|
|
|
|
, m_impl->m_receive_buffer_size);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// remove the packets from the receive_buffer that we already copied over
|
|
|
|
// and freed
|
|
|
|
m_impl->m_receive_buffer.erase(m_impl->m_receive_buffer.begin()
|
|
|
|
, m_impl->m_receive_buffer.begin() + pop_packets);
|
|
|
|
// we exited either because we ran out of bytes to copy
|
|
|
|
// or because we ran out of space to copy the bytes to
|
|
|
|
TORRENT_ASSERT(m_impl->m_receive_buffer_size == 0
|
|
|
|
|| m_impl->m_read_buffer.empty());
|
|
|
|
|
2012-06-25 07:57:59 +02:00
|
|
|
UTP_LOGV("%8p: %d packets moved from buffer to user space (%d bytes)\n"
|
2015-09-06 23:29:01 +02:00
|
|
|
, static_cast<void*>(m_impl), pop_packets, int(ret));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (clear_buffers)
|
|
|
|
{
|
2015-09-06 23:29:01 +02:00
|
|
|
m_impl->m_read_buffer_size = 0;
|
|
|
|
m_impl->m_read_buffer.clear();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2014-07-06 21:18:00 +02:00
|
|
|
TORRENT_ASSERT(ret > 0 || m_impl->m_null_buffers);
|
2010-11-29 02:33:05 +01:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
// this is called when all user provided write buffers have been
|
|
|
|
// added. Start trying to send packets with the payload immediately.
|
2015-02-15 01:25:56 +01:00
|
|
|
void utp_stream::issue_write()
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: new write handler. %d bytes to write\n"
|
2015-09-06 23:29:01 +02:00
|
|
|
, static_cast<void*>(m_impl), m_impl->m_write_buffer_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(m_impl->m_write_buffer_size > 0);
|
2015-02-15 01:25:56 +01:00
|
|
|
TORRENT_ASSERT(m_impl->m_write_handler == false);
|
2011-08-24 07:13:25 +02:00
|
|
|
TORRENT_ASSERT(m_impl->m_userdata);
|
2015-02-15 01:25:56 +01:00
|
|
|
|
|
|
|
m_impl->m_write_handler = true;
|
2010-11-29 02:33:05 +01:00
|
|
|
m_impl->m_written = 0;
|
|
|
|
if (m_impl->test_socket_state()) return;
|
|
|
|
|
|
|
|
// try to write. send_pkt returns false if there's
|
|
|
|
// no more payload to send or if the congestion window
|
|
|
|
// is full and we can't send more packets right now
|
2012-06-28 05:53:52 +02:00
|
|
|
while (m_impl->send_pkt());
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// if there was an error in send_pkt(), m_impl may be
|
|
|
|
// 0 at this point
|
2013-02-06 05:38:30 +01:00
|
|
|
if (m_impl) m_impl->maybe_trigger_send_callback();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
void utp_stream::do_connect(tcp::endpoint const& ep)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
int link_mtu, utp_mtu;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_impl->m_sm.mtu_for_dest(ep.address(), link_mtu, utp_mtu);
|
2010-11-29 02:33:05 +01:00
|
|
|
m_impl->init_mtu(link_mtu, utp_mtu);
|
2015-02-15 01:25:56 +01:00
|
|
|
TORRENT_ASSERT(m_impl->m_connect_handler == false);
|
2010-11-29 02:33:05 +01:00
|
|
|
m_impl->m_remote_address = ep.address();
|
|
|
|
m_impl->m_port = ep.port();
|
2015-02-15 01:25:56 +01:00
|
|
|
|
|
|
|
m_impl->m_connect_handler = true;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (m_impl->test_socket_state()) return;
|
|
|
|
m_impl->send_syn();
|
|
|
|
}
|
|
|
|
|
|
|
|
// =========== utp_socket_impl ============
|
|
|
|
|
|
|
|
utp_socket_impl::~utp_socket_impl()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(!m_attached);
|
2012-06-21 17:05:57 +02:00
|
|
|
TORRENT_ASSERT(!m_deferred_ack);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::num_utp_idle + m_state, -1);
|
2015-01-04 22:31:02 +01:00
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: destroying utp socket state\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// free any buffers we're holding
|
2016-11-25 17:17:25 +01:00
|
|
|
for (std::uint16_t i = std::uint16_t(m_inbuf.cursor()), end((m_inbuf.cursor()
|
2010-11-29 02:33:05 +01:00
|
|
|
+ m_inbuf.capacity()) & ACK_MASK);
|
|
|
|
i != end; i = (i + 1) & ACK_MASK)
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr p = m_inbuf.remove(i);
|
|
|
|
release_packet(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2016-11-25 17:17:25 +01:00
|
|
|
for (std::uint16_t i = std::uint16_t(m_outbuf.cursor()), end((m_outbuf.cursor()
|
2010-11-29 02:33:05 +01:00
|
|
|
+ m_outbuf.capacity()) & ACK_MASK);
|
|
|
|
i != end; i = (i + 1) & ACK_MASK)
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr p = m_outbuf.remove(i);
|
|
|
|
release_packet(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
for (auto& p : m_receive_buffer)
|
|
|
|
release_packet(std::move(p));
|
2012-07-21 16:33:01 +02:00
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
release_packet(std::move(m_nagle_packet));
|
|
|
|
m_nagle_packet.reset();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
bool utp_socket_impl::should_delete() const
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// if the socket state is not attached anymore we're free
|
|
|
|
// to delete it from the client's point of view. The other
|
|
|
|
// endpoint however might still need to be told that we're
|
|
|
|
// closing the socket. Only delete the state if we're not
|
|
|
|
// attached and we're in a state where the other end doesn't
|
|
|
|
// expect the socket to still be alive
|
2012-11-20 02:58:09 +01:00
|
|
|
// when m_stalled is true, it means the socket manager has a
|
|
|
|
// pointer to this socket, waiting for the UDP socket to
|
|
|
|
// become writable again. We have to wait for that, so that
|
|
|
|
// the pointer is removed from that queue. Otherwise we would
|
|
|
|
// leave a dangling pointer in the socket manager
|
2010-11-29 02:33:05 +01:00
|
|
|
bool ret = (m_state >= UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_NONE)
|
2012-11-20 02:58:09 +01:00
|
|
|
&& !m_attached && !m_stalled;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (ret)
|
|
|
|
{
|
2015-09-06 23:29:01 +02:00
|
|
|
UTP_LOGV("%8p: should_delete() = true\n", static_cast<void const*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
void utp_socket_impl::maybe_trigger_receive_callback()
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
if (m_read_handler == false) return;
|
2014-07-06 21:18:00 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// nothing has been read or there's no outstanding read operation
|
2014-07-06 21:18:00 +02:00
|
|
|
if (m_null_buffers && m_receive_buffer_size == 0) return;
|
|
|
|
else if (!m_null_buffers && m_read == 0) return;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: calling read handler read:%d\n", static_cast<void*>(this), m_read);
|
2015-02-15 01:25:56 +01:00
|
|
|
m_read_handler = false;
|
2017-02-18 12:52:46 +01:00
|
|
|
utp_stream::on_read(m_userdata, aux::numeric_cast<std::size_t>(m_read), m_error, false);
|
2013-02-06 05:38:30 +01:00
|
|
|
m_read = 0;
|
|
|
|
m_read_buffer_size = 0;
|
|
|
|
m_read_buffer.clear();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
void utp_socket_impl::maybe_trigger_send_callback()
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// nothing has been written or there's no outstanding write operation
|
2015-02-15 01:25:56 +01:00
|
|
|
if (m_written == 0 || m_write_handler == false) return;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: calling write handler written:%d\n", static_cast<void*>(this), m_written);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
m_write_handler = false;
|
2017-02-18 12:52:46 +01:00
|
|
|
utp_stream::on_write(m_userdata, aux::numeric_cast<std::size_t>(m_written), m_error, false);
|
2013-02-06 05:38:30 +01:00
|
|
|
m_written = 0;
|
|
|
|
m_write_buffer_size = 0;
|
|
|
|
m_write_buffer.clear();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-04 02:27:31 +01:00
|
|
|
void utp_socket_impl::set_close_reason(close_reason_t code)
|
2015-02-28 20:51:15 +01:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: set_close_reason: %d\n"
|
2017-02-04 02:27:31 +01:00
|
|
|
, static_cast<void*>(this), static_cast<int>(m_close_reason));
|
2015-02-28 20:51:15 +01:00
|
|
|
#endif
|
|
|
|
m_close_reason = code;
|
|
|
|
}
|
|
|
|
|
2011-07-14 05:46:03 +02:00
|
|
|
bool utp_socket_impl::destroy()
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2015-02-28 20:51:15 +01:00
|
|
|
UTP_LOGV("%8p: destroy state:%s (close-reason: %d)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), socket_state_names[m_state], int(m_close_reason));
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
2016-07-09 22:26:26 +02:00
|
|
|
if (m_userdata == nullptr) return false;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (m_state == UTP_STATE_CONNECTED)
|
|
|
|
send_fin();
|
|
|
|
|
2015-06-06 07:22:53 +02:00
|
|
|
bool cancelled = cancel_handlers(boost::asio::error::operation_aborted, true);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-07-09 22:26:26 +02:00
|
|
|
m_userdata = nullptr;
|
2011-08-24 07:13:25 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
m_read_buffer.clear();
|
|
|
|
m_read_buffer_size = 0;
|
|
|
|
|
|
|
|
m_write_buffer.clear();
|
|
|
|
m_write_buffer_size = 0;
|
|
|
|
|
|
|
|
if ((m_state == UTP_STATE_ERROR_WAIT
|
|
|
|
|| m_state == UTP_STATE_NONE
|
|
|
|
|| m_state == UTP_STATE_SYN_SENT) && cancelled)
|
|
|
|
{
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_DELETE);
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: state:%s\n", static_cast<void*>(this), socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2011-07-14 05:46:03 +02:00
|
|
|
return cancelled;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// #error our end is closing. Wait for everything to be acked
|
|
|
|
}
|
|
|
|
|
|
|
|
void utp_socket_impl::detach()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: detach()\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_attached = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void utp_socket_impl::send_syn()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2016-11-25 17:17:25 +01:00
|
|
|
m_seq_nr = std::uint16_t(random(0xffff));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_acked_seq_nr = (m_seq_nr - 1) & ACK_MASK;
|
|
|
|
m_loss_seq_nr = m_acked_seq_nr;
|
|
|
|
m_ack_nr = 0;
|
|
|
|
m_fast_resend_seq_nr = m_seq_nr;
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr p = acquire_packet(sizeof(utp_header));
|
2010-11-29 02:33:05 +01:00
|
|
|
p->size = sizeof(utp_header);
|
|
|
|
p->header_size = sizeof(utp_header);
|
2012-07-01 20:44:46 +02:00
|
|
|
p->num_transmissions = 0;
|
2016-03-12 07:07:17 +01:00
|
|
|
p->mtu_probe = false;
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2013-12-29 20:35:39 +01:00
|
|
|
p->num_fast_resend = 0;
|
|
|
|
#endif
|
2010-11-29 02:33:05 +01:00
|
|
|
p->need_resend = false;
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* h = reinterpret_cast<utp_header*>(p->buf);
|
2010-11-29 02:33:05 +01:00
|
|
|
h->type_ver = (ST_SYN << 4) | 1;
|
2015-02-28 20:51:15 +01:00
|
|
|
h->extension = utp_no_extension;
|
2010-11-29 02:33:05 +01:00
|
|
|
// using recv_id here is intentional! This is an odd
|
|
|
|
// thing in uTP. The syn packet is sent with the connection
|
|
|
|
// ID that it expects to receive the syn ack on. All
|
|
|
|
// subsequent connection IDs will be this plus one.
|
|
|
|
h->connection_id = m_recv_id;
|
|
|
|
h->timestamp_difference_microseconds = m_reply_micro;
|
|
|
|
h->wnd_size = 0;
|
|
|
|
h->seq_nr = m_seq_nr;
|
|
|
|
h->ack_nr = 0;
|
|
|
|
|
2016-12-10 20:15:25 +01:00
|
|
|
time_point const now = clock_type::now();
|
2010-11-29 02:33:05 +01:00
|
|
|
p->send_time = now;
|
2016-06-18 20:01:38 +02:00
|
|
|
h->timestamp_microseconds = std::uint32_t(
|
2014-07-06 21:18:00 +02:00
|
|
|
total_microseconds(now.time_since_epoch()) & 0xffffffff);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: send_syn seq_nr:%d id:%d target:%s\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(m_seq_nr), int(m_recv_id)
|
2010-11-29 02:33:05 +01:00
|
|
|
, print_endpoint(udp::endpoint(m_remote_address, m_port)).c_str());
|
|
|
|
#endif
|
|
|
|
|
|
|
|
error_code ec;
|
2017-04-26 05:00:41 +02:00
|
|
|
m_sm.send_packet(m_sock, udp::endpoint(m_remote_address, m_port)
|
2015-08-17 15:01:43 +02:00
|
|
|
, reinterpret_cast<char const*>(h) , sizeof(utp_header), ec);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-11-20 03:08:12 +01:00
|
|
|
if (ec == error::would_block || ec == error::try_again)
|
2012-07-01 20:44:46 +02:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: socket stalled\n", static_cast<void*>(this));
|
2012-07-01 20:44:46 +02:00
|
|
|
#endif
|
|
|
|
if (!m_stalled)
|
|
|
|
{
|
|
|
|
m_stalled = true;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.subscribe_writable(this);
|
2012-07-01 20:44:46 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (ec)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
release_packet(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_error = ec;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
if (!m_stalled)
|
|
|
|
++p->num_transmissions;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(!m_outbuf.at(m_seq_nr));
|
2012-06-27 01:52:48 +02:00
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
2016-06-18 20:01:38 +02:00
|
|
|
TORRENT_ASSERT(p->buf == reinterpret_cast<std::uint8_t*>(h));
|
2017-02-11 20:21:48 +01:00
|
|
|
m_outbuf.insert(m_seq_nr, std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
m_seq_nr = (m_seq_nr + 1) & ACK_MASK;
|
|
|
|
|
|
|
|
TORRENT_ASSERT(!m_error);
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_SYN_SENT);
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: state:%s\n", static_cast<void*>(this), socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
// if a send ever failed with EWOULDBLOCK, we
|
|
|
|
// subscribe to the udp socket and will be
|
|
|
|
// signalled with this function.
|
|
|
|
void utp_socket_impl::writable()
|
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: writable\n", static_cast<void*>(this));
|
2012-07-01 20:44:46 +02:00
|
|
|
#endif
|
2012-11-20 02:58:09 +01:00
|
|
|
if (should_delete()) return;
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
while(send_pkt());
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
maybe_trigger_send_callback();
|
2012-07-01 20:44:46 +02:00
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
void utp_socket_impl::send_fin()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2012-06-28 05:53:52 +02:00
|
|
|
send_pkt(pkt_fin);
|
|
|
|
// unless there was an error, we're now
|
|
|
|
// in FIN-SENT state
|
|
|
|
if (!m_error)
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_FIN_SENT);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: state:%s\n", static_cast<void*>(this), socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2015-08-19 01:39:01 +02:00
|
|
|
void utp_socket_impl::send_reset(utp_header const* ph)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
utp_header h;
|
|
|
|
h.type_ver = (ST_RESET << 4) | 1;
|
2015-02-28 20:51:15 +01:00
|
|
|
h.extension = utp_no_extension;
|
2010-11-29 02:33:05 +01:00
|
|
|
h.connection_id = m_send_id;
|
|
|
|
h.timestamp_difference_microseconds = m_reply_micro;
|
|
|
|
h.wnd_size = 0;
|
2016-11-25 17:17:25 +01:00
|
|
|
h.seq_nr = std::uint16_t(random(0xffff));
|
2010-11-29 02:33:05 +01:00
|
|
|
h.ack_nr = ph->seq_nr;
|
2016-12-10 20:15:25 +01:00
|
|
|
time_point const now = clock_type::now();
|
2016-06-18 20:01:38 +02:00
|
|
|
h.timestamp_microseconds = std::uint32_t(
|
2014-07-06 21:18:00 +02:00
|
|
|
total_microseconds(now.time_since_epoch()) & 0xffffffff);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: send_reset seq_nr:%d id:%d ack_nr:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(h.seq_nr), int(m_send_id), int(ph->seq_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// ignore errors here
|
|
|
|
error_code ec;
|
2017-04-26 05:00:41 +02:00
|
|
|
m_sm.send_packet(m_sock, udp::endpoint(m_remote_address, m_port)
|
2015-08-17 15:01:43 +02:00
|
|
|
, reinterpret_cast<char const*>(&h), sizeof(h), ec);
|
2016-03-12 07:07:17 +01:00
|
|
|
if (ec)
|
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: socket error: %s\n"
|
|
|
|
, static_cast<void*>(this)
|
|
|
|
, ec.message().c_str());
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
std::size_t utp_socket_impl::available() const
|
|
|
|
{
|
2017-02-18 12:52:46 +01:00
|
|
|
return aux::numeric_cast<std::size_t>(m_receive_buffer_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
void utp_socket_impl::parse_close_reason(std::uint8_t const* ptr, int const size)
|
2015-02-28 20:51:15 +01:00
|
|
|
{
|
|
|
|
if (size != 4) return;
|
|
|
|
// skip reserved bytes
|
|
|
|
ptr += 2;
|
2017-02-04 02:27:31 +01:00
|
|
|
close_reason_t incoming_close_reason = static_cast<close_reason_t>(detail::read_uint16(ptr));
|
2015-02-28 20:51:15 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: incoming close_reason: %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(incoming_close_reason));
|
2015-02-28 20:51:15 +01:00
|
|
|
|
2016-07-09 22:26:26 +02:00
|
|
|
if (m_userdata == nullptr) return;
|
2015-02-28 20:51:15 +01:00
|
|
|
|
|
|
|
utp_stream::on_close_reason(m_userdata, incoming_close_reason);
|
|
|
|
}
|
|
|
|
|
2017-02-18 19:16:55 +01:00
|
|
|
// returns (rtt, acked_bytes)
|
|
|
|
std::pair<std::uint32_t, int> utp_socket_impl::parse_sack(std::uint16_t const packet_ack
|
|
|
|
, std::uint8_t const* ptr, int size, time_point const now)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2017-02-18 19:16:55 +01:00
|
|
|
if (size == 0) return { 0u, 0 };
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// this is the sequence number the current bit represents
|
2017-02-26 18:08:35 +01:00
|
|
|
std::uint32_t ack_nr = (packet_ack + 2) & ACK_MASK;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-25 07:57:59 +02:00
|
|
|
#if TORRENT_VERBOSE_UTP_LOG
|
2010-11-29 02:33:05 +01:00
|
|
|
std::string bitmask;
|
2012-06-25 07:57:59 +02:00
|
|
|
bitmask.reserve(size);
|
2016-06-18 20:01:38 +02:00
|
|
|
for (std::uint8_t const* b = ptr, *end = ptr + size; b != end; ++b)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
unsigned char bitfield = unsigned(*b);
|
|
|
|
unsigned char mask = 1;
|
|
|
|
// for each bit
|
|
|
|
for (int i = 0; i < 8; ++i)
|
|
|
|
{
|
|
|
|
bitmask += (mask & bitfield) ? "1" : "0";
|
|
|
|
mask <<= 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
UTP_LOGV("%8p: got SACK first:%d %s our_seq_nr:%u\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), ack_nr, bitmask.c_str(), m_seq_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
// the number of acked packets past the fast re-send sequence number
|
|
|
|
// this is used to determine if we should trigger more fast re-sends
|
|
|
|
int dups = 0;
|
|
|
|
|
|
|
|
// the sequence number of the last ACKed packet
|
2017-02-26 18:08:35 +01:00
|
|
|
std::uint32_t last_ack = packet_ack;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-18 19:16:55 +01:00
|
|
|
int acked_bytes = 0;
|
|
|
|
std::uint32_t min_rtt = std::numeric_limits<std::uint32_t>::max();
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// for each byte
|
2016-06-18 20:01:38 +02:00
|
|
|
for (std::uint8_t const* end = ptr + size; ptr != end; ++ptr)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2016-11-25 17:17:25 +01:00
|
|
|
std::uint8_t bitfield = *ptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
unsigned char mask = 1;
|
|
|
|
// for each bit
|
|
|
|
for (int i = 0; i < 8; ++i)
|
|
|
|
{
|
|
|
|
if (mask & bitfield)
|
|
|
|
{
|
|
|
|
last_ack = ack_nr;
|
|
|
|
if (m_fast_resend_seq_nr == ack_nr)
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
|
|
|
|
2011-11-06 01:11:33 +01:00
|
|
|
if (compare_less_wrap(m_fast_resend_seq_nr, ack_nr, ACK_MASK)) ++dups;
|
2010-11-29 02:33:05 +01:00
|
|
|
// this bit was set, ack_nr was received
|
2017-02-18 12:52:46 +01:00
|
|
|
packet_ptr p = m_outbuf.remove(aux::numeric_cast<packet_buffer::index_type>(ack_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
if (p)
|
|
|
|
{
|
2017-02-18 19:16:55 +01:00
|
|
|
acked_bytes += p->size - p->header_size;
|
2010-11-29 02:33:05 +01:00
|
|
|
// each ACKed packet counts as a duplicate ack
|
|
|
|
UTP_LOGV("%8p: duplicate_acks:%u fast_resend_seq_nr:%u\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_duplicate_acks, m_fast_resend_seq_nr);
|
2017-02-18 19:16:55 +01:00
|
|
|
min_rtt = std::min(min_rtt, ack_packet(std::move(p), now, std::uint16_t(ack_nr)));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2010-12-05 04:03:56 +01:00
|
|
|
else
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2010-12-05 04:03:56 +01:00
|
|
|
// this packet might have been acked by a previous
|
2010-11-29 02:33:05 +01:00
|
|
|
// selective ack
|
2010-12-05 04:03:56 +01:00
|
|
|
maybe_inc_acked_seq_nr();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
mask <<= 1;
|
|
|
|
ack_nr = (ack_nr + 1) & ACK_MASK;
|
|
|
|
|
|
|
|
// we haven't sent packets past this point.
|
|
|
|
// if there are any more bits set, we have to
|
|
|
|
// ignore them anyway
|
|
|
|
if (ack_nr == m_seq_nr) break;
|
|
|
|
}
|
|
|
|
if (ack_nr == m_seq_nr) break;
|
|
|
|
}
|
|
|
|
|
2010-12-05 04:03:56 +01:00
|
|
|
TORRENT_ASSERT(m_outbuf.at((m_acked_seq_nr + 1) & ACK_MASK) || ((m_seq_nr - m_acked_seq_nr) & ACK_MASK) <= 1);
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// we received more than dup_ack_limit ACKs in this SACK message.
|
|
|
|
// trigger fast re-send
|
2011-11-06 01:11:33 +01:00
|
|
|
if (dups >= dup_ack_limit && compare_less_wrap(m_fast_resend_seq_nr, last_ack, ACK_MASK))
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
experienced_loss(m_fast_resend_seq_nr);
|
|
|
|
int num_resent = 0;
|
2013-12-29 20:35:39 +01:00
|
|
|
while (m_fast_resend_seq_nr != last_ack)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-08-17 15:01:43 +02:00
|
|
|
packet* p = m_outbuf.at(m_fast_resend_seq_nr);
|
2013-12-29 20:35:39 +01:00
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
2010-11-29 02:33:05 +01:00
|
|
|
if (!p) continue;
|
|
|
|
++num_resent;
|
|
|
|
if (!resend_packet(p, true)) break;
|
|
|
|
m_duplicate_acks = 0;
|
|
|
|
if (num_resent >= sack_resend_limit) break;
|
|
|
|
}
|
|
|
|
}
|
2017-02-18 19:16:55 +01:00
|
|
|
|
|
|
|
return { min_rtt, acked_bytes };
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// copies data from the write buffer into the packet
|
|
|
|
// pointed to by ptr
|
2016-06-18 20:01:38 +02:00
|
|
|
void utp_socket_impl::write_payload(std::uint8_t* ptr, int size)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2010-11-29 02:33:05 +01:00
|
|
|
int write_buffer_size = 0;
|
2017-03-12 16:34:42 +01:00
|
|
|
for (auto const& i : m_write_buffer)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2017-03-12 16:34:42 +01:00
|
|
|
TORRENT_ASSERT(std::numeric_limits<int>::max() - int(i.len) > write_buffer_size);
|
|
|
|
write_buffer_size += int(i.len);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
TORRENT_ASSERT(m_write_buffer_size == write_buffer_size);
|
|
|
|
#endif
|
|
|
|
TORRENT_ASSERT(!m_write_buffer.empty() || size == 0);
|
|
|
|
TORRENT_ASSERT(m_write_buffer_size >= size);
|
2017-03-12 16:34:42 +01:00
|
|
|
auto i = m_write_buffer.begin();
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (size == 0) return;
|
|
|
|
|
|
|
|
int buffers_to_clear = 0;
|
|
|
|
while (size > 0)
|
|
|
|
{
|
|
|
|
// i points to the iovec we'll start copying from
|
2017-02-11 20:21:48 +01:00
|
|
|
int to_copy = std::min(size, int(i->len));
|
2011-02-21 06:24:41 +01:00
|
|
|
TORRENT_ASSERT(to_copy >= 0);
|
|
|
|
TORRENT_ASSERT(to_copy < INT_MAX / 2 && m_written < INT_MAX / 2);
|
2017-02-18 12:52:46 +01:00
|
|
|
std::memcpy(ptr, static_cast<char const*>(i->buf), std::size_t(to_copy));
|
2010-11-29 02:33:05 +01:00
|
|
|
size -= to_copy;
|
|
|
|
m_written += to_copy;
|
|
|
|
ptr += to_copy;
|
2017-02-18 12:52:46 +01:00
|
|
|
i->len -= std::size_t(to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(m_write_buffer_size >= to_copy);
|
|
|
|
m_write_buffer_size -= to_copy;
|
2015-08-19 01:39:01 +02:00
|
|
|
i->buf = static_cast<char*>(i->buf) + to_copy;
|
2010-11-29 02:33:05 +01:00
|
|
|
if (i->len == 0) ++buffers_to_clear;
|
|
|
|
++i;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (buffers_to_clear)
|
|
|
|
m_write_buffer.erase(m_write_buffer.begin()
|
|
|
|
, m_write_buffer.begin() + buffers_to_clear);
|
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2010-11-29 02:33:05 +01:00
|
|
|
write_buffer_size = 0;
|
2017-03-12 16:34:42 +01:00
|
|
|
for (auto const& j : m_write_buffer)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2017-03-12 16:34:42 +01:00
|
|
|
TORRENT_ASSERT(std::numeric_limits<int>::max() - int(j.len) > write_buffer_size);
|
|
|
|
write_buffer_size += int(j.len);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
TORRENT_ASSERT(m_write_buffer_size == write_buffer_size);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
void utp_socket_impl::subscribe_drained()
|
|
|
|
{
|
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
|
|
|
if (m_subscribe_drained) return;
|
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: subscribe drained\n", static_cast<void*>(this));
|
2013-02-06 05:38:30 +01:00
|
|
|
m_subscribe_drained = true;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.subscribe_drained(this);
|
2013-02-06 05:38:30 +01:00
|
|
|
}
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
void utp_socket_impl::defer_ack()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
if (m_deferred_ack) return;
|
2012-06-25 07:57:59 +02:00
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: defer ack\n", static_cast<void*>(this));
|
2012-06-21 17:05:57 +02:00
|
|
|
m_deferred_ack = true;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.defer_ack(this);
|
2012-06-21 17:05:57 +02:00
|
|
|
}
|
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
void utp_socket_impl::remove_sack_header(packet* p)
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
// remove the sack header
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t* ptr = p->buf + sizeof(utp_header);
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* h = reinterpret_cast<utp_header*>(p->buf);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
2015-02-28 20:51:15 +01:00
|
|
|
TORRENT_ASSERT(h->extension == utp_sack);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
h->extension = ptr[0];
|
|
|
|
int sack_size = ptr[1];
|
2015-02-28 20:51:15 +01:00
|
|
|
TORRENT_ASSERT(h->extension == utp_no_extension
|
|
|
|
|| h->extension == utp_close_reason);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: removing SACK header, %d bytes\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), sack_size + 2);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
2012-06-26 06:10:53 +02:00
|
|
|
TORRENT_ASSERT(p->size >= p->header_size);
|
2017-02-18 12:52:46 +01:00
|
|
|
TORRENT_ASSERT(p->header_size >= sizeof(utp_header) + aux::numeric_cast<std::size_t>(sack_size) + 2);
|
|
|
|
std::memmove(ptr, ptr + sack_size + 2, p->size - p->header_size);
|
2016-12-13 16:30:36 +01:00
|
|
|
p->header_size -= std::uint16_t(sack_size + 2);
|
|
|
|
p->size -= std::uint16_t(sack_size + 2);
|
2012-06-25 08:27:37 +02:00
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// sends a packet, pulls data from the write buffer (if there's any)
|
|
|
|
// if ack is true, we need to send a packet regardless of if there's
|
|
|
|
// any data. Returns true if we could send more data (i.e. call
|
|
|
|
// send_pkt() again)
|
2014-02-15 03:45:04 +01:00
|
|
|
// returns true if there is more space for payload in our
|
|
|
|
// congestion window, false if there is no more space.
|
2016-05-02 06:17:17 +02:00
|
|
|
bool utp_socket_impl::send_pkt(int const flags)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-07-20 17:20:17 +02:00
|
|
|
#ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
2015-07-20 17:20:17 +02:00
|
|
|
#endif
|
2012-06-26 05:42:01 +02:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
bool const force = (flags & pkt_ack) || (flags & pkt_fin);
|
2012-06-28 05:53:52 +02:00
|
|
|
|
|
|
|
// TORRENT_ASSERT(m_state != UTP_STATE_FIN_SENT || (flags & pkt_ack));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// first see if we need to resend any packets
|
|
|
|
|
2015-02-28 20:51:15 +01:00
|
|
|
// TODO: this loop is not very efficient. It could be fixed by having
|
|
|
|
// a separate list of sequence numbers that need resending
|
2010-11-29 02:33:05 +01:00
|
|
|
for (int i = (m_acked_seq_nr + 1) & ACK_MASK; i != m_seq_nr; i = (i + 1) & ACK_MASK)
|
|
|
|
{
|
2017-02-18 12:52:46 +01:00
|
|
|
packet* p = m_outbuf.at(aux::numeric_cast<packet_buffer::index_type>(i));
|
2010-11-29 02:33:05 +01:00
|
|
|
if (!p) continue;
|
|
|
|
if (!p->need_resend) continue;
|
|
|
|
if (!resend_packet(p))
|
|
|
|
{
|
|
|
|
// we couldn't resend the packet. It probably doesn't
|
2012-06-28 05:53:52 +02:00
|
|
|
// fit in our cwnd. If force is set, we need to continue
|
2013-09-14 12:06:48 +02:00
|
|
|
// to send our packet anyway, if we don't have force set,
|
|
|
|
// we might as well return
|
2012-06-28 05:53:52 +02:00
|
|
|
if (!force) return false;
|
2010-11-29 02:33:05 +01:00
|
|
|
// resend_packet might have failed
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// don't fast-resend this packet
|
|
|
|
if (m_fast_resend_seq_nr == i)
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
|
|
|
}
|
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// MTU DISCOVERY
|
2015-02-28 20:51:15 +01:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// under these conditions, the next packet we send should be an MTU probe.
|
|
|
|
// MTU probes get to use the mid-point packet size, whereas other packets
|
|
|
|
// use a conservative packet size of the largest known to work. The reason
|
|
|
|
// for the cwnd condition is to make sure the probe is surrounded by non-
|
|
|
|
// probes, to be able to distinguish a loss of the probe vs. just loss in
|
|
|
|
// general.
|
|
|
|
bool const mtu_probe = (m_mtu_seq == 0
|
|
|
|
&& m_write_buffer_size >= m_mtu_floor * 3
|
|
|
|
&& m_seq_nr != 0
|
|
|
|
&& (m_cwnd >> 16) > m_mtu_floor * 3);
|
2018-03-07 16:44:35 +01:00
|
|
|
// for non MTU-probes, use the conservative packet size
|
|
|
|
int const effective_mtu = mtu_probe ? m_mtu : m_mtu_floor;
|
|
|
|
|
2018-04-12 16:21:20 +02:00
|
|
|
std::uint32_t const close_reason = static_cast<std::uint32_t>(m_close_reason);
|
2018-03-07 16:44:35 +01:00
|
|
|
|
|
|
|
int sack = 0;
|
|
|
|
if (m_inbuf.size())
|
|
|
|
{
|
|
|
|
const int max_sack_size = effective_mtu
|
2018-04-12 16:21:20 +02:00
|
|
|
- int(sizeof(utp_header))
|
2018-03-07 16:44:35 +01:00
|
|
|
- 2 // for sack padding/header
|
|
|
|
- (close_reason ? 6 : 0);
|
|
|
|
|
|
|
|
// the SACK bitfield should ideally fit all
|
|
|
|
// the pieces we have successfully received
|
|
|
|
sack = (m_inbuf.span() + 7) / 8;
|
|
|
|
if (sack > max_sack_size) sack = max_sack_size;
|
|
|
|
}
|
2016-03-12 07:07:17 +01:00
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
int const header_size = int(sizeof(utp_header))
|
2015-02-28 20:51:15 +01:00
|
|
|
+ (sack ? sack + 2 : 0)
|
|
|
|
+ (close_reason ? 6 : 0);
|
2016-03-12 07:07:17 +01:00
|
|
|
|
|
|
|
// for non MTU-probes, use the conservative packet size
|
2017-02-11 20:21:48 +01:00
|
|
|
int payload_size = std::min(m_write_buffer_size
|
2016-03-12 07:07:17 +01:00
|
|
|
, effective_mtu - header_size);
|
2016-09-13 08:36:23 +02:00
|
|
|
TORRENT_ASSERT(payload_size >= 0);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// if we have one MSS worth of data, make sure it fits in our
|
2016-04-26 06:37:47 +02:00
|
|
|
// congestion window and the advertised receive window from
|
2010-11-29 02:33:05 +01:00
|
|
|
// the other end.
|
2017-02-11 20:21:48 +01:00
|
|
|
if (m_bytes_in_flight + payload_size > std::min(int(m_cwnd >> 16)
|
2018-02-28 11:50:41 +01:00
|
|
|
, int(m_adv_wnd)))
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
// this means there's not enough room in the send window for
|
|
|
|
// another packet. We have to hold off sending this data.
|
|
|
|
// we still need to send an ACK though
|
2012-06-28 05:53:52 +02:00
|
|
|
// if we're trying to send a FIN, make an exception
|
|
|
|
if ((flags & pkt_fin) == 0) payload_size = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2011-11-05 10:30:38 +01:00
|
|
|
// we're constrained by the window size
|
|
|
|
m_cwnd_full = true;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: no space in window send_buffer_size:%d cwnd:%d "
|
2012-06-25 08:27:37 +02:00
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_write_buffer_size, int(m_cwnd >> 16)
|
2012-06-25 08:27:37 +02:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu);
|
|
|
|
|
2012-06-28 05:53:52 +02:00
|
|
|
if (!force)
|
2012-06-25 08:27:37 +02:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: skipping send seq_nr:%d ack_nr:%d "
|
|
|
|
"id:%d target:%s header_size:%d error:%s send_buffer_size:%d cwnd:%d "
|
2016-03-12 07:07:17 +01:00
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d effective-mtu:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(m_seq_nr), int(m_ack_nr)
|
2012-06-25 08:27:37 +02:00
|
|
|
, m_send_id, print_endpoint(udp::endpoint(m_remote_address, m_port)).c_str()
|
|
|
|
, header_size, m_error.message().c_str(), m_write_buffer_size, int(m_cwnd >> 16)
|
2016-03-12 07:07:17 +01:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu, effective_mtu);
|
2012-06-25 08:27:37 +02:00
|
|
|
#endif
|
|
|
|
return false;
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// if we don't have any data to send, or can't send any data
|
2012-06-28 05:53:52 +02:00
|
|
|
// and we don't have any data to force, don't send a packet
|
|
|
|
if (payload_size == 0 && !force && !m_nagle_packet)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2012-06-28 05:53:52 +02:00
|
|
|
UTP_LOGV("%8p: skipping send (no payload and no force) seq_nr:%d ack_nr:%d "
|
2010-11-29 02:33:05 +01:00
|
|
|
"id:%d target:%s header_size:%d error:%s send_buffer_size:%d cwnd:%d "
|
2012-06-25 08:27:37 +02:00
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(m_seq_nr), int(m_ack_nr)
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_send_id, print_endpoint(udp::endpoint(m_remote_address, m_port)).c_str()
|
2011-07-14 05:46:03 +02:00
|
|
|
, header_size, m_error.message().c_str(), m_write_buffer_size, int(m_cwnd >> 16)
|
2012-06-25 08:27:37 +02:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
int packet_size = header_size + payload_size;
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr p;
|
2016-06-20 17:32:06 +02:00
|
|
|
std::uint8_t* ptr = nullptr;
|
|
|
|
utp_header* h = nullptr;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
// payload size being zero means we're just sending
|
2012-06-28 05:53:52 +02:00
|
|
|
// an force. We should not pick up the nagle packet
|
|
|
|
if (!m_nagle_packet || (payload_size == 0 && force))
|
2012-06-25 08:27:37 +02:00
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
p = acquire_packet(effective_mtu);
|
|
|
|
|
2015-08-19 01:39:01 +02:00
|
|
|
if (payload_size)
|
2012-06-25 08:27:37 +02:00
|
|
|
{
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_payload_pkts_out);
|
2012-06-25 08:27:37 +02:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-11-25 17:17:25 +01:00
|
|
|
p->size = std::uint16_t(packet_size);
|
|
|
|
p->header_size = std::uint16_t(packet_size - payload_size);
|
2012-06-25 08:27:37 +02:00
|
|
|
p->num_transmissions = 0;
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2013-12-29 20:35:39 +01:00
|
|
|
p->num_fast_resend = 0;
|
|
|
|
#endif
|
2016-09-13 08:36:23 +02:00
|
|
|
p->mtu_probe = false;
|
2012-06-25 08:27:37 +02:00
|
|
|
p->need_resend = false;
|
|
|
|
ptr = p->buf;
|
2015-08-19 01:39:01 +02:00
|
|
|
h = reinterpret_cast<utp_header*>(ptr);
|
2012-06-25 08:27:37 +02:00
|
|
|
ptr += sizeof(utp_header);
|
|
|
|
|
2016-12-13 16:30:36 +01:00
|
|
|
h->extension = std::uint8_t(sack ? utp_sack
|
|
|
|
: close_reason ? utp_close_reason : utp_no_extension);
|
2012-06-25 08:27:37 +02:00
|
|
|
h->connection_id = m_send_id;
|
|
|
|
// seq_nr is ignored for ST_STATE packets, so it doesn't
|
|
|
|
// matter that we say this is a sequence number we haven't
|
|
|
|
// actually sent yet
|
|
|
|
h->seq_nr = m_seq_nr;
|
2016-11-25 17:17:25 +01:00
|
|
|
h->type_ver = std::uint8_t(((payload_size ? ST_DATA : ST_STATE) << 4) | 1);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
write_payload(p->buf + p->header_size, payload_size);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// pick up the nagle packet and keep adding bytes to it
|
2017-02-11 20:21:48 +01:00
|
|
|
p = std::move(m_nagle_packet);
|
|
|
|
m_nagle_packet.reset();
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
ptr = p->buf + sizeof(utp_header);
|
2015-08-19 01:39:01 +02:00
|
|
|
h = reinterpret_cast<utp_header*>(p->buf);
|
2012-06-28 05:53:52 +02:00
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
2012-06-28 05:53:52 +02:00
|
|
|
// if the packet has a selective force header, we'll need
|
2012-06-25 08:27:37 +02:00
|
|
|
// to update it
|
2015-02-28 20:51:15 +01:00
|
|
|
if (h->extension == utp_sack)
|
2012-06-25 08:27:37 +02:00
|
|
|
{
|
|
|
|
sack = ptr[1];
|
2012-10-20 16:53:54 +02:00
|
|
|
// if we no longer have any out-of-order packets waiting
|
|
|
|
// to be delivered, there's no selective ack to be sent.
|
|
|
|
if (m_inbuf.size() == 0)
|
2012-06-25 08:27:37 +02:00
|
|
|
{
|
|
|
|
// we need to remove the sack header
|
2017-02-11 20:21:48 +01:00
|
|
|
remove_sack_header(p.get());
|
2012-06-25 08:27:37 +02:00
|
|
|
sack = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
sack = 0;
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
std::int32_t const size_left = std::min(p->allocated - p->size
|
2016-03-12 07:07:17 +01:00
|
|
|
, m_write_buffer_size);
|
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
write_payload(p->buf + p->size, size_left);
|
2016-12-13 16:30:36 +01:00
|
|
|
p->size += std::uint16_t(size_left);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: NAGLE appending %d bytes to nagle packet. new size: %d allocated: %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), size_left, p->size, p->allocated);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
// did we fill up the whole mtu?
|
|
|
|
// if we didn't, we may still send it if there's
|
|
|
|
// no bytes in flight
|
|
|
|
if (m_bytes_in_flight > 0
|
2017-02-11 20:21:48 +01:00
|
|
|
&& p->size < std::min(p->allocated, m_mtu_floor)
|
2012-06-28 05:53:52 +02:00
|
|
|
&& !force
|
2012-06-25 08:27:37 +02:00
|
|
|
&& m_nagle)
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
// the packet is still not a full MSS, so put it back into the nagle
|
|
|
|
// packet
|
|
|
|
m_nagle_packet = std::move(p);
|
2012-06-25 08:27:37 +02:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
packet_size = p->size;
|
|
|
|
payload_size = p->size - p->header_size;
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (sack)
|
|
|
|
{
|
2016-12-13 16:30:36 +01:00
|
|
|
*ptr++ = std::uint8_t(close_reason ? utp_close_reason : utp_no_extension);
|
2016-11-25 17:17:25 +01:00
|
|
|
*ptr++ = std::uint8_t(sack); // bytes for SACK bitfield
|
2010-11-29 02:33:05 +01:00
|
|
|
write_sack(ptr, sack);
|
|
|
|
ptr += sack;
|
2012-06-25 08:27:37 +02:00
|
|
|
TORRENT_ASSERT(ptr <= p->buf + p->header_size);
|
|
|
|
}
|
|
|
|
|
2015-02-28 20:51:15 +01:00
|
|
|
if (close_reason)
|
|
|
|
{
|
|
|
|
*ptr++ = utp_no_extension;
|
|
|
|
*ptr++ = 4;
|
|
|
|
detail::write_uint32(close_reason, ptr);
|
|
|
|
}
|
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
if (m_bytes_in_flight > 0
|
2012-06-28 05:53:52 +02:00
|
|
|
&& p->size < p->allocated
|
|
|
|
&& !force
|
2012-06-25 08:27:37 +02:00
|
|
|
&& m_nagle)
|
|
|
|
{
|
|
|
|
// this is nagle. If we don't have a full packet
|
|
|
|
// worth of payload to send AND we have at least
|
|
|
|
// one outstanding packet, hold off. Once the
|
|
|
|
// outstanding packet is acked, we'll send this
|
|
|
|
// payload
|
|
|
|
UTP_LOGV("%8p: NAGLE not enough payload send_buffer_size:%d cwnd:%d "
|
2016-03-12 07:07:17 +01:00
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d effective_mtu:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_write_buffer_size, int(m_cwnd >> 16)
|
2016-03-12 07:07:17 +01:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu, effective_mtu);
|
2017-02-11 20:21:48 +01:00
|
|
|
TORRENT_ASSERT(!m_nagle_packet);
|
2012-06-28 05:53:52 +02:00
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
2017-02-11 20:21:48 +01:00
|
|
|
m_nagle_packet = std::move(p);
|
2012-06-25 08:27:37 +02:00
|
|
|
return false;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// for ST_DATA packets, payload size is 0. Such packets do not have unique
|
|
|
|
// sequence numbers and should never be used as mtu probes
|
2016-05-02 06:17:17 +02:00
|
|
|
if ((mtu_probe || p->mtu_probe) && payload_size > m_mtu_floor)
|
2012-06-28 05:53:52 +02:00
|
|
|
{
|
|
|
|
p->mtu_probe = true;
|
|
|
|
m_mtu_seq = m_seq_nr;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
p->mtu_probe = false;
|
|
|
|
}
|
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
h->timestamp_difference_microseconds = m_reply_micro;
|
2018-04-12 16:21:20 +02:00
|
|
|
h->wnd_size = static_cast<std::uint32_t>(std::max(
|
|
|
|
m_receive_buffer_capacity - m_buffered_incoming_bytes
|
2017-02-27 17:50:53 +01:00
|
|
|
- m_receive_buffer_size, 0));
|
2012-06-25 08:27:37 +02:00
|
|
|
h->ack_nr = m_ack_nr;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-28 05:53:52 +02:00
|
|
|
// if this is a FIN packet, override the type
|
|
|
|
if (flags & pkt_fin)
|
|
|
|
h->type_ver = (ST_FIN << 4) | 1;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// fill in the timestamp as late as possible
|
2016-12-10 20:15:25 +01:00
|
|
|
time_point const now = clock_type::now();
|
2010-11-29 02:33:05 +01:00
|
|
|
p->send_time = now;
|
2016-06-18 20:01:38 +02:00
|
|
|
h->timestamp_microseconds = std::uint32_t(
|
2014-07-06 21:18:00 +02:00
|
|
|
total_microseconds(now.time_since_epoch()) & 0xffffffff);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
2012-06-25 07:38:29 +02:00
|
|
|
UTP_LOG("%8p: sending packet seq_nr:%d ack_nr:%d type:%s "
|
2010-11-29 02:33:05 +01:00
|
|
|
"id:%d target:%s size:%d error:%s send_buffer_size:%d cwnd:%d "
|
2012-06-25 07:38:29 +02:00
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d timestamp:%u time_diff:%u "
|
|
|
|
"mtu_probe:%d extension:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(h->seq_nr), int(h->ack_nr), packet_type_names[h->get_type()]
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_send_id, print_endpoint(udp::endpoint(m_remote_address, m_port)).c_str()
|
2012-06-29 21:40:27 +02:00
|
|
|
, p->size, m_error.message().c_str(), m_write_buffer_size, int(m_cwnd >> 16)
|
2016-06-18 20:01:38 +02:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu, std::uint32_t(h->timestamp_microseconds)
|
|
|
|
, std::uint32_t(h->timestamp_difference_microseconds), int(p->mtu_probe)
|
2012-06-25 07:38:29 +02:00
|
|
|
, h->extension);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
error_code ec;
|
2017-04-26 05:00:41 +02:00
|
|
|
m_sm.send_packet(m_sock, udp::endpoint(m_remote_address, m_port)
|
2015-08-16 18:17:23 +02:00
|
|
|
, reinterpret_cast<char const*>(h), p->size, ec
|
2017-07-26 19:38:40 +02:00
|
|
|
, p->mtu_probe ? udp_socket::dont_fragment : udp_send_flags_t{});
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
++m_out_packets;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_packets_out);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
if (ec == error::message_size)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2013-02-25 06:09:35 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2016-03-12 07:07:17 +01:00
|
|
|
UTP_LOGV("%8p: error sending packet: %s\n"
|
|
|
|
, static_cast<void*>(this)
|
|
|
|
, ec.message().c_str());
|
2013-02-25 06:09:35 +01:00
|
|
|
#endif
|
|
|
|
// if we fail even though this is not a probe, we're screwed
|
|
|
|
// since we'd have to repacketize
|
|
|
|
TORRENT_ASSERT(p->mtu_probe);
|
2012-06-29 21:40:27 +02:00
|
|
|
m_mtu_ceiling = p->size - 1;
|
|
|
|
if (m_mtu_floor > m_mtu_ceiling) m_mtu_floor = m_mtu_ceiling;
|
2010-11-29 02:33:05 +01:00
|
|
|
update_mtu_limits();
|
2014-01-20 10:20:47 +01:00
|
|
|
// resend the packet immediately without
|
2010-11-29 02:33:05 +01:00
|
|
|
// it being an MTU probe
|
2013-02-25 06:09:35 +01:00
|
|
|
p->mtu_probe = false;
|
2016-03-12 07:07:17 +01:00
|
|
|
m_mtu_seq = 0;
|
2013-02-25 06:09:35 +01:00
|
|
|
ec.clear();
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: re-sending\n", static_cast<void*>(this));
|
2013-02-25 06:09:35 +01:00
|
|
|
#endif
|
2017-04-26 05:00:41 +02:00
|
|
|
m_sm.send_packet(m_sock, udp::endpoint(m_remote_address, m_port)
|
2017-07-26 19:38:40 +02:00
|
|
|
, reinterpret_cast<char const*>(h), p->size, ec, {});
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2013-02-25 06:09:35 +01:00
|
|
|
|
|
|
|
if (ec == error::would_block || ec == error::try_again)
|
2012-07-01 20:44:46 +02:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: socket stalled\n", static_cast<void*>(this));
|
2012-07-01 20:44:46 +02:00
|
|
|
#endif
|
|
|
|
if (!m_stalled)
|
|
|
|
{
|
|
|
|
m_stalled = true;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.subscribe_writable(this);
|
2012-07-01 20:44:46 +02:00
|
|
|
}
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
else if (ec)
|
|
|
|
{
|
|
|
|
m_error = ec;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
2017-02-11 20:21:48 +01:00
|
|
|
release_packet(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
if (!m_stalled)
|
|
|
|
++p->num_transmissions;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// if we have payload, we need to save the packet until it's acked
|
|
|
|
// and progress m_seq_nr
|
2012-06-28 05:53:52 +02:00
|
|
|
if (p->size > p->header_size)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-28 05:53:52 +02:00
|
|
|
// if we're sending a payload packet, there should not
|
|
|
|
// be a nagle packet waiting for more data
|
2017-02-11 20:21:48 +01:00
|
|
|
TORRENT_ASSERT(!m_nagle_packet);
|
2012-06-28 05:53:52 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
#if !TORRENT_UT_SEQ
|
|
|
|
// if the other end closed the connection immediately
|
|
|
|
// our FIN packet will end up having the same sequence
|
|
|
|
// number as the SYN, so this assert is invalid
|
|
|
|
TORRENT_ASSERT(!m_outbuf.at(m_seq_nr));
|
|
|
|
#endif
|
2012-06-28 05:53:52 +02:00
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
2014-01-18 09:15:13 +01:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// 0 is a special sequence number, since it's also used as "uninitialized".
|
|
|
|
// we never send an mtu probe for sequence number 0
|
|
|
|
TORRENT_ASSERT(p->mtu_probe == (m_seq_nr == m_mtu_seq)
|
|
|
|
|| m_seq_nr == 0);
|
|
|
|
|
2014-01-18 09:15:13 +01:00
|
|
|
// release the buffer, we're saving it in the circular
|
|
|
|
// buffer of outgoing packets
|
2017-02-11 20:21:48 +01:00
|
|
|
int const new_in_flight = p->size - p->header_size;
|
|
|
|
packet_ptr old = m_outbuf.insert(m_seq_nr, std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
if (old)
|
|
|
|
{
|
2016-07-02 06:41:48 +02:00
|
|
|
// TORRENT_ASSERT(reinterpret_cast<utp_header*>(old->buf)->seq_nr == m_seq_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (!old->need_resend) m_bytes_in_flight -= old->size - old->header_size;
|
2017-02-11 20:21:48 +01:00
|
|
|
release_packet(std::move(old));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2012-06-27 01:52:48 +02:00
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
m_seq_nr = (m_seq_nr + 1) & ACK_MASK;
|
|
|
|
TORRENT_ASSERT(payload_size >= 0);
|
2017-02-11 20:21:48 +01:00
|
|
|
m_bytes_in_flight += new_in_flight;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2012-06-28 05:53:52 +02:00
|
|
|
else
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(h->seq_nr == m_seq_nr);
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
// if the socket is stalled, always return false, don't
|
|
|
|
// try to write more packets. We'll keep writing once
|
|
|
|
// the underlying UDP socket becomes writable
|
|
|
|
return m_write_buffer_size > 0 && !m_cwnd_full && !m_stalled;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// size is in bytes
|
2017-02-18 12:52:46 +01:00
|
|
|
void utp_socket_impl::write_sack(std::uint8_t* buf, int const size) const
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(m_inbuf.size());
|
|
|
|
int ack_nr = (m_ack_nr + 2) & ACK_MASK;
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t* end = buf + size;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
for (; buf != end; ++buf)
|
|
|
|
{
|
|
|
|
*buf = 0;
|
|
|
|
int mask = 1;
|
|
|
|
for (int i = 0; i < 8; ++i)
|
|
|
|
{
|
2017-02-18 12:52:46 +01:00
|
|
|
if (m_inbuf.at(aux::numeric_cast<packet_buffer::index_type>(ack_nr))) *buf |= mask;
|
2010-11-29 02:33:05 +01:00
|
|
|
mask <<= 1;
|
|
|
|
ack_nr = (ack_nr + 1) & ACK_MASK;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool utp_socket_impl::resend_packet(packet* p, bool fast_resend)
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// for fast re-sends the packet hasn't been marked as needing resending
|
|
|
|
TORRENT_ASSERT(p->need_resend || fast_resend);
|
|
|
|
|
2012-07-25 18:14:21 +02:00
|
|
|
if (m_error) return false;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-26 05:41:04 +02:00
|
|
|
if (((m_acked_seq_nr + 1) & ACK_MASK) == m_mtu_seq
|
2010-11-29 02:33:05 +01:00
|
|
|
&& m_mtu_seq != 0)
|
|
|
|
{
|
2012-06-26 05:41:04 +02:00
|
|
|
m_mtu_seq = 0;
|
|
|
|
p->mtu_probe = false;
|
2010-11-29 02:33:05 +01:00
|
|
|
// we got multiple acks for the packet before our probe, assume
|
|
|
|
// it was dropped because it was too big
|
2012-06-29 21:40:27 +02:00
|
|
|
m_mtu_ceiling = p->size - 1;
|
2010-11-29 02:33:05 +01:00
|
|
|
update_mtu_limits();
|
|
|
|
}
|
|
|
|
|
|
|
|
// we can only resend the packet if there's
|
|
|
|
// enough space in our congestion window
|
2013-02-25 06:09:35 +01:00
|
|
|
// since we can't re-packetize, some packets that are
|
|
|
|
// larger than the congestion window must be allowed through
|
|
|
|
// but only if we don't have any outstanding bytes
|
2018-02-20 14:14:09 +01:00
|
|
|
int const window_size_left = std::min(int(m_cwnd >> 16), int(m_adv_wnd)) - m_bytes_in_flight;
|
2013-02-25 06:09:35 +01:00
|
|
|
if (!fast_resend
|
|
|
|
&& p->size - p->header_size > window_size_left
|
|
|
|
&& m_bytes_in_flight > 0)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2011-11-05 10:30:38 +01:00
|
|
|
m_cwnd_full = true;
|
2010-11-29 02:33:05 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// plus one since we have fast-resend as well, which doesn't
|
|
|
|
// necessarily trigger by a timeout
|
2017-02-18 21:40:19 +01:00
|
|
|
TORRENT_ASSERT(p->num_transmissions < m_sm.num_resends() + 1);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(p->size - p->header_size >= 0);
|
|
|
|
if (p->need_resend) m_bytes_in_flight += p->size - p->header_size;
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_packet_resend);
|
|
|
|
if (fast_resend) m_sm.inc_stats_counter(counters::utp_fast_retransmit);
|
2013-09-14 12:06:48 +02:00
|
|
|
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2013-12-29 20:35:39 +01:00
|
|
|
if (fast_resend) ++p->num_fast_resend;
|
|
|
|
#endif
|
2010-11-29 02:33:05 +01:00
|
|
|
p->need_resend = false;
|
2018-01-11 01:35:15 +01:00
|
|
|
auto* h = reinterpret_cast<utp_header*>(p->buf);
|
2010-11-29 02:33:05 +01:00
|
|
|
// update packet header
|
|
|
|
h->timestamp_difference_microseconds = m_reply_micro;
|
2015-03-12 05:34:54 +01:00
|
|
|
p->send_time = clock_type::now();
|
2016-06-18 20:01:38 +02:00
|
|
|
h->timestamp_microseconds = std::uint32_t(
|
2014-07-06 21:18:00 +02:00
|
|
|
total_microseconds(p->send_time.time_since_epoch()) & 0xffffffff);
|
2012-06-25 08:27:37 +02:00
|
|
|
|
|
|
|
// if the packet has a selective ack header, we'll need
|
|
|
|
// to update it
|
2015-02-28 20:51:15 +01:00
|
|
|
if (h->extension == utp_sack && h->ack_nr != m_ack_nr)
|
2010-12-05 04:03:56 +01:00
|
|
|
{
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t* ptr = p->buf + sizeof(utp_header);
|
2012-06-25 08:27:37 +02:00
|
|
|
int sack_size = ptr[1];
|
|
|
|
if (m_inbuf.size())
|
|
|
|
{
|
|
|
|
// update the sack header
|
|
|
|
write_sack(ptr + 2, sack_size);
|
|
|
|
TORRENT_ASSERT(ptr + sack_size + 2 <= p->buf + p->header_size);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
remove_sack_header(p);
|
|
|
|
}
|
2010-12-05 04:03:56 +01:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-25 08:27:37 +02:00
|
|
|
h->ack_nr = m_ack_nr;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
error_code ec;
|
2017-04-26 05:00:41 +02:00
|
|
|
m_sm.send_packet(m_sock, udp::endpoint(m_remote_address, m_port)
|
2015-08-19 01:39:01 +02:00
|
|
|
, reinterpret_cast<char const*>(p->buf), p->size, ec);
|
2010-11-29 02:33:05 +01:00
|
|
|
++m_out_packets;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_packets_out);
|
2013-09-14 12:06:48 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: re-sending packet seq_nr:%d ack_nr:%d type:%s "
|
|
|
|
"id:%d target:%s size:%d error:%s send_buffer_size:%d cwnd:%d "
|
|
|
|
"adv_wnd:%d in-flight:%d mtu:%d timestamp:%u time_diff:%u\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(h->seq_nr), int(h->ack_nr), packet_type_names[h->get_type()]
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_send_id, print_endpoint(udp::endpoint(m_remote_address, m_port)).c_str()
|
|
|
|
, p->size, ec.message().c_str(), m_write_buffer_size, int(m_cwnd >> 16)
|
2016-06-18 20:01:38 +02:00
|
|
|
, m_adv_wnd, m_bytes_in_flight, m_mtu, std::uint32_t(h->timestamp_microseconds)
|
|
|
|
, std::uint32_t(h->timestamp_difference_microseconds));
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
2012-11-20 03:08:12 +01:00
|
|
|
if (ec == error::would_block || ec == error::try_again)
|
2012-07-01 20:44:46 +02:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: socket stalled\n", static_cast<void*>(this));
|
2012-07-01 20:44:46 +02:00
|
|
|
#endif
|
|
|
|
if (!m_stalled)
|
|
|
|
{
|
|
|
|
m_stalled = true;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.subscribe_writable(this);
|
2012-07-01 20:44:46 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (ec)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
m_error = ec;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-07-01 20:44:46 +02:00
|
|
|
if (!m_stalled)
|
|
|
|
++p->num_transmissions;
|
|
|
|
|
|
|
|
return !m_stalled;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-26 18:08:35 +01:00
|
|
|
void utp_socket_impl::experienced_loss(std::uint32_t const seq_nr)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// the window size could go below one MMS here, if it does,
|
|
|
|
// we'll get a timeout in about one second
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_packet_loss);
|
2016-03-12 07:07:17 +01:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// since loss often comes in bursts, we only cut the
|
|
|
|
// window in half once per RTT. This is implemented
|
|
|
|
// by limiting which packets can cause us to cut the
|
|
|
|
// window size. The first packet that's lost will
|
|
|
|
// update the limit to the last sequence number we sent.
|
|
|
|
// i.e. only packet sent after this loss can cause another
|
2013-09-14 12:06:48 +02:00
|
|
|
// window size cut. The +1 is to turn the comparison into
|
|
|
|
// less than or equal to. If we experience loss of the
|
|
|
|
// same packet again, ignore it.
|
|
|
|
if (compare_less_wrap(seq_nr, m_loss_seq_nr + 1, ACK_MASK)) return;
|
2015-07-20 17:20:17 +02:00
|
|
|
|
|
|
|
// cut window size in 2
|
2017-02-18 21:40:19 +01:00
|
|
|
m_cwnd = std::max(m_cwnd * m_sm.loss_multiplier() / 100
|
2016-10-08 20:17:51 +02:00
|
|
|
, std::int64_t(m_mtu) * (1 << 16));
|
2015-07-20 17:20:17 +02:00
|
|
|
m_loss_seq_nr = m_seq_nr;
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: Lost packet %d caused cwnd cut\n", static_cast<void*>(this), seq_nr);
|
2015-07-20 17:20:17 +02:00
|
|
|
|
2014-08-20 09:02:03 +02:00
|
|
|
// if we happen to be in slow-start mode, we need to leave it
|
2015-07-20 17:20:17 +02:00
|
|
|
// note that we set ssthres to the window size _after_ reducing it. Next slow
|
|
|
|
// start should end before we over shoot.
|
2014-08-20 09:02:03 +02:00
|
|
|
if (m_slow_start)
|
|
|
|
{
|
2016-12-05 14:39:53 +01:00
|
|
|
m_ssthres = std::int32_t(m_cwnd >> 16);
|
2014-08-20 09:02:03 +02:00
|
|
|
m_slow_start = false;
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: experienced loss, slow_start -> 0\n", static_cast<void*>(this));
|
2014-08-20 09:02:03 +02:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2015-01-04 22:31:02 +01:00
|
|
|
void utp_socket_impl::set_state(int s)
|
|
|
|
{
|
|
|
|
if (s == m_state) return;
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::num_utp_idle + m_state, -1);
|
2016-11-25 17:17:25 +01:00
|
|
|
m_state = std::uint8_t(s);
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::num_utp_idle + m_state, 1);
|
2015-01-04 22:31:02 +01:00
|
|
|
}
|
|
|
|
|
2010-12-05 04:03:56 +01:00
|
|
|
void utp_socket_impl::maybe_inc_acked_seq_nr()
|
|
|
|
{
|
2015-07-20 17:20:17 +02:00
|
|
|
#ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
2015-07-20 17:20:17 +02:00
|
|
|
#endif
|
2012-06-26 05:42:01 +02:00
|
|
|
|
2010-12-05 04:03:56 +01:00
|
|
|
bool incremented = false;
|
|
|
|
// don't pass m_seq_nr, since we move into sequence
|
|
|
|
// numbers that haven't been sent yet, and aren't
|
|
|
|
// supposed to be in m_outbuf
|
2011-11-06 01:11:33 +01:00
|
|
|
// if the slot in m_outbuf is 0, it means the
|
|
|
|
// packet has been ACKed and removed from the send buffer
|
2010-12-05 04:03:56 +01:00
|
|
|
while (((m_acked_seq_nr + 1) & ACK_MASK) != m_seq_nr
|
2016-07-09 22:26:26 +02:00
|
|
|
&& m_outbuf.at((m_acked_seq_nr + 1) & ACK_MASK) == nullptr)
|
2010-12-05 04:03:56 +01:00
|
|
|
{
|
|
|
|
// increment the fast resend sequence number
|
|
|
|
if (m_fast_resend_seq_nr == m_acked_seq_nr)
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
|
|
|
|
|
|
|
m_acked_seq_nr = (m_acked_seq_nr + 1) & ACK_MASK;
|
|
|
|
incremented = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!incremented) return;
|
|
|
|
|
|
|
|
// update loss seq number if it's less than the packet
|
|
|
|
// that was just acked. If loss seq nr is greater, it suggests
|
|
|
|
// that we're still in a window that has experienced loss
|
|
|
|
if (compare_less_wrap(m_loss_seq_nr, m_acked_seq_nr, ACK_MASK))
|
|
|
|
m_loss_seq_nr = m_acked_seq_nr;
|
|
|
|
m_duplicate_acks = 0;
|
|
|
|
}
|
|
|
|
|
2017-02-18 19:16:55 +01:00
|
|
|
// returns RTT
|
|
|
|
std::uint32_t utp_socket_impl::ack_packet(packet_ptr p, time_point const& receive_time
|
|
|
|
, std::uint16_t seq_nr)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-07-20 17:20:17 +02:00
|
|
|
#ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
2015-07-20 17:20:17 +02:00
|
|
|
#endif
|
2012-06-26 05:42:01 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(p);
|
2010-12-05 04:03:56 +01:00
|
|
|
|
|
|
|
// verify that the packet we're removing was in fact sent
|
|
|
|
// with the sequence number we expect
|
2016-07-02 06:41:48 +02:00
|
|
|
// TORRENT_ASSERT(reinterpret_cast<utp_header*>(p->buf)->seq_nr == seq_nr);
|
2010-12-05 04:03:56 +01:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (!p->need_resend)
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(m_bytes_in_flight >= p->size - p->header_size);
|
|
|
|
m_bytes_in_flight -= p->size - p->header_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (seq_nr == m_mtu_seq && m_mtu_seq != 0)
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(p->mtu_probe);
|
|
|
|
// our mtu probe was acked!
|
2017-02-11 20:21:48 +01:00
|
|
|
m_mtu_floor = std::max(m_mtu_floor, p->size);
|
2012-06-30 10:06:20 +02:00
|
|
|
if (m_mtu_ceiling < m_mtu_floor) m_mtu_ceiling = m_mtu_floor;
|
2010-11-29 02:33:05 +01:00
|
|
|
update_mtu_limits();
|
|
|
|
}
|
|
|
|
|
|
|
|
// increment the acked sequence number counter
|
2010-12-05 04:03:56 +01:00
|
|
|
maybe_inc_acked_seq_nr();
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t rtt = std::uint32_t(total_microseconds(receive_time - p->send_time));
|
2010-11-29 02:33:05 +01:00
|
|
|
if (receive_time < p->send_time)
|
|
|
|
{
|
|
|
|
// this means our clock is not monotonic. Just assume the RTT was 100 ms
|
|
|
|
rtt = 100000;
|
|
|
|
|
2016-04-26 06:37:47 +02:00
|
|
|
// the clock for this platform is not monotonic!
|
2016-05-02 18:36:21 +02:00
|
|
|
TORRENT_ASSERT_FAIL();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
UTP_LOGV("%8p: acked packet %d (%d bytes) (rtt:%u)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), seq_nr, p->size - p->header_size, rtt / 1000);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
m_rtt.add_sample(rtt / 1000);
|
2017-02-11 20:21:48 +01:00
|
|
|
release_packet(std::move(p));
|
2017-02-18 19:16:55 +01:00
|
|
|
return rtt;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
void utp_socket_impl::incoming(std::uint8_t const* buf, int size, packet_ptr p
|
2015-03-12 05:34:54 +01:00
|
|
|
, time_point /* now */)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-07-20 17:20:17 +02:00
|
|
|
#ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
2015-07-20 17:20:17 +02:00
|
|
|
#endif
|
2012-06-26 05:42:01 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
while (!m_read_buffer.empty())
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: incoming: have user buffer (%d)\n", static_cast<void*>(this), m_read_buffer_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (p)
|
|
|
|
{
|
|
|
|
buf = p->buf + p->header_size;
|
|
|
|
TORRENT_ASSERT(p->size - p->header_size >= size);
|
|
|
|
}
|
|
|
|
iovec_t* target = &m_read_buffer.front();
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
int const to_copy = std::min(size, aux::numeric_cast<int>(target->len));
|
|
|
|
TORRENT_ASSERT(to_copy >= 0);
|
|
|
|
std::memcpy(target->buf, buf, std::size_t(to_copy));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_read += to_copy;
|
2016-06-18 20:01:38 +02:00
|
|
|
target->buf = reinterpret_cast<std::uint8_t*>(target->buf) + to_copy;
|
2017-02-18 12:52:46 +01:00
|
|
|
target->len -= std::size_t(to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
buf += to_copy;
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: copied %d bytes into user receive buffer\n", static_cast<void*>(this), to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(m_read_buffer_size >= to_copy);
|
|
|
|
m_read_buffer_size -= to_copy;
|
|
|
|
size -= to_copy;
|
|
|
|
if (target->len == 0) m_read_buffer.erase(m_read_buffer.begin());
|
|
|
|
if (p)
|
|
|
|
{
|
2016-12-13 16:30:36 +01:00
|
|
|
p->header_size += std::uint16_t(to_copy);
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(p->header_size <= p->size);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (size == 0)
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
TORRENT_ASSERT(!p || p->header_size == p->size);
|
|
|
|
release_packet(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_read_buffer_size == 0);
|
|
|
|
|
|
|
|
if (!p)
|
|
|
|
{
|
|
|
|
TORRENT_ASSERT(buf);
|
2017-02-10 14:25:03 +01:00
|
|
|
p = acquire_packet(size);
|
2016-11-25 17:17:25 +01:00
|
|
|
p->size = std::uint16_t(size);
|
2010-11-29 02:33:05 +01:00
|
|
|
p->header_size = 0;
|
2017-02-18 12:52:46 +01:00
|
|
|
std::memcpy(p->buf, buf, aux::numeric_cast<std::size_t>(size));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
// save this packet until the client issues another read
|
|
|
|
m_receive_buffer_size += p->size - p->header_size;
|
2017-02-11 20:21:48 +01:00
|
|
|
m_receive_buffer.emplace_back(std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: incoming: saving packet in receive buffer (%d)\n", static_cast<void*>(this), m_receive_buffer_size);
|
2014-07-06 21:18:00 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
check_receive_buffers();
|
|
|
|
}
|
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
bool utp_socket_impl::cancel_handlers(error_code const& ec, bool shutdown)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(ec);
|
|
|
|
bool ret = m_read_handler || m_write_handler || m_connect_handler;
|
2016-03-12 07:07:17 +01:00
|
|
|
|
2011-08-24 07:13:25 +02:00
|
|
|
// calling the callbacks with m_userdata being 0 will just crash
|
2016-06-20 17:32:06 +02:00
|
|
|
TORRENT_ASSERT((ret && m_userdata != nullptr) || !ret);
|
2011-08-24 07:13:25 +02:00
|
|
|
|
2015-02-15 01:25:56 +01:00
|
|
|
bool read = m_read_handler;
|
|
|
|
bool write = m_write_handler;
|
|
|
|
bool connect = m_connect_handler;
|
|
|
|
m_read_handler = false;
|
|
|
|
m_write_handler = false;
|
|
|
|
m_connect_handler = false;
|
|
|
|
|
2017-09-17 13:12:54 +02:00
|
|
|
if (read) utp_stream::on_read(m_userdata, 0, ec, shutdown);
|
|
|
|
if (write) utp_stream::on_write(m_userdata, 0, ec, shutdown);
|
|
|
|
if (connect) utp_stream::on_connect(m_userdata, ec, shutdown);
|
2010-11-29 02:33:05 +01:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool utp_socket_impl::consume_incoming_data(
|
2017-02-18 12:52:46 +01:00
|
|
|
utp_header const* ph, std::uint8_t const* ptr, int const payload_size
|
2017-02-13 01:43:21 +01:00
|
|
|
, time_point const now)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (ph->get_type() != ST_DATA) return false;
|
|
|
|
|
|
|
|
if (m_eof && m_ack_nr == m_eof_seq_nr)
|
|
|
|
{
|
|
|
|
// What?! We've already received a FIN and everything up
|
|
|
|
// to it has been acked. Ignore this packet
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOG("%8p: ERROR: ignoring packet on shut down socket\n"
|
|
|
|
, static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2013-12-29 20:35:39 +01:00
|
|
|
if (m_read_buffer_size == 0
|
2018-03-07 16:44:35 +01:00
|
|
|
&& m_receive_buffer_size >= m_receive_buffer_capacity - m_buffered_incoming_bytes)
|
2013-12-29 20:35:39 +01:00
|
|
|
{
|
|
|
|
// if we don't have a buffer from the upper layer, and the
|
|
|
|
// number of queued up bytes, waiting for the upper layer,
|
2016-04-26 06:37:47 +02:00
|
|
|
// exceeds the advertised receive window, start ignoring
|
2013-12-29 20:35:39 +01:00
|
|
|
// more data packets
|
|
|
|
UTP_LOG("%8p: ERROR: our advertized window is not honored. "
|
2013-12-29 21:26:17 +01:00
|
|
|
"recv_buf: %d buffered_in: %d max_size: %d\n"
|
2018-03-07 16:44:35 +01:00
|
|
|
, static_cast<void*>(this), m_receive_buffer_size, m_buffered_incoming_bytes, m_receive_buffer_capacity);
|
2013-12-29 20:35:39 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (ph->seq_nr == ((m_ack_nr + 1) & ACK_MASK))
|
|
|
|
{
|
2017-02-13 01:43:21 +01:00
|
|
|
TORRENT_ASSERT(m_inbuf.at(m_ack_nr) == nullptr);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2018-03-07 16:44:35 +01:00
|
|
|
if (m_buffered_incoming_bytes + m_receive_buffer_size + payload_size > m_receive_buffer_capacity)
|
2013-09-25 03:37:12 +02:00
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: other end is not honoring our advertised window, dropping packet\n"
|
|
|
|
, static_cast<void*>(this));
|
2013-09-25 03:37:12 +02:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// we received a packet in order
|
2017-02-11 20:21:48 +01:00
|
|
|
incoming(ptr, payload_size, packet_ptr(), now);
|
2010-11-29 02:33:05 +01:00
|
|
|
m_ack_nr = (m_ack_nr + 1) & ACK_MASK;
|
|
|
|
|
|
|
|
// If this packet was previously in the reorder buffer
|
|
|
|
// it would have been acked when m_ack_nr-1 was acked.
|
2017-02-11 20:21:48 +01:00
|
|
|
TORRENT_ASSERT(!m_inbuf.at(m_ack_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: remove inbuf: %d (%d)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_ack_nr, int(m_inbuf.size()));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
for (;;)
|
|
|
|
{
|
|
|
|
int const next_ack_nr = (m_ack_nr + 1) & ACK_MASK;
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
packet_ptr p = m_inbuf.remove(aux::numeric_cast<packet_buffer::index_type>(next_ack_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2013-09-25 03:37:12 +02:00
|
|
|
if (!p) break;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-13 01:43:21 +01:00
|
|
|
TORRENT_ASSERT(p->size >= p->header_size);
|
|
|
|
int const size = p->size - p->header_size;
|
|
|
|
m_buffered_incoming_bytes -= size;
|
|
|
|
incoming(nullptr, size, std::move(p), now);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2016-11-25 17:17:25 +01:00
|
|
|
m_ack_nr = std::uint16_t(next_ack_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: reordered remove inbuf: %d (%d)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_ack_nr, int(m_inbuf.size()));
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// this packet was received out of order. Stick it in the
|
|
|
|
// reorder buffer until it can be delivered in order
|
|
|
|
|
|
|
|
// have we already received this packet and passed it on
|
|
|
|
// to the client?
|
|
|
|
if (!compare_less_wrap(m_ack_nr, ph->seq_nr, ACK_MASK))
|
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: already received seq_nr: %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// do we already have this packet? If so, just ignore it
|
|
|
|
if (m_inbuf.at(ph->seq_nr))
|
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: already received seq_nr: %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr));
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-03-07 16:44:35 +01:00
|
|
|
if (m_buffered_incoming_bytes + m_receive_buffer_size + payload_size > m_receive_buffer_capacity)
|
2013-09-25 03:37:12 +02:00
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: other end is not honoring our advertised window, dropping packet %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr));
|
2013-09-25 03:37:12 +02:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// we don't need to save the packet header, just the payload
|
2017-02-11 20:21:48 +01:00
|
|
|
packet_ptr p = acquire_packet(payload_size);
|
2016-11-25 17:17:25 +01:00
|
|
|
p->size = std::uint16_t(payload_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
p->header_size = 0;
|
|
|
|
p->num_transmissions = 0;
|
2016-07-02 01:46:59 +02:00
|
|
|
#if TORRENT_USE_ASSERTS
|
2013-12-29 20:35:39 +01:00
|
|
|
p->num_fast_resend = 0;
|
|
|
|
#endif
|
2010-11-29 02:33:05 +01:00
|
|
|
p->need_resend = false;
|
2017-02-18 12:52:46 +01:00
|
|
|
std::memcpy(p->buf, ptr, aux::numeric_cast<std::size_t>(payload_size));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_buffered_incoming_bytes += p->size;
|
2017-02-11 20:21:48 +01:00
|
|
|
m_inbuf.insert(ph->seq_nr, std::move(p));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
UTP_LOGV("%8p: out of order. insert inbuf: %d (%d) m_ack_nr: %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr), int(m_inbuf.size()), m_ack_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// returns true of the socket was closed
|
|
|
|
bool utp_socket_impl::test_socket_state()
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// if the socket is in a state where it's dead, just waiting to
|
|
|
|
// tell the client that it's closed. Do that and transition into
|
|
|
|
// the deleted state, where it will be deleted
|
|
|
|
// it might be possible to get here twice, in which we need to
|
|
|
|
// cancel any new handlers as well, even though we're already
|
|
|
|
// in the delete state
|
|
|
|
if (!m_error) return false;
|
|
|
|
TORRENT_ASSERT(m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE);
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: state:%s error:%s\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), socket_state_names[m_state], m_error.message().c_str());
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
if (cancel_handlers(m_error, true))
|
|
|
|
{
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_DELETE);
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: state:%s\n", static_cast<void*>(this), socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void utp_socket_impl::init_mtu(int link_mtu, int utp_mtu)
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2015-08-20 02:40:21 +02:00
|
|
|
if (link_mtu > TORRENT_ETHERNET_MTU)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
// we can't use larger packets than this since we're
|
|
|
|
// not allocating any more memory for socket buffers
|
|
|
|
int decrease = link_mtu - TORRENT_ETHERNET_MTU;
|
|
|
|
utp_mtu -= decrease;
|
|
|
|
link_mtu -= decrease;
|
|
|
|
}
|
|
|
|
|
2012-08-12 17:54:59 +02:00
|
|
|
// set the ceiling to what we found out from the interface
|
2016-11-25 17:17:25 +01:00
|
|
|
m_mtu_ceiling = std::uint16_t(utp_mtu);
|
2012-08-12 17:54:59 +02:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
// start in the middle of the PMTU search space
|
|
|
|
m_mtu = (m_mtu_ceiling + m_mtu_floor) / 2;
|
2012-08-12 17:54:59 +02:00
|
|
|
if (m_mtu > m_mtu_ceiling) m_mtu = m_mtu_ceiling;
|
2016-11-25 17:17:25 +01:00
|
|
|
if (m_mtu_floor > utp_mtu) m_mtu_floor = std::uint16_t(utp_mtu);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// if the window size is smaller than one packet size
|
|
|
|
// set it to one
|
2016-10-08 20:17:51 +02:00
|
|
|
if ((m_cwnd >> 16) < m_mtu) m_cwnd = std::int64_t(m_mtu) * (1 << 16);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
UTP_LOGV("%8p: initializing MTU to: %d [%d, %d]\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_mtu, m_mtu_floor, m_mtu_ceiling);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// return false if this is an invalid packet
|
2016-07-22 18:31:42 +02:00
|
|
|
bool utp_socket_impl::incoming_packet(span<std::uint8_t const> buf
|
2015-03-12 05:34:54 +01:00
|
|
|
, udp::endpoint const& ep, time_point receive_time)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2018-01-11 01:35:15 +01:00
|
|
|
auto const* ph = reinterpret_cast<utp_header const*>(buf.data());
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_packets_in);
|
2013-09-14 12:06:48 +02:00
|
|
|
|
2018-03-21 11:14:25 +01:00
|
|
|
if (buf.size() < sizeof(utp_header))
|
2018-02-05 01:29:14 +01:00
|
|
|
{
|
|
|
|
UTP_LOG("%8p: ERROR: incoming packet size too small:%d (ignored)\n"
|
2018-03-21 11:14:25 +01:00
|
|
|
, static_cast<void*>(this), int(buf.size()));
|
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2018-02-05 01:29:14 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (ph->get_version() != 1)
|
|
|
|
{
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: incoming packet version:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->get_version()));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// SYN packets have special (reverse) connection ids
|
|
|
|
if (ph->get_type() != ST_SYN && ph->connection_id != m_recv_id)
|
|
|
|
{
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: incoming packet id:%d expected:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->connection_id), int(m_recv_id));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ph->get_type() >= NUM_TYPES)
|
|
|
|
{
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: incoming packet type:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->get_type()));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (m_state == UTP_STATE_NONE && ph->get_type() == ST_SYN)
|
|
|
|
{
|
|
|
|
m_remote_address = ep.address();
|
|
|
|
m_port = ep.port();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (m_state != UTP_STATE_NONE && ph->get_type() == ST_SYN)
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOG("%8p: ERROR: incoming packet type:ST_SYN (ignored)\n"
|
|
|
|
, static_cast<void*>(this));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool step = false;
|
|
|
|
if (receive_time - m_last_history_step > minutes(1))
|
|
|
|
{
|
|
|
|
step = true;
|
|
|
|
m_last_history_step = receive_time;
|
|
|
|
}
|
|
|
|
|
|
|
|
// this is the difference between their send time and our receive time
|
|
|
|
// 0 means no sample yet
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t their_delay = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
if (ph->timestamp_microseconds != 0)
|
|
|
|
{
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t timestamp = std::uint32_t(total_microseconds(
|
2014-07-06 21:18:00 +02:00
|
|
|
receive_time.time_since_epoch()) & 0xffffffff);
|
2014-02-05 10:38:32 +01:00
|
|
|
m_reply_micro = timestamp - ph->timestamp_microseconds;
|
2017-02-27 17:50:53 +01:00
|
|
|
std::uint32_t const prev_base = m_their_delay_hist.initialized() ? m_their_delay_hist.base() : 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
their_delay = m_their_delay_hist.add_sample(m_reply_micro, step);
|
2017-02-27 17:50:53 +01:00
|
|
|
int const base_change = int(m_their_delay_hist.base() - prev_base);
|
2010-11-29 02:33:05 +01:00
|
|
|
UTP_LOGV("%8p: their_delay::add_sample:%u prev_base:%u new_base:%u\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_reply_micro, prev_base, m_their_delay_hist.base());
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2010-12-25 00:54:03 +01:00
|
|
|
if (prev_base && base_change < 0 && base_change > -10000 && m_delay_hist.initialized())
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
// their base delay went down. This is caused by clock drift. To compensate,
|
|
|
|
// adjust our base delay upwards
|
|
|
|
// don't adjust more than 10 ms. If the change is that big, something is probably wrong
|
|
|
|
m_delay_hist.adjust_base(-base_change);
|
|
|
|
}
|
|
|
|
|
|
|
|
UTP_LOGV("%8p: incoming packet reply_micro:%u base_change:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_reply_micro, prev_base ? base_change : 0);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// is this ACK valid? If the other end is ACKing
|
|
|
|
// a packet that hasn't been sent yet
|
|
|
|
// just ignore it. A 3rd party could easily inject a packet
|
|
|
|
// like this in a stream, don't sever it because of it.
|
|
|
|
// since m_seq_nr is the sequence number of the next packet
|
|
|
|
// we'll send (and m_seq_nr-1 was the last packet we sent),
|
|
|
|
// if the ACK we got is greater than the last packet we sent
|
|
|
|
// something is wrong.
|
|
|
|
// If our state is state_none, this packet must be a syn packet
|
|
|
|
// and the ack_nr should be ignored
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint16_t cmp_seq_nr = (m_seq_nr - 1) & ACK_MASK;
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UT_SEQ
|
|
|
|
if (m_state == UTP_STATE_SYN_SENT && ph->get_type() == ST_STATE)
|
|
|
|
cmp_seq_nr = m_seq_nr;
|
|
|
|
#endif
|
2015-07-19 03:55:26 +02:00
|
|
|
if ((m_state != UTP_STATE_NONE || ph->get_type() != ST_SYN)
|
|
|
|
&& (compare_less_wrap(cmp_seq_nr, ph->ack_nr, ACK_MASK)
|
|
|
|
|| compare_less_wrap(ph->ack_nr, m_acked_seq_nr
|
|
|
|
- dup_ack_limit, ACK_MASK)))
|
|
|
|
{
|
|
|
|
UTP_LOG("%8p: ERROR: incoming packet ack_nr:%d our seq_nr:%d our "
|
|
|
|
"acked_seq_nr:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->ack_nr), m_seq_nr, m_acked_seq_nr);
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_redundant_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// check to make sure the sequence number of this packet
|
|
|
|
// is reasonable. If it's a data packet and we've already
|
|
|
|
// received it, ignore it. This is either a stray old packet
|
|
|
|
// that finally made it here (after having been re-sent) or
|
|
|
|
// an attempt to interfere with the connection from a 3rd party
|
|
|
|
// in both cases, we can safely ignore the timestamp and ACK
|
|
|
|
// information in this packet
|
|
|
|
/*
|
|
|
|
// even if we've already received this packet, we need to
|
|
|
|
// send another ack to it, since it may be a resend caused by
|
|
|
|
// our ack getting dropped
|
|
|
|
if (m_state != UTP_STATE_SYN_SENT
|
|
|
|
&& ph->get_type() == ST_DATA
|
|
|
|
&& !compare_less_wrap(m_ack_nr, ph->seq_nr, ACK_MASK))
|
|
|
|
{
|
|
|
|
// we've already received this packet
|
|
|
|
UTP_LOGV("%8p: incoming packet seq_nr:%d our ack_nr:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr), m_ack_nr);
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_redundant_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
*/
|
|
|
|
|
|
|
|
// if the socket is closing, always ignore any packet
|
|
|
|
// with a higher sequence number than the FIN sequence number
|
|
|
|
if (m_eof && compare_less_wrap(m_eof_seq_nr, ph->seq_nr, ACK_MASK))
|
|
|
|
{
|
2013-12-29 20:35:39 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOG("%8p: ERROR: incoming packet type: %s seq_nr:%d eof_seq_nr:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), packet_type_names[ph->get_type()], int(ph->seq_nr), m_eof_seq_nr);
|
2013-12-29 20:35:39 +01:00
|
|
|
#endif
|
|
|
|
return true;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2013-09-14 12:06:48 +02:00
|
|
|
if (ph->get_type() == ST_DATA)
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_payload_pkts_in);
|
2013-09-14 12:06:48 +02:00
|
|
|
|
2018-03-01 10:41:29 +01:00
|
|
|
// the number of packets that'll fit in the reorder buffer
|
2018-04-12 16:21:20 +02:00
|
|
|
std::uint32_t const max_packets_reorder
|
|
|
|
= static_cast<std::uint32_t>(std::max(16, m_receive_buffer_capacity / 1100));
|
2018-03-01 10:41:29 +01:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_state != UTP_STATE_NONE
|
|
|
|
&& m_state != UTP_STATE_SYN_SENT
|
|
|
|
&& compare_less_wrap((m_ack_nr + max_packets_reorder) & ACK_MASK, ph->seq_nr, ACK_MASK))
|
|
|
|
{
|
|
|
|
// this is too far out to fit in our reorder buffer. Drop it
|
|
|
|
// This is either an attack to try to break the connection
|
2016-04-26 06:37:47 +02:00
|
|
|
// or a seriously damaged connection that lost a lot of
|
2010-11-29 02:33:05 +01:00
|
|
|
// packets. Neither is very likely, and it should be OK
|
|
|
|
// to drop the timestamp information.
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: incoming packet seq_nr:%d our ack_nr:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr), m_ack_nr);
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_redundant_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2013-12-29 20:35:39 +01:00
|
|
|
if (ph->get_type() == ST_RESET)
|
|
|
|
{
|
|
|
|
if (compare_less_wrap(cmp_seq_nr, ph->ack_nr, ACK_MASK))
|
|
|
|
{
|
|
|
|
UTP_LOG("%8p: ERROR: invalid RESET packet, ack_nr:%d our seq_nr:%d (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->ack_nr), m_seq_nr);
|
2013-12-29 20:35:39 +01:00
|
|
|
return true;
|
|
|
|
}
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: incoming packet type:RESET\n", static_cast<void*>(this));
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::connection_reset;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2013-12-29 20:35:39 +01:00
|
|
|
test_socket_state();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
++m_in_packets;
|
|
|
|
|
|
|
|
// this is a valid incoming packet, update the timeout timer
|
|
|
|
m_num_timeouts = 0;
|
|
|
|
m_timeout = receive_time + milliseconds(packet_timeout());
|
|
|
|
UTP_LOGV("%8p: updating timeout to: now + %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), packet_timeout());
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// the test for INT_MAX here is a work-around for a bug in uTorrent where
|
|
|
|
// it's sometimes sent as INT_MAX when it is in fact uninitialized
|
2016-06-18 20:01:38 +02:00
|
|
|
const std::uint32_t sample = ph->timestamp_difference_microseconds == INT_MAX
|
2010-11-29 02:33:05 +01:00
|
|
|
? 0 : ph->timestamp_difference_microseconds;
|
|
|
|
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t delay = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
if (sample != 0)
|
|
|
|
{
|
|
|
|
delay = m_delay_hist.add_sample(sample, step);
|
|
|
|
m_delay_sample_hist[m_delay_sample_idx++] = delay;
|
2016-10-08 18:12:33 +02:00
|
|
|
if (m_delay_sample_idx >= m_delay_sample_hist.size())
|
|
|
|
m_delay_sample_idx = 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
int acked_bytes = 0;
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_bytes_in_flight >= 0);
|
|
|
|
int prev_bytes_in_flight = m_bytes_in_flight;
|
|
|
|
|
|
|
|
m_adv_wnd = ph->wnd_size;
|
|
|
|
|
|
|
|
// if we get an ack for the same sequence number as
|
|
|
|
// was last ACKed, and we have outstanding packets,
|
2016-03-12 07:07:17 +01:00
|
|
|
// it counts as a duplicate ack. The reason to not count ST_DATA packets as
|
|
|
|
// duplicate ACKs is because we may be receiving a stream of those
|
|
|
|
// regardless of our outgoing traffic, which makes their ACK number not
|
|
|
|
// indicative of a dropped packet
|
|
|
|
if (ph->ack_nr == m_acked_seq_nr
|
|
|
|
&& m_outbuf.size()
|
|
|
|
&& ph->get_type() == ST_STATE)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
++m_duplicate_acks;
|
|
|
|
}
|
|
|
|
|
2017-02-18 12:52:46 +01:00
|
|
|
std::uint32_t min_rtt = std::numeric_limits<std::uint32_t>::max();
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
TORRENT_ASSERT(m_outbuf.at((m_acked_seq_nr + 1) & ACK_MASK) || ((m_seq_nr - m_acked_seq_nr) & ACK_MASK) <= 1);
|
|
|
|
|
|
|
|
// has this packet already been ACKed?
|
|
|
|
// if the ACK we just got is less than the max ACKed
|
|
|
|
// sequence number, it doesn't tell us anything.
|
|
|
|
// So, only act on it if the ACK is greater than the last acked
|
|
|
|
// sequence number
|
|
|
|
if (m_state != UTP_STATE_NONE && compare_less_wrap(m_acked_seq_nr, ph->ack_nr, ACK_MASK))
|
|
|
|
{
|
|
|
|
int const next_ack_nr = ph->ack_nr;
|
|
|
|
|
|
|
|
for (int ack_nr = (m_acked_seq_nr + 1) & ACK_MASK;
|
|
|
|
ack_nr != ((next_ack_nr + 1) & ACK_MASK);
|
|
|
|
ack_nr = (ack_nr + 1) & ACK_MASK)
|
|
|
|
{
|
|
|
|
if (m_fast_resend_seq_nr == ack_nr)
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
2017-02-18 12:52:46 +01:00
|
|
|
packet_ptr p = m_outbuf.remove(aux::numeric_cast<packet_buffer::index_type>(ack_nr));
|
2010-12-05 04:03:56 +01:00
|
|
|
|
|
|
|
if (!p) continue;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
acked_bytes += p->size - p->header_size;
|
2017-02-18 19:16:55 +01:00
|
|
|
std::uint32_t const rtt = ack_packet(std::move(p), receive_time, std::uint16_t(ack_nr));
|
|
|
|
min_rtt = std::min(min_rtt, rtt);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2010-12-05 04:03:56 +01:00
|
|
|
maybe_inc_acked_seq_nr();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// look for extended headers
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint8_t const* ptr = buf.data();
|
2016-05-02 18:36:21 +02:00
|
|
|
int const size = int(buf.size());
|
2010-11-29 02:33:05 +01:00
|
|
|
ptr += sizeof(utp_header);
|
|
|
|
|
2017-02-26 18:08:35 +01:00
|
|
|
std::uint8_t extension = ph->extension;
|
2010-11-29 02:33:05 +01:00
|
|
|
while (extension)
|
|
|
|
{
|
|
|
|
// invalid packet. It says it has an extension header
|
|
|
|
// but the packet is too short
|
2016-04-28 14:20:10 +02:00
|
|
|
if (ptr - buf.data() + 2 > size)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOG("%8p: ERROR: invalid extension header\n", static_cast<void*>(this));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
2017-02-26 18:08:35 +01:00
|
|
|
std::uint8_t const next_extension = *ptr++;
|
2017-02-27 17:50:53 +01:00
|
|
|
int const len = *ptr++;
|
2012-06-25 08:27:37 +02:00
|
|
|
if (len < 0)
|
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: invalid extension length:%d packet:%d\n"
|
2016-04-28 14:20:10 +02:00
|
|
|
, static_cast<void*>(this), len, int(ptr - buf.data()));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2012-06-25 08:27:37 +02:00
|
|
|
return true;
|
|
|
|
}
|
2017-02-26 18:08:35 +01:00
|
|
|
if (ptr - buf.data() + len > size)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: invalid extension header size:%d packet:%d\n"
|
2016-04-28 14:20:10 +02:00
|
|
|
, static_cast<void*>(this), len, int(ptr - buf.data()));
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_invalid_pkts_in);
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
switch(extension)
|
|
|
|
{
|
2015-02-28 20:51:15 +01:00
|
|
|
case utp_sack: // selective ACKs
|
2017-02-18 19:16:55 +01:00
|
|
|
{
|
|
|
|
std::uint32_t rtt;
|
|
|
|
std::tie(rtt, acked_bytes) = parse_sack(ph->ack_nr, ptr, len, receive_time);
|
|
|
|
min_rtt = std::min(min_rtt, rtt);
|
2010-11-29 02:33:05 +01:00
|
|
|
break;
|
2017-02-18 19:16:55 +01:00
|
|
|
}
|
2015-02-28 20:51:15 +01:00
|
|
|
case utp_close_reason:
|
|
|
|
parse_close_reason(ptr, len);
|
|
|
|
break;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
ptr += len;
|
|
|
|
extension = next_extension;
|
|
|
|
}
|
2015-07-19 03:55:26 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the send operation in parse_sack() may have set the socket to an error
|
|
|
|
// state, in which case we shouldn't continue
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
|
|
|
|
if (m_duplicate_acks >= dup_ack_limit
|
|
|
|
&& ((m_acked_seq_nr + 1) & ACK_MASK) == m_fast_resend_seq_nr)
|
|
|
|
{
|
|
|
|
// LOSS
|
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: Packet %d lost. (%d duplicate acks, trigger fast-resend)\n"
|
|
|
|
, static_cast<void*>(this), m_fast_resend_seq_nr, m_duplicate_acks);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// resend the lost packet
|
2015-08-17 15:01:43 +02:00
|
|
|
packet* p = m_outbuf.at(m_fast_resend_seq_nr);
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(p);
|
2013-12-29 20:35:39 +01:00
|
|
|
|
|
|
|
// don't fast-resend this again
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (p)
|
|
|
|
{
|
|
|
|
experienced_loss(m_fast_resend_seq_nr);
|
|
|
|
resend_packet(p, true);
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// ptr points to the payload of the packet
|
|
|
|
// size is the packet size, payload is the
|
|
|
|
// number of payload bytes are in this packet
|
2016-12-05 14:39:53 +01:00
|
|
|
const int header_size = int(ptr - buf.data());
|
2010-11-29 02:33:05 +01:00
|
|
|
const int payload_size = size - header_size;
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: incoming packet seq_nr:%d ack_nr:%d type:%s id:%d size:%d timestampdiff:%u timestamp:%u "
|
|
|
|
"our ack_nr:%d our seq_nr:%d our acked_seq_nr:%d our state:%s\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->seq_nr), int(ph->ack_nr), packet_type_names[ph->get_type()]
|
2016-06-18 20:01:38 +02:00
|
|
|
, int(ph->connection_id), payload_size, std::uint32_t(ph->timestamp_difference_microseconds)
|
|
|
|
, std::uint32_t(ph->timestamp_microseconds), m_ack_nr, m_seq_nr, m_acked_seq_nr, socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
if (ph->get_type() == ST_FIN)
|
|
|
|
{
|
|
|
|
// We ignore duplicate FIN packets, but we still need to ACK them.
|
|
|
|
if (ph->seq_nr == ((m_ack_nr + 1) & ACK_MASK)
|
|
|
|
|| ph->seq_nr == m_ack_nr)
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: FIN received in order\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// The FIN arrived in order, nothing else is in the
|
|
|
|
// reorder buffer.
|
|
|
|
|
|
|
|
// TORRENT_ASSERT(m_inbuf.size() == 0);
|
|
|
|
m_ack_nr = ph->seq_nr;
|
|
|
|
|
|
|
|
// Transition to UTP_STATE_FIN_SENT. The sent FIN is also an ack
|
|
|
|
// to the FIN we received. Once we're in UTP_STATE_FIN_SENT we
|
|
|
|
// just need to wait for our FIN to be acked.
|
|
|
|
|
|
|
|
if (m_state == UTP_STATE_FIN_SENT)
|
|
|
|
{
|
2012-07-01 20:44:46 +02:00
|
|
|
send_pkt(pkt_ack);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
send_fin();
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (m_eof)
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: duplicate FIN packet (ignoring)\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
m_eof = true;
|
|
|
|
m_eof_seq_nr = ph->seq_nr;
|
|
|
|
|
|
|
|
// we will respond with a fin once we have received everything up to m_eof_seq_nr
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (m_state)
|
|
|
|
{
|
|
|
|
case UTP_STATE_NONE:
|
|
|
|
{
|
|
|
|
if (ph->get_type() == ST_SYN)
|
|
|
|
{
|
|
|
|
// if we're in state_none, the only thing
|
|
|
|
// we accept are SYN packets.
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_CONNECTED);
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
m_remote_address = ep.address();
|
|
|
|
m_port = ep.port();
|
|
|
|
|
|
|
|
m_ack_nr = ph->seq_nr;
|
2016-11-25 17:17:25 +01:00
|
|
|
m_seq_nr = std::uint16_t(random(0xffff));
|
2010-11-29 02:33:05 +01:00
|
|
|
m_acked_seq_nr = (m_seq_nr - 1) & ACK_MASK;
|
|
|
|
m_loss_seq_nr = m_acked_seq_nr;
|
2011-11-06 01:11:33 +01:00
|
|
|
m_fast_resend_seq_nr = m_seq_nr;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-25 07:57:59 +02:00
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: received ST_SYN state:%s seq_nr:%d ack_nr:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), socket_state_names[m_state], m_seq_nr, m_ack_nr);
|
2012-06-25 07:57:59 +02:00
|
|
|
#endif
|
2018-02-05 01:29:14 +01:00
|
|
|
if (m_send_id != ph->connection_id)
|
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: received invalid connection_id:%d expected: %d\n"
|
|
|
|
, static_cast<void*>(this), int(ph->connection_id), int(m_send_id));
|
|
|
|
#endif
|
|
|
|
return false;
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(m_recv_id == ((m_send_id + 1) & 0xffff));
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
defer_ack();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
2013-12-29 20:35:39 +01:00
|
|
|
UTP_LOG("%8p: ERROR: type:%s state:%s (ignored)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), packet_type_names[ph->get_type()], socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case UTP_STATE_SYN_SENT:
|
|
|
|
{
|
|
|
|
// just wait for an ack to our SYN, ignore everything else
|
|
|
|
if (ph->ack_nr != ((m_seq_nr - 1) & ACK_MASK))
|
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: incorrect ack_nr (%d) waiting for %d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(ph->ack_nr), (m_seq_nr - 1) & ACK_MASK);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
2015-08-19 01:39:01 +02:00
|
|
|
break;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
TORRENT_ASSERT(!m_error);
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_CONNECTED);
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: state:%s\n", static_cast<void*>(this), socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
// only progress our ack_nr on ST_DATA messages
|
|
|
|
// since our m_ack_nr is uninitialized at this point
|
|
|
|
// we still need to set it to something regardless
|
|
|
|
if (ph->get_type() == ST_DATA)
|
|
|
|
m_ack_nr = ph->seq_nr;
|
|
|
|
else
|
|
|
|
m_ack_nr = (ph->seq_nr - 1) & ACK_MASK;
|
|
|
|
|
|
|
|
// notify the client that the socket connected
|
|
|
|
if (m_connect_handler)
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: calling connect handler\n", static_cast<void*>(this));
|
2015-02-15 01:25:56 +01:00
|
|
|
m_connect_handler = false;
|
|
|
|
utp_stream::on_connect(m_userdata, m_error, false);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2016-07-31 03:53:11 +02:00
|
|
|
BOOST_FALLTHROUGH;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2017-10-23 16:50:24 +02:00
|
|
|
// fall through
|
2010-11-29 02:33:05 +01:00
|
|
|
case UTP_STATE_CONNECTED:
|
|
|
|
{
|
|
|
|
// the lowest seen RTT can be used to clamp the delay
|
|
|
|
// within reasonable bounds. The one-way delay is never
|
|
|
|
// higher than the round-trip time.
|
|
|
|
|
|
|
|
if (sample && acked_bytes && prev_bytes_in_flight)
|
2011-06-21 09:44:13 +02:00
|
|
|
{
|
2015-05-04 00:21:19 +02:00
|
|
|
// only use the minimum from the last 3 delay measurements
|
2016-10-08 18:12:33 +02:00
|
|
|
delay = *std::min_element(m_delay_sample_hist.begin()
|
|
|
|
, m_delay_sample_hist.end());
|
2015-05-04 00:21:19 +02:00
|
|
|
|
2012-06-25 16:17:51 +02:00
|
|
|
// it's impossible for delay to be more than the RTT, so make
|
|
|
|
// sure to clamp it as a sanity check
|
|
|
|
if (delay > min_rtt) delay = min_rtt;
|
2015-07-20 17:20:17 +02:00
|
|
|
|
2017-02-27 17:50:53 +01:00
|
|
|
do_ledbat(acked_bytes, int(delay), prev_bytes_in_flight);
|
|
|
|
m_send_delay = std::int32_t(delay);
|
2011-06-21 09:44:13 +02:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-27 17:50:53 +01:00
|
|
|
m_recv_delay = std::int32_t(std::min(their_delay, min_rtt));
|
2011-09-28 02:03:12 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
consume_incoming_data(ph, ptr, payload_size, receive_time);
|
|
|
|
|
|
|
|
// the parameter to send_pkt tells it if we're acking data
|
|
|
|
// If we are, we'll send an ACK regardless of if we have any
|
|
|
|
// space left in our send window or not. If we just got an ACK
|
|
|
|
// (i.e. ST_STATE) we're not ACKing anything. If we just
|
|
|
|
// received a FIN packet, we need to ack that as well
|
|
|
|
bool has_ack = ph->get_type() == ST_DATA || ph->get_type() == ST_FIN || ph->get_type() == ST_SYN;
|
2016-06-18 20:01:38 +02:00
|
|
|
std::uint32_t prev_out_packets = m_out_packets;
|
2012-06-21 17:05:57 +02:00
|
|
|
|
2015-07-19 03:55:26 +02:00
|
|
|
// the connection is connected and this packet made it past all the
|
|
|
|
// checks. We can now assume the other end is not spoofing it's IP.
|
|
|
|
if (ph->get_type() != ST_SYN) m_confirmed = true;
|
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
// try to send more data as long as we can
|
|
|
|
// if send_pkt returns true
|
2012-07-01 20:44:46 +02:00
|
|
|
while (send_pkt());
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-21 17:05:57 +02:00
|
|
|
if (has_ack && prev_out_packets == m_out_packets)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-21 17:05:57 +02:00
|
|
|
// we need to ack some data we received, and we didn't
|
|
|
|
// end up sending any payload packets in the loop
|
2016-04-26 06:37:47 +02:00
|
|
|
// above (because m_out_packets would have been incremented
|
2012-06-21 17:05:57 +02:00
|
|
|
// in that case). This means we need to send an ack.
|
|
|
|
// don't do it right away, because we may still receive
|
|
|
|
// more packets. defer the ack to send as few acks as possible
|
|
|
|
defer_ack();
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
2012-06-21 17:05:57 +02:00
|
|
|
|
2013-02-06 05:38:30 +01:00
|
|
|
// we may want to call the user callback function at the end
|
|
|
|
// of this round. Subscribe to that event
|
|
|
|
subscribe_drained();
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
|
2016-04-26 06:37:47 +02:00
|
|
|
// Everything up to the FIN has been received, respond with a FIN
|
2010-11-29 02:33:05 +01:00
|
|
|
// from our side.
|
|
|
|
if (m_eof && m_ack_nr == ((m_eof_seq_nr - 1) & ACK_MASK))
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: incoming stream consumed\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// This transitions to the UTP_STATE_FIN_SENT state.
|
|
|
|
send_fin();
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
if (sample && acked_bytes && prev_bytes_in_flight)
|
|
|
|
{
|
|
|
|
char their_delay_base[20];
|
|
|
|
if (m_their_delay_hist.initialized())
|
2016-05-17 15:24:06 +02:00
|
|
|
std::snprintf(their_delay_base, sizeof(their_delay_base), "%u", m_their_delay_hist.base());
|
2010-11-29 02:33:05 +01:00
|
|
|
else
|
|
|
|
strcpy(their_delay_base, "-");
|
|
|
|
|
|
|
|
char our_delay_base[20];
|
|
|
|
if (m_delay_hist.initialized())
|
2016-05-17 15:24:06 +02:00
|
|
|
std::snprintf(our_delay_base, sizeof(our_delay_base), "%u", m_delay_hist.base());
|
2010-11-29 02:33:05 +01:00
|
|
|
else
|
|
|
|
strcpy(our_delay_base, "-");
|
|
|
|
|
|
|
|
UTP_LOG("%8p: "
|
|
|
|
"actual_delay:%u "
|
|
|
|
"our_delay:%f "
|
|
|
|
"their_delay:%f "
|
|
|
|
"off_target:%f "
|
|
|
|
"max_window:%u "
|
|
|
|
"upload_rate:%d "
|
|
|
|
"delay_base:%s "
|
|
|
|
"delay_sum:%f "
|
|
|
|
"target_delay:%d "
|
|
|
|
"acked_bytes:%d "
|
|
|
|
"cur_window:%d "
|
|
|
|
"scaled_gain:%f "
|
|
|
|
"rtt:%u "
|
|
|
|
"rate:%d "
|
|
|
|
"quota:%d "
|
|
|
|
"wnduser:%u "
|
|
|
|
"rto:%d "
|
|
|
|
"timeout:%d "
|
|
|
|
"get_microseconds:%u "
|
|
|
|
"cur_window_packets:%u "
|
|
|
|
"packet_size:%d "
|
|
|
|
"their_delay_base:%s "
|
|
|
|
"their_actual_delay:%u "
|
|
|
|
"seq_nr:%u "
|
|
|
|
"acked_seq_nr:%u "
|
|
|
|
"reply_micro:%u "
|
|
|
|
"min_rtt:%u "
|
|
|
|
"send_buffer:%d "
|
|
|
|
"recv_buffer:%d "
|
2011-11-06 01:11:33 +01:00
|
|
|
"fast_resend_seq_nr:%d "
|
2014-08-20 09:02:03 +02:00
|
|
|
"ssthres:%d "
|
2010-11-29 02:33:05 +01:00
|
|
|
"\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this)
|
2010-11-29 02:33:05 +01:00
|
|
|
, sample
|
2016-08-03 06:35:40 +02:00
|
|
|
, delay / 1000.0
|
|
|
|
, their_delay / 1000.0
|
2017-02-18 21:40:19 +01:00
|
|
|
, int(m_sm.target_delay() - delay) / 1000.0
|
2016-06-18 20:01:38 +02:00
|
|
|
, std::uint32_t(m_cwnd >> 16)
|
2010-11-29 02:33:05 +01:00
|
|
|
, 0
|
|
|
|
, our_delay_base
|
2016-08-03 06:35:40 +02:00
|
|
|
, (delay + their_delay) / 1000.0
|
2017-02-18 21:40:19 +01:00
|
|
|
, m_sm.target_delay() / 1000
|
2010-11-29 02:33:05 +01:00
|
|
|
, acked_bytes
|
|
|
|
, m_bytes_in_flight
|
2016-08-03 06:35:40 +02:00
|
|
|
, 0.0 // float(scaled_gain)
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_rtt.mean()
|
2011-11-05 19:28:32 +01:00
|
|
|
, int((m_cwnd * 1000 / (m_rtt.mean()?m_rtt.mean():50)) >> 16)
|
2010-11-29 02:33:05 +01:00
|
|
|
, 0
|
|
|
|
, m_adv_wnd
|
|
|
|
, packet_timeout()
|
|
|
|
, int(total_milliseconds(m_timeout - receive_time))
|
2014-07-06 21:18:00 +02:00
|
|
|
, int(total_microseconds(receive_time.time_since_epoch()))
|
2010-11-29 02:33:05 +01:00
|
|
|
, (m_seq_nr - m_acked_seq_nr) & ACK_MASK
|
|
|
|
, m_mtu
|
|
|
|
, their_delay_base
|
2016-06-18 20:01:38 +02:00
|
|
|
, std::uint32_t(m_reply_micro)
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_seq_nr
|
|
|
|
, m_acked_seq_nr
|
|
|
|
, m_reply_micro
|
|
|
|
, min_rtt / 1000
|
|
|
|
, m_write_buffer_size
|
2011-11-06 01:11:33 +01:00
|
|
|
, m_read_buffer_size
|
2014-08-20 09:02:03 +02:00
|
|
|
, m_fast_resend_seq_nr
|
|
|
|
, m_ssthres);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2015-08-19 01:39:01 +02:00
|
|
|
break;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
case UTP_STATE_FIN_SENT:
|
|
|
|
{
|
|
|
|
// There are two ways we can end up in this state:
|
|
|
|
//
|
|
|
|
// 1. If the socket has been explicitly closed on our
|
|
|
|
// side, in which case m_eof is false.
|
|
|
|
//
|
|
|
|
// 2. If we received a FIN from the remote side, in which
|
|
|
|
// case m_eof is true. If this is the case, we don't
|
|
|
|
// come here until everything up to the FIN has been
|
|
|
|
// received.
|
|
|
|
//
|
|
|
|
//
|
|
|
|
//
|
|
|
|
|
|
|
|
// At this point m_seq_nr - 1 is the FIN sequence number.
|
|
|
|
|
|
|
|
// We can receive both ST_DATA and ST_STATE here, because after
|
|
|
|
// we have closed our end of the socket, the remote end might
|
|
|
|
// have data in the pipeline. We don't really care about the
|
|
|
|
// data, but we do have to ack it. Or rather, we have to ack
|
|
|
|
// the FIN that will come after the data.
|
|
|
|
|
|
|
|
// Case 1:
|
|
|
|
// ---------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// If we are here because the local endpoint was closed, we need
|
|
|
|
// to first wait for all of our messages to be acked:
|
|
|
|
//
|
|
|
|
// if (m_acked_seq_nr == ((m_seq_nr - 1) & ACK_MASK))
|
|
|
|
//
|
|
|
|
// `m_seq_nr - 1` is the ST_FIN message that we sent.
|
|
|
|
//
|
|
|
|
// ----------------------
|
|
|
|
//
|
|
|
|
// After that has happened we need to wait for the remote side
|
|
|
|
// to send its ST_FIN message. When we receive that we send an
|
|
|
|
// ST_STATE back to ack, and wait for a sufficient period.
|
|
|
|
// During this wait we keep acking incoming ST_FIN's. This is
|
|
|
|
// all handled at the top of this function.
|
|
|
|
//
|
|
|
|
// Note that the user handlers are all cancelled when the initial
|
|
|
|
// close() call happens, so nothing will happen on the user side
|
|
|
|
// after that.
|
|
|
|
|
|
|
|
// Case 2:
|
|
|
|
// ---------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// If we are here because we received a ST_FIN message, and then
|
|
|
|
// sent our own ST_FIN to ack that, we need to wait for our ST_FIN
|
|
|
|
// to be acked:
|
|
|
|
//
|
|
|
|
// if (m_acked_seq_nr == ((m_seq_nr - 1) & ACK_MASK))
|
|
|
|
//
|
|
|
|
// `m_seq_nr - 1` is the ST_FIN message that we sent.
|
|
|
|
//
|
|
|
|
// After that has happened we know the remote side has all our
|
|
|
|
// data, and we can gracefully shut down.
|
|
|
|
|
|
|
|
if (consume_incoming_data(ph, ptr, payload_size, receive_time))
|
2015-08-19 01:39:01 +02:00
|
|
|
{
|
|
|
|
break;
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (m_acked_seq_nr == ((m_seq_nr - 1) & ACK_MASK))
|
|
|
|
{
|
|
|
|
// When this happens we know that the remote side has
|
|
|
|
// received all of our packets.
|
|
|
|
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: FIN acked\n", static_cast<void*>(this));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
if (!m_attached)
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: close initiated here, delete socket\n"
|
|
|
|
, static_cast<void*>(this));
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::eof;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_DELETE);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: closing socket\n", static_cast<void*>(this));
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::eof;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-19 01:39:01 +02:00
|
|
|
break;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
case UTP_STATE_DELETE:
|
|
|
|
default:
|
|
|
|
{
|
|
|
|
// respond with a reset
|
|
|
|
send_reset(ph);
|
2015-08-19 01:39:01 +02:00
|
|
|
break;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
}
|
2015-08-19 01:39:01 +02:00
|
|
|
return true;
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2015-05-03 05:28:39 +02:00
|
|
|
void utp_socket_impl::do_ledbat(const int acked_bytes, const int delay
|
|
|
|
, const int in_flight)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// the portion of the in-flight bytes that were acked. This is used to make
|
|
|
|
// the gain factor be scaled by the rtt. The formula is applied once per
|
2016-04-26 06:37:47 +02:00
|
|
|
// rtt, or on every ACK scaled by the number of ACKs per rtt
|
2010-11-29 02:33:05 +01:00
|
|
|
TORRENT_ASSERT(in_flight > 0);
|
|
|
|
TORRENT_ASSERT(acked_bytes > 0);
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
const int target_delay = std::max(1, m_sm.target_delay());
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2013-09-14 12:06:48 +02:00
|
|
|
// true if the upper layer is pushing enough data down the socket to be
|
|
|
|
// limited by the cwnd. If this is not the case, we should not adjust cwnd.
|
2015-05-03 05:28:39 +02:00
|
|
|
const bool cwnd_saturated = (m_bytes_in_flight + acked_bytes + m_mtu > (m_cwnd >> 16));
|
2013-09-14 12:06:48 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// all of these are fixed points with 16 bits fraction portion
|
2016-10-08 20:17:51 +02:00
|
|
|
const std::int64_t window_factor = (std::int64_t(acked_bytes) * (1 << 16)) / in_flight;
|
|
|
|
const std::int64_t delay_factor = (std::int64_t(target_delay - delay) * (1 << 16)) / target_delay;
|
2016-06-18 20:01:38 +02:00
|
|
|
std::int64_t scaled_gain;
|
2015-07-20 17:20:17 +02:00
|
|
|
|
2012-06-21 18:41:05 +02:00
|
|
|
if (delay >= target_delay)
|
2011-11-05 10:30:38 +01:00
|
|
|
{
|
2014-08-19 01:05:10 +02:00
|
|
|
if (m_slow_start)
|
2014-08-20 09:02:03 +02:00
|
|
|
{
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: off_target: %d slow_start -> 0\n"
|
|
|
|
, static_cast<void*>(this), target_delay - delay);
|
2016-12-05 14:39:53 +01:00
|
|
|
m_ssthres = std::int32_t((m_cwnd >> 16) / 2);
|
2014-08-20 09:02:03 +02:00
|
|
|
m_slow_start = false;
|
|
|
|
}
|
2014-08-19 01:05:10 +02:00
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_samples_above_target);
|
2011-11-05 10:30:38 +01:00
|
|
|
}
|
2013-09-14 12:06:48 +02:00
|
|
|
else
|
|
|
|
{
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_samples_below_target);
|
2013-09-14 12:06:48 +02:00
|
|
|
}
|
2011-11-05 10:30:38 +01:00
|
|
|
|
2018-02-20 14:14:09 +01:00
|
|
|
std::int64_t const linear_gain = ((window_factor * delay_factor) >> 16)
|
|
|
|
* std::int64_t(m_sm.gain_factor());
|
2011-11-05 10:30:38 +01:00
|
|
|
|
2013-09-14 12:06:48 +02:00
|
|
|
// if the user is not saturating the link (i.e. not filling the
|
|
|
|
// congestion window), don't adjust it at all.
|
|
|
|
if (cwnd_saturated)
|
2011-11-05 10:30:38 +01:00
|
|
|
{
|
2018-02-20 14:14:09 +01:00
|
|
|
std::int64_t const exponential_gain = std::int64_t(acked_bytes) * (1 << 16);
|
2013-09-14 12:06:48 +02:00
|
|
|
if (m_slow_start)
|
|
|
|
{
|
|
|
|
// mimic TCP slow-start by adding the number of acked
|
|
|
|
// bytes to cwnd
|
2014-08-20 09:02:03 +02:00
|
|
|
if (m_ssthres != 0 && ((m_cwnd + exponential_gain) >> 16) > m_ssthres)
|
|
|
|
{
|
2016-04-26 06:37:47 +02:00
|
|
|
// if we would exceed the slow start threshold by growing the cwnd
|
2014-08-20 09:02:03 +02:00
|
|
|
// exponentially, don't do it, and leave slow-start mode. This
|
|
|
|
// make us avoid causing more delay and/or packet loss by being too
|
|
|
|
// aggressive
|
|
|
|
m_slow_start = false;
|
|
|
|
scaled_gain = linear_gain;
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: cwnd > ssthres (%d) slow_start -> 0\n"
|
|
|
|
, static_cast<void*>(this), m_ssthres);
|
2014-08-20 09:02:03 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2017-02-11 20:21:48 +01:00
|
|
|
scaled_gain = std::max(exponential_gain, linear_gain);
|
2014-08-20 09:02:03 +02:00
|
|
|
}
|
2013-09-14 12:06:48 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
scaled_gain = linear_gain;
|
|
|
|
}
|
2011-11-05 10:30:38 +01:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2013-09-14 12:06:48 +02:00
|
|
|
scaled_gain = 0;
|
2011-11-05 10:30:38 +01:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2012-06-21 18:41:05 +02:00
|
|
|
// make sure we don't wrap the cwnd
|
2018-02-20 14:14:09 +01:00
|
|
|
if (scaled_gain >= std::numeric_limits<std::int64_t>::max() - m_cwnd)
|
|
|
|
scaled_gain = std::numeric_limits<std::int64_t>::max() - m_cwnd - 1;
|
2012-06-21 18:41:05 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
UTP_LOGV("%8p: do_ledbat delay:%d off_target: %d window_factor:%f target_factor:%f "
|
2011-11-05 10:30:38 +01:00
|
|
|
"scaled_gain:%f cwnd:%d slow_start:%d\n"
|
2016-08-03 06:35:40 +02:00
|
|
|
, static_cast<void*>(this), delay, target_delay - delay, window_factor / double(1 << 16)
|
|
|
|
, delay_factor / double(1 << 16)
|
|
|
|
, scaled_gain / double(1 << 16), int(m_cwnd >> 16)
|
2011-11-05 10:30:38 +01:00
|
|
|
, int(m_slow_start));
|
2010-11-29 02:33:05 +01:00
|
|
|
|
|
|
|
// if scaled_gain + m_cwnd <= 0, set m_cwnd to 0
|
|
|
|
if (-scaled_gain >= m_cwnd)
|
|
|
|
{
|
|
|
|
m_cwnd = 0;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
m_cwnd += scaled_gain;
|
|
|
|
TORRENT_ASSERT(m_cwnd > 0);
|
|
|
|
}
|
2011-11-05 10:30:38 +01:00
|
|
|
|
2012-06-21 18:41:05 +02:00
|
|
|
TORRENT_ASSERT(m_cwnd >= 0);
|
2011-11-05 10:30:38 +01:00
|
|
|
|
2018-02-20 14:14:09 +01:00
|
|
|
int const window_size_left = std::min(int(m_cwnd >> 16), int(m_adv_wnd)) - in_flight + acked_bytes;
|
2011-11-05 10:30:38 +01:00
|
|
|
if (window_size_left >= m_mtu)
|
|
|
|
{
|
|
|
|
UTP_LOGV("%8p: mtu:%d in_flight:%d adv_wnd:%d cwnd:%d acked_bytes:%d cwnd_full -> 0\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_mtu, in_flight, int(m_adv_wnd), int(m_cwnd >> 16), acked_bytes);
|
2011-11-05 10:30:38 +01:00
|
|
|
m_cwnd_full = false;
|
|
|
|
}
|
2012-06-21 18:41:05 +02:00
|
|
|
|
|
|
|
if ((m_cwnd >> 16) >= m_adv_wnd)
|
2014-08-20 09:02:03 +02:00
|
|
|
{
|
2012-06-21 18:41:05 +02:00
|
|
|
m_slow_start = false;
|
2014-08-20 09:02:03 +02:00
|
|
|
UTP_LOGV("%8p: cwnd > advertized wnd (%d) slow_start -> 0\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), m_adv_wnd);
|
2014-08-20 09:02:03 +02:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2015-05-18 03:30:32 +02:00
|
|
|
void utp_stream::bind(endpoint_type const&, error_code&) { }
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2014-10-06 05:03:01 +02:00
|
|
|
void utp_stream::cancel_handlers(error_code const& ec)
|
|
|
|
{
|
|
|
|
if (!m_impl) return;
|
|
|
|
m_impl->cancel_handlers(ec, false);
|
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
// returns the number of milliseconds a packet would have before
|
|
|
|
// it would time-out if it was sent right now. Takes the RTT estimate
|
|
|
|
// into account
|
|
|
|
int utp_socket_impl::packet_timeout() const
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// SYN packets have a bit longer timeout, since we don't
|
|
|
|
// have an RTT estimate yet, make a conservative guess
|
|
|
|
if (m_state == UTP_STATE_NONE) return 3000;
|
|
|
|
|
2012-02-27 07:26:11 +01:00
|
|
|
// avoid overflow by simply capping based on number of timeouts as well
|
|
|
|
if (m_num_timeouts >= 7) return 60000;
|
|
|
|
|
2017-02-18 21:40:19 +01:00
|
|
|
int timeout = std::max(m_sm.min_timeout(), m_rtt.mean() + m_rtt.avg_deviation() * 2);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_num_timeouts > 0) timeout += (1 << (int(m_num_timeouts) - 1)) * 1000;
|
2014-07-06 21:18:00 +02:00
|
|
|
|
|
|
|
// timeouts over 1 minute are capped
|
|
|
|
if (timeout > 60000) timeout = 60000;
|
2010-11-29 02:33:05 +01:00
|
|
|
return timeout;
|
|
|
|
}
|
|
|
|
|
2015-03-12 05:34:54 +01:00
|
|
|
void utp_socket_impl::tick(time_point now)
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: tick:%s r: %d (%s) w: %d (%s)\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), socket_state_names[m_state], m_read, m_read_handler ? "handler" : "no handler"
|
2010-11-29 02:33:05 +01:00
|
|
|
, m_written, m_write_handler ? "handler" : "no handler");
|
|
|
|
#endif
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_outbuf.at((m_acked_seq_nr + 1) & ACK_MASK) || ((m_seq_nr - m_acked_seq_nr) & ACK_MASK) <= 1);
|
|
|
|
|
|
|
|
// if we're already in an error state, we're just waiting for the
|
|
|
|
// client to perform an operation so that we can communicate the
|
|
|
|
// error. No need to do anything else with this socket
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return;
|
|
|
|
|
|
|
|
if (now > m_timeout)
|
|
|
|
{
|
|
|
|
// TIMEOUT!
|
|
|
|
// set cwnd to 1 MSS
|
|
|
|
|
2016-08-02 06:46:15 +02:00
|
|
|
// the close_reason here is a bit of a hack. When it's set, it indicates
|
|
|
|
// that the upper layer intends to close the socket. However, it has been
|
|
|
|
// observed that the SSL shutdown sometimes can hang in a state where
|
|
|
|
// there's no outstanding data, and it won't receive any more from the
|
|
|
|
// other end. This catches that case and let the socket time out.
|
2017-02-04 02:27:31 +01:00
|
|
|
if (m_outbuf.size() || m_close_reason != close_reason_t::none)
|
2016-08-02 06:46:15 +02:00
|
|
|
{
|
|
|
|
++m_num_timeouts;
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.inc_stats_counter(counters::utp_timeout);
|
2016-08-02 06:46:15 +02:00
|
|
|
}
|
2012-05-06 01:54:41 +02:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
UTP_LOGV("%8p: timeout num-timeouts: %d max-resends: %d confirmed: %d "
|
|
|
|
" acked-seq-num: %d mtu-seq: %d\n"
|
|
|
|
, static_cast<void*>(this)
|
|
|
|
, m_num_timeouts
|
2017-02-18 21:40:19 +01:00
|
|
|
, m_sm.num_resends()
|
2016-03-12 07:07:17 +01:00
|
|
|
, m_confirmed
|
|
|
|
, m_acked_seq_nr
|
|
|
|
, m_mtu_seq);
|
|
|
|
|
2015-07-19 03:55:26 +02:00
|
|
|
// a socket that has not been confirmed to actually have a live remote end
|
|
|
|
// (the IP may have been spoofed) fail on the first timeout. If we had
|
|
|
|
// heard anything from this peer, it would have been confirmed.
|
2017-02-18 21:40:19 +01:00
|
|
|
if (m_num_timeouts > m_sm.num_resends()
|
2015-08-13 05:05:33 +02:00
|
|
|
|| (m_num_timeouts > 0 && !m_confirmed))
|
2012-05-06 01:54:41 +02:00
|
|
|
{
|
|
|
|
// the connection is dead
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::timed_out;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2012-05-06 01:54:41 +02:00
|
|
|
test_socket_state();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
if (((m_acked_seq_nr + 1) & ACK_MASK) == m_mtu_seq
|
|
|
|
&& ((m_seq_nr - 1) & ACK_MASK) == m_mtu_seq
|
|
|
|
&& m_mtu_seq != 0)
|
|
|
|
{
|
|
|
|
// we timed out, and the only outstanding packet
|
|
|
|
// we had was the probe. Assume it was dropped
|
|
|
|
// because it was too big
|
|
|
|
m_mtu_ceiling = m_mtu - 1;
|
2012-06-29 21:40:27 +02:00
|
|
|
if (m_mtu_floor > m_mtu_ceiling) m_mtu_floor = m_mtu_ceiling;
|
2010-11-29 02:33:05 +01:00
|
|
|
update_mtu_limits();
|
|
|
|
}
|
|
|
|
|
2012-08-13 18:49:28 +02:00
|
|
|
if (m_bytes_in_flight == 0 && (m_cwnd >> 16) >= m_mtu)
|
|
|
|
{
|
|
|
|
// this is just a timeout because this direction of
|
|
|
|
// the stream is idle. Don't reset the cwnd, just decay it
|
2016-10-08 20:17:51 +02:00
|
|
|
m_cwnd = std::max(m_cwnd * 2 / 3, std::int64_t(m_mtu) * (1 << 16));
|
2012-08-13 18:49:28 +02:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// we timed out because a packet was not ACKed or because
|
|
|
|
// the cwnd was made smaller than one packet
|
2016-10-08 20:17:51 +02:00
|
|
|
m_cwnd = std::int64_t(m_mtu) * (1 << 16);
|
2012-08-13 18:49:28 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_cwnd >= 0);
|
|
|
|
|
|
|
|
m_timeout = now + milliseconds(packet_timeout());
|
2015-07-19 03:55:26 +02:00
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
UTP_LOGV("%8p: resetting cwnd:%d\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), int(m_cwnd >> 16));
|
2012-08-13 18:49:28 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// we dropped all packets, that includes the mtu probe
|
|
|
|
m_mtu_seq = 0;
|
|
|
|
|
|
|
|
// since we've already timed out now, don't count
|
|
|
|
// loss that we might detect for packets that just
|
|
|
|
// timed out
|
|
|
|
m_loss_seq_nr = m_seq_nr;
|
|
|
|
|
2014-08-20 09:02:03 +02:00
|
|
|
// when we time out, the cwnd is reset to 1 MSS, which means we
|
|
|
|
// need to ramp it up quickly again. enter slow start mode. This time
|
|
|
|
// we're very likely to have an ssthres set, which will make us leave
|
|
|
|
// slow start before inducing more delay or loss.
|
|
|
|
m_slow_start = true;
|
2016-03-12 07:07:17 +01:00
|
|
|
UTP_LOGV("%8p: slow_start -> 1\n", static_cast<void*>(this));
|
2014-08-20 09:02:03 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// we need to go one past m_seq_nr to cover the case
|
|
|
|
// where we just sent a SYN packet and then adjusted for
|
|
|
|
// the uTorrent sequence number reuse
|
|
|
|
for (int i = m_acked_seq_nr & ACK_MASK;
|
|
|
|
i != ((m_seq_nr + 1) & ACK_MASK);
|
|
|
|
i = (i + 1) & ACK_MASK)
|
|
|
|
{
|
2017-02-18 12:52:46 +01:00
|
|
|
packet* p = m_outbuf.at(aux::numeric_cast<packet_buffer::index_type>(i));
|
2010-11-29 02:33:05 +01:00
|
|
|
if (!p) continue;
|
|
|
|
if (p->need_resend) continue;
|
|
|
|
p->need_resend = true;
|
|
|
|
TORRENT_ASSERT(m_bytes_in_flight >= p->size - p->header_size);
|
|
|
|
m_bytes_in_flight -= p->size - p->header_size;
|
2015-09-06 22:47:10 +02:00
|
|
|
UTP_LOGV("%8p: Packet %d lost (timeout).\n", static_cast<void*>(this), i);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
TORRENT_ASSERT(m_bytes_in_flight == 0);
|
|
|
|
|
|
|
|
// if we have a packet that needs re-sending, resend it
|
2015-08-17 15:01:43 +02:00
|
|
|
packet* p = m_outbuf.at((m_acked_seq_nr + 1) & ACK_MASK);
|
2010-11-29 02:33:05 +01:00
|
|
|
if (p)
|
|
|
|
{
|
2017-02-18 21:40:19 +01:00
|
|
|
if (p->num_transmissions >= m_sm.num_resends()
|
|
|
|
|| (m_state == UTP_STATE_SYN_SENT && p->num_transmissions >= m_sm.syn_resends())
|
|
|
|
|| (m_state == UTP_STATE_FIN_SENT && p->num_transmissions >= m_sm.fin_resends()))
|
2010-11-29 02:33:05 +01:00
|
|
|
{
|
|
|
|
#if TORRENT_UTP_LOG
|
|
|
|
UTP_LOGV("%8p: %d failed sends in a row. Socket timed out. state:%s\n"
|
2015-09-06 22:47:10 +02:00
|
|
|
, static_cast<void*>(this), p->num_transmissions, socket_state_names[m_state]);
|
2010-11-29 02:33:05 +01:00
|
|
|
#endif
|
|
|
|
|
2016-03-12 07:07:17 +01:00
|
|
|
if (p->size > m_mtu_floor)
|
|
|
|
{
|
|
|
|
// the packet that caused the connection to fail was an mtu probe
|
|
|
|
// (note that the mtu_probe field won't be set at this point because
|
|
|
|
// it's cleared when the packet is re-sent). This suggests that
|
|
|
|
// perhaps our network throws away oversized packets without
|
|
|
|
// fragmenting them. Tell the socket manager to be more conservative
|
|
|
|
// about mtu ceiling in the future
|
2017-02-18 21:40:19 +01:00
|
|
|
m_sm.restrict_mtu(m_mtu);
|
2016-03-12 07:07:17 +01:00
|
|
|
}
|
2010-11-29 02:33:05 +01:00
|
|
|
// the connection is dead
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::timed_out;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
return;
|
|
|
|
}
|
2015-06-06 07:22:53 +02:00
|
|
|
|
2010-11-29 02:33:05 +01:00
|
|
|
// don't fast-resend this packet
|
|
|
|
if (m_fast_resend_seq_nr == ((m_acked_seq_nr + 1) & ACK_MASK))
|
|
|
|
m_fast_resend_seq_nr = (m_fast_resend_seq_nr + 1) & ACK_MASK;
|
|
|
|
|
|
|
|
// the packet timed out, resend it
|
|
|
|
resend_packet(p);
|
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return;
|
|
|
|
}
|
|
|
|
else if (m_state < UTP_STATE_FIN_SENT)
|
|
|
|
{
|
2012-07-01 20:44:46 +02:00
|
|
|
send_pkt();
|
2010-11-29 02:33:05 +01:00
|
|
|
if (m_state == UTP_STATE_ERROR_WAIT || m_state == UTP_STATE_DELETE) return;
|
|
|
|
}
|
|
|
|
else if (m_state == UTP_STATE_FIN_SENT)
|
|
|
|
{
|
|
|
|
// the connection is dead
|
2015-06-06 07:22:53 +02:00
|
|
|
m_error = boost::asio::error::eof;
|
2015-01-04 22:31:02 +01:00
|
|
|
set_state(UTP_STATE_ERROR_WAIT);
|
2010-11-29 02:33:05 +01:00
|
|
|
test_socket_state();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (m_state)
|
|
|
|
{
|
|
|
|
case UTP_STATE_NONE:
|
|
|
|
case UTP_STATE_DELETE:
|
|
|
|
return;
|
|
|
|
// case UTP_STATE_SYN_SENT:
|
|
|
|
//
|
|
|
|
// break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void utp_socket_impl::check_receive_buffers() const
|
|
|
|
{
|
2012-06-26 05:42:01 +02:00
|
|
|
INVARIANT_CHECK;
|
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
int size = 0;
|
|
|
|
for (auto const& p : m_receive_buffer)
|
|
|
|
size += p ? p->size - p->header_size : 0;
|
2010-11-29 02:33:05 +01:00
|
|
|
|
2017-02-11 20:21:48 +01:00
|
|
|
TORRENT_ASSERT(size == m_receive_buffer_size);
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|
|
|
|
|
2014-01-21 20:26:09 +01:00
|
|
|
#if TORRENT_USE_INVARIANT_CHECKS
|
2012-06-26 05:42:01 +02:00
|
|
|
void utp_socket_impl::check_invariant() const
|
|
|
|
{
|
2017-02-27 17:50:53 +01:00
|
|
|
for (packet_buffer::index_type i = m_outbuf.cursor();
|
|
|
|
i != ((m_outbuf.cursor() + m_outbuf.span()) & ACK_MASK);
|
2012-06-26 05:42:01 +02:00
|
|
|
i = (i + 1) & ACK_MASK)
|
|
|
|
{
|
2017-02-27 17:50:53 +01:00
|
|
|
packet* p = m_outbuf.at(i);
|
2015-07-20 17:20:17 +02:00
|
|
|
if (!p) continue;
|
|
|
|
if (m_mtu_seq == i && m_mtu_seq != 0)
|
2012-06-26 05:42:01 +02:00
|
|
|
{
|
2012-06-28 05:53:52 +02:00
|
|
|
TORRENT_ASSERT(p->mtu_probe);
|
2012-06-26 05:42:01 +02:00
|
|
|
}
|
2015-08-19 01:39:01 +02:00
|
|
|
TORRENT_ASSERT(reinterpret_cast<utp_header*>(p->buf)->seq_nr == i);
|
2012-06-26 05:42:01 +02:00
|
|
|
}
|
2012-06-28 05:53:52 +02:00
|
|
|
|
|
|
|
if (m_nagle_packet)
|
|
|
|
{
|
|
|
|
// if this packet is full, it should have been sent
|
|
|
|
TORRENT_ASSERT(m_nagle_packet->size < m_nagle_packet->allocated);
|
|
|
|
}
|
2012-06-26 05:42:01 +02:00
|
|
|
}
|
|
|
|
#endif
|
2010-11-29 02:33:05 +01:00
|
|
|
}
|