/* Copyright (c) 2014-2015, Arvid Norberg All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the author nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "libtorrent/kademlia/dht_settings.hpp" #include "libtorrent/io_service.hpp" #include "libtorrent/deadline_timer.hpp" #include "libtorrent/address.hpp" #include "libtorrent/time.hpp" #include "libtorrent/kademlia/node.hpp" #include "libtorrent/kademlia/dht_observer.hpp" #include "libtorrent/aux_/session_impl.hpp" #include "setup_transfer.hpp" #include // for unique_ptr #include #include "libtorrent/socket_io.hpp" // print_endpoint #include "libtorrent/random.hpp" #include "libtorrent/crc32c.hpp" #include "libtorrent/alert_types.hpp" // for dht_routing_bucket #include "libtorrent/aux_/listen_socket_handle.hpp" #include "setup_dht.hpp" using namespace sim; using namespace lt; #ifndef TORRENT_DISABLE_DHT namespace { // this is the IP address assigned to node 'idx' asio::ip::address addr_from_int(int /* idx */) { return rand_v4(); } asio::ip::address addr6_from_int(int /* idx */) { asio::ip::address_v6::bytes_type bytes; for (uint8_t& b : bytes) b = uint8_t(lt::random(0xff)); return asio::ip::address_v6(bytes); } // this is the node ID assigned to node 'idx' dht::node_id id_from_addr(lt::address const& addr) { return dht::generate_id_impl(addr, 0); } std::shared_ptr sim_listen_socket(tcp::endpoint ep) { auto ls = std::make_shared(); ls->external_address.cast_vote(ep.address() , lt::aux::session_interface::source_dht, lt::address()); ls->local_endpoint = ep; return ls; } } // anonymous namespace struct dht_node final : lt::dht::socket_manager { dht_node(sim::simulation& sim, lt::dht::settings const& sett, lt::counters& cnt , int const idx, std::uint32_t const flags) : m_io_service(sim, (flags & dht_network::bind_ipv6) ? addr6_from_int(idx) : addr_from_int(idx)) , m_dht_storage(lt::dht::dht_default_storage_constructor(sett)) , m_add_dead_nodes((flags & dht_network::add_dead_nodes) != 0) , m_ipv6((flags & dht_network::bind_ipv6) != 0) , m_socket(m_io_service) , m_ls(sim_listen_socket(tcp::endpoint(m_io_service.get_ips().front(), 6881))) , m_dht(m_ls, this, sett, id_from_addr(m_io_service.get_ips().front()) , nullptr, cnt , [](lt::dht::node_id const&, std::string const&) -> lt::dht::node* { return nullptr; } , *m_dht_storage) { m_dht_storage->update_node_ids({id_from_addr(m_io_service.get_ips().front())}); sock().open(m_ipv6 ? asio::ip::udp::v6() : asio::ip::udp::v4()); sock().bind(asio::ip::udp::endpoint( m_ipv6 ? lt::address(lt::address_v6::any()) : lt::address(lt::address_v4::any()), 6881)); sock().non_blocking(true); sock().async_receive_from(asio::mutable_buffers_1(m_buffer, sizeof(m_buffer)) , m_ep, [&](lt::error_code const& ec, std::size_t bytes_transferred) { this->on_read(ec, bytes_transferred); }); } // This type is not copyable, because the socket and the dht node is not // copyable. dht_node(dht_node const&) = delete; dht_node& operator=(dht_node const&) = delete; // it's also not movable, because it passes in its this-pointer to the async // receive function, which pins this object down. However, std::vector cannot // hold non-movable and non-copyable types. dht_node(dht_node&& n) = delete; dht_node& operator=(dht_node&&) = delete; void on_read(lt::error_code const& ec, std::size_t bytes_transferred) { if (ec) return; using lt::entry; using lt::bdecode; int pos; error_code err; // since the simulation is single threaded, we can get away with just // allocating a single of these static bdecode_node msg; int const ret = bdecode(m_buffer, m_buffer + bytes_transferred, msg, err, &pos, 10, 500); if (ret != 0) return; if (msg.type() != bdecode_node::dict_t) return; lt::dht::msg m(msg, m_ep); dht().incoming(m_ls, m); sock().async_receive_from(asio::mutable_buffers_1(m_buffer, sizeof(m_buffer)) , m_ep, [&](lt::error_code const& ec, std::size_t bytes_transferred) { this->on_read(ec, bytes_transferred); }); } bool has_quota() override { return true; } bool send_packet(lt::aux::listen_socket_handle const&, entry& e, udp::endpoint const& addr) override { // since the simulaton is single threaded, we can get away with allocating // just a single send buffer static std::vector send_buf; send_buf.clear(); bencode(std::back_inserter(send_buf), e); sock().send_to(boost::asio::const_buffers_1(send_buf.data(), int(send_buf.size())), addr); return true; } // the node_id and IP address of this node std::pair node_info() const { return std::make_pair(dht().nid(), lt::udp::endpoint(m_io_service.get_ips().front(), 6881)); } void bootstrap(std::vector> const& nodes) { // we don't want to tell every node about every other node. That's way too // expensive. instead. pick a random subset of nodes proportionate to the // bucket it would fall into dht::node_id const id = dht().nid(); // the number of slots left per bucket std::array nodes_per_bucket; nodes_per_bucket.fill(8); // when we use the larger routing table, the low buckets are larger nodes_per_bucket[0] = 128; nodes_per_bucket[1] = 64; nodes_per_bucket[2] = 32; nodes_per_bucket[3] = 16; // pick nodes in random order to provide good connectivity std::vector order(nodes.size()); for (size_t i = 0; i < order.size(); ++i) order[i] = i; while (!order.empty()) { auto const idx = lt::random(static_cast(order.size() - 1)); assert(idx >= 0 && idx < order.size()); auto const& n = nodes[order[idx]]; if (idx < order.size() - 1) order[idx] = order.back(); order.pop_back(); if (n.first == id) continue; int const bucket = 159 - dht::distance_exp(id, n.first); /* std::printf("%s ^ %s = %s %d\n" , to_hex(id.to_string()).c_str() , to_hex(n.first.to_string()).c_str() , to_hex(dht::distance(id, n.first).to_string()).c_str() , bucket); */ // there are no more slots in this bucket, just move on if (nodes_per_bucket[bucket] == 0) continue; --nodes_per_bucket[bucket]; bool const added = dht().m_table.node_seen(n.first, n.second, lt::random(300) + 10); TEST_CHECK(added); if (m_add_dead_nodes) { // generate a random node ID that would fall in `bucket` dht::node_id const mask = dht::generate_prefix_mask(bucket + 1); udp::endpoint const ep = rand_udp_ep(m_ipv6 ? rand_v6 : rand_v4); dht::node_id target = dht::generate_id_impl(ep.address(), 0) & ~mask; target |= id & mask; dht().m_table.node_seen(target, ep, lt::random(300) + 10); } } /* for (int i = 0; i < 40; ++i) { std::printf("%d ", nodes_per_bucket[i]); } std::printf("\n"); */ //#error add invalid IPs as well, to simulate churn } void stop() { sock().close(); } lt::dht::node& dht() { return m_dht; } lt::dht::node const& dht() const { return m_dht; } private: asio::io_service m_io_service; std::shared_ptr m_dht_storage; bool const m_add_dead_nodes; bool const m_ipv6; lt::udp::socket m_socket; lt::udp::socket& sock() { return m_socket; } std::shared_ptr m_ls; lt::dht::node m_dht; lt::udp::endpoint m_ep; char m_buffer[1300]; }; dht_network::dht_network(sim::simulation& sim, int num_nodes, std::uint32_t flags) { m_sett.ignore_dark_internet = false; m_sett.restrict_routing_ips = false; // TODO: how can we introduce churn among peers? std::vector> all_nodes; all_nodes.reserve(num_nodes); for (int i = 0; i < num_nodes; ++i) { // node 0 is the one we log m_nodes.emplace_back(sim, m_sett, m_cnt, i, flags); all_nodes.push_back(m_nodes.back().node_info()); } for (auto& n : m_nodes) n.bootstrap(all_nodes); } dht_network::~dht_network() = default; void print_routing_table(std::vector const& rt) { int bucket = 0; for (std::vector::const_iterator i = rt.begin() , end(rt.end()); i != end; ++i, ++bucket) { char const* progress_bar = "################################" "################################" "################################" "################################"; char const* short_progress_bar = "--------"; std::printf("%3d [%3d, %d] %s%s\n" , bucket, i->num_nodes, i->num_replacements , progress_bar + (128 - i->num_nodes) , short_progress_bar + (8 - std::min(8, i->num_replacements))); } } std::vector dht_network::router_nodes() const { int idx = 0; std::vector ret; ret.reserve(8); for (auto const& n : m_nodes) { if (idx >= 8) break; ++idx; ret.push_back(n.node_info().second); } return ret; } void dht_network::stop() { for (auto& n : m_nodes) n.stop(); } #endif // TORRENT_DISABLE_DHT