premiere-libtorrent/include/libtorrent/disk_io_thread.hpp

513 lines
16 KiB
C++
Raw Normal View History

2007-06-10 22:46:09 +02:00
/*
Copyright (c) 2007-2012, Arvid Norberg
2007-06-10 22:46:09 +02:00
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the distribution.
* Neither the name of the author nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef TORRENT_DISK_IO_THREAD
#define TORRENT_DISK_IO_THREAD
2007-06-10 22:46:09 +02:00
#include "libtorrent/storage.hpp"
#include "libtorrent/allocator.hpp"
#include "libtorrent/io_service.hpp"
#include "libtorrent/sliding_average.hpp"
#include <boost/function/function0.hpp>
#include <boost/function/function2.hpp>
2007-06-10 22:46:09 +02:00
#include <boost/noncopyable.hpp>
2008-02-08 11:22:05 +01:00
#include <boost/shared_array.hpp>
#include <deque>
2007-06-10 22:46:09 +02:00
#include "libtorrent/config.hpp"
#include "libtorrent/thread.hpp"
#include "libtorrent/disk_buffer_pool.hpp"
2007-06-10 22:46:09 +02:00
#include <boost/multi_index_container.hpp>
#include <boost/multi_index/member.hpp>
#include <boost/multi_index/ordered_index.hpp>
#include <boost/multi_index/mem_fun.hpp>
2007-06-10 22:46:09 +02:00
namespace libtorrent
{
using boost::multi_index::multi_index_container;
using boost::multi_index::ordered_non_unique;
using boost::multi_index::ordered_unique;
using boost::multi_index::indexed_by;
using boost::multi_index::member;
using boost::multi_index::const_mem_fun;
2008-02-08 11:22:05 +01:00
struct cached_piece_info
{
// the piece index for this cache entry.
2008-02-08 11:22:05 +01:00
int piece;
// holds one entry for each block in this piece. ``true`` represents
// the data for that block being in the disk cache and ``false`` means it's not.
2008-02-08 11:22:05 +01:00
std::vector<bool> blocks;
// the time when a block was last written to this piece. The older
// a piece is, the more likely it is to be flushed to disk.
ptime last_use;
// The index of the next block that needs to be hashed.
// Blocks are hashed as they are downloaded in order to not
// have to re-read them from disk once the piece is complete, to
// compare its hash against the hashes in the .torrent file.
int next_to_hash;
enum kind_t { read_cache = 0, write_cache = 1 };
// specifies if this piece is part of the read cache or the write cache.
kind_t kind;
2008-02-08 11:22:05 +01:00
};
2007-06-10 22:46:09 +02:00
struct disk_io_job
{
disk_io_job()
: action(read)
, buffer(0)
, buffer_size(0)
, piece(0)
, offset(0)
, max_cache_line(0)
, cache_min_time(0)
2007-06-10 22:46:09 +02:00
{}
enum action_t
{
read
, write
, hash
, move_storage
, release_files
, delete_files
2008-03-08 07:06:31 +01:00
, check_fastresume
, check_files
, save_resume_data
, rename_file
, abort_thread
, clear_read_cache
, abort_torrent
, update_settings
, read_and_hash
, cache_piece
2013-02-16 09:26:55 +01:00
#ifndef TORRENT_NO_DEPRECATE
, finalize_file
2013-02-16 09:26:55 +01:00
#endif
2007-06-10 22:46:09 +02:00
};
action_t action;
char* buffer;
2008-04-05 23:18:27 +02:00
int buffer_size;
2007-06-10 22:46:09 +02:00
boost::intrusive_ptr<piece_manager> storage;
// arguments used for read and write
// piece is used as flags for move_storage
2007-06-10 22:46:09 +02:00
int piece, offset;
// used for move_storage and rename_file. On errors, this is set
2007-06-10 22:46:09 +02:00
// to the error message
std::string str;
2008-04-13 00:08:07 +02:00
// on error, this is set to the path of the
// file the disk operation failed on
std::string error_file;
// if this is > 0, it specifies the max number of blocks to read
// ahead in the read cache for this access. This is only valid
// for 'read' actions
int max_cache_line;
// if this is > 0, it may increase the minimum time the cache
// line caused by this operation stays in the cache
int cache_min_time;
2007-09-16 03:34:06 +02:00
boost::shared_ptr<entry> resume_data;
// the error code from the file operation
error_code error;
2007-06-10 22:46:09 +02:00
// this is called when operation completes
boost::function<void(int, disk_io_job const&)> callback;
// the time when this job was issued. This is used to
// keep track of disk I/O congestion
ptime start_time;
2007-06-10 22:46:09 +02:00
};
// returns true if the fundamental operation
// of the given disk job is a read operation
bool is_read_operation(disk_io_job const& j);
// this is true if the buffer field in the disk_io_job
// points to a disk buffer
bool operation_has_buffer(disk_io_job const& j);
2013-08-04 11:02:19 +02:00
struct TORRENT_EXPORT cache_status
2008-02-08 11:22:05 +01:00
{
cache_status()
: blocks_written(0)
, writes(0)
, blocks_read(0)
, blocks_read_hit(0)
, reads(0)
, queued_bytes(0)
, cache_size(0)
, read_cache_size(0)
, total_used_buffers(0)
2010-03-04 04:43:50 +01:00
, average_queue_time(0)
, average_read_time(0)
2011-03-11 08:37:12 +01:00
, average_write_time(0)
, average_hash_time(0)
, average_job_time(0)
2011-03-15 03:21:28 +01:00
, average_sort_time(0)
2010-03-04 04:43:50 +01:00
, job_queue_length(0)
2011-03-19 23:23:58 +01:00
, cumulative_job_time(0)
, cumulative_read_time(0)
, cumulative_write_time(0)
, cumulative_hash_time(0)
, cumulative_sort_time(0)
, total_read_back(0)
2011-03-20 10:04:03 +01:00
, read_queue_size(0)
{}
// the total number of 16 KiB blocks written to disk
// since this session was started.
2008-02-08 11:22:05 +01:00
size_type blocks_written;
// the total number of write operations performed since this
// session was started.
//
// The ratio (``blocks_written`` - ``writes``) / ``blocks_written`` represents
// the number of saved write operations per total write operations. i.e. a kind
// of cache hit ratio for the write cahe.
2008-02-08 11:22:05 +01:00
size_type writes;
// the number of blocks that were requested from the
// bittorrent engine (from peers), that were served from disk or cache.
size_type blocks_read;
// the number of blocks that was just copied from the read cache
//
// The ratio ``blocks_read_hit`` / ``blocks_read`` is the cache hit ratio
// for the read cache.
size_type blocks_read_hit;
// the number of read operations used
size_type reads;
mutable size_type queued_bytes;
// the number of 16 KiB blocks currently in the disk cache (both read and write).
// This includes both read and write cache.
int cache_size;
// the number of 16KiB blocks in the read cache.
int read_cache_size;
// the total number of buffers currently in use.
// This includes the read/write disk cache as well as send and receive buffers
// used in peer connections.
mutable int total_used_buffers;
// the number of microseconds an average disk I/O job
// has to wait in the job queue before it get processed.
int average_queue_time;
// the time read jobs takes on average to complete
// (not including the time in the queue), in microseconds. This only measures
// read cache misses.
int average_read_time;
// the time write jobs takes to complete, on average,
// in microseconds. This does not include the time the job sits in the disk job
// queue or in the write cache, only blocks that are flushed to disk.
2011-03-11 08:37:12 +01:00
int average_write_time;
// the time hash jobs takes to complete on average, in
// microseconds. Hash jobs include running SHA-1 on the data (which for the most
// part is done incrementally) and sometimes reading back parts of the piece. It
// also includes checking files without valid resume data.
int average_hash_time;
int average_job_time;
2011-03-15 03:21:28 +01:00
int average_sort_time;
// the number of jobs in the job queue.
int job_queue_length;
2011-03-19 23:23:58 +01:00
boost::uint32_t cumulative_job_time;
boost::uint32_t cumulative_read_time;
boost::uint32_t cumulative_write_time;
boost::uint32_t cumulative_hash_time;
boost::uint32_t cumulative_sort_time;
int total_read_back;
2011-03-20 10:04:03 +01:00
int read_queue_size;
2008-02-08 11:22:05 +01:00
};
// this is a singleton consisting of the thread and a queue
// of disk io jobs
struct TORRENT_EXTRA_EXPORT disk_io_thread : disk_buffer_pool
{
disk_io_thread(io_service& ios
, boost::function<void()> const& queue_callback
2010-01-23 04:02:32 +01:00
, file_pool& fp
, int block_size = 16 * 1024);
~disk_io_thread();
void abort();
void join();
2007-06-10 22:46:09 +02:00
// aborts read operations
void stop(boost::intrusive_ptr<piece_manager> s);
// returns the disk write queue size
int add_job(disk_io_job const& j
2007-06-10 22:46:09 +02:00
, boost::function<void(int, disk_io_job const&)> const& f
= boost::function<void(int, disk_io_job const&)>());
// keep track of the number of bytes in the job queue
// at any given time. i.e. the sum of all buffer_size.
// this is used to slow down the download global download
// speed when the queue buffer size is too big.
size_type queue_buffer_size() const;
2011-02-13 23:27:02 +01:00
bool can_write() const;
2007-06-10 22:46:09 +02:00
2008-02-08 11:22:05 +01:00
void get_cache_info(sha1_hash const& ih
, std::vector<cached_piece_info>& ret) const;
cache_status status() const;
void thread_fun();
2007-06-10 22:46:09 +02:00
#if defined TORRENT_DEBUG && !defined TORRENT_DISABLE_INVARIANT_CHECKS
void check_invariant() const;
#endif
struct cached_block_entry
{
cached_block_entry(): buf(0) {}
// the buffer pointer (this is a disk_pool buffer)
// or 0
char* buf;
// callback for when this block is flushed to disk
boost::function<void(int, disk_io_job const&)> callback;
};
2008-02-08 11:22:05 +01:00
struct cached_piece_entry
{
int piece;
// storage this piece belongs to
boost::intrusive_ptr<piece_manager> storage;
// the pointers to the block data
boost::shared_array<cached_block_entry> blocks;
2008-02-08 11:22:05 +01:00
// the last time a block was writting to this piece
// plus the minimum amount of time the block is guaranteed
// to stay in the cache
ptime expire;
2008-02-08 11:22:05 +01:00
// the number of blocks in the cache for this piece
int num_blocks;
2011-03-16 08:21:58 +01:00
// used to determine if this piece should be flushed
int num_contiguous_blocks;
// this is the first block that has not yet been hashed
// by the partial hasher. When minimizing read-back, this
// is used to determine if flushing a range would force us
// to read it back later when hashing
int next_block_to_hash;
std::pair<void*, int> storage_piece_pair() const
{ return std::pair<void*, int>(storage.get(), piece); }
2008-02-08 11:22:05 +01:00
};
typedef multi_index_container<
cached_piece_entry, indexed_by<
ordered_unique<const_mem_fun<cached_piece_entry, std::pair<void*, int>
, &cached_piece_entry::storage_piece_pair> >
, ordered_non_unique<member<cached_piece_entry, ptime
, &cached_piece_entry::expire> >
>
> cache_t;
typedef cache_t::nth_index<0>::type cache_piece_index_t;
typedef cache_t::nth_index<1>::type cache_lru_index_t;
2008-02-25 00:14:10 +01:00
private:
int add_job(disk_io_job const& j
, mutex::scoped_lock& l
, boost::function<void(int, disk_io_job const&)> const& f
= boost::function<void(int, disk_io_job const&)>());
bool test_error(disk_io_job& j);
void post_callback(disk_io_job const& j, int ret);
// cache operations
cache_piece_index_t::iterator find_cached_piece(
2008-02-25 00:14:10 +01:00
cache_t& cache, disk_io_job const& j
, mutex::scoped_lock& l);
bool is_cache_hit(cached_piece_entry& p
, disk_io_job const& j, mutex::scoped_lock& l);
int copy_from_piece(cached_piece_entry& p, bool& hit
, disk_io_job const& j, mutex::scoped_lock& l);
2012-04-07 02:35:25 +02:00
struct ignore_t
{
ignore_t(): piece(-1), storage(0) {}
ignore_t(int idx, piece_manager* st): piece(idx), storage(st) {}
int piece;
piece_manager* storage;
};
// write cache operations
enum options_t { dont_flush_write_blocks = 1, ignore_cache_size = 2 };
int flush_cache_blocks(mutex::scoped_lock& l
2012-04-07 02:35:25 +02:00
, int blocks, ignore_t ignore = ignore_t(), int options = 0);
void flush_expired_pieces();
int flush_contiguous_blocks(cached_piece_entry& p
2011-09-27 06:05:05 +02:00
, mutex::scoped_lock& l, int lower_limit = 0, bool avoid_readback = false);
int flush_range(cached_piece_entry& p, int start, int end, mutex::scoped_lock& l);
int cache_block(disk_io_job& j
, boost::function<void(int,disk_io_job const&)>& handler
, int cache_expire
, mutex::scoped_lock& l);
2008-02-08 11:22:05 +01:00
// read cache operations
2012-04-07 02:35:25 +02:00
int clear_oldest_read_piece(int num_blocks, ignore_t ignore
, mutex::scoped_lock& l);
int read_into_piece(cached_piece_entry& p, int start_block
, int options, int num_blocks, mutex::scoped_lock& l);
int cache_read_block(disk_io_job const& j, mutex::scoped_lock& l);
int free_piece(cached_piece_entry& p, mutex::scoped_lock& l);
2010-05-17 01:14:47 +02:00
int drain_piece_bufs(cached_piece_entry& p, std::vector<char*>& buf
2010-05-13 17:01:20 +02:00
, mutex::scoped_lock& l);
enum cache_flags_t {
cache_only = 1
};
int try_read_from_cache(disk_io_job const& j, bool& hit, int flags = 0);
int read_piece_from_cache_and_hash(disk_io_job const& j, sha1_hash& h);
int cache_piece(disk_io_job const& j, cache_piece_index_t::iterator& p
, bool& hit, int options, mutex::scoped_lock& l);
// this mutex only protects m_jobs, m_queue_buffer_size,
// m_exceeded_write_queue and m_abort
mutable mutex m_queue_mutex;
event m_signal;
2007-06-10 22:46:09 +02:00
bool m_abort;
bool m_waiting_to_shutdown;
std::deque<disk_io_job> m_jobs;
2007-06-10 22:46:09 +02:00
size_type m_queue_buffer_size;
ptime m_last_file_check;
// this protects the piece cache and related members
mutable mutex m_piece_mutex;
// write cache
2008-02-25 00:14:10 +01:00
cache_t m_pieces;
// read cache
2008-02-25 00:14:10 +01:00
cache_t m_read_pieces;
void flip_stats(ptime now);
// total number of blocks in use by both the read
// and the write cache. This is not supposed to
// exceed m_cache_size
cache_status m_cache_stats;
// keeps average queue time for disk jobs (in microseconds)
average_accumulator m_queue_time;
// average read time for cache misses (in microseconds)
average_accumulator m_read_time;
2011-03-11 08:37:12 +01:00
// average write time (in microseconds)
average_accumulator m_write_time;
// average hash time (in microseconds)
average_accumulator m_hash_time;
2011-03-11 08:37:12 +01:00
// average time to serve a job (any job) in microseconds
average_accumulator m_job_time;
2011-03-15 02:44:32 +01:00
2011-03-15 03:21:28 +01:00
// average time to ask for physical offset on disk
// and insert into queue
average_accumulator m_sort_time;
// the last time we reset the average time and store the
// latest value in m_cache_stats
ptime m_last_stats_flip;
2011-03-15 03:21:28 +01:00
typedef std::multimap<size_type, disk_io_job> read_jobs_t;
read_jobs_t m_sorted_read_jobs;
#ifdef TORRENT_DISK_STATS
std::ofstream m_log;
#endif
2008-02-08 11:22:05 +01:00
// the amount of physical ram in the machine
boost::uint64_t m_physical_ram;
// if we exceeded the max queue disk write size
// this is set to true. It remains true until the
// queue is smaller than the low watermark
bool m_exceeded_write_queue;
2008-05-03 18:05:42 +02:00
io_service& m_ios;
2008-02-08 11:22:05 +01:00
boost::function<void()> m_queue_callback;
// this keeps the io_service::run() call blocked from
// returning. When shutting down, it's possible that
// the event queue is drained before the disk_io_thread
// has posted its last callback. When this happens, the
// io_service will have a pending callback from the
// disk_io_thread, but the event loop is not running.
// this means that the event is destructed after the
// disk_io_thread. If the event refers to a disk buffer
// it will try to free it, but the buffer pool won't
// exist anymore, and crash. This prevents that.
boost::optional<io_service::work> m_work;
2010-01-23 04:02:32 +01:00
// reference to the file_pool which is a member of
// the session_impl object
file_pool& m_file_pool;
// when completion notifications are queued, they're stuck
// in this list
std::list<std::pair<disk_io_job, int> > m_queued_completions;
2007-06-10 22:46:09 +02:00
// thread for performing blocking disk io operations
thread m_disk_io_thread;
2007-06-10 22:46:09 +02:00
};
}
#endif