/* Copyright (c) 2003-2012, Arvid Norberg All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the author nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "libtorrent/pch.hpp" #include #include #include #include #include #ifdef _MSC_VER #pragma warning(push, 1) #endif #include #include #ifdef _MSC_VER #pragma warning(pop) #endif #include "libtorrent/config.hpp" #include "libtorrent/torrent_handle.hpp" #include "libtorrent/session.hpp" #include "libtorrent/torrent_info.hpp" #include "libtorrent/tracker_manager.hpp" #include "libtorrent/parse_url.hpp" #include "libtorrent/bencode.hpp" #include "libtorrent/hasher.hpp" #include "libtorrent/entry.hpp" #include "libtorrent/peer.hpp" #include "libtorrent/peer_connection.hpp" #include "libtorrent/bt_peer_connection.hpp" #include "libtorrent/web_peer_connection.hpp" #include "libtorrent/http_seed_connection.hpp" #include "libtorrent/peer_id.hpp" #include "libtorrent/alert.hpp" #include "libtorrent/identify_client.hpp" #include "libtorrent/alert_types.hpp" #include "libtorrent/extensions.hpp" #include "libtorrent/aux_/session_impl.hpp" #include "libtorrent/instantiate_connection.hpp" #include "libtorrent/assert.hpp" #include "libtorrent/broadcast_socket.hpp" #include "libtorrent/kademlia/dht_tracker.hpp" #include "libtorrent/peer_info.hpp" #include "libtorrent/enum_net.hpp" #include "libtorrent/http_connection.hpp" #include "libtorrent/gzip.hpp" // for inflate_gzip #include "libtorrent/random.hpp" #include "libtorrent/string_util.hpp" // for allocate_string_copy #ifdef TORRENT_USE_OPENSSL #include "libtorrent/ssl_stream.hpp" #include #if BOOST_VERSION >= 104700 #include #endif // BOOST_VERSION #endif // TORRENT_USE_OPENSSL #include "../../src/util.h" #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING #include "libtorrent/struct_debug.hpp" #endif using namespace libtorrent; using boost::tuples::tuple; using boost::tuples::get; using boost::tuples::make_tuple; using libtorrent::aux::session_impl; namespace { size_type collect_free_download( torrent::peer_iterator start , torrent::peer_iterator end) { size_type accumulator = 0; for (torrent::peer_iterator i = start; i != end; ++i) { // if the peer is interested in us, it means it may // want to trade it's surplus uploads for downloads itself // (and we should not consider it free). If the share diff is // negative, there's no free download to get from this peer. size_type diff = (*i)->share_diff(); TORRENT_ASSERT(diff < (std::numeric_limits::max)()); if ((*i)->is_peer_interested() || diff <= 0) continue; TORRENT_ASSERT(diff > 0); (*i)->add_free_upload(-diff); accumulator += diff; TORRENT_ASSERT(accumulator > 0); } TORRENT_ASSERT(accumulator >= 0); return accumulator; } // returns the amount of free upload left after // it has been distributed to the peers boost::uint32_t distribute_free_upload( torrent::peer_iterator start , torrent::peer_iterator end , size_type free_upload) { TORRENT_ASSERT(free_upload >= 0); if (free_upload <= 0) return 0; int num_peers = 0; size_type total_diff = 0; for (torrent::peer_iterator i = start; i != end; ++i) { size_type d = (*i)->share_diff(); TORRENT_ASSERT(d < (std::numeric_limits::max)()); total_diff += d; if (!(*i)->is_peer_interested() || (*i)->share_diff() >= 0) continue; ++num_peers; } if (num_peers == 0) return boost::uint32_t(free_upload); size_type upload_share; if (total_diff >= 0) { upload_share = (std::min)(free_upload, total_diff) / num_peers; } else { upload_share = (free_upload + total_diff) / num_peers; } if (upload_share < 0) return boost::uint32_t(free_upload); for (torrent::peer_iterator i = start; i != end; ++i) { peer_connection* p = *i; if (!p->is_peer_interested() || p->share_diff() >= 0) continue; p->add_free_upload(upload_share); free_upload -= upload_share; } return (std::min)(free_upload, size_type(UINT_MAX)); } struct find_peer_by_ip { find_peer_by_ip(tcp::endpoint const& a, const torrent* t) : ip(a) , tor(t) { TORRENT_ASSERT(t != 0); } bool operator()(session_impl::connection_map::value_type const& c) const { tcp::endpoint const& sender = c->remote(); if (sender.address() != ip.address()) return false; if (tor != c->associated_torrent().lock().get()) return false; return true; } tcp::endpoint const& ip; torrent const* tor; }; struct peer_by_id { peer_by_id(const peer_id& i): pid(i) {} bool operator()(session_impl::connection_map::value_type const& p) const { if (p->pid() != pid) return false; // have a special case for all zeros. We can have any number // of peers with that pid, since it's used to indicate no pid. if (pid.is_all_zeros()) return false; return true; } peer_id const& pid; }; void nop() {} } namespace libtorrent { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING void torrent::print_size(logger& l) { char tmp[300]; int temp = 0; int prev_size = 0; PRINT_SIZEOF(torrent) PRINT_OFFSETOF(torrent, m_tracker_address) PRINT_OFFSETOF(torrent, m_manager) PRINT_OFFSETOF(torrent, m_policy) PRINT_OFFSETOF(torrent, m_total_uploaded) PRINT_OFFSETOF(torrent, m_total_downloaded) PRINT_OFFSETOF(torrent, m_started) PRINT_OFFSETOF(torrent, m_torrent_file) PRINT_OFFSETOF(torrent, m_owning_storage) PRINT_OFFSETOF(torrent, m_storage) PRINT_OFFSETOF(torrent, m_connections) PRINT_OFFSETOF(torrent, m_web_seeds) #ifndef TORRENT_DISABLE_EXTENSIONS PRINT_OFFSETOF(torrent, m_extensions) #endif PRINT_OFFSETOF(torrent, m_tracker_timer) PRINT_OFFSETOF(torrent, m_stat) // some compilers don't like using offsetof on references it seems #ifndef _MSC_VER PRINT_OFFSETOF(torrent, m_ses) #endif PRINT_OFFSETOF(torrent, m_file_priority) PRINT_OFFSETOF(torrent, m_file_progress) PRINT_OFFSETOF(torrent, m_picker) PRINT_OFFSETOF(torrent, m_trackers) PRINT_OFFSETOF(torrent, m_time_critical_pieces) PRINT_OFFSETOF(torrent, m_username) PRINT_OFFSETOF(torrent, m_password) PRINT_OFFSETOF(torrent, m_net_interfaces) PRINT_OFFSETOF(torrent, m_save_path) PRINT_OFFSETOF(torrent, m_url) PRINT_OFFSETOF(torrent, m_uuid) PRINT_OFFSETOF(torrent, m_source_feed_url) PRINT_OFFSETOF(torrent, m_torrent_file_buf) PRINT_OFFSETOF(torrent, m_verified) PRINT_OFFSETOF(torrent, m_error) PRINT_OFFSETOF(torrent, m_error_file) PRINT_OFFSETOF(torrent, m_resume_data) PRINT_OFFSETOF(torrent, m_resume_entry) PRINT_OFFSETOF(torrent, m_name) PRINT_OFFSETOF(torrent, m_storage_constructor) PRINT_OFFSETOF(torrent, m_added_time) PRINT_OFFSETOF(torrent, m_completed_time) PRINT_OFFSETOF(torrent, m_last_saved_resume) PRINT_OFFSETOF(torrent, m_last_seen_complete) PRINT_OFFSETOF(torrent, m_swarm_last_seen_complete) PRINT_OFFSETOF(torrent, m_num_verified) #ifndef TORRENT_DISABLE_ENCRYPTION PRINT_OFFSETOF(torrent, m_obfuscated_hash) #endif PRINT_OFFSETOF(torrent, m_ratio) PRINT_OFFSETOF(torrent, m_available_free_upload) PRINT_OFFSETOF(torrent, m_average_piece_time) PRINT_OFFSETOF(torrent, m_piece_time_deviation) PRINT_OFFSETOF(torrent, m_total_failed_bytes) PRINT_OFFSETOF(torrent, m_total_redundant_bytes) // PRINT_OFFSETOF(torrent, m_upload_mode_time:24) // PRINT_OFFSETOF(torrent, m_state:3) // PRINT_OFFSETOF(torrent, m_storage_mode:2) // PRINT_OFFSETOF(torrent, m_announcing:1) // PRINT_OFFSETOF(torrent, m_waiting_tracker:1) // PRINT_OFFSETOF(torrent, m_seed_mode:1) // PRINT_OFFSETOF(torrent, m_active_time:24) PRINT_OFFSETOF(torrent, m_last_working_tracker) // PRINT_OFFSETOF(torrent, m_finished_time:24) // PRINT_OFFSETOF(torrent, m_sequential_download:1) // PRINT_OFFSETOF(torrent, m_got_tracker_response:1) // PRINT_OFFSETOF(torrent, m_connections_initialized:1) // PRINT_OFFSETOF(torrent, m_super_seeding:1) // PRINT_OFFSETOF(torrent, m_override_resume_data:1) // PRINT_OFFSETOF(torrent, m_resolving_country:1) // PRINT_OFFSETOF(torrent, m_resolve_countries:1) // PRINT_OFFSETOF(torrent, m_need_save_resume_data:1) // PRINT_OFFSETOF(torrent, m_seeding_time:24) PRINT_OFFSETOF(torrent, m_time_scaler) // PRINT_OFFSETOF(torrent, m_max_uploads:24) // PRINT_OFFSETOF(torrent, m_num_uploads:24) // PRINT_OFFSETOF(torrent, m_block_size_shift:5) // PRINT_OFFSETOF(torrent, m_has_incoming:1) // PRINT_OFFSETOF(torrent, m_files_checked:1) // PRINT_OFFSETOF(torrent, m_queued_for_checking:1) // PRINT_OFFSETOF(torrent, m_max_connections:24) // PRINT_OFFSETOF(torrent, m_padding:24) PRINT_OFFSETOF(torrent, m_sequence_number) // PRINT_OFFSETOF(torrent, m_complete:24) PRINT_OFFSETOF(torrent, m_priority) // PRINT_OFFSETOF(torrent, m_incomplete:24) // PRINT_OFFSETOF(torrent, m_progress_ppm:20) // PRINT_OFFSETOF(torrent, m_abort:1) // PRINT_OFFSETOF(torrent, m_announce_to_dht:1) // PRINT_OFFSETOF(torrent, m_announce_to_trackers:1) // PRINT_OFFSETOF(torrent, m_announce_to_lsd:1) // PRINT_OFFSETOF(torrent, m_allow_peers:1) // PRINT_OFFSETOF(torrent, m_upload_mode:1) // PRINT_OFFSETOF(torrent, m_auto_managed:1) PRINT_OFFSETOF(torrent, m_last_scrape) } #undef PRINT_SIZEOF #undef PRINT_OFFSETOF #endif int root2(int x) { int ret = 0; x >>= 1; while (x > 0) { // if this assert triggers, the block size // is not an even 2 exponent! TORRENT_ASSERT(x == 1 || (x & 1) == 0); ++ret; x >>= 1; } return ret; } #ifndef TORRENT_DISABLE_EXTENSIONS // defined in ut_pex.cpp bool was_introduced_by(peer_plugin const*, tcp::endpoint const&); #endif torrent::torrent( session_impl& ses , tcp::endpoint const& net_interface , int block_size , int seq , add_torrent_params const& p , sha1_hash const& info_hash) : m_policy(this) , m_total_uploaded(0) , m_total_downloaded(0) , m_started(time_now()) , m_storage(0) , m_num_connecting(0) , m_tracker_timer(ses.m_io_service) , m_ses(ses) , m_host_resolver(ses.m_io_service) , m_trackerid(p.trackerid) , m_save_path(complete(p.save_path)) , m_url(p.url) , m_uuid(p.uuid) , m_source_feed_url(p.source_feed_url) , m_storage_constructor(p.storage) , m_added_time(time(0)) , m_completed_time(0) , m_last_saved_resume(time(0)) , m_last_seen_complete(0) , m_swarm_last_seen_complete(0) , m_num_verified(0) , m_ratio(0.f) , m_available_free_upload(0) , m_average_piece_time(0) , m_piece_time_deviation(0) , m_total_failed_bytes(0) , m_total_redundant_bytes(0) , m_sequence_number(seq) , m_upload_mode_time(0) , m_state(torrent_status::checking_resume_data) , m_storage_mode(p.storage_mode) , m_announcing(false) , m_waiting_tracker(false) , m_seed_mode(false) , m_active_time(0) , m_last_working_tracker(-1) , m_finished_time(0) , m_sequential_download(false) , m_got_tracker_response(false) , m_connections_initialized(false) , m_super_seeding(false) , m_following(false) , m_override_resume_data(p.flags & add_torrent_params::flag_override_resume_data) #ifndef TORRENT_DISABLE_RESOLVE_COUNTRIES , m_resolving_country(false) , m_resolve_countries(false) #endif , m_need_save_resume_data(true) , m_seeding_time(0) , m_time_scaler(0) , m_max_uploads((1<<24)-1) , m_save_resume_flags(0) , m_num_uploads(0) , m_block_size_shift(root2(block_size)) , m_has_incoming(false) , m_files_checked(false) , m_queued_for_checking(false) , m_max_connections((1<<24)-1) , m_padding(0) , m_complete(0xffffff) , m_priority(0) , m_incomplete(0xffffff) , m_progress_ppm(0) , m_abort(false) , m_announce_to_dht((p.flags & add_torrent_params::flag_paused) == 0) , m_announce_to_trackers((p.flags & add_torrent_params::flag_paused) == 0) , m_announce_to_lsd((p.flags & add_torrent_params::flag_paused) == 0) , m_allow_peers((p.flags & add_torrent_params::flag_paused) == 0) , m_upload_mode(p.flags & add_torrent_params::flag_upload_mode) , m_auto_managed(p.flags & add_torrent_params::flag_auto_managed) , m_share_mode(p.flags & add_torrent_params::flag_share_mode) , m_last_download(0) , m_last_upload(0) , m_last_scrape(0) , m_downloaded(0xffffff) , m_interface_index(0) , m_graceful_pause_mode(false) , m_need_connect_boost(true) , m_lsd_seq(0) , m_magnet_link(false) , m_apply_ip_filter(p.flags & add_torrent_params::flag_apply_ip_filter) , m_merge_resume_trackers(p.flags & add_torrent_params::flag_merge_resume_trackers) , m_state_subscription(p.flags & add_torrent_params::flag_update_subscribe) , m_in_state_updates(false) , m_is_active_download(false) , m_is_active_finished(false) { if (!p.name.empty()) m_name.reset(new std::string(p.name)); // if there is resume data already, we don't need to trigger the initial save // resume data if (!p.resume_data.empty() && (p.flags & add_torrent_params::flag_override_resume_data) == 0) m_need_save_resume_data = false; #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS m_resume_data_loaded = false; #endif #if TORRENT_USE_UNC_PATHS m_save_path = canonicalize_path(m_save_path); #endif if (!m_apply_ip_filter) ++m_ses.m_non_filtered_torrents; update_guage(); if (!p.ti || !p.ti->is_valid()) { // we don't have metadata for this torrent. We'll download // it either through the URL passed in, or through a metadata // extension. Make sure that when we save resume data for this // torrent, we also save the metadata m_magnet_link = true; } if (!m_torrent_file) m_torrent_file = (p.ti ? p.ti : new torrent_info(info_hash, p.name)); m_trackers = m_torrent_file->trackers(); if (m_torrent_file->is_valid()) { m_seed_mode = p.flags & add_torrent_params::flag_seed_mode; m_connections_initialized = true; m_block_size_shift = root2((std::min)(block_size, m_torrent_file->piece_length())); } if (!m_url.empty() && m_uuid.empty()) m_uuid = m_url; TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("creating torrent: %s", torrent_file().name().c_str()); #endif m_net_interfaces.push_back(tcp::endpoint(net_interface.address(), 0)); m_file_priority = p.file_priorities; if (m_seed_mode) m_verified.resize(m_torrent_file->num_pieces(), false); m_resume_data = p.resume_data; #ifndef TORRENT_DISABLE_ENCRYPTION hasher h; h.update("req2", 4); h.update((char*)&m_torrent_file->info_hash()[0], 20); m_obfuscated_hash = h.final(); #endif #ifdef TORRENT_DEBUG m_files_checked = false; #endif INVARIANT_CHECK; if (p.flags & add_torrent_params::flag_sequential_download) m_sequential_download = true; if (p.flags & add_torrent_params::flag_super_seeding) m_super_seeding = true; set_max_uploads(p.max_uploads, false); set_max_connections(p.max_connections, false); set_upload_limit(p.upload_limit, false); set_download_limit(p.download_limit, false); if (!m_name && !m_url.empty()) m_name.reset(new std::string(m_url)); #ifndef TORRENT_NO_DEPRECATE if (p.tracker_url && std::strlen(p.tracker_url) > 0) { m_trackers.push_back(announce_entry(p.tracker_url)); m_trackers.back().fail_limit = 0; m_trackers.back().source = announce_entry::source_magnet_link; m_torrent_file->add_tracker(p.tracker_url); } #endif for (std::vector::const_iterator i = p.trackers.begin() , end(p.trackers.end()); i != end; ++i) { m_trackers.push_back(announce_entry(*i)); m_trackers.back().fail_limit = 0; m_trackers.back().source = announce_entry::source_magnet_link; m_torrent_file->add_tracker(*i); } if (settings().prefer_udp_trackers) prioritize_udp_trackers(); } #if 0 // NON BOTTLED VERSION. SUPPORTS PROGRESS REPORTING // since this download is not bottled, this callback will // be called every time we receive another piece of the // .torrent file void torrent::on_torrent_download(error_code const& ec , http_parser const& parser , char const* data, int size) { if (m_abort) return; if (ec && ec != asio::error::eof) { set_error(ec, m_url); pause(); return; } if (size > 0) { m_torrent_file_buf.insert(m_torrent_file_buf.end(), data, data + size); if (parser.content_length() > 0) set_progress_ppm(boost::int64_t(m_torrent_file_buf.size()) * 1000000 / parser.content_length()); } if (parser.header_finished() && parser.status_code() != 200) { set_error(error_code(parser.status_code(), get_http_category()), parser.message()); pause(); return; } if (!ec) return; // if this was received with chunked encoding, we need to strip out // the chunk headers size = parser.collapse_chunk_headers((char*)&m_torrent_file_buf[0], m_torrent_file_buf.size()); m_torrent_file_buf.resize(size); std::string const& encoding = parser.header("content-encoding"); if ((encoding == "gzip" || encoding == "x-gzip") && m_torrent_file_buf.size()) { std::vector buf; std::string error; if (inflate_gzip(&m_torrent_file_buf[0], m_torrent_file_buf.size() , buf, 4 * 1024 * 1024, error)) { set_error(errors::http_failed_decompress, m_url); pause(); std::vector().swap(m_torrent_file_buf); return; } m_torrent_file_buf.swap(buf); } // we're done! error_code e; intrusive_ptr tf(new torrent_info( &m_torrent_file_buf[0], m_torrent_file_buf.size(), e)); if (e) { set_error(e, m_url); pause(); std::vector().swap(m_torrent_file_buf); return; } std::vector().swap(m_torrent_file_buf); // update our torrent_info object and move the // torrent from the old info-hash to the new one // as we replace the torrent_info object #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS int num_torrents = m_ses.m_torrents.size(); #endif // we're about to erase the session's reference to this // torrent, create another reference boost::shared_ptr me(shared_from_this()); m_ses.remove_torrent_impl(me, 0); m_torrent_file = tf; // now, we might already have this torrent in the session. session_impl::torrent_map::iterator i = m_ses.m_torrents.find(m_torrent_file->info_hash()); if (i != m_ses.m_torrents.end()) { if (!m_uuid.empty() && i->second->uuid().empty()) i->second->set_uuid(m_uuid); if (!m_url.empty() && i->second->url().empty()) i->second->set_url(m_url); if (!m_source_feed_url.empty() && i->second->source_feed_url().empty()) i->second->set_source_feed_url(m_source_feed_url); // insert this torrent in the uuid index if (!m_uuid.empty() || !m_url.empty()) { m_ses.m_uuids.insert(std::make_pair(m_uuid.empty() ? m_url : m_uuid, i->second)); } set_error(error_code(errors::duplicate_torrent, get_libtorrent_category()), ""); abort(); return; } m_ses.m_torrents.insert(std::make_pair(m_torrent_file->info_hash(), me)); if (!m_uuid.empty()) m_ses.m_uuids.insert(std::make_pair(m_uuid, me)); TORRENT_ASSERT(num_torrents == int(m_ses.m_torrents.size())); // if the user added any trackers while downloading the // .torrent file, serge them into the new tracker list std::vector new_trackers = m_torrent_file->trackers(); for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { // if we already have this tracker, ignore it if (std::find_if(new_trackers.begin(), new_trackers.end() , boost::bind(&announce_entry::url, _1) == i->url) != new_trackers.end()) continue; // insert the tracker ordered by tier new_trackers.insert(std::find_if(new_trackers.begin(), new_trackers.end() , boost::bind(&announce_entry::tier, _1) >= i->tier), *i); } m_trackers.swap(new_trackers); #ifndef TORRENT_DISABLE_ENCRYPTION hasher h; h.update("req2", 4); h.update((char*)&m_torrent_file->info_hash()[0], 20); m_obfuscated_hash = h.final(); #endif if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(metadata_received_alert( get_handle())); } state_updated(); set_state(torrent_status::downloading); m_override_resume_data = true; init(); } #else void torrent::on_torrent_download(error_code const& ec , http_parser const& parser, char const* data, int size) { } #endif void torrent::start() { TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("starting torrent"); #endif TORRENT_ASSERT(!m_picker); if (!m_seed_mode) { m_picker.reset(new piece_picker()); std::fill(m_file_progress.begin(), m_file_progress.end(), 0); if (!m_resume_data.empty()) { int pos; error_code ec; if (lazy_bdecode(&m_resume_data[0], &m_resume_data[0] + m_resume_data.size(), m_resume_entry, ec, &pos) != 0) { std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("resume data rejected: %s pos: %d", ec.message().c_str(), pos); #endif if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(fastresume_rejected_alert(get_handle(), ec)); } } } } if (!m_torrent_file->is_valid() && !m_url.empty()) { // we need to download the .torrent file from m_url start_download_url(); } else if (m_torrent_file->is_valid()) { init(); } else { // we need to start announcing since we don't have any // metadata. To receive peers to ask for it. set_state(torrent_status::downloading_metadata); start_announcing(); } } bool torrent::is_active_download() const { return (m_state == torrent_status::downloading || m_state == torrent_status::downloading_metadata) && m_allow_peers && !m_abort; } bool torrent::is_active_finished() const { return (m_state == torrent_status::finished || m_state == torrent_status::seeding) && m_allow_peers && !m_abort; } void torrent::update_guage() { bool is_active_download = (m_state == torrent_status::downloading || m_state == torrent_status::downloading_metadata) && m_allow_peers && !m_abort; bool is_active_finished = (m_state == torrent_status::finished || m_state == torrent_status::seeding) && m_allow_peers && !m_abort; // update finished and downloading counters if (is_active_download != m_is_active_download) { if (is_active_download) m_ses.inc_active_downloading(); else m_ses.dec_active_downloading(); m_is_active_download = is_active_download; } if (is_active_finished != m_is_active_finished) { if (is_active_finished) m_ses.inc_active_finished(); else m_ses.dec_active_finished(); m_is_active_finished = is_active_finished; } } void torrent::start_download_url() { } void torrent::set_apply_ip_filter(bool b) { if (b == m_apply_ip_filter) return; if (b) { TORRENT_ASSERT(m_ses.m_non_filtered_torrents > 0); --m_ses.m_non_filtered_torrents; } else { ++m_ses.m_non_filtered_torrents; } m_apply_ip_filter = b; m_policy.ip_filter_updated(); state_updated(); } #ifndef TORRENT_DISABLE_DHT bool torrent::should_announce_dht() const { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_ses.m_listen_sockets.empty()) return false; if (!m_ses.m_dht) return false; if (m_torrent_file->is_valid() && !m_files_checked) return false; if (!m_announce_to_dht) return false; if (!m_allow_peers) return false; // if we don't have the metadata, and we're waiting // for a web server to serve it to us, no need to announce // because the info-hash is just the URL hash if (!m_torrent_file->is_valid() && !m_url.empty()) return false; // try to reduce the level of useless dht tracker requests by not // announcing empty torrents we do not follow. if (!m_following && last_have() == -1 ) return false; // don't announce private torrents if (m_torrent_file->is_valid() && m_torrent_file->priv()) return false; if (m_trackers.empty()) return true; if (!settings().use_dht_as_fallback) return true; int verified_trackers = 0; for (std::vector::const_iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) if (i->verified) ++verified_trackers; return verified_trackers == 0; } #endif torrent::~torrent() { if (!m_apply_ip_filter) { TORRENT_ASSERT(m_ses.m_non_filtered_torrents > 0); --m_ses.m_non_filtered_torrents; m_apply_ip_filter = true; } TORRENT_ASSERT(m_ses.is_network_thread()); // The invariant can't be maintained here, since the torrent // is being destructed, all weak references to it have been // reset, which means that all its peers already have an // invalidated torrent pointer (so it cannot be verified to be correct) // i.e. the invariant can only be maintained if all connections have // been closed by the time the torrent is destructed. And they are // supposed to be closed. So we can still do the invariant check. INVARIANT_CHECK; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING log_to_all_peers("DESTRUCTING TORRENT"); #endif TORRENT_ASSERT(m_abort); TORRENT_ASSERT(m_connections.empty()); if (!m_connections.empty()) disconnect_all(errors::torrent_aborted); } void torrent::read_piece(int piece) { if (m_abort) { // failed m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), piece, error_code(boost::system::errc::operation_canceled, get_system_category()))); return; } TORRENT_ASSERT(piece >= 0 && piece < m_torrent_file->num_pieces()); int piece_size = m_torrent_file->piece_size(piece); int blocks_in_piece = (piece_size + block_size() - 1) / block_size(); // if blocks_in_piece is 0, rp will leak TORRENT_ASSERT(blocks_in_piece > 0); TORRENT_ASSERT(piece_size > 0); read_piece_struct* rp = new read_piece_struct; rp->piece_data.reset(new (std::nothrow) char[piece_size]); rp->blocks_left = 0; rp->fail = false; peer_request r; r.piece = piece; r.start = 0; for (int i = 0; i < blocks_in_piece; ++i, r.start += block_size()) { r.length = (std::min)(piece_size - r.start, block_size()); filesystem().async_read(r, boost::bind(&torrent::on_disk_read_complete , shared_from_this(), _1, _2, r, rp)); ++rp->blocks_left; } } void torrent::get_pieces(std::vector *pieces, int count, int max_id, int since_id, uint32_t filter_flags, mutex *mut, condition_variable *cond, int *reqs) { if( !m_picker ) return; max_id = std::min( max_id, m_picker->last_have() ); int piece_size = m_torrent_file->piece_size(0); for( int i = max_id; i >= 0 && i > since_id && (*reqs) < count; i--) { if( m_picker->have_piece(i) && (m_picker->post_flags(i) & filter_flags) == m_picker->post_flags(i) ) { (*reqs)++; peer_request r; r.piece = i; r.start = 0; r.length = piece_size; filesystem().async_read(r, boost::bind(&torrent::on_disk_read_get_piece_complete , shared_from_this(), _1, _2, pieces, mut, cond, reqs)); } } } void torrent::on_disk_read_get_piece_complete(int ret, disk_io_job const& j, std::vector *pieces, mutex *mut, condition_variable *cond, int *reqs) { TORRENT_ASSERT(m_ses.is_network_thread()); mutex::scoped_lock l(*mut); //printf("on_disk_read_get_piece_complete ret=%d reqs=%d\n", ret, *reqs); if (ret > 0) { pieces->push_back( std::string(j.buffer, ret)); } else { printf("piece read error (database corrupt?) - setting we_dont_have(%d)\n", j.piece); we_dont_have(j.piece); } (*reqs)--; if (!(*reqs)) cond->notify_all(); } void torrent::send_share_mode() { #ifndef TORRENT_DISABLE_EXTENSIONS for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { if ((*i)->type() != peer_connection::bittorrent_connection) continue; bt_peer_connection* p = (bt_peer_connection*)*i; p->write_share_mode(); } #endif } void torrent::send_upload_only() { #ifndef TORRENT_DISABLE_EXTENSIONS if (share_mode()) return; if (super_seeding()) return; for (std::set::iterator i = m_connections.begin(); i != m_connections.end();) { // since the call to disconnect_if_redundant() may // delete the entry from this container, make sure // to increment the iterator early bt_peer_connection* p = (bt_peer_connection*)*i; ++i; if (p->type() == peer_connection::bittorrent_connection) p->write_upload_only(); p->disconnect_if_redundant(); } #endif } void torrent::set_share_mode(bool s) { if (s == m_share_mode) return; m_share_mode = s; // in share mode, all pieces have their priorities initialized to 0 std::fill(m_file_priority.begin(), m_file_priority.end(), !m_share_mode); update_piece_priorities(); if (m_share_mode) recalc_share_mode(); } void torrent::set_upload_mode(bool b) { if (b == m_upload_mode) return; m_upload_mode = b; state_updated(); send_upload_only(); if (m_upload_mode) { // clear request queues of all peers for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { peer_connection* p = (*i); p->cancel_all_requests(); } // this is used to try leaving upload only mode periodically m_upload_mode_time = 0; } else { // reset last_connected, to force fast reconnect after leaving upload mode for (policy::iterator i = m_policy.begin_peer() , end(m_policy.end_peer()); i != end; ++i) { (*i)->last_connected = 0; } // send_block_requests on all peers for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { peer_connection* p = (*i); p->send_block_requests(); } } } void torrent::handle_disk_error(disk_io_job const& j, peer_connection* c) { TORRENT_ASSERT(m_ses.is_network_thread()); if (!j.error) return; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("disk error: (%d) %s in file: %s", j.error.value(), j.error.message().c_str() , j.error_file.c_str()); #endif TORRENT_ASSERT(j.piece >= 0); piece_block block_finished(j.piece, j.offset / block_size()); if (j.action == disk_io_job::write) { // we failed to write j.piece to disk tell the piece picker if (has_picker() && j.piece >= 0) picker().write_failed(block_finished); } if (j.error == #if BOOST_VERSION == 103500 error_code(boost::system::posix_error::not_enough_memory, get_posix_category()) #elif BOOST_VERSION > 103500 error_code(boost::system::errc::not_enough_memory, boost::system::generic_category()) #else asio::error::no_memory #endif ) { if (alerts().should_post()) alerts().post_alert(file_error_alert(j.error_file, get_handle(), j.error)); if (c) c->disconnect(errors::no_memory); return; } // notify the user of the error if (alerts().should_post()) alerts().post_alert(file_error_alert(j.error_file, get_handle(), j.error)); // put the torrent in an error-state set_error(j.error, j.error_file); if (j.action == disk_io_job::write && (j.error == boost::system::errc::read_only_file_system || j.error == boost::system::errc::permission_denied || j.error == boost::system::errc::operation_not_permitted || j.error == boost::system::errc::no_space_on_device || j.error == boost::system::errc::file_too_large)) { // if we failed to write, stop downloading and just // keep seeding. // TODO: 1 make this depend on the error and on the filesystem the // files are being downloaded to. If the error is no_space_left_on_device // and the filesystem doesn't support sparse files, only zero the priorities // of the pieces that are at the tails of all files, leaving everything // up to the highest written piece in each file set_upload_mode(true); return; } // if the error appears to be more serious than a full disk, just pause the torrent pause(); } void torrent::on_disk_read_complete(int ret, disk_io_job const& j, peer_request r, read_piece_struct* rp) { TORRENT_ASSERT(m_ses.is_network_thread()); disk_buffer_holder buffer(m_ses, j.buffer); --rp->blocks_left; if (ret != r.length) { rp->fail = true; rp->error = j.error; handle_disk_error(j); } else { std::memcpy(rp->piece_data.get() + r.start, j.buffer, r.length); } if (rp->blocks_left == 0) { int size = m_torrent_file->piece_size(r.piece); if (rp->fail) { m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), r.piece, rp->error)); } else { m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), r.piece, rp->piece_data, size)); } delete rp; } } // [MF] test only? void torrent::add_piece(int piece, char const* data, int size, int flags) { TORRENT_ASSERT(m_ses.is_network_thread()); //[MF] increase num pieces to ensure we are not a seeder increase_num_pieces(piece+2); TORRENT_ASSERT(piece >= 0 && piece < m_torrent_file->num_pieces()); int piece_size = size; // avoid crash trying to access the picker when there is none if (!has_picker()) return; if (picker().have_piece(piece) && (flags & torrent::overwrite_existing) == 0) return; peer_request p; p.piece = piece; p.start = 0; picker().inc_refcount(piece, 0); p.length = piece_size; char* buffer = m_ses.allocate_disk_buffer("add piece"); // out of memory if (buffer == 0) { picker().dec_refcount(piece, 0); return; } disk_buffer_holder holder(m_ses, buffer); std::memcpy(buffer, data + p.start, p.length); filesystem().async_write(p, holder, boost::bind(&torrent::on_disk_write_complete , shared_from_this(), _1, _2, p)); piece_block block(piece, 0); picker().mark_as_downloading(block, 0, piece_picker::fast); picker().mark_as_writing(block, 0); async_verify_piece(piece, boost::bind(&torrent::piece_finished , shared_from_this(), piece, _1, _2, p.length)); picker().dec_refcount(piece, 0); } void torrent::on_disk_write_complete(int ret, disk_io_job const& j , peer_request p) { //printf("on_disk_write_complete: size %d\n", j.buffer_size); TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_abort) { piece_block block_finished(p.piece, p.start / block_size()); return; } piece_block block_finished(p.piece, p.start / block_size()); if (ret == -1) { handle_disk_error(j); return; } if (!has_picker()) return; // if we already have this block, just ignore it. // this can happen if the same block is passed in through // add_piece() multiple times if (picker().is_finished(block_finished)) return; picker().mark_as_finished(block_finished, 0); } void torrent::on_disk_cache_complete(int ret, disk_io_job const& j) { // suggest this piece to all peers for (peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) (*i)->send_suggest(j.piece); } bool torrent::add_merkle_nodes(std::map const& nodes, int piece) { return m_torrent_file->add_merkle_nodes(nodes, piece); } peer_request torrent::to_req(piece_block const& p) const { int block_offset = p.block_index * block_size(); int block = (std::min)(torrent_file().piece_size( p.piece_index) - block_offset, int(block_size())); TORRENT_ASSERT(block > 0); TORRENT_ASSERT(block <= block_size()); peer_request r; r.piece = p.piece_index; r.start = block_offset; r.length = block; return r; } std::string torrent::name() const { if (valid_metadata()) return m_torrent_file->name(); if (m_name) return *m_name; return ""; } #ifndef TORRENT_DISABLE_EXTENSIONS void torrent::add_extension(boost::shared_ptr ext) { m_extensions.push_back(ext); } void torrent::add_extension(boost::function(torrent*, void*)> const& ext , void* userdata) { boost::shared_ptr tp(ext(this, userdata)); if (!tp) return; add_extension(tp); for (peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { peer_connection* p = *i; boost::shared_ptr pp(tp->new_connection(p)); if (pp) p->add_extension(pp); } // if files are checked for this torrent, call the extension // to let it initialize itself if (m_connections_initialized) tp->on_files_checked(); } #endif #ifdef TORRENT_USE_OPENSSL #if BOOST_VERSION >= 104700 bool torrent::verify_peer_cert(bool preverified, boost::asio::ssl::verify_context& ctx) { // if the cert wasn't signed by the correct CA, fail the verification if (!preverified) return false; // we're only interested in checking the certificate at the end of the chain. int depth = X509_STORE_CTX_get_error_depth(ctx.native_handle()); if (depth > 0) return true; X509* cert = X509_STORE_CTX_get_current_cert(ctx.native_handle()); // Go through the alternate names in the certificate looking for matching DNS entries GENERAL_NAMES* gens = static_cast( X509_get_ext_d2i(cert, NID_subject_alt_name, 0, 0)); #if defined(TORRENT_VERBOSE_LOGGING) || defined(TORRENT_LOGGING) std::string names; bool match = false; #endif for (int i = 0; i < sk_GENERAL_NAME_num(gens); ++i) { GENERAL_NAME* gen = sk_GENERAL_NAME_value(gens, i); if (gen->type != GEN_DNS) continue; ASN1_IA5STRING* domain = gen->d.dNSName; if (domain->type != V_ASN1_IA5STRING || !domain->data || !domain->length) continue; const char* torrent_name = reinterpret_cast(domain->data); std::size_t name_length = domain->length; #if defined(TORRENT_VERBOSE_LOGGING) || defined(TORRENT_LOGGING) if (i > 1) names += " | n: "; names.append(torrent_name, name_length); #endif if (strncmp(torrent_name, "*", name_length) == 0 || strncmp(torrent_name, m_torrent_file->name().c_str(), name_length) == 0) { #if defined(TORRENT_VERBOSE_LOGGING) || defined(TORRENT_LOGGING) match = true; // if we're logging, keep looping over all names, // for completeness of the log continue; #endif return true; } } // no match in the alternate names, so try the common names. We should only // use the "most specific" common name, which is the last one in the list. X509_NAME* name = X509_get_subject_name(cert); int i = -1; ASN1_STRING* common_name = 0; while ((i = X509_NAME_get_index_by_NID(name, NID_commonName, i)) >= 0) { X509_NAME_ENTRY* name_entry = X509_NAME_get_entry(name, i); common_name = X509_NAME_ENTRY_get_data(name_entry); } if (common_name && common_name->data && common_name->length) { const char* torrent_name = reinterpret_cast(common_name->data); std::size_t name_length = common_name->length; #if defined(TORRENT_VERBOSE_LOGGING) || defined(TORRENT_LOGGING) if (!names.empty()) names += " | n: "; names.append(torrent_name, name_length); #endif if (strncmp(torrent_name, "*", name_length) == 0 || strncmp(torrent_name, m_torrent_file->name().c_str(), name_length) == 0) { #if !defined(TORRENT_VERBOSE_LOGGING) && !defined(TORRENT_LOGGING) return true; #else match = true; #endif } } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("<== incoming SSL CONNECTION [ n: %s | match: %s ]" , names.c_str(), match?"yes":"no"); return match; #endif return false; } #endif // BOOST_VERSION void torrent::init_ssl(std::string const& cert) { using boost::asio::ssl::context; // this is needed for openssl < 1.0 to decrypt keys created by openssl 1.0+ OpenSSL_add_all_algorithms(); boost::uint64_t now = total_microseconds(time_now_hires() - min_time()); // assume 9 bits of entropy (i.e. about 1 millisecond) RAND_add(&now, 8, 1.125); RAND_add(&info_hash()[0], 20, 3); // entropy is also added on incoming and completed connection attempts TORRENT_ASSERT(RAND_status() == 1); #if BOOST_VERSION >= 104700 // create the SSL context for this torrent. We need to // inject the root certificate, and no other, to // verify other peers against boost::shared_ptr ctx( new (std::nothrow) context(m_ses.m_io_service, context::sslv23)); if (!ctx) { error_code ec(::ERR_get_error(), asio::error::get_ssl_category()); set_error(ec, "SSL context"); pause(); return; } ctx->set_options(context::default_workarounds | boost::asio::ssl::context::no_sslv2 | boost::asio::ssl::context::single_dh_use); error_code ec; ctx->set_verify_mode(context::verify_peer | context::verify_fail_if_no_peer_cert | context::verify_client_once, ec); if (ec) { set_error(ec, "SSL verify mode"); pause(); return; } // the verification function verifies the distinguished name // of a peer certificate to make sure it matches the info-hash // of the torrent, or that it's a "star-cert" ctx->set_verify_callback(boost::bind(&torrent::verify_peer_cert, this, _1, _2), ec); if (ec) { set_error(ec, "SSL verify callback"); pause(); return; } SSL_CTX* ssl_ctx = ctx->impl(); // create a new x.509 certificate store X509_STORE* cert_store = X509_STORE_new(); if (!cert_store) { error_code ec(::ERR_get_error(), asio::error::get_ssl_category()); set_error(ec, "x.509 certificate store"); pause(); return; } // wrap the PEM certificate in a BIO, for openssl to read BIO* bp = BIO_new_mem_buf((void*)cert.c_str(), cert.size()); // parse the certificate into OpenSSL's internal // representation X509* certificate = PEM_read_bio_X509_AUX(bp, 0, 0, 0); BIO_free(bp); if (!certificate) { error_code ec(::ERR_get_error(), asio::error::get_ssl_category()); X509_STORE_free(cert_store); set_error(ec, "x.509 certificate"); pause(); return; } // add cert to cert_store X509_STORE_add_cert(cert_store, certificate); X509_free(certificate); // and lastly, replace the default cert store with ours SSL_CTX_set_cert_store(ssl_ctx, cert_store); #if 0 char filename[100]; snprintf(filename, sizeof(filename), "/tmp/%d.pem", rand()); FILE* f = fopen(filename, "w+"); fwrite(cert.c_str(), cert.size(), 1, f); fclose(f); ctx->load_verify_file(filename); #endif // if all went well, set the torrent ssl context to this one m_ssl_ctx = ctx; // tell the client we need a cert for this torrent alerts().post_alert(torrent_need_cert_alert(get_handle())); #else set_error(asio::error::operation_not_supported, "x.509 certificate"); pause(); #endif } #endif // TORRENT_OPENSSL peer_connection* torrent::find_lowest_ranking_peer() const { const_peer_iterator lowest_rank = end(); for (const_peer_iterator i = begin(); i != end(); ++i) { // disconnecting peers don't count if ((*i)->is_disconnecting()) continue; if (lowest_rank == end() || (*lowest_rank)->peer_rank() > (*i)->peer_rank()) lowest_rank = i; } if (lowest_rank == end()) return NULL; return *lowest_rank; } // this may not be called from a constructor because of the call to // shared_from_this() void torrent::init() { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(m_torrent_file->is_valid()); TORRENT_ASSERT(m_torrent_file->num_files() > 0); TORRENT_ASSERT(m_torrent_file->total_size() >= 0); if (m_file_priority.size() > m_torrent_file->num_files()) m_file_priority.resize(m_torrent_file->num_files()); #ifdef TORRENT_USE_OPENSSL std::string cert = m_torrent_file->ssl_cert(); if (!cert.empty()) init_ssl(cert); #endif m_file_priority.resize(m_torrent_file->num_files(), 1); m_file_progress.resize(m_torrent_file->num_files(), 0); m_block_size_shift = root2((std::min)(int(block_size()), m_torrent_file->piece_length())); if (m_torrent_file->num_pieces() > piece_picker::max_pieces) { set_error(errors::too_many_pieces_in_torrent, ""); pause(); return; } if (m_torrent_file->num_pieces() == 0) { set_error(errors::torrent_invalid_length, ""); pause(); return; } // the shared_from_this() will create an intentional // cycle of ownership, se the hpp file for description. m_owning_storage = new piece_manager(shared_from_this(), m_torrent_file , m_save_path, m_ses.m_swarmDb, m_ses.m_disk_thread, m_storage_constructor , (storage_mode_t)m_storage_mode, m_file_priority); m_storage = m_owning_storage.get(); if (has_picker()) { int blocks_per_piece = (m_torrent_file->piece_length() + block_size() - 1) / block_size(); int blocks_in_last_piece = ((m_torrent_file->total_size() % m_torrent_file->piece_length()) + block_size() - 1) / block_size(); m_picker->init(blocks_per_piece, blocks_in_last_piece, m_torrent_file->num_pieces()); } if (m_share_mode) { // in share mode, all pieces have their priorities initialized to 0 std::fill(m_file_priority.begin(), m_file_priority.end(), 0); } if (!m_connections_initialized) { m_connections_initialized = true; // all peer connections have to initialize themselves now that the metadata // is available for (torrent::peer_iterator i = m_connections.begin(); i != m_connections.end();) { peer_connection* pc = *i; ++i; if (pc->is_disconnecting()) continue; pc->on_metadata_impl(); if (pc->is_disconnecting()) continue; pc->init(); } } // in case file priorities were passed in via the add_torrent_params // ans also in the case of share mode, we need to update the priorities update_piece_priorities(); std::vector const& web_seeds = m_torrent_file->web_seeds(); m_web_seeds.insert(m_web_seeds.end(), web_seeds.begin(), web_seeds.end()); if (m_seed_mode) { m_ses.m_io_service.post(boost::bind(&torrent::files_checked, shared_from_this())); std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS m_resume_data_loaded = true; #endif return; } set_state(torrent_status::checking_resume_data); if (m_resume_entry.type() == lazy_entry::dict_t) { int ev = 0; if (m_resume_entry.dict_find_string_value("file-format") != "libtorrent resume file") ev = errors::invalid_file_tag; if (!ev && m_resume_entry.dict_find_int_value("file-version") < 2 ) ev = errors::invalid_file_tag; std::string info_hash = m_resume_entry.dict_find_string_value("info-hash"); if (!ev && info_hash.empty()) ev = errors::missing_info_hash; if (!ev && sha1_hash(info_hash) != m_torrent_file->info_hash()) ev = errors::mismatching_info_hash; if (ev && m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(fastresume_rejected_alert(get_handle() , error_code(ev, get_libtorrent_category()))); } if (ev) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("fastresume data rejected: %s" , error_code(ev, get_libtorrent_category()).message().c_str()); #endif std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); } else { read_resume_data(m_resume_entry); } } #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS m_resume_data_loaded = true; #endif TORRENT_ASSERT(block_size() > 0); m_storage->async_check_fastresume(&m_resume_entry , boost::bind(&torrent::on_resume_data_checked , shared_from_this(), _1, _2)); } bt_peer_connection* torrent::find_introducer(tcp::endpoint const& ep) const { #ifndef TORRENT_DISABLE_EXTENSIONS for (const_peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { if ((*i)->type() != peer_connection::bittorrent_connection) continue; bt_peer_connection* p = (bt_peer_connection*)(*i); if (!p->supports_holepunch()) continue; peer_plugin const* pp = p->find_plugin("ut_pex"); if (!pp) continue; if (was_introduced_by(pp, ep)) return (bt_peer_connection*)p; } #endif return 0; } bt_peer_connection* torrent::find_peer(tcp::endpoint const& ep) const { for (const_peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { peer_connection* p = *i; if (p->type() != peer_connection::bittorrent_connection) continue; if (p->remote() == ep) return (bt_peer_connection*)p; } return 0; } void torrent::on_resume_data_checked(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (ret == piece_manager::fatal_disk_error) { handle_disk_error(j); auto_managed(false); pause(); set_state(torrent_status::queued_for_checking); std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); return; } state_updated(); if (m_resume_entry.type() == lazy_entry::dict_t) { using namespace libtorrent::detail; // for read_*_endpoint() peer_id id(0); if (lazy_entry const* peers_entry = m_resume_entry.dict_find_string("peers")) { int num_peers = peers_entry->string_length() / (sizeof(address_v4::bytes_type) + 2); char const* ptr = peers_entry->string_ptr(); for (int i = 0; i < num_peers; ++i) { m_policy.add_peer(read_v4_endpoint(ptr) , id, peer_info::resume_data, 0); } } if (lazy_entry const* banned_peers_entry = m_resume_entry.dict_find_string("banned_peers")) { int num_peers = banned_peers_entry->string_length() / (sizeof(address_v4::bytes_type) + 2); char const* ptr = banned_peers_entry->string_ptr(); for (int i = 0; i < num_peers; ++i) { policy::peer* p = m_policy.add_peer(read_v4_endpoint(ptr) , id, peer_info::resume_data, 0); if (p) m_policy.ban_peer(p); } } #if TORRENT_USE_IPV6 if (lazy_entry const* peers6_entry = m_resume_entry.dict_find_string("peers6")) { int num_peers = peers6_entry->string_length() / (sizeof(address_v6::bytes_type) + 2); char const* ptr = peers6_entry->string_ptr(); for (int i = 0; i < num_peers; ++i) { m_policy.add_peer(read_v6_endpoint(ptr) , id, peer_info::resume_data, 0); } } if (lazy_entry const* banned_peers6_entry = m_resume_entry.dict_find_string("banned_peers6")) { int num_peers = banned_peers6_entry->string_length() / (sizeof(address_v6::bytes_type) + 2); char const* ptr = banned_peers6_entry->string_ptr(); for (int i = 0; i < num_peers; ++i) { policy::peer* p = m_policy.add_peer(read_v6_endpoint(ptr) , id, peer_info::resume_data, 0); if (p) m_policy.ban_peer(p); } } #endif // parse out "peers" from the resume data and add them to the peer list if (lazy_entry const* peers_entry = m_resume_entry.dict_find_list("peers")) { for (int i = 0; i < peers_entry->list_size(); ++i) { lazy_entry const* e = peers_entry->list_at(i); if (e->type() != lazy_entry::dict_t) continue; std::string ip = e->dict_find_string_value("ip"); int port = e->dict_find_int_value("port"); if (ip.empty() || port == 0) continue; error_code ec; tcp::endpoint a(address::from_string(ip, ec), (unsigned short)port); if (ec) continue; m_policy.add_peer(a, id, peer_info::resume_data, 0); } } // parse out "banned_peers" and add them as banned if (lazy_entry const* banned_peers_entry = m_resume_entry.dict_find_list("banned_peers")) { for (int i = 0; i < banned_peers_entry->list_size(); ++i) { lazy_entry const* e = banned_peers_entry->list_at(i); if (e->type() != lazy_entry::dict_t) continue; std::string ip = e->dict_find_string_value("ip"); int port = e->dict_find_int_value("port"); if (ip.empty() || port == 0) continue; error_code ec; tcp::endpoint a(address::from_string(ip, ec), (unsigned short)port); if (ec) continue; policy::peer* p = m_policy.add_peer(a, id, peer_info::resume_data, 0); if (p) m_policy.ban_peer(p); } } } // only report this error if the user actually provided resume data if ((j.error || ret != 0) && !m_resume_data.empty() && m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(fastresume_rejected_alert(get_handle(), j.error)); } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING if (ret != 0) { debug_log("fastresume data rejected: ret: %d (%d) %s" , ret, j.error.value(), j.error.message().c_str()); } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING else debug_log("fastresume data accepted"); #endif #endif // if ret != 0, it means we need a full check. We don't necessarily need // that when the resume data check fails. For instance, if the resume data // is incorrect, but we don't have any files, we skip the check and initialize // the storage to not have anything. if (ret == 0) { // there are either no files for this torrent // or the resume_data was accepted if (!j.error && m_resume_entry.type() == lazy_entry::dict_t) { // parse have bitmask lazy_entry const* pieces = m_resume_entry.dict_find("pieces"); if (pieces && pieces->type() == lazy_entry::string_t && int(pieces->string_length()) == m_torrent_file->num_pieces()) { char const* pieces_str = pieces->string_ptr(); for (int i = 0, end(pieces->string_length()); i < end; ++i) { if (pieces_str[i] & 1) we_have(i, pieces_str[i]>>1); if (m_seed_mode && (pieces_str[i] & 2)) m_verified.set_bit(i); } } // parse unfinished pieces int num_blocks_per_piece = static_cast(torrent_file().piece_length()) / block_size(); } files_checked(); } else { // either the fastresume data was rejected or there are // some files set_state(torrent_status::queued_for_checking); if (should_check_files()) queue_torrent_check(); } std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); } void torrent::queue_torrent_check() { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_queued_for_checking) return; m_queued_for_checking = true; m_ses.queue_check_torrent(shared_from_this()); } void torrent::dequeue_torrent_check() { TORRENT_ASSERT(m_ses.is_network_thread()); if (!m_queued_for_checking) return; m_queued_for_checking = false; m_ses.dequeue_check_torrent(shared_from_this()); } void torrent::force_recheck() { if (!valid_metadata()) return; // if the torrent is already queued to check its files // don't do anything if (should_check_files() || m_state == torrent_status::checking_resume_data) return; clear_error(); disconnect_all(errors::stopping_torrent); stop_announcing(); m_owning_storage->async_release_files(); if (!m_picker) m_picker.reset(new piece_picker()); std::fill(m_file_progress.begin(), m_file_progress.end(), 0); int blocks_per_piece = (m_torrent_file->piece_length() + block_size() - 1) / block_size(); int blocks_in_last_piece = ((m_torrent_file->total_size() % m_torrent_file->piece_length()) + block_size() - 1) / block_size(); m_picker->init(blocks_per_piece, blocks_in_last_piece, m_torrent_file->num_pieces()); // assume that we don't have anything TORRENT_ASSERT(m_picker->num_have() == 0); m_files_checked = false; set_state(torrent_status::checking_resume_data); m_policy.recalculate_connect_candidates(); if (m_auto_managed && !is_finished()) set_queue_position((std::numeric_limits::max)()); std::vector().swap(m_resume_data); lazy_entry().swap(m_resume_entry); m_storage->async_check_fastresume(&m_resume_entry , boost::bind(&torrent::on_force_recheck , shared_from_this(), _1, _2)); } void torrent::on_force_recheck(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); state_updated(); if (ret == piece_manager::fatal_disk_error) { handle_disk_error(j); return; } if (ret == 0) { // if there are no files, just start files_checked(); } else { set_state(torrent_status::queued_for_checking); if (should_check_files()) queue_torrent_check(); } } void torrent::start_checking() { TORRENT_ASSERT(should_check_files()); set_state(torrent_status::checking_files); m_storage->async_check_files(boost::bind( &torrent::on_piece_checked , shared_from_this(), _1, _2)); } void torrent::on_piece_checked(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; state_updated(); if (ret == piece_manager::disk_check_aborted) { dequeue_torrent_check(); pause(); return; } if (ret == piece_manager::fatal_disk_error) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(file_error_alert(j.error_file, get_handle(), j.error)); } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("fatal disk error: (%d) %s", j.error.value(), j.error.message().c_str()); #endif auto_managed(false); pause(); set_error(j.error, j.error_file); // recalculate auto-managed torrents sooner // in order to start checking the next torrent m_ses.trigger_auto_manage(); return; } m_progress_ppm = size_type(j.piece) * 1000000 / torrent_file().num_pieces(); TORRENT_ASSERT(m_picker); if (j.offset >= 0 && !m_picker->have_piece(j.offset)) { we_have(j.offset, j.post_flags); remove_time_critical_piece(j.offset); } // we're not done checking yet // this handler will be called repeatedly until // we're done, or encounter a failure if (ret == piece_manager::need_full_check) return; dequeue_torrent_check(); files_checked(); } void torrent::use_interface(std::string net_interfaces) { INVARIANT_CHECK; m_net_interfaces.clear(); char* str = allocate_string_copy(net_interfaces.c_str()); char* ptr = str; while (ptr) { char* space = strchr(ptr, ','); if (space) *space++ = 0; error_code ec; address a(address::from_string(ptr, ec)); ptr = space; if (ec) continue; m_net_interfaces.push_back(tcp::endpoint(a, 0)); } free(str); } tcp::endpoint torrent::get_interface() const { if (m_net_interfaces.empty()) return tcp::endpoint(address_v4(), 0); if (m_interface_index >= m_net_interfaces.size()) m_interface_index = 0; return m_net_interfaces[m_interface_index++]; } void torrent::on_tracker_announce_disp(boost::weak_ptr p , error_code const& e) { #if defined TORRENT_ASIO_DEBUGGING complete_async("tracker::on_tracker_announce_disp"); #endif if (e) return; boost::shared_ptr t = p.lock(); if (!t) return; t->on_tracker_announce(); } void torrent::on_tracker_announce() { TORRENT_ASSERT(m_ses.is_network_thread()); m_waiting_tracker = false; if (m_abort) return; announce_with_tracker(); } void torrent::lsd_announce() { if (m_abort) return; // if the files haven't been checked yet, we're // not ready for peers if (!m_files_checked) return; if (!m_announce_to_lsd) return; if (m_torrent_file->is_valid() && (m_torrent_file->priv() || (torrent_file().is_i2p() && !settings().allow_i2p_mixed))) return; if (is_paused()) return; #ifdef TORRENT_USE_OPENSSL int port = is_ssl_torrent() ? m_ses.ssl_listen_port() : m_ses.listen_port(); #else int port = m_ses.listen_port(); #endif // announce with the local discovery service m_ses.announce_lsd(m_torrent_file->info_hash(), port , m_ses.settings().broadcast_lsd && m_lsd_seq == 0); ++m_lsd_seq; } #ifndef TORRENT_DISABLE_DHT void torrent::dht_announce() { TORRENT_ASSERT(m_ses.is_network_thread()); if (!m_ses.m_dht) return; if (!m_name) return; if (!should_announce_dht()) return; TORRENT_ASSERT(m_allow_peers); // [MF] use m_dht->announce with myself=false to update dht tracker with peers we know { policy::const_iterator i = get_policy().begin_peer(); policy::const_iterator end = get_policy().end_peer(); for (; i != end; ++i) { policy::peer const* p = *i; bool connect_recently = !p->banned && int(p->failcount) < settings().max_failcount && p->last_connected && (m_ses.session_time() - p->last_connected) < (4*3600); if( p->connectable && ( p->connection || connect_recently) ) { m_ses.m_dht->announce(name(), m_torrent_file->info_hash() , p->address(), p->port, p->seed, false, m_policy.num_peers() , boost::bind(&nop)); } } } #ifdef TORRENT_USE_OPENSSL int port = is_ssl_torrent() ? m_ses.ssl_listen_port() : m_ses.listen_port(); #else int port = m_ses.listen_port(); #endif boost::weak_ptr self(shared_from_this()); m_ses.m_dht->announce(name(), m_torrent_file->info_hash() , m_ses.external_address().external_address(address_v4()), port, is_seed(), true, m_policy.num_peers() , boost::bind(&torrent::on_dht_announce_response_disp, self, _1)); } void torrent::on_dht_announce_response_disp(boost::weak_ptr t , std::vector const& peers) { boost::shared_ptr tor = t.lock(); if (!tor) return; tor->on_dht_announce_response(peers); } void torrent::on_dht_announce_response(std::vector const& peers) { TORRENT_ASSERT(m_ses.is_network_thread()); if (peers.empty()) return; if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(dht_reply_alert( get_handle(), peers.size())); } if (torrent_file().priv() || (torrent_file().is_i2p() && !settings().allow_i2p_mixed)) return; #ifdef TORRENT_USE_OPENSSL int port = is_ssl_torrent() ? m_ses.ssl_listen_port() : m_ses.listen_port(); #else int port = m_ses.listen_port(); #endif tcp::endpoint localpeer; localpeer.address(m_ses.external_address().external_address(address_v4())); localpeer.port(port); BOOST_FOREACH(tcp::endpoint const& p, peers) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING //debug_log("on_dht_announce_response %s:%d (local=%d)", p.address().to_string().c_str(), p.port(), p==localpeer); #endif if( p != localpeer ) { m_policy.add_peer(p, peer_id(0), peer_info::dht, 0); } } /* std::for_each(peers.begin(), peers.end(), boost::bind( &policy::add_peer, boost::ref(m_policy), _1, peer_id(0) , peer_info::dht, 0)); */ do_connect_boost(); } #endif void torrent::announce_with_tracker(tracker_request::event_t e , address const& bind_interface) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_trackers.empty()) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING //debug_log("*** announce_with_tracker: no trackers"); #endif return; } if (m_abort) e = tracker_request::stopped; // if we're not announcing to trackers, only allow // stopping if (e != tracker_request::stopped && !m_announce_to_trackers) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("*** announce_with_tracker: event != stopped && !m_announce_to_trackers"); #endif return; } TORRENT_ASSERT(m_allow_peers || e == tracker_request::stopped); if (e == tracker_request::none && is_finished() && !is_seed()) e = tracker_request::paused; tracker_request req; req.apply_ip_filter = m_apply_ip_filter && m_ses.m_settings.apply_ip_filter_to_trackers; req.info_hash = m_torrent_file->info_hash(); req.pid = m_ses.get_peer_id(); req.downloaded = m_stat.total_payload_download() - m_total_failed_bytes; req.uploaded = m_stat.total_payload_upload(); req.corrupt = m_total_failed_bytes; req.left = bytes_left(); if (req.left == -1) req.left = 16*1024; #ifdef TORRENT_USE_OPENSSL // if this torrent contains an SSL certificate, make sure // any SSL tracker presents a certificate signed by it req.ssl_ctx = m_ssl_ctx.get(); #endif // exclude redundant bytes if we should if (!settings().report_true_downloaded) req.downloaded -= m_total_redundant_bytes; if (req.downloaded < 0) req.downloaded = 0; req.event = e; error_code ec; // if we are aborting. we don't want any new peers req.num_want = (req.event == tracker_request::stopped) ?0:settings().num_want; // SSL torrents use their own listen socket #ifdef TORRENT_USE_OPENSSL if (is_ssl_torrent()) req.listen_port = m_ses.ssl_listen_port(); else #endif req.listen_port = m_ses.listen_port(); req.key = m_ses.m_key; ptime now = time_now_hires(); // the tier is kept as INT_MAX until we find the first // tracker that works, then it's set to that tracker's // tier. int tier = INT_MAX; // have we sent an announce in this tier yet? bool sent_announce = false; for (int i = 0; i < int(m_trackers.size()); ++i) { announce_entry& ae = m_trackers[i]; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING char msg[1000]; snprintf(msg, sizeof(msg), "*** announce with tracker: considering \"%s\" " "[ announce_to_all_tiers: %d announce_to_all_trackers: %d" " i->tier: %d tier: %d " " is_working: %d fails: %d fail_limit: %d updating: %d" " can_announce: %d sent_announce: %d ]" , ae.url.c_str(), settings().announce_to_all_tiers , settings().announce_to_all_trackers , ae.tier, tier, ae.is_working(), ae.fails, ae.fail_limit , ae.updating, ae.can_announce(now, is_seed()), sent_announce); debug_log(msg); #endif // if trackerid is not specified for tracker use default one, probably set explicitly req.trackerid = ae.trackerid.empty() ? m_trackerid : ae.trackerid; if (settings().announce_to_all_tiers && !settings().announce_to_all_trackers && sent_announce && ae.tier <= tier && tier != INT_MAX) continue; if (ae.tier > tier && sent_announce && !settings().announce_to_all_tiers) break; if (ae.is_working()) { tier = ae.tier; sent_announce = false; } if (!ae.can_announce(now, is_seed())) { // this counts if (ae.is_working()) sent_announce = true; continue; } req.url = ae.url; req.event = e; if (req.event == tracker_request::none) { if (!ae.start_sent) req.event = tracker_request::started; else if (!ae.complete_sent && is_seed()) req.event = tracker_request::completed; } if (!is_any(bind_interface)) req.bind_ip = bind_interface; else req.bind_ip = m_ses.m_listen_interface.address(); if (settings().force_proxy) { // in force_proxy mode we don't talk directly to trackers // we only allow trackers if there is a proxy and issue // a warning if there isn't one std::string protocol = req.url.substr(0, req.url.find(':')); int proxy_type = m_ses.m_proxy.type; // http can run over any proxy, so as long as one is used // it's OK. If no proxy is configured, skip this tracker if ((protocol == "http" || protocol == "https") && proxy_type == proxy_settings::none) { ae.next_announce = now + minutes(10); if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( anonymous_mode_alert(get_handle() , anonymous_mode_alert::tracker_not_anonymous, req.url)); } continue; } // for UDP, only socks5 and i2p proxies will work. // if we're not using one of those proxues with a UDP // tracker, skip it if (protocol == "udp" && proxy_type != proxy_settings::socks5 && proxy_type != proxy_settings::socks5_pw && proxy_type != proxy_settings::i2p_proxy) { ae.next_announce = now + minutes(10); if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( anonymous_mode_alert(get_handle() , anonymous_mode_alert::tracker_not_anonymous, req.url)); } continue; } } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("==> TRACKER REQUEST \"%s\" event: %s abort: %d" , req.url.c_str() , (req.event==tracker_request::stopped?"stopped" :req.event==tracker_request::started?"started":"") , m_abort); if (m_abort) { boost::shared_ptr tl(new aux::tracker_logger(m_ses)); m_ses.m_tracker_manager.queue_request(m_ses.m_io_service, m_ses.m_half_open, req , tracker_login(), tl); } else #endif { m_ses.m_tracker_manager.queue_request(m_ses.m_io_service, m_ses.m_half_open, req , tracker_login() , shared_from_this()); } ae.updating = true; ae.next_announce = now + seconds(20); ae.min_announce = now + seconds(10); if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( tracker_announce_alert(get_handle(), req.url, req.event)); } sent_announce = true; if (ae.is_working() && !settings().announce_to_all_trackers && !settings().announce_to_all_tiers) break; } update_tracker_timer(now); } void torrent::scrape_tracker() { TORRENT_ASSERT(m_ses.is_network_thread()); m_last_scrape = 0; if (m_trackers.empty()) return; int i = m_last_working_tracker; if (i == -1) i = 0; tracker_request req; req.apply_ip_filter = m_apply_ip_filter && m_ses.m_settings.apply_ip_filter_to_trackers; req.info_hash = m_torrent_file->info_hash(); req.kind = tracker_request::scrape_request; req.url = m_trackers[i].url; req.bind_ip = m_ses.m_listen_interface.address(); m_ses.m_tracker_manager.queue_request(m_ses.m_io_service, m_ses.m_half_open, req , tracker_login(), shared_from_this()); } void torrent::tracker_warning(tracker_request const& req, std::string const& msg) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(tracker_warning_alert(get_handle(), req.url, msg)); } void torrent::tracker_scrape_response(tracker_request const& req , int complete, int incomplete, int downloaded, int downloaders) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(req.kind == tracker_request::scrape_request); announce_entry* ae = find_tracker(req); if (ae) { if (incomplete >= 0) ae->scrape_incomplete = incomplete; if (complete >= 0) ae->scrape_complete = complete; if (downloaded >= 0) ae->scrape_downloaded = downloaded; update_scrape_state(); } if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(scrape_reply_alert( get_handle(), incomplete, complete, req.url)); } } void torrent::update_scrape_state() { // loop over all trackers and find the largest numbers for each scrape field // then update the torrent-wide understanding of number of downloaders and seeds int complete = -1; int incomplete = -1; int downloaded = -1; for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { complete = (std::max)(i->scrape_complete, complete); incomplete = (std::max)(i->scrape_incomplete, incomplete); downloaded = (std::max)(i->scrape_downloaded, downloaded); } if ((complete >= 0 && m_complete != complete) || (incomplete >= 0 && m_incomplete != incomplete) || (downloaded >= 0 && m_downloaded != downloaded)) state_updated(); m_complete = complete; m_incomplete = incomplete; m_downloaded = downloaded; } void torrent::tracker_response( tracker_request const& r , address const& tracker_ip // this is the IP we connected to , std::list
const& tracker_ips // these are all the IPs it resolved to , std::vector& peer_list , int interval , int min_interval , int complete , int incomplete , int downloaded , address const& external_ip , const std::string& trackerid) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(r.kind == tracker_request::announce_request); if (external_ip != address() && !tracker_ips.empty()) m_ses.set_external_address(external_ip, aux::session_impl::source_tracker , *tracker_ips.begin()); ptime now = time_now(); if (interval < settings().min_announce_interval) interval = settings().min_announce_interval; announce_entry* ae = find_tracker(r); if (ae) { if (incomplete >= 0) ae->scrape_incomplete = incomplete; if (complete >= 0) ae->scrape_complete = complete; if (downloaded >= 0) ae->scrape_downloaded = downloaded; if (!ae->start_sent && r.event == tracker_request::started) ae->start_sent = true; if (!ae->complete_sent && r.event == tracker_request::completed) ae->complete_sent = true; ae->verified = true; ae->updating = false; ae->fails = 0; ae->next_announce = now + seconds(interval); ae->min_announce = now + seconds(min_interval); int tracker_index = ae - &m_trackers[0]; m_last_working_tracker = prioritize_tracker(tracker_index); if ((!trackerid.empty()) && (ae->trackerid != trackerid)) { ae->trackerid = trackerid; if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(trackerid_alert(get_handle(), r.url, trackerid)); } update_scrape_state(); } update_tracker_timer(now); if (complete >= 0 && incomplete >= 0) m_last_scrape = 0; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("TRACKER RESPONSE\n" "interval: %d\n" "external ip: %s\n" "we connected to: %s\n" "peers:" , interval , print_address(external_ip).c_str() , print_address(tracker_ip).c_str()); for (std::vector::const_iterator i = peer_list.begin(); i != peer_list.end(); ++i) { debug_log(" %16s %5d %s %s", i->ip.c_str(), i->port , i->pid.is_all_zeros()?"":to_hex(i->pid.to_string()).c_str() , identify_client(i->pid).c_str()); } #endif // for each of the peers we got from the tracker for (std::vector::iterator i = peer_list.begin(); i != peer_list.end(); ++i) { // don't make connections to ourself if (i->pid == m_ses.get_peer_id()) continue; error_code ec; tcp::endpoint a(address::from_string(i->ip, ec), i->port); if (ec) { // assume this is because we got a hostname instead of // an ip address from the tracker #if TORRENT_USE_I2P char const* top_domain = strrchr(i->ip.c_str(), '.'); if (top_domain && strcmp(top_domain, ".i2p") == 0 && m_ses.m_i2p_conn.is_open()) { // this is an i2p name, we need to use the sam connection // to do the name lookup /* m_ses.m_i2p_conn.async_name_lookup(i->ip.c_str() , boost::bind(&torrent::on_i2p_resolve , shared_from_this(), _1, _2)); */ // it seems like you're not supposed to do a name lookup // on the peers returned from the tracker, but just strip // the .i2p and use it as a destination i->ip.resize(i->ip.size() - 4); m_policy.add_i2p_peer(i->ip.c_str(), peer_info::tracker, 0); } else #endif { #if defined TORRENT_ASIO_DEBUGGING add_outstanding_async("torrent::on_peer_name_lookup"); #endif tcp::resolver::query q(i->ip, to_string(i->port).elems); m_host_resolver.async_resolve(q, boost::bind(&torrent::on_peer_name_lookup, shared_from_this(), _1, _2, i->pid)); } } else { // ignore local addresses from the tracker (unless the tracker is local too) // there are 2 reasons to allow this: // 1. retrackers are popular in russia, where an ISP runs a tracker within // the AS (but not on the local network) giving out peers only from the // local network // 2. it might make sense to have a tracker extension in the future where // trackers records a peer's internal and external IP, and match up // peers on the same local network // if (is_local(a.address()) && !is_local(tracker_ip)) continue; m_policy.add_peer(a, i->pid, peer_info::tracker, 0); } } if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(tracker_reply_alert( get_handle(), peer_list.size(), r.url)); } m_got_tracker_response = true; // we're listening on an interface type that was not used // when talking to the tracker. If there is a matching interface // type in the tracker IP list, make another tracker request // using that interface // in order to avoid triggering this case over and over, don't // do it if the bind IP for the tracker request that just completed // matches one of the listen interfaces, since that means this // announce was the second one // don't connect twice just to tell it we're stopping if (((!is_any(m_ses.m_ipv6_interface.address()) && tracker_ip.is_v4()) || (!is_any(m_ses.m_ipv4_interface.address()) && tracker_ip.is_v6())) && r.bind_ip != m_ses.m_ipv4_interface.address() && r.bind_ip != m_ses.m_ipv6_interface.address() && r.event != tracker_request::stopped) { std::list
::const_iterator i = std::find_if(tracker_ips.begin() , tracker_ips.end(), boost::bind(&address::is_v4, _1) != tracker_ip.is_v4()); if (i != tracker_ips.end()) { // the tracker did resolve to a different type of address, so announce // to that as well // tell the tracker to bind to the opposite protocol type address bind_interface = tracker_ip.is_v4() ?m_ses.m_ipv6_interface.address() :m_ses.m_ipv4_interface.address(); announce_with_tracker(r.event, bind_interface); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("announce again using %s as the bind interface" , print_address(bind_interface).c_str()); #endif } } do_connect_boost(); state_updated(); } void torrent::do_connect_boost() { if (!m_need_connect_boost) return; m_need_connect_boost = false; // this is the first tracker response for this torrent // instead of waiting one second for session_impl::on_tick() // to be called, connect to a few peers immediately int conns = (std::min)((std::min)((std::min)(m_ses.m_settings.torrent_connect_boost , m_ses.m_settings.connections_limit - m_ses.num_connections()) , m_ses.m_half_open.free_slots()) , m_ses.m_boost_connections - m_ses.m_settings.connection_speed); while (want_more_peers() && conns > 0) { if (!m_policy.connect_one_peer(m_ses.session_time())) break; // increase m_ses.m_boost_connections for each connection // attempt. This will be deducted from the connect speed // the next time session_impl::on_tick() is triggered --conns; ++m_ses.m_boost_connections; } if (want_more_peers()) m_ses.prioritize_connections(shared_from_this()); } ptime torrent::next_announce() const { return m_waiting_tracker?m_tracker_timer.expires_at():min_time(); } void torrent::force_tracker_request() { force_tracker_request(time_now_hires()); } void torrent::force_tracker_request(ptime t) { if (is_paused()) return; for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) i->next_announce = (std::max)(t, i->min_announce) + seconds(1); update_tracker_timer(time_now_hires()); } void torrent::set_tracker_login( std::string const& name , std::string const& pw) { m_username = name; m_password = pw; } #if TORRENT_USE_I2P void torrent::on_i2p_resolve(error_code const& ec, char const* dest) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; #if defined TORRENT_LOGGING if (ec) debug_log("i2p_resolve error: %s", ec.message().c_str()); #endif if (ec || m_ses.is_aborted()) return; m_policy.add_i2p_peer(dest, peer_info::tracker, 0); } #endif void torrent::on_peer_name_lookup(error_code const& e, tcp::resolver::iterator host , peer_id pid) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; #if defined TORRENT_ASIO_DEBUGGING complete_async("torrent::on_peer_name_lookup"); #endif #if defined TORRENT_LOGGING if (e) debug_log("peer name lookup error: %s", e.message().c_str()); #endif if (e || host == tcp::resolver::iterator() || m_ses.is_aborted()) return; if (m_apply_ip_filter && m_ses.m_ip_filter.access(host->endpoint().address()) & ip_filter::blocked) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING error_code ec; debug_log("blocked ip from tracker: %s", host->endpoint().address().to_string(ec).c_str()); #endif if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(peer_blocked_alert(get_handle(), host->endpoint().address())); return; } m_policy.add_peer(*host, pid, peer_info::tracker, 0); } size_type torrent::bytes_left() const { // if we don't have the metadata yet, we // cannot tell how big the torrent is. if (!valid_metadata()) return -1; return m_torrent_file->total_size() - quantized_bytes_done(); } size_type torrent::quantized_bytes_done() const { // INVARIANT_CHECK; if (!valid_metadata()) return 0; if (m_torrent_file->num_pieces() == 0) return 0; if (is_seed()) return m_torrent_file->total_size(); const int last_piece = m_torrent_file->num_pieces() - 1; size_type total_done = size_type(num_have()) * m_torrent_file->piece_length(); // if we have the last piece, we have to correct // the amount we have, since the first calculation // assumed all pieces were of equal size if (m_picker->have_piece(last_piece)) { int corr = m_torrent_file->piece_size(last_piece) - m_torrent_file->piece_length(); total_done += corr; } return total_done; } // returns the number of bytes we are interested // in for the given block. This returns block_size() // for all blocks except the last one (if it's smaller // than block_size()) and blocks that overlap a padding // file int torrent::block_bytes_wanted(piece_block const& p) const { file_storage const& fs = m_torrent_file->files(); int piece_size = m_torrent_file->piece_size(p.piece_index); int offset = p.block_index * block_size(); if (m_padding == 0) return (std::min)(piece_size - offset, int(block_size())); std::vector files = fs.map_block( p.piece_index, offset, (std::min)(piece_size - offset, int(block_size()))); int ret = 0; for (std::vector::iterator i = files.begin() , end(files.end()); i != end; ++i) { if (fs.pad_file_at(i->file_index)) continue; ret += i->size; } TORRENT_ASSERT(ret <= (std::min)(piece_size - offset, int(block_size()))); return ret; } // fills in total_wanted, total_wanted_done and total_done void torrent::bytes_done(torrent_status& st, bool accurate) const { INVARIANT_CHECK; st.total_done = 0; st.total_wanted_done = 0; st.total_wanted = m_torrent_file->total_size(); TORRENT_ASSERT(st.total_wanted >= m_padding); TORRENT_ASSERT(st.total_wanted >= 0); if (!valid_metadata() || m_torrent_file->num_pieces() == 0) return; TORRENT_ASSERT(st.total_wanted >= size_type(m_torrent_file->piece_length()) * (m_torrent_file->num_pieces() - 1)); const int last_piece = m_torrent_file->num_pieces() - 1; const int piece_size = m_torrent_file->piece_length(); if (is_seed()) { st.total_done = m_torrent_file->total_size() - m_padding; st.total_wanted_done = st.total_done; st.total_wanted = st.total_done; return; } TORRENT_ASSERT(num_have() >= m_picker->num_have_filtered()); st.total_wanted_done = size_type(num_have() - m_picker->num_have_filtered()) * piece_size; TORRENT_ASSERT(st.total_wanted_done >= 0); st.total_done = size_type(num_have()) * piece_size; TORRENT_ASSERT(num_have() < m_torrent_file->num_pieces()); int num_filtered_pieces = m_picker->num_filtered() + m_picker->num_have_filtered(); int last_piece_index = m_torrent_file->num_pieces() - 1; if (m_picker->piece_priority(last_piece_index) == 0) { st.total_wanted -= m_torrent_file->piece_size(last_piece_index); --num_filtered_pieces; } st.total_wanted -= size_type(num_filtered_pieces) * piece_size; // if we have the last piece, we have to correct // the amount we have, since the first calculation // assumed all pieces were of equal size if (m_picker->have_piece(last_piece)) { TORRENT_ASSERT(st.total_done >= piece_size); int corr = m_torrent_file->piece_size(last_piece) - piece_size; TORRENT_ASSERT(corr <= 0); TORRENT_ASSERT(corr > -piece_size); st.total_done += corr; if (m_picker->piece_priority(last_piece) != 0) { TORRENT_ASSERT(st.total_wanted_done >= piece_size); st.total_wanted_done += corr; } } TORRENT_ASSERT(st.total_wanted >= st.total_wanted_done); // subtract padding files if (m_padding > 0 && accurate) { file_storage const& files = m_torrent_file->files(); int fileno = 0; for (file_storage::iterator i = files.begin() , end(files.end()); i != end; ++i, ++fileno) { if (!i->pad_file) continue; peer_request p = files.map_file(fileno, 0, i->size); for (int j = p.piece; p.length > 0; ++j) { int deduction = (std::min)(p.length, piece_size - p.start); bool done = m_picker->have_piece(j); bool wanted = m_picker->piece_priority(j) > 0; if (done) st.total_done -= deduction; if (wanted) st.total_wanted -= deduction; if (wanted && done) st.total_wanted_done -= deduction; TORRENT_ASSERT(st.total_done >= 0); TORRENT_ASSERT(st.total_wanted >= 0); TORRENT_ASSERT(st.total_wanted_done >= 0); p.length -= piece_size - p.start; p.start = 0; ++p.piece; } } } TORRENT_ASSERT(st.total_done <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_wanted_done <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_wanted_done >= 0); TORRENT_ASSERT(st.total_done >= st.total_wanted_done); // this is expensive, we might not want to do it all the time if (!accurate) return; const std::vector& dl_queue = m_picker->get_download_queue(); const int blocks_per_piece = (piece_size + block_size() - 1) / block_size(); // look at all unfinished pieces and add the completed // blocks to our 'done' counter for (std::vector::const_iterator i = dl_queue.begin(); i != dl_queue.end(); ++i) { int corr = 0; int index = i->index; // completed pieces are already accounted for if (m_picker->have_piece(index)) continue; TORRENT_ASSERT(i->finished <= m_picker->blocks_in_piece(index)); #ifdef TORRENT_DEBUG for (std::vector::const_iterator j = boost::next(i); j != dl_queue.end(); ++j) { TORRENT_ASSERT(j->index != index); } #endif for (int j = 0; j < blocks_per_piece; ++j) { #ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS TORRENT_ASSERT(m_picker->is_finished(piece_block(index, j)) == (i->info[j].state == piece_picker::block_info::state_finished)); #endif if (i->info[j].state == piece_picker::block_info::state_finished) { corr += block_bytes_wanted(piece_block(index, j)); } TORRENT_ASSERT(corr >= 0); TORRENT_ASSERT(index != last_piece || j < m_picker->blocks_in_last_piece() || i->info[j].state != piece_picker::block_info::state_finished); } st.total_done += corr; if (m_picker->piece_priority(index) > 0) st.total_wanted_done += corr; } TORRENT_ASSERT(st.total_wanted <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_done <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_wanted_done <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_wanted_done >= 0); TORRENT_ASSERT(st.total_done >= st.total_wanted_done); std::map downloading_piece; for (const_peer_iterator i = begin(); i != end(); ++i) { peer_connection* pc = *i; boost::optional p = pc->downloading_piece_progress(); if (!p) continue; if (m_picker->have_piece(p->piece_index)) continue; piece_block block(p->piece_index, p->block_index); if (m_picker->is_finished(block)) continue; std::map::iterator dp = downloading_piece.find(block); if (dp != downloading_piece.end()) { if (dp->second < p->bytes_downloaded) dp->second = p->bytes_downloaded; } else { downloading_piece[block] = p->bytes_downloaded; } #ifdef TORRENT_DEBUG TORRENT_ASSERT(p->bytes_downloaded <= p->full_block_bytes); TORRENT_ASSERT(p->full_block_bytes == to_req(piece_block( p->piece_index, p->block_index)).length); #endif } for (std::map::iterator i = downloading_piece.begin(); i != downloading_piece.end(); ++i) { int done = (std::min)(block_bytes_wanted(i->first), i->second); st.total_done += done; if (m_picker->piece_priority(i->first.piece_index) != 0) st.total_wanted_done += done; } TORRENT_ASSERT(st.total_done <= m_torrent_file->total_size() - m_padding); TORRENT_ASSERT(st.total_wanted_done <= m_torrent_file->total_size() - m_padding); #ifdef TORRENT_DEBUG if (st.total_done >= m_torrent_file->total_size()) { // Thist happens when a piece has been downloaded completely // but not yet verified against the hash fprintf(stderr, "num_have: %d\nunfinished:\n", num_have()); for (std::vector::const_iterator i = dl_queue.begin(); i != dl_queue.end(); ++i) { fprintf(stderr, " %d ", i->index); for (int j = 0; j < blocks_per_piece; ++j) { char const* state = i->info[j].state == piece_picker::block_info::state_finished ? "1" : "0"; fputs(state, stderr); } fputs("\n", stderr); } fputs("downloading pieces:\n", stderr); for (std::map::iterator i = downloading_piece.begin(); i != downloading_piece.end(); ++i) { fprintf(stderr, " %d:%d %d\n", int(i->first.piece_index), int(i->first.block_index), i->second); } } TORRENT_ASSERT(st.total_done <= m_torrent_file->total_size()); TORRENT_ASSERT(st.total_wanted_done <= m_torrent_file->total_size()); #endif TORRENT_ASSERT(st.total_done >= st.total_wanted_done); } // passed_hash_check // 0: success, piece passed check // -1: disk failure // -2: piece failed check void torrent::piece_finished(int index, int passed_hash_check, boost::uint32_t post_flags, int piece_size) { TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("*** PIECE_FINISHED [ p: %d | chk: %s | size: %d ]" , index, ((passed_hash_check == 0) ?"passed":passed_hash_check == -1 ?"disk failed":"failed") , piece_size); #endif TORRENT_ASSERT(valid_metadata()); // it's possible to get here if the last piece was downloaded // from peers and inserted with add_piece at the same time. // if we're a seed, we won't have a piece picker, and can't continue if (is_seed()) return; TORRENT_ASSERT(!m_picker->have_piece(index)); state_updated(); // even though the piece passed the hash-check // it might still have failed being written to disk // if so, piece_picker::write_failed() has been // called, and the piece is no longer finished. // in this case, we have to ignore the fact that // it passed the check if (!m_picker->is_piece_finished(index)) return; if (passed_hash_check == 0) { // the following call may cause picker to become invalid // in case we just became a seed piece_passed(index, post_flags); // if we're in seed mode, we just acquired this piece // mark it as verified if (m_seed_mode) verified(index); } else if (passed_hash_check == -2) { // piece_failed() will restore the piece piece_failed(index); } else { TORRENT_ASSERT(passed_hash_check == -1); m_picker->restore_piece(index); restore_piece_state(index); } } void torrent::update_sparse_piece_prio(int i, int start, int end) { TORRENT_ASSERT(m_picker); if (m_picker->have_piece(i) || m_picker->piece_priority(i) == 0) return; bool have_before = i == 0 || m_picker->have_piece(i - 1); bool have_after = i == end - 1 || m_picker->have_piece(i + 1); if (have_after && have_before) m_picker->set_piece_priority(i, 7); else if (have_after || have_before) m_picker->set_piece_priority(i, 6); } void on_disk_read_recheck_piece_complete(int ret, disk_io_job const& j, peer_request r) { // [MF] FIXME: implement cond_wakeup here so that recheck_pieces would wait } void torrent::recheck_pieces(uint32_t piece_flags) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(m_picker); for( int i = 0; i <= last_have(); i++) { if( m_picker->have_piece(i) && m_picker->post_flags(i) == piece_flags ) { peer_request r; r.piece = i; r.start = 0; r.length = torrent_file().piece_size(i); filesystem().async_read_and_hash(r, boost::bind(&on_disk_read_recheck_piece_complete , _1, _2, r), 1); } } } void torrent::we_have(int index, boost::uint32_t post_flags) { TORRENT_ASSERT(m_ses.is_network_thread()); // update m_file_progress TORRENT_ASSERT(m_picker); TORRENT_ASSERT(!have_piece(index)); TORRENT_ASSERT(!m_picker->have_piece(index)); remove_time_critical_piece(index, true); m_picker->we_have(index, post_flags); } void torrent::we_dont_have(int index) { TORRENT_ASSERT(m_ses.is_network_thread()); // update m_file_progress TORRENT_ASSERT(m_picker); m_picker->we_dont_have(index); } void torrent::piece_passed(int index, boost::uint32_t post_flags) { // INVARIANT_CHECK; TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); #ifdef TORRENT_DEBUG // make sure all blocks were successfully written before we // declare the piece as "we have". piece_picker::downloading_piece dp; m_picker->piece_info(index, dp); int blocks_in_piece = m_picker->blocks_in_piece(index); TORRENT_ASSERT(dp.finished == blocks_in_piece); TORRENT_ASSERT(dp.writing == 0); TORRENT_ASSERT(dp.requested == 0); TORRENT_ASSERT(dp.index == index); #endif if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(piece_finished_alert(get_handle() , index)); } m_need_save_resume_data = true; state_updated(); remove_time_critical_piece(index, true); bool was_finished = m_picker->num_filtered() + num_have() == torrent_file().num_pieces(); std::vector downloaders; m_picker->get_downloaders(downloaders, index); // increase the trust point of all peers that sent // parts of this piece. std::set peers; // these policy::peer pointers are owned by m_policy and they may be // invalidated if a peer disconnects. We cannot keep them across any // significant operations, but we should use them right away // ignore NULL pointers std::remove_copy(downloaders.begin(), downloaders.end() , std::inserter(peers, peers.begin()), (policy::peer*)0); for (std::set::iterator i = peers.begin() , end(peers.end()); i != end; ++i) { policy::peer* p = static_cast(*i); TORRENT_ASSERT(p != 0); if (p == 0) continue; TORRENT_ASSERT(p->in_use); p->on_parole = false; int trust_points = p->trust_points; ++trust_points; if (trust_points > 8) trust_points = 8; p->trust_points = trust_points; if (p->connection) { TORRENT_ASSERT(p->connection->m_in_use == 1337); p->connection->received_valid_data(index); } } // announcing a piece may invalidate the policy::peer pointers // so we can't use them anymore downloaders.clear(); peers.clear(); we_have(index, post_flags); for (peer_iterator i = m_connections.begin(); i != m_connections.end();) { intrusive_ptr p = *i; ++i; p->announce_piece(index); } if (settings().max_sparse_regions > 0 && m_picker->sparse_regions() > settings().max_sparse_regions) { // we have too many sparse regions. Prioritize pieces // that won't introduce new sparse regions // prioritize pieces that will reduce the number of sparse // regions even higher int start = m_picker->cursor(); int end = m_picker->reverse_cursor(); if (index > start) update_sparse_piece_prio(index - 1, start, end); if (index < end - 1) update_sparse_piece_prio(index + 1, start, end); } #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->on_piece_pass(index); } TORRENT_CATCH (std::exception&) {} } #endif // since this piece just passed, we might have // become uninterested in some peers where this // was the last piece we were interested in for (peer_iterator i = m_connections.begin(); i != m_connections.end();) { peer_connection* p = *i; // update_interest may disconnect the peer and // invalidate the iterator ++i; // if we're not interested already, no need to check if (!p->is_interesting()) continue; // if the peer doesn't have the piece we just got, it // wouldn't affect our interest if (!p->has_piece(index)) continue; p->update_interest(); } if (!was_finished && is_finished()) { // torrent finished // i.e. all the pieces we're interested in have // been downloaded. Release the files (they will open // in read only mode if needed) finished(); // if we just became a seed, picker is now invalid, since it // is deallocated by the torrent once it starts seeding } m_last_download = 0; if (m_share_mode) recalc_share_mode(); } void torrent::piece_failed(int index) { // if the last piece fails the peer connection will still // think that it has received all of it until this function // resets the download queue. So, we cannot do the // invariant check here since it assumes: // (total_done == m_torrent_file->total_size()) => is_seed() INVARIANT_CHECK; TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(m_storage); TORRENT_ASSERT(m_storage->refcount() > 0); TORRENT_ASSERT(m_picker.get()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(hash_failed_alert(get_handle(), index)); // increase the total amount of failed bytes add_failed_bytes(m_torrent_file->piece_size(index)); std::vector downloaders; m_picker->get_downloaders(downloaders, index); // decrease the trust point of all peers that sent // parts of this piece. // first, build a set of all peers that participated std::set peers; std::copy(downloaders.begin(), downloaders.end(), std::inserter(peers, peers.begin())); #ifdef TORRENT_DEBUG for (std::vector::iterator i = downloaders.begin() , end(downloaders.end()); i != end; ++i) { policy::peer* p = (policy::peer*)*i; if (p && p->connection) { p->connection->piece_failed = true; } } #endif #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->on_piece_failed(index); } TORRENT_CATCH (std::exception&) {} } #endif // did we receive this piece from a single peer? bool single_peer = peers.size() == 1; for (std::set::iterator i = peers.begin() , end(peers.end()); i != end; ++i) { policy::peer* p = static_cast(*i); if (p == 0) continue; TORRENT_ASSERT(p->in_use); bool allow_disconnect = true; if (p->connection) { TORRENT_ASSERT(p->connection->m_in_use == 1337); // the peer implementation can ask not to be disconnected. // this is used for web seeds for instance, to instead of // disconnecting, mark the file as not being haved. allow_disconnect = p->connection->received_invalid_data(index, single_peer); } if (m_ses.settings().use_parole_mode) p->on_parole = true; int hashfails = p->hashfails; int trust_points = p->trust_points; // we decrease more than we increase, to keep the // allowed failed/passed ratio low. trust_points -= 2; ++hashfails; if (trust_points < -7) trust_points = -7; p->trust_points = trust_points; if (hashfails > 255) hashfails = 255; p->hashfails = hashfails; // either, we have received too many failed hashes // or this was the only peer that sent us this piece. // if we have failed more than 3 pieces from this peer, // don't trust it regardless. if (p->trust_points <= -7 || (single_peer && allow_disconnect)) { // we don't trust this peer anymore // ban it. if (m_ses.m_alerts.should_post()) { peer_id pid(0); if (p->connection) pid = p->connection->pid(); m_ses.m_alerts.post_alert(peer_ban_alert( get_handle(), p->ip(), pid)); } // mark the peer as banned m_policy.ban_peer(p); #ifdef TORRENT_STATS ++m_ses.m_banned_for_hash_failure; #endif if (p->connection) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("*** BANNING PEER: \"%s\" Too many corrupt pieces" , print_endpoint(p->ip()).c_str()); #endif #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->connection->peer_log("*** BANNING PEER: Too many corrupt pieces"); #endif p->connection->disconnect(errors::too_many_corrupt_pieces); } } } // we have to let the piece_picker know that // this piece failed the check as it can restore it // and mark it as being interesting for download m_picker->restore_piece(index); // we might still have outstanding requests to this // piece that hasn't been received yet. If this is the // case, we need to re-open the piece and mark any // blocks we're still waiting for as requested restore_piece_state(index); TORRENT_ASSERT(m_storage); TORRENT_ASSERT(m_picker->have_piece(index) == false); #ifdef TORRENT_DEBUG for (std::vector::iterator i = downloaders.begin() , end(downloaders.end()); i != end; ++i) { policy::peer* p = (policy::peer*)*i; if (p && p->connection) { p->connection->piece_failed = false; } } #endif } void torrent::restore_piece_state(int index) { TORRENT_ASSERT(has_picker()); for (peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { peer_connection* p = *i; std::vector const& dq = p->download_queue(); std::vector const& rq = p->request_queue(); for (std::vector::const_iterator k = dq.begin() , end(dq.end()); k != end; ++k) { if (k->timed_out || k->not_wanted) continue; if (int(k->block.piece_index) != index) continue; m_picker->mark_as_downloading(k->block, p->peer_info_struct() , (piece_picker::piece_state_t)p->peer_speed()); } for (std::vector::const_iterator k = rq.begin() , end(rq.end()); k != end; ++k) { if (int(k->block.piece_index) != index) continue; m_picker->mark_as_downloading(k->block, p->peer_info_struct() , (piece_picker::piece_state_t)p->peer_speed()); } } } void torrent::abort() { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_abort) return; m_abort = true; update_guage(); // if the torrent is paused, it doesn't need // to announce with even=stopped again. if (!is_paused()) { stop_announcing(); } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING log_to_all_peers("ABORTING TORRENT"); #endif // disconnect all peers and close all // files belonging to the torrents disconnect_all(errors::torrent_aborted); // post a message to the main thread to destruct // the torrent object from there if (m_owning_storage.get()) { m_storage->abort_disk_io(); m_storage->async_release_files( boost::bind(&torrent::on_cache_flushed, shared_from_this(), _1, _2)); } else { if (alerts().should_post()) alerts().post_alert(cache_flushed_alert(get_handle())); } dequeue_torrent_check(); if (m_state == torrent_status::checking_files) set_state(torrent_status::queued_for_checking); m_owning_storage = 0; m_host_resolver.cancel(); } void torrent::super_seeding(bool on) { if (on == m_super_seeding) return; m_super_seeding = on; if (m_super_seeding) return; // disable super seeding for all peers for (peer_iterator i = begin(); i != end(); ++i) { (*i)->superseed_piece(-1, -1); } } void torrent::set_following(bool on) { m_following = on; } int torrent::get_piece_to_super_seed(bitfield const& bits) { // return a piece with low availability that is not in // the bitfield and that is not currently being super // seeded by any peer TORRENT_ASSERT(m_super_seeding); // do a linear search from the first piece int min_availability = 9999; std::vector avail_vec; for (int i = 0; i < m_torrent_file->num_pieces(); ++i) { if (bits[i]) continue; int availability = 0; for (const_peer_iterator j = begin(); j != end(); ++j) { if ((*j)->super_seeded_piece(i)) { // avoid superseeding the same piece to more than one // peer if we can avoid it. Do this by artificially // increase the availability availability = 999; break; } if ((*j)->has_piece(i)) ++availability; } if (availability > min_availability) continue; if (availability == min_availability) { avail_vec.push_back(i); continue; } TORRENT_ASSERT(availability < min_availability); min_availability = availability; avail_vec.clear(); avail_vec.push_back(i); } if (min_availability > 1) { // if the minimum availability is 2 or more, // we shouldn't be super seeding any more super_seeding(false); return -1; } return avail_vec[random() % avail_vec.size()]; } void torrent::on_files_deleted(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (ret != 0) { alerts().post_alert(torrent_delete_failed_alert(get_handle(), j.error)); } else { alerts().post_alert(torrent_deleted_alert(get_handle(), m_torrent_file->info_hash())); } } void torrent::on_files_released(int ret, disk_io_job const& j) { } void torrent::on_save_resume_data(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (!j.resume_data) { alerts().post_alert(save_resume_data_failed_alert(get_handle(), j.error)); } else { m_need_save_resume_data = false; m_last_saved_resume = time(0); write_resume_data(*j.resume_data); alerts().post_alert(save_resume_data_alert(j.resume_data , get_handle())); state_updated(); } } void torrent::on_file_renamed(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (ret == 0) { if (alerts().should_post()) alerts().post_alert(file_renamed_alert(get_handle(), j.str, j.piece)); m_torrent_file->rename_file(j.piece, j.str); } else { if (alerts().should_post()) alerts().post_alert(file_rename_failed_alert(get_handle() , j.piece, j.error)); } } void torrent::on_torrent_paused(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (alerts().should_post()) alerts().post_alert(torrent_paused_alert(get_handle())); } std::string torrent::tracker_login() const { if (m_username.empty() && m_password.empty()) return ""; return m_username + ":" + m_password; } void torrent::set_piece_deadline(int piece, int t, int flags) { INVARIANT_CHECK; if (m_abort) { // failed if (flags & torrent_handle::alert_when_available) { m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), piece, error_code(boost::system::errc::operation_canceled, get_system_category()))); } return; } ptime deadline = time_now() + milliseconds(t); if (is_seed() || m_picker->have_piece(piece)) { if (flags & torrent_handle::alert_when_available) read_piece(piece); return; } for (std::deque::iterator i = m_time_critical_pieces.begin() , end(m_time_critical_pieces.end()); i != end; ++i) { if (i->piece != piece) continue; i->deadline = deadline; i->flags = flags; // resort i since deadline might have changed while (boost::next(i) != m_time_critical_pieces.end() && i->deadline > boost::next(i)->deadline) { std::iter_swap(i, boost::next(i)); ++i; } while (i != m_time_critical_pieces.begin() && i->deadline < boost::prior(i)->deadline) { std::iter_swap(i, boost::prior(i)); --i; } // just in case this piece had priority 0 if (m_picker->piece_priority(piece) == 0) m_picker->set_piece_priority(piece, 1); return; } time_critical_piece p; p.first_requested = min_time(); p.last_requested = min_time(); p.flags = flags; p.deadline = deadline; p.peers = 0; p.piece = piece; std::deque::iterator i = std::upper_bound(m_time_critical_pieces.begin() , m_time_critical_pieces.end(), p); m_time_critical_pieces.insert(i, p); // just in case this piece had priority 0 if (m_picker->piece_priority(piece) == 0) m_picker->set_piece_priority(piece, 1); piece_picker::downloading_piece pi; m_picker->piece_info(piece, pi); if (pi.requested == 0) return; // this means we have outstanding requests (or queued // up requests that haven't been sent yet). Promote them // to deadline pieces immediately std::vector downloaders; m_picker->get_downloaders(downloaders, piece); int block = 0; for (std::vector::iterator i = downloaders.begin() , end(downloaders.end()); i != end; ++i, ++block) { policy::peer* p = (policy::peer*)*i; if (p == 0 || p->connection == 0) continue; p->connection->make_time_critical(piece_block(piece, block)); } } void torrent::reset_piece_deadline(int piece) { remove_time_critical_piece(piece); } void torrent::remove_time_critical_piece(int piece, bool finished) { for (std::deque::iterator i = m_time_critical_pieces.begin() , end(m_time_critical_pieces.end()); i != end; ++i) { if (i->piece != piece) continue; if (finished) { if (i->flags & torrent_handle::alert_when_available) { read_piece(i->piece); } // if first_requested is min_time(), it wasn't requested as a critical piece // and we shouldn't adjust any average download times if (i->first_requested != min_time()) { // update the average download time and average // download time deviation int dl_time = total_milliseconds(time_now() - i->first_requested); if (m_average_piece_time == 0) { m_average_piece_time = dl_time; } else { int diff = abs(int(dl_time - m_average_piece_time)); if (m_piece_time_deviation == 0) m_piece_time_deviation = diff; else m_piece_time_deviation = (m_piece_time_deviation * 6 + diff * 4) / 10; m_average_piece_time = (m_average_piece_time * 6 + dl_time * 4) / 10; } } } else if (i->flags & torrent_handle::alert_when_available) { // post an empty read_piece_alert to indicate it failed m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), piece, error_code(boost::system::errc::operation_canceled, get_system_category()))); } m_time_critical_pieces.erase(i); return; } } // remove time critical pieces where priority is 0 void torrent::remove_time_critical_pieces(std::vector const& priority) { for (std::deque::iterator i = m_time_critical_pieces.begin(); i != m_time_critical_pieces.end();) { if (priority[i->piece] == 0) { if (i->flags & torrent_handle::alert_when_available) { // post an empty read_piece_alert to indicate it failed m_ses.m_alerts.post_alert(read_piece_alert( get_handle(), i->piece, error_code(boost::system::errc::operation_canceled, get_system_category()))); } i = m_time_critical_pieces.erase(i); continue; } ++i; } } void torrent::piece_availability(std::vector& avail) const { INVARIANT_CHECK; TORRENT_ASSERT(valid_metadata()); if (!has_picker()) { avail.clear(); return; } m_picker->get_availability(avail); } void torrent::set_piece_priority(int index, int priority) { // INVARIANT_CHECK; TORRENT_ASSERT(valid_metadata()); if (is_seed()) return; // this call is only valid on torrents with metadata TORRENT_ASSERT(m_picker.get()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); if (index < 0 || index >= m_torrent_file->num_pieces()) return; bool was_finished = is_finished(); bool filter_updated = m_picker->set_piece_priority(index, priority); TORRENT_ASSERT(num_have() >= m_picker->num_have_filtered()); if (filter_updated) { update_peer_interest(was_finished); if (priority == 0) remove_time_critical_piece(index); } } int torrent::piece_priority(int index) const { // INVARIANT_CHECK; TORRENT_ASSERT(valid_metadata()); if (is_seed()) return 1; // this call is only valid on torrents with metadata TORRENT_ASSERT(m_picker.get()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); if (index < 0 || index >= m_torrent_file->num_pieces()) return 0; return m_picker->piece_priority(index); } void torrent::prioritize_pieces(std::vector const& pieces) { INVARIANT_CHECK; // this call is only valid on torrents with metadata TORRENT_ASSERT(valid_metadata()); if (is_seed()) return; TORRENT_ASSERT(m_picker.get()); int index = 0; bool filter_updated = false; bool was_finished = is_finished(); for (std::vector::const_iterator i = pieces.begin() , end(pieces.end()); i != end; ++i, ++index) { TORRENT_ASSERT(*i >= 0); TORRENT_ASSERT(*i <= 7); filter_updated |= m_picker->set_piece_priority(index, *i); TORRENT_ASSERT(num_have() >= m_picker->num_have_filtered()); } if (filter_updated) { // we need to save this new state m_need_save_resume_data = true; update_peer_interest(was_finished); remove_time_critical_pieces(pieces); } state_updated(); } void torrent::piece_priorities(std::vector* pieces) const { INVARIANT_CHECK; // this call is only valid on torrents with metadata TORRENT_ASSERT(valid_metadata()); if (is_seed()) { pieces->clear(); pieces->resize(m_torrent_file->num_pieces(), 1); return; } TORRENT_ASSERT(m_picker.get()); m_picker->piece_priorities(*pieces); } namespace { void set_if_greater(int& piece_prio, int file_prio) { if (file_prio > piece_prio) piece_prio = file_prio; } } void torrent::prioritize_files(std::vector const& files) { INVARIANT_CHECK; // this call is only valid on torrents with metadata if (!valid_metadata() || is_seed()) return; // the bitmask need to have exactly one bit for every file // in the torrent TORRENT_ASSERT(int(files.size()) == m_torrent_file->num_files()); if (m_torrent_file->num_pieces() == 0) return; int limit = int(files.size()); if (valid_metadata() && limit > m_torrent_file->num_files()) limit = m_torrent_file->num_files(); if (m_file_priority.size() < limit) m_file_priority.resize(limit); std::copy(files.begin(), files.begin() + limit, m_file_priority.begin()); if (valid_metadata() && m_torrent_file->num_files() > int(m_file_priority.size())) m_file_priority.resize(m_torrent_file->num_files(), 1); update_piece_priorities(); } void torrent::set_file_priority(int index, int prio) { INVARIANT_CHECK; // this call is only valid on torrents with metadata if (!valid_metadata() || is_seed()) return; if (index < 0 || index >= m_torrent_file->num_files()) return; if (prio < 0) prio = 0; else if (prio > 7) prio = 7; if (m_file_priority.size() <= index) { if (prio == 1) return; m_file_priority.resize(m_torrent_file->num_files(), 1); } if (m_file_priority[index] == prio) return; m_file_priority[index] = prio; update_piece_priorities(); } int torrent::file_priority(int index) const { // this call is only valid on torrents with metadata if (!valid_metadata()) return 1; if (index < 0 || index >= m_torrent_file->num_files()) return 0; if (m_file_priority.size() <= index) return 1; return m_file_priority[index]; } void torrent::file_priorities(std::vector* files) const { INVARIANT_CHECK; if (!valid_metadata()) { files->resize(m_file_priority.size()); std::copy(m_file_priority.begin(), m_file_priority.end(), files->begin()); return; } files->resize(m_torrent_file->num_files(), 1); TORRENT_ASSERT(m_file_priority.size() <= m_torrent_file->num_files()); std::copy(m_file_priority.begin(), m_file_priority.end(), files->begin()); } void torrent::update_piece_priorities() { INVARIANT_CHECK; if (m_torrent_file->num_pieces() == 0) return; size_type position = 0; int piece_length = m_torrent_file->piece_length(); // initialize the piece priorities to 0, then only allow // setting higher priorities std::vector pieces(m_torrent_file->num_pieces(), 0); // [MF] just one "file" // mark all pieces of the file with this file's priority // but only if the priority is higher than the pieces // already set (to avoid problems with overlapping pieces) int start_piece = 0; int last_piece = m_torrent_file->num_pieces() - 1; TORRENT_ASSERT(last_piece < int(pieces.size())); std::for_each(pieces.begin() + start_piece , pieces.begin() + last_piece + 1 , boost::bind(&set_if_greater, _1, m_file_priority[0])); prioritize_pieces(pieces); } // this is called when piece priorities have been updated // updates the interested flag in peers void torrent::update_peer_interest(bool was_finished) { for (peer_iterator i = begin(); i != end();) { peer_connection* p = *i; // update_interest may disconnect the peer and // invalidate the iterator ++i; p->update_interest(); } // the torrent just became finished if (is_finished() && !was_finished) { finished(); } else if (!is_finished() && was_finished) { // if we used to be finished, but we aren't anymore // we may need to connect to peers again resume_download(); } } void torrent::filter_piece(int index, bool filter) { INVARIANT_CHECK; TORRENT_ASSERT(valid_metadata()); if (is_seed()) return; // this call is only valid on torrents with metadata TORRENT_ASSERT(m_picker.get()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); if (index < 0 || index >= m_torrent_file->num_pieces()) return; bool was_finished = is_finished(); m_picker->set_piece_priority(index, filter ? 1 : 0); update_peer_interest(was_finished); } void torrent::filter_pieces(std::vector const& bitmask) { INVARIANT_CHECK; // this call is only valid on torrents with metadata TORRENT_ASSERT(valid_metadata()); if (is_seed()) return; TORRENT_ASSERT(m_picker.get()); bool was_finished = is_finished(); int index = 0; for (std::vector::const_iterator i = bitmask.begin() , end(bitmask.end()); i != end; ++i, ++index) { if ((m_picker->piece_priority(index) == 0) == *i) continue; if (*i) m_picker->set_piece_priority(index, 0); else m_picker->set_piece_priority(index, 1); } update_peer_interest(was_finished); } bool torrent::is_piece_filtered(int index) const { // this call is only valid on torrents with metadata TORRENT_ASSERT(valid_metadata()); if (is_seed()) return false; TORRENT_ASSERT(m_picker.get()); TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < m_torrent_file->num_pieces()); if (index < 0 || index >= m_torrent_file->num_pieces()) return true; return m_picker->piece_priority(index) == 0; } void torrent::filtered_pieces(std::vector& bitmask) const { INVARIANT_CHECK; // this call is only valid on torrents with metadata TORRENT_ASSERT(valid_metadata()); if (is_seed()) { bitmask.clear(); bitmask.resize(m_torrent_file->num_pieces(), false); return; } TORRENT_ASSERT(m_picker.get()); m_picker->filtered_pieces(bitmask); } void torrent::filter_files(std::vector const& bitmask) { // [MF] nop } void torrent::replace_trackers(std::vector const& urls) { m_trackers.clear(); std::remove_copy_if(urls.begin(), urls.end(), back_inserter(m_trackers) , boost::bind(&std::string::empty, boost::bind(&announce_entry::url, _1))); m_last_working_tracker = -1; for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) if (i->source == 0) i->source = announce_entry::source_client; if (settings().prefer_udp_trackers) prioritize_udp_trackers(); if (!m_trackers.empty()) announce_with_tracker(); m_need_save_resume_data = true; } void torrent::prioritize_udp_trackers() { // look for udp-trackers for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { if (i->url.substr(0, 6) != "udp://") continue; // now, look for trackers with the same hostname // that is has higher priority than this one // if we find one, swap with the udp-tracker error_code ec; std::string udp_hostname; using boost::tuples::ignore; boost::tie(ignore, ignore, udp_hostname, ignore, ignore) = parse_url_components(i->url, ec); for (std::vector::iterator j = m_trackers.begin(); j != i; ++j) { std::string hostname; boost::tie(ignore, ignore, hostname, ignore, ignore) = parse_url_components(j->url, ec); if (hostname != udp_hostname) continue; if (j->url.substr(0, 6) == "udp://") continue; using std::swap; using std::iter_swap; swap(i->tier, j->tier); iter_swap(i, j); break; } } } void torrent::add_tracker(announce_entry const& url) { std::vector::iterator k = std::find_if(m_trackers.begin() , m_trackers.end(), boost::bind(&announce_entry::url, _1) == url.url); if (k != m_trackers.end()) { k->source |= url.source; return; } k = std::upper_bound(m_trackers.begin(), m_trackers.end(), url , boost::bind(&announce_entry::tier, _1) < boost::bind(&announce_entry::tier, _2)); if (k - m_trackers.begin() < m_last_working_tracker) ++m_last_working_tracker; k = m_trackers.insert(k, url); if (k->source == 0) k->source = announce_entry::source_client; if (!m_trackers.empty()) announce_with_tracker(); } bool torrent::choke_peer(peer_connection& c) { INVARIANT_CHECK; TORRENT_ASSERT(!c.is_choked()); TORRENT_ASSERT(!c.ignore_unchoke_slots()); TORRENT_ASSERT(m_num_uploads > 0); if (!c.send_choke()) return false; --m_num_uploads; state_updated(); return true; } bool torrent::unchoke_peer(peer_connection& c, bool optimistic) { INVARIANT_CHECK; TORRENT_ASSERT(!m_graceful_pause_mode); TORRENT_ASSERT(c.is_choked()); TORRENT_ASSERT(!c.ignore_unchoke_slots()); // when we're unchoking the optimistic slots, we might // exceed the limit temporarily while we're iterating // over the peers if (m_num_uploads >= m_max_uploads && !optimistic) return false; if (!c.send_unchoke()) return false; ++m_num_uploads; state_updated(); return true; } void torrent::cancel_block(piece_block block) { INVARIANT_CHECK; for (peer_iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { (*i)->cancel_request(block); } } #ifdef TORRENT_USE_OPENSSL std::string password_callback(int length, boost::asio::ssl::context::password_purpose p , std::string pw) { if (p != boost::asio::ssl::context::for_reading) return ""; return pw; } // certificate is a filename to a .pem file which is our // certificate. The certificate must be signed by the root // cert of the torrent file. any peer we connect to or that // connect to use must present a valid certificate signed // by the torrent root cert as well void torrent::set_ssl_cert(std::string const& certificate , std::string const& private_key , std::string const& dh_params , std::string const& passphrase) { if (!m_ssl_ctx) return; using boost::asio::ssl::context; error_code ec; m_ssl_ctx->set_password_callback(boost::bind(&password_callback, _1, _2, passphrase), ec); if (ec) { if (alerts().should_post()) alerts().post_alert(torrent_error_alert(get_handle(), ec)); } m_ssl_ctx->use_certificate_file(certificate, context::pem, ec); if (ec) { if (alerts().should_post()) alerts().post_alert(torrent_error_alert(get_handle(), ec)); } m_ssl_ctx->use_private_key_file(private_key, context::pem, ec); if (ec) { if (alerts().should_post()) alerts().post_alert(torrent_error_alert(get_handle(), ec)); } m_ssl_ctx->use_tmp_dh_file(dh_params, ec); if (ec) { if (alerts().should_post()) alerts().post_alert(torrent_error_alert(get_handle(), ec)); } } #endif void torrent::remove_peer(peer_connection* p) { // INVARIANT_CHECK; TORRENT_ASSERT(p != 0); TORRENT_ASSERT(m_ses.is_network_thread()); peer_iterator i = m_connections.find(p); if (i == m_connections.end()) { TORRENT_ASSERT(false); return; } if (ready_for_connections()) { TORRENT_ASSERT(p->associated_torrent().lock().get() == NULL || p->associated_torrent().lock().get() == this); if (p->is_seed()) { if (m_picker.get()) { m_picker->dec_refcount_all(p); } } else { if (m_picker.get()) { bitfield const& pieces = p->get_bitfield(); TORRENT_ASSERT(pieces.count() <= int(pieces.size())); m_picker->dec_refcount(pieces, p); } } } if (!p->is_choked() && !p->ignore_unchoke_slots()) { --m_num_uploads; m_ses.m_unchoke_time_scaler = 0; } policy::peer* pp = p->peer_info_struct(); if (pp) { if (pp->optimistically_unchoked) m_ses.m_optimistic_unchoke_time_scaler = 0; // if the share ratio is 0 (infinite), the // m_available_free_upload isn't used, // because it isn't necessary. if (ratio() != 0.f) { TORRENT_ASSERT(p->associated_torrent().lock().get() == this); TORRENT_ASSERT(p->share_diff() < (std::numeric_limits::max)()); add_free_upload(p->share_diff()); } TORRENT_ASSERT(pp->prev_amount_upload == 0); TORRENT_ASSERT(pp->prev_amount_download == 0); pp->prev_amount_download += p->statistics().total_payload_download() >> 10; pp->prev_amount_upload += p->statistics().total_payload_upload() >> 10; } m_policy.connection_closed(*p, m_ses.session_time()); p->set_peer_info(0); TORRENT_ASSERT(i != m_connections.end()); m_connections.erase(i); } void torrent::remove_web_seed(std::list::iterator web) { if (web->resolving) { web->removed = true; return; } peer_connection * peer = web->peer_info.connection; if (peer) { TORRENT_ASSERT(peer->m_in_use == 1337); peer->set_peer_info(0); } if (has_picker()) picker().clear_peer(&web->peer_info); m_web_seeds.erase(web); } void torrent::connect_to_url_seed(std::list::iterator web) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(!web->resolving); if (web->resolving) return; if (int(m_connections.size()) >= m_max_connections || m_ses.num_connections() >= m_ses.settings().connections_limit) return; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("resolving web seed: %s", web->url.c_str()); #endif std::string protocol; std::string auth; std::string hostname; int port; std::string path; error_code ec; boost::tie(protocol, auth, hostname, port, path) = parse_url_components(web->url, ec); if (port == -1) { port = protocol == "http" ? 80 : 443; } if (ec) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("failed to parse web seed url: %s", ec.message().c_str()); #endif if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, ec)); } // never try it again remove_web_seed(web); return; } if (web->peer_info.banned) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("banned web seed: %s", web->url.c_str()); #endif if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url , error_code(libtorrent::errors::peer_banned, get_libtorrent_category()))); } // never try it again remove_web_seed(web); return; } #ifdef TORRENT_USE_OPENSSL if (protocol != "http" && protocol != "https") #else if (protocol != "http") #endif { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, errors::unsupported_url_protocol)); } // never try it again remove_web_seed(web); return; } if (hostname.empty()) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, errors::invalid_hostname)); } // never try it again remove_web_seed(web); return; } if (port == 0) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, errors::invalid_port)); } // never try it again remove_web_seed(web); return; } if (m_ses.m_port_filter.access(port) & port_filter::blocked) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, errors::port_blocked)); } // never try it again remove_web_seed(web); return; } if (web->endpoint.port() != 0) { connect_web_seed(web, web->endpoint); return; } proxy_settings const& ps = m_ses.proxy(); if (ps.type == proxy_settings::http || ps.type == proxy_settings::http_pw) { // use proxy web->resolving = true; tcp::resolver::query q(ps.hostname, to_string(ps.port).elems); m_host_resolver.async_resolve(q, boost::bind(&torrent::on_proxy_name_lookup, shared_from_this(), _1, _2, web)); } else if (ps.proxy_hostnames && (ps.type == proxy_settings::socks5 || ps.type == proxy_settings::socks5_pw)) { connect_web_seed(web, tcp::endpoint(address(), port)); } else { web->resolving = true; tcp::resolver::query q(hostname, to_string(port).elems); m_host_resolver.async_resolve(q, boost::bind(&torrent::on_name_lookup, shared_from_this(), _1, _2, web , tcp::endpoint())); } } void torrent::on_proxy_name_lookup(error_code const& e, tcp::resolver::iterator host , std::list::iterator web) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(web->resolving == true); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("completed resolve proxy hostname for: %s", web->url.c_str()); #endif #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING if (e) debug_log("proxy name lookup error: %s", e.message().c_str()); #endif web->resolving = false; if (web->removed) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("removed web seed"); #endif remove_web_seed(web); return; } if (m_abort) return; if (e || host == tcp::resolver::iterator()) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, e)); } // the name lookup failed for the http host. Don't try // this host again remove_web_seed(web); return; } if (m_ses.is_aborted()) return; #ifndef TORRENT_DISABLE_GEO_IP int as = m_ses.as_for_ip(host->endpoint().address()); #ifdef TORRENT_DEBUG web->peer_info.inet_as_num = as; #endif web->peer_info.inet_as = m_ses.lookup_as(as); #endif if (int(m_connections.size()) >= m_max_connections || m_ses.num_connections() >= m_ses.settings().connections_limit) return; tcp::endpoint a(host->endpoint()); using boost::tuples::ignore; std::string hostname; int port; error_code ec; std::string protocol; boost::tie(protocol, ignore, hostname, port, ignore) = parse_url_components(web->url, ec); if (port == -1) port = protocol == "http" ? 80 : 443; if (ec) { if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert( url_seed_alert(get_handle(), web->url, ec)); } remove_web_seed(web); return; } if (m_apply_ip_filter && m_ses.m_ip_filter.access(a.address()) & ip_filter::blocked) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(peer_blocked_alert(get_handle(), a.address())); return; } web->resolving = true; tcp::resolver::query q(hostname, to_string(port).elems); m_host_resolver.async_resolve(q, boost::bind(&torrent::on_name_lookup, shared_from_this(), _1, _2, web, a)); } void torrent::on_name_lookup(error_code const& e, tcp::resolver::iterator host , std::list::iterator web, tcp::endpoint proxy) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(web->resolving == true); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("completed resolve: %s", web->url.c_str()); #endif web->resolving = false; if (web->removed) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("removed web seed"); #endif remove_web_seed(web); return; } if (m_abort) return; if (e || host == tcp::resolver::iterator()) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(url_seed_alert(get_handle(), web->url, e)); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("*** HOSTNAME LOOKUP FAILED: %s: (%d) %s" , web->url.c_str(), e.value(), e.message().c_str()); #endif // unavailable, retry in 30 minutes web->retry = time_now() + minutes(30); return; } if (int(m_connections.size()) >= m_max_connections || m_ses.num_connections() >= m_ses.settings().connections_limit) return; tcp::endpoint a(host->endpoint()); connect_web_seed(web, a); } void torrent::connect_web_seed(std::list::iterator web, tcp::endpoint a) { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_apply_ip_filter && m_ses.m_ip_filter.access(a.address()) & ip_filter::blocked) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(peer_blocked_alert(get_handle(), a.address())); return; } TORRENT_ASSERT(web->resolving == false); TORRENT_ASSERT(web->peer_info.connection == 0); web->endpoint = a; if (is_paused()) return; if (m_ses.is_aborted()) return; boost::shared_ptr s(new (std::nothrow) socket_type(m_ses.m_io_service)); if (!s) return; void* userdata = 0; #ifdef TORRENT_USE_OPENSSL bool ssl = string_begins_no_case("https://", web->url.c_str()); if (ssl) { userdata = m_ssl_ctx.get(); if (!userdata) userdata = &m_ses.m_ssl_ctx; } #endif bool ret = instantiate_connection(m_ses.m_io_service, m_ses.proxy(), *s, userdata, 0, true); (void)ret; TORRENT_ASSERT(ret); proxy_settings const& ps = m_ses.proxy(); if (s->get()) { // the web seed connection will talk immediately to // the proxy, without requiring CONNECT support s->get()->set_no_connect(true); } using boost::tuples::ignore; std::string hostname; error_code ec; boost::tie(ignore, ignore, hostname, ignore, ignore) = parse_url_components(web->url, ec); if (ec) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(url_seed_alert(get_handle(), web->url, ec)); return; } if (ps.proxy_hostnames && (ps.type == proxy_settings::socks5 || ps.type == proxy_settings::socks5_pw)) { // we're using a socks proxy and we're resolving // hostnames through it socks5_stream* str = #ifdef TORRENT_USE_OPENSSL ssl ? &s->get >()->next_layer() : #endif s->get(); TORRENT_ASSERT(str); str->set_dst_name(hostname); } setup_ssl_hostname(*s, hostname, ec); if (ec) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(url_seed_alert(get_handle(), web->url, ec)); return; } boost::intrusive_ptr c; if (web->type == web_seed_entry::url_seed) { c = new (std::nothrow) web_peer_connection( m_ses, shared_from_this(), s, a, web->url, &web->peer_info, web->auth, web->extra_headers); } else if (web->type == web_seed_entry::http_seed) { c = new (std::nothrow) http_seed_connection( m_ses, shared_from_this(), s, a, web->url, &web->peer_info, web->auth, web->extra_headers); } if (!c) return; #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS c->m_in_constructor = false; #endif #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { boost::shared_ptr pp((*i)->new_connection(c.get())); if (pp) c->add_extension(pp); } #endif TORRENT_TRY { // add the newly connected peer to this torrent's peer list m_connections.insert(boost::get_pointer(c)); m_ses.m_connections.insert(c); TORRENT_ASSERT(!web->peer_info.connection); web->peer_info.connection = c.get(); #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS web->peer_info.in_use = true; #endif c->add_stat(size_type(web->peer_info.prev_amount_download) << 10 , size_type(web->peer_info.prev_amount_upload) << 10); web->peer_info.prev_amount_download = 0; web->peer_info.prev_amount_upload = 0; #if defined TORRENT_VERBOSE_LOGGING debug_log("web seed connection started: %s", web->url.c_str()); #endif c->start(); if (c->is_disconnecting()) return; m_ses.m_half_open.enqueue( boost::bind(&peer_connection::on_connect, c, _1) , boost::bind(&peer_connection::on_timeout, c) , seconds(settings().peer_connect_timeout)); } TORRENT_CATCH (std::exception& e) { TORRENT_DECLARE_DUMMY(std::exception, e); (void)e; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("*** HOST NAME LOOKUP FAILED: %s", e.what()); #endif c->disconnect(errors::no_error, 1); } } #ifndef TORRENT_DISABLE_RESOLVE_COUNTRIES namespace { unsigned long swap_bytes(unsigned long a) { return (a >> 24) | ((a & 0xff0000) >> 8) | ((a & 0xff00) << 8) | ((a & 0xff) << 24); } } void torrent::resolve_peer_country(boost::intrusive_ptr const& p) const { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_resolving_country || is_local(p->remote().address()) || p->has_country() || p->is_connecting() || p->is_queued() || p->in_handshake() || p->remote().address().is_v6()) return; asio::ip::address_v4 reversed(swap_bytes(p->remote().address().to_v4().to_ulong())); error_code ec; tcp::resolver::query q(reversed.to_string(ec) + ".zz.countries.nerd.dk", "0"); if (ec) { p->set_country("!!"); return; } m_resolving_country = true; m_host_resolver.async_resolve(q, boost::bind(&torrent::on_country_lookup, shared_from_this(), _1, _2, p)); } namespace { struct country_entry { int code; char const* name; }; } void torrent::on_country_lookup(error_code const& error, tcp::resolver::iterator i , intrusive_ptr p) const { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; m_resolving_country = false; if (m_abort) return; // must be ordered in increasing order static const country_entry country_map[] = { { 4, "AF"}, { 8, "AL"}, { 10, "AQ"}, { 12, "DZ"}, { 16, "AS"} , { 20, "AD"}, { 24, "AO"}, { 28, "AG"}, { 31, "AZ"}, { 32, "AR"} , { 36, "AU"}, { 40, "AT"}, { 44, "BS"}, { 48, "BH"}, { 50, "BD"} , { 51, "AM"}, { 52, "BB"}, { 56, "BE"}, { 60, "BM"}, { 64, "BT"} , { 68, "BO"}, { 70, "BA"}, { 72, "BW"}, { 74, "BV"}, { 76, "BR"} , { 84, "BZ"}, { 86, "IO"}, { 90, "SB"}, { 92, "VG"}, { 96, "BN"} , {100, "BG"}, {104, "MM"}, {108, "BI"}, {112, "BY"}, {116, "KH"} , {120, "CM"}, {124, "CA"}, {132, "CV"}, {136, "KY"}, {140, "CF"} , {144, "LK"}, {148, "TD"}, {152, "CL"}, {156, "CN"}, {158, "TW"} , {162, "CX"}, {166, "CC"}, {170, "CO"}, {174, "KM"}, {175, "YT"} , {178, "CG"}, {180, "CD"}, {184, "CK"}, {188, "CR"}, {191, "HR"} , {192, "CU"}, {203, "CZ"}, {204, "BJ"}, {208, "DK"}, {212, "DM"} , {214, "DO"}, {218, "EC"}, {222, "SV"}, {226, "GQ"}, {231, "ET"} , {232, "ER"}, {233, "EE"}, {234, "FO"}, {238, "FK"}, {239, "GS"} , {242, "FJ"}, {246, "FI"}, {248, "AX"}, {250, "FR"}, {254, "GF"} , {258, "PF"}, {260, "TF"}, {262, "DJ"}, {266, "GA"}, {268, "GE"} , {270, "GM"}, {275, "PS"}, {276, "DE"}, {288, "GH"}, {292, "GI"} , {296, "KI"}, {300, "GR"}, {304, "GL"}, {308, "GD"}, {312, "GP"} , {316, "GU"}, {320, "GT"}, {324, "GN"}, {328, "GY"}, {332, "HT"} , {334, "HM"}, {336, "VA"}, {340, "HN"}, {344, "HK"}, {348, "HU"} , {352, "IS"}, {356, "IN"}, {360, "ID"}, {364, "IR"}, {368, "IQ"} , {372, "IE"}, {376, "IL"}, {380, "IT"}, {384, "CI"}, {388, "JM"} , {392, "JP"}, {398, "KZ"}, {400, "JO"}, {404, "KE"}, {408, "KP"} , {410, "KR"}, {414, "KW"}, {417, "KG"}, {418, "LA"}, {422, "LB"} , {426, "LS"}, {428, "LV"}, {430, "LR"}, {434, "LY"}, {438, "LI"} , {440, "LT"}, {442, "LU"}, {446, "MO"}, {450, "MG"}, {454, "MW"} , {458, "MY"}, {462, "MV"}, {466, "ML"}, {470, "MT"}, {474, "MQ"} , {478, "MR"}, {480, "MU"}, {484, "MX"}, {492, "MC"}, {496, "MN"} , {498, "MD"}, {500, "MS"}, {504, "MA"}, {508, "MZ"}, {512, "OM"} , {516, "NA"}, {520, "NR"}, {524, "NP"}, {528, "NL"}, {530, "AN"} , {533, "AW"}, {540, "NC"}, {548, "VU"}, {554, "NZ"}, {558, "NI"} , {562, "NE"}, {566, "NG"}, {570, "NU"}, {574, "NF"}, {578, "NO"} , {580, "MP"}, {581, "UM"}, {583, "FM"}, {584, "MH"}, {585, "PW"} , {586, "PK"}, {591, "PA"}, {598, "PG"}, {600, "PY"}, {604, "PE"} , {608, "PH"}, {612, "PN"}, {616, "PL"}, {620, "PT"}, {624, "GW"} , {626, "TL"}, {630, "PR"}, {634, "QA"}, {634, "QA"}, {638, "RE"} , {642, "RO"}, {643, "RU"}, {646, "RW"}, {654, "SH"}, {659, "KN"} , {660, "AI"}, {662, "LC"}, {666, "PM"}, {670, "VC"}, {674, "SM"} , {678, "ST"}, {682, "SA"}, {686, "SN"}, {690, "SC"}, {694, "SL"} , {702, "SG"}, {703, "SK"}, {704, "VN"}, {705, "SI"}, {706, "SO"} , {710, "ZA"}, {716, "ZW"}, {724, "ES"}, {732, "EH"}, {736, "SD"} , {740, "SR"}, {744, "SJ"}, {748, "SZ"}, {752, "SE"}, {756, "CH"} , {760, "SY"}, {762, "TJ"}, {764, "TH"}, {768, "TG"}, {772, "TK"} , {776, "TO"}, {780, "TT"}, {784, "AE"}, {788, "TN"}, {792, "TR"} , {795, "TM"}, {796, "TC"}, {798, "TV"}, {800, "UG"}, {804, "UA"} , {807, "MK"}, {818, "EG"}, {826, "GB"}, {834, "TZ"}, {840, "US"} , {850, "VI"}, {854, "BF"}, {858, "UY"}, {860, "UZ"}, {862, "VE"} , {876, "WF"}, {882, "WS"}, {887, "YE"}, {891, "CS"}, {894, "ZM"} }; if (error || i == tcp::resolver::iterator()) { // this is used to indicate that we shouldn't // try to resolve it again p->set_country("--"); return; } while (i != tcp::resolver::iterator() && !i->endpoint().address().is_v4()) ++i; if (i != tcp::resolver::iterator()) { // country is an ISO 3166 country code int country = i->endpoint().address().to_v4().to_ulong() & 0xffff; // look up the country code in the map const int size = sizeof(country_map)/sizeof(country_map[0]); country_entry tmp = {country, ""}; country_entry const* j = std::lower_bound(country_map, country_map + size, tmp , boost::bind(&country_entry::code, _1) < boost::bind(&country_entry::code, _2)); if (j == country_map + size || j->code != country) { // unknown country! p->set_country("!!"); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("IP \"%s\" was mapped to unknown country: %d" , print_address(p->remote().address()).c_str(), country); #endif return; } p->set_country(j->name); } } #endif // [MF] num_pieces increases with new twister posts void torrent::increase_num_pieces(int num_pieces) { if( m_torrent_file->num_pieces() > num_pieces ) return; m_torrent_file->increase_num_pieces(num_pieces); if(has_picker()) { m_picker->increase_num_pieces(num_pieces); } if (m_connections_initialized) { // all peer connections have to update num_pieces for (torrent::peer_iterator i = m_connections.begin(); i != m_connections.end();) { peer_connection* pc = *i; ++i; if (pc->is_disconnecting()) continue; pc->on_metadata_impl(); //if (pc->is_disconnecting()) continue; //pc->init(); } } } void torrent::read_resume_data(lazy_entry const& rd) { int num_pieces = rd.dict_find_int_value("num_pieces"); increase_num_pieces(num_pieces); m_total_uploaded = rd.dict_find_int_value("total_uploaded"); m_total_downloaded = rd.dict_find_int_value("total_downloaded"); m_active_time = rd.dict_find_int_value("active_time"); m_finished_time = rd.dict_find_int_value("finished_time"); m_seeding_time = rd.dict_find_int_value("seeding_time"); m_last_seen_complete = rd.dict_find_int_value("last_seen_complete"); m_complete = rd.dict_find_int_value("num_complete", 0xffffff); m_incomplete = rd.dict_find_int_value("num_incomplete", 0xffffff); m_downloaded = rd.dict_find_int_value("num_downloaded", 0xffffff); set_upload_limit(rd.dict_find_int_value("upload_rate_limit", -1)); set_download_limit(rd.dict_find_int_value("download_rate_limit", -1)); set_max_connections(rd.dict_find_int_value("max_connections", -1)); set_max_uploads(rd.dict_find_int_value("max_uploads", -1)); m_seed_mode = rd.dict_find_int_value("seed_mode", 0) && m_torrent_file->is_valid(); if (m_seed_mode) m_verified.resize(m_torrent_file->num_pieces(), false); super_seeding(rd.dict_find_int_value("super_seeding", 0)); m_last_scrape = rd.dict_find_int_value("last_scrape", 0); m_last_download = rd.dict_find_int_value("last_download", 0); m_last_upload = rd.dict_find_int_value("last_upload", 0); m_url = rd.dict_find_string_value("url"); m_uuid = rd.dict_find_string_value("uuid"); m_source_feed_url = rd.dict_find_string_value("feed"); if (!m_uuid.empty() || !m_url.empty()) { boost::shared_ptr me(shared_from_this()); // insert this torrent in the uuid index m_ses.m_uuids.insert(std::make_pair(m_uuid.empty() ? m_url : m_uuid, me)); } m_added_time = rd.dict_find_int_value("added_time", m_added_time); m_completed_time = rd.dict_find_int_value("completed_time", m_completed_time); if (m_completed_time != 0 && m_completed_time < m_added_time) m_completed_time = m_added_time; lazy_entry const* file_priority = rd.dict_find_list("file_priority"); if (file_priority && file_priority->list_size() == m_torrent_file->num_files()) { for (int i = 0; i < file_priority->list_size(); ++i) m_file_priority[i] = file_priority->list_int_value_at(i, 1); update_piece_priorities(); } lazy_entry const* piece_priority = rd.dict_find_string("piece_priority"); if (piece_priority && piece_priority->string_length() == m_torrent_file->num_pieces()) { char const* p = piece_priority->string_ptr(); for (int i = 0; i < piece_priority->string_length(); ++i) m_picker->set_piece_priority(i, p[i]); m_policy.recalculate_connect_candidates(); } if (!m_override_resume_data) { int auto_managed_ = rd.dict_find_int_value("auto_managed", -1); if (auto_managed_ != -1) m_auto_managed = auto_managed_; } int sequential_ = rd.dict_find_int_value("sequential_download", -1); if (sequential_ != -1) set_sequential_download(sequential_); if (!m_override_resume_data) { int paused_ = rd.dict_find_int_value("paused", -1); if (paused_ != -1) { set_allow_peers(!paused_); m_announce_to_dht = !paused_; m_announce_to_trackers = !paused_; m_announce_to_lsd = !paused_; } int dht_ = rd.dict_find_int_value("announce_to_dht", -1); if (dht_ != -1) m_announce_to_dht = dht_; int lsd_ = rd.dict_find_int_value("announce_to_lsd", -1); if (lsd_ != -1) m_announce_to_lsd = lsd_; int track_ = rd.dict_find_int_value("announce_to_trackers", -1); if (track_ != -1) m_announce_to_trackers = track_; } lazy_entry const* trackers = rd.dict_find_list("trackers"); if (trackers) { if (!m_merge_resume_trackers) m_trackers.clear(); int tier = 0; for (int i = 0; i < trackers->list_size(); ++i) { lazy_entry const* tier_list = trackers->list_at(i); if (tier_list == 0 || tier_list->type() != lazy_entry::list_t) continue; for (int j = 0; j < tier_list->list_size(); ++j) { announce_entry e(tier_list->list_string_value_at(j)); if (std::find_if(m_trackers.begin(), m_trackers.end() , boost::bind(&announce_entry::url, _1) == e.url) != m_trackers.end()) continue; e.tier = tier; e.fail_limit = 0; m_trackers.push_back(e); } ++tier; } std::sort(m_trackers.begin(), m_trackers.end(), boost::bind(&announce_entry::tier, _1) < boost::bind(&announce_entry::tier, _2)); if (settings().prefer_udp_trackers) prioritize_udp_trackers(); } lazy_entry const* url_list = rd.dict_find_list("url-list"); if (url_list) { for (int i = 0; i < url_list->list_size(); ++i) { std::string url = url_list->list_string_value_at(i); if (url.empty()) continue; if (m_torrent_file->num_files() > 1 && url[url.size()-1] != '/') url += '/'; add_web_seed(url, web_seed_entry::url_seed); } } lazy_entry const* httpseeds = rd.dict_find_list("httpseeds"); if (httpseeds) { for (int i = 0; i < httpseeds->list_size(); ++i) { std::string url = httpseeds->list_string_value_at(i); if (url.empty()) continue; add_web_seed(url, web_seed_entry::http_seed); } } if (m_torrent_file->is_merkle_torrent()) { lazy_entry const* mt = rd.dict_find_string("merkle tree"); if (mt) { std::vector tree; tree.resize(m_torrent_file->merkle_tree().size()); std::memcpy(&tree[0], mt->string_ptr() , (std::min)(mt->string_length(), int(tree.size()) * 20)); if (mt->string_length() < int(tree.size()) * 20) std::memset(&tree[0] + mt->string_length() / 20, 0 , tree.size() - mt->string_length() / 20); m_torrent_file->set_merkle_tree(tree); } else { // TODO: 0 if this is a merkle torrent and we can't // restore the tree, we need to wipe all the // bits in the have array, but not necessarily // we might want to do a full check to see if we have // all the pieces. This is low priority since almost // no one uses merkle torrents TORRENT_ASSERT(false); } } } boost::intrusive_ptr torrent::get_torrent_copy() { if (!m_torrent_file->is_valid()) return boost::intrusive_ptr(); // copy the torrent_info object return boost::intrusive_ptr(new torrent_info(*m_torrent_file)); } void torrent::write_resume_data(entry& ret) const { using namespace libtorrent::detail; // for write_*_endpoint() ret["file-format"] = "libtorrent resume file"; ret["file-version"] = 2; ret["libtorrent-version"] = LIBTORRENT_VERSION; ret["num_pieces"] = m_torrent_file->num_pieces(); ret["total_uploaded"] = m_total_uploaded; ret["total_downloaded"] = m_total_downloaded; ret["active_time"] = m_active_time; ret["finished_time"] = m_finished_time; ret["seeding_time"] = m_seeding_time; ret["last_seen_complete"] = m_last_seen_complete; ret["num_complete"] = m_complete; ret["num_incomplete"] = m_incomplete; ret["num_downloaded"] = m_downloaded; ret["sequential_download"] = m_sequential_download; ret["seed_mode"] = m_seed_mode; ret["super_seeding"] = m_super_seeding; ret["added_time"] = m_added_time; ret["completed_time"] = m_completed_time; ret["last_scrape"] = m_last_scrape; ret["last_download"] = m_last_download; ret["last_upload"] = m_last_upload; if (!m_url.empty()) ret["url"] = m_url; if (!m_uuid.empty()) ret["uuid"] = m_uuid; if (!m_source_feed_url.empty()) ret["feed"] = m_source_feed_url; const sha1_hash& info_hash = torrent_file().info_hash(); ret["info-hash"] = std::string((char*)info_hash.begin(), (char*)info_hash.end()); /* [MF] don't save metadata if (valid_metadata()) { if (m_magnet_link || (m_save_resume_flags & torrent_handle::save_info_dict)) ret["info"] = bdecode(&torrent_file().metadata()[0] , &torrent_file().metadata()[0] + torrent_file().metadata_size()); } */ // blocks per piece int num_blocks_per_piece = static_cast(torrent_file().piece_length()) / block_size(); ret["blocks per piece"] = num_blocks_per_piece; if (m_torrent_file->is_merkle_torrent()) { // we need to save the whole merkle hash tree // in order to resume std::string& tree_str = ret["merkle tree"].string(); std::vector const& tree = m_torrent_file->merkle_tree(); tree_str.resize(tree.size() * 20); std::memcpy(&tree_str[0], &tree[0], tree.size() * 20); } // save trackers if (!m_trackers.empty()) { entry::list_type& tr_list = ret["trackers"].list(); tr_list.push_back(entry::list_type()); int tier = 0; for (std::vector::const_iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { // don't save trackers we can't trust // TODO: 1 save the send_stats state instead of throwing them away // it may pose an issue when downgrading though if (i->send_stats == false) continue; if (i->tier == tier) { tr_list.back().list().push_back(i->url); } else { tr_list.push_back(entry::list_t); tr_list.back().list().push_back(i->url); tier = i->tier; } } } // save web seeds if (!m_web_seeds.empty()) { entry::list_type& url_list = ret["url-list"].list(); entry::list_type& httpseed_list = ret["httpseeds"].list(); for (std::list::const_iterator i = m_web_seeds.begin() , end(m_web_seeds.end()); i != end; ++i) { if (i->type == web_seed_entry::url_seed) url_list.push_back(i->url); else if (i->type == web_seed_entry::http_seed) httpseed_list.push_back(i->url); } } // write have bitmask // the pieces string has one byte per piece. Each // byte is a bitmask representing different properties // for the piece // bit 0: set if we have the piece // bit 1: set if we have verified the piece (in seed mode) entry::string_type& pieces = ret["pieces"].string(); pieces.resize(m_torrent_file->num_pieces()); if (is_seed()) { std::memset(&pieces[0], 1, pieces.size()); } else { for (int i = 0, end(pieces.size()); i < end; ++i) pieces[i] = m_picker->have_piece(i) ? 1 + (m_picker->post_flags(i)<<1) : 0; } if (m_seed_mode) { TORRENT_ASSERT(m_verified.size() == pieces.size()); for (int i = 0, end(pieces.size()); i < end; ++i) pieces[i] |= m_verified[i] ? 2 : 0; } // write local peers std::back_insert_iterator peers(ret["peers"].string()); std::back_insert_iterator banned_peers(ret["banned_peers"].string()); #if TORRENT_USE_IPV6 std::back_insert_iterator peers6(ret["peers6"].string()); std::back_insert_iterator banned_peers6(ret["banned_peers6"].string()); #endif // failcount is a 5 bit value int max_failcount = (std::min)(settings().max_failcount, 31); int num_saved_peers = 0; for (policy::const_iterator i = m_policy.begin_peer() , end(m_policy.end_peer()); i != end; ++i) { error_code ec; policy::peer const* p = *i; address addr = p->address(); if (p->banned) { #if TORRENT_USE_IPV6 if (addr.is_v6()) { write_address(addr, banned_peers6); write_uint16(p->port, banned_peers6); } else #endif { write_address(addr, banned_peers); write_uint16(p->port, banned_peers); } continue; } // we cannot save remote connection // since we don't know their listen port // unless they gave us their listen port // through the extension handshake // so, if the peer is not connectable (i.e. we // don't know its listen port) or if it has // been banned, don't save it. if (!p->connectable) continue; // don't save peers that don't work if (int(p->failcount) >= max_failcount) continue; // the more peers we've saved, the more picky we get // about which ones are worth saving if (num_saved_peers > 10 && int (p->failcount) > 0 && int(p->failcount) > (40 - (num_saved_peers - 10)) * max_failcount / 40) continue; // if we have 40 peers, don't save any peers whom // we've only heard from through the resume data if (num_saved_peers > 40 && p->source == peer_info::resume_data) continue; #if TORRENT_USE_IPV6 if (addr.is_v6()) { write_address(addr, peers6); write_uint16(p->port, peers6); } else #endif { write_address(addr, peers); write_uint16(p->port, peers); } ++num_saved_peers; } ret["upload_rate_limit"] = upload_limit(); ret["download_rate_limit"] = download_limit(); ret["max_connections"] = max_connections(); ret["max_uploads"] = max_uploads(); ret["paused"] = is_torrent_paused(); ret["announce_to_dht"] = m_announce_to_dht; ret["announce_to_trackers"] = m_announce_to_trackers; ret["announce_to_lsd"] = m_announce_to_lsd; ret["auto_managed"] = m_auto_managed; // write piece priorities entry::string_type& piece_priority = ret["piece_priority"].string(); piece_priority.resize(m_torrent_file->num_pieces()); if (is_seed()) { std::memset(&piece_priority[0], 1, pieces.size()); } else { for (int i = 0, end(piece_priority.size()); i < end; ++i) piece_priority[i] = m_picker->piece_priority(i); } // write file priorities entry::list_type& file_priority = ret["file_priority"].list(); file_priority.clear(); for (int i = 0, end(m_file_priority.size()); i < end; ++i) file_priority.push_back(m_file_priority[i]); } void torrent::get_full_peer_list(std::vector& v) const { v.clear(); v.reserve(m_policy.num_peers()); for (policy::const_iterator i = m_policy.begin_peer(); i != m_policy.end_peer(); ++i) { peer_list_entry e; e.ip = (*i)->ip(); e.flags = (*i)->banned ? peer_list_entry::banned : 0; e.failcount = (*i)->failcount; e.source = (*i)->source; v.push_back(e); } } void torrent::get_peer_info(std::vector& v) { v.clear(); for (peer_iterator i = begin(); i != end(); ++i) { peer_connection* peer = *i; TORRENT_ASSERT(peer->m_in_use == 1337); // incoming peers that haven't finished the handshake should // not be included in this list if (peer->associated_torrent().expired()) continue; v.push_back(peer_info()); peer_info& p = v.back(); peer->get_peer_info(p); #ifndef TORRENT_DISABLE_RESOLVE_COUNTRIES if (resolving_countries()) resolve_peer_country(intrusive_ptr(peer)); #endif } } void torrent::get_download_queue(std::vector* queue) { TORRENT_ASSERT(m_ses.is_network_thread()); queue->clear(); std::vector& blk = m_ses.m_block_info_storage; blk.clear(); if (!valid_metadata() || !has_picker()) return; piece_picker const& p = picker(); std::vector const& q = p.get_download_queue(); const int blocks_per_piece = m_picker->blocks_in_piece(0); blk.resize(q.size() * blocks_per_piece); // for some weird reason valgrind claims these are uninitialized // unless it's zeroed out here (block_info has a construct that's // supposed to initialize it) if (!blk.empty()) memset(&blk[0], 0, sizeof(blk[0]) * blk.size()); int counter = 0; for (std::vector::const_iterator i = q.begin(); i != q.end(); ++i, ++counter) { partial_piece_info pi; pi.piece_state = (partial_piece_info::state_t)i->state; pi.blocks_in_piece = p.blocks_in_piece(i->index); pi.finished = (int)i->finished; pi.writing = (int)i->writing; pi.requested = (int)i->requested; TORRENT_ASSERT(counter * blocks_per_piece + pi.blocks_in_piece <= int(blk.size())); pi.blocks = &blk[counter * blocks_per_piece]; int piece_size = int(torrent_file().piece_size(i->index)); for (int j = 0; j < pi.blocks_in_piece; ++j) { block_info& bi = pi.blocks[j]; bi.state = i->info[j].state; bi.block_size = j < pi.blocks_in_piece - 1 ? block_size() : piece_size - (j * block_size()); bool complete = bi.state == block_info::writing || bi.state == block_info::finished; if (i->info[j].peer == 0) { bi.set_peer(tcp::endpoint()); bi.bytes_progress = complete ? bi.block_size : 0; } else { policy::peer* p = static_cast(i->info[j].peer); if (p->connection) { bi.set_peer(p->connection->remote()); if (bi.state == block_info::requested) { boost::optional pbp = p->connection->downloading_piece_progress(); if (pbp && pbp->piece_index == i->index && pbp->block_index == j) { bi.bytes_progress = pbp->bytes_downloaded; TORRENT_ASSERT(bi.bytes_progress <= bi.block_size); } else { bi.bytes_progress = 0; } } else { bi.bytes_progress = complete ? bi.block_size : 0; } } else { bi.set_peer(p->ip()); bi.bytes_progress = complete ? bi.block_size : 0; } } pi.blocks[j].num_peers = i->info[j].num_peers; } pi.piece_index = i->index; queue->push_back(pi); } } bool torrent::connect_to_peer(policy::peer* peerinfo, bool ignore_limit) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; TORRENT_ASSERT(peerinfo); TORRENT_ASSERT(peerinfo->connection == 0); peerinfo->last_connected = m_ses.session_time(); #ifdef TORRENT_DEBUG if (!settings().allow_multiple_connections_per_ip) { // this asserts that we don't have duplicates in the policy's peer list peer_iterator i_ = std::find_if(m_connections.begin(), m_connections.end() , boost::bind(&peer_connection::remote, _1) == peerinfo->ip()); #if TORRENT_USE_I2P TORRENT_ASSERT(i_ == m_connections.end() || (*i_)->type() != peer_connection::bittorrent_connection || peerinfo->is_i2p_addr); #else TORRENT_ASSERT(i_ == m_connections.end() || (*i_)->type() != peer_connection::bittorrent_connection); #endif } #endif // extend connect timeout by this many seconds int timeout_extend = 0; TORRENT_ASSERT(want_more_peers() || ignore_limit); TORRENT_ASSERT(m_ses.num_connections() < m_ses.settings().connections_limit || ignore_limit); tcp::endpoint a(peerinfo->ip()); TORRENT_ASSERT(!m_apply_ip_filter || (m_ses.m_ip_filter.access(peerinfo->address()) & ip_filter::blocked) == 0); boost::shared_ptr s(new socket_type(m_ses.m_io_service)); #if TORRENT_USE_I2P bool i2p = peerinfo->is_i2p_addr; if (i2p) { bool ret = instantiate_connection(m_ses.m_io_service, m_ses.i2p_proxy(), *s); (void)ret; TORRENT_ASSERT(ret); s->get()->set_destination(static_cast(peerinfo)->destination); s->get()->set_command(i2p_stream::cmd_connect); s->get()->set_session_id(m_ses.m_i2p_conn.session_id()); // i2p setups are slow timeout_extend = 20; } else #endif { // this is where we determine if we open a regular TCP connection // or a uTP connection. If the m_utp_socket_manager pointer is not passed in // we'll instantiate a TCP connection utp_socket_manager* sm = 0; if (m_ses.m_settings.enable_outgoing_utp && (!m_ses.m_settings.enable_outgoing_tcp || peerinfo->supports_utp || peerinfo->confirmed_supports_utp)) sm = &m_ses.m_utp_socket_manager; // don't make a TCP connection if it's disabled if (sm == 0 && !m_ses.m_settings.enable_outgoing_tcp) return false; void* userdata = 0; #ifdef TORRENT_USE_OPENSSL if (is_ssl_torrent()) { userdata = m_ssl_ctx.get(); // SSL handshakes are slow timeout_extend = 10; // we don't support SSL over uTP yet sm = 0; } #endif bool ret = instantiate_connection(m_ses.m_io_service, m_ses.proxy(), *s, userdata, sm, true); (void)ret; TORRENT_ASSERT(ret); #if defined TORRENT_USE_OPENSSL && BOOST_VERSION >= 104700 if (is_ssl_torrent()) { // for ssl sockets, set the hostname std::string host_name = to_hex(m_torrent_file->info_hash().to_string()); #define CASE(t) case socket_type_int_impl >::value: \ s->get >()->set_host_name(host_name); break; switch (s->type()) { CASE(stream_socket) CASE(socks5_stream) CASE(http_stream) CASE(utp_stream) default: break; }; } #undef CASE #endif } m_ses.setup_socket_buffers(*s); boost::intrusive_ptr c(new bt_peer_connection( m_ses, s, a, peerinfo, shared_from_this(), true)); #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS c->m_in_constructor = false; #endif c->add_stat(size_type(peerinfo->prev_amount_download) << 10 , size_type(peerinfo->prev_amount_upload) << 10); peerinfo->prev_amount_download = 0; peerinfo->prev_amount_upload = 0; #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { boost::shared_ptr pp((*i)->new_connection(c.get())); if (pp) c->add_extension(pp); } TORRENT_CATCH (std::exception&) {} } #endif // add the newly connected peer to this torrent's peer list m_connections.insert(boost::get_pointer(c)); m_ses.m_connections.insert(c); m_policy.set_connection(peerinfo, c.get()); c->start(); int timeout = settings().peer_connect_timeout; if (peerinfo) timeout += 3 * peerinfo->failcount; timeout += timeout_extend; TORRENT_TRY { m_ses.m_half_open.enqueue( boost::bind(&peer_connection::on_connect, c, _1) , boost::bind(&peer_connection::on_timeout, c) , seconds(timeout)); } TORRENT_CATCH (std::exception&) { std::set::iterator i = m_connections.find(boost::get_pointer(c)); if (i != m_connections.end()) m_connections.erase(i); c->disconnect(errors::no_error, 1); return false; } if (m_share_mode) recalc_share_mode(); return peerinfo->connection; } bool torrent::set_metadata(char const* metadata_buf, int metadata_size) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_torrent_file->is_valid()) return false; hasher h; h.update(metadata_buf, metadata_size); sha1_hash info_hash = h.final(); if (info_hash != m_torrent_file->info_hash()) { if (alerts().should_post()) { alerts().post_alert(metadata_failed_alert(get_handle())); } return false; } lazy_entry metadata; error_code ec; int ret = lazy_bdecode(metadata_buf, metadata_buf + metadata_size, metadata, ec); if (ret != 0 || !m_torrent_file->parse_info_section(metadata, ec, 0)) { // this means the metadata is correct, since we // verified it against the info-hash, but we // failed to parse it. Pause the torrent if (alerts().should_post()) { // TODO: 2 pass in ec along with the alert alerts().post_alert(metadata_failed_alert(get_handle())); } set_error(errors::invalid_swarm_metadata, ""); pause(); return false; } if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(metadata_received_alert( get_handle())); } // this makes the resume data "paused" and // "auto_managed" fields be ignored. If the paused // field is not ignored, the invariant check will fail // since we will be paused but without having disconnected // any of the peers. m_override_resume_data = true; // we have to initialize the torrent before we start // disconnecting redundant peers, otherwise we'll think // we're a seed, because we have all 0 pieces init(); // disconnect redundant peers for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end;) { std::set::iterator p = i++; (*p)->disconnect_if_redundant(); } m_need_save_resume_data = true; return true; } bool torrent::attach_peer(peer_connection* p) { // INVARIANT_CHECK; #ifdef TORRENT_USE_OPENSSL #if BOOST_VERSION >= 104700 if (is_ssl_torrent()) { // if this is an SSL torrent, don't allow non SSL peers on it boost::shared_ptr s = p->get_socket(); // #define SSL(t) socket_type_int_impl >::value: \ ssl_conn = s->get >()->native_handle(); \ break; SSL* ssl_conn = 0; switch (s->type()) { case SSL(stream_socket) case SSL(socks5_stream) case SSL(http_stream) case SSL(utp_stream) }; #undef SSL if (ssl_conn == 0) { // don't allow non SSL peers on SSL torrents p->disconnect(errors::requires_ssl_connection); return false; } if (SSL_get_SSL_CTX(ssl_conn) != m_ssl_ctx->native_handle()) { // if the SSL_CTX associated with this connection is // not the one belonging to this torrent, the SSL handshake // connected to one torrent, and the BitTorrent protocol // to a different one. This is probably an attempt to circumvent // access control. Don't allow it. p->disconnect(errors::invalid_ssl_cert); return false; } } #else // BOOST_VERSION if (is_ssl_torrent()) { p->disconnect(asio::error::operation_not_supported); return false; } #endif #endif // TORRENT_USE_OPENSSL TORRENT_ASSERT(p != 0); TORRENT_ASSERT(!p->is_outgoing()); m_has_incoming = true; if (m_apply_ip_filter && m_ses.m_ip_filter.access(p->remote().address()) & ip_filter::blocked) { if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(peer_blocked_alert(get_handle(), p->remote().address())); p->disconnect(errors::banned_by_ip_filter); return false; } if ((m_state == torrent_status::queued_for_checking || m_state == torrent_status::checking_files || m_state == torrent_status::checking_resume_data) && valid_metadata()) { p->disconnect(errors::torrent_not_ready); return false; } if (m_ses.m_connections.find(p) == m_ses.m_connections.end()) { p->disconnect(errors::peer_not_constructed); return false; } if (m_ses.is_aborted()) { p->disconnect(errors::session_closing); return false; } bool maybe_replace_peer = false; if (m_connections.size() >= m_max_connections) { // if more than 10% of the connections are outgoing // connection attempts that haven't completed yet, // disconnect one of them and let this incoming // connection through. if (m_num_connecting > m_max_connections / 10) { // find one of the connecting peers and disconnect it // find any peer that's connecting (i.e. a half-open TCP connection) // that's also not disconnecting // TODO: 1 ideally, we would disconnect the oldest connection // i.e. the one that has waited the longest to connect. std::set::iterator i = std::find_if(begin(), end() , boost::bind(&peer_connection::is_connecting, _1) && !boost::bind(&peer_connection::is_disconnecting, _1)); if (i == end()) { // this seems odd, but we might as well handle it p->disconnect(errors::too_many_connections); return false; } (*i)->disconnect(errors::too_many_connections); // if this peer was let in via connections slack, // it has done its duty of causing the disconnection // of another peer p->peer_disconnected_other(); } else { maybe_replace_peer = true; } } TORRENT_TRY { #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { boost::shared_ptr pp((*i)->new_connection(p)); if (pp) p->add_extension(pp); } #endif if (!m_policy.new_connection(*p, m_ses.session_time())) { #if defined TORRENT_LOGGING debug_log("CLOSING CONNECTION \"%s\" peer list full" , print_endpoint(p->remote()).c_str()); #endif p->disconnect(errors::too_many_connections); return false; } } TORRENT_CATCH (std::exception& e) { TORRENT_DECLARE_DUMMY(std::exception, e); (void)e; #if defined TORRENT_LOGGING debug_log("CLOSING CONNECTION \"%s\" caught exception: %s" , print_endpoint(p->remote()).c_str(), e.what()); #endif p->disconnect(errors::no_error); return false; } TORRENT_ASSERT(m_connections.find(p) == m_connections.end()); m_connections.insert(p); #ifdef TORRENT_DEBUG error_code ec; TORRENT_ASSERT(p->remote() == p->get_socket()->remote_endpoint(ec) || ec); #endif TORRENT_ASSERT(p->peer_info_struct() != NULL); // we need to do this after we've added the peer to the policy // since that's when the peer is assigned its peer_info object, // which holds the rank if (maybe_replace_peer) { // now, find the lowest rank peer and disconnect that // if it's lower rank than the incoming connection peer_connection* peer = find_lowest_ranking_peer(); // TODO: 3 if peer is a really good peer, maybe we shouldn't disconnect it if (peer && peer->peer_rank() < p->peer_rank()) { peer->disconnect(errors::too_many_connections); p->peer_disconnected_other(); } else { p->disconnect(errors::too_many_connections); // we have to do this here because from the peer's point of // it wasn't really attached to the torrent, but we do need // to let policy know we're removing it remove_peer(p); return false; } } #if defined TORRENT_DEBUG && !defined TORRENT_DISABLE_INVARIANT_CHECKS m_policy.check_invariant(); #endif if (m_share_mode) recalc_share_mode(); return true; } bool torrent::want_more_peers() const { return m_connections.size() < m_max_connections && !is_paused() && ((m_state != torrent_status::checking_files && m_state != torrent_status::checking_resume_data && m_state != torrent_status::queued_for_checking) || !valid_metadata()) && m_policy.num_connect_candidates() > 0 && !m_abort && (m_ses.settings().seeding_outgoing_connections || (m_state != torrent_status::seeding && m_state != torrent_status::finished)); } void torrent::disconnect_all(error_code const& ec) { // doesn't work with the !m_allow_peers -> m_num_peers == 0 condition // INVARIANT_CHECK; while (!m_connections.empty()) { peer_connection* p = *m_connections.begin(); TORRENT_ASSERT(p->associated_torrent().lock().get() == this); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->peer_log("*** CLOSING CONNECTION \"%s\"", ec.message().c_str()); #endif #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS std::size_t size = m_connections.size(); #endif if (p->is_disconnecting()) m_connections.erase(m_connections.begin()); else p->disconnect(ec); TORRENT_ASSERT(m_connections.size() <= size); } } // this returns true if lhs is a better disconnect candidate than rhs bool compare_disconnect_peer(peer_connection const* lhs, peer_connection const* rhs) { // prefer to disconnect peers that are already disconnecting if (lhs->is_disconnecting() != rhs->is_disconnecting()) return lhs->is_disconnecting(); // prefer to disconnect peers we're not interested in if (lhs->is_interesting() != rhs->is_interesting()) return rhs->is_interesting(); // prefer to disconnect peers that are not seeds if (lhs->is_seed() != rhs->is_seed()) return rhs->is_seed(); // prefer to disconnect peers that are on parole if (lhs->on_parole() != rhs->on_parole()) return lhs->on_parole(); // prefer to disconnect peers that send data at a lower rate size_type lhs_transferred = lhs->statistics().total_payload_download(); size_type rhs_transferred = rhs->statistics().total_payload_download(); ptime now = time_now(); size_type lhs_time_connected = total_seconds(now - lhs->connected_time()); size_type rhs_time_connected = total_seconds(now - rhs->connected_time()); lhs_transferred /= lhs_time_connected + 1; rhs_transferred /= (rhs_time_connected + 1); if (lhs_transferred != rhs_transferred) return lhs_transferred < rhs_transferred; // prefer to disconnect peers that chokes us if (lhs->is_choked() != rhs->is_choked()) return lhs->is_choked(); return lhs->last_received() < rhs->last_received(); } int torrent::disconnect_peers(int num, error_code const& ec) { INVARIANT_CHECK; #ifdef TORRENT_DEBUG for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { // make sure this peer is not a dangling pointer TORRENT_ASSERT(m_ses.has_peer(*i)); } #endif int ret = 0; while (ret < num && !m_connections.empty()) { std::set::iterator i = std::min_element( m_connections.begin(), m_connections.end(), compare_disconnect_peer); peer_connection* p = *i; ++ret; TORRENT_ASSERT(p->associated_torrent().lock().get() == this); #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS int num_conns = m_connections.size(); #endif p->disconnect(ec); TORRENT_ASSERT(int(m_connections.size()) == num_conns - 1); } return ret; } int torrent::bandwidth_throttle(int channel) const { return m_bandwidth_channel[channel].throttle(); } // called when torrent is finished (all interesting // pieces have been downloaded) void torrent::finished() { INVARIANT_CHECK; TORRENT_ASSERT(is_finished()); TORRENT_ASSERT(m_state != torrent_status::finished && m_state != torrent_status::seeding); set_state(torrent_status::finished); set_queue_position(-1); // we have to call completed() before we start // disconnecting peers, since there's an assert // to make sure we're cleared the piece picker if (is_seed()) completed(); send_upload_only(); state_updated(); m_completed_time = time(0); // disconnect all seeds if (settings().close_redundant_connections) { // TODO: 1 should disconnect all peers that have the pieces we have // not just seeds. It would be pretty expensive to check all pieces // for all peers though std::vector seeds; for (peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { peer_connection* p = *i; TORRENT_ASSERT(p->associated_torrent().lock().get() == this); if (p->upload_only()) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->peer_log("*** SEED, CLOSING CONNECTION"); #endif seeds.push_back(p); } } std::for_each(seeds.begin(), seeds.end() , boost::bind(&peer_connection::disconnect, _1, errors::torrent_finished, 0)); } if (m_abort) return; m_policy.recalculate_connect_candidates(); TORRENT_ASSERT(m_storage); // we need to keep the object alive during this operation m_storage->async_release_files( boost::bind(&torrent::on_files_released, shared_from_this(), _1, _2)); // this torrent just completed downloads, which means it will fall // under a different limit with the auto-manager. Make sure we // update auto-manage torrents in that case if (m_auto_managed) m_ses.trigger_auto_manage(); } // this is called when we were finished, but some files were // marked for downloading, and we are no longer finished void torrent::resume_download() { INVARIANT_CHECK; TORRENT_ASSERT(!is_finished()); set_state(torrent_status::downloading); set_queue_position((std::numeric_limits::max)()); m_policy.recalculate_connect_candidates(); m_completed_time = 0; send_upload_only(); } // called when torrent is complete (all pieces downloaded) void torrent::completed() { m_picker.reset(); set_state(torrent_status::seeding); if (!m_announcing) return; ptime now = time_now(); for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { if (i->complete_sent) continue; i->next_announce = now; i->min_announce = now; } announce_with_tracker(); } // this will move the tracker with the given index // to a prioritized position in the list (move it towards // the begining) and return the new index to the tracker. int torrent::prioritize_tracker(int index) { INVARIANT_CHECK; TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < int(m_trackers.size())); if (index >= (int)m_trackers.size()) return -1; while (index > 0 && m_trackers[index].tier == m_trackers[index-1].tier) { using std::swap; swap(m_trackers[index], m_trackers[index-1]); if (m_last_working_tracker == index) --m_last_working_tracker; else if (m_last_working_tracker == index - 1) ++m_last_working_tracker; --index; } return index; } int torrent::deprioritize_tracker(int index) { INVARIANT_CHECK; TORRENT_ASSERT(index >= 0); TORRENT_ASSERT(index < int(m_trackers.size())); if (index >= (int)m_trackers.size()) return -1; while (index < int(m_trackers.size()) - 1 && m_trackers[index].tier == m_trackers[index + 1].tier) { using std::swap; swap(m_trackers[index], m_trackers[index + 1]); if (m_last_working_tracker == index) ++m_last_working_tracker; else if (m_last_working_tracker == index + 1) --m_last_working_tracker; ++index; } return index; } void torrent::files_checked() { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(m_torrent_file->is_valid()); if (m_abort) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("files_checked(), paused"); #endif return; } // we might be finished already, in which case we should // not switch to downloading mode. If all files are // filtered, we're finished when we start. if (m_state != torrent_status::finished && m_state != torrent_status::seeding) set_state(torrent_status::downloading); INVARIANT_CHECK; if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(torrent_checked_alert( get_handle())); } // calling pause will also trigger the auto managed // recalculation // if we just got here by downloading the metadata, // just keep going, no need to disconnect all peers just // to restart the torrent in a second if (m_auto_managed) { // if this is an auto managed torrent, force a recalculation // of which torrents to have active m_ses.trigger_auto_manage(); } if (!is_seed()) { // turn off super seeding if we're not a seed if (m_super_seeding) m_super_seeding = false; // if we just finished checking and we're not a seed, we are // likely to be unpaused m_ses.trigger_auto_manage(); if (is_finished() && m_state != torrent_status::finished) finished(); } else { for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) i->complete_sent = true; if (m_state != torrent_status::finished && m_state != torrent_status::seeding) finished(); } #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->on_files_checked(); } TORRENT_CATCH (std::exception&) {} } #endif if (!m_connections_initialized) { m_connections_initialized = true; // all peer connections have to initialize themselves now that the metadata // is available for (torrent::peer_iterator i = m_connections.begin(); i != m_connections.end();) { peer_connection* pc = *i; ++i; if (pc->is_disconnecting()) continue; pc->on_metadata_impl(); //if (pc->is_disconnecting()) continue; //pc->init(); } } m_files_checked = true; start_announcing(); } alert_manager& torrent::alerts() const { TORRENT_ASSERT(m_ses.is_network_thread()); return m_ses.m_alerts; } std::string torrent::save_path() const { return m_save_path; } bool torrent::rename_file(int index, std::string const& name) { } void torrent::move_storage(std::string const& save_path, int flags) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_owning_storage.get()) { #if TORRENT_USE_UNC_PATHS std::string path = canonicalize_path(save_path); #else std::string const& path = save_path; #endif m_owning_storage->async_move_storage(path, flags , boost::bind(&torrent::on_storage_moved, shared_from_this(), _1, _2)); } else { #if TORRENT_USE_UNC_PATHS m_save_path = canonicalize_path(save_path); #else m_save_path = save_path; #endif if (alerts().should_post()) { alerts().post_alert(storage_moved_alert(get_handle(), m_save_path)); } } } void torrent::on_storage_moved(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (ret == piece_manager::no_error || ret == piece_manager::need_full_check) { if (alerts().should_post()) alerts().post_alert(storage_moved_alert(get_handle(), j.str)); m_save_path = j.str; if (ret == piece_manager::need_full_check) force_recheck(); } else { if (alerts().should_post()) alerts().post_alert(storage_moved_failed_alert(get_handle(), j.error)); } } piece_manager& torrent::filesystem() { TORRENT_ASSERT(m_owning_storage.get()); TORRENT_ASSERT(m_storage); return *m_storage; } torrent_handle torrent::get_handle() { TORRENT_ASSERT(m_ses.is_network_thread()); return torrent_handle(shared_from_this()); } session_settings const& torrent::settings() const { TORRENT_ASSERT(m_ses.is_network_thread()); return m_ses.settings(); } #if defined TORRENT_DEBUG && !defined TORRENT_DISABLE_INVARIANT_CHECKS void torrent::check_invariant() const { for (std::deque::const_iterator i = m_time_critical_pieces.begin() , end(m_time_critical_pieces.end()); i != end; ++i) { TORRENT_ASSERT(!is_seed()); TORRENT_ASSERT(!has_picker() || !m_picker->have_piece(i->piece)); } TORRENT_ASSERT(m_ses.is_network_thread()); if (is_paused()) TORRENT_ASSERT(num_peers() == 0 || m_graceful_pause_mode); if (!should_check_files()) TORRENT_ASSERT(m_state != torrent_status::checking_files); else TORRENT_ASSERT(m_queued_for_checking); if (!m_ses.m_queued_for_checking.empty()) { // if there are torrents waiting to be checked // assert that there's a torrent that is being // processed right now int found = 0; int found_active = 0; for (aux::session_impl::torrent_map::iterator i = m_ses.m_torrents.begin() , end(m_ses.m_torrents.end()); i != end; ++i) if (i->second->m_state == torrent_status::checking_files) { ++found; if (i->second->should_check_files()) ++found_active; } // if the session is paused, there might still be some torrents // in the checking_files state that haven't been dequeued yet if (m_ses.is_paused()) { TORRENT_ASSERT(found_active == 0); } else { // the case of 2 is in the special case where one switches over from // checking to complete. TORRENT_ASSERT(found_active >= 1); TORRENT_ASSERT(found_active <= 2); TORRENT_ASSERT(found >= 1); } } TORRENT_ASSERT(m_resume_entry.type() == lazy_entry::dict_t || m_resume_entry.type() == lazy_entry::none_t); int num_uploads = 0; std::map num_requests; for (const_peer_iterator i = begin(); i != end(); ++i) { #ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS // make sure this peer is not a dangling pointer TORRENT_ASSERT(m_ses.has_peer(*i)); #endif peer_connection const& p = *(*i); for (std::vector::const_iterator i = p.request_queue().begin() , end(p.request_queue().end()); i != end; ++i) ++num_requests[i->block]; for (std::vector::const_iterator i = p.download_queue().begin() , end(p.download_queue().end()); i != end; ++i) if (!i->not_wanted && !i->timed_out) ++num_requests[i->block]; if (!p.is_choked() && !p.ignore_unchoke_slots()) ++num_uploads; torrent* associated_torrent = p.associated_torrent().lock().get(); if (associated_torrent != this && associated_torrent != 0) TORRENT_ASSERT(false); } TORRENT_ASSERT(num_uploads == int(m_num_uploads)); if (has_picker()) { for (std::map::iterator i = num_requests.begin() , end(num_requests.end()); i != end; ++i) { piece_block b = i->first; int count = i->second; int picker_count = m_picker->num_peers(b); // if we're no longer downloading the piece // (for instance, it may be fully downloaded and waiting // for the hash check to return), the piece picker always // returns 0 requests, regardless of how many peers may still // have the block in their queue if (!m_picker->is_downloaded(b) && m_picker->is_downloading(b.piece_index)) TORRENT_ASSERT(picker_count == count); } TORRENT_ASSERT(num_have() >= m_picker->num_have_filtered()); } if (valid_metadata()) { TORRENT_ASSERT(m_abort || m_error || !m_picker || m_picker->num_pieces() == m_torrent_file->num_pieces()); } else { TORRENT_ASSERT(m_abort || m_error || !m_picker || m_picker->num_pieces() == 0); } #ifdef TORRENT_EXPENSIVE_INVARIANT_CHECKS // make sure we haven't modified the peer object // in a way that breaks the sort order if (m_policy.begin_peer() != m_policy.end_peer()) { policy::const_iterator i = m_policy.begin_peer(); policy::const_iterator prev = i++; policy::const_iterator end(m_policy.end_peer()); policy::peer_address_compare cmp; for (; i != end; ++i, ++prev) { TORRENT_ASSERT(!cmp(*i, *prev)); } } #endif size_type total_done = quantized_bytes_done(); if (m_torrent_file->is_valid()) { if (is_seed()) TORRENT_ASSERT(total_done == m_torrent_file->total_size()); else TORRENT_ASSERT(total_done != m_torrent_file->total_size() || !m_files_checked); TORRENT_ASSERT(block_size() <= m_torrent_file->piece_length()); } else { TORRENT_ASSERT(total_done == 0); } if (m_picker && !m_abort) { // make sure that pieces that have completed the download // of all their blocks are in the disk io thread's queue // to be checked. const std::vector& dl_queue = m_picker->get_download_queue(); for (std::vector::const_iterator i = dl_queue.begin(); i != dl_queue.end(); ++i) { const int blocks_per_piece = m_picker->blocks_in_piece(i->index); bool complete = true; for (int j = 0; j < blocks_per_piece; ++j) { if (i->info[j].state == piece_picker::block_info::state_finished) continue; complete = false; break; } } } if (m_files_checked && valid_metadata()) { TORRENT_ASSERT(block_size() > 0); } for (std::vector::const_iterator i = m_file_progress.begin() , end(m_file_progress.end()); i != end; ++i) { int index = i - m_file_progress.begin(); TORRENT_ASSERT(*i <= m_torrent_file->files().file_size(index)); } } #endif void torrent::set_sequential_download(bool sd) { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_sequential_download == sd) return; m_sequential_download = sd; m_need_save_resume_data = true; state_updated(); } void torrent::queue_up() { set_queue_position(queue_position() == 0 ? queue_position() : queue_position() - 1); } void torrent::queue_down() { set_queue_position(queue_position() + 1); } void torrent::set_queue_position(int p) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT((p == -1) == is_finished() || (!m_auto_managed && p == -1) || (m_abort && p == -1)); if (is_finished() && p != -1) return; if (p == m_sequence_number) return; state_updated(); session_impl::torrent_map& torrents = m_ses.m_torrents; if (p >= 0 && m_sequence_number == -1) { int max_seq = -1; for (session_impl::torrent_map::iterator i = torrents.begin() , end(torrents.end()); i != end; ++i) { torrent* t = i->second.get(); if (t->m_sequence_number > max_seq) max_seq = t->m_sequence_number; if (t->m_sequence_number >= p) { ++t->m_sequence_number; t->state_updated(); } } m_sequence_number = (std::min)(max_seq + 1, p); } else if (p < 0) { for (session_impl::torrent_map::iterator i = torrents.begin() , end(torrents.end()); i != end; ++i) { torrent* t = i->second.get(); if (t == this) continue; if (t->m_sequence_number >= m_sequence_number && t->m_sequence_number != -1) { --t->m_sequence_number; t->state_updated(); } } m_sequence_number = p; } else if (p < m_sequence_number) { for (session_impl::torrent_map::iterator i = torrents.begin() , end(torrents.end()); i != end; ++i) { torrent* t = i->second.get(); if (t == this) continue; if (t->m_sequence_number >= p && t->m_sequence_number < m_sequence_number && t->m_sequence_number != -1) { ++t->m_sequence_number; t->state_updated(); } } m_sequence_number = p; } else if (p > m_sequence_number) { int max_seq = 0; for (session_impl::torrent_map::iterator i = torrents.begin() , end(torrents.end()); i != end; ++i) { torrent* t = i->second.get(); int pos = t->m_sequence_number; if (pos > max_seq) max_seq = pos; if (t == this) continue; if (pos <= p && pos > m_sequence_number && pos != -1) { --t->m_sequence_number; t->state_updated(); } } m_sequence_number = (std::min)(max_seq, p); } m_ses.m_auto_manage_time_scaler = 2; } void torrent::set_max_uploads(int limit, bool state_update) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); if (limit <= 0) limit = (1<<24)-1; if (m_max_uploads != limit && state_update) state_updated(); m_max_uploads = limit; m_need_save_resume_data = true; } void torrent::set_max_connections(int limit, bool state_update) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); if (limit <= 0) limit = (1<<24)-1; if (m_max_connections != limit && state_update) state_updated(); m_max_connections = limit; if (num_peers() > int(m_max_connections)) { disconnect_peers(num_peers() - m_max_connections , error_code(errors::too_many_connections, get_libtorrent_category())); } m_need_save_resume_data = true; } int torrent::get_peer_upload_limit(tcp::endpoint ip) const { TORRENT_ASSERT(m_ses.is_network_thread()); const_peer_iterator i = std::find_if(m_connections.begin(), m_connections.end() , boost::bind(&peer_connection::remote, _1) == ip); if (i == m_connections.end()) return -1; return (*i)->get_upload_limit(); } int torrent::get_peer_download_limit(tcp::endpoint ip) const { TORRENT_ASSERT(m_ses.is_network_thread()); const_peer_iterator i = std::find_if(m_connections.begin(), m_connections.end() , boost::bind(&peer_connection::remote, _1) == ip); if (i == m_connections.end()) return -1; return (*i)->get_download_limit(); } void torrent::set_peer_upload_limit(tcp::endpoint ip, int limit) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); peer_iterator i = std::find_if(m_connections.begin(), m_connections.end() , boost::bind(&peer_connection::remote, _1) == ip); if (i == m_connections.end()) return; (*i)->set_upload_limit(limit); } void torrent::set_peer_download_limit(tcp::endpoint ip, int limit) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); peer_iterator i = std::find_if(m_connections.begin(), m_connections.end() , boost::bind(&peer_connection::remote, _1) == ip); if (i == m_connections.end()) return; (*i)->set_download_limit(limit); } void torrent::set_upload_limit(int limit, bool state_update) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); if (limit <= 0) limit = 0; if (m_bandwidth_channel[peer_connection::upload_channel].throttle() != limit && state_update) state_updated(); m_bandwidth_channel[peer_connection::upload_channel].throttle(limit); m_need_save_resume_data = true; } int torrent::upload_limit() const { TORRENT_ASSERT(m_ses.is_network_thread()); int limit = m_bandwidth_channel[peer_connection::upload_channel].throttle(); if (limit == (std::numeric_limits::max)()) limit = -1; return limit; } void torrent::set_download_limit(int limit, bool state_update) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(limit >= -1); if (limit <= 0) limit = 0; if (m_bandwidth_channel[peer_connection::download_channel].throttle() != limit && state_update) state_updated(); m_bandwidth_channel[peer_connection::download_channel].throttle(limit); m_need_save_resume_data = true; } int torrent::download_limit() const { TORRENT_ASSERT(m_ses.is_network_thread()); int limit = m_bandwidth_channel[peer_connection::download_channel].throttle(); if (limit == (std::numeric_limits::max)()) limit = -1; return limit; } bool torrent::delete_files() { TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING log_to_all_peers("DELETING FILES IN TORRENT"); #endif disconnect_all(errors::torrent_removed); stop_announcing(); if (m_owning_storage.get()) { TORRENT_ASSERT(m_storage); m_storage->async_delete_files( boost::bind(&torrent::on_files_deleted, shared_from_this(), _1, _2)); return true; } return false; } void torrent::clear_error() { TORRENT_ASSERT(m_ses.is_network_thread()); if (!m_error) return; bool checking_files = should_check_files(); m_ses.trigger_auto_manage(); m_error = error_code(); m_error_file.clear(); state_updated(); // if we haven't downloaded the metadata from m_url, try again if (!m_url.empty() && !m_torrent_file->is_valid()) { start_download_url(); return; } // if the error happened during initialization, try again now if (!m_storage) init(); if (!checking_files && should_check_files()) queue_torrent_check(); } void torrent::set_error(error_code const& ec, std::string const& error_file) { TORRENT_ASSERT(m_ses.is_network_thread()); bool checking_files = should_check_files(); m_error = ec; m_error_file = error_file; if (alerts().should_post()) alerts().post_alert(torrent_error_alert(get_handle(), ec)); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING if (ec) { char buf[1024]; snprintf(buf, sizeof(buf), "TORRENT ERROR: %s: %s", ec.message().c_str(), error_file.c_str()); log_to_all_peers(buf); } #endif if (checking_files && !should_check_files()) { // stop checking m_storage->abort_disk_io(); dequeue_torrent_check(); set_state(torrent_status::queued_for_checking); } state_updated(); } void torrent::auto_managed(bool a) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_auto_managed == a) return; bool checking_files = should_check_files(); m_auto_managed = a; state_updated(); // we need to save this new state as well m_need_save_resume_data = true; // recalculate which torrents should be // paused m_ses.trigger_auto_manage(); if (!checking_files && should_check_files()) { queue_torrent_check(); } else if (checking_files && !should_check_files()) { // stop checking m_storage->abort_disk_io(); dequeue_torrent_check(); set_state(torrent_status::queued_for_checking); } // if this torrent is running and just became auto-managed // we might want to pause it in favor of some other torrent if (m_auto_managed && !is_paused()) m_ses.m_auto_manage_time_scaler = 2; } // the higher seed rank, the more important to seed int torrent::seed_rank(session_settings const& s) const { TORRENT_ASSERT(m_ses.is_network_thread()); enum flags { seed_ratio_not_met = 0x40000000, no_seeds = 0x20000000, recently_started = 0x10000000, prio_mask = 0x0fffffff }; if (!is_finished()) return 0; int scale = 1000; if (!is_seed()) scale = 500; int ret = 0; ptime now = time_now(); int finished_time = m_finished_time; int download_time = int(m_active_time) - finished_time; // if we haven't yet met the seed limits, set the seed_ratio_not_met // flag. That will make this seed prioritized // downloaded may be 0 if the torrent is 0-sized size_type downloaded = (std::max)(m_total_downloaded, m_torrent_file->total_size()); if (finished_time < s.seed_time_limit && (download_time > 1 && finished_time / download_time < s.seed_time_ratio_limit) && downloaded > 0 && m_total_uploaded / downloaded < s.share_ratio_limit) ret |= seed_ratio_not_met; // if this torrent is running, and it was started less // than 30 minutes ago, give it priority, to avoid oscillation if (!is_paused() && now - m_started < minutes(30)) ret |= recently_started; // if we have any scrape data, use it to calculate // seed rank int seeds = 0; int downloaders = 0; if (m_complete != 0xffffff) seeds = m_complete; else seeds = m_policy.num_seeds(); if (m_incomplete != 0xffffff) downloaders = m_incomplete; else downloaders = m_policy.num_peers() - m_policy.num_seeds(); if (seeds == 0) { ret |= no_seeds; ret |= downloaders & prio_mask; } else { ret |= ((1 + downloaders) * scale / seeds) & prio_mask; } return ret; } // this is an async operation triggered by the client void torrent::save_resume_data(int flags) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (!valid_metadata()) { alerts().post_alert(save_resume_data_failed_alert(get_handle() , errors::no_metadata)); return; } if (!m_owning_storage.get()) { alerts().post_alert(save_resume_data_failed_alert(get_handle() , errors::destructing_torrent)); return; } m_need_save_resume_data = false; m_last_saved_resume = time(0); m_save_resume_flags = boost::uint8_t(flags); state_updated(); TORRENT_ASSERT(m_storage); if (m_state == torrent_status::queued_for_checking || m_state == torrent_status::checking_files || m_state == torrent_status::checking_resume_data) { boost::shared_ptr rd(new entry); write_resume_data(*rd); alerts().post_alert(save_resume_data_alert(rd , get_handle())); return; } if (flags & torrent_handle::flush_disk_cache) m_storage->async_release_files(); m_storage->async_save_resume_data( boost::bind(&torrent::on_save_resume_data, shared_from_this(), _1, _2)); } bool torrent::should_check_files() const { TORRENT_ASSERT(m_ses.is_network_thread()); // #error should m_allow_peers really affect checking? return (m_state == torrent_status::checking_files || m_state == torrent_status::queued_for_checking) && (m_allow_peers || m_auto_managed) && !has_error() && !m_abort && !m_graceful_pause_mode && !m_ses.is_paused(); } void torrent::flush_cache() { TORRENT_ASSERT(m_ses.is_network_thread()); m_storage->async_release_files( boost::bind(&torrent::on_cache_flushed, shared_from_this(), _1, _2)); } void torrent::on_cache_flushed(int ret, disk_io_job const& j) { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_ses.is_aborted()) return; if (alerts().should_post()) alerts().post_alert(cache_flushed_alert(get_handle())); } bool torrent::is_paused() const { TORRENT_ASSERT(m_ses.is_network_thread()); return !m_allow_peers || m_ses.is_paused() || m_graceful_pause_mode; } void torrent::pause(bool graceful) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (!m_allow_peers) return; if (!graceful) set_allow_peers(false); m_announce_to_dht = false; m_announce_to_trackers = false; m_announce_to_lsd = false; // we need to save this new state m_need_save_resume_data = true; state_updated(); bool prev_graceful = m_graceful_pause_mode; m_graceful_pause_mode = graceful; if (!m_ses.is_paused() || (prev_graceful && !m_graceful_pause_mode)) { do_pause(); m_ses.trigger_auto_manage(); } } void torrent::do_pause() { TORRENT_ASSERT(m_ses.is_network_thread()); if (!is_paused()) return; #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { if ((*i)->on_pause()) return; } TORRENT_CATCH (std::exception&) {} } #endif state_updated(); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING log_to_all_peers("PAUSING TORRENT"); #endif // this will make the storage close all // files and flush all cached data if (m_owning_storage.get()) { TORRENT_ASSERT(m_storage); m_storage->async_release_files( boost::bind(&torrent::on_torrent_paused, shared_from_this(), _1, _2)); m_storage->async_clear_read_cache(); } else { if (alerts().should_post()) alerts().post_alert(torrent_paused_alert(get_handle())); } if (!m_graceful_pause_mode) { disconnect_all(errors::torrent_paused); } else { // disconnect all peers with no outstanding data to receive // and choke all remaining peers to prevent responding to new // requests for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end;) { std::set::iterator j = i++; peer_connection* p = *j; TORRENT_ASSERT(p->associated_torrent().lock().get() == this); if (p->is_disconnecting()) { m_connections.erase(j); continue; } if (p->outstanding_bytes() > 0) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->peer_log("*** CHOKING PEER: torrent graceful paused"); #endif // remove any un-sent requests from the queue p->clear_request_queue(); // don't accept new requests from the peer if (!p->is_choked()) m_ses.choke_peer(*p); continue; } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->peer_log("*** CLOSING CONNECTION: torrent_paused"); #endif p->disconnect(errors::torrent_paused); } } stop_announcing(); if (m_queued_for_checking && !should_check_files()) { // stop checking m_storage->abort_disk_io(); dequeue_torrent_check(); set_state(torrent_status::queued_for_checking); TORRENT_ASSERT(!m_queued_for_checking); } } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING || defined TORRENT_LOGGING void torrent::log_to_all_peers(char const* message) { TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING for (peer_iterator i = m_connections.begin(); i != m_connections.end(); ++i) { (*i)->peer_log("*** %s", message); } #endif debug_log("%s", message); } #endif // add or remove a url that will be attempted for // finding the file(s) in this torrent. void torrent::add_web_seed(std::string const& url, web_seed_entry::type_t type) { web_seed_entry ent(url, type); // don't add duplicates if (std::find(m_web_seeds.begin(), m_web_seeds.end(), ent) != m_web_seeds.end()) return; m_web_seeds.push_back(ent); } void torrent::add_web_seed(std::string const& url, web_seed_entry::type_t type , std::string const& auth, web_seed_entry::headers_t const& extra_headers) { web_seed_entry ent(url, type, auth, extra_headers); // don't add duplicates if (std::find(m_web_seeds.begin(), m_web_seeds.end(), ent) != m_web_seeds.end()) return; m_web_seeds.push_back(ent); } void torrent::set_allow_peers(bool b, bool graceful) { TORRENT_ASSERT(m_ses.is_network_thread()); if (m_allow_peers == b && (m_graceful_pause_mode == graceful || m_ses.is_paused()) ) return; m_allow_peers = b; if (!m_ses.is_paused()) m_graceful_pause_mode = graceful; if (!b) { m_announce_to_dht = false; m_announce_to_trackers = false; m_announce_to_lsd = false; do_pause(); } else { do_resume(); } update_guage(); } void torrent::resume() { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; if (m_allow_peers && m_announce_to_dht && m_announce_to_trackers && m_announce_to_lsd) return; m_announce_to_dht = true; m_announce_to_trackers = true; m_announce_to_lsd = true; // this call will trigger a tracker announce, that's // why it's important to set announce_to_trackers to // true first set_allow_peers(true); if (!m_ses.is_paused()) m_graceful_pause_mode = false; // we need to save this new state m_need_save_resume_data = true; do_resume(); } void torrent::do_resume() { TORRENT_ASSERT(m_ses.is_network_thread()); if (is_paused()) return; #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { if ((*i)->on_resume()) return; } TORRENT_CATCH (std::exception&) {} } #endif if (alerts().should_post()) alerts().post_alert(torrent_resumed_alert(get_handle())); state_updated(); m_started = time_now(); clear_error(); start_announcing(); if (!m_queued_for_checking && should_check_files()) queue_torrent_check(); } void torrent::update_tracker_timer(ptime now) { TORRENT_ASSERT(m_ses.is_network_thread()); if (!m_announcing) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("*** update tracker timer: not announcing"); #endif return; } ptime next_announce = max_time(); int tier = INT_MAX; bool found_working = false; for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING char msg[1000]; snprintf(msg, sizeof(msg), "*** update tracker timer: considering \"%s\" " "[ announce_to_all_tiers: %d announce_to_all_trackers: %d" " found_working: %d i->tier: %d tier: %d " " is_working: %d fails: %d fail_limit: %d updating: %d ]" , i->url.c_str(), settings().announce_to_all_tiers , settings().announce_to_all_trackers, found_working , i->tier, tier, i->is_working(), i->fails, i->fail_limit , i->updating); debug_log(msg); #endif if (settings().announce_to_all_tiers && found_working && i->tier <= tier && tier != INT_MAX) continue; if (i->tier > tier && !settings().announce_to_all_tiers) break; if (i->is_working()) { tier = i->tier; found_working = false; } if (i->fails >= i->fail_limit && i->fail_limit != 0) continue; if (i->updating) { found_working = true; } else { ptime next_tracker_announce = (std::max)(i->next_announce, i->min_announce); if (next_tracker_announce < next_announce && (!found_working || i->is_working())) next_announce = next_tracker_announce; } if (i->is_working()) found_working = true; if (found_working && !settings().announce_to_all_trackers && !settings().announce_to_all_tiers) break; } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING char msg[200]; snprintf(msg, sizeof(msg), "*** update tracker timer: next_announce < now %d" " m_waiting_tracker: %d next_announce_in: %d" , next_announce <= now, m_waiting_tracker, total_seconds(now - next_announce)); debug_log(msg); #endif if (next_announce <= now) next_announce = now; // don't re-issue the timer if it's the same expiration time as last time // if m_waiting_tracker is false, expires_at() is undefined if (m_waiting_tracker && m_tracker_timer.expires_at() == next_announce) return; m_waiting_tracker = true; error_code ec; boost::weak_ptr self(shared_from_this()); #if defined TORRENT_ASIO_DEBUGGING add_outstanding_async("tracker::on_tracker_announce_disp"); #endif m_tracker_timer.expires_at(next_announce, ec); m_tracker_timer.async_wait(boost::bind(&torrent::on_tracker_announce_disp, self, _1)); } void torrent::start_announcing() { TORRENT_ASSERT(m_ses.is_network_thread()); if (is_paused()) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("start_announcing(), paused"); #endif return; } // if we don't have metadata, we need to announce // before checking files, to get peers to // request the metadata from if (!m_files_checked && valid_metadata()) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("start_announcing(), files not checked (with valid metadata)"); #endif return; } if (!m_torrent_file->is_valid() && !m_url.empty()) { #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("start_announcing(), downloading URL"); #endif return; } if (m_announcing) return; m_announcing = true; #ifndef TORRENT_DISABLE_DHT if (m_policy.num_peers() == 0 && m_ses.m_dht) { // we don't have any peers, prioritize // announcing this torrent with the DHT m_ses.prioritize_dht(shared_from_this()); } #endif if (!m_trackers.empty()) { // tell the tracker that we're back std::for_each(m_trackers.begin(), m_trackers.end() , boost::bind(&announce_entry::reset, _1)); } // reset the stats, since from the tracker's // point of view, this is a new session m_total_failed_bytes = 0; m_total_redundant_bytes = 0; m_stat.clear(); announce_with_tracker(); // private torrents are never announced on LSD // or on DHT, we don't need this timer. if (!m_torrent_file->is_valid() || (!m_torrent_file->priv() && (!m_torrent_file->is_i2p() || settings().allow_i2p_mixed))) { if (m_ses.m_lsd) lsd_announce(); } } void torrent::stop_announcing() { TORRENT_ASSERT(m_ses.is_network_thread()); if (!m_announcing) return; error_code ec; m_tracker_timer.cancel(ec); m_announcing = false; ptime now = time_now(); for (std::vector::iterator i = m_trackers.begin() , end(m_trackers.end()); i != end; ++i) { i->next_announce = now; i->min_announce = now; } announce_with_tracker(tracker_request::stopped); } void torrent::second_tick(stat& accumulator, int tick_interval_ms) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->tick(); } TORRENT_CATCH (std::exception&) {} } #endif m_time_scaler--; if (m_time_scaler <= 0) { m_time_scaler = 10; if (settings().max_sparse_regions > 0 && m_picker && m_picker->sparse_regions() > settings().max_sparse_regions) { // we have too many sparse regions. Prioritize pieces // that won't introduce new sparse regions // prioritize pieces that will reduce the number of sparse // regions even higher int start = m_picker->cursor(); int end = m_picker->reverse_cursor(); for (int i = start; i < end; ++i) update_sparse_piece_prio(i, start, end); } // ------------------------ // upload shift // ------------------------ // this part will shift downloads // from peers that are seeds and peers // that don't want to download from us // to peers that cannot upload anything // to us. The shifting will make sure // that the torrent's share ratio // will be maintained // if the share ratio is 0 (infinite) // m_available_free_upload isn't used // because it isn't necessary if (ratio() != 0.f) { // accumulate all the free download we get // and add it to the available free upload add_free_upload(collect_free_download( this->begin(), this->end())); // distribute the free upload among the peers m_available_free_upload = distribute_free_upload( this->begin(), this->end(), m_available_free_upload); } } // if we're in upload only mode and we're auto-managed // leave upload mode every 10 minutes hoping that the error // condition has been fixed if (m_upload_mode && m_auto_managed && int(m_upload_mode_time) >= settings().optimistic_disk_retry) { set_upload_mode(false); } if (is_paused()) { // let the stats fade out to 0 accumulator += m_stat; m_stat.second_tick(tick_interval_ms); // if the rate is 0, there's no update because of network transfers if (m_stat.low_pass_upload_rate() > 0 || m_stat.low_pass_download_rate() > 0) state_updated(); return; } if (settings().rate_limit_ip_overhead) { int up_limit = m_bandwidth_channel[peer_connection::upload_channel].throttle(); int down_limit = m_bandwidth_channel[peer_connection::download_channel].throttle(); if (down_limit > 0 && m_stat.download_ip_overhead() >= down_limit && alerts().should_post()) { alerts().post_alert(performance_alert(get_handle() , performance_alert::download_limit_too_low)); } if (up_limit > 0 && m_stat.upload_ip_overhead() >= up_limit && alerts().should_post()) { alerts().post_alert(performance_alert(get_handle() , performance_alert::upload_limit_too_low)); } } int seconds_since_last_tick = 1; if (m_ses.m_tick_residual >= 1000) ++seconds_since_last_tick; if (is_seed()) m_seeding_time += seconds_since_last_tick; if (is_finished()) m_finished_time += seconds_since_last_tick; if (m_upload_mode) m_upload_mode_time += seconds_since_last_tick; m_last_scrape += seconds_since_last_tick; m_active_time += seconds_since_last_tick; m_last_download += seconds_since_last_tick; m_last_upload += seconds_since_last_tick; // ---- TIME CRITICAL PIECES ---- if (!m_time_critical_pieces.empty()) { request_time_critical_pieces(); } // ---- WEB SEEDS ---- // if we have everything we want we don't need to connect to any web-seed if (!is_finished() && !m_web_seeds.empty() && m_files_checked && int(m_connections.size()) < m_max_connections && m_ses.num_connections() < m_ses.settings().connections_limit) { // keep trying web-seeds if there are any // first find out which web seeds we are connected to for (std::list::iterator i = m_web_seeds.begin(); i != m_web_seeds.end();) { std::list::iterator w = i++; if (w->peer_info.connection) continue; if (w->retry > time_now()) continue; if (w->resolving) continue; connect_to_url_seed(w); } } m_swarm_last_seen_complete = m_last_seen_complete; for (peer_iterator i = m_connections.begin(); i != m_connections.end();) { peer_connection* p = *i; ++i; // look for the peer that saw a seed most recently m_swarm_last_seen_complete = (std::max)(p->last_seen_complete(), m_swarm_last_seen_complete); if (!p->ignore_stats()) m_stat += p->statistics(); // updates the peer connection's ul/dl bandwidth // resource requests TORRENT_TRY { p->second_tick(tick_interval_ms); } TORRENT_CATCH (std::exception& e) { TORRENT_DECLARE_DUMMY(std::exception, e); (void)e; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING p->peer_log("*** ERROR %s", e.what()); #endif p->disconnect(errors::no_error, 1); } } if (m_ses.m_alerts.should_post()) m_ses.m_alerts.post_alert(stats_alert(get_handle(), tick_interval_ms, m_stat)); accumulator += m_stat; m_total_uploaded += m_stat.last_payload_uploaded(); m_total_downloaded += m_stat.last_payload_downloaded(); m_stat.second_tick(tick_interval_ms); // if the rate is 0, there's no update because of network transfers if (m_stat.low_pass_upload_rate() > 0 || m_stat.low_pass_download_rate() > 0) state_updated(); } void torrent::recalc_share_mode() { TORRENT_ASSERT(share_mode()); if (is_seed()) return; int pieces_in_torrent = m_torrent_file->num_pieces(); int num_seeds = 0; int num_peers = 0; int num_downloaders = 0; int missing_pieces = 0; int num_interested = 0; for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { peer_connection* p = *i; if (p->is_connecting()) continue; ++num_peers; if (p->is_seed()) { ++num_seeds; continue; } if (p->share_mode()) continue; if ((*i)->is_peer_interested()) ++num_interested; ++num_downloaders; missing_pieces += pieces_in_torrent - p->num_have_pieces(); } if (num_peers == 0) return; if (num_seeds * 100 / num_peers > 50 && (num_peers * 100 / m_max_connections > 90 || num_peers > 20)) { // we are connected to more than 90% seeds (and we're beyond // 90% of the max number of connections). That will // limit our ability to upload. We need more downloaders. // disconnect some seeds so that we don't have more than 50% int to_disconnect = num_seeds - num_peers / 2; std::vector seeds; seeds.reserve(num_seeds); for (std::set::iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) { peer_connection* p = *i; if (p->is_seed()) seeds.push_back(p); } std::random_shuffle(seeds.begin(), seeds.end()); TORRENT_ASSERT(to_disconnect <= int(seeds.size())); for (int i = 0; i < to_disconnect; ++i) seeds[i]->disconnect(errors::upload_upload_connection); } if (num_downloaders == 0) return; // assume that the seeds are about as fast as us. During the time // we can download one piece, and upload one piece, each seed // can upload two pieces. missing_pieces -= 2 * num_seeds; if (missing_pieces <= 0) return; // missing_pieces represents our opportunity to download pieces // and share them more than once each // now, download at least one piece, otherwise download one more // piece if our downloaded (and downloading) pieces is less than 50% // of the uploaded bytes int num_downloaded_pieces = (std::max)(m_picker->num_have() , pieces_in_torrent - m_picker->num_filtered()); if (num_downloaded_pieces * m_torrent_file->piece_length() * settings().share_mode_target > m_total_uploaded && num_downloaded_pieces > 0) return; // don't have more pieces downloading in parallel than 5% of the total // number of pieces we have downloaded if (int(m_picker->get_download_queue().size()) > num_downloaded_pieces / 20) return; // one more important property is that there are enough pieces // that more than one peer wants to download // make sure that there are enough downloaders for the rarest // piece. Go through all pieces, figure out which one is the rarest // and how many peers that has that piece std::vector rarest_pieces; int num_pieces = m_torrent_file->num_pieces(); int rarest_rarity = INT_MAX; bool prio_updated = false; for (int i = 0; i < num_pieces; ++i) { piece_picker::piece_pos const& pp = m_picker->piece_stats(i); if (pp.peer_count == 0) continue; if (pp.filtered() && (pp.have() || pp.downloading)) { m_picker->set_piece_priority(i, 1); prio_updated = true; continue; } // don't count pieces we already have or are downloading if (!pp.filtered() || pp.have()) continue; if (int(pp.peer_count) > rarest_rarity) continue; if (int(pp.peer_count) == rarest_rarity) { rarest_pieces.push_back(i); continue; } rarest_pieces.clear(); rarest_rarity = pp.peer_count; rarest_pieces.push_back(i); } if (prio_updated) m_policy.recalculate_connect_candidates(); // now, rarest_pieces is a list of all pieces that are the rarest ones. // and rarest_rarity is the number of peers that have the rarest pieces // if there's only a single peer that doesn't have the rarest piece // it's impossible for us to download one piece and upload it // twice. i.e. we cannot get a positive share ratio if (num_peers - rarest_rarity < settings().share_mode_target) return; // we might be able to do better than a share ratio of 2 if there are // enough downloaders of the pieces we already have. // TODO: go through the pieces we have and count the total number // of downloaders we have. Only count peers that are interested in us // since some peers might not send have messages for pieces we have // it num_interested == 0, we need to pick a new piece // now, pick one of the rarest pieces to download int pick = random() % rarest_pieces.size(); bool was_finished = is_finished(); m_picker->set_piece_priority(rarest_pieces[pick], 1); update_peer_interest(was_finished); m_policy.recalculate_connect_candidates(); } void torrent::refresh_explicit_cache(int cache_size) { TORRENT_ASSERT(m_ses.is_network_thread()); if (!ready_for_connections()) return; // rotate the cached pieces // add blocks_per_piece / 2 in order to round to closest whole piece int blocks_per_piece = m_torrent_file->piece_length() / block_size(); int num_cache_pieces = (cache_size + blocks_per_piece / 2) / blocks_per_piece; if (num_cache_pieces > m_torrent_file->num_pieces()) num_cache_pieces = m_torrent_file->num_pieces(); std::vector avail_vec; if (has_picker()) { m_picker->get_availability(avail_vec); } else { // we don't keep track of availability, do it the expensive way // do a linear search from the first piece for (int i = 0; i < m_torrent_file->num_pieces(); ++i) { int availability = 0; if (!have_piece(i)) { avail_vec.push_back(INT_MAX); continue; } for (const_peer_iterator j = this->begin(); j != this->end(); ++j) if ((*j)->has_piece(i)) ++availability; avail_vec.push_back(availability); } } // now pick the num_cache_pieces rarest pieces from avail_vec std::vector > pieces(m_torrent_file->num_pieces()); for (int i = 0; i < m_torrent_file->num_pieces(); ++i) { pieces[i].second = i; if (!have_piece(i)) pieces[i].first = INT_MAX; else pieces[i].first = avail_vec[i]; } // decrease the availability of the pieces that are // already in the read cache, to move them closer to // the beginning of the pieces list, and more likely // to be included in this round of cache pieces std::vector ret; m_ses.m_disk_thread.get_cache_info(info_hash(), ret); // remove write cache entries ret.erase(std::remove_if(ret.begin(), ret.end() , boost::bind(&cached_piece_info::kind, _1) == cached_piece_info::write_cache) , ret.end()); for (std::vector::iterator i = ret.begin() , end(ret.end()); i != end; ++i) { --pieces[i->piece].first; } std::random_shuffle(pieces.begin(), pieces.end()); std::stable_sort(pieces.begin(), pieces.end() , boost::bind(&std::pair::first, _1) < boost::bind(&std::pair::first, _2)); avail_vec.clear(); for (int i = 0; i < num_cache_pieces; ++i) { if (pieces[i].first == INT_MAX) break; avail_vec.push_back(pieces[i].second); } if (!avail_vec.empty()) { // the number of pieces to cache for this torrent is proportional // the number of peers it has, divided by the total number of peers. // Each peer gets an equal share of the cache avail_vec.resize((std::min)(num_cache_pieces, int(avail_vec.size()))); for (std::vector::iterator i = avail_vec.begin() , end(avail_vec.end()); i != end; ++i) filesystem().async_cache(*i, boost::bind(&torrent::on_disk_cache_complete , shared_from_this(), _1, _2)); } } void torrent::get_suggested_pieces(std::vector& s) const { TORRENT_ASSERT(m_ses.is_network_thread()); if (settings().suggest_mode == session_settings::no_piece_suggestions) { s.clear(); return; } std::vector ret; m_ses.m_disk_thread.get_cache_info(info_hash(), ret); // remove write cache entries ret.erase(std::remove_if(ret.begin(), ret.end() , boost::bind(&cached_piece_info::kind, _1) == cached_piece_info::write_cache) , ret.end()); // sort by how new the cached entry is, new pieces first std::sort(ret.begin(), ret.end() , boost::bind(&cached_piece_info::last_use, _1) < boost::bind(&cached_piece_info::last_use, _2)); // cut off the oldest pieces that we don't want to suggest // if we have an explicit cache, it's much more likely to // stick around, so we should suggest all pieces int num_pieces_to_suggest = int(ret.size()); if (num_pieces_to_suggest == 0) return; if (!settings().explicit_read_cache) num_pieces_to_suggest = (std::max)(1, int(ret.size() / 2)); ret.resize(num_pieces_to_suggest); std::transform(ret.begin(), ret.end(), std::back_inserter(s) , boost::bind(&cached_piece_info::piece, _1)); } void torrent::add_stats(stat const& s) { TORRENT_ASSERT(m_ses.is_network_thread()); // these stats are propagated to the session // stats the next time second_tick is called m_stat += s; } void torrent::request_time_critical_pieces() { TORRENT_ASSERT(m_ses.is_network_thread()); // build a list of peers and sort it by download_queue_time // we use this sorted list to determine which peer we should // request a block from. The higher up a peer is in the list, // the sooner we will fully download the block we request. std::vector peers; peers.reserve(m_connections.size()); std::remove_copy_if(m_connections.begin(), m_connections.end() , std::back_inserter(peers), !boost::bind(&peer_connection::can_request_time_critical, _1)); std::sort(peers.begin(), peers.end() , boost::bind(&peer_connection::download_queue_time, _1, 16*1024) < boost::bind(&peer_connection::download_queue_time, _2, 16*1024)); std::set peers_with_requests; std::vector interesting_blocks; std::vector backup1; std::vector backup2; std::vector ignore; // peers that should be temporarily ignored for a specific piece // in order to give priority to other peers. They should be used for // subsequent pieces, so they are stored in this vector until the // piece is done std::vector ignore_peers; ptime now = time_now_hires(); // now, iterate over all time critical pieces, in order of importance, and // request them from the peers, in order of responsiveness. i.e. request // the most time critical pieces from the fastest peers. for (std::deque::iterator i = m_time_critical_pieces.begin() , end(m_time_critical_pieces.end()); i != end; ++i) { if (peers.empty()) break; // the +1000 is to compensate for the fact that we only call this functions // once per second, so if we need to request it 500 ms from now, we should request // it right away if (i != m_time_critical_pieces.begin() && i->deadline > now + milliseconds(m_average_piece_time + m_piece_time_deviation * 4 + 1000)) { // don't request pieces whose deadline is too far in the future // this is one of the termination conditions. We don't want to // send requests for all pieces in the torrent right away break; } piece_picker::downloading_piece pi; m_picker->piece_info(i->piece, pi); int free_to_request = m_picker->blocks_in_piece(i->piece) - pi.finished - pi.writing - pi.requested; if (free_to_request == 0) { // every block in this piece is already requested // there's no need to consider this piece, unless it // appears to be stalled. if (pi.requested == 0 || i->last_requested + milliseconds(m_average_piece_time) > now) { // if requested is 0, it meants all blocks have been received, and // we're just waiting for it to flush them to disk. // if last_requested is recent enough, we should give it some // more time break; } } // loop until every block has been requested from this piece (i->piece) do { // pick the peer with the lowest download_queue_time that has i->piece std::vector::iterator p = std::find_if(peers.begin(), peers.end() , boost::bind(&peer_connection::has_piece, _1, i->piece)); // obviously we'll have to skip it if we don't have a peer that has this piece if (p == peers.end()) break; peer_connection& c = **p; interesting_blocks.clear(); backup1.clear(); backup2.clear(); // specifically request blocks with no affinity towards fast or slow // pieces. If we would, the picked block might end up in one of // the backup lists m_picker->add_blocks(i->piece, c.get_bitfield(), interesting_blocks , backup1, backup2, 1, 0, c.peer_info_struct() , ignore, piece_picker::none, 0); std::vector const& rq = c.request_queue(); std::vector const& dq = c.download_queue(); bool added_request = false; if (!interesting_blocks.empty()) { bool already_requested = std::find_if(dq.begin(), dq.end() , has_block(interesting_blocks.front())) != dq.end(); if (already_requested) { // if the piece is stalled, we may end up picking a block // that we've already requested from this peer. If so, we should // simply disregard this peer from this piece, since this peer // is likely to be causing the stall. We should request it // from the next peer in the list ignore_peers.push_back(*p); peers.erase(p); continue; } bool already_in_queue = std::find_if(rq.begin(), rq.end() , has_block(interesting_blocks.front())) != rq.end(); if (already_in_queue) { c.make_time_critical(interesting_blocks.front()); added_request = true; } else { if (!c.add_request(interesting_blocks.front(), peer_connection::req_time_critical)) { peers.erase(p); continue; } added_request = true; } } if (added_request) { peers_with_requests.insert(peers_with_requests.begin(), &c); if (i->first_requested == min_time()) i->first_requested = now; if (!c.can_request_time_critical()) { peers.erase(p); } else { // resort p, since it will have a higher download_queue_time now while (p != peers.end()-1 && (*p)->download_queue_time() > (*(p+1))->download_queue_time()) { std::iter_swap(p, p+1); ++p; } } } // TODO: 2 will pick_pieces ever return an empty set? } while (!interesting_blocks.empty()); peers.insert(peers.begin(), ignore_peers.begin(), ignore_peers.end()); ignore_peers.clear(); } // commit all the time critical requests for (std::set::iterator i = peers_with_requests.begin() , end(peers_with_requests.end()); i != end; ++i) { (*i)->send_block_requests(); } } std::set torrent::web_seeds(web_seed_entry::type_t type) const { TORRENT_ASSERT(m_ses.is_network_thread()); std::set ret; for (std::list::const_iterator i = m_web_seeds.begin() , end(m_web_seeds.end()); i != end; ++i) { if (i->type != type) continue; ret.insert(i->url); } return ret; } void torrent::remove_web_seed(std::string const& url, web_seed_entry::type_t type) { std::list::iterator i = std::find_if(m_web_seeds.begin(), m_web_seeds.end() , (boost::bind(&web_seed_entry::url, _1) == url && boost::bind(&web_seed_entry::type, _1) == type)); if (i != m_web_seeds.end()) remove_web_seed(i); } void torrent::disconnect_web_seed(peer_connection* p) { std::list::iterator i = std::find_if(m_web_seeds.begin(), m_web_seeds.end() , (boost::bind(&policy::peer::connection, boost::bind(&web_seed_entry::peer_info, _1)) == p)); // this happens if the web server responded with a redirect // or with something incorrect, so that we removed the web seed // immediately, before we disconnected if (i == m_web_seeds.end()) return; TORRENT_ASSERT(i->resolving == false); #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("disconnect web seed: \"%s\"", i->url.c_str()); #endif TORRENT_ASSERT(i->peer_info.connection); i->peer_info.connection = 0; } void torrent::remove_web_seed(peer_connection* p) { std::list::iterator i = std::find_if(m_web_seeds.begin(), m_web_seeds.end() , (boost::bind(&policy::peer::connection, boost::bind(&web_seed_entry::peer_info, _1)) == p)); TORRENT_ASSERT(i != m_web_seeds.end()); if (i == m_web_seeds.end()) return; p->set_peer_info(0); if (has_picker()) picker().clear_peer(&i->peer_info); m_web_seeds.erase(i); } void torrent::retry_web_seed(peer_connection* p, int retry) { TORRENT_ASSERT(m_ses.is_network_thread()); std::list::iterator i = std::find_if(m_web_seeds.begin(), m_web_seeds.end() , (boost::bind(&policy::peer::connection, boost::bind(&web_seed_entry::peer_info, _1)) == p)); TORRENT_ASSERT(i != m_web_seeds.end()); if (i == m_web_seeds.end()) return; if (retry == 0) retry = m_ses.settings().urlseed_wait_retry; i->retry = time_now() + seconds(retry); } bool torrent::try_connect_peer() { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(want_more_peers()); bool ret = m_policy.connect_one_peer(m_ses.session_time()); return ret; } void torrent::add_peer(tcp::endpoint const& adr, int source) { TORRENT_ASSERT(m_ses.is_network_thread()); peer_id id(0); m_policy.add_peer(adr, id, source, 0); state_updated(); } void torrent::async_verify_piece(int piece_index, boost::function const& f) { TORRENT_ASSERT(m_ses.is_network_thread()); // INVARIANT_CHECK; TORRENT_ASSERT(m_storage); TORRENT_ASSERT(m_storage->refcount() > 0); TORRENT_ASSERT(piece_index >= 0); TORRENT_ASSERT(piece_index < m_torrent_file->num_pieces()); TORRENT_ASSERT(piece_index < (int)m_picker->num_pieces()); TORRENT_ASSERT(!m_picker || !m_picker->have_piece(piece_index)); #ifdef TORRENT_DEBUG if (m_picker) { int blocks_in_piece = m_picker->blocks_in_piece(piece_index); for (int i = 0; i < blocks_in_piece; ++i) { TORRENT_ASSERT(m_picker->num_peers(piece_block(piece_index, i)) == 0); } } #endif m_storage->async_hash(piece_index, boost::bind(&torrent::on_piece_verified , shared_from_this(), _1, _2, f)); #if defined TORRENT_DEBUG && !defined TORRENT_DISABLE_INVARIANT_CHECKS check_invariant(); #endif } void torrent::on_piece_verified(int ret, disk_io_job const& j , boost::function f) { TORRENT_ASSERT(m_ses.is_network_thread()); // return value: // 0: success, piece passed hash check // -1: disk failure // -2: hash check failed state_updated(); if (ret == -1) handle_disk_error(j); f(ret, j.post_flags); } tcp::endpoint torrent::current_tracker() const { return m_tracker_address; } announce_entry* torrent::find_tracker(tracker_request const& r) { std::vector::iterator i = std::find_if( m_trackers.begin(), m_trackers.end() , boost::bind(&announce_entry::url, _1) == r.url); if (i == m_trackers.end()) return 0; return &*i; } #if !TORRENT_NO_FPU void torrent::file_progress(std::vector& fp) const { fp.clear(); if (!valid_metadata()) return; fp.resize(m_torrent_file->num_files(), 1.f); if (is_seed()) return; std::vector progress; file_progress(progress); for (int i = 0; i < m_torrent_file->num_files(); ++i) { file_entry const& f = m_torrent_file->file_at(i); if (f.size == 0) fp[i] = 1.f; else fp[i] = float(progress[i]) / f.size; } } #endif void torrent::file_progress(std::vector& fp, int flags) const { if (!valid_metadata()) { fp.clear(); return; } fp.resize(m_torrent_file->num_files(), 0); if (flags & torrent_handle::piece_granularity) { std::copy(m_file_progress.begin(), m_file_progress.end(), fp.begin()); return; } if (is_seed()) { for (int i = 0; i < m_torrent_file->num_files(); ++i) fp[i] = m_torrent_file->files().at(i).size; return; } TORRENT_ASSERT(has_picker()); for (int i = 0; i < m_torrent_file->num_files(); ++i) { // zero sized files are considered // 100% done all the time fp[i] = 0; } } void torrent::set_state(torrent_status::state_t s) { TORRENT_ASSERT(m_ses.is_network_thread()); #if defined TORRENT_DEBUG || TORRENT_RELEASE_ASSERTS if (s != torrent_status::checking_files && s != torrent_status::queued_for_checking) { // the only valid transition away from queued_for_checking // is to checking_files. One exception is to finished // in case all the files are marked with priority 0 if (m_queued_for_checking) { std::vector pieces; m_picker->piece_priorities(pieces); // make sure all pieces have priority 0 TORRENT_ASSERT(std::accumulate(pieces.begin(), pieces.end(), 0) == 0); } } if (s == torrent_status::seeding) TORRENT_ASSERT(is_seed()); if (s == torrent_status::seeding) TORRENT_ASSERT(is_seed()); if (s == torrent_status::finished) TORRENT_ASSERT(is_finished()); if (s == torrent_status::downloading && m_state == torrent_status::finished) TORRENT_ASSERT(!is_finished()); #endif if (int(m_state) == s) return; if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(state_changed_alert(get_handle() , s, (torrent_status::state_t)m_state)); } if (s == torrent_status::finished && m_ses.m_alerts.should_post()) { alerts().post_alert(torrent_finished_alert( get_handle())); } m_state = s; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING debug_log("set_state() %d", m_state); #endif update_guage(); state_updated(); #ifndef TORRENT_DISABLE_EXTENSIONS for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->on_state(m_state); } TORRENT_CATCH (std::exception&) {} } #endif } #ifndef TORRENT_DISABLE_EXTENSIONS void torrent::notify_extension_add_peer(tcp::endpoint const& ip , int src, int flags) { for (extension_list_t::iterator i = m_extensions.begin() , end(m_extensions.end()); i != end; ++i) { TORRENT_TRY { (*i)->on_add_peer(ip, src, flags); } TORRENT_CATCH (std::exception&) {} } } #endif void torrent::state_updated() { // if this fails, this function is probably called // from within the torrent constructor, which it // shouldn't be. Whichever function ends up calling // this should probably be moved to torrent::start() TORRENT_ASSERT(shared_from_this()); // we can't call state_updated() while the session // is building the status update alert TORRENT_ASSERT(!m_ses.m_posting_torrent_updates); // we're either not subscribing to this torrent, or // it has already been updated this round, no need to // add it to the list twice if (!m_state_subscription) return; if (m_in_state_updates) { TORRENT_ASSERT(m_ses.in_state_updates(shared_from_this())); return; } m_ses.add_to_update_queue(shared_from_this()); m_in_state_updates = true; } void torrent::status(torrent_status* st, boost::uint32_t flags) { INVARIANT_CHECK; ptime now = time_now(); st->handle = get_handle(); st->info_hash = info_hash(); if (flags & torrent_handle::query_name) st->name = name(); if (flags & torrent_handle::query_save_path) st->save_path = save_path(); if (flags & torrent_handle::query_torrent_file) st->torrent_file = m_torrent_file; st->listen_port = 0; #ifdef TORRENT_USE_OPENSSL if (is_ssl_torrent()) st->listen_port = m_ses.ssl_listen_port(); #endif st->has_incoming = m_has_incoming; if (m_error) st->error = convert_from_native(m_error.message()) + ": " + m_error_file; st->seed_mode = m_seed_mode; st->added_time = m_added_time; st->completed_time = m_completed_time; st->last_scrape = m_last_scrape; st->share_mode = m_share_mode; st->upload_mode = m_upload_mode; st->up_bandwidth_queue = 0; st->down_bandwidth_queue = 0; st->priority = m_priority; st->num_peers = int(m_connections.size()) - m_num_connecting; st->list_peers = m_policy.num_peers(); st->list_seeds = m_policy.num_seeds(); st->connect_candidates = m_policy.num_connect_candidates(); st->seed_rank = seed_rank(settings()); st->all_time_upload = m_total_uploaded; st->all_time_download = m_total_downloaded; // activity time st->finished_time = m_finished_time; st->active_time = m_active_time; st->seeding_time = m_seeding_time; st->time_since_upload = m_last_upload; st->time_since_download = m_last_download; st->storage_mode = (storage_mode_t)m_storage_mode; st->num_complete = (m_complete == 0xffffff) ? -1 : m_complete; st->num_incomplete = (m_incomplete == 0xffffff) ? -1 : m_incomplete; st->paused = is_torrent_paused(); st->auto_managed = m_auto_managed; st->sequential_download = m_sequential_download; st->is_seeding = is_seed(); st->is_finished = is_finished(); st->super_seeding = m_super_seeding; st->has_metadata = valid_metadata(); /* [MF] bytes_done(*st, flags & torrent_handle::query_accurate_download_counters); TORRENT_ASSERT(st->total_wanted_done >= 0); TORRENT_ASSERT(st->total_done >= st->total_wanted_done); */ // payload transfer st->total_payload_download = m_stat.total_payload_download(); st->total_payload_upload = m_stat.total_payload_upload(); // total transfer st->total_download = m_stat.total_payload_download() + m_stat.total_protocol_download(); st->total_upload = m_stat.total_payload_upload() + m_stat.total_protocol_upload(); // failed bytes st->total_failed_bytes = m_total_failed_bytes; st->total_redundant_bytes = m_total_redundant_bytes; // transfer rate st->download_rate = m_stat.download_rate(); st->upload_rate = m_stat.upload_rate(); st->download_payload_rate = m_stat.download_payload_rate(); st->upload_payload_rate = m_stat.upload_payload_rate(); if (m_waiting_tracker && !is_paused()) st->next_announce = boost::posix_time::seconds( total_seconds(next_announce() - now)); else st->next_announce = boost::posix_time::seconds(0); if (st->next_announce.is_negative()) st->next_announce = boost::posix_time::seconds(0); st->announce_interval = boost::posix_time::seconds(0); st->current_tracker.clear(); if (m_last_working_tracker >= 0) { TORRENT_ASSERT(m_last_working_tracker < int(m_trackers.size())); st->current_tracker = m_trackers[m_last_working_tracker].url; } else { std::vector::const_iterator i; for (i = m_trackers.begin(); i != m_trackers.end(); ++i) { if (!i->updating) continue; st->current_tracker = i->url; break; } } if ((flags & torrent_handle::query_verified_pieces)) { st->verified_pieces = m_verified; } st->num_uploads = m_num_uploads; st->uploads_limit = m_max_uploads == (1<<24)-1 ? -1 : m_max_uploads; st->num_connections = int(m_connections.size()); st->connections_limit = m_max_connections == (1<<24)-1 ? -1 : m_max_connections; // if we don't have any metadata, stop here st->queue_position = queue_position(); st->need_save_resume = need_save_resume_data(); st->ip_filter_applies = m_apply_ip_filter; st->state = (torrent_status::state_t)m_state; if (!valid_metadata()) { st->state = torrent_status::downloading_metadata; st->progress_ppm = m_progress_ppm; #if !TORRENT_NO_FPU st->progress = m_progress_ppm / 1000000.f; #endif st->block_size = 0; return; } st->block_size = block_size(); if (m_state == torrent_status::checking_files) { st->progress_ppm = m_progress_ppm; #if !TORRENT_NO_FPU st->progress = m_progress_ppm / 1000000.f; #endif } else if (st->total_wanted == 0) { st->progress_ppm = 1000000; st->progress = 1.f; } else { st->progress_ppm = st->total_wanted_done * 1000000 / st->total_wanted; #if !TORRENT_NO_FPU st->progress = st->progress_ppm / 1000000.f; #endif } if (has_picker() && (flags & torrent_handle::query_pieces)) { st->sparse_regions = m_picker->sparse_regions(); int num_pieces = m_picker->num_pieces(); st->pieces.resize(num_pieces, false); for (int i = 0; i < num_pieces; ++i) if (m_picker->have_piece(i)) st->pieces.set_bit(i); } else if (is_seed()) { int num_pieces = m_torrent_file->num_pieces(); st->pieces.resize(num_pieces, true); } st->num_pieces = num_have(); st->last_have = last_have(); st->num_seeds = num_seeds(); if ((flags & torrent_handle::query_distributed_copies) && m_picker.get()) { boost::tie(st->distributed_full_copies, st->distributed_fraction) = m_picker->distributed_copies(); #if TORRENT_NO_FPU st->distributed_copies = -1.f; #else st->distributed_copies = st->distributed_full_copies + float(st->distributed_fraction) / 1000; #endif } else { st->distributed_full_copies = -1; st->distributed_fraction = -1; st->distributed_copies = -1.f; } st->last_seen_complete = m_swarm_last_seen_complete; } void torrent::add_redundant_bytes(int b, torrent::wasted_reason_t reason) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(b > 0); m_total_redundant_bytes += b; m_ses.add_redundant_bytes(b, reason); // TORRENT_ASSERT(m_total_redundant_bytes + m_total_failed_bytes // <= m_stat.total_payload_download()); } void torrent::add_failed_bytes(int b) { TORRENT_ASSERT(m_ses.is_network_thread()); TORRENT_ASSERT(b > 0); m_total_failed_bytes += b; m_ses.add_failed_bytes(b); // TORRENT_ASSERT(m_total_redundant_bytes + m_total_failed_bytes // <= m_stat.total_payload_download()); } int torrent::num_seeds() const { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; int ret = 0; for (std::set::const_iterator i = m_connections.begin() , end(m_connections.end()); i != end; ++i) if ((*i)->is_seed()) ++ret; return ret; } void torrent::tracker_request_error(tracker_request const& r , int response_code, error_code const& ec, const std::string& msg , int retry_interval) { TORRENT_ASSERT(m_ses.is_network_thread()); INVARIANT_CHECK; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("*** tracker error: (%d) %s %s", ec.value(), ec.message().c_str(), msg.c_str()); #endif if (r.kind == tracker_request::announce_request) { announce_entry* ae = find_tracker(r); if (ae) { ae->failed(settings(), retry_interval); ae->last_error = ec; ae->message = msg; int tracker_index = ae - &m_trackers[0]; #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING debug_log("*** increment tracker fail count [%d]", ae->fails); #endif // never talk to this tracker again if (response_code == 410) ae->fail_limit = 1; deprioritize_tracker(tracker_index); } if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(tracker_error_alert(get_handle() , ae?ae->fails:0, response_code, r.url, ec, msg)); } } else if (r.kind == tracker_request::scrape_request) { if (response_code == 410) { // never talk to this tracker again announce_entry* ae = find_tracker(r); if (ae) ae->fail_limit = 1; } if (m_ses.m_alerts.should_post()) { m_ses.m_alerts.post_alert(scrape_failed_alert(get_handle(), r.url, ec)); } } // announce to the next working tracker if ((!m_abort && !is_paused()) || r.event == tracker_request::stopped) announce_with_tracker(r.event); update_tracker_timer(time_now()); } #if defined TORRENT_VERBOSE_LOGGING || defined TORRENT_LOGGING || defined TORRENT_ERROR_LOGGING void torrent::debug_log(const char* fmt, ...) const { if (!m_ses.m_logger) return; va_list v; va_start(v, fmt); char usr[1024]; vsnprintf(usr, sizeof(usr), fmt, v); va_end(v); char buf[1280]; snprintf(buf, sizeof(buf), "%s: %s: %s\n", time_now_string() , to_hex(info_hash().to_string()).substr(0, 6).c_str(), usr); //(*m_ses.m_logger) << buf; // [MF] printf("%s", buf); } #endif }