twisterp2pnetworkbittorrentblockchainmicrobloggingipv6social-networkdhtdecentralizedtwister-servertwister-ipv6twister-coretwisterarmyp2p-network
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
658 lines
18 KiB
658 lines
18 KiB
/* |
|
|
|
Copyright (c) 2008-2012, Arvid Norberg |
|
All rights reserved. |
|
|
|
Redistribution and use in source and binary forms, with or without |
|
modification, are permitted provided that the following conditions |
|
are met: |
|
|
|
* Redistributions of source code must retain the above copyright |
|
notice, this list of conditions and the following disclaimer. |
|
* Redistributions in binary form must reproduce the above copyright |
|
notice, this list of conditions and the following disclaimer in |
|
the documentation and/or other materials provided with the distribution. |
|
* Neither the name of the author nor the names of its |
|
contributors may be used to endorse or promote products derived |
|
from this software without specific prior written permission. |
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
|
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
|
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
|
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
|
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
|
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
|
POSSIBILITY OF SUCH DAMAGE. |
|
|
|
*/ |
|
|
|
#include "libtorrent/create_torrent.hpp" |
|
#include "libtorrent/file_pool.hpp" |
|
#include "libtorrent/storage.hpp" |
|
#include "libtorrent/escape_string.hpp" |
|
#include "libtorrent/torrent_info.hpp" // for merkle_*() |
|
|
|
#include <boost/bind.hpp> |
|
#include <boost/next_prior.hpp> |
|
|
|
#include <sys/types.h> |
|
#include <sys/stat.h> |
|
|
|
#define MAX_SYMLINK_PATH 200 |
|
|
|
namespace libtorrent |
|
{ |
|
|
|
namespace detail |
|
{ |
|
int get_file_attributes(std::string const& p) |
|
{ |
|
#ifdef TORRENT_WINDOWS |
|
|
|
#if TORRENT_USE_WSTRING |
|
std::wstring path = convert_to_wstring(p); |
|
DWORD attr = GetFileAttributesW(path.c_str()); |
|
#else |
|
std::string path = convert_to_native(p); |
|
DWORD attr = GetFileAttributesA(path.c_str()); |
|
#endif // TORRENT_USE_WSTRING |
|
if (attr == INVALID_FILE_ATTRIBUTES) return 0; |
|
if (attr & FILE_ATTRIBUTE_HIDDEN) return file_storage::attribute_hidden; |
|
return 0; |
|
#else |
|
struct stat s; |
|
if (lstat(convert_to_native(p).c_str(), &s) < 0) return 0; |
|
int file_attr = 0; |
|
if (s.st_mode & S_IXUSR) |
|
file_attr += file_storage::attribute_executable; |
|
if (S_ISLNK(s.st_mode)) |
|
file_attr += file_storage::attribute_symlink; |
|
return file_attr; |
|
#endif |
|
} |
|
|
|
#ifndef TORRENT_WINDOWS |
|
std::string get_symlink_path_impl(char const* path) |
|
{ |
|
char buf[MAX_SYMLINK_PATH]; |
|
std::string f = convert_to_native(path); |
|
int char_read = readlink(f.c_str(),buf,MAX_SYMLINK_PATH); |
|
if (char_read < 0) return ""; |
|
if (char_read < MAX_SYMLINK_PATH) buf[char_read] = 0; |
|
else buf[0] = 0; |
|
return convert_from_native(buf); |
|
} |
|
#endif |
|
|
|
std::string get_symlink_path(std::string const& p) |
|
{ |
|
#if defined TORRENT_WINDOWS |
|
return ""; |
|
#else |
|
std::string path = convert_to_native(p); |
|
return get_symlink_path_impl(p.c_str()); |
|
#endif |
|
} |
|
|
|
void add_files_impl(file_storage& fs, std::string const& p |
|
, std::string const& l, boost::function<bool(std::string)> pred, boost::uint32_t flags) |
|
{ |
|
std::string f = combine_path(p, l); |
|
if (!pred(f)) return; |
|
error_code ec; |
|
file_status s; |
|
stat_file(f, &s, ec, (flags & create_torrent::symlinks) ? dont_follow_links : 0); |
|
if (ec) return; |
|
|
|
// recurse into directories |
|
bool recurse = (s.mode & file_status::directory) != 0; |
|
|
|
// if the file is not a link or we're following links, and it's a directory |
|
// only then should we recurse |
|
#ifndef TORRENT_WINDOWS |
|
if ((s.mode & file_status::link) && (flags & create_torrent::symlinks)) |
|
recurse = false; |
|
#endif |
|
|
|
if (recurse) |
|
{ |
|
for (directory i(f, ec); !i.done(); i.next(ec)) |
|
{ |
|
std::string leaf = i.file(); |
|
if (ignore_subdir(leaf)) continue; |
|
add_files_impl(fs, p, combine_path(l, leaf), pred, flags); |
|
} |
|
} |
|
else |
|
{ |
|
// #error use the fields from s |
|
int file_flags = get_file_attributes(f); |
|
|
|
// mask all bits to check if the file is a symlink |
|
if ((file_flags & file_storage::attribute_symlink) |
|
&& (flags & create_torrent::symlinks)) |
|
{ |
|
std::string sym_path = get_symlink_path(f); |
|
fs.add_file(l, 0, file_flags, s.mtime, sym_path); |
|
} |
|
else |
|
{ |
|
fs.add_file(l, s.file_size, file_flags, s.mtime); |
|
} |
|
} |
|
} |
|
} |
|
|
|
struct piece_holder |
|
{ |
|
piece_holder(int bytes): m_piece(page_aligned_allocator::malloc(bytes)) {} |
|
~piece_holder() { page_aligned_allocator::free(m_piece); } |
|
char* bytes() { return m_piece; } |
|
private: |
|
char* m_piece; |
|
}; |
|
|
|
#if TORRENT_USE_WSTRING |
|
void set_piece_hashes(create_torrent& t, std::wstring const& p |
|
, boost::function<void(int)> const& f, error_code& ec) |
|
{ |
|
/* |
|
file_pool fp; |
|
std::string utf8; |
|
wchar_utf8(p, utf8); |
|
#if TORRENT_USE_UNC_PATHS |
|
utf8 = canonicalize_path(utf8); |
|
#endif |
|
boost::scoped_ptr<storage_interface> st( |
|
default_storage_constructor(const_cast<file_storage&>(t.files()), 0, utf8, fp |
|
, std::vector<boost::uint8_t>())); |
|
|
|
// calculate the hash for all pieces |
|
int num = t.num_pieces(); |
|
std::vector<char> buf(t.piece_length()); |
|
for (int i = 0; i < num; ++i) |
|
{ |
|
// read hits the disk and will block. Progress should |
|
// be updated in between reads |
|
st->read(&buf[0], i, 0, t.piece_size(i)); |
|
if (st->error()) |
|
{ |
|
ec = st->error(); |
|
return; |
|
} |
|
hasher h(&buf[0], t.piece_size(i)); |
|
t.set_hash(i, h.final()); |
|
f(i); |
|
} |
|
*/ |
|
} |
|
#endif |
|
|
|
void set_piece_hashes(create_torrent& t, std::string const& p |
|
, boost::function<void(int)> f, error_code& ec) |
|
{ |
|
/* |
|
file_pool fp; |
|
#if TORRENT_USE_UNC_PATHS |
|
std::string path = canonicalize_path(p); |
|
#else |
|
std::string const& path = p; |
|
#endif |
|
|
|
boost::scoped_ptr<storage_interface> st( |
|
default_storage_constructor(const_cast<file_storage&>(t.files()), 0, path, fp |
|
, std::vector<boost::uint8_t>())); |
|
|
|
// if we're calculating file hashes as well, use this hasher |
|
hasher filehash; |
|
int file_idx = 0; |
|
size_type left_in_file = t.files().at(0).size; |
|
|
|
// calculate the hash for all pieces |
|
int num = t.num_pieces(); |
|
piece_holder buf(t.piece_length()); |
|
for (int i = 0; i < num; ++i) |
|
{ |
|
// read hits the disk and will block. Progress should |
|
// be updated in between reads |
|
st->read(buf.bytes(), i, 0, t.piece_size(i)); |
|
if (st->error()) |
|
{ |
|
ec = st->error(); |
|
return; |
|
} |
|
|
|
if (t.should_add_file_hashes()) |
|
{ |
|
int left_in_piece = t.piece_size(i); |
|
int this_piece_size = left_in_piece; |
|
// the number of bytes from this file we just read |
|
while (left_in_piece > 0) |
|
{ |
|
int to_hash_for_file = int((std::min)(size_type(left_in_piece), left_in_file)); |
|
if (to_hash_for_file > 0) |
|
{ |
|
int offset = this_piece_size - left_in_piece; |
|
filehash.update(buf.bytes() + offset, to_hash_for_file); |
|
} |
|
left_in_file -= to_hash_for_file; |
|
left_in_piece -= to_hash_for_file; |
|
if (left_in_file == 0) |
|
{ |
|
if (!t.files().at(file_idx).pad_file) |
|
t.set_file_hash(file_idx, filehash.final()); |
|
filehash.reset(); |
|
file_idx++; |
|
if (file_idx >= t.files().num_files()) break; |
|
left_in_file = t.files().at(file_idx).size; |
|
} |
|
} |
|
} |
|
|
|
hasher h(buf.bytes(), t.piece_size(i)); |
|
t.set_hash(i, h.final()); |
|
f(i); |
|
} |
|
*/ |
|
} |
|
|
|
create_torrent::~create_torrent() {} |
|
|
|
create_torrent::create_torrent(file_storage& fs, int piece_size, int pad_file_limit, int flags, int alignment) |
|
: m_files(fs) |
|
, m_creation_date(time(0)) |
|
, m_multifile(fs.num_files() > 1) |
|
, m_private(false) |
|
, m_merkle_torrent((flags & merkle) != 0) |
|
, m_include_mtime((flags & modification_time) != 0) |
|
, m_include_symlinks((flags & symlinks) != 0) |
|
, m_calculate_file_hashes((flags & calculate_file_hashes) != 0) |
|
{ |
|
TORRENT_ASSERT(fs.num_files() > 0); |
|
|
|
// return instead of crash in release mode |
|
if (fs.num_files() == 0) return; |
|
|
|
if (!m_multifile && has_parent_path(m_files.file_path(*m_files.begin()))) m_multifile = true; |
|
|
|
// a piece_size of 0 means automatic |
|
if (piece_size == 0 && !m_merkle_torrent) |
|
{ |
|
const int target_size = 40 * 1024; |
|
piece_size = int(fs.total_size() / (target_size / 20)); |
|
|
|
int i = 16*1024; |
|
for (; i < 2*1024*1024; i *= 2) |
|
{ |
|
if (piece_size > i) continue; |
|
break; |
|
} |
|
piece_size = i; |
|
} |
|
else if (piece_size == 0 && m_merkle_torrent) |
|
{ |
|
piece_size = 64*1024; |
|
} |
|
|
|
// make sure the size is an even power of 2 |
|
#ifndef NDEBUG |
|
for (int i = 0; i < 32; ++i) |
|
{ |
|
if (piece_size & (1 << i)) |
|
{ |
|
TORRENT_ASSERT((piece_size & ~(1 << i)) == 0); |
|
break; |
|
} |
|
} |
|
#endif |
|
m_files.set_piece_length(piece_size); |
|
if (flags & optimize) |
|
m_files.optimize(pad_file_limit, alignment); |
|
m_files.set_num_pieces(static_cast<int>( |
|
(m_files.total_size() + m_files.piece_length() - 1) / m_files.piece_length())); |
|
m_piece_hash.resize(m_files.num_pieces()); |
|
} |
|
|
|
create_torrent::create_torrent(torrent_info const& ti) |
|
: m_files(const_cast<file_storage&>(ti.files())) |
|
, m_creation_date(time(0)) |
|
, m_multifile(ti.num_files() > 1) |
|
, m_private(ti.priv()) |
|
, m_merkle_torrent(ti.is_merkle_torrent()) |
|
, m_include_mtime(false) |
|
, m_include_symlinks(false) |
|
, m_calculate_file_hashes(false) |
|
{ |
|
TORRENT_ASSERT(ti.is_valid()); |
|
if (ti.creation_date()) m_creation_date = *ti.creation_date(); |
|
|
|
if (!ti.creator().empty()) set_creator(ti.creator().c_str()); |
|
if (!ti.comment().empty()) set_comment(ti.comment().c_str()); |
|
|
|
torrent_info::nodes_t const& nodes = ti.nodes(); |
|
for (torrent_info::nodes_t::const_iterator i = nodes.begin() |
|
, end(nodes.end()); i != end; ++i) |
|
add_node(*i); |
|
|
|
std::vector<libtorrent::announce_entry> const& trackers = ti.trackers(); |
|
for (std::vector<libtorrent::announce_entry>::const_iterator i = trackers.begin() |
|
, end(trackers.end()); i != end; ++i) |
|
add_tracker(i->url, i->tier); |
|
|
|
std::vector<web_seed_entry> const& web_seeds = ti.web_seeds(); |
|
for (std::vector<web_seed_entry>::const_iterator i = web_seeds.begin() |
|
, end(web_seeds.end()); i != end; ++i) |
|
{ |
|
if (i->type == web_seed_entry::url_seed) |
|
add_url_seed(i->url); |
|
else if (i->type == web_seed_entry::http_seed) |
|
add_http_seed(i->url); |
|
} |
|
|
|
m_piece_hash.resize(m_files.num_pieces()); |
|
for (int i = 0; i < num_pieces(); ++i) set_hash(i, ti.hash_for_piece(i)); |
|
|
|
m_info_dict = bdecode(&ti.metadata()[0], &ti.metadata()[0] + ti.metadata_size()); |
|
m_info_hash = ti.info_hash(); |
|
} |
|
|
|
entry create_torrent::generate() const |
|
{ |
|
TORRENT_ASSERT(m_files.piece_length() > 0); |
|
|
|
entry dict; |
|
|
|
if (m_files.num_files() == 0) |
|
return dict; |
|
|
|
if (!m_urls.empty()) dict["announce"] = m_urls.front().first; |
|
|
|
if (!m_nodes.empty()) |
|
{ |
|
entry& nodes = dict["nodes"]; |
|
entry::list_type& nodes_list = nodes.list(); |
|
for (nodes_t::const_iterator i = m_nodes.begin() |
|
, end(m_nodes.end()); i != end; ++i) |
|
{ |
|
entry::list_type node; |
|
node.push_back(entry(i->first)); |
|
node.push_back(entry(i->second)); |
|
nodes_list.push_back(entry(node)); |
|
} |
|
} |
|
|
|
if (m_urls.size() > 1) |
|
{ |
|
entry trackers(entry::list_t); |
|
entry tier(entry::list_t); |
|
int current_tier = m_urls.front().second; |
|
for (std::vector<announce_entry>::const_iterator i = m_urls.begin(); |
|
i != m_urls.end(); ++i) |
|
{ |
|
if (i->second != current_tier) |
|
{ |
|
current_tier = i->second; |
|
trackers.list().push_back(tier); |
|
tier.list().clear(); |
|
} |
|
tier.list().push_back(entry(i->first)); |
|
} |
|
trackers.list().push_back(tier); |
|
dict["announce-list"] = trackers; |
|
} |
|
|
|
if (!m_comment.empty()) |
|
dict["comment"] = m_comment; |
|
|
|
dict["creation date"] = m_creation_date; |
|
|
|
if (!m_created_by.empty()) |
|
dict["created by"] = m_created_by; |
|
|
|
if (!m_url_seeds.empty()) |
|
{ |
|
if (m_url_seeds.size() == 1) |
|
{ |
|
dict["url-list"] = m_url_seeds.front(); |
|
} |
|
else |
|
{ |
|
entry& list = dict["url-list"]; |
|
for (std::vector<std::string>::const_iterator i |
|
= m_url_seeds.begin(); i != m_url_seeds.end(); ++i) |
|
{ |
|
list.list().push_back(entry(*i)); |
|
} |
|
} |
|
} |
|
|
|
if (!m_http_seeds.empty()) |
|
{ |
|
if (m_http_seeds.size() == 1) |
|
{ |
|
dict["httpseeds"] = m_http_seeds.front(); |
|
} |
|
else |
|
{ |
|
entry& list = dict["httpseeds"]; |
|
for (std::vector<std::string>::const_iterator i |
|
= m_http_seeds.begin(); i != m_http_seeds.end(); ++i) |
|
{ |
|
list.list().push_back(entry(*i)); |
|
} |
|
} |
|
} |
|
|
|
entry& info = dict["info"]; |
|
if (m_info_dict.type() == entry::dictionary_t) |
|
{ |
|
info = m_info_dict; |
|
return dict; |
|
} |
|
|
|
info["name"] = m_files.name(); |
|
|
|
if (!m_root_cert.empty()) |
|
info["ssl-cert"] = m_root_cert; |
|
|
|
if (m_private) info["private"] = 1; |
|
|
|
if (!m_multifile) |
|
{ |
|
file_entry e = m_files.at(0); |
|
if (m_include_mtime) info["mtime"] = e.mtime; |
|
info["length"] = e.size; |
|
if (e.pad_file |
|
|| e.hidden_attribute |
|
|| e.executable_attribute |
|
|| e.symlink_attribute) |
|
{ |
|
std::string& attr = info["attr"].string(); |
|
if (e.pad_file) attr += 'p'; |
|
if (e.hidden_attribute) attr += 'h'; |
|
if (e.executable_attribute) attr += 'x'; |
|
if (m_include_symlinks && e.symlink_attribute) attr += 'l'; |
|
} |
|
if (m_include_symlinks |
|
&& e.symlink_attribute) |
|
{ |
|
entry& sympath_e = info["symlink path"]; |
|
|
|
std::string split = split_path(e.symlink_path); |
|
for (char const* e = split.c_str(); e != 0; e = next_path_element(e)) |
|
sympath_e.list().push_back(entry(e)); |
|
} |
|
if (!m_filehashes.empty()) |
|
{ |
|
info["sha1"] = m_filehashes[0].to_string(); |
|
} |
|
} |
|
else |
|
{ |
|
if (!info.find_key("files")) |
|
{ |
|
entry& files = info["files"]; |
|
|
|
for (file_storage::iterator i = m_files.begin(); |
|
i != m_files.end(); ++i) |
|
{ |
|
files.list().push_back(entry()); |
|
entry& file_e = files.list().back(); |
|
if (m_include_mtime && m_files.mtime(*i)) file_e["mtime"] = m_files.mtime(*i); |
|
file_e["length"] = i->size; |
|
entry& path_e = file_e["path"]; |
|
|
|
TORRENT_ASSERT(has_parent_path(m_files.file_path(*i))); |
|
|
|
std::string split = split_path(m_files.file_path(*i)); |
|
TORRENT_ASSERT(split.c_str() == m_files.name()); |
|
|
|
for (char const* e = next_path_element(split.c_str()); |
|
e != 0; e = next_path_element(e)) |
|
path_e.list().push_back(entry(e)); |
|
|
|
if (i->pad_file |
|
|| i->hidden_attribute |
|
|| i->executable_attribute |
|
|| i->symlink_attribute) |
|
{ |
|
std::string& attr = file_e["attr"].string(); |
|
if (i->pad_file) attr += 'p'; |
|
if (i->hidden_attribute) attr += 'h'; |
|
if (i->executable_attribute) attr += 'x'; |
|
if (m_include_symlinks && i->symlink_attribute) attr += 'l'; |
|
} |
|
if (m_include_symlinks |
|
&& i->symlink_attribute |
|
&& i->symlink_index != -1) |
|
{ |
|
entry& sympath_e = file_e["symlink path"]; |
|
|
|
std::string split = split_path(m_files.symlink(*i)); |
|
for (char const* e = split.c_str(); e != 0; e = next_path_element(e)) |
|
sympath_e.list().push_back(entry(e)); |
|
} |
|
int file_index = i - m_files.begin(); |
|
if (!m_filehashes.empty() && m_filehashes[file_index] != sha1_hash()) |
|
{ |
|
file_e["sha1"] = m_filehashes[file_index].to_string(); |
|
} |
|
} |
|
} |
|
} |
|
|
|
info["piece length"] = m_files.piece_length(); |
|
if (m_merkle_torrent) |
|
{ |
|
int num_leafs = merkle_num_leafs(m_files.num_pieces()); |
|
int num_nodes = merkle_num_nodes(num_leafs); |
|
int first_leaf = num_nodes - num_leafs; |
|
m_merkle_tree.resize(num_nodes); |
|
int num_pieces = m_piece_hash.size(); |
|
for (int i = 0; i < num_pieces; ++i) |
|
m_merkle_tree[first_leaf + i] = m_piece_hash[i]; |
|
sha1_hash filler(0); |
|
for (int i = num_pieces; i < num_leafs; ++i) |
|
m_merkle_tree[first_leaf + i] = filler; |
|
|
|
// now that we have initialized all leaves, build |
|
// each level bottom-up |
|
int level_start = first_leaf; |
|
int level_size = num_leafs; |
|
while (level_start > 0) |
|
{ |
|
int parent = merkle_get_parent(level_start); |
|
for (int i = level_start; i < level_start + level_size; i += 2, ++parent) |
|
{ |
|
hasher h; |
|
h.update((char const*)&m_merkle_tree[i][0], 20); |
|
h.update((char const*)&m_merkle_tree[i+1][0], 20); |
|
m_merkle_tree[parent] = h.final(); |
|
} |
|
level_start = merkle_get_parent(level_start); |
|
level_size /= 2; |
|
} |
|
TORRENT_ASSERT(level_size == 1); |
|
std::string& p = info["root hash"].string(); |
|
p.assign((char const*)&m_merkle_tree[0][0], 20); |
|
} |
|
else |
|
{ |
|
std::string& p = info["pieces"].string(); |
|
|
|
for (std::vector<sha1_hash>::const_iterator i = m_piece_hash.begin(); |
|
i != m_piece_hash.end(); ++i) |
|
{ |
|
p.append((char*)i->begin(), sha1_hash::size); |
|
} |
|
} |
|
|
|
std::vector<char> buf; |
|
bencode(std::back_inserter(buf), info); |
|
m_info_hash = hasher(&buf[0], buf.size()).final(); |
|
|
|
return dict; |
|
|
|
} |
|
|
|
void create_torrent::add_tracker(std::string const& url, int tier) |
|
{ |
|
m_urls.push_back(announce_entry(url, tier)); |
|
|
|
std::sort(m_urls.begin(), m_urls.end() |
|
, boost::bind(&announce_entry::second, _1) < boost::bind(&announce_entry::second, _2)); |
|
} |
|
|
|
void create_torrent::set_root_cert(std::string const& cert) |
|
{ |
|
m_root_cert = cert; |
|
} |
|
|
|
void create_torrent::set_hash(int index, sha1_hash const& h) |
|
{ |
|
TORRENT_ASSERT(index >= 0); |
|
TORRENT_ASSERT(index < (int)m_piece_hash.size()); |
|
m_piece_hash[index] = h; |
|
} |
|
|
|
void create_torrent::set_file_hash(int index, sha1_hash const& h) |
|
{ |
|
TORRENT_ASSERT(index >= 0); |
|
TORRENT_ASSERT(index < (int)m_files.num_files()); |
|
if (m_filehashes.empty()) m_filehashes.resize(m_files.num_files()); |
|
m_filehashes[index] = h; |
|
} |
|
|
|
void create_torrent::add_node(std::pair<std::string, int> const& node) |
|
{ |
|
m_nodes.push_back(node); |
|
} |
|
|
|
void create_torrent::add_url_seed(std::string const& url) |
|
{ |
|
m_url_seeds.push_back(url); |
|
} |
|
|
|
void create_torrent::add_http_seed(std::string const& url) |
|
{ |
|
m_http_seeds.push_back(url); |
|
} |
|
|
|
void create_torrent::set_comment(char const* str) |
|
{ |
|
if (str == 0) m_comment.clear(); |
|
else m_comment = str; |
|
} |
|
|
|
void create_torrent::set_creator(char const* str) |
|
{ |
|
if (str == 0) m_created_by.clear(); |
|
else m_created_by = str; |
|
} |
|
|
|
} |
|
|
|
|