mirror of git://erdgeist.org/opentracker
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
763 lines
26 KiB
763 lines
26 KiB
/* This software was written by Dirk Engling <erdgeist@erdgeist.org> |
|
It is considered beerware. Prost. Skol. Cheers or whatever. |
|
|
|
$id$ */ |
|
|
|
/* System */ |
|
#include <arpa/inet.h> |
|
#include <pthread.h> |
|
#include <stdio.h> |
|
#include <stdlib.h> |
|
#include <string.h> |
|
#include <sys/types.h> |
|
#include <unistd.h> |
|
|
|
/* Libowfat */ |
|
#include "array.h" |
|
#include "byte.h" |
|
#include "case.h" |
|
#include "iob.h" |
|
#include "ip6.h" |
|
#include "scan.h" |
|
|
|
/* Opentracker */ |
|
#include "ot_accesslist.h" |
|
#include "ot_fullscrape.h" |
|
#include "ot_http.h" |
|
#include "ot_iovec.h" |
|
#include "ot_mutex.h" |
|
#include "ot_stats.h" |
|
#include "scan_urlencoded_query.h" |
|
#include "trackerlogic.h" |
|
|
|
#ifdef WANT_NO_AUTO_FREE |
|
#define OT_IOB_INIT iob_init |
|
#else |
|
#define OT_IOB_INIT iob_init_autofree |
|
#endif |
|
|
|
#define OT_MAXMULTISCRAPE_COUNT 64 |
|
#define OT_BATCH_LIMIT (1024 * 1024 * 16) |
|
extern char *g_redirecturl; |
|
|
|
char *g_stats_path; |
|
ssize_t g_stats_path_len; |
|
|
|
enum { SUCCESS_HTTP_HEADER_LENGTH = 80, SUCCESS_HTTP_SIZE_OFF = 17 }; |
|
|
|
static void http_senddata(const int64 sock, struct ot_workstruct *ws) { |
|
struct http_data *cookie = io_getcookie(sock); |
|
ssize_t written_size; |
|
|
|
if (!cookie) { |
|
io_close(sock); |
|
return; |
|
} |
|
|
|
/* whoever sends data is not interested in its input-array */ |
|
if (ws->keep_alive && ws->header_size != ws->request_size) { |
|
size_t rest = ws->request_size - ws->header_size; |
|
if (array_start(&cookie->request)) { |
|
memmove(array_start(&cookie->request), ws->request + ws->header_size, rest); |
|
array_truncate(&cookie->request, 1, rest); |
|
} else |
|
array_catb(&cookie->request, ws->request + ws->header_size, rest); |
|
} else |
|
array_reset(&cookie->request); |
|
|
|
written_size = write(sock, ws->reply, ws->reply_size); |
|
if ((written_size < 0) || ((written_size == ws->reply_size) && !ws->keep_alive)) { |
|
array_reset(&cookie->request); |
|
free(cookie); |
|
io_close(sock); |
|
return; |
|
} |
|
|
|
if (written_size < ws->reply_size) { |
|
char *outbuf; |
|
tai6464 t; |
|
|
|
if (!(outbuf = malloc(ws->reply_size - written_size))) { |
|
array_reset(&cookie->request); |
|
free(cookie); |
|
io_close(sock); |
|
return; |
|
} |
|
|
|
memcpy(outbuf, ws->reply + written_size, ws->reply_size - written_size); |
|
if (!cookie->batch) { |
|
cookie->batch = malloc(sizeof(io_batch)); |
|
OT_IOB_INIT(cookie->batch, 0); |
|
cookie->batches = 1; |
|
} |
|
|
|
iob_addbuf_free(cookie->batch, outbuf, ws->reply_size - written_size); |
|
|
|
/* writeable short data sockets just have a tcp timeout */ |
|
if (!ws->keep_alive) { |
|
taia_uint(&t, 0); |
|
io_timeout(sock, t); |
|
io_dontwantread(sock); |
|
} |
|
io_wantwrite(sock); |
|
} |
|
} |
|
|
|
#define HTTPERROR_302 return http_issue_error(sock, ws, CODE_HTTPERROR_302) |
|
#define HTTPERROR_400 return http_issue_error(sock, ws, CODE_HTTPERROR_400) |
|
#define HTTPERROR_400_PARAM return http_issue_error(sock, ws, CODE_HTTPERROR_400_PARAM) |
|
#define HTTPERROR_400_COMPACT return http_issue_error(sock, ws, CODE_HTTPERROR_400_COMPACT) |
|
#define HTTPERROR_400_DOUBLEHASH return http_issue_error(sock, ws, CODE_HTTPERROR_400_PARAM) |
|
#define HTTPERROR_402_NOTMODEST return http_issue_error(sock, ws, CODE_HTTPERROR_402_NOTMODEST) |
|
#define HTTPERROR_403_IP return http_issue_error(sock, ws, CODE_HTTPERROR_403_IP) |
|
#define HTTPERROR_404 return http_issue_error(sock, ws, CODE_HTTPERROR_404) |
|
#define HTTPERROR_500 return http_issue_error(sock, ws, CODE_HTTPERROR_500) |
|
ssize_t http_issue_error(const int64 sock, struct ot_workstruct *ws, int code) { |
|
char *error_code[] = {"302 Found", "400 Invalid Request", "400 Invalid Request", "400 Invalid Request", "402 Payment Required", |
|
"403 Not Modest", "403 Access Denied", "404 Not Found", "500 Internal Server Error"}; |
|
char *title = error_code[code]; |
|
|
|
ws->reply = ws->outbuf; |
|
if (code == CODE_HTTPERROR_302) |
|
ws->reply_size = snprintf(ws->reply, G_OUTBUF_SIZE, "HTTP/1.0 302 Found\r\nContent-Length: 0\r\nLocation: %s\r\n\r\n", g_redirecturl); |
|
else |
|
ws->reply_size = snprintf(ws->reply, G_OUTBUF_SIZE, "HTTP/1.0 %s\r\nContent-Type: text/html\r\nContent-Length: %zd\r\n\r\n<title>%s</title>\n", title, |
|
strlen(title) + 16 - 4, title + 4); |
|
|
|
#ifdef _DEBUG_HTTPERROR |
|
fprintf(stderr, "DEBUG: invalid request was: %s\n", ws->debugbuf); |
|
#endif |
|
stats_issue_event(EVENT_FAILED, FLAG_TCP, code); |
|
http_senddata(sock, ws); |
|
return ws->reply_size = -2; |
|
} |
|
|
|
ssize_t http_sendiovecdata(const int64 sock, struct ot_workstruct *ws, int iovec_entries, struct iovec *iovector, int is_partial) { |
|
struct http_data *cookie = io_getcookie(sock); |
|
io_batch *current; |
|
char *header; |
|
const char *encoding = ""; |
|
int i; |
|
size_t header_size, size = iovec_length(&iovec_entries, (const struct iovec **)&iovector); |
|
tai6464 t; |
|
|
|
/* No cookie? Bad socket. Leave. */ |
|
if (!cookie) { |
|
iovec_free(&iovec_entries, &iovector); |
|
HTTPERROR_500; |
|
} |
|
|
|
/* If this socket collected request in a buffer, free it now */ |
|
array_reset(&cookie->request); |
|
|
|
/* If we came here, wait for the answer is over */ |
|
if (cookie->flag & STRUCT_HTTP_FLAG_WAITINGFORTASK) { |
|
io_dontwantread(sock); |
|
cookie->flag &= ~STRUCT_HTTP_FLAG_WAITINGFORTASK; |
|
} |
|
|
|
if (iovec_entries) { |
|
|
|
if (cookie->flag & STRUCT_HTTP_FLAG_GZIP) |
|
encoding = "Content-Encoding: gzip\r\n"; |
|
else if (cookie->flag & STRUCT_HTTP_FLAG_BZIP2) |
|
encoding = "Content-Encoding: bzip2\r\n"; |
|
|
|
if (!(cookie->flag & STRUCT_HTTP_FLAG_CHUNKED)) |
|
header_size = asprintf(&header, "HTTP/1.0 200 OK\r\nContent-Type: text/plain\r\n%sContent-Length: %zd\r\n\r\n", encoding, size); |
|
else { |
|
if (!(cookie->flag & STRUCT_HTTP_FLAG_CHUNKED_IN_TRANSFER)) { |
|
header_size = |
|
asprintf(&header, "HTTP/1.0 200 OK\r\nContent-Type: application/octet-stream\r\n%sTransfer-Encoding: chunked\r\n\r\n%zx\r\n", encoding, size); |
|
cookie->flag |= STRUCT_HTTP_FLAG_CHUNKED_IN_TRANSFER; |
|
} else |
|
header_size = asprintf(&header, "%zx\r\n", size); |
|
} |
|
if (!header) { |
|
iovec_free(&iovec_entries, &iovector); |
|
HTTPERROR_500; |
|
} |
|
|
|
if (!cookie->batch) { |
|
cookie->batch = malloc(sizeof(io_batch)); |
|
if (!cookie->batch) { |
|
free(header); |
|
iovec_free(&iovec_entries, &iovector); |
|
HTTPERROR_500; |
|
} |
|
OT_IOB_INIT(cookie->batch, 0); |
|
cookie->batches = 1; |
|
} |
|
current = cookie->batch + cookie->batches - 1; |
|
iob_addbuf_free(current, header, header_size); |
|
|
|
/* Split huge iovectors into separate io_batches */ |
|
for (i = 0; i < iovec_entries; ++i) { |
|
/* If the current batch's limit is reached, try to reallocate a new batch to work on */ |
|
if (current->bytesleft > OT_BATCH_LIMIT) { |
|
io_batch *new_batch = realloc(cookie->batch, (cookie->batches + 1) * sizeof(io_batch)); |
|
if (new_batch) { |
|
cookie->batch = new_batch; |
|
current = cookie->batch + cookie->batches++; |
|
OT_IOB_INIT(current, 0); |
|
} |
|
} |
|
iob_addbuf_free(current, iovector[i].iov_base, iovector[i].iov_len); |
|
} |
|
free(iovector); |
|
if (cookie->flag & STRUCT_HTTP_FLAG_CHUNKED_IN_TRANSFER) |
|
iob_addbuf(current, "\r\n", 2); |
|
} |
|
|
|
if ((cookie->flag & STRUCT_HTTP_FLAG_CHUNKED_IN_TRANSFER) && cookie->batch && !is_partial) { |
|
current = cookie->batch + cookie->batches - 1; |
|
iob_addbuf(current, "0\r\n\r\n", 5); |
|
cookie->flag &= ~STRUCT_HTTP_FLAG_CHUNKED_IN_TRANSFER; |
|
} |
|
|
|
/* writeable sockets timeout after 10 minutes */ |
|
taia_now(&t); |
|
taia_addsec(&t, &t, OT_CLIENT_TIMEOUT_SEND); |
|
io_timeout(sock, t); |
|
io_wantwrite(sock); |
|
return 0; |
|
} |
|
|
|
static ssize_t http_handle_stats(const int64 sock, struct ot_workstruct *ws, char *read_ptr) { |
|
static const ot_keywords keywords_main[] = {{"mode", 1}, {"format", 2}, {"info_hash", 3}, {NULL, -3}}; |
|
static const ot_keywords keywords_mode[] = {{"peer", TASK_STATS_PEERS}, |
|
{"conn", TASK_STATS_CONNS}, |
|
{"scrp", TASK_STATS_SCRAPE}, |
|
{"udp4", TASK_STATS_UDP}, |
|
{"tcp4", TASK_STATS_TCP}, |
|
{"busy", TASK_STATS_BUSY_NETWORKS}, |
|
{"torr", TASK_STATS_TORRENTS}, |
|
{"fscr", TASK_STATS_FULLSCRAPE}, |
|
{"s24s", TASK_STATS_SLASH24S}, |
|
{"tpbs", TASK_STATS_TPB}, |
|
{"herr", TASK_STATS_HTTPERRORS}, |
|
{"completed", TASK_STATS_COMPLETED}, |
|
{"top100", TASK_STATS_TOP100}, |
|
{"top10", TASK_STATS_TOP10}, |
|
{"renew", TASK_STATS_RENEW}, |
|
{"syncs", TASK_STATS_SYNCS}, |
|
{"version", TASK_STATS_VERSION}, |
|
{"everything", TASK_STATS_EVERYTHING}, |
|
{"statedump", TASK_FULLSCRAPE_TRACKERSTATE}, |
|
{"fulllog", TASK_STATS_FULLLOG}, |
|
{"woodpeckers", TASK_STATS_WOODPECKERS}, |
|
#ifdef WANT_LOG_NUMWANT |
|
{"numwants", TASK_STATS_NUMWANTS}, |
|
#endif |
|
{NULL, -3}}; |
|
static const ot_keywords keywords_format[] = {{"bin", TASK_FULLSCRAPE_TPB_BINARY}, {"ben", TASK_FULLSCRAPE}, |
|
{"url", TASK_FULLSCRAPE_TPB_URLENCODED}, {"txt", TASK_FULLSCRAPE_TPB_ASCII}, |
|
{"txtp", TASK_FULLSCRAPE_TPB_ASCII_PLUS}, {NULL, -3}}; |
|
|
|
int mode = TASK_STATS_PEERS, scanon = 1, format = 0; |
|
|
|
#ifdef WANT_RESTRICT_STATS |
|
struct http_data *cookie = io_getcookie(sock); |
|
|
|
if (!cookie || !accesslist_is_blessed(cookie->ip, OT_PERMISSION_MAY_STAT)) |
|
HTTPERROR_403_IP; |
|
#endif |
|
|
|
while (scanon) { |
|
switch (scan_find_keywords(keywords_main, &read_ptr, SCAN_SEARCHPATH_PARAM)) { |
|
case -2: |
|
scanon = 0; |
|
break; /* TERMINATOR */ |
|
case -1: |
|
HTTPERROR_400_PARAM; /* PARSE ERROR */ |
|
case -3: |
|
scan_urlencoded_skipvalue(&read_ptr); |
|
break; |
|
case 1: /* matched "mode" */ |
|
if ((mode = scan_find_keywords(keywords_mode, &read_ptr, SCAN_SEARCHPATH_VALUE)) <= 0) |
|
HTTPERROR_400_PARAM; |
|
break; |
|
case 2: /* matched "format" */ |
|
if ((format = scan_find_keywords(keywords_format, &read_ptr, SCAN_SEARCHPATH_VALUE)) <= 0) |
|
HTTPERROR_400_PARAM; |
|
break; |
|
case 3: |
|
HTTPERROR_400_PARAM; /* If the stats URL was mistakenly added as announce URL, return a 400 */ |
|
} |
|
} |
|
|
|
#ifdef WANT_FULLSCRAPE |
|
if (mode == TASK_FULLSCRAPE_TRACKERSTATE) { |
|
format = mode; |
|
mode = TASK_STATS_TPB; |
|
} |
|
|
|
if (mode == TASK_STATS_TPB) { |
|
struct http_data *cookie = io_getcookie(sock); |
|
tai6464 t; |
|
#ifdef WANT_COMPRESSION_GZIP |
|
ws->request[ws->request_size] = 0; |
|
#ifndef WANT_COMPRESSION_GZIP_ALWAYS |
|
if (strstr(read_ptr - 1, "gzip")) { |
|
#endif |
|
cookie->flag |= STRUCT_HTTP_FLAG_GZIP; |
|
format |= TASK_FLAG_GZIP; |
|
#ifndef WANT_COMPRESSION_GZIP_ALWAYS |
|
} |
|
#endif |
|
#endif |
|
/* Pass this task to the worker thread */ |
|
cookie->flag |= STRUCT_HTTP_FLAG_WAITINGFORTASK | STRUCT_HTTP_FLAG_CHUNKED; |
|
|
|
/* Clients waiting for us should not easily timeout */ |
|
taia_uint(&t, 0); |
|
io_timeout(sock, t); |
|
fullscrape_deliver(sock, format); |
|
io_dontwantread(sock); |
|
return ws->reply_size = -2; |
|
} |
|
#endif |
|
|
|
/* default format for now */ |
|
if ((mode & TASK_CLASS_MASK) == TASK_STATS) { |
|
tai6464 t; |
|
/* Complex stats also include expensive memory debugging tools */ |
|
taia_uint(&t, 0); |
|
io_timeout(sock, t); |
|
stats_deliver(sock, mode); |
|
return ws->reply_size = -2; |
|
} |
|
|
|
/* Simple stats can be answerred immediately */ |
|
return ws->reply_size = return_stats_for_tracker(ws->reply, mode, 0); |
|
} |
|
|
|
#ifdef WANT_MODEST_FULLSCRAPES |
|
static pthread_mutex_t g_modest_fullscrape_mutex = PTHREAD_MUTEX_INITIALIZER; |
|
static ot_vector g_modest_fullscrape_timeouts; |
|
typedef struct { |
|
ot_ip6 ip; |
|
ot_time last_fullscrape; |
|
} ot_scrape_log; |
|
#endif |
|
|
|
#ifdef WANT_FULLSCRAPE |
|
static ssize_t http_handle_fullscrape(const int64 sock, struct ot_workstruct *ws) { |
|
struct http_data *cookie = io_getcookie(sock); |
|
int format = 0; |
|
tai6464 t; |
|
|
|
#ifdef WANT_MODEST_FULLSCRAPES |
|
{ |
|
ot_scrape_log this_peer, *new_peer; |
|
int exactmatch; |
|
memcpy(this_peer.ip, cookie->ip, sizeof(ot_ip6)); |
|
this_peer.last_fullscrape = g_now_seconds; |
|
pthread_mutex_lock(&g_modest_fullscrape_mutex); |
|
new_peer = vector_find_or_insert(&g_modest_fullscrape_timeouts, &this_peer, sizeof(ot_scrape_log), sizeof(ot_ip6), &exactmatch); |
|
if (!new_peer) { |
|
pthread_mutex_unlock(&g_modest_fullscrape_mutex); |
|
HTTPERROR_500; |
|
} |
|
if (exactmatch && (this_peer.last_fullscrape - new_peer->last_fullscrape) < OT_MODEST_PEER_TIMEOUT) { |
|
pthread_mutex_unlock(&g_modest_fullscrape_mutex); |
|
HTTPERROR_402_NOTMODEST; |
|
} |
|
memcpy(new_peer, &this_peer, sizeof(ot_scrape_log)); |
|
pthread_mutex_unlock(&g_modest_fullscrape_mutex); |
|
} |
|
#endif |
|
|
|
#ifdef WANT_COMPRESSION_GZIP |
|
ws->request[ws->request_size - 1] = 0; |
|
#ifndef WANT_COMPRESSION_GZIP_ALWAYS |
|
if (strstr(ws->request, "gzip")) { |
|
#endif |
|
cookie->flag |= STRUCT_HTTP_FLAG_GZIP; |
|
format = TASK_FLAG_GZIP; |
|
stats_issue_event(EVENT_FULLSCRAPE_REQUEST_GZIP, 0, (uintptr_t)cookie->ip); |
|
#ifndef WANT_COMPRESSION_GZIP_ALWAYS |
|
} else |
|
#endif |
|
#endif |
|
stats_issue_event(EVENT_FULLSCRAPE_REQUEST, 0, (uintptr_t)cookie->ip); |
|
|
|
#ifdef _DEBUG_HTTPERROR |
|
fprintf(stderr, "%s", ws->debugbuf); |
|
#endif |
|
|
|
/* Pass this task to the worker thread */ |
|
cookie->flag |= STRUCT_HTTP_FLAG_WAITINGFORTASK | STRUCT_HTTP_FLAG_CHUNKED; |
|
/* Clients waiting for us should not easily timeout */ |
|
taia_uint(&t, 0); |
|
io_timeout(sock, t); |
|
fullscrape_deliver(sock, TASK_FULLSCRAPE | format); |
|
io_dontwantread(sock); |
|
return ws->reply_size = -2; |
|
} |
|
#endif |
|
|
|
static ssize_t http_handle_scrape(const int64 sock, struct ot_workstruct *ws, char *read_ptr) { |
|
static const ot_keywords keywords_scrape[] = {{"info_hash", 1}, {NULL, -3}}; |
|
|
|
ot_hash *multiscrape_buf = (ot_hash *)ws->request; |
|
int scanon = 1, numwant = 0; |
|
|
|
/* This is to hack around stupid clients that send "scrape ?info_hash" */ |
|
if (read_ptr[-1] != '?') { |
|
while ((*read_ptr != '?') && (*read_ptr != '\n')) |
|
++read_ptr; |
|
if (*read_ptr == '\n') |
|
HTTPERROR_400_PARAM; |
|
++read_ptr; |
|
} |
|
|
|
while (scanon) { |
|
switch (scan_find_keywords(keywords_scrape, &read_ptr, SCAN_SEARCHPATH_PARAM)) { |
|
case -2: |
|
scanon = 0; |
|
break; /* TERMINATOR */ |
|
default: |
|
HTTPERROR_400_PARAM; /* PARSE ERROR */ |
|
case -3: |
|
scan_urlencoded_skipvalue(&read_ptr); |
|
break; |
|
case 1: /* matched "info_hash" */ |
|
/* ignore this, when we have less than 20 bytes */ |
|
if (scan_urlencoded_query(&read_ptr, (char *)(multiscrape_buf + numwant++), SCAN_SEARCHPATH_VALUE) != (ssize_t)sizeof(ot_hash)) |
|
HTTPERROR_400_PARAM; |
|
break; |
|
} |
|
} |
|
|
|
/* No info_hash found? Inform user */ |
|
if (!numwant) |
|
HTTPERROR_400_PARAM; |
|
|
|
/* Limit number of hashes to process */ |
|
if (numwant > OT_MAXMULTISCRAPE_COUNT) |
|
numwant = OT_MAXMULTISCRAPE_COUNT; |
|
|
|
/* Enough for http header + whole scrape string */ |
|
ws->reply_size = return_tcp_scrape_for_torrent((const ot_hash *)multiscrape_buf, numwant, ws->reply); |
|
stats_issue_event(EVENT_SCRAPE, FLAG_TCP, ws->reply_size); |
|
return ws->reply_size; |
|
} |
|
|
|
#ifdef WANT_LOG_NUMWANT |
|
unsigned long long numwants[201]; |
|
#endif |
|
|
|
#if defined(WANT_KEEPALIVE) || defined(WANT_IP_FROM_PROXY) |
|
static char *http_header(char *data, size_t byte_count, char *header) { |
|
size_t i; |
|
long sl = strlen(header); |
|
for (i = 0; i + sl + 2 < byte_count; ++i) { |
|
if (data[i] != '\n' || data[i + sl + 1] != ':') |
|
continue; |
|
if (!case_equalb(data + i + 1, sl, header)) |
|
continue; |
|
data += i + sl + 2; |
|
while (*data == ' ' || *data == '\t') |
|
++data; |
|
return data; |
|
} |
|
return 0; |
|
} |
|
#endif |
|
|
|
static ot_keywords keywords_announce[] = {{"port", 1}, {"left", 2}, {"event", 3}, {"numwant", 4}, {"compact", 5}, {"compact6", 5}, {"info_hash", 6}, |
|
#ifdef WANT_IP_FROM_QUERY_STRING |
|
{"ip", 7}, |
|
#endif |
|
#ifdef WANT_FULLLOG_NETWORKS |
|
{"lognet", 8}, |
|
#endif |
|
{"peer_id", 9}, {NULL, -3}}; |
|
static ot_keywords keywords_announce_event[] = {{"completed", 1}, {"stopped", 2}, {NULL, -3}}; |
|
static ssize_t http_handle_announce(const int64 sock, struct ot_workstruct *ws, char *read_ptr) { |
|
int numwant, tmp, scanon; |
|
unsigned short port = 0; |
|
char *write_ptr; |
|
ssize_t len; |
|
struct http_data *cookie = io_getcookie(sock); |
|
|
|
/* This is to hack around stupid clients that send "announce ?info_hash" */ |
|
if (read_ptr[-1] != '?') { |
|
while ((*read_ptr != '?') && (*read_ptr != '\n')) |
|
++read_ptr; |
|
if (*read_ptr == '\n') |
|
HTTPERROR_400_PARAM; |
|
++read_ptr; |
|
} |
|
|
|
#ifdef WANT_IP_FROM_PROXY |
|
if (accesslist_is_blessed(cookie->ip, OT_PERMISSION_MAY_PROXY)) { |
|
ot_ip6 proxied_ip; |
|
char *fwd = http_header(ws->request, ws->header_size, "x-forwarded-for"); |
|
if (fwd && scan_ip6(fwd, proxied_ip)) { |
|
OT_SETIP(ws->peer, proxied_ip); |
|
} else |
|
OT_SETIP(ws->peer, cookie->ip); |
|
} else |
|
#endif |
|
OT_SETIP(ws->peer, cookie->ip); |
|
|
|
ws->peer_id = NULL; |
|
ws->hash = NULL; |
|
|
|
OT_SETPORT(ws->peer, &port); |
|
OT_PEERFLAG(ws->peer) = 0; |
|
numwant = 50; |
|
scanon = 1; |
|
|
|
while (scanon) { |
|
switch (scan_find_keywords(keywords_announce, &read_ptr, SCAN_SEARCHPATH_PARAM)) { |
|
case -2: |
|
scanon = 0; |
|
break; /* TERMINATOR */ |
|
case -1: |
|
HTTPERROR_400_PARAM; /* PARSE ERROR */ |
|
case -3: |
|
scan_urlencoded_skipvalue(&read_ptr); |
|
break; |
|
case 1: /* matched "port" */ |
|
len = scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE); |
|
if ((len <= 0) || scan_fixed_int(write_ptr, len, &tmp) || (tmp > 0xffff)) |
|
HTTPERROR_400_PARAM; |
|
port = htons(tmp); |
|
OT_SETPORT(&ws->peer, &port); |
|
break; |
|
case 2: /* matched "left" */ |
|
if ((len = scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE)) <= 0) |
|
HTTPERROR_400_PARAM; |
|
if (scan_fixed_int(write_ptr, len, &tmp)) |
|
tmp = 0; |
|
if (!tmp) |
|
OT_PEERFLAG(&ws->peer) |= PEER_FLAG_SEEDING; |
|
break; |
|
case 3: /* matched "event" */ |
|
switch (scan_find_keywords(keywords_announce_event, &read_ptr, SCAN_SEARCHPATH_VALUE)) { |
|
case -1: |
|
HTTPERROR_400_PARAM; |
|
case 1: /* matched "completed" */ |
|
OT_PEERFLAG(&ws->peer) |= PEER_FLAG_COMPLETED; |
|
break; |
|
case 2: /* matched "stopped" */ |
|
OT_PEERFLAG(&ws->peer) |= PEER_FLAG_STOPPED; |
|
break; |
|
default: |
|
break; |
|
} |
|
break; |
|
case 4: /* matched "numwant" */ |
|
len = scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE); |
|
if ((len <= 0) || scan_fixed_int(write_ptr, len, &numwant)) |
|
HTTPERROR_400_PARAM; |
|
if (numwant < 0) |
|
numwant = 50; |
|
if (numwant > 200) |
|
numwant = 200; |
|
break; |
|
case 5: /* matched "compact" */ |
|
len = scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE); |
|
if ((len <= 0) || scan_fixed_int(write_ptr, len, &tmp)) |
|
HTTPERROR_400_PARAM; |
|
if (!tmp) |
|
HTTPERROR_400_COMPACT; |
|
break; |
|
case 6: /* matched "info_hash" */ |
|
if (ws->hash) |
|
HTTPERROR_400_DOUBLEHASH; |
|
/* ignore this, when we have less than 20 bytes */ |
|
if (scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE) != 20) |
|
HTTPERROR_400_PARAM; |
|
ws->hash = (ot_hash *)write_ptr; |
|
break; |
|
#ifdef WANT_IP_FROM_QUERY_STRING |
|
case 7: /* matched "ip" */ |
|
{ |
|
char *tmp_buf1 = ws->reply, *tmp_buf2 = ws->reply + 16; |
|
len = scan_urlencoded_query(&read_ptr, tmp_buf2, SCAN_SEARCHPATH_VALUE); |
|
tmp_buf2[len] = 0; |
|
if ((len <= 0) || !scan_ip6(tmp_buf2, tmp_buf1)) |
|
HTTPERROR_400_PARAM; |
|
OT_SETIP(&ws->peer, tmp_buf1); |
|
} break; |
|
#endif |
|
#ifdef WANT_FULLLOG_NETWORKS |
|
case 8: /* matched "lognet" */ |
|
{ |
|
// if( accesslist_is_blessed( cookie->ip, OT_PERMISSION_MAY_STAT ) ) { |
|
char *tmp_buf = ws->reply; |
|
ot_net net; |
|
signed short parsed, bits; |
|
|
|
len = scan_urlencoded_query(&read_ptr, tmp_buf, SCAN_SEARCHPATH_VALUE); |
|
tmp_buf[len] = 0; |
|
if (len <= 0) |
|
HTTPERROR_400_PARAM; |
|
if (*tmp_buf == '-') { |
|
loglist_reset(); |
|
return ws->reply_size = sprintf(ws->reply, "Successfully removed.\n"); |
|
} |
|
parsed = scan_ip6(tmp_buf, net.address); |
|
if (!parsed) |
|
HTTPERROR_400_PARAM; |
|
if (tmp_buf[parsed++] != '/') |
|
bits = 128; |
|
else { |
|
parsed = scan_short(tmp_buf + parsed, &bits); |
|
if (!parsed) |
|
HTTPERROR_400_PARAM; |
|
if (ip6_isv4mapped(net.address)) |
|
bits += 96; |
|
} |
|
net.bits = bits; |
|
loglist_add_network(&net); |
|
return ws->reply_size = sprintf(ws->reply, "Successfully added.\n"); |
|
//} |
|
} break; |
|
#endif |
|
case 9: /* matched "peer_id" */ |
|
/* ignore this, when we have less than 20 bytes */ |
|
if (scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_SEARCHPATH_VALUE) != 20) |
|
HTTPERROR_400_PARAM; |
|
ws->peer_id = write_ptr; |
|
break; |
|
} |
|
} |
|
|
|
#ifdef WANT_LOG_NUMWANT |
|
numwants[numwant]++; |
|
#endif |
|
|
|
/* XXX DEBUG |
|
stats_issue_event( EVENT_ACCEPT, FLAG_TCP, (uintptr_t)ws->reply ); |
|
*/ |
|
|
|
/* Scanned whole query string */ |
|
if (!ws->hash) |
|
return ws->reply_size = sprintf(ws->reply, "d14:failure reason80:Your client forgot to send your torrent's info_hash. Please upgrade your client.e"); |
|
|
|
if (OT_PEERFLAG(&ws->peer) & PEER_FLAG_STOPPED) |
|
ws->reply_size = remove_peer_from_torrent(FLAG_TCP, ws); |
|
else |
|
ws->reply_size = add_peer_to_torrent_and_return_peers(FLAG_TCP, ws, numwant); |
|
|
|
stats_issue_event(EVENT_ANNOUNCE, FLAG_TCP, ws->reply_size); |
|
return ws->reply_size; |
|
} |
|
|
|
ssize_t http_handle_request(const int64 sock, struct ot_workstruct *ws) { |
|
ssize_t reply_off, len; |
|
char *read_ptr = ws->request, *write_ptr; |
|
|
|
#ifdef WANT_FULLLOG_NETWORKS |
|
struct http_data *cookie = io_getcookie(sock); |
|
if (loglist_check_address(cookie->ip)) { |
|
ot_log *log = malloc(sizeof(ot_log)); |
|
if (log) { |
|
log->size = ws->request_size; |
|
log->data = malloc(ws->request_size); |
|
log->next = 0; |
|
log->time = g_now_seconds; |
|
memcpy(log->ip, cookie->ip, sizeof(ot_ip6)); |
|
if (log->data) { |
|
memcpy(log->data, ws->request, ws->request_size); |
|
if (!g_logchain_first) |
|
g_logchain_first = g_logchain_last = log; |
|
else { |
|
g_logchain_last->next = log; |
|
g_logchain_last = log; |
|
} |
|
} else |
|
free(log); |
|
} |
|
} |
|
#endif |
|
|
|
#ifdef _DEBUG_HTTPERROR |
|
reply_off = ws->request_size; |
|
if (ws->request_size >= G_DEBUGBUF_SIZE) |
|
reply_off = G_DEBUGBUF_SIZE - 1; |
|
memcpy(ws->debugbuf, ws->request, reply_off); |
|
ws->debugbuf[reply_off] = 0; |
|
#endif |
|
|
|
/* Tell subroutines where to put reply data */ |
|
ws->reply = ws->outbuf + SUCCESS_HTTP_HEADER_LENGTH; |
|
|
|
/* This one implicitely tests strlen < 5, too -- remember, it is \n terminated */ |
|
if (memcmp(read_ptr, "GET /", 5)) |
|
HTTPERROR_400; |
|
|
|
/* Skip leading '/' */ |
|
for (read_ptr += 4; *read_ptr == '/'; ++read_ptr) |
|
; |
|
|
|
/* Try to parse the request. |
|
In reality we abandoned requiring the url to be correct. This now |
|
only decodes url encoded characters, we check for announces and |
|
scrapes by looking for "a*" or "sc" */ |
|
len = scan_urlencoded_query(&read_ptr, write_ptr = read_ptr, SCAN_PATH); |
|
|
|
/* If parsing returned an error, leave with not found */ |
|
if (g_redirecturl && (len == -2)) |
|
HTTPERROR_302; |
|
if (len <= 0) |
|
HTTPERROR_404; |
|
|
|
/* This is the hardcore match for announce*/ |
|
if ((*write_ptr == 'a') || (*write_ptr == '?')) |
|
http_handle_announce(sock, ws, read_ptr); |
|
#ifdef WANT_FULLSCRAPE |
|
else if (!memcmp(write_ptr, "scrape HTTP/", 12)) |
|
http_handle_fullscrape(sock, ws); |
|
#endif |
|
/* This is the hardcore match for scrape */ |
|
else if (!memcmp(write_ptr, "sc", 2)) |
|
http_handle_scrape(sock, ws, read_ptr); |
|
/* All the rest is matched the standard way */ |
|
else if (len == g_stats_path_len && !memcmp(write_ptr, g_stats_path, len)) |
|
http_handle_stats(sock, ws, read_ptr); |
|
else |
|
HTTPERROR_404; |
|
|
|
/* Find out if the client wants to keep this connection alive */ |
|
ws->keep_alive = 0; |
|
#ifdef WANT_KEEPALIVE |
|
read_ptr = http_header(ws->request, ws->header_size, "connection"); |
|
if (read_ptr && (*read_ptr == 'K' || *read_ptr == 'k')) |
|
ws->keep_alive = 1; |
|
#endif |
|
|
|
/* If routines handled sending themselves, just return */ |
|
if (ws->reply_size == -2) |
|
return 0; |
|
/* If routine failed, let http error take over */ |
|
if (ws->reply_size <= 0) |
|
HTTPERROR_500; |
|
|
|
/* This one is rather ugly, so I take you step by step through it. |
|
|
|
1. In order to avoid having two buffers, one for header and one for content, we allow all above functions from trackerlogic to |
|
write to a fixed location, leaving SUCCESS_HTTP_HEADER_LENGTH bytes in our work buffer, which is enough for the static string |
|
plus dynamic space needed to expand our Content-Length value. We reserve SUCCESS_HTTP_SIZE_OFF for its expansion and calculate |
|
the space NOT needed to expand in reply_off |
|
*/ |
|
reply_off = SUCCESS_HTTP_SIZE_OFF - snprintf(ws->outbuf, 0, "%zd", ws->reply_size); |
|
ws->reply = ws->outbuf + reply_off; |
|
|
|
/* 2. Now we sprintf our header so that sprintf writes its terminating '\0' exactly one byte before content starts. Complete |
|
packet size is increased by size of header plus one byte '\n', we will copy over '\0' in next step */ |
|
ws->reply_size += 1 + sprintf(ws->reply, "HTTP/1.1 200 OK\r\nContent-Type: text/plain\r\nContent-Length: %zd\r\n\r", ws->reply_size); |
|
|
|
/* 3. Finally we join both blocks neatly */ |
|
ws->outbuf[SUCCESS_HTTP_HEADER_LENGTH - 1] = '\n'; |
|
|
|
http_senddata(sock, ws); |
|
return ws->reply_size; |
|
} |
|
|
|
const char *g_version_http_c = "$Source$: $Revision$\n";
|
|
|