mirror of https://github.com/YGGverse/qBt_SE.git
imDMG
6 years ago
commit
d2103e1759
7 changed files with 1625 additions and 0 deletions
@ -0,0 +1,244 @@ |
|||||||
|
# VERSION: 1.0 |
||||||
|
# AUTHORS: imDMG |
||||||
|
|
||||||
|
# LICENSING INFORMATION |
||||||
|
|
||||||
|
import tempfile |
||||||
|
import os |
||||||
|
import logging |
||||||
|
import time |
||||||
|
|
||||||
|
from urllib.request import build_opener, HTTPCookieProcessor, ProxyHandler |
||||||
|
from urllib.parse import urlencode, quote, unquote |
||||||
|
from urllib.error import URLError, HTTPError |
||||||
|
from http.cookiejar import CookieJar |
||||||
|
from html.parser import HTMLParser |
||||||
|
from novaprinter import prettyPrinter |
||||||
|
|
||||||
|
# setup logging into qBittorrent/logs |
||||||
|
logging.basicConfig(level=logging.DEBUG, |
||||||
|
format='%(asctime)s %(name)-12s %(levelname)-8s %(message)s', |
||||||
|
datefmt='%m-%d %H:%M', |
||||||
|
filename=os.path.abspath( |
||||||
|
os.path.join(os.path.dirname(__file__), '../..', 'logs')) + "/kinozal_se.log", |
||||||
|
# filename="kinozal_se.log", |
||||||
|
filemode='w') |
||||||
|
|
||||||
|
# benchmark |
||||||
|
# start_time = time.time() |
||||||
|
|
||||||
|
|
||||||
|
class kinozal(object): |
||||||
|
name = 'Kinozal' |
||||||
|
url = 'http://kinozal.tv' |
||||||
|
supported_categories = {'all': '0', |
||||||
|
'movies': '1002', |
||||||
|
'tv': '1001', |
||||||
|
'music': '1004', |
||||||
|
'games': '23', |
||||||
|
'anime': '20', |
||||||
|
'software': '32'} |
||||||
|
|
||||||
|
# Set proxies (default false) |
||||||
|
# make sure that proxies keys is'nt empty |
||||||
|
proxy = True |
||||||
|
proxies = { |
||||||
|
'http': '', |
||||||
|
'https': '', |
||||||
|
} |
||||||
|
|
||||||
|
# credentials |
||||||
|
username = "USERNAME" |
||||||
|
password = "PASSWORD" |
||||||
|
ua = 'Mozilla/5.0 (X11; Linux i686; rv:38.0) Gecko/20100101 Firefox/38.0' |
||||||
|
|
||||||
|
def __init__(self): |
||||||
|
# establish connection |
||||||
|
# |
||||||
|
# make cookie |
||||||
|
cj = CookieJar() |
||||||
|
self.session = build_opener(HTTPCookieProcessor(cj)) |
||||||
|
|
||||||
|
# add proxy handler if needed |
||||||
|
if self.proxy and any(self.proxies.keys()): |
||||||
|
self.session.add_handler(ProxyHandler(self.proxies)) |
||||||
|
|
||||||
|
# change user-agent |
||||||
|
self.session.addheaders.pop() |
||||||
|
self.session.addheaders.append(('User-Agent', self.ua)) |
||||||
|
|
||||||
|
form_data = {"username": self.username, "password": self.password} |
||||||
|
data_encoded = urlencode(form_data).encode('cp1251') |
||||||
|
|
||||||
|
try: |
||||||
|
response = self.session.open(self.url + '/takelogin.php', data_encoded) |
||||||
|
# Only continue if response status is OK. |
||||||
|
if response.getcode() != 200: |
||||||
|
raise HTTPError(response.geturl(), response.getcode(), |
||||||
|
"HTTP request to {} failed with status: {}".format(self.url, response.getcode()), |
||||||
|
response.info(), None) |
||||||
|
except (URLError, HTTPError) as e: |
||||||
|
logging.error(e) |
||||||
|
raise e |
||||||
|
|
||||||
|
if 'uid' not in [cookie.name for cookie in cj]: |
||||||
|
logging.debug(cj) |
||||||
|
|
||||||
|
class WorstParser(HTMLParser): |
||||||
|
def __init__(self, url=''): |
||||||
|
HTMLParser.__init__(self) |
||||||
|
self.url = url |
||||||
|
self.torrent = {'link': '', |
||||||
|
'name': '', |
||||||
|
'size': '', |
||||||
|
'seeds': '', |
||||||
|
'leech': '', |
||||||
|
'desc_link': '', } |
||||||
|
|
||||||
|
# we need a page markup to know when stop and collect data, |
||||||
|
# because available methods, in this class, do not communicate each other |
||||||
|
# as a result, we make markup to transfer information |
||||||
|
# from one method to another, along a chain |
||||||
|
# |
||||||
|
# markup on result table |
||||||
|
self.result_table = False # table with results is found |
||||||
|
self.torrent_row = False # found torrent row for collect data |
||||||
|
self.index_td = 0 # td counter in torrent row |
||||||
|
self.write = None # trigger to detecting when to collect data |
||||||
|
|
||||||
|
# markup pagination |
||||||
|
self.paginator = False # found more pages in result |
||||||
|
self.pages = 0 # page counter |
||||||
|
|
||||||
|
self.found_torrents = 0 |
||||||
|
|
||||||
|
def handle_starttag(self, tag, attrs): |
||||||
|
# search result table by class t_peer |
||||||
|
if tag == 'table': |
||||||
|
for name, value in attrs: |
||||||
|
if name == 'class' and 't_peer' in value: |
||||||
|
self.result_table = True |
||||||
|
|
||||||
|
# search for torrent row by class bg |
||||||
|
if self.result_table and tag == 'tr': |
||||||
|
for name, value in attrs: |
||||||
|
if name == 'class' and 'bg' in value: |
||||||
|
self.torrent_row = True |
||||||
|
|
||||||
|
# count td for find right td |
||||||
|
if self.torrent_row and tag == 'td': |
||||||
|
if self.index_td == 3: |
||||||
|
self.write = "size" |
||||||
|
elif self.index_td == 4: |
||||||
|
self.write = "seeds" |
||||||
|
elif self.index_td == 5: |
||||||
|
self.write = "leech" |
||||||
|
|
||||||
|
self.index_td += 1 |
||||||
|
|
||||||
|
# search for torrent link by classes r0 or r1 |
||||||
|
if self.torrent_row and tag == 'a': |
||||||
|
for name, value in attrs: |
||||||
|
if name == 'class' and 'r' in value: |
||||||
|
self.torrent['link'] = 'http://dl.kinozal.tv/download.php?id=' + attrs[0][1].split('=')[1] |
||||||
|
self.torrent['desc_link'] = self.url + attrs[0][1] |
||||||
|
self.write = "name" |
||||||
|
|
||||||
|
# search for right div with class paginator |
||||||
|
if self.found_torrents == 50 and tag == 'div': |
||||||
|
for name, value in attrs: |
||||||
|
if name == 'class' and value == 'paginator': |
||||||
|
self.paginator = True |
||||||
|
|
||||||
|
# search for block with page numbers |
||||||
|
if self.paginator and tag == 'li': |
||||||
|
self.pages += 1 |
||||||
|
|
||||||
|
def handle_endtag(self, tag): |
||||||
|
# detecting that torrent row is closed and print all collected data |
||||||
|
if self.torrent_row and tag == 'tr': |
||||||
|
self.torrent["engine_url"] = self.url |
||||||
|
logging.debug('tr: ' + str(self.torrent)) |
||||||
|
prettyPrinter(self.torrent) |
||||||
|
self.torrent = {key: '' for key in self.torrent} |
||||||
|
self.index_td = 0 |
||||||
|
self.torrent_row = False |
||||||
|
self.found_torrents += 1 |
||||||
|
|
||||||
|
# detecting that table with result is close |
||||||
|
if self.result_table and tag == 'table': |
||||||
|
self.result_table = False |
||||||
|
|
||||||
|
# detecting that we found all pagination |
||||||
|
if self.paginator and tag == 'ul': |
||||||
|
self.paginator = False |
||||||
|
|
||||||
|
def handle_data(self, data: str): |
||||||
|
# detecting that we need write data at this moment |
||||||
|
if self.write and self.result_table: |
||||||
|
if self.write == 'size': |
||||||
|
data = self.units_convert(data) |
||||||
|
self.torrent[self.write] = data.strip() |
||||||
|
self.write = None |
||||||
|
|
||||||
|
@staticmethod |
||||||
|
def units_convert(unit): |
||||||
|
# replace size units |
||||||
|
table = {'ТБ': 'TB', 'ГБ': 'GB', 'МБ': 'MB', 'КБ': 'KB'} |
||||||
|
x = unit.split(" ") |
||||||
|
x[1] = table[x[1]] |
||||||
|
|
||||||
|
return " ".join(x) |
||||||
|
|
||||||
|
def error(self, message): |
||||||
|
pass |
||||||
|
|
||||||
|
def download_torrent(self, url): |
||||||
|
# Create a torrent file |
||||||
|
file, path = tempfile.mkstemp('.torrent') |
||||||
|
file = os.fdopen(file, "wb") |
||||||
|
|
||||||
|
# Download url |
||||||
|
try: |
||||||
|
response = self.session.open(url) |
||||||
|
# Only continue if response status is OK. |
||||||
|
if response.getcode() != 200: |
||||||
|
raise HTTPError(response.geturl(), response.getcode(), |
||||||
|
"HTTP request to {} failed with status: {}".format(url, response.getcode()), |
||||||
|
response.info(), None) |
||||||
|
except (URLError, HTTPError) as e: |
||||||
|
logging.error(e) |
||||||
|
raise e |
||||||
|
|
||||||
|
# Write it to a file |
||||||
|
file.write(response.read()) |
||||||
|
file.close() |
||||||
|
|
||||||
|
# return file path |
||||||
|
logging.debug(path + " " + url) |
||||||
|
print(path + " " + url) |
||||||
|
|
||||||
|
def search(self, what, cat='all'): |
||||||
|
query = '%s/browse.php?s=%s&c=%s' % (self.url, unquote(quote(what)), self.supported_categories[cat]) |
||||||
|
response = self.session.open(query) |
||||||
|
parser = self.WorstParser(self.url) |
||||||
|
parser.feed(response.read().decode('cp1251')) |
||||||
|
parser.close() |
||||||
|
|
||||||
|
# if first request return that we have pages, we do cycle |
||||||
|
if parser.pages: |
||||||
|
for x in range(1, parser.pages): |
||||||
|
response = self.session.open('%s&page=%s' % (query, x)) |
||||||
|
parser.feed(response.read().decode('cp1251')) |
||||||
|
parser.close() |
||||||
|
|
||||||
|
|
||||||
|
# logging.debug("--- %s seconds ---" % (time.time() - start_time)) |
||||||
|
if __name__ == "__main__": |
||||||
|
kinozal_se = kinozal() |
||||||
|
# print(os.path.abspath(os.path.join(os.path.dirname(__file__), '../..', 'logs'))) |
||||||
|
# print(kinozal_se.WorstParser.units_convert("500 КБ")) |
||||||
|
# kinozal_se.search('terror lostfilm', 'tv') |
||||||
|
# kinozal_se._handle_connection(True) |
||||||
|
# kinozal_se.download_torrent('http://dl.kinozal.tv/download.php?id=1609776') |
||||||
|
# print("--- %s seconds ---" % (time.time() - start_time)) |
@ -0,0 +1,123 @@ |
|||||||
|
#VERSION: 1.42 |
||||||
|
|
||||||
|
# Author: |
||||||
|
# Christophe DUMEZ (chris@qbittorrent.org) |
||||||
|
|
||||||
|
# Redistribution and use in source and binary forms, with or without |
||||||
|
# modification, are permitted provided that the following conditions are met: |
||||||
|
# |
||||||
|
# * Redistributions of source code must retain the above copyright notice, |
||||||
|
# this list of conditions and the following disclaimer. |
||||||
|
# * Redistributions in binary form must reproduce the above copyright |
||||||
|
# notice, this list of conditions and the following disclaimer in the |
||||||
|
# documentation and/or other materials provided with the distribution. |
||||||
|
# * Neither the name of the author nor the names of its contributors may be |
||||||
|
# used to endorse or promote products derived from this software without |
||||||
|
# specific prior written permission. |
||||||
|
# |
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
||||||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
||||||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
||||||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
||||||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
||||||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
||||||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
||||||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
||||||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
||||||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
||||||
|
# POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
import gzip |
||||||
|
import html.entities |
||||||
|
import io |
||||||
|
import os |
||||||
|
import re |
||||||
|
import socket |
||||||
|
import socks |
||||||
|
import tempfile |
||||||
|
import urllib.error |
||||||
|
import urllib.parse |
||||||
|
import urllib.request |
||||||
|
|
||||||
|
# Some sites blocks default python User-agent |
||||||
|
user_agent = 'Mozilla/5.0 (X11; Linux i686; rv:38.0) Gecko/20100101 Firefox/38.0' |
||||||
|
headers = {'User-Agent': user_agent} |
||||||
|
# SOCKS5 Proxy support |
||||||
|
if "sock_proxy" in os.environ and len(os.environ["sock_proxy"].strip()) > 0: |
||||||
|
proxy_str = os.environ["sock_proxy"].strip() |
||||||
|
m = re.match(r"^(?:(?P<username>[^:]+):(?P<password>[^@]+)@)?(?P<host>[^:]+):(?P<port>\w+)$", |
||||||
|
proxy_str) |
||||||
|
if m is not None: |
||||||
|
socks.setdefaultproxy(socks.PROXY_TYPE_SOCKS5, m.group('host'), |
||||||
|
int(m.group('port')), True, m.group('username'), m.group('password')) |
||||||
|
socket.socket = socks.socksocket |
||||||
|
|
||||||
|
|
||||||
|
def htmlentitydecode(s): |
||||||
|
# First convert alpha entities (such as é) |
||||||
|
# (Inspired from http://mail.python.org/pipermail/python-list/2007-June/443813.html) |
||||||
|
def entity2char(m): |
||||||
|
entity = m.group(1) |
||||||
|
if entity in html.entities.name2codepoint: |
||||||
|
return chr(html.entities.name2codepoint[entity]) |
||||||
|
return " " # Unknown entity: We replace with a space. |
||||||
|
t = re.sub('&(%s);' % '|'.join(html.entities.name2codepoint), entity2char, s) |
||||||
|
|
||||||
|
# Then convert numerical entities (such as é) |
||||||
|
t = re.sub(r'&#(\d+);', lambda x: chr(int(x.group(1))), t) |
||||||
|
|
||||||
|
# Then convert hexa entities (such as é) |
||||||
|
return re.sub(r'&#x(\w+);', lambda x: chr(int(x.group(1), 16)), t) |
||||||
|
|
||||||
|
|
||||||
|
def retrieve_url(url): |
||||||
|
""" Return the content of the url page as a string """ |
||||||
|
req = urllib.request.Request(url, headers=headers) |
||||||
|
try: |
||||||
|
response = urllib.request.urlopen(req) |
||||||
|
except urllib.error.URLError as errno: |
||||||
|
print(" ".join(("Connection error:", str(errno.reason)))) |
||||||
|
return "" |
||||||
|
dat = response.read() |
||||||
|
# Check if it is gzipped |
||||||
|
if dat[:2] == b'\x1f\x8b': |
||||||
|
# Data is gzip encoded, decode it |
||||||
|
compressedstream = io.BytesIO(dat) |
||||||
|
gzipper = gzip.GzipFile(fileobj=compressedstream) |
||||||
|
extracted_data = gzipper.read() |
||||||
|
dat = extracted_data |
||||||
|
info = response.info() |
||||||
|
charset = 'utf-8' |
||||||
|
try: |
||||||
|
ignore, charset = info['Content-Type'].split('charset=') |
||||||
|
except Exception: |
||||||
|
pass |
||||||
|
dat = dat.decode(charset, 'replace') |
||||||
|
dat = htmlentitydecode(dat) |
||||||
|
# return dat.encode('utf-8', 'replace') |
||||||
|
return dat |
||||||
|
|
||||||
|
|
||||||
|
def download_file(url, referer=None): |
||||||
|
""" Download file at url and write it to a file, return the path to the file and the url """ |
||||||
|
file, path = tempfile.mkstemp() |
||||||
|
file = os.fdopen(file, "wb") |
||||||
|
# Download url |
||||||
|
req = urllib.request.Request(url, headers=headers) |
||||||
|
if referer is not None: |
||||||
|
req.add_header('referer', referer) |
||||||
|
response = urllib.request.urlopen(req) |
||||||
|
dat = response.read() |
||||||
|
# Check if it is gzipped |
||||||
|
if dat[:2] == b'\x1f\x8b': |
||||||
|
# Data is gzip encoded, decode it |
||||||
|
compressedstream = io.BytesIO(dat) |
||||||
|
gzipper = gzip.GzipFile(fileobj=compressedstream) |
||||||
|
extracted_data = gzipper.read() |
||||||
|
dat = extracted_data |
||||||
|
|
||||||
|
# Write it to a file |
||||||
|
file.write(dat) |
||||||
|
file.close() |
||||||
|
# return file path |
||||||
|
return (path + " " + url) |
@ -0,0 +1,190 @@ |
|||||||
|
#VERSION: 1.43 |
||||||
|
|
||||||
|
# Author: |
||||||
|
# Fabien Devaux <fab AT gnux DOT info> |
||||||
|
# Contributors: |
||||||
|
# Christophe Dumez <chris@qbittorrent.org> (qbittorrent integration) |
||||||
|
# Thanks to gab #gcu @ irc.freenode.net (multipage support on PirateBay) |
||||||
|
# Thanks to Elias <gekko04@users.sourceforge.net> (torrentreactor and isohunt search engines) |
||||||
|
# |
||||||
|
# Licence: BSD |
||||||
|
|
||||||
|
# Redistribution and use in source and binary forms, with or without |
||||||
|
# modification, are permitted provided that the following conditions are met: |
||||||
|
# |
||||||
|
# * Redistributions of source code must retain the above copyright notice, |
||||||
|
# this list of conditions and the following disclaimer. |
||||||
|
# * Redistributions in binary form must reproduce the above copyright |
||||||
|
# notice, this list of conditions and the following disclaimer in the |
||||||
|
# documentation and/or other materials provided with the distribution. |
||||||
|
# * Neither the name of the author nor the names of its contributors may be |
||||||
|
# used to endorse or promote products derived from this software without |
||||||
|
# specific prior written permission. |
||||||
|
# |
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
||||||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
||||||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
||||||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
||||||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
||||||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
||||||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
||||||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
||||||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
||||||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
||||||
|
# POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
import urllib.parse |
||||||
|
from os import path |
||||||
|
from glob import glob |
||||||
|
from sys import argv |
||||||
|
from multiprocessing import Pool, cpu_count |
||||||
|
|
||||||
|
THREADED = True |
||||||
|
try: |
||||||
|
MAX_THREADS = cpu_count() |
||||||
|
except NotImplementedError: |
||||||
|
MAX_THREADS = 1 |
||||||
|
|
||||||
|
CATEGORIES = {'all', 'movies', 'tv', 'music', 'games', 'anime', 'software', 'pictures', 'books'} |
||||||
|
|
||||||
|
################################################################################ |
||||||
|
# Every engine should have a "search" method taking |
||||||
|
# a space-free string as parameter (ex. "family+guy") |
||||||
|
# it should call prettyPrinter() with a dict as parameter. |
||||||
|
# The keys in the dict must be: link,name,size,seeds,leech,engine_url |
||||||
|
# As a convention, try to list results by decreasing number of seeds or similar |
||||||
|
################################################################################ |
||||||
|
|
||||||
|
|
||||||
|
def initialize_engines(): |
||||||
|
""" Import available engines |
||||||
|
|
||||||
|
Return list of available engines |
||||||
|
""" |
||||||
|
supported_engines = [] |
||||||
|
|
||||||
|
engines = glob(path.join(path.dirname(__file__), 'engines', '*.py')) |
||||||
|
for engine in engines: |
||||||
|
engi = path.basename(engine).split('.')[0].strip() |
||||||
|
if len(engi) == 0 or engi.startswith('_'): |
||||||
|
continue |
||||||
|
try: |
||||||
|
# import engines.[engine] |
||||||
|
engine_module = __import__(".".join(("engines", engi))) |
||||||
|
# get low-level module |
||||||
|
engine_module = getattr(engine_module, engi) |
||||||
|
# bind class name |
||||||
|
globals()[engi] = getattr(engine_module, engi) |
||||||
|
supported_engines.append(engi) |
||||||
|
except Exception: |
||||||
|
pass |
||||||
|
|
||||||
|
return supported_engines |
||||||
|
|
||||||
|
|
||||||
|
def engines_to_xml(supported_engines): |
||||||
|
""" Generates xml for supported engines """ |
||||||
|
tab = " " * 4 |
||||||
|
|
||||||
|
for short_name in supported_engines: |
||||||
|
search_engine = globals()[short_name]() |
||||||
|
|
||||||
|
supported_categories = "" |
||||||
|
if hasattr(search_engine, "supported_categories"): |
||||||
|
supported_categories = " ".join((key |
||||||
|
for key in search_engine.supported_categories.keys() |
||||||
|
if key != "all")) |
||||||
|
|
||||||
|
yield "".join((tab, "<", short_name, ">\n", |
||||||
|
tab, tab, "<name>", search_engine.name, "</name>\n", |
||||||
|
tab, tab, "<url>", search_engine.url, "</url>\n", |
||||||
|
tab, tab, "<categories>", supported_categories, "</categories>\n", |
||||||
|
tab, "</", short_name, ">\n")) |
||||||
|
|
||||||
|
|
||||||
|
def displayCapabilities(supported_engines): |
||||||
|
""" |
||||||
|
Display capabilities in XML format |
||||||
|
<capabilities> |
||||||
|
<engine_short_name> |
||||||
|
<name>long name</name> |
||||||
|
<url>http://example.com</url> |
||||||
|
<categories>movies music games</categories> |
||||||
|
</engine_short_name> |
||||||
|
</capabilities> |
||||||
|
""" |
||||||
|
xml = "".join(("<capabilities>\n", |
||||||
|
"".join(engines_to_xml(supported_engines)), |
||||||
|
"</capabilities>")) |
||||||
|
print(xml) |
||||||
|
|
||||||
|
|
||||||
|
def run_search(engine_list): |
||||||
|
""" Run search in engine |
||||||
|
|
||||||
|
@param engine_list List with engine, query and category |
||||||
|
|
||||||
|
@retval False if any exceptions occurred |
||||||
|
@retval True otherwise |
||||||
|
""" |
||||||
|
engine, what, cat = engine_list |
||||||
|
try: |
||||||
|
engine = engine() |
||||||
|
# avoid exceptions due to invalid category |
||||||
|
if hasattr(engine, 'supported_categories'): |
||||||
|
if cat in engine.supported_categories: |
||||||
|
engine.search(what, cat) |
||||||
|
else: |
||||||
|
engine.search(what) |
||||||
|
|
||||||
|
return True |
||||||
|
except Exception: |
||||||
|
return False |
||||||
|
|
||||||
|
|
||||||
|
def main(args): |
||||||
|
supported_engines = initialize_engines() |
||||||
|
|
||||||
|
if not args: |
||||||
|
raise SystemExit("./nova2.py [all|engine1[,engine2]*] <category> <keywords>\n" |
||||||
|
"available engines: %s" % (','.join(supported_engines))) |
||||||
|
|
||||||
|
elif args[0] == "--capabilities": |
||||||
|
displayCapabilities(supported_engines) |
||||||
|
return |
||||||
|
|
||||||
|
elif len(args) < 3: |
||||||
|
raise SystemExit("./nova2.py [all|engine1[,engine2]*] <category> <keywords>\n" |
||||||
|
"available engines: %s" % (','.join(supported_engines))) |
||||||
|
|
||||||
|
# get only unique engines with set |
||||||
|
engines_list = set(e.lower() for e in args[0].strip().split(',')) |
||||||
|
|
||||||
|
if 'all' in engines_list: |
||||||
|
engines_list = supported_engines |
||||||
|
else: |
||||||
|
# discard un-supported engines |
||||||
|
engines_list = [engine for engine in engines_list |
||||||
|
if engine in supported_engines] |
||||||
|
|
||||||
|
if not engines_list: |
||||||
|
# engine list is empty. Nothing to do here |
||||||
|
return |
||||||
|
|
||||||
|
cat = args[1].lower() |
||||||
|
|
||||||
|
if cat not in CATEGORIES: |
||||||
|
raise SystemExit(" - ".join(('Invalid category', cat))) |
||||||
|
|
||||||
|
what = urllib.parse.quote(' '.join(args[2:])) |
||||||
|
if THREADED: |
||||||
|
# child process spawning is controlled min(number of searches, number of cpu) |
||||||
|
with Pool(min(len(engines_list), MAX_THREADS)) as pool: |
||||||
|
pool.map(run_search, ([globals()[engine], what, cat] for engine in engines_list)) |
||||||
|
else: |
||||||
|
# py3 note: map is needed to be evaluated for content to be executed |
||||||
|
all(map(run_search, ([globals()[engine], what, cat] for engine in engines_list))) |
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__": |
||||||
|
main(argv[1:]) |
@ -0,0 +1,63 @@ |
|||||||
|
#VERSION: 1.22 |
||||||
|
|
||||||
|
# Author: |
||||||
|
# Christophe DUMEZ (chris@qbittorrent.org) |
||||||
|
|
||||||
|
# Redistribution and use in source and binary forms, with or without |
||||||
|
# modification, are permitted provided that the following conditions are met: |
||||||
|
# |
||||||
|
# * Redistributions of source code must retain the above copyright notice, |
||||||
|
# this list of conditions and the following disclaimer. |
||||||
|
# * Redistributions in binary form must reproduce the above copyright |
||||||
|
# notice, this list of conditions and the following disclaimer in the |
||||||
|
# documentation and/or other materials provided with the distribution. |
||||||
|
# * Neither the name of the author nor the names of its contributors may be |
||||||
|
# used to endorse or promote products derived from this software without |
||||||
|
# specific prior written permission. |
||||||
|
# |
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
||||||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
||||||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
||||||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
||||||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
||||||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
||||||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
||||||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
||||||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
||||||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
||||||
|
# POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
import sys |
||||||
|
import os |
||||||
|
import glob |
||||||
|
from helpers import download_file |
||||||
|
|
||||||
|
supported_engines = dict() |
||||||
|
|
||||||
|
engines = glob.glob(os.path.join(os.path.dirname(__file__), 'engines', '*.py')) |
||||||
|
for engine in engines: |
||||||
|
e = engine.split(os.sep)[-1][:-3] |
||||||
|
if len(e.strip()) == 0: |
||||||
|
continue |
||||||
|
if e.startswith('_'): |
||||||
|
continue |
||||||
|
try: |
||||||
|
exec("from engines.%s import %s" % (e, e)) |
||||||
|
exec("engine_url = %s.url" % e) |
||||||
|
supported_engines[engine_url] = e |
||||||
|
except Exception: |
||||||
|
pass |
||||||
|
|
||||||
|
if __name__ == '__main__': |
||||||
|
if len(sys.argv) < 3: |
||||||
|
raise SystemExit('./nova2dl.py engine_url download_parameter') |
||||||
|
engine_url = sys.argv[1].strip() |
||||||
|
download_param = sys.argv[2].strip() |
||||||
|
if engine_url not in list(supported_engines.keys()): |
||||||
|
raise SystemExit('./nova2dl.py: this engine_url was not recognized') |
||||||
|
exec("engine = %s()" % supported_engines[engine_url]) |
||||||
|
if hasattr(engine, 'download_torrent'): |
||||||
|
engine.download_torrent(download_param) |
||||||
|
else: |
||||||
|
print(download_file(download_param)) |
||||||
|
sys.exit(0) |
@ -0,0 +1,67 @@ |
|||||||
|
#VERSION: 1.46 |
||||||
|
|
||||||
|
# Redistribution and use in source and binary forms, with or without |
||||||
|
# modification, are permitted provided that the following conditions are met: |
||||||
|
# |
||||||
|
# * Redistributions of source code must retain the above copyright notice, |
||||||
|
# this list of conditions and the following disclaimer. |
||||||
|
# * Redistributions in binary form must reproduce the above copyright |
||||||
|
# notice, this list of conditions and the following disclaimer in the |
||||||
|
# documentation and/or other materials provided with the distribution. |
||||||
|
# * Neither the name of the author nor the names of its contributors may be |
||||||
|
# used to endorse or promote products derived from this software without |
||||||
|
# specific prior written permission. |
||||||
|
# |
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
||||||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
||||||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
||||||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
||||||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
||||||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
||||||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
||||||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
||||||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
||||||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
||||||
|
# POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
|
||||||
|
def prettyPrinter(dictionary): |
||||||
|
dictionary['size'] = anySizeToBytes(dictionary['size']) |
||||||
|
outtext = "|".join((dictionary["link"], dictionary["name"].replace("|", " "), |
||||||
|
str(dictionary["size"]), str(dictionary["seeds"]), |
||||||
|
str(dictionary["leech"]), dictionary["engine_url"])) |
||||||
|
if 'desc_link' in dictionary: |
||||||
|
outtext = "|".join((outtext, dictionary["desc_link"])) |
||||||
|
|
||||||
|
# fd 1 is stdout |
||||||
|
with open(1, 'w', encoding='utf-8', closefd=False) as utf8stdout: |
||||||
|
print(outtext, file=utf8stdout) |
||||||
|
|
||||||
|
|
||||||
|
def anySizeToBytes(size_string): |
||||||
|
""" |
||||||
|
Convert a string like '1 KB' to '1024' (bytes) |
||||||
|
""" |
||||||
|
# separate integer from unit |
||||||
|
try: |
||||||
|
size, unit = size_string.split() |
||||||
|
except: |
||||||
|
try: |
||||||
|
size = size_string.strip() |
||||||
|
unit = ''.join([c for c in size if c.isalpha()]) |
||||||
|
if len(unit) > 0: |
||||||
|
size = size[:-len(unit)] |
||||||
|
except: |
||||||
|
return -1 |
||||||
|
if len(size) == 0: |
||||||
|
return -1 |
||||||
|
size = float(size) |
||||||
|
if len(unit) == 0: |
||||||
|
return int(size) |
||||||
|
short_unit = unit.upper()[0] |
||||||
|
|
||||||
|
# convert |
||||||
|
units_dict = {'T': 40, 'G': 30, 'M': 20, 'K': 10} |
||||||
|
if short_unit in units_dict: |
||||||
|
size = size * 2**units_dict[short_unit] |
||||||
|
return int(size) |
@ -0,0 +1,547 @@ |
|||||||
|
"""A parser for SGML, using the derived class as a static DTD.""" |
||||||
|
|
||||||
|
# XXX This only supports those SGML features used by HTML. |
||||||
|
|
||||||
|
# XXX There should be a way to distinguish between PCDATA (parsed |
||||||
|
# character data -- the normal case), RCDATA (replaceable character |
||||||
|
# data -- only char and entity references and end tags are special) |
||||||
|
# and CDATA (character data -- only end tags are special). RCDATA is |
||||||
|
# not supported at all. |
||||||
|
|
||||||
|
import _markupbase |
||||||
|
import re |
||||||
|
|
||||||
|
__all__ = ["SGMLParser", "SGMLParseError"] |
||||||
|
|
||||||
|
# Regular expressions used for parsing |
||||||
|
|
||||||
|
interesting = re.compile('[&<]') |
||||||
|
incomplete = re.compile('&([a-zA-Z][a-zA-Z0-9]*|#[0-9]*)?|' |
||||||
|
'<([a-zA-Z][^<>]*|' |
||||||
|
'/([a-zA-Z][^<>]*)?|' |
||||||
|
'![^<>]*)?') |
||||||
|
|
||||||
|
entityref = re.compile('&([a-zA-Z][-.a-zA-Z0-9]*)[^a-zA-Z0-9]') |
||||||
|
charref = re.compile('&#([0-9]+)[^0-9]') |
||||||
|
|
||||||
|
starttagopen = re.compile('<[>a-zA-Z]') |
||||||
|
shorttagopen = re.compile('<[a-zA-Z][-.a-zA-Z0-9]*/') |
||||||
|
shorttag = re.compile('<([a-zA-Z][-.a-zA-Z0-9]*)/([^/]*)/') |
||||||
|
piclose = re.compile('>') |
||||||
|
endbracket = re.compile('[<>]') |
||||||
|
tagfind = re.compile('[a-zA-Z][-_.a-zA-Z0-9]*') |
||||||
|
attrfind = re.compile( |
||||||
|
r'\s*([a-zA-Z_][-:.a-zA-Z_0-9]*)(\s*=\s*' |
||||||
|
r'(\'[^\']*\'|"[^"]*"|[][\-a-zA-Z0-9./,:;+*%?!&$\(\)_#=~\'"@]*))?') |
||||||
|
|
||||||
|
|
||||||
|
class SGMLParseError(RuntimeError): |
||||||
|
"""Exception raised for all parse errors.""" |
||||||
|
pass |
||||||
|
|
||||||
|
|
||||||
|
# SGML parser base class -- find tags and call handler functions. |
||||||
|
# Usage: p = SGMLParser(); p.feed(data); ...; p.close(). |
||||||
|
# The dtd is defined by deriving a class which defines methods |
||||||
|
# with special names to handle tags: start_foo and end_foo to handle |
||||||
|
# <foo> and </foo>, respectively, or do_foo to handle <foo> by itself. |
||||||
|
# (Tags are converted to lower case for this purpose.) The data |
||||||
|
# between tags is passed to the parser by calling self.handle_data() |
||||||
|
# with some data as argument (the data may be split up in arbitrary |
||||||
|
# chunks). Entity references are passed by calling |
||||||
|
# self.handle_entityref() with the entity reference as argument. |
||||||
|
|
||||||
|
class SGMLParser(_markupbase.ParserBase): |
||||||
|
# Definition of entities -- derived classes may override |
||||||
|
entity_or_charref = re.compile('&(?:' |
||||||
|
'([a-zA-Z][-.a-zA-Z0-9]*)|#([0-9]+)' |
||||||
|
')(;?)') |
||||||
|
|
||||||
|
def __init__(self, verbose=0): |
||||||
|
"""Initialize and reset this instance.""" |
||||||
|
self.verbose = verbose |
||||||
|
self.reset() |
||||||
|
|
||||||
|
def reset(self): |
||||||
|
"""Reset this instance. Loses all unprocessed data.""" |
||||||
|
self.__starttag_text = None |
||||||
|
self.rawdata = '' |
||||||
|
self.stack = [] |
||||||
|
self.lasttag = '???' |
||||||
|
self.nomoretags = 0 |
||||||
|
self.literal = 0 |
||||||
|
_markupbase.ParserBase.reset(self) |
||||||
|
|
||||||
|
def setnomoretags(self): |
||||||
|
"""Enter literal mode (CDATA) till EOF. |
||||||
|
|
||||||
|
Intended for derived classes only. |
||||||
|
""" |
||||||
|
self.nomoretags = self.literal = 1 |
||||||
|
|
||||||
|
def setliteral(self, *args): |
||||||
|
"""Enter literal mode (CDATA). |
||||||
|
|
||||||
|
Intended for derived classes only. |
||||||
|
""" |
||||||
|
self.literal = 1 |
||||||
|
|
||||||
|
def feed(self, data): |
||||||
|
"""Feed some data to the parser. |
||||||
|
|
||||||
|
Call this as often as you want, with as little or as much text |
||||||
|
as you want (may include '\n'). (This just saves the text, |
||||||
|
all the processing is done by goahead().) |
||||||
|
""" |
||||||
|
|
||||||
|
self.rawdata = self.rawdata + data |
||||||
|
self.goahead(0) |
||||||
|
|
||||||
|
def close(self): |
||||||
|
"""Handle the remaining data.""" |
||||||
|
self.goahead(1) |
||||||
|
|
||||||
|
def error(self, message): |
||||||
|
raise SGMLParseError(message) |
||||||
|
|
||||||
|
# Internal -- handle data as far as reasonable. May leave state |
||||||
|
# and data to be processed by a subsequent call. If 'end' is |
||||||
|
# true, force handling all data as if followed by EOF marker. |
||||||
|
def goahead(self, end): |
||||||
|
rawdata = self.rawdata |
||||||
|
i = 0 |
||||||
|
n = len(rawdata) |
||||||
|
while i < n: |
||||||
|
if self.nomoretags: |
||||||
|
self.handle_data(rawdata[i:n]) |
||||||
|
i = n |
||||||
|
break |
||||||
|
match = interesting.search(rawdata, i) |
||||||
|
if match: j = match.start() |
||||||
|
else: j = n |
||||||
|
if i < j: |
||||||
|
self.handle_data(rawdata[i:j]) |
||||||
|
i = j |
||||||
|
if i == n: break |
||||||
|
if rawdata[i] == '<': |
||||||
|
if starttagopen.match(rawdata, i): |
||||||
|
if self.literal: |
||||||
|
self.handle_data(rawdata[i]) |
||||||
|
i = i+1 |
||||||
|
continue |
||||||
|
k = self.parse_starttag(i) |
||||||
|
if k < 0: break |
||||||
|
i = k |
||||||
|
continue |
||||||
|
if rawdata.startswith("</", i): |
||||||
|
k = self.parse_endtag(i) |
||||||
|
if k < 0: break |
||||||
|
i = k |
||||||
|
self.literal = 0 |
||||||
|
continue |
||||||
|
if self.literal: |
||||||
|
if n > (i + 1): |
||||||
|
self.handle_data("<") |
||||||
|
i = i+1 |
||||||
|
else: |
||||||
|
# incomplete |
||||||
|
break |
||||||
|
continue |
||||||
|
if rawdata.startswith("<!--", i): |
||||||
|
# Strictly speaking, a comment is --.*-- |
||||||
|
# within a declaration tag <!...>. |
||||||
|
# This should be removed, |
||||||
|
# and comments handled only in parse_declaration. |
||||||
|
k = self.parse_comment(i) |
||||||
|
if k < 0: break |
||||||
|
i = k |
||||||
|
continue |
||||||
|
if rawdata.startswith("<?", i): |
||||||
|
k = self.parse_pi(i) |
||||||
|
if k < 0: break |
||||||
|
i = i+k |
||||||
|
continue |
||||||
|
if rawdata.startswith("<!", i): |
||||||
|
# This is some sort of declaration; in "HTML as |
||||||
|
# deployed," this should only be the document type |
||||||
|
# declaration ("<!DOCTYPE html...>"). |
||||||
|
k = self.parse_declaration(i) |
||||||
|
if k < 0: break |
||||||
|
i = k |
||||||
|
continue |
||||||
|
elif rawdata[i] == '&': |
||||||
|
if self.literal: |
||||||
|
self.handle_data(rawdata[i]) |
||||||
|
i = i+1 |
||||||
|
continue |
||||||
|
match = charref.match(rawdata, i) |
||||||
|
if match: |
||||||
|
name = match.group(1) |
||||||
|
self.handle_charref(name) |
||||||
|
i = match.end(0) |
||||||
|
if rawdata[i-1] != ';': i = i-1 |
||||||
|
continue |
||||||
|
match = entityref.match(rawdata, i) |
||||||
|
if match: |
||||||
|
name = match.group(1) |
||||||
|
self.handle_entityref(name) |
||||||
|
i = match.end(0) |
||||||
|
if rawdata[i-1] != ';': i = i-1 |
||||||
|
continue |
||||||
|
else: |
||||||
|
self.error('neither < nor & ??') |
||||||
|
# We get here only if incomplete matches but |
||||||
|
# nothing else |
||||||
|
match = incomplete.match(rawdata, i) |
||||||
|
if not match: |
||||||
|
self.handle_data(rawdata[i]) |
||||||
|
i = i+1 |
||||||
|
continue |
||||||
|
j = match.end(0) |
||||||
|
if j == n: |
||||||
|
break # Really incomplete |
||||||
|
self.handle_data(rawdata[i:j]) |
||||||
|
i = j |
||||||
|
# end while |
||||||
|
if end and i < n: |
||||||
|
self.handle_data(rawdata[i:n]) |
||||||
|
i = n |
||||||
|
self.rawdata = rawdata[i:] |
||||||
|
# XXX if end: check for empty stack |
||||||
|
|
||||||
|
# Extensions for the DOCTYPE scanner: |
||||||
|
_decl_otherchars = '=' |
||||||
|
|
||||||
|
# Internal -- parse processing instr, return length or -1 if not terminated |
||||||
|
def parse_pi(self, i): |
||||||
|
rawdata = self.rawdata |
||||||
|
if rawdata[i:i+2] != '<?': |
||||||
|
self.error('unexpected call to parse_pi()') |
||||||
|
match = piclose.search(rawdata, i+2) |
||||||
|
if not match: |
||||||
|
return -1 |
||||||
|
j = match.start(0) |
||||||
|
self.handle_pi(rawdata[i+2: j]) |
||||||
|
j = match.end(0) |
||||||
|
return j-i |
||||||
|
|
||||||
|
def get_starttag_text(self): |
||||||
|
return self.__starttag_text |
||||||
|
|
||||||
|
# Internal -- handle starttag, return length or -1 if not terminated |
||||||
|
def parse_starttag(self, i): |
||||||
|
self.__starttag_text = None |
||||||
|
start_pos = i |
||||||
|
rawdata = self.rawdata |
||||||
|
if shorttagopen.match(rawdata, i): |
||||||
|
# SGML shorthand: <tag/data/ == <tag>data</tag> |
||||||
|
# XXX Can data contain &... (entity or char refs)? |
||||||
|
# XXX Can data contain < or > (tag characters)? |
||||||
|
# XXX Can there be whitespace before the first /? |
||||||
|
match = shorttag.match(rawdata, i) |
||||||
|
if not match: |
||||||
|
return -1 |
||||||
|
tag, data = match.group(1, 2) |
||||||
|
self.__starttag_text = '<%s/' % tag |
||||||
|
tag = tag.lower() |
||||||
|
k = match.end(0) |
||||||
|
self.finish_shorttag(tag, data) |
||||||
|
self.__starttag_text = rawdata[start_pos:match.end(1) + 1] |
||||||
|
return k |
||||||
|
# XXX The following should skip matching quotes (' or ") |
||||||
|
# As a shortcut way to exit, this isn't so bad, but shouldn't |
||||||
|
# be used to locate the actual end of the start tag since the |
||||||
|
# < or > characters may be embedded in an attribute value. |
||||||
|
match = endbracket.search(rawdata, i+1) |
||||||
|
if not match: |
||||||
|
return -1 |
||||||
|
j = match.start(0) |
||||||
|
# Now parse the data between i+1 and j into a tag and attrs |
||||||
|
attrs = [] |
||||||
|
if rawdata[i:i+2] == '<>': |
||||||
|
# SGML shorthand: <> == <last open tag seen> |
||||||
|
k = j |
||||||
|
tag = self.lasttag |
||||||
|
else: |
||||||
|
match = tagfind.match(rawdata, i+1) |
||||||
|
if not match: |
||||||
|
self.error('unexpected call to parse_starttag') |
||||||
|
k = match.end(0) |
||||||
|
tag = rawdata[i+1:k].lower() |
||||||
|
self.lasttag = tag |
||||||
|
while k < j: |
||||||
|
match = attrfind.match(rawdata, k) |
||||||
|
if not match: break |
||||||
|
attrname, rest, attrvalue = match.group(1, 2, 3) |
||||||
|
if not rest: |
||||||
|
attrvalue = attrname |
||||||
|
else: |
||||||
|
if (attrvalue[:1] == "'" == attrvalue[-1:] or |
||||||
|
attrvalue[:1] == '"' == attrvalue[-1:]): |
||||||
|
# strip quotes |
||||||
|
attrvalue = attrvalue[1:-1] |
||||||
|
attrvalue = self.entity_or_charref.sub( |
||||||
|
self._convert_ref, attrvalue) |
||||||
|
attrs.append((attrname.lower(), attrvalue)) |
||||||
|
k = match.end(0) |
||||||
|
if rawdata[j] == '>': |
||||||
|
j = j+1 |
||||||
|
self.__starttag_text = rawdata[start_pos:j] |
||||||
|
self.finish_starttag(tag, attrs) |
||||||
|
return j |
||||||
|
|
||||||
|
# Internal -- convert entity or character reference |
||||||
|
def _convert_ref(self, match): |
||||||
|
if match.group(2): |
||||||
|
return self.convert_charref(match.group(2)) or \ |
||||||
|
'&#%s%s' % match.groups()[1:] |
||||||
|
elif match.group(3): |
||||||
|
return self.convert_entityref(match.group(1)) or \ |
||||||
|
'&%s;' % match.group(1) |
||||||
|
else: |
||||||
|
return '&%s' % match.group(1) |
||||||
|
|
||||||
|
# Internal -- parse endtag |
||||||
|
def parse_endtag(self, i): |
||||||
|
rawdata = self.rawdata |
||||||
|
match = endbracket.search(rawdata, i+1) |
||||||
|
if not match: |
||||||
|
return -1 |
||||||
|
j = match.start(0) |
||||||
|
tag = rawdata[i+2:j].strip().lower() |
||||||
|
if rawdata[j] == '>': |
||||||
|
j = j+1 |
||||||
|
self.finish_endtag(tag) |
||||||
|
return j |
||||||
|
|
||||||
|
# Internal -- finish parsing of <tag/data/ (same as <tag>data</tag>) |
||||||
|
def finish_shorttag(self, tag, data): |
||||||
|
self.finish_starttag(tag, []) |
||||||
|
self.handle_data(data) |
||||||
|
self.finish_endtag(tag) |
||||||
|
|
||||||
|
# Internal -- finish processing of start tag |
||||||
|
# Return -1 for unknown tag, 0 for open-only tag, 1 for balanced tag |
||||||
|
def finish_starttag(self, tag, attrs): |
||||||
|
try: |
||||||
|
method = getattr(self, 'start_' + tag) |
||||||
|
except AttributeError: |
||||||
|
try: |
||||||
|
method = getattr(self, 'do_' + tag) |
||||||
|
except AttributeError: |
||||||
|
self.unknown_starttag(tag, attrs) |
||||||
|
return -1 |
||||||
|
else: |
||||||
|
self.handle_starttag(tag, method, attrs) |
||||||
|
return 0 |
||||||
|
else: |
||||||
|
self.stack.append(tag) |
||||||
|
self.handle_starttag(tag, method, attrs) |
||||||
|
return 1 |
||||||
|
|
||||||
|
# Internal -- finish processing of end tag |
||||||
|
def finish_endtag(self, tag): |
||||||
|
if not tag: |
||||||
|
found = len(self.stack) - 1 |
||||||
|
if found < 0: |
||||||
|
self.unknown_endtag(tag) |
||||||
|
return |
||||||
|
else: |
||||||
|
if tag not in self.stack: |
||||||
|
try: |
||||||
|
method = getattr(self, 'end_' + tag) |
||||||
|
except AttributeError: |
||||||
|
self.unknown_endtag(tag) |
||||||
|
else: |
||||||
|
self.report_unbalanced(tag) |
||||||
|
return |
||||||
|
found = len(self.stack) |
||||||
|
for i in range(found): |
||||||
|
if self.stack[i] == tag: found = i |
||||||
|
while len(self.stack) > found: |
||||||
|
tag = self.stack[-1] |
||||||
|
try: |
||||||
|
method = getattr(self, 'end_' + tag) |
||||||
|
except AttributeError: |
||||||
|
method = None |
||||||
|
if method: |
||||||
|
self.handle_endtag(tag, method) |
||||||
|
else: |
||||||
|
self.unknown_endtag(tag) |
||||||
|
del self.stack[-1] |
||||||
|
|
||||||
|
# Overridable -- handle start tag |
||||||
|
def handle_starttag(self, tag, method, attrs): |
||||||
|
method(attrs) |
||||||
|
|
||||||
|
# Overridable -- handle end tag |
||||||
|
def handle_endtag(self, tag, method): |
||||||
|
method() |
||||||
|
|
||||||
|
# Example -- report an unbalanced </...> tag. |
||||||
|
def report_unbalanced(self, tag): |
||||||
|
if self.verbose: |
||||||
|
print('*** Unbalanced </' + tag + '>') |
||||||
|
print('*** Stack:', self.stack) |
||||||
|
|
||||||
|
def convert_charref(self, name): |
||||||
|
"""Convert character reference, may be overridden.""" |
||||||
|
try: |
||||||
|
n = int(name) |
||||||
|
except ValueError: |
||||||
|
return |
||||||
|
if not 0 <= n <= 127: |
||||||
|
return |
||||||
|
return self.convert_codepoint(n) |
||||||
|
|
||||||
|
def convert_codepoint(self, codepoint): |
||||||
|
return chr(codepoint) |
||||||
|
|
||||||
|
def handle_charref(self, name): |
||||||
|
"""Handle character reference, no need to override.""" |
||||||
|
replacement = self.convert_charref(name) |
||||||
|
if replacement is None: |
||||||
|
self.unknown_charref(name) |
||||||
|
else: |
||||||
|
self.handle_data(replacement) |
||||||
|
|
||||||
|
# Definition of entities -- derived classes may override |
||||||
|
entitydefs = \ |
||||||
|
{'lt': '<', 'gt': '>', 'amp': '&', 'quot': '"', 'apos': '\''} |
||||||
|
|
||||||
|
def convert_entityref(self, name): |
||||||
|
"""Convert entity references. |
||||||
|
|
||||||
|
As an alternative to overriding this method; one can tailor the |
||||||
|
results by setting up the self.entitydefs mapping appropriately. |
||||||
|
""" |
||||||
|
table = self.entitydefs |
||||||
|
if name in table: |
||||||
|
return table[name] |
||||||
|
else: |
||||||
|
return |
||||||
|
|
||||||
|
def handle_entityref(self, name): |
||||||
|
"""Handle entity references, no need to override.""" |
||||||
|
replacement = self.convert_entityref(name) |
||||||
|
if replacement is None: |
||||||
|
self.unknown_entityref(name) |
||||||
|
else: |
||||||
|
self.handle_data(replacement) |
||||||
|
|
||||||
|
# Example -- handle data, should be overridden |
||||||
|
def handle_data(self, data): |
||||||
|
pass |
||||||
|
|
||||||
|
# Example -- handle comment, could be overridden |
||||||
|
def handle_comment(self, data): |
||||||
|
pass |
||||||
|
|
||||||
|
# Example -- handle declaration, could be overridden |
||||||
|
def handle_decl(self, decl): |
||||||
|
pass |
||||||
|
|
||||||
|
# Example -- handle processing instruction, could be overridden |
||||||
|
def handle_pi(self, data): |
||||||
|
pass |
||||||
|
|
||||||
|
# To be overridden -- handlers for unknown objects |
||||||
|
def unknown_starttag(self, tag, attrs): pass |
||||||
|
def unknown_endtag(self, tag): pass |
||||||
|
def unknown_charref(self, ref): pass |
||||||
|
def unknown_entityref(self, ref): pass |
||||||
|
|
||||||
|
|
||||||
|
class TestSGMLParser(SGMLParser): |
||||||
|
|
||||||
|
def __init__(self, verbose=0): |
||||||
|
self.testdata = "" |
||||||
|
SGMLParser.__init__(self, verbose) |
||||||
|
|
||||||
|
def handle_data(self, data): |
||||||
|
self.testdata = self.testdata + data |
||||||
|
if len(repr(self.testdata)) >= 70: |
||||||
|
self.flush() |
||||||
|
|
||||||
|
def flush(self): |
||||||
|
data = self.testdata |
||||||
|
if data: |
||||||
|
self.testdata = "" |
||||||
|
print('data:', repr(data)) |
||||||
|
|
||||||
|
def handle_comment(self, data): |
||||||
|
self.flush() |
||||||
|
r = repr(data) |
||||||
|
if len(r) > 68: |
||||||
|
r = r[:32] + '...' + r[-32:] |
||||||
|
print('comment:', r) |
||||||
|
|
||||||
|
def unknown_starttag(self, tag, attrs): |
||||||
|
self.flush() |
||||||
|
if not attrs: |
||||||
|
print('start tag: <' + tag + '>') |
||||||
|
else: |
||||||
|
print('start tag: <' + tag, end=' ') |
||||||
|
for name, value in attrs: |
||||||
|
print(name + '=' + '"' + value + '"', end=' ') |
||||||
|
print('>') |
||||||
|
|
||||||
|
def unknown_endtag(self, tag): |
||||||
|
self.flush() |
||||||
|
print('end tag: </' + tag + '>') |
||||||
|
|
||||||
|
def unknown_entityref(self, ref): |
||||||
|
self.flush() |
||||||
|
print('*** unknown entity ref: &' + ref + ';') |
||||||
|
|
||||||
|
def unknown_charref(self, ref): |
||||||
|
self.flush() |
||||||
|
print('*** unknown char ref: &#' + ref + ';') |
||||||
|
|
||||||
|
def unknown_decl(self, data): |
||||||
|
self.flush() |
||||||
|
print('*** unknown decl: [' + data + ']') |
||||||
|
|
||||||
|
def close(self): |
||||||
|
SGMLParser.close(self) |
||||||
|
self.flush() |
||||||
|
|
||||||
|
|
||||||
|
def test(args = None): |
||||||
|
import sys |
||||||
|
|
||||||
|
if args is None: |
||||||
|
args = sys.argv[1:] |
||||||
|
|
||||||
|
if args and args[0] == '-s': |
||||||
|
args = args[1:] |
||||||
|
klass = SGMLParser |
||||||
|
else: |
||||||
|
klass = TestSGMLParser |
||||||
|
|
||||||
|
if args: |
||||||
|
file = args[0] |
||||||
|
else: |
||||||
|
file = 'test.html' |
||||||
|
|
||||||
|
if file == '-': |
||||||
|
f = sys.stdin |
||||||
|
else: |
||||||
|
try: |
||||||
|
f = open(file, 'r') |
||||||
|
except IOError as msg: |
||||||
|
print(file, ":", msg) |
||||||
|
sys.exit(1) |
||||||
|
|
||||||
|
data = f.read() |
||||||
|
if f is not sys.stdin: |
||||||
|
f.close() |
||||||
|
|
||||||
|
x = klass() |
||||||
|
for c in data: |
||||||
|
x.feed(c) |
||||||
|
x.close() |
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__': |
||||||
|
test() |
@ -0,0 +1,391 @@ |
|||||||
|
"""SocksiPy - Python SOCKS module. |
||||||
|
Version 1.01 |
||||||
|
|
||||||
|
Copyright 2006 Dan-Haim. All rights reserved. |
||||||
|
Various fixes by Christophe DUMEZ <chris@qbittorrent.org> - 2010 |
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without modification, |
||||||
|
are permitted provided that the following conditions are met: |
||||||
|
1. Redistributions of source code must retain the above copyright notice, this |
||||||
|
list of conditions and the following disclaimer. |
||||||
|
2. Redistributions in binary form must reproduce the above copyright notice, |
||||||
|
this list of conditions and the following disclaimer in the documentation |
||||||
|
and/or other materials provided with the distribution. |
||||||
|
3. Neither the name of Dan Haim nor the names of his contributors may be used |
||||||
|
to endorse or promote products derived from this software without specific |
||||||
|
prior written permission. |
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY DAN HAIM "AS IS" AND ANY EXPRESS OR IMPLIED |
||||||
|
WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF |
||||||
|
MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
||||||
|
EVENT SHALL DAN HAIM OR HIS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, |
||||||
|
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA |
||||||
|
OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF |
||||||
|
LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT |
||||||
|
OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMANGE. |
||||||
|
|
||||||
|
|
||||||
|
This module provides a standard socket-like interface for Python |
||||||
|
for tunneling connections through SOCKS proxies. |
||||||
|
|
||||||
|
""" |
||||||
|
|
||||||
|
import socket |
||||||
|
import struct |
||||||
|
|
||||||
|
PROXY_TYPE_SOCKS4 = 1 |
||||||
|
PROXY_TYPE_SOCKS5 = 2 |
||||||
|
PROXY_TYPE_HTTP = 3 |
||||||
|
|
||||||
|
_defaultproxy = None |
||||||
|
_orgsocket = socket.socket |
||||||
|
|
||||||
|
class ProxyError(Exception): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
class GeneralProxyError(ProxyError): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
class Socks5AuthError(ProxyError): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
class Socks5Error(ProxyError): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
class Socks4Error(ProxyError): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
class HTTPError(ProxyError): |
||||||
|
def __init__(self, value): |
||||||
|
self.value = value |
||||||
|
def __str__(self): |
||||||
|
return repr(self.value) |
||||||
|
|
||||||
|
_generalerrors = ("success", |
||||||
|
"invalid data", |
||||||
|
"not connected", |
||||||
|
"not available", |
||||||
|
"bad proxy type", |
||||||
|
"bad input") |
||||||
|
|
||||||
|
_socks5errors = ("succeeded", |
||||||
|
"general SOCKS server failure", |
||||||
|
"connection not allowed by ruleset", |
||||||
|
"Network unreachable", |
||||||
|
"Host unreachable", |
||||||
|
"Connection refused", |
||||||
|
"TTL expired", |
||||||
|
"Command not supported", |
||||||
|
"Address type not supported", |
||||||
|
"Unknown error") |
||||||
|
|
||||||
|
_socks5autherrors = ("succeeded", |
||||||
|
"authentication is required", |
||||||
|
"all offered authentication methods were rejected", |
||||||
|
"unknown username or invalid password", |
||||||
|
"unknown error") |
||||||
|
|
||||||
|
_socks4errors = ("request granted", |
||||||
|
"request rejected or failed", |
||||||
|
"request rejected because SOCKS server cannot connect to identd on the client", |
||||||
|
"request rejected because the client program and identd report different user-ids", |
||||||
|
"unknown error") |
||||||
|
|
||||||
|
def setdefaultproxy(proxytype=None,addr=None,port=None,rdns=True,username=None,password=None): |
||||||
|
"""setdefaultproxy(proxytype, addr[, port[, rdns[, username[, password]]]]) |
||||||
|
Sets a default proxy which all further socksocket objects will use, |
||||||
|
unless explicitly changed. |
||||||
|
""" |
||||||
|
global _defaultproxy |
||||||
|
_defaultproxy = (proxytype,addr,port,rdns,username,password) |
||||||
|
|
||||||
|
class socksocket(socket.socket): |
||||||
|
"""socksocket([family[, type[, proto]]]) -> socket object |
||||||
|
|
||||||
|
Open a SOCKS enabled socket. The parameters are the same as |
||||||
|
those of the standard socket init. In order for SOCKS to work, |
||||||
|
you must specify family=AF_INET, type=SOCK_STREAM and proto=0. |
||||||
|
""" |
||||||
|
|
||||||
|
def __init__(self, family=socket.AF_INET, type=socket.SOCK_STREAM, proto=0, _sock=None): |
||||||
|
_orgsocket.__init__(self,family,type,proto,_sock) |
||||||
|
if _defaultproxy != None: |
||||||
|
self.__proxy = _defaultproxy |
||||||
|
else: |
||||||
|
self.__proxy = (None, None, None, None, None, None) |
||||||
|
self.__proxysockname = None |
||||||
|
self.__proxypeername = None |
||||||
|
|
||||||
|
def __recvall(self, bytes): |
||||||
|
"""__recvall(bytes) -> data |
||||||
|
Receive EXACTLY the number of bytes requested from the socket. |
||||||
|
Blocks until the required number of bytes have been received. |
||||||
|
""" |
||||||
|
data = "" |
||||||
|
while len(data) < bytes: |
||||||
|
d = self.recv(bytes-len(data)) |
||||||
|
if not d: |
||||||
|
raise GeneralProxyError("connection closed unexpectedly") |
||||||
|
data = data + d |
||||||
|
return data |
||||||
|
|
||||||
|
def setproxy(self,proxytype=None,addr=None,port=None,rdns=True,username=None,password=None): |
||||||
|
"""setproxy(proxytype, addr[, port[, rdns[, username[, password]]]]) |
||||||
|
Sets the proxy to be used. |
||||||
|
proxytype - The type of the proxy to be used. Three types |
||||||
|
are supported: PROXY_TYPE_SOCKS4 (including socks4a), |
||||||
|
PROXY_TYPE_SOCKS5 and PROXY_TYPE_HTTP |
||||||
|
addr - The address of the server (IP or DNS). |
||||||
|
port - The port of the server. Defaults to 1080 for SOCKS |
||||||
|
servers and 8080 for HTTP proxy servers. |
||||||
|
rdns - Should DNS queries be preformed on the remote side |
||||||
|
(rather than the local side). The default is True. |
||||||
|
Note: This has no effect with SOCKS4 servers. |
||||||
|
username - Username to authenticate with to the server. |
||||||
|
The default is no authentication. |
||||||
|
password - Password to authenticate with to the server. |
||||||
|
Only relevant when username is also provided. |
||||||
|
""" |
||||||
|
self.__proxy = (proxytype,addr,port,rdns,username,password) |
||||||
|
|
||||||
|
def __negotiatesocks5(self,destaddr,destport): |
||||||
|
"""__negotiatesocks5(self,destaddr,destport) |
||||||
|
Negotiates a connection through a SOCKS5 server. |
||||||
|
""" |
||||||
|
# First we'll send the authentication packages we support. |
||||||
|
if (self.__proxy[4]!=None) and (self.__proxy[5]!=None): |
||||||
|
# The username/password details were supplied to the |
||||||
|
# setproxy method so we support the USERNAME/PASSWORD |
||||||
|
# authentication (in addition to the standard none). |
||||||
|
self.sendall("\x05\x02\x00\x02") |
||||||
|
else: |
||||||
|
# No username/password were entered, therefore we |
||||||
|
# only support connections with no authentication. |
||||||
|
self.sendall("\x05\x01\x00") |
||||||
|
# We'll receive the server's response to determine which |
||||||
|
# method was selected |
||||||
|
chosenauth = self.__recvall(2) |
||||||
|
if chosenauth[0] != "\x05": |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
# Check the chosen authentication method |
||||||
|
if chosenauth[1] == "\x00": |
||||||
|
# No authentication is required |
||||||
|
pass |
||||||
|
elif chosenauth[1] == "\x02": |
||||||
|
# Okay, we need to perform a basic username/password |
||||||
|
# authentication. |
||||||
|
self.sendall("\x01" + chr(len(self.__proxy[4])) + self.__proxy[4] + chr(len(self.__proxy[5])) + self.__proxy[5]) |
||||||
|
authstat = self.__recvall(2) |
||||||
|
if authstat[0] != "\x01": |
||||||
|
# Bad response |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
if authstat[1] != "\x00": |
||||||
|
# Authentication failed |
||||||
|
self.close() |
||||||
|
raise Socks5AuthError((3,_socks5autherrors[3])) |
||||||
|
# Authentication succeeded |
||||||
|
else: |
||||||
|
# Reaching here is always bad |
||||||
|
self.close() |
||||||
|
if chosenauth[1] == "\xFF": |
||||||
|
raise Socks5AuthError((2,_socks5autherrors[2])) |
||||||
|
else: |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
# Now we can request the actual connection |
||||||
|
req = "\x05\x01\x00" |
||||||
|
# If the given destination address is an IP address, we'll |
||||||
|
# use the IPv4 address request even if remote resolving was specified. |
||||||
|
try: |
||||||
|
ipaddr = socket.inet_aton(destaddr) |
||||||
|
req = req + "\x01" + ipaddr |
||||||
|
except socket.error: |
||||||
|
# Well it's not an IP number, so it's probably a DNS name. |
||||||
|
if self.__proxy[3]==True: |
||||||
|
# Resolve remotely |
||||||
|
ipaddr = None |
||||||
|
req = req + "\x03" + chr(len(destaddr)) + destaddr |
||||||
|
else: |
||||||
|
# Resolve locally |
||||||
|
ipaddr = socket.inet_aton(socket.gethostbyname(destaddr)) |
||||||
|
req = req + "\x01" + ipaddr |
||||||
|
req = req + struct.pack(">H",destport) |
||||||
|
self.sendall(req) |
||||||
|
# Get the response |
||||||
|
resp = self.__recvall(4) |
||||||
|
if resp[0] != "\x05": |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
elif resp[1] != "\x00": |
||||||
|
# Connection failed |
||||||
|
self.close() |
||||||
|
if ord(resp[1])<=8: |
||||||
|
raise Socks5Error((ord(resp[1]),_generalerrors[ord(resp[1])])) |
||||||
|
else: |
||||||
|
raise Socks5Error((9,_generalerrors[9])) |
||||||
|
# Get the bound address/port |
||||||
|
elif resp[3] == "\x01": |
||||||
|
boundaddr = self.__recvall(4) |
||||||
|
elif resp[3] == "\x03": |
||||||
|
resp = resp + self.recv(1) |
||||||
|
boundaddr = self.__recvall(ord(resp[4])) |
||||||
|
else: |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
boundport = struct.unpack(">H",self.__recvall(2))[0] |
||||||
|
self.__proxysockname = (boundaddr,boundport) |
||||||
|
if ipaddr != None: |
||||||
|
self.__proxypeername = (socket.inet_ntoa(ipaddr),destport) |
||||||
|
else: |
||||||
|
self.__proxypeername = (destaddr,destport) |
||||||
|
|
||||||
|
def getproxysockname(self): |
||||||
|
"""getsockname() -> address info |
||||||
|
Returns the bound IP address and port number at the proxy. |
||||||
|
""" |
||||||
|
return self.__proxysockname |
||||||
|
|
||||||
|
def getproxypeername(self): |
||||||
|
"""getproxypeername() -> address info |
||||||
|
Returns the IP and port number of the proxy. |
||||||
|
""" |
||||||
|
return _orgsocket.getpeername(self) |
||||||
|
|
||||||
|
def getpeername(self): |
||||||
|
"""getpeername() -> address info |
||||||
|
Returns the IP address and port number of the destination |
||||||
|
machine (note: getproxypeername returns the proxy) |
||||||
|
""" |
||||||
|
return self.__proxypeername |
||||||
|
|
||||||
|
def __negotiatesocks4(self,destaddr,destport): |
||||||
|
"""__negotiatesocks4(self,destaddr,destport) |
||||||
|
Negotiates a connection through a SOCKS4 server. |
||||||
|
""" |
||||||
|
# Check if the destination address provided is an IP address |
||||||
|
rmtrslv = False |
||||||
|
try: |
||||||
|
ipaddr = socket.inet_aton(destaddr) |
||||||
|
except socket.error: |
||||||
|
# It's a DNS name. Check where it should be resolved. |
||||||
|
if self.__proxy[3]==True: |
||||||
|
ipaddr = "\x00\x00\x00\x01" |
||||||
|
rmtrslv = True |
||||||
|
else: |
||||||
|
ipaddr = socket.inet_aton(socket.gethostbyname(destaddr)) |
||||||
|
# Construct the request packet |
||||||
|
req = "\x04\x01" + struct.pack(">H",destport) + ipaddr |
||||||
|
# The username parameter is considered userid for SOCKS4 |
||||||
|
if self.__proxy[4] != None: |
||||||
|
req = req + self.__proxy[4] |
||||||
|
req = req + "\x00" |
||||||
|
# DNS name if remote resolving is required |
||||||
|
# NOTE: This is actually an extension to the SOCKS4 protocol |
||||||
|
# called SOCKS4A and may not be supported in all cases. |
||||||
|
if rmtrslv==True: |
||||||
|
req = req + destaddr + "\x00" |
||||||
|
self.sendall(req) |
||||||
|
# Get the response from the server |
||||||
|
resp = self.__recvall(8) |
||||||
|
if resp[0] != "\x00": |
||||||
|
# Bad data |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
if resp[1] != "\x5A": |
||||||
|
# Server returned an error |
||||||
|
self.close() |
||||||
|
if ord(resp[1]) in (91,92,93): |
||||||
|
self.close() |
||||||
|
raise Socks4Error((ord(resp[1]),_socks4errors[ord(resp[1])-90])) |
||||||
|
else: |
||||||
|
raise Socks4Error((94,_socks4errors[4])) |
||||||
|
# Get the bound address/port |
||||||
|
self.__proxysockname = (socket.inet_ntoa(resp[4:]),struct.unpack(">H",resp[2:4])[0]) |
||||||
|
if rmtrslv != None: |
||||||
|
self.__proxypeername = (socket.inet_ntoa(ipaddr),destport) |
||||||
|
else: |
||||||
|
self.__proxypeername = (destaddr,destport) |
||||||
|
|
||||||
|
def __negotiatehttp(self,destaddr,destport): |
||||||
|
"""__negotiatehttp(self,destaddr,destport) |
||||||
|
Negotiates a connection through an HTTP server. |
||||||
|
""" |
||||||
|
# If we need to resolve locally, we do this now |
||||||
|
if self.__proxy[3] == False: |
||||||
|
addr = socket.gethostbyname(destaddr) |
||||||
|
else: |
||||||
|
addr = destaddr |
||||||
|
self.sendall("CONNECT " + addr + ":" + str(destport) + " HTTP/1.1\r\n" + "Host: " + destaddr + "\r\n\r\n") |
||||||
|
# We read the response until we get the string "\r\n\r\n" |
||||||
|
resp = self.recv(1) |
||||||
|
while resp.find("\r\n\r\n")==-1: |
||||||
|
resp = resp + self.recv(1) |
||||||
|
# We just need the first line to check if the connection |
||||||
|
# was successful |
||||||
|
statusline = resp.splitlines()[0].split(" ",2) |
||||||
|
if statusline[0] not in ("HTTP/1.0","HTTP/1.1"): |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
try: |
||||||
|
statuscode = int(statusline[1]) |
||||||
|
except ValueError: |
||||||
|
self.close() |
||||||
|
raise GeneralProxyError((1,_generalerrors[1])) |
||||||
|
if statuscode != 200: |
||||||
|
self.close() |
||||||
|
raise HTTPError((statuscode,statusline[2])) |
||||||
|
self.__proxysockname = ("0.0.0.0",0) |
||||||
|
self.__proxypeername = (addr,destport) |
||||||
|
|
||||||
|
def connect(self,destpair): |
||||||
|
"""connect(self,despair) |
||||||
|
Connects to the specified destination through a proxy. |
||||||
|
destpar - A tuple of the IP/DNS address and the port number. |
||||||
|
(identical to socket's connect). |
||||||
|
To select the proxy server use setproxy(). |
||||||
|
""" |
||||||
|
# Do a minimal input check first |
||||||
|
if (type(destpair) in (list,tuple)==False) or (len(destpair)<2) or (type(destpair[0])!=str) or (type(destpair[1])!=int): |
||||||
|
raise GeneralProxyError((5,_generalerrors[5])) |
||||||
|
if self.__proxy[0] == PROXY_TYPE_SOCKS5: |
||||||
|
if self.__proxy[2] != None: |
||||||
|
portnum = self.__proxy[2] |
||||||
|
else: |
||||||
|
portnum = 1080 |
||||||
|
_orgsocket.connect(self,(self.__proxy[1],portnum)) |
||||||
|
self.__negotiatesocks5(destpair[0],destpair[1]) |
||||||
|
elif self.__proxy[0] == PROXY_TYPE_SOCKS4: |
||||||
|
if self.__proxy[2] != None: |
||||||
|
portnum = self.__proxy[2] |
||||||
|
else: |
||||||
|
portnum = 1080 |
||||||
|
_orgsocket.connect(self,(self.__proxy[1],portnum)) |
||||||
|
self.__negotiatesocks4(destpair[0],destpair[1]) |
||||||
|
elif self.__proxy[0] == PROXY_TYPE_HTTP: |
||||||
|
if self.__proxy[2] != None: |
||||||
|
portnum = self.__proxy[2] |
||||||
|
else: |
||||||
|
portnum = 8080 |
||||||
|
_orgsocket.connect(self,(self.__proxy[1],portnum)) |
||||||
|
self.__negotiatehttp(destpair[0],destpair[1]) |
||||||
|
elif self.__proxy[0] == None: |
||||||
|
_orgsocket.connect(self,(destpair[0],destpair[1])) |
||||||
|
else: |
||||||
|
raise GeneralProxyError((4,_generalerrors[4])) |
Loading…
Reference in new issue