mirror of https://github.com/YGGverse/qBt_SE.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
244 lines
8.9 KiB
244 lines
8.9 KiB
# VERSION: 1.0 |
|
# AUTHORS: imDMG |
|
|
|
# LICENSING INFORMATION |
|
|
|
import tempfile |
|
import os |
|
import logging |
|
import time |
|
|
|
from urllib.request import build_opener, HTTPCookieProcessor, ProxyHandler |
|
from urllib.parse import urlencode, quote, unquote |
|
from urllib.error import URLError, HTTPError |
|
from http.cookiejar import CookieJar |
|
from html.parser import HTMLParser |
|
from novaprinter import prettyPrinter |
|
|
|
# setup logging into qBittorrent/logs |
|
logging.basicConfig(level=logging.DEBUG, |
|
format='%(asctime)s %(name)-12s %(levelname)-8s %(message)s', |
|
datefmt='%m-%d %H:%M', |
|
filename=os.path.abspath( |
|
os.path.join(os.path.dirname(__file__), '../..', 'logs')) + "/kinozal_se.log", |
|
# filename="kinozal_se.log", |
|
filemode='w') |
|
|
|
# benchmark |
|
# start_time = time.time() |
|
|
|
|
|
class kinozal(object): |
|
name = 'Kinozal' |
|
url = 'http://kinozal.tv' |
|
supported_categories = {'all': '0', |
|
'movies': '1002', |
|
'tv': '1001', |
|
'music': '1004', |
|
'games': '23', |
|
'anime': '20', |
|
'software': '32'} |
|
|
|
# Set proxies (default false) |
|
# make sure that proxies keys is'nt empty |
|
proxy = True |
|
proxies = { |
|
'http': '', |
|
'https': '', |
|
} |
|
|
|
# credentials |
|
username = "USERNAME" |
|
password = "PASSWORD" |
|
ua = 'Mozilla/5.0 (X11; Linux i686; rv:38.0) Gecko/20100101 Firefox/38.0' |
|
|
|
def __init__(self): |
|
# establish connection |
|
# |
|
# make cookie |
|
cj = CookieJar() |
|
self.session = build_opener(HTTPCookieProcessor(cj)) |
|
|
|
# add proxy handler if needed |
|
if self.proxy and any(self.proxies.keys()): |
|
self.session.add_handler(ProxyHandler(self.proxies)) |
|
|
|
# change user-agent |
|
self.session.addheaders.pop() |
|
self.session.addheaders.append(('User-Agent', self.ua)) |
|
|
|
form_data = {"username": self.username, "password": self.password} |
|
data_encoded = urlencode(form_data).encode('cp1251') |
|
|
|
try: |
|
response = self.session.open(self.url + '/takelogin.php', data_encoded) |
|
# Only continue if response status is OK. |
|
if response.getcode() != 200: |
|
raise HTTPError(response.geturl(), response.getcode(), |
|
"HTTP request to {} failed with status: {}".format(self.url, response.getcode()), |
|
response.info(), None) |
|
except (URLError, HTTPError) as e: |
|
logging.error(e) |
|
raise e |
|
|
|
if 'uid' not in [cookie.name for cookie in cj]: |
|
logging.debug(cj) |
|
|
|
class WorstParser(HTMLParser): |
|
def __init__(self, url=''): |
|
HTMLParser.__init__(self) |
|
self.url = url |
|
self.torrent = {'link': '', |
|
'name': '', |
|
'size': '', |
|
'seeds': '', |
|
'leech': '', |
|
'desc_link': '', } |
|
|
|
# we need a page markup to know when stop and collect data, |
|
# because available methods, in this class, do not communicate each other |
|
# as a result, we make markup to transfer information |
|
# from one method to another, along a chain |
|
# |
|
# markup on result table |
|
self.result_table = False # table with results is found |
|
self.torrent_row = False # found torrent row for collect data |
|
self.index_td = 0 # td counter in torrent row |
|
self.write = None # trigger to detecting when to collect data |
|
|
|
# markup pagination |
|
self.paginator = False # found more pages in result |
|
self.pages = 0 # page counter |
|
|
|
self.found_torrents = 0 |
|
|
|
def handle_starttag(self, tag, attrs): |
|
# search result table by class t_peer |
|
if tag == 'table': |
|
for name, value in attrs: |
|
if name == 'class' and 't_peer' in value: |
|
self.result_table = True |
|
|
|
# search for torrent row by class bg |
|
if self.result_table and tag == 'tr': |
|
for name, value in attrs: |
|
if name == 'class' and 'bg' in value: |
|
self.torrent_row = True |
|
|
|
# count td for find right td |
|
if self.torrent_row and tag == 'td': |
|
if self.index_td == 3: |
|
self.write = "size" |
|
elif self.index_td == 4: |
|
self.write = "seeds" |
|
elif self.index_td == 5: |
|
self.write = "leech" |
|
|
|
self.index_td += 1 |
|
|
|
# search for torrent link by classes r0 or r1 |
|
if self.torrent_row and tag == 'a': |
|
for name, value in attrs: |
|
if name == 'class' and 'r' in value: |
|
self.torrent['link'] = 'http://dl.kinozal.tv/download.php?id=' + attrs[0][1].split('=')[1] |
|
self.torrent['desc_link'] = self.url + attrs[0][1] |
|
self.write = "name" |
|
|
|
# search for right div with class paginator |
|
if self.found_torrents == 50 and tag == 'div': |
|
for name, value in attrs: |
|
if name == 'class' and value == 'paginator': |
|
self.paginator = True |
|
|
|
# search for block with page numbers |
|
if self.paginator and tag == 'li': |
|
self.pages += 1 |
|
|
|
def handle_endtag(self, tag): |
|
# detecting that torrent row is closed and print all collected data |
|
if self.torrent_row and tag == 'tr': |
|
self.torrent["engine_url"] = self.url |
|
logging.debug('tr: ' + str(self.torrent)) |
|
prettyPrinter(self.torrent) |
|
self.torrent = {key: '' for key in self.torrent} |
|
self.index_td = 0 |
|
self.torrent_row = False |
|
self.found_torrents += 1 |
|
|
|
# detecting that table with result is close |
|
if self.result_table and tag == 'table': |
|
self.result_table = False |
|
|
|
# detecting that we found all pagination |
|
if self.paginator and tag == 'ul': |
|
self.paginator = False |
|
|
|
def handle_data(self, data: str): |
|
# detecting that we need write data at this moment |
|
if self.write and self.result_table: |
|
if self.write == 'size': |
|
data = self.units_convert(data) |
|
self.torrent[self.write] = data.strip() |
|
self.write = None |
|
|
|
@staticmethod |
|
def units_convert(unit): |
|
# replace size units |
|
table = {'ТБ': 'TB', 'ГБ': 'GB', 'МБ': 'MB', 'КБ': 'KB'} |
|
x = unit.split(" ") |
|
x[1] = table[x[1]] |
|
|
|
return " ".join(x) |
|
|
|
def error(self, message): |
|
pass |
|
|
|
def download_torrent(self, url): |
|
# Create a torrent file |
|
file, path = tempfile.mkstemp('.torrent') |
|
file = os.fdopen(file, "wb") |
|
|
|
# Download url |
|
try: |
|
response = self.session.open(url) |
|
# Only continue if response status is OK. |
|
if response.getcode() != 200: |
|
raise HTTPError(response.geturl(), response.getcode(), |
|
"HTTP request to {} failed with status: {}".format(url, response.getcode()), |
|
response.info(), None) |
|
except (URLError, HTTPError) as e: |
|
logging.error(e) |
|
raise e |
|
|
|
# Write it to a file |
|
file.write(response.read()) |
|
file.close() |
|
|
|
# return file path |
|
logging.debug(path + " " + url) |
|
print(path + " " + url) |
|
|
|
def search(self, what, cat='all'): |
|
query = '%s/browse.php?s=%s&c=%s' % (self.url, unquote(quote(what)), self.supported_categories[cat]) |
|
response = self.session.open(query) |
|
parser = self.WorstParser(self.url) |
|
parser.feed(response.read().decode('cp1251')) |
|
parser.close() |
|
|
|
# if first request return that we have pages, we do cycle |
|
if parser.pages: |
|
for x in range(1, parser.pages): |
|
response = self.session.open('%s&page=%s' % (query, x)) |
|
parser.feed(response.read().decode('cp1251')) |
|
parser.close() |
|
|
|
|
|
# logging.debug("--- %s seconds ---" % (time.time() - start_time)) |
|
if __name__ == "__main__": |
|
kinozal_se = kinozal() |
|
# print(os.path.abspath(os.path.join(os.path.dirname(__file__), '../..', 'logs'))) |
|
# print(kinozal_se.WorstParser.units_convert("500 КБ")) |
|
# kinozal_se.search('terror lostfilm', 'tv') |
|
# kinozal_se._handle_connection(True) |
|
# kinozal_se.download_torrent('http://dl.kinozal.tv/download.php?id=1609776') |
|
# print("--- %s seconds ---" % (time.time() - start_time))
|
|
|