qBittorrent search engine plugins
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

244 lines
8.9 KiB

# VERSION: 1.0
# AUTHORS: imDMG
# LICENSING INFORMATION
import tempfile
import os
import logging
import time
from urllib.request import build_opener, HTTPCookieProcessor, ProxyHandler
from urllib.parse import urlencode, quote, unquote
from urllib.error import URLError, HTTPError
from http.cookiejar import CookieJar
from html.parser import HTMLParser
from novaprinter import prettyPrinter
# setup logging into qBittorrent/logs
logging.basicConfig(level=logging.DEBUG,
format='%(asctime)s %(name)-12s %(levelname)-8s %(message)s',
datefmt='%m-%d %H:%M',
filename=os.path.abspath(
os.path.join(os.path.dirname(__file__), '../..', 'logs')) + "/kinozal_se.log",
# filename="kinozal_se.log",
filemode='w')
# benchmark
# start_time = time.time()
class kinozal(object):
name = 'Kinozal'
url = 'http://kinozal.tv'
supported_categories = {'all': '0',
'movies': '1002',
'tv': '1001',
'music': '1004',
'games': '23',
'anime': '20',
'software': '32'}
# Set proxies (default false)
# make sure that proxies keys is'nt empty
proxy = True
proxies = {
'http': '',
'https': '',
}
# credentials
username = "USERNAME"
password = "PASSWORD"
ua = 'Mozilla/5.0 (X11; Linux i686; rv:38.0) Gecko/20100101 Firefox/38.0'
def __init__(self):
# establish connection
#
# make cookie
cj = CookieJar()
self.session = build_opener(HTTPCookieProcessor(cj))
# add proxy handler if needed
if self.proxy and any(self.proxies.keys()):
self.session.add_handler(ProxyHandler(self.proxies))
# change user-agent
self.session.addheaders.pop()
self.session.addheaders.append(('User-Agent', self.ua))
form_data = {"username": self.username, "password": self.password}
data_encoded = urlencode(form_data).encode('cp1251')
try:
response = self.session.open(self.url + '/takelogin.php', data_encoded)
# Only continue if response status is OK.
if response.getcode() != 200:
raise HTTPError(response.geturl(), response.getcode(),
"HTTP request to {} failed with status: {}".format(self.url, response.getcode()),
response.info(), None)
except (URLError, HTTPError) as e:
logging.error(e)
raise e
if 'uid' not in [cookie.name for cookie in cj]:
logging.debug(cj)
class WorstParser(HTMLParser):
def __init__(self, url=''):
HTMLParser.__init__(self)
self.url = url
self.torrent = {'link': '',
'name': '',
'size': '',
'seeds': '',
'leech': '',
'desc_link': '', }
# we need a page markup to know when stop and collect data,
# because available methods, in this class, do not communicate each other
# as a result, we make markup to transfer information
# from one method to another, along a chain
#
# markup on result table
self.result_table = False # table with results is found
self.torrent_row = False # found torrent row for collect data
self.index_td = 0 # td counter in torrent row
self.write = None # trigger to detecting when to collect data
# markup pagination
self.paginator = False # found more pages in result
self.pages = 0 # page counter
self.found_torrents = 0
def handle_starttag(self, tag, attrs):
# search result table by class t_peer
if tag == 'table':
for name, value in attrs:
if name == 'class' and 't_peer' in value:
self.result_table = True
# search for torrent row by class bg
if self.result_table and tag == 'tr':
for name, value in attrs:
if name == 'class' and 'bg' in value:
self.torrent_row = True
# count td for find right td
if self.torrent_row and tag == 'td':
if self.index_td == 3:
self.write = "size"
elif self.index_td == 4:
self.write = "seeds"
elif self.index_td == 5:
self.write = "leech"
self.index_td += 1
# search for torrent link by classes r0 or r1
if self.torrent_row and tag == 'a':
for name, value in attrs:
if name == 'class' and 'r' in value:
self.torrent['link'] = 'http://dl.kinozal.tv/download.php?id=' + attrs[0][1].split('=')[1]
self.torrent['desc_link'] = self.url + attrs[0][1]
self.write = "name"
# search for right div with class paginator
if self.found_torrents == 50 and tag == 'div':
for name, value in attrs:
if name == 'class' and value == 'paginator':
self.paginator = True
# search for block with page numbers
if self.paginator and tag == 'li':
self.pages += 1
def handle_endtag(self, tag):
# detecting that torrent row is closed and print all collected data
if self.torrent_row and tag == 'tr':
self.torrent["engine_url"] = self.url
logging.debug('tr: ' + str(self.torrent))
prettyPrinter(self.torrent)
self.torrent = {key: '' for key in self.torrent}
self.index_td = 0
self.torrent_row = False
self.found_torrents += 1
# detecting that table with result is close
if self.result_table and tag == 'table':
self.result_table = False
# detecting that we found all pagination
if self.paginator and tag == 'ul':
self.paginator = False
def handle_data(self, data: str):
# detecting that we need write data at this moment
if self.write and self.result_table:
if self.write == 'size':
data = self.units_convert(data)
self.torrent[self.write] = data.strip()
self.write = None
@staticmethod
def units_convert(unit):
# replace size units
table = {'ТБ': 'TB', 'ГБ': 'GB', 'МБ': 'MB', 'КБ': 'KB'}
x = unit.split(" ")
x[1] = table[x[1]]
return " ".join(x)
def error(self, message):
pass
def download_torrent(self, url):
# Create a torrent file
file, path = tempfile.mkstemp('.torrent')
file = os.fdopen(file, "wb")
# Download url
try:
response = self.session.open(url)
# Only continue if response status is OK.
if response.getcode() != 200:
raise HTTPError(response.geturl(), response.getcode(),
"HTTP request to {} failed with status: {}".format(url, response.getcode()),
response.info(), None)
except (URLError, HTTPError) as e:
logging.error(e)
raise e
# Write it to a file
file.write(response.read())
file.close()
# return file path
logging.debug(path + " " + url)
print(path + " " + url)
def search(self, what, cat='all'):
query = '%s/browse.php?s=%s&c=%s' % (self.url, unquote(quote(what)), self.supported_categories[cat])
response = self.session.open(query)
parser = self.WorstParser(self.url)
parser.feed(response.read().decode('cp1251'))
parser.close()
# if first request return that we have pages, we do cycle
if parser.pages:
for x in range(1, parser.pages):
response = self.session.open('%s&page=%s' % (query, x))
parser.feed(response.read().decode('cp1251'))
parser.close()
# logging.debug("--- %s seconds ---" % (time.time() - start_time))
if __name__ == "__main__":
kinozal_se = kinozal()
# print(os.path.abspath(os.path.join(os.path.dirname(__file__), '../..', 'logs')))
# print(kinozal_se.WorstParser.units_convert("500 КБ"))
# kinozal_se.search('terror lostfilm', 'tv')
# kinozal_se._handle_connection(True)
# kinozal_se.download_torrent('http://dl.kinozal.tv/download.php?id=1609776')
# print("--- %s seconds ---" % (time.time() - start_time))