ngosang
9 years ago
6 changed files with 196 additions and 0 deletions
After Width: | Height: | Size: 561 B |
@ -0,0 +1,97 @@ |
|||||||
|
#VERSION: 2.0 |
||||||
|
#AUTHORS: Douman (custparasite@gmx.se) |
||||||
|
#CONTRIBUTORS: Diego de las Heras (ngosang@hotmail.es) |
||||||
|
|
||||||
|
from novaprinter import prettyPrinter |
||||||
|
from helpers import retrieve_url, download_file |
||||||
|
from re import compile as re_compile |
||||||
|
from HTMLParser import HTMLParser |
||||||
|
|
||||||
|
class torlock(object): |
||||||
|
url = "https://www.torlock.com" |
||||||
|
name = "TorLock" |
||||||
|
supported_categories = {'all' : 'all', |
||||||
|
'anime' : 'anime', |
||||||
|
'software' : 'software', |
||||||
|
'games' : 'game', |
||||||
|
'movies' : 'movie', |
||||||
|
'music' : 'music', |
||||||
|
'tv' : 'television', |
||||||
|
'books' : 'ebooks'} |
||||||
|
|
||||||
|
def download_torrent(self, info): |
||||||
|
print(download_file(info)) |
||||||
|
|
||||||
|
class MyHtmlParser(HTMLParser): |
||||||
|
""" Sub-class for parsing results """ |
||||||
|
def __init__(self, url): |
||||||
|
HTMLParser.__init__(self) |
||||||
|
self.url = url |
||||||
|
self.article_found = False #true when <article> with results is found |
||||||
|
self.item_found = False |
||||||
|
self.item_bad = False #set to True for malicious links |
||||||
|
self.current_item = None #dict for found item |
||||||
|
self.item_name = None #key's name in current_item dict |
||||||
|
self.parser_class = {"ts" : "size", |
||||||
|
"tul" : "seeds", |
||||||
|
"tdl" : "leech"} |
||||||
|
|
||||||
|
def handle_starttag(self, tag, attrs): |
||||||
|
params = dict(attrs) |
||||||
|
if self.item_found: |
||||||
|
if tag == "td": |
||||||
|
if "class" in params: |
||||||
|
self.item_name = self.parser_class.get(params["class"], None) |
||||||
|
if self.item_name: |
||||||
|
self.current_item[self.item_name] = "" |
||||||
|
|
||||||
|
elif self.article_found and tag == "a": |
||||||
|
if "href" in params: |
||||||
|
link = params["href"] |
||||||
|
if link.startswith("/torrent"): |
||||||
|
self.current_item["desc_link"] = "".join((self.url, link)) |
||||||
|
self.current_item["link"] = "".join((self.url, "/tor/", link.split('/')[2], ".torrent")) |
||||||
|
self.current_item["engine_url"] = self.url |
||||||
|
self.item_found = True |
||||||
|
self.item_name = "name" |
||||||
|
self.current_item["name"] = "" |
||||||
|
self.item_bad = "rel" in params and params["rel"] == "nofollow" |
||||||
|
|
||||||
|
elif tag == "article": |
||||||
|
self.article_found = True |
||||||
|
self.current_item = {} |
||||||
|
|
||||||
|
def handle_data(self, data): |
||||||
|
if self.item_name: |
||||||
|
self.current_item[self.item_name] += data |
||||||
|
|
||||||
|
def handle_endtag(self, tag): |
||||||
|
if tag == "article": |
||||||
|
self.article_found = False |
||||||
|
elif self.item_name and (tag == "a" or tag == "td"): |
||||||
|
self.item_name = None |
||||||
|
elif self.item_found and tag == "tr": |
||||||
|
self.item_found = False |
||||||
|
if not self.item_bad: |
||||||
|
prettyPrinter(self.current_item) |
||||||
|
self.current_item = {} |
||||||
|
|
||||||
|
def search(self, query, cat='all'): |
||||||
|
""" Performs search """ |
||||||
|
query = query.replace("%20", "-") |
||||||
|
|
||||||
|
parser = self.MyHtmlParser(self.url) |
||||||
|
page = "".join((self.url, "/", self.supported_categories[cat], "/torrents/", query, ".html?sort=seeds&page=1")) |
||||||
|
html = retrieve_url(page) |
||||||
|
parser.feed(html) |
||||||
|
|
||||||
|
counter = 1 |
||||||
|
additional_pages = re_compile("/{0}/torrents/{1}.html\?sort=seeds&page=[0-9]+".format(self.supported_categories[cat], query)) |
||||||
|
list_searches = additional_pages.findall(html)[:-1] #last link is next(i.e. second) |
||||||
|
for page in map(lambda link: "".join((self.url, link)), list_searches): |
||||||
|
html = retrieve_url(page) |
||||||
|
parser.feed(html) |
||||||
|
counter += 1 |
||||||
|
if counter > 3: |
||||||
|
break |
||||||
|
parser.close() |
After Width: | Height: | Size: 561 B |
@ -0,0 +1,97 @@ |
|||||||
|
#VERSION: 2.0 |
||||||
|
#AUTHORS: Douman (custparasite@gmx.se) |
||||||
|
#CONTRIBUTORS: Diego de las Heras (ngosang@hotmail.es) |
||||||
|
|
||||||
|
from novaprinter import prettyPrinter |
||||||
|
from helpers import retrieve_url, download_file |
||||||
|
from re import compile as re_compile |
||||||
|
from html.parser import HTMLParser |
||||||
|
|
||||||
|
class torlock(object): |
||||||
|
url = "https://www.torlock.com" |
||||||
|
name = "TorLock" |
||||||
|
supported_categories = {'all' : 'all', |
||||||
|
'anime' : 'anime', |
||||||
|
'software' : 'software', |
||||||
|
'games' : 'game', |
||||||
|
'movies' : 'movie', |
||||||
|
'music' : 'music', |
||||||
|
'tv' : 'television', |
||||||
|
'books' : 'ebooks'} |
||||||
|
|
||||||
|
def download_torrent(self, info): |
||||||
|
print(download_file(info)) |
||||||
|
|
||||||
|
class MyHtmlParser(HTMLParser): |
||||||
|
""" Sub-class for parsing results """ |
||||||
|
def __init__(self, url): |
||||||
|
HTMLParser.__init__(self) |
||||||
|
self.url = url |
||||||
|
self.article_found = False #true when <article> with results is found |
||||||
|
self.item_found = False |
||||||
|
self.item_bad = False #set to True for malicious links |
||||||
|
self.current_item = None #dict for found item |
||||||
|
self.item_name = None #key's name in current_item dict |
||||||
|
self.parser_class = {"ts" : "size", |
||||||
|
"tul" : "seeds", |
||||||
|
"tdl" : "leech"} |
||||||
|
|
||||||
|
def handle_starttag(self, tag, attrs): |
||||||
|
params = dict(attrs) |
||||||
|
if self.item_found: |
||||||
|
if tag == "td": |
||||||
|
if "class" in params: |
||||||
|
self.item_name = self.parser_class.get(params["class"], None) |
||||||
|
if self.item_name: |
||||||
|
self.current_item[self.item_name] = "" |
||||||
|
|
||||||
|
elif self.article_found and tag == "a": |
||||||
|
if "href" in params: |
||||||
|
link = params["href"] |
||||||
|
if link.startswith("/torrent"): |
||||||
|
self.current_item["desc_link"] = "".join((self.url, link)) |
||||||
|
self.current_item["link"] = "".join((self.url, "/tor/", link.split('/')[2], ".torrent")) |
||||||
|
self.current_item["engine_url"] = self.url |
||||||
|
self.item_found = True |
||||||
|
self.item_name = "name" |
||||||
|
self.current_item["name"] = "" |
||||||
|
self.item_bad = "rel" in params and params["rel"] == "nofollow" |
||||||
|
|
||||||
|
elif tag == "article": |
||||||
|
self.article_found = True |
||||||
|
self.current_item = {} |
||||||
|
|
||||||
|
def handle_data(self, data): |
||||||
|
if self.item_name: |
||||||
|
self.current_item[self.item_name] += data |
||||||
|
|
||||||
|
def handle_endtag(self, tag): |
||||||
|
if tag == "article": |
||||||
|
self.article_found = False |
||||||
|
elif self.item_name and (tag == "a" or tag == "td"): |
||||||
|
self.item_name = None |
||||||
|
elif self.item_found and tag == "tr": |
||||||
|
self.item_found = False |
||||||
|
if not self.item_bad: |
||||||
|
prettyPrinter(self.current_item) |
||||||
|
self.current_item = {} |
||||||
|
|
||||||
|
def search(self, query, cat='all'): |
||||||
|
""" Performs search """ |
||||||
|
query = query.replace("%20", "-") |
||||||
|
|
||||||
|
parser = self.MyHtmlParser(self.url) |
||||||
|
page = "".join((self.url, "/", self.supported_categories[cat], "/torrents/", query, ".html?sort=seeds&page=1")) |
||||||
|
html = retrieve_url(page) |
||||||
|
parser.feed(html) |
||||||
|
|
||||||
|
counter = 1 |
||||||
|
additional_pages = re_compile("/{0}/torrents/{1}.html\?sort=seeds&page=[0-9]+".format(self.supported_categories[cat], query)) |
||||||
|
list_searches = additional_pages.findall(html)[:-1] #last link is next(i.e. second) |
||||||
|
for page in map(lambda link: "".join((self.url, link)), list_searches): |
||||||
|
html = retrieve_url(page) |
||||||
|
parser.feed(html) |
||||||
|
counter += 1 |
||||||
|
if counter > 3: |
||||||
|
break |
||||||
|
parser.close() |
Loading…
Reference in new issue