Compare commits
No commits in common. "41e51a0b12e4130ba75b97d409b3fe31db8450bc" and "8169ffbf8e9b2fc26db277762d9a482fa441edcc" have entirely different histories.
41e51a0b12
...
8169ffbf8e
@ -1,5 +1,5 @@
|
||||
from pynyaata.config import app
|
||||
from pynyaata.connectors.core import curl_content
|
||||
from pynyaata.connectors import curl_content
|
||||
from pynyaata.models import AnimeLink
|
||||
|
||||
app.config['SQLALCHEMY_ECHO'] = False
|
||||
|
@ -1,19 +1,26 @@
|
||||
from functools import wraps
|
||||
from operator import attrgetter, itemgetter
|
||||
|
||||
from flask import redirect, render_template, request, url_for, abort
|
||||
|
||||
from .config import app, auth, ADMIN_USERNAME, ADMIN_PASSWORD, MYSQL_ENABLED
|
||||
from .config import app, auth, ADMIN_USERNAME, ADMIN_PASSWORD
|
||||
from .connectors import *
|
||||
from .connectors.core import ConnectorCore, ConnectorLang, ConnectorReturn
|
||||
from .forms import SearchForm, DeleteForm, EditForm
|
||||
from .utils import boldify, clean_model
|
||||
|
||||
if MYSQL_ENABLED:
|
||||
from .config import db
|
||||
from .models import AnimeFolder, AnimeTitle, AnimeLink
|
||||
|
||||
|
||||
def clean_model(obj):
|
||||
for attr in dir(obj):
|
||||
if not attr.startswith('_') and getattr(obj, attr) is None:
|
||||
try:
|
||||
setattr(obj, attr, '')
|
||||
except AttributeError:
|
||||
pass
|
||||
return obj
|
||||
|
||||
|
||||
def mysql_required(f):
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
@ -32,11 +39,11 @@ def verify_password(username, password):
|
||||
@app.template_filter('boldify')
|
||||
def boldify(name):
|
||||
query = request.args.get('q', '')
|
||||
name = utils.boldify(name, query)
|
||||
name = Connector.boldify(name, query)
|
||||
if MYSQL_ENABLED:
|
||||
for keyword in db.session.query(AnimeTitle.keyword.distinct()).all():
|
||||
if keyword[0].lower() != query.lower():
|
||||
name = utils.boldify(name, keyword[0])
|
||||
name = Connector.boldify(name, keyword[0])
|
||||
return name
|
||||
|
||||
|
||||
@ -47,7 +54,7 @@ def flagify(is_vf):
|
||||
|
||||
@app.template_filter('colorify')
|
||||
def colorify(model):
|
||||
return get_instance(model.link, model.title.keyword).color
|
||||
return Connector.get_instance(model.link, model.title.keyword).color
|
||||
|
||||
|
||||
@app.route('/')
|
||||
@ -62,20 +69,34 @@ def search():
|
||||
if not query:
|
||||
return redirect(url_for('home'))
|
||||
|
||||
return render_template('search.html', search_form=SearchForm(), connectors=run_all(query),
|
||||
results = [
|
||||
Nyaa(query).run(),
|
||||
Pantsu(query).run(),
|
||||
YggTorrent(query).run(),
|
||||
YggAnimation(query).run(),
|
||||
AnimeUltime(query).run(),
|
||||
]
|
||||
|
||||
return render_template('search.html', search_form=SearchForm(), connectors=results,
|
||||
mysql_disabled=not MYSQL_ENABLED)
|
||||
|
||||
|
||||
@app.route('/latest')
|
||||
@app.route('/latest/<int:page>')
|
||||
def latest(page=1):
|
||||
torrents = run_all('', return_type=ConnectorReturn.HISTORY, page=page)
|
||||
torrents = [
|
||||
Nyaa('', return_type=ConnectorReturn.HISTORY, page=page).run(),
|
||||
Pantsu('', return_type=ConnectorReturn.HISTORY, page=page).run(),
|
||||
YggTorrent('', return_type=ConnectorReturn.HISTORY, page=page).run(),
|
||||
YggAnimation('', return_type=ConnectorReturn.HISTORY, page=page).run(),
|
||||
AnimeUltime('', return_type=ConnectorReturn.HISTORY, page=page).run(),
|
||||
]
|
||||
|
||||
results = []
|
||||
for torrent in torrents:
|
||||
results = results + torrent.data
|
||||
for result in results:
|
||||
result['self'] = get_instance(result['href'], '')
|
||||
result['self'] = Connector.get_instance(result['href'], '')
|
||||
results.sort(key=itemgetter('date'), reverse=True)
|
||||
|
||||
return render_template('latest.html', search_form=SearchForm(), torrents=results, page=page,
|
||||
|
531
pynyaata/connectors.py
Normal file
531
pynyaata/connectors.py
Normal file
@ -0,0 +1,531 @@
|
||||
import re
|
||||
from abc import ABC, abstractmethod
|
||||
from datetime import datetime, timedelta
|
||||
from enum import Enum
|
||||
from functools import wraps
|
||||
from logging import getLogger
|
||||
from urllib.parse import quote
|
||||
|
||||
from bs4 import BeautifulSoup
|
||||
from cloudscraper import create_scraper
|
||||
from cloudscraper.exceptions import CloudflareException
|
||||
from dateparser import parse
|
||||
from requests import RequestException
|
||||
|
||||
from .config import IS_DEBUG, MYSQL_ENABLED, CACHE_TIMEOUT, BLACKLIST_WORDS
|
||||
|
||||
scraper = create_scraper()
|
||||
|
||||
|
||||
class ConnectorReturn(Enum):
|
||||
SEARCH = 1
|
||||
HISTORY = 2
|
||||
|
||||
|
||||
class ConnectorLang(Enum):
|
||||
FR = '🇫🇷'
|
||||
JP = '🇯🇵'
|
||||
|
||||
|
||||
class Cache:
|
||||
CACHE_DATA = {}
|
||||
|
||||
def cache_data(self, f):
|
||||
@wraps(f)
|
||||
def wrapper(*args, **kwds):
|
||||
connector = args[0]
|
||||
timestamp = datetime.now().timestamp()
|
||||
|
||||
# clear old data
|
||||
for connector_class in list(self.CACHE_DATA):
|
||||
for connector_func in list(self.CACHE_DATA[connector_class]):
|
||||
for connector_query in list(self.CACHE_DATA[connector_class][connector_func]):
|
||||
for connector_page in list(self.CACHE_DATA[connector_class][connector_func][connector_query]):
|
||||
if self.CACHE_DATA[connector_class][connector_func][connector_query][connector_page][
|
||||
'timeout'
|
||||
] < timestamp:
|
||||
del self.CACHE_DATA[connector_class][connector_func][connector_query][connector_page]
|
||||
|
||||
if connector.__class__.__name__ not in self.CACHE_DATA:
|
||||
self.CACHE_DATA[connector.__class__.__name__] = {}
|
||||
if f.__name__ not in self.CACHE_DATA[connector.__class__.__name__]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__] = {}
|
||||
if connector.query not in self.CACHE_DATA[connector.__class__.__name__][f.__name__]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query] = {}
|
||||
if connector.page not in self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page] = {
|
||||
'timeout': 0.0
|
||||
}
|
||||
|
||||
cached_data = self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page]
|
||||
if cached_data['timeout'] > timestamp:
|
||||
connector.data = cached_data['data']
|
||||
connector.is_more = cached_data['is_more']
|
||||
connector.on_error = False
|
||||
return
|
||||
|
||||
ret = f(*args, **kwds)
|
||||
if not connector.on_error:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page] = {
|
||||
'data': connector.data,
|
||||
'timeout': timestamp + CACHE_TIMEOUT,
|
||||
'is_more': connector.is_more
|
||||
}
|
||||
return ret
|
||||
|
||||
return wrapper
|
||||
|
||||
|
||||
ConnectorCache = Cache()
|
||||
|
||||
|
||||
def curl_content(url, params=None, ajax=False):
|
||||
if ajax:
|
||||
headers = {'X-Requested-With': 'XMLHttpRequest'}
|
||||
else:
|
||||
headers = {}
|
||||
|
||||
try:
|
||||
if params is not None:
|
||||
response = scraper.post(url, params, timeout=5, headers=headers)
|
||||
else:
|
||||
response = scraper.get(url, timeout=5, headers=headers)
|
||||
|
||||
output = response.text
|
||||
http_code = response.status_code
|
||||
except (RequestException, CloudflareException) as e:
|
||||
output = ''
|
||||
http_code = 500
|
||||
if IS_DEBUG:
|
||||
getLogger().exception(e)
|
||||
|
||||
return {'http_code': http_code, 'output': output}
|
||||
|
||||
|
||||
def link_exist_in_db(href):
|
||||
if MYSQL_ENABLED:
|
||||
from .models import AnimeLink
|
||||
return AnimeLink.query.filter_by(link=href).first()
|
||||
return False
|
||||
|
||||
|
||||
def parse_date(str_to_parse, date_format=''):
|
||||
if str_to_parse is None:
|
||||
return datetime.fromtimestamp(0)
|
||||
elif isinstance(str_to_parse, datetime):
|
||||
return str_to_parse
|
||||
else:
|
||||
date = parse(str_to_parse, date_formats=[date_format])
|
||||
if date:
|
||||
return date
|
||||
else:
|
||||
return datetime.fromtimestamp(0)
|
||||
|
||||
|
||||
class Connector(ABC):
|
||||
@property
|
||||
@abstractmethod
|
||||
def color(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def title(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def favicon(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def base_url(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def is_light(self):
|
||||
pass
|
||||
|
||||
def __init__(self, query, page=1, return_type=ConnectorReturn.SEARCH):
|
||||
self.query = query
|
||||
self.data = []
|
||||
self.is_more = False
|
||||
self.on_error = True
|
||||
self.page = page
|
||||
self.return_type = return_type
|
||||
|
||||
@abstractmethod
|
||||
def get_full_search_url(self):
|
||||
pass
|
||||
|
||||
@abstractmethod
|
||||
def search(self):
|
||||
pass
|
||||
|
||||
@abstractmethod
|
||||
def get_history(self):
|
||||
pass
|
||||
|
||||
def run(self):
|
||||
if self.on_error:
|
||||
if self.return_type is ConnectorReturn.SEARCH:
|
||||
self.search()
|
||||
elif self.return_type is ConnectorReturn.HISTORY:
|
||||
self.get_history()
|
||||
return self
|
||||
|
||||
@staticmethod
|
||||
def get_instance(url, query):
|
||||
if 'nyaa.si' in url:
|
||||
return Nyaa(query)
|
||||
elif 'nyaa.net' in url:
|
||||
return Pantsu(query)
|
||||
elif 'anime-ultime' in url:
|
||||
return AnimeUltime(query)
|
||||
elif 'ygg' in url:
|
||||
return YggTorrent(query)
|
||||
else:
|
||||
return Other(query)
|
||||
|
||||
@staticmethod
|
||||
def get_lang(str_to_test):
|
||||
if re.search('(vf|multi|french)', str_to_test, re.IGNORECASE):
|
||||
return ConnectorLang.FR
|
||||
else:
|
||||
return ConnectorLang.JP
|
||||
|
||||
@staticmethod
|
||||
def boldify(str_to_replace, keyword):
|
||||
if keyword:
|
||||
return re.sub('(%s)' % keyword, r'<b>\1</b>', str_to_replace, flags=re.IGNORECASE)
|
||||
else:
|
||||
return str_to_replace
|
||||
|
||||
|
||||
class Nyaa(Connector):
|
||||
color = 'is-link'
|
||||
title = 'Nyaa'
|
||||
favicon = 'nyaa.png'
|
||||
base_url = 'https://nyaa.si'
|
||||
is_light = False
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 'size'
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 'id'
|
||||
|
||||
to_query = '(%s vf)|(%s vostfr)|(%s multi)|(%s french)' % (self.query, self.query, self.query, self.query)
|
||||
return '%s/?f=0&c=1_3&s=%s&o=desc&q=%s&p=%s' % (self.base_url, sort_type, to_query, self.page)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('table.torrent-list tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[7].string)
|
||||
check_seeds = int(tds[5].string)
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
urls = tds[1].findAll('a')
|
||||
|
||||
if len(urls) > 1:
|
||||
url = urls[1]
|
||||
has_comment = True
|
||||
else:
|
||||
url = urls[0]
|
||||
has_comment = False
|
||||
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
href = '%s%s' % (self.base_url, url['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': href,
|
||||
'name': url_safe,
|
||||
'comment': str(urls[0]).replace('/view/',
|
||||
'%s%s' % (self.base_url, '/view/')) if has_comment else '',
|
||||
'link': tds[2].decode_contents().replace('/download/',
|
||||
'%s%s' % (self.base_url, '/download/')),
|
||||
'size': tds[3].string,
|
||||
'date': parse_date(tds[4].string, '%Y-%m-%d %H:%M'),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[6].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(href) else 'is-%s' % tr['class'][0]
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
||||
|
||||
|
||||
class Pantsu(Connector):
|
||||
color = 'is-info'
|
||||
title = 'Pantsu'
|
||||
favicon = 'pantsu.png'
|
||||
base_url = 'https://nyaa.net'
|
||||
is_light = False
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 4
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 2
|
||||
|
||||
to_query = '(%s vf)|(%s vostfr)|(%s multi)|(%s french)' % (self.query, self.query, self.query, self.query)
|
||||
return '%s/search/%s?c=3_13&order=false&q=%s&sort=%s' % (self.base_url, self.page, to_query, sort_type)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('div.results tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[6].string.replace('-', '0'))
|
||||
check_seeds = int(tds[4].string.replace('-', '0'))
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
url = tds[1].a
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
href = '%s%s' % (self.base_url, url['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': href,
|
||||
'name': url_safe,
|
||||
'comment': '',
|
||||
'link': tds[2].decode_contents().replace('icon-magnet', 'fa fa-fw fa-magnet').replace(
|
||||
'icon-floppy', 'fa fa-fw fa-download'),
|
||||
'size': tds[3].string,
|
||||
'date': parse_date(tds[7]['title'][:-6], '%m/%d/%Y, %I:%M:%S %p'),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[5].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(href) else 'is-%s' % tr['class'][0]
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
||||
|
||||
|
||||
class YggTorrent(Connector):
|
||||
color = 'is-success'
|
||||
title = 'YggTorrent'
|
||||
favicon = 'yggtorrent.png'
|
||||
base_url = 'https://www2.yggtorrent.se'
|
||||
is_light = False
|
||||
category = 2179
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 'size'
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 'publish_date'
|
||||
sort_page = '&page=%s' % (self.page * 50) if self.page > 1 else ''
|
||||
|
||||
return '%s/engine/search?name=%s&category=2145&sub_category=%s&do=search&order=desc&sort=%s%s' % (
|
||||
self.base_url, self.query, self.category, sort_type, sort_page
|
||||
)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
if self.category:
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('table.table tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[6].string)
|
||||
check_seeds = int(tds[7].string)
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
url = tds[1].a
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': url['href'],
|
||||
'name': url_safe,
|
||||
'comment': '<a href="%s#comm" target="_blank"><i class="fa fa-comments-o"></i>%s</a>' %
|
||||
(url['href'], tds[3].decode_contents()),
|
||||
'link': '<a href="%s/engine/download_torrent?id=%s">'
|
||||
'<i class="fa fa-fw fa-download"></i>'
|
||||
'</a>' % (self.base_url, re.search(r'/(\d+)', url['href']).group(1)),
|
||||
'size': tds[5].string,
|
||||
'date': parse_date(datetime.fromtimestamp(int(tds[4].div.string))),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[8].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(quote(url['href'], '/+:')) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
||||
|
||||
|
||||
class YggAnimation(YggTorrent):
|
||||
title = 'YggAnimation'
|
||||
category = 2178
|
||||
|
||||
|
||||
class AnimeUltime(Connector):
|
||||
color = 'is-warning'
|
||||
title = 'Anime-Ultime'
|
||||
favicon = 'animeultime.png'
|
||||
base_url = 'http://www.anime-ultime.net'
|
||||
is_light = True
|
||||
|
||||
def get_full_search_url(self):
|
||||
from_date = ''
|
||||
sort_type = 'search'
|
||||
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
try:
|
||||
page_date = datetime.now() - timedelta((int(self.page) - 1) * 365 / 12)
|
||||
except OverflowError:
|
||||
page_date = datetime.fromtimestamp(0)
|
||||
from_date = page_date.strftime('%m%Y')
|
||||
sort_type = 'history'
|
||||
|
||||
return '%s/%s-0-1/%s' % (self.base_url, sort_type, from_date)
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url(), {'search': self.query})
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
title = html.select('div.title')
|
||||
|
||||
if 'Recherche' in title[0].string:
|
||||
trs = html.select('table.jtable tr')
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
|
||||
if len(tds) < 2:
|
||||
continue
|
||||
|
||||
url = tds[0].a
|
||||
href = '%s/%s' % (self.base_url, url['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/%s' % (self.base_url, url['href']),
|
||||
'name': url.get_text(),
|
||||
'type': tds[1].string,
|
||||
'date': parse_date(None),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
else:
|
||||
player = html.select('div.AUVideoPlayer')
|
||||
name = html.select('h1')
|
||||
ani_type = html.select('div.titre')
|
||||
href = '%s/file-0-1/%s' % (self.base_url, player[0]['data-serie'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/file-0-1/%s' % (self.base_url, player[0]['data-serie']),
|
||||
'name': name[0].string,
|
||||
'type': ani_type[0].string.replace(':', ''),
|
||||
'date': parse_date(None),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def get_history(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
tables = html.select('table.jtable')
|
||||
h3s = html.findAll('h3')
|
||||
|
||||
for i, table in enumerate(tables):
|
||||
for j, tr in enumerate(table.findAll('tr')):
|
||||
if not j:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
link = tds[0].a
|
||||
href = '%s/%s' % (self.base_url, link['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/%s' % (self.base_url, link['href']),
|
||||
'name': link.string,
|
||||
'type': tds[4].string,
|
||||
'date': parse_date(h3s[i].string[:-3], '%A %d %B %Y'),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
|
||||
|
||||
class Other(Connector):
|
||||
color = 'is-danger'
|
||||
title = 'Other'
|
||||
favicon = 'blank.png'
|
||||
base_url = ''
|
||||
is_light = True
|
||||
|
||||
def get_full_search_url(self):
|
||||
pass
|
||||
|
||||
def search(self):
|
||||
pass
|
||||
|
||||
def get_history(self):
|
||||
pass
|
@ -1,28 +0,0 @@
|
||||
from .animeultime import AnimeUltime
|
||||
from .core import Other
|
||||
from .nyaa import Nyaa
|
||||
from .pantsu import Pantsu
|
||||
from .yggtorrent import YggTorrent, YggAnimation
|
||||
|
||||
|
||||
def run_all(*args, **kwargs):
|
||||
return [
|
||||
Nyaa(*args, **kwargs).run(),
|
||||
Pantsu(*args, **kwargs).run(),
|
||||
YggTorrent(*args, **kwargs).run(),
|
||||
YggAnimation(*args, **kwargs).run(),
|
||||
AnimeUltime(*args, **kwargs).run(),
|
||||
]
|
||||
|
||||
|
||||
def get_instance(url, query):
|
||||
if 'nyaa.si' in url:
|
||||
return Nyaa(query)
|
||||
elif 'nyaa.net' in url:
|
||||
return Pantsu(query)
|
||||
elif 'anime-ultime' in url:
|
||||
return AnimeUltime(query)
|
||||
elif 'ygg' in url:
|
||||
return YggTorrent(query)
|
||||
else:
|
||||
return Other(query)
|
@ -1,105 +0,0 @@
|
||||
from datetime import datetime, timedelta
|
||||
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
from .core import ConnectorCore, ConnectorReturn, ConnectorCache, ConnectorLang, curl_content
|
||||
from ..utils import parse_date, link_exist_in_db
|
||||
|
||||
|
||||
class AnimeUltime(ConnectorCore):
|
||||
color = 'is-warning'
|
||||
title = 'Anime-Ultime'
|
||||
favicon = 'animeultime.png'
|
||||
base_url = 'http://www.anime-ultime.net'
|
||||
is_light = True
|
||||
|
||||
def get_full_search_url(self):
|
||||
from_date = ''
|
||||
sort_type = 'search'
|
||||
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
try:
|
||||
page_date = datetime.now() - timedelta((int(self.page) - 1) * 365 / 12)
|
||||
except OverflowError:
|
||||
page_date = datetime.fromtimestamp(0)
|
||||
from_date = page_date.strftime('%m%Y')
|
||||
sort_type = 'history'
|
||||
|
||||
return '%s/%s-0-1/%s' % (self.base_url, sort_type, from_date)
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url(), {'search': self.query})
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
title = html.select('div.title')
|
||||
|
||||
if 'Recherche' in title[0].string:
|
||||
trs = html.select('table.jtable tr')
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
|
||||
if len(tds) < 2:
|
||||
continue
|
||||
|
||||
url = tds[0].a
|
||||
href = '%s/%s' % (self.base_url, url['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/%s' % (self.base_url, url['href']),
|
||||
'name': url.get_text(),
|
||||
'type': tds[1].string,
|
||||
'date': parse_date(None),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
else:
|
||||
player = html.select('div.AUVideoPlayer')
|
||||
name = html.select('h1')
|
||||
ani_type = html.select('div.titre')
|
||||
href = '%s/file-0-1/%s' % (self.base_url, player[0]['data-serie'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/file-0-1/%s' % (self.base_url, player[0]['data-serie']),
|
||||
'name': name[0].string,
|
||||
'type': ani_type[0].string.replace(':', ''),
|
||||
'date': parse_date(None),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def get_history(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
tables = html.select('table.jtable')
|
||||
h3s = html.findAll('h3')
|
||||
|
||||
for i, table in enumerate(tables):
|
||||
for j, tr in enumerate(table.findAll('tr')):
|
||||
if not j:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
link = tds[0].a
|
||||
href = '%s/%s' % (self.base_url, link['href'])
|
||||
|
||||
self.data.append({
|
||||
'lang': ConnectorLang.JP,
|
||||
'href': '%s/%s' % (self.base_url, link['href']),
|
||||
'name': link.string,
|
||||
'type': tds[4].string,
|
||||
'date': parse_date(h3s[i].string[:-3], '%A %d %B %Y'),
|
||||
'class': self.color if link_exist_in_db(href) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
@ -1,177 +0,0 @@
|
||||
import re
|
||||
from abc import ABC, abstractmethod
|
||||
from datetime import datetime
|
||||
from enum import Enum
|
||||
from functools import wraps
|
||||
from logging import getLogger
|
||||
|
||||
from cloudscraper import create_scraper
|
||||
from cloudscraper.exceptions import CloudflareException
|
||||
from requests import RequestException
|
||||
|
||||
from ..config import CACHE_TIMEOUT, IS_DEBUG
|
||||
|
||||
scraper = create_scraper()
|
||||
|
||||
|
||||
class ConnectorReturn(Enum):
|
||||
SEARCH = 1
|
||||
HISTORY = 2
|
||||
|
||||
|
||||
class ConnectorLang(Enum):
|
||||
FR = '🇫🇷'
|
||||
JP = '🇯🇵'
|
||||
|
||||
|
||||
class Cache:
|
||||
CACHE_DATA = {}
|
||||
|
||||
def cache_data(self, f):
|
||||
@wraps(f)
|
||||
def wrapper(*args, **kwds):
|
||||
connector = args[0]
|
||||
timestamp = datetime.now().timestamp()
|
||||
|
||||
for connector_class in list(self.CACHE_DATA):
|
||||
for connector_func in list(self.CACHE_DATA[connector_class]):
|
||||
for connector_query in list(self.CACHE_DATA[connector_class][connector_func]):
|
||||
for connector_page in list(self.CACHE_DATA[connector_class][connector_func][connector_query]):
|
||||
if self.CACHE_DATA[connector_class][connector_func][connector_query][connector_page][
|
||||
'timeout'
|
||||
] < timestamp:
|
||||
del self.CACHE_DATA[connector_class][connector_func][connector_query][connector_page]
|
||||
|
||||
if connector.__class__.__name__ not in self.CACHE_DATA:
|
||||
self.CACHE_DATA[connector.__class__.__name__] = {}
|
||||
if f.__name__ not in self.CACHE_DATA[connector.__class__.__name__]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__] = {}
|
||||
if connector.query not in self.CACHE_DATA[connector.__class__.__name__][f.__name__]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query] = {}
|
||||
if connector.page not in self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query]:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page] = {
|
||||
'timeout': 0.0
|
||||
}
|
||||
|
||||
cached_data = self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page]
|
||||
if cached_data['timeout'] > timestamp:
|
||||
connector.data = cached_data['data']
|
||||
connector.is_more = cached_data['is_more']
|
||||
connector.on_error = False
|
||||
return
|
||||
|
||||
ret = f(*args, **kwds)
|
||||
if not connector.on_error:
|
||||
self.CACHE_DATA[connector.__class__.__name__][f.__name__][connector.query][connector.page] = {
|
||||
'data': connector.data,
|
||||
'timeout': timestamp + CACHE_TIMEOUT,
|
||||
'is_more': connector.is_more
|
||||
}
|
||||
return ret
|
||||
|
||||
return wrapper
|
||||
|
||||
|
||||
ConnectorCache = Cache()
|
||||
|
||||
|
||||
def curl_content(url, params=None, ajax=False):
|
||||
if ajax:
|
||||
headers = {'X-Requested-With': 'XMLHttpRequest'}
|
||||
else:
|
||||
headers = {}
|
||||
|
||||
try:
|
||||
if params is not None:
|
||||
response = scraper.post(url, params, timeout=5, headers=headers)
|
||||
else:
|
||||
response = scraper.get(url, timeout=5, headers=headers)
|
||||
|
||||
output = response.text
|
||||
http_code = response.status_code
|
||||
except (RequestException, CloudflareException) as e:
|
||||
output = ''
|
||||
http_code = 500
|
||||
if IS_DEBUG:
|
||||
getLogger().exception(e)
|
||||
|
||||
return {'http_code': http_code, 'output': output}
|
||||
|
||||
|
||||
class ConnectorCore(ABC):
|
||||
@property
|
||||
@abstractmethod
|
||||
def color(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def title(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def favicon(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def base_url(self):
|
||||
pass
|
||||
|
||||
@property
|
||||
@abstractmethod
|
||||
def is_light(self):
|
||||
pass
|
||||
|
||||
def __init__(self, query, page=1, return_type=ConnectorReturn.SEARCH):
|
||||
self.query = query
|
||||
self.data = []
|
||||
self.is_more = False
|
||||
self.on_error = True
|
||||
self.page = page
|
||||
self.return_type = return_type
|
||||
|
||||
@abstractmethod
|
||||
def get_full_search_url(self):
|
||||
pass
|
||||
|
||||
@abstractmethod
|
||||
def search(self):
|
||||
pass
|
||||
|
||||
@abstractmethod
|
||||
def get_history(self):
|
||||
pass
|
||||
|
||||
@staticmethod
|
||||
def get_lang(str_to_test):
|
||||
if re.search('(vf|multi|french)', str_to_test, re.IGNORECASE):
|
||||
return ConnectorLang.FR
|
||||
else:
|
||||
return ConnectorLang.JP
|
||||
|
||||
def run(self):
|
||||
if self.on_error:
|
||||
if self.return_type is ConnectorReturn.SEARCH:
|
||||
self.search()
|
||||
elif self.return_type is ConnectorReturn.HISTORY:
|
||||
self.get_history()
|
||||
return self
|
||||
|
||||
|
||||
class Other(ConnectorCore):
|
||||
color = 'is-danger'
|
||||
title = 'Other'
|
||||
favicon = 'blank.png'
|
||||
base_url = ''
|
||||
is_light = True
|
||||
|
||||
def get_full_search_url(self):
|
||||
pass
|
||||
|
||||
def search(self):
|
||||
pass
|
||||
|
||||
def get_history(self):
|
||||
pass
|
@ -1,76 +0,0 @@
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
from .core import ConnectorCore, ConnectorReturn, ConnectorCache, curl_content
|
||||
from ..config import BLACKLIST_WORDS
|
||||
from ..utils import parse_date, link_exist_in_db
|
||||
|
||||
|
||||
class Nyaa(ConnectorCore):
|
||||
color = 'is-link'
|
||||
title = 'Nyaa'
|
||||
favicon = 'nyaa.png'
|
||||
base_url = 'https://nyaa.si'
|
||||
is_light = False
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 'size'
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 'id'
|
||||
|
||||
to_query = '(%s vf)|(%s vostfr)|(%s multi)|(%s french)' % (self.query, self.query, self.query, self.query)
|
||||
return '%s/?f=0&c=1_3&s=%s&o=desc&q=%s&p=%s' % (self.base_url, sort_type, to_query, self.page)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('table.torrent-list tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[7].string)
|
||||
check_seeds = int(tds[5].string)
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
urls = tds[1].findAll('a')
|
||||
|
||||
if len(urls) > 1:
|
||||
url = urls[1]
|
||||
has_comment = True
|
||||
else:
|
||||
url = urls[0]
|
||||
has_comment = False
|
||||
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
href = self.base_url + url['href']
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': href,
|
||||
'name': url_safe,
|
||||
'comment': str(urls[0]).replace('/view/', self.base_url + '/view/') if has_comment else '',
|
||||
'link': tds[2].decode_contents().replace('/download/', self.base_url + '/download/'),
|
||||
'size': tds[3].string,
|
||||
'date': parse_date(tds[4].string, '%Y-%m-%d %H:%M'),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[6].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(href) else 'is-%s' % tr['class'][0]
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
@ -1,69 +0,0 @@
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
from .core import ConnectorCore, ConnectorReturn, ConnectorCache, curl_content
|
||||
from ..config import BLACKLIST_WORDS
|
||||
from ..utils import parse_date, link_exist_in_db
|
||||
|
||||
|
||||
class Pantsu(ConnectorCore):
|
||||
color = 'is-info'
|
||||
title = 'Pantsu'
|
||||
favicon = 'pantsu.png'
|
||||
base_url = 'https://nyaa.net'
|
||||
is_light = False
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 4
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 2
|
||||
|
||||
to_query = '(%s vf)|(%s vostfr)|(%s multi)|(%s french)' % (self.query, self.query, self.query, self.query)
|
||||
return '%s/search/%s?c=3_13&order=false&q=%s&sort=%s' % (self.base_url, self.page, to_query, sort_type)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('div.results tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[6].string.replace('-', '0'))
|
||||
check_seeds = int(tds[4].string.replace('-', '0'))
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
url = tds[1].a
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
href = self.base_url + url['href']
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': href,
|
||||
'name': url_safe,
|
||||
'comment': '',
|
||||
'link': tds[2].decode_contents().replace('icon-magnet', 'fa fa-fw fa-magnet').replace(
|
||||
'icon-floppy', 'fa fa-fw fa-download'),
|
||||
'size': tds[3].string,
|
||||
'date': parse_date(tds[7]['title'][:-6], '%m/%d/%Y, %I:%M:%S %p'),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[5].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(href) else 'is-%s' % tr['class'][0]
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
@ -1,83 +0,0 @@
|
||||
import re
|
||||
from datetime import datetime
|
||||
from urllib.parse import quote
|
||||
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
from .core import ConnectorCore, ConnectorReturn, ConnectorCache, curl_content
|
||||
from ..config import BLACKLIST_WORDS
|
||||
from ..utils import parse_date, link_exist_in_db
|
||||
|
||||
|
||||
class YggTorrent(ConnectorCore):
|
||||
color = 'is-success'
|
||||
title = 'YggTorrent'
|
||||
favicon = 'yggtorrent.png'
|
||||
base_url = 'https://www2.yggtorrent.se'
|
||||
is_light = False
|
||||
category = 2179
|
||||
|
||||
def get_full_search_url(self):
|
||||
sort_type = 'size'
|
||||
if self.return_type is ConnectorReturn.HISTORY:
|
||||
sort_type = 'publish_date'
|
||||
sort_page = '&page=%s' % ((self.page - 1) * 50) if self.page > 1 else ''
|
||||
|
||||
return '%s/engine/search?name=%s&category=2145&sub_category=%s&do=search&order=desc&sort=%s%s' % (
|
||||
self.base_url, self.query, self.category, sort_type, sort_page
|
||||
)
|
||||
|
||||
def get_history(self):
|
||||
self.search()
|
||||
|
||||
@ConnectorCache.cache_data
|
||||
def search(self):
|
||||
if self.category:
|
||||
response = curl_content(self.get_full_search_url())
|
||||
|
||||
if response['http_code'] == 200:
|
||||
html = BeautifulSoup(response['output'], 'html.parser')
|
||||
trs = html.select('table.table tr')
|
||||
valid_trs = 0
|
||||
|
||||
for i, tr in enumerate(trs):
|
||||
if not i:
|
||||
continue
|
||||
|
||||
tds = tr.findAll('td')
|
||||
check_downloads = int(tds[6].string)
|
||||
check_seeds = int(tds[7].string)
|
||||
|
||||
if check_downloads or check_seeds:
|
||||
url = tds[1].a
|
||||
url_safe = url.get_text()
|
||||
|
||||
if any(word.lower() in url_safe.lower() for word in BLACKLIST_WORDS):
|
||||
continue
|
||||
|
||||
valid_trs = valid_trs + 1
|
||||
|
||||
self.data.append({
|
||||
'lang': self.get_lang(url_safe),
|
||||
'href': url['href'],
|
||||
'name': url_safe,
|
||||
'comment': '<a href="%s#comm" target="_blank"><i class="fa fa-comments-o"></i>%s</a>' %
|
||||
(url['href'], tds[3].decode_contents()),
|
||||
'link': '<a href="%s/engine/download_torrent?id=%s">'
|
||||
'<i class="fa fa-fw fa-download"></i>'
|
||||
'</a>' % (self.base_url, re.search(r'/(\d+)', url['href']).group(1)),
|
||||
'size': tds[5].string,
|
||||
'date': parse_date(datetime.fromtimestamp(int(tds[4].div.string))),
|
||||
'seeds': check_seeds,
|
||||
'leechs': tds[8].string,
|
||||
'downloads': check_downloads,
|
||||
'class': self.color if link_exist_in_db(quote(url['href'], '/+:')) else ''
|
||||
})
|
||||
|
||||
self.on_error = False
|
||||
self.is_more = valid_trs and valid_trs is not len(trs) - 1
|
||||
|
||||
|
||||
class YggAnimation(YggTorrent):
|
||||
title = 'YggAnimation'
|
||||
category = 2178
|
@ -1,43 +0,0 @@
|
||||
import re
|
||||
from datetime import datetime
|
||||
|
||||
from dateparser import parse
|
||||
|
||||
from . import MYSQL_ENABLED
|
||||
|
||||
|
||||
def link_exist_in_db(href):
|
||||
if MYSQL_ENABLED:
|
||||
from .models import AnimeLink
|
||||
return AnimeLink.query.filter_by(link=href).first()
|
||||
return False
|
||||
|
||||
|
||||
def parse_date(str_to_parse, date_format=''):
|
||||
if str_to_parse is None:
|
||||
return datetime.fromtimestamp(0)
|
||||
elif isinstance(str_to_parse, datetime):
|
||||
return str_to_parse
|
||||
else:
|
||||
date = parse(str_to_parse, date_formats=[date_format])
|
||||
if date:
|
||||
return date
|
||||
else:
|
||||
return datetime.fromtimestamp(0)
|
||||
|
||||
|
||||
def boldify(str_to_replace, keyword):
|
||||
if keyword:
|
||||
return re.sub('(%s)' % keyword, r'<b>\1</b>', str_to_replace, flags=re.IGNORECASE)
|
||||
else:
|
||||
return str_to_replace
|
||||
|
||||
|
||||
def clean_model(obj):
|
||||
for attr in dir(obj):
|
||||
if not attr.startswith('_') and getattr(obj, attr) is None:
|
||||
try:
|
||||
setattr(obj, attr, '')
|
||||
except AttributeError:
|
||||
pass
|
||||
return obj
|
Reference in New Issue
Block a user