diff --git a/gui/slick/images/providers/torrentbytes.png b/gui/slick/images/providers/torrentbytes.png new file mode 100644 index 00000000..0c092ee7 Binary files /dev/null and b/gui/slick/images/providers/torrentbytes.png differ diff --git a/sickbeard/__init__.py b/sickbeard/__init__.py index c35bcc80..5749e666 100644 --- a/sickbeard/__init__.py +++ b/sickbeard/__init__.py @@ -32,7 +32,7 @@ import sys from sickbeard import providers, metadata, config, webserveInit from sickbeard.providers.generic import GenericProvider from providers import ezrss, tvtorrents, btn, newznab, womble, thepiratebay, torrentleech, kat, iptorrents, \ - omgwtfnzbs, scc, hdtorrents, torrentday, hdbits, nextgen, speedcd, nyaatorrents, fanzub + omgwtfnzbs, scc, hdtorrents, torrentday, hdbits, nextgen, speedcd, nyaatorrents, fanzub, torrentbytes from sickbeard.config import CheckSection, check_setting_int, check_setting_str, check_setting_float, ConfigMigrator, \ naming_ep_type from sickbeard import searchBacklog, showUpdater, versionChecker, properFinder, autoPostProcesser, \ @@ -1751,4 +1751,4 @@ def getEpList(epIDs, showid=None): curEpObj = curShowObj.getEpisode(int(curEp["season"]), int(curEp["episode"])) epList.append(curEpObj) - return epList \ No newline at end of file + return epList diff --git a/sickbeard/providers/__init__.py b/sickbeard/providers/__init__.py index cba5ffbf..7bcdec2c 100644 --- a/sickbeard/providers/__init__.py +++ b/sickbeard/providers/__init__.py @@ -32,7 +32,8 @@ __all__ = ['ezrss', 'nextgen', 'speedcd', 'nyaatorrents', - 'fanzub' + 'fanzub', + 'torrentbytes' ] import sickbeard diff --git a/sickbeard/providers/torrentbytes.py b/sickbeard/providers/torrentbytes.py new file mode 100644 index 00000000..24f3dfd9 --- /dev/null +++ b/sickbeard/providers/torrentbytes.py @@ -0,0 +1,347 @@ +# Author: Idan Gutman +# URL: http://code.google.com/p/sickbeard/ +# +# This file is part of SickRage. +# +# SickRage is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# SickRage is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with SickRage. If not, see . + +import re +import traceback +import datetime +import urlparse +import time +import sickbeard +import generic +from sickbeard.common import Quality, cpu_presets +from sickbeard import logger +from sickbeard import tvcache +from sickbeard import db +from sickbeard import classes +from sickbeard import helpers +from sickbeard import show_name_helpers +from sickbeard.common import Overview +from sickbeard.exceptions import ex +from sickbeard import clients +from lib import requests +from lib.requests import exceptions +from bs4 import BeautifulSoup +from lib.unidecode import unidecode +from sickbeard.helpers import sanitizeSceneName + + +class TorrentBytesProvider(generic.TorrentProvider): + urls = {'base_url': 'https://www.torrentbytes.net', + 'login': 'https://www.torrentbytes.net/takelogin.php', + 'detail': 'https://www.torrentbytes.net/details.php?id=%s', + 'search': 'https://www.torrentbytes.net/browse.php?search=%s%s', + 'download': 'https://www.torrentbytes.net/download.php?id=%s&name=%s', + } + + def __init__(self): + + generic.TorrentProvider.__init__(self, "TorrentBytes") + + self.supportsBacklog = True + + self.enabled = False + self.username = None + self.password = None + self.ratio = None + self.minseed = None + self.minleech = None + + self.cache = TorrentBytesCache(self) + + self.url = self.urls['base_url'] + + self.categories = "&c41=1&c33=1&c38=1&c32=1&c37=1" + + def isEnabled(self): + return self.enabled + + def imageName(self): + return 'torrentbytes.png' + + def getQuality(self, item, anime=False): + + quality = Quality.sceneQuality(item[0], anime) + return quality + + def _doLogin(self): + + login_params = {'username': self.username, + 'password': self.password, + 'login': 'submit' + } + + self.session = requests.Session() + + try: + response = self.session.post(self.urls['login'], data=login_params, timeout=30) + except (requests.exceptions.ConnectionError, requests.exceptions.HTTPError), e: + logger.log(u'Unable to connect to ' + self.name + ' provider: ' + ex(e), logger.ERROR) + return False + + if re.search('Username or password incorrect', response.text): + + logger.log(u'Invalid username or password for ' + self.name + ' Check your settings', logger.ERROR) + return False + + return True + + def _get_season_search_strings(self, ep_obj): + + search_string = {'Season': []} + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + if ep_obj.show.air_by_date or ep_obj.show.sports: + ep_string = show_name + '.' + str(ep_obj.airdate).split('-')[0] + elif ep_obj.show.anime: + ep_string = show_name + '.' + "%d" % ep_obj.scene_absolute_number + else: + ep_string = show_name + '.S%02d' % int(ep_obj.scene_season) #1) showName SXX + + search_string['Season'].append(ep_string) + + return [search_string] + + def _get_episode_search_strings(self, ep_obj, add_string=''): + + search_string = {'Episode': []} + + if not ep_obj: + return [] + + if self.show.air_by_date: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + str(ep_obj.airdate).replace('-', '|') + search_string['Episode'].append(ep_string) + elif self.show.sports: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + str(ep_obj.airdate).replace('-', '|') + '|' + \ + ep_obj.airdate.strftime('%b') + search_string['Episode'].append(ep_string) + elif self.show.anime: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + "%i" % int(ep_obj.scene_absolute_number) + search_string['Episode'].append(ep_string) + else: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = show_name_helpers.sanitizeSceneName(show_name) + ' ' + \ + sickbeard.config.naming_ep_type[2] % {'seasonnumber': ep_obj.scene_season, + 'episodenumber': ep_obj.scene_episode} + + search_string['Episode'].append(re.sub('\s+', ' ', ep_string)) + + return [search_string] + + def _doSearch(self, search_params, epcount=0, age=0): + + results = [] + items = {'Season': [], 'Episode': [], 'RSS': []} + + if not self._doLogin(): + return [] + + for mode in search_params.keys(): + for search_string in search_params[mode]: + + if isinstance(search_string, unicode): + search_string = unidecode(search_string) + + searchURL = self.urls['search'] % (search_string, self.categories) + + logger.log(u"Search string: " + searchURL, logger.DEBUG) + + data = self.getURL(searchURL) + if not data: + continue + + try: + html = BeautifulSoup(data) + + torrent_table = html.find('table', attrs={'border': '1'}) + torrent_rows = torrent_table.find_all('tr') if torrent_table else [] + + #Continue only if one Release is found + if len(torrent_rows) < 2: + logger.log(u"The Data returned from " + self.name + " do not contains any torrent", + logger.DEBUG) + continue + + for result in torrent_rows[1:]: + cells = result.find_all('td') + + link = cells[1].find('a', attrs = {'class': 'index'}) + + full_id = link['href'].replace('details.php?id=', '') + torrent_id = full_id[:6] + + try: + if link.has_key('title'): + title = cells[1].find('a', {'class': 'index'})['title'] + else: + title = link.contents[0] + download_url = self.urls['download'] % (torrent_id, link.contents[0]) + id = int(torrent_id) + seeders = int(cells[8].find('span').contents[0]) + leechers = int(cells[9].find('span').contents[0]) + except (AttributeError, TypeError): + continue + + #Filter unseeded torrent + if mode != 'RSS' and (seeders == 0 or seeders < self.minseed or leechers < self.minleech): + continue + + if not title or not download_url: + continue + + item = title, download_url, id, seeders, leechers + logger.log(u"Found result: " + title + "(" + searchURL + ")", logger.DEBUG) + + items[mode].append(item) + + except Exception, e: + logger.log(u"Failed parsing " + self.name + " Traceback: " + traceback.format_exc(), logger.ERROR) + + #For each search mode sort all the items by seeders + items[mode].sort(key=lambda tup: tup[3], reverse=True) + + results += items[mode] + + return results + + def _get_title_and_url(self, item): + + title, url, id, seeders, leechers = item + + if url: + url = str(url).replace('&', '&') + + return (title, url) + + def getURL(self, url, post_data=None, headers=None, json=False): + + if not self.session: + self._doLogin() + + if not headers: + headers = [] + + try: + # Remove double-slashes from url + parsed = list(urlparse.urlparse(url)) + parsed[2] = re.sub("/{2,}", "/", parsed[2]) # replace two or more / with one + url = urlparse.urlunparse(parsed) + + response = self.session.get(url, verify=False) + except (requests.exceptions.ConnectionError, requests.exceptions.HTTPError), e: + logger.log(u"Error loading " + self.name + " URL: " + ex(e), logger.ERROR) + return None + + if response.status_code != 200: + logger.log(self.name + u" page requested with url " + url + " returned status code is " + str( + response.status_code) + ': ' + clients.http_error_code[response.status_code], logger.WARNING) + return None + + return response.content + + def findPropers(self, search_date=datetime.datetime.today()): + + results = [] + + myDB = db.DBConnection() + sqlResults = myDB.select( + 'SELECT s.show_name, e.showid, e.season, e.episode, e.status, e.airdate FROM tv_episodes AS e' + + ' INNER JOIN tv_shows AS s ON (e.showid = s.indexer_id)' + + ' WHERE e.airdate >= ' + str(search_date.toordinal()) + + ' AND (e.status IN (' + ','.join([str(x) for x in Quality.DOWNLOADED]) + ')' + + ' OR (e.status IN (' + ','.join([str(x) for x in Quality.SNATCHED]) + ')))' + ) + + if not sqlResults: + return [] + + for sqlshow in sqlResults: + self.show = helpers.findCertainShow(sickbeard.showList, int(sqlshow["showid"])) + if self.show: + curEp = self.show.getEpisode(int(sqlshow["season"]), int(sqlshow["episode"])) + + searchString = self._get_episode_search_strings(curEp, add_string='PROPER|REPACK') + + for item in self._doSearch(searchString[0]): + title, url = self._get_title_and_url(item) + results.append(classes.Proper(title, url, datetime.datetime.today())) + + return results + + def seedRatio(self): + return self.ratio + + +class TorrentBytesCache(tvcache.TVCache): + def __init__(self, provider): + + tvcache.TVCache.__init__(self, provider) + + # only poll TorrentBytes every 20 minutes max + self.minTime = 20 + + def updateCache(self): + + # delete anything older then 7 days + logger.log(u"Clearing " + self.provider.name + " cache") + self._clearCache() + + if not self.shouldUpdate(): + return + + search_params = {'RSS': ['']} + rss_results = self.provider._doSearch(search_params) + + if rss_results: + self.setLastUpdate() + else: + return [] + + cl = [] + for result in rss_results: + + item = (result[0], result[1]) + ci = self._parseItem(item) + if ci is not None: + cl.append(ci) + + + + if cl: + myDB = self._getDB() + myDB.mass_action(cl) + + + def _parseItem(self, item): + + (title, url) = item + + if not title or not url: + return None + + logger.log(u"Attempting to cache item:[" + title +"]", logger.DEBUG) + + return self._addCacheEntry(title, url) + + +provider = TorrentBytesProvider()