From 662353e936ae11c50c6e3dc4f281176f4305a186 Mon Sep 17 00:00:00 2001 From: marseneault Date: Sun, 20 Jul 2014 11:14:39 +0800 Subject: [PATCH] Add support for Bitsoup --- gui/slick/images/providers/bitsoup.png | Bin 0 -> 828 bytes sickbeard/__init__.py | 2 +- sickbeard/providers/__init__.py | 3 +- sickbeard/providers/bitsoup.py | 342 +++++++++++++++++++++++++ 4 files changed, 345 insertions(+), 2 deletions(-) create mode 100644 gui/slick/images/providers/bitsoup.png create mode 100644 sickbeard/providers/bitsoup.py diff --git a/gui/slick/images/providers/bitsoup.png b/gui/slick/images/providers/bitsoup.png new file mode 100644 index 0000000000000000000000000000000000000000..8d7c5eb1c8f5a6e6c2a25180c8847614de8d4bdf GIT binary patch literal 828 zcmV-C1H=4@P)1LF}dc-Z7TabAI1t%elqRl5HEvpksF9|KE! zEzq8b$Bn-5M6X9LEXH@qg*X^KbsBVDfpj{J(mrkFu&|tArmOW?P}bNpyO&4Z6&ws! zY`fdu4grH6x<1eOP%UkPlM%NlkW40#bE84qhr?kov&G(5I-CQCJsx+KY=C`{yDXH^ z#N_kp?0Lw~-^a&Ia3aLBB%NLZb)CD;c?GiBEJ{eFQdlSy3ZS6@(y+lyC6$$H8mXu% zD2rZ;L-fcARG?>4frBO$FioycMj~J$kpP`npjxd$0eio0<~Z|TzG1I2A9?zX7@S<1 z!sirwF4&a|Ma({>lht8xHpYr$(`Uiy7$>67=L4Nr;KzMH^wJn&&kx(I^xji?;`u7) zLm#a-hb1qQ>u8QG`WtM_#. + +import re +import traceback +import datetime +import urlparse +import time +import sickbeard +import generic +from sickbeard.common import Quality, cpu_presets +from sickbeard import logger +from sickbeard import tvcache +from sickbeard import db +from sickbeard import classes +from sickbeard import helpers +from sickbeard import show_name_helpers +from sickbeard.common import Overview +from sickbeard.exceptions import ex +from sickbeard import clients +from lib import requests +from lib.requests import exceptions +from bs4 import BeautifulSoup +from lib.unidecode import unidecode +from sickbeard.helpers import sanitizeSceneName + + +class BitSoupProvider(generic.TorrentProvider): + urls = {'base_url': 'https://www.bitsoup.me', + 'login': 'https://www.bitsoup.me/takelogin.php', + 'detail': 'https://www.bitsoup.me/details.php?id=%s', + 'search': 'https://www.bitsoup.me/browse.php?search=%s%s', + 'download': 'https://bitsoup.me/%s', + } + + def __init__(self): + + generic.TorrentProvider.__init__(self, "BitSoup") + + self.supportsBacklog = True + + self.enabled = False + self.username = None + self.password = None + self.ratio = None + self.minseed = None + self.minleech = None + + self.cache = BitSoupCache(self) + + self.url = self.urls['base_url'] + + self.categories = "&c42=1&c45=1&c49=1&c7=1" + + def isEnabled(self): + return self.enabled + + def imageName(self): + return 'bitsoup.png' + + def getQuality(self, item, anime=False): + + quality = Quality.sceneQuality(item[0], anime) + return quality + + def _doLogin(self): + + login_params = {'username': self.username, + 'password': self.password, + 'ssl': 'yes' + } + + self.session = requests.Session() + + try: + response = self.session.post(self.urls['login'], data=login_params, timeout=30) + except (requests.exceptions.ConnectionError, requests.exceptions.HTTPError), e: + logger.log(u'Unable to connect to ' + self.name + ' provider: ' + ex(e), logger.ERROR) + return False + + if re.search('Username or password incorrect', response.text): + logger.log(u'Invalid username or password for ' + self.name + ' Check your settings', logger.ERROR) + return False + + return True + + def _get_season_search_strings(self, ep_obj): + + search_string = {'Season': []} + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + if ep_obj.show.air_by_date or ep_obj.show.sports: + ep_string = show_name + ' ' + str(ep_obj.airdate).split('-')[0] + elif ep_obj.show.anime: + ep_string = show_name + ' ' + "%d" % ep_obj.scene_absolute_number + else: + ep_string = show_name + ' S%02d' % int(ep_obj.scene_season) #1) showName SXX + + search_string['Season'].append(ep_string) + + return [search_string] + + def _get_episode_search_strings(self, ep_obj, add_string=''): + + search_string = {'Episode': []} + + if not ep_obj: + return [] + + if self.show.air_by_date: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + str(ep_obj.airdate).replace('-', '|') + search_string['Episode'].append(ep_string) + elif self.show.sports: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + str(ep_obj.airdate).replace('-', '|') + '|' + \ + ep_obj.airdate.strftime('%b') + search_string['Episode'].append(ep_string) + elif self.show.anime: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = sanitizeSceneName(show_name) + ' ' + \ + "%i" % int(ep_obj.scene_absolute_number) + search_string['Episode'].append(ep_string) + else: + for show_name in set(show_name_helpers.allPossibleShowNames(self.show)): + ep_string = show_name_helpers.sanitizeSceneName(show_name) + ' ' + \ + sickbeard.config.naming_ep_type[2] % {'seasonnumber': ep_obj.scene_season, + 'episodenumber': ep_obj.scene_episode} + + search_string['Episode'].append(re.sub('\s+', ' ', ep_string)) + + return [search_string] + + def _doSearch(self, search_params, epcount=0, age=0): + + results = [] + items = {'Season': [], 'Episode': [], 'RSS': []} + + if not self._doLogin(): + return [] + + for mode in search_params.keys(): + for search_string in search_params[mode]: + + if isinstance(search_string, unicode): + search_string = unidecode(search_string) + + searchURL = self.urls['search'] % (search_string, self.categories) + + logger.log(u"Search string: " + searchURL, logger.DEBUG) + + data = self.getURL(searchURL) + if not data: + continue + + try: + html = BeautifulSoup(data, "html.parser") + + torrent_table = html.find('table', attrs={'class': 'koptekst'}) + torrent_rows = torrent_table.find_all('tr') if torrent_table else [] + + #Continue only if one Release is found + if len(torrent_rows) < 2: + logger.log(u"The Data returned from " + self.name + " do not contains any torrent", + logger.DEBUG) + continue + + for result in torrent_rows[1:]: + cells = result.find_all('td') + + link = cells[1].find('a') + download_url = self.urls['download'] % cells[3].find('a')['href'] + + id = link['href'] + id = id.replace('details.php?id=','') + id = id.replace('&hit=1', '') + + try: + title = link.getText() + id = int(id) + seeders = int(cells[9].getText()) + leechers = int(cells[10].getText()) + except (AttributeError, TypeError): + continue + + #Filter unseeded torrent + if mode != 'RSS' and (seeders < self.minseed or leechers < self.minleech): + continue + + if not title or not download_url: + continue + + item = title, download_url, id, seeders, leechers + logger.log(u"Found result: " + title + "(" + searchURL + ")", logger.DEBUG) + + items[mode].append(item) + + except Exception, e: + logger.log(u"Failed parsing " + self.name + " Traceback: " + traceback.format_exc(), logger.ERROR) + + #For each search mode sort all the items by seeders + items[mode].sort(key=lambda tup: tup[3], reverse=True) + + results += items[mode] + + return results + + def _get_title_and_url(self, item): + + title, url, id, seeders, leechers = item + + if url: + url = str(url).replace('&', '&') + + return (title, url) + + def getURL(self, url, post_data=None, headers=None, json=False): + + if not self.session: + self._doLogin() + + if not headers: + headers = [] + + try: + # Remove double-slashes from url + parsed = list(urlparse.urlparse(url)) + parsed[2] = re.sub("/{2,}", "/", parsed[2]) # replace two or more / with one + url = urlparse.urlunparse(parsed) + + response = self.session.get(url, verify=False) + except (requests.exceptions.ConnectionError, requests.exceptions.HTTPError), e: + logger.log(u"Error loading " + self.name + " URL: " + ex(e), logger.ERROR) + return None + + if response.status_code != 200: + logger.log(self.name + u" page requested with url " + url + " returned status code is " + str( + response.status_code) + ': ' + clients.http_error_code[response.status_code], logger.WARNING) + return None + + return response.content + + def findPropers(self, search_date=datetime.datetime.today()): + + results = [] + + myDB = db.DBConnection() + sqlResults = myDB.select( + 'SELECT s.show_name, e.showid, e.season, e.episode, e.status, e.airdate FROM tv_episodes AS e' + + ' INNER JOIN tv_shows AS s ON (e.showid = s.indexer_id)' + + ' WHERE e.airdate >= ' + str(search_date.toordinal()) + + ' AND (e.status IN (' + ','.join([str(x) for x in Quality.DOWNLOADED]) + ')' + + ' OR (e.status IN (' + ','.join([str(x) for x in Quality.SNATCHED]) + ')))' + ) + + if not sqlResults: + return [] + + for sqlshow in sqlResults: + self.show = helpers.findCertainShow(sickbeard.showList, int(sqlshow["showid"])) + if self.show: + curEp = self.show.getEpisode(int(sqlshow["season"]), int(sqlshow["episode"])) + + searchString = self._get_episode_search_strings(curEp, add_string='PROPER|REPACK') + + for item in self._doSearch(searchString[0]): + title, url = self._get_title_and_url(item) + results.append(classes.Proper(title, url, datetime.datetime.today())) + + return results + + def seedRatio(self): + return self.ratio + + +class BitSoupCache(tvcache.TVCache): + def __init__(self, provider): + + tvcache.TVCache.__init__(self, provider) + + # only poll TorrentBytes every 20 minutes max + self.minTime = 20 + + def updateCache(self): + + # delete anything older then 7 days + logger.log(u"Clearing " + self.provider.name + " cache") + self._clearCache() + + if not self.shouldUpdate(): + return + + search_params = {'RSS': ['']} + rss_results = self.provider._doSearch(search_params) + + if rss_results: + self.setLastUpdate() + else: + return [] + + cl = [] + for result in rss_results: + + item = (result[0], result[1]) + ci = self._parseItem(item) + if ci is not None: + cl.append(ci) + + if cl: + myDB = self._getDB() + myDB.mass_action(cl) + + + def _parseItem(self, item): + + (title, url) = item + + if not title or not url: + return None + + logger.log(u"Attempting to cache item:[" + title + "]", logger.DEBUG) + + return self._addCacheEntry(title, url) + + +provider = BitSoupProvider()