mirror of
https://github.com/moparisthebest/SickRage
synced 2024-12-12 11:02:21 -05:00
Fix for debug logging on console.
Fix for RSS feeds for all providers.
This commit is contained in:
parent
5e38813a73
commit
99bbd0649c
@ -98,7 +98,7 @@ class SBRotatingLogHandler(object):
|
||||
# define a Handler which writes INFO messages or higher to the sys.stderr
|
||||
console = logging.StreamHandler()
|
||||
|
||||
console.setLevel(logging.DEBUG)
|
||||
console.setLevel(logging.INFO)
|
||||
|
||||
# set a format which is simpler for console use
|
||||
console.setFormatter(DispatchingFormatter(
|
||||
|
@ -211,4 +211,25 @@ sports_regexs = [
|
||||
(?P<extra_info>.*?(?<![. _-])(?<!WEB))[. _-]+
|
||||
(?P<release_group>.*?)$
|
||||
'''),
|
||||
|
||||
('sports_parts_without_event',
|
||||
# Show.Name.Event.Nov.23rd.2010.Source.Quality.Etc-Group
|
||||
'''
|
||||
^(?P<series_name>.*?(UEFA|MLB|ESPN|WWE|MMA|UFC|TNA|EPL|NASCAR|NBA|NFL|NHL|NRL|PGA|SUPER LEAGUE|FORMULA|FIFA|NETBALL|MOTOGP).*?)[. _-]+
|
||||
(?P<parts>\d{1,3}\d{1,3}.*?)[. _-]+
|
||||
(?P<air_day>\d{1,2}[a-zA-Z]{2})[. _-]+
|
||||
(?P<air_month>[a-zA-Z]{3,})[. _-]+
|
||||
(?P<air_year>\d{4})[. _-]+
|
||||
(?P<extra_info>.*?(?<![. _-])(?<!WEB))[. _-]+
|
||||
(?P<release_group>.*?)$
|
||||
'''),
|
||||
|
||||
('sports_bare',
|
||||
# Show.Name.Event.Nov.23rd.2010.Source.Quality.Etc-Group
|
||||
'''
|
||||
^(?P<series_name>.*?(UEFA|MLB|ESPN|WWE|MMA|UFC|TNA|EPL|NASCAR|NBA|NFL|NHL|NRL|PGA|SUPER LEAGUE|FORMULA|FIFA|NETBALL|MOTOGP).*?)[. _-]+
|
||||
(?P<parts>\d{1,3}\d{1,3}.*?)[. _-]+
|
||||
(?P<extra_info>.*?(?<![. _-])(?<!WEB))[. _-]+
|
||||
(?P<release_group>.*?)$
|
||||
'''),
|
||||
]
|
@ -357,9 +357,7 @@ class BTNCache(tvcache.TVCache):
|
||||
logger.WARNING)
|
||||
seconds_since_last_update = 86400
|
||||
|
||||
data = self.provider._doSearch(search_params=None, age=seconds_since_last_update)
|
||||
|
||||
return data
|
||||
return self.provider._doSearch(search_params=None, age=seconds_since_last_update)
|
||||
|
||||
def _parseItem(self, item):
|
||||
(title, url) = self.provider._get_title_and_url(item)
|
||||
|
@ -126,8 +126,7 @@ class DTTCache(tvcache.TVCache):
|
||||
|
||||
url = self.provider.url + 'rss/allshows?' + urllib.urlencode(params)
|
||||
logger.log(u"DTT cache update URL: " + url, logger.DEBUG)
|
||||
data = self.provider.getRSSFeed(url)
|
||||
return data
|
||||
return self.provider.getRSSFeed(url)
|
||||
|
||||
def _parseItem(self, item):
|
||||
title, url = self.provider._get_title_and_url(item)
|
||||
|
@ -118,7 +118,6 @@ class EZRSSProvider(generic.TorrentProvider):
|
||||
data = self.getRSSFeed(search_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + search_url, logger.ERROR)
|
||||
return []
|
||||
|
||||
items = data.entries
|
||||
@ -172,13 +171,7 @@ class EZRSSCache(tvcache.TVCache):
|
||||
rss_url = self.provider.url + 'feed/'
|
||||
logger.log(self.provider.name + " cache update URL: " + rss_url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(rss_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + rss_url, logger.ERROR)
|
||||
return None
|
||||
|
||||
return data
|
||||
return self.provider.getRSSFeed(rss_url)
|
||||
|
||||
def _parseItem(self, item):
|
||||
|
||||
|
@ -44,7 +44,6 @@ from lib.hachoir_parser import createParser
|
||||
from sickbeard.name_parser.parser import NameParser, InvalidNameException
|
||||
from sickbeard.common import Quality, Overview
|
||||
|
||||
|
||||
class GenericProvider:
|
||||
NZB = "nzb"
|
||||
TORRENT = "torrent"
|
||||
@ -137,6 +136,12 @@ class GenericProvider:
|
||||
if not f:
|
||||
logger.log(u"Error loading " + self.name + " URL: " + url, logger.ERROR)
|
||||
return None
|
||||
elif 'error' in f.feed:
|
||||
logger.log(u"Newznab ERROR:[%s] CODE:[%s]" % (f.feed['error']['description'], f.feed['error']['code']), logger.DEBUG)
|
||||
return None
|
||||
elif not f.entries:
|
||||
logger.log(u"Error loading RSS feed items for " + self.name + " using URL: " + url, logger.ERROR)
|
||||
return None
|
||||
|
||||
return f
|
||||
|
||||
|
@ -323,9 +323,7 @@ class NewzbinProvider(generic.NZBProvider):
|
||||
url = self.url + "search/?%s" % urllib.urlencode(params)
|
||||
logger.log("Newzbin search URL: " + url, logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(url)
|
||||
|
||||
return data
|
||||
return self.getRSSFeed(url)
|
||||
|
||||
def _checkAuth(self):
|
||||
if sickbeard.NEWZBIN_USERNAME in (None, "") or sickbeard.NEWZBIN_PASSWORD in (None, ""):
|
||||
@ -342,9 +340,7 @@ class NewzbinCache(tvcache.TVCache):
|
||||
|
||||
def _getRSSData(self):
|
||||
|
||||
data = self.provider._getRSSData()
|
||||
|
||||
return data
|
||||
return self.provider._getRSSData()
|
||||
|
||||
def _parseItem(self, item):
|
||||
|
||||
|
@ -166,19 +166,9 @@ class NewznabProvider(generic.NZBProvider):
|
||||
if data is None:
|
||||
return self._checkAuth()
|
||||
|
||||
if len(data.feed) > 0:
|
||||
if 'error' in data.feed:
|
||||
if data.feed.error == 100:
|
||||
raise AuthException("Your API key for " + self.name + " is incorrect, check your config.")
|
||||
elif data.feed.error == 101:
|
||||
raise AuthException("Your account on " + self.name + " has been suspended, contact the administrator.")
|
||||
elif data.feed.error == 102:
|
||||
raise AuthException(
|
||||
"Your account isn't allowed to use the API on " + self.name + ", contact the administrator")
|
||||
else:
|
||||
logger.log(u"Unknown error given from " + self.name + ": " + data.feed.title,
|
||||
logger.ERROR)
|
||||
return False
|
||||
if 'error' in data.feed:
|
||||
logger.log(u"Newznab ERROR:[%s] CODE:[%s]" % (data.feed['error']['description'], data.feed['error']['code']), logger.DEBUG)
|
||||
raise AuthException("%s" % data.feed['error']['description'])
|
||||
|
||||
return True
|
||||
|
||||
@ -206,9 +196,7 @@ class NewznabProvider(generic.NZBProvider):
|
||||
logger.log(u"Search url: " + search_url, logger.DEBUG)
|
||||
|
||||
data = self.getRSSFeed(search_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + search_url, logger.ERROR)
|
||||
return []
|
||||
|
||||
if self._checkAuthFromData(data):
|
||||
@ -279,13 +267,7 @@ class NewznabCache(tvcache.TVCache):
|
||||
|
||||
logger.log(self.provider.name + " cache update URL: " + rss_url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(rss_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + rss_url, logger.ERROR)
|
||||
return None
|
||||
|
||||
return data
|
||||
return self.provider.getRSSFeed(rss_url)
|
||||
|
||||
def _checkAuth(self, data):
|
||||
return self.provider._checkAuthFromData(data)
|
||||
|
@ -133,9 +133,7 @@ class NyaaCache(tvcache.TVCache):
|
||||
|
||||
logger.log(u"NyaaTorrents cache update URL: " + url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(url)
|
||||
|
||||
return data
|
||||
return self.provider.getRSSFeed(url)
|
||||
|
||||
def _parseItem(self, item):
|
||||
(title, url) = self.provider._get_title_and_url(item)
|
||||
|
@ -151,9 +151,7 @@ class NZBsCache(tvcache.TVCache):
|
||||
|
||||
logger.log(u"NZBs cache update URL: " + url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getURL(url)
|
||||
|
||||
return data
|
||||
return self.provider.getURL(url)
|
||||
|
||||
def _checkItemAuth(self, title, url):
|
||||
if "&i=" not in url and "&h=" not in url:
|
||||
|
@ -107,8 +107,7 @@ class NZBsRUSCache(tvcache.TVCache):
|
||||
url += urllib.urlencode(urlArgs)
|
||||
logger.log(u"NZBs'R'US cache update URL: " + url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(url)
|
||||
return data
|
||||
return self.provider.getRSSFeed(url)
|
||||
|
||||
def _checkAuth(self, data):
|
||||
return data != 'Invalid Link'
|
||||
|
@ -164,13 +164,7 @@ class OmgwtfnzbsCache(tvcache.TVCache):
|
||||
|
||||
logger.log(self.provider.name + u" cache update URL: " + rss_url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(rss_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + rss_url, logger.ERROR)
|
||||
return None
|
||||
|
||||
return data
|
||||
return self.provider.getRSSFeed(rss_url)
|
||||
|
||||
def _checkAuth(self, data):
|
||||
return self.provider._checkAuthFromData(data)
|
||||
|
@ -82,7 +82,6 @@ class TorrentRssProvider(generic.TorrentProvider):
|
||||
try:
|
||||
|
||||
data = self.cache._getRSSData()
|
||||
|
||||
if not data:
|
||||
return (False, 'No data returned from url: ' + self.url)
|
||||
|
||||
|
@ -122,7 +122,6 @@ class ThePirateBayProvider(generic.TorrentProvider):
|
||||
fileURL = self.proxy._buildURL(self.url + 'ajax_details_filelist.php?id=' + str(torrent_id))
|
||||
|
||||
data = self.getURL(fileURL)
|
||||
|
||||
if not data:
|
||||
return None
|
||||
|
||||
|
@ -84,13 +84,7 @@ class TvTorrentsCache(tvcache.TVCache):
|
||||
rss_url = self.provider.url + 'RssServlet?digest=' + sickbeard.TVTORRENTS_DIGEST + '&hash=' + sickbeard.TVTORRENTS_HASH + '&fname=true&exclude=(' + ignore_regex + ')'
|
||||
logger.log(self.provider.name + u" cache update URL: " + rss_url, logger.DEBUG)
|
||||
|
||||
data = self.provider.getRSSFeed(rss_url)
|
||||
|
||||
if not data:
|
||||
logger.log(u"No data returned from " + rss_url, logger.ERROR)
|
||||
return None
|
||||
|
||||
return data
|
||||
return self.provider.getRSSFeed(rss_url)
|
||||
|
||||
def _checkAuth(self, data):
|
||||
return self.provider._checkAuthFromData(data)
|
||||
|
@ -42,8 +42,7 @@ class WombleCache(tvcache.TVCache):
|
||||
def _getRSSData(self):
|
||||
url = self.provider.url + 'rss/?sec=TV-x264&fr=false'
|
||||
logger.log(u"Womble's Index cache update URL: " + url, logger.DEBUG)
|
||||
data = self.provider.getRSSFeed(url)
|
||||
return data
|
||||
return self.provider.getRSSFeed(url)
|
||||
|
||||
def _checkAuth(self, data):
|
||||
return data != 'Invalid Link'
|
||||
|
@ -56,8 +56,8 @@ class XEMBasicTests(test.SickbeardTestDBCase):
|
||||
|
||||
ep_date_formated = ep.airdate.strftime('%b')
|
||||
|
||||
show_name = 'UFC'
|
||||
if show_name.lower() in sickbeard.showList:
|
||||
show_name = None
|
||||
if show_name:
|
||||
print 'good'
|
||||
else:
|
||||
print 'bad'
|
||||
|
Loading…
Reference in New Issue
Block a user