mirror of
https://github.com/moparisthebest/SickRage
synced 2024-11-06 01:15:05 -05:00
0d9fbc1ad7
This version of SickBeard uses both TVDB and TVRage to search and gather it's series data from allowing you to now have access to and download shows that you couldn't before because of being locked into only what TheTVDB had to offer. Also this edition is based off the code we used in our XEM editon so it does come with scene numbering support as well as all the other features our XEM edition has to offer. Please before using this with your existing database (sickbeard.db) please make a backup copy of it and delete any other database files such as cache.db and failed.db if present, we HIGHLY recommend starting out with no database files at all to make this a fresh start but the choice is at your own risk! Enjoy!
91 lines
4.8 KiB
Python
91 lines
4.8 KiB
Python
# -*- coding: utf-8 -*-
|
|
# Copyright 2011-2012 Antoine Bertin <diaoulael@gmail.com>
|
|
#
|
|
# This file is part of subliminal.
|
|
#
|
|
# subliminal is free software; you can redistribute it and/or modify it under
|
|
# the terms of the GNU Lesser General Public License as published by
|
|
# the Free Software Foundation; either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# subliminal is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU Lesser General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU Lesser General Public License
|
|
# along with subliminal. If not, see <http://www.gnu.org/licenses/>.
|
|
from . import ServiceBase
|
|
from ..language import language_set, Language
|
|
from ..subtitles import get_subtitle_path, ResultSubtitle
|
|
from ..utils import get_keywords, split_keyword
|
|
from ..videos import Episode
|
|
from bs4 import BeautifulSoup
|
|
import logging
|
|
import re
|
|
import unicodedata
|
|
import urllib
|
|
|
|
|
|
logger = logging.getLogger("subliminal")
|
|
|
|
|
|
class Subtitulos(ServiceBase):
|
|
server_url = 'http://www.subtitulos.es'
|
|
site_url = 'http://www.subtitulos.es'
|
|
api_based = False
|
|
languages = language_set(['eng-US', 'eng-GB', 'eng', 'fre', 'pob', 'por', 'spa-ES', u'spa', u'ita', u'cat'])
|
|
language_map = {u'Español': Language('spa'), u'Español (España)': Language('spa'), #u'Español (Latinoamérica)': Language('spa'),
|
|
u'Català': Language('cat'), u'Brazilian': Language('pob'), u'English (US)': Language('eng-US'),
|
|
u'English (UK)': Language('eng-GB'), 'Galego': Language('glg')}
|
|
language_code = 'name'
|
|
videos = [Episode]
|
|
require_video = False
|
|
required_features = ['permissive']
|
|
# the '.+' in the pattern for Version allows us to match both 'ó'
|
|
# and the 'ó' char directly. This is because now BS4 converts the html
|
|
# code chars into their equivalent unicode char
|
|
release_pattern = re.compile('Versi.+n (.+) ([0-9]+).([0-9])+ megabytes')
|
|
extra_keywords_pattern = re.compile("(?:con|para)\s(?:720p)?(?:\-|\s)?([A-Za-z]+)(?:\-|\s)?(?:720p)?(?:\s|\.)(?:y\s)?(?:720p)?(?:\-\s)?([A-Za-z]+)?(?:\-\s)?(?:720p)?(?:\.)?");
|
|
|
|
def list_checked(self, video, languages):
|
|
return self.query(video.path or video.release, languages, get_keywords(video.guess), video.series, video.season, video.episode)
|
|
|
|
def query(self, filepath, languages, keywords, series, season, episode):
|
|
request_series = series.lower().replace(' ', '-').replace('&', '@').replace('(','').replace(')','')
|
|
if isinstance(request_series, unicode):
|
|
request_series = unicodedata.normalize('NFKD', request_series).encode('ascii', 'ignore')
|
|
logger.debug(u'Getting subtitles for %s season %d episode %d with languages %r' % (series, season, episode, languages))
|
|
r = self.session.get('%s/%s/%sx%.2d' % (self.server_url, urllib.quote(request_series), season, episode))
|
|
if r.status_code == 404:
|
|
logger.debug(u'Could not find subtitles for %s season %d episode %d with languages %r' % (series, season, episode, languages))
|
|
return []
|
|
if r.status_code != 200:
|
|
logger.error(u'Request %s returned status code %d' % (r.url, r.status_code))
|
|
return []
|
|
soup = BeautifulSoup(r.content, self.required_features)
|
|
subtitles = []
|
|
for sub in soup('div', {'id': 'version'}):
|
|
sub_keywords = split_keyword(self.release_pattern.search(sub.find('p', {'class': 'title-sub'}).contents[1]).group(1).lower())
|
|
if keywords and not keywords & sub_keywords:
|
|
logger.debug(u'None of subtitle keywords %r in %r' % (sub_keywords, keywords))
|
|
continue
|
|
for html_language in sub.findAllNext('ul', {'class': 'sslist'}):
|
|
language = self.get_language(html_language.findNext('li', {'class': 'li-idioma'}).find('strong').contents[0].string.strip())
|
|
if language not in languages:
|
|
logger.debug(u'Language %r not in wanted languages %r' % (language, languages))
|
|
continue
|
|
html_status = html_language.findNext('li', {'class': 'li-estado green'})
|
|
status = html_status.contents[0].string.strip()
|
|
if status != 'Completado':
|
|
logger.debug(u'Wrong subtitle status %s' % status)
|
|
continue
|
|
path = get_subtitle_path(filepath, language, self.config.multi)
|
|
subtitle = ResultSubtitle(path, language, self.__class__.__name__.lower(), html_status.findNext('span', {'class': 'descargar green'}).find('a')['href'],
|
|
keywords=sub_keywords)
|
|
subtitles.append(subtitle)
|
|
return subtitles
|
|
|
|
|
|
Service = Subtitulos
|