mirror of
https://github.com/moparisthebest/SickRage
synced 2024-11-17 14:55:07 -05:00
0d9fbc1ad7
This version of SickBeard uses both TVDB and TVRage to search and gather it's series data from allowing you to now have access to and download shows that you couldn't before because of being locked into only what TheTVDB had to offer. Also this edition is based off the code we used in our XEM editon so it does come with scene numbering support as well as all the other features our XEM edition has to offer. Please before using this with your existing database (sickbeard.db) please make a backup copy of it and delete any other database files such as cache.db and failed.db if present, we HIGHLY recommend starting out with no database files at all to make this a fresh start but the choice is at your own risk! Enjoy!
1893 lines
77 KiB
Python
1893 lines
77 KiB
Python
"""
|
|
parser.http.movieParser module (imdb package).
|
|
|
|
This module provides the classes (and the instances), used to parse the
|
|
IMDb pages on the akas.imdb.com server about a movie.
|
|
E.g., for Brian De Palma's "The Untouchables", the referred
|
|
pages would be:
|
|
combined details: http://akas.imdb.com/title/tt0094226/combined
|
|
plot summary: http://akas.imdb.com/title/tt0094226/plotsummary
|
|
...and so on...
|
|
|
|
Copyright 2004-2012 Davide Alberani <da@erlug.linux.it>
|
|
2008 H. Turgut Uyar <uyar@tekir.org>
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
"""
|
|
|
|
import re
|
|
import urllib
|
|
|
|
from imdb import imdbURL_base
|
|
from imdb.Person import Person
|
|
from imdb.Movie import Movie
|
|
from imdb.Company import Company
|
|
from imdb.utils import analyze_title, split_company_name_notes, _Container
|
|
from utils import build_person, DOMParserBase, Attribute, Extractor, \
|
|
analyze_imdbid
|
|
|
|
|
|
# Dictionary used to convert some section's names.
|
|
_SECT_CONV = {
|
|
'directed': 'director',
|
|
'directed by': 'director',
|
|
'directors': 'director',
|
|
'editors': 'editor',
|
|
'writing credits': 'writer',
|
|
'writers': 'writer',
|
|
'produced': 'producer',
|
|
'cinematography': 'cinematographer',
|
|
'film editing': 'editor',
|
|
'casting': 'casting director',
|
|
'costume design': 'costume designer',
|
|
'makeup department': 'make up',
|
|
'production management': 'production manager',
|
|
'second unit director or assistant director': 'assistant director',
|
|
'costume and wardrobe department': 'costume department',
|
|
'sound department': 'sound crew',
|
|
'stunts': 'stunt performer',
|
|
'other crew': 'miscellaneous crew',
|
|
'also known as': 'akas',
|
|
'country': 'countries',
|
|
'runtime': 'runtimes',
|
|
'language': 'languages',
|
|
'certification': 'certificates',
|
|
'genre': 'genres',
|
|
'created': 'creator',
|
|
'creators': 'creator',
|
|
'color': 'color info',
|
|
'plot': 'plot outline',
|
|
'seasons': 'number of seasons',
|
|
'art directors': 'art direction',
|
|
'assistant directors': 'assistant director',
|
|
'set decorators': 'set decoration',
|
|
'visual effects department': 'visual effects',
|
|
'production managers': 'production manager',
|
|
'miscellaneous': 'miscellaneous crew',
|
|
'make up department': 'make up',
|
|
'plot summary': 'plot outline',
|
|
'cinematographers': 'cinematographer',
|
|
'camera department': 'camera and electrical department',
|
|
'costume designers': 'costume designer',
|
|
'production designers': 'production design',
|
|
'production managers': 'production manager',
|
|
'music original': 'original music',
|
|
'casting directors': 'casting director',
|
|
'other companies': 'miscellaneous companies',
|
|
'producers': 'producer',
|
|
'special effects by': 'special effects department',
|
|
'special effects': 'special effects companies'
|
|
}
|
|
|
|
|
|
def _manageRoles(mo):
|
|
"""Perform some transformation on the html, so that roleIDs can
|
|
be easily retrieved."""
|
|
firstHalf = mo.group(1)
|
|
secondHalf = mo.group(2)
|
|
newRoles = []
|
|
roles = secondHalf.split(' / ')
|
|
for role in roles:
|
|
role = role.strip()
|
|
if not role:
|
|
continue
|
|
roleID = analyze_imdbid(role)
|
|
if roleID is None:
|
|
roleID = u'/'
|
|
else:
|
|
roleID += u'/'
|
|
newRoles.append(u'<div class="_imdbpyrole" roleid="%s">%s</div>' % \
|
|
(roleID, role.strip()))
|
|
return firstHalf + u' / '.join(newRoles) + mo.group(3)
|
|
|
|
|
|
_reRolesMovie = re.compile(r'(<td class="char">)(.*?)(</td>)',
|
|
re.I | re.M | re.S)
|
|
|
|
def _replaceBR(mo):
|
|
"""Replaces <br> tags with '::' (useful for some akas)"""
|
|
txt = mo.group(0)
|
|
return txt.replace('<br>', '::')
|
|
|
|
_reAkas = re.compile(r'<h5>also known as:</h5>.*?</div>', re.I | re.M | re.S)
|
|
|
|
def makeSplitter(lstrip=None, sep='|', comments=True,
|
|
origNotesSep=' (', newNotesSep='::(', strip=None):
|
|
"""Return a splitter function suitable for a given set of data."""
|
|
def splitter(x):
|
|
if not x: return x
|
|
x = x.strip()
|
|
if not x: return x
|
|
if lstrip is not None:
|
|
x = x.lstrip(lstrip).lstrip()
|
|
lx = x.split(sep)
|
|
lx[:] = filter(None, [j.strip() for j in lx])
|
|
if comments:
|
|
lx[:] = [j.replace(origNotesSep, newNotesSep, 1) for j in lx]
|
|
if strip:
|
|
lx[:] = [j.strip(strip) for j in lx]
|
|
return lx
|
|
return splitter
|
|
|
|
|
|
def _toInt(val, replace=()):
|
|
"""Return the value, converted to integer, or None; if present, 'replace'
|
|
must be a list of tuples of values to replace."""
|
|
for before, after in replace:
|
|
val = val.replace(before, after)
|
|
try:
|
|
return int(val)
|
|
except (TypeError, ValueError):
|
|
return None
|
|
|
|
|
|
class DOMHTMLMovieParser(DOMParserBase):
|
|
"""Parser for the "combined details" (and if instance.mdparse is
|
|
True also for the "main details") page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
mparser = DOMHTMLMovieParser()
|
|
result = mparser.parse(combined_details_html_string)
|
|
"""
|
|
_containsObjects = True
|
|
|
|
extractors = [Extractor(label='title',
|
|
path="//h1",
|
|
attrs=Attribute(key='title',
|
|
path=".//text()",
|
|
postprocess=analyze_title)),
|
|
|
|
Extractor(label='glossarysections',
|
|
group="//a[@class='glossary']",
|
|
group_key="./@name",
|
|
group_key_normalize=lambda x: x.replace('_', ' '),
|
|
path="../../../..//tr",
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={'person': ".//text()",
|
|
'link': "./td[1]/a[@href]/@href"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('person') or u'',
|
|
personID=analyze_imdbid(x.get('link')))
|
|
)),
|
|
|
|
Extractor(label='cast',
|
|
path="//table[@class='cast']//tr",
|
|
attrs=Attribute(key="cast",
|
|
multi=True,
|
|
path={'person': ".//text()",
|
|
'link': "td[2]/a/@href",
|
|
'roleID': \
|
|
"td[4]/div[@class='_imdbpyrole']/@roleid"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('person') or u'',
|
|
personID=analyze_imdbid(x.get('link')),
|
|
roleID=(x.get('roleID') or u'').split('/'))
|
|
)),
|
|
|
|
Extractor(label='genres',
|
|
path="//div[@class='info']//a[starts-with(@href," \
|
|
" '/Sections/Genres')]",
|
|
attrs=Attribute(key="genres",
|
|
multi=True,
|
|
path="./text()")),
|
|
|
|
Extractor(label='h5sections',
|
|
path="//div[@class='info']/h5/..",
|
|
attrs=[
|
|
Attribute(key="plot summary",
|
|
path="./h5[starts-with(text(), " \
|
|
"'Plot:')]/../div/text()",
|
|
postprocess=lambda x: \
|
|
x.strip().rstrip('|').rstrip()),
|
|
Attribute(key="aspect ratio",
|
|
path="./h5[starts-with(text()," \
|
|
" 'Aspect')]/../div/text()",
|
|
postprocess=lambda x: x.strip()),
|
|
Attribute(key="mpaa",
|
|
path="./h5/a[starts-with(text()," \
|
|
" 'MPAA')]/../../div/text()",
|
|
postprocess=lambda x: x.strip()),
|
|
Attribute(key="countries",
|
|
path="./h5[starts-with(text(), " \
|
|
"'Countr')]/../div[@class='info-content']//text()",
|
|
postprocess=makeSplitter('|')),
|
|
Attribute(key="language",
|
|
path="./h5[starts-with(text(), " \
|
|
"'Language')]/..//text()",
|
|
postprocess=makeSplitter('Language:')),
|
|
Attribute(key='color info',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Color')]/..//text()",
|
|
postprocess=makeSplitter('Color:')),
|
|
Attribute(key='sound mix',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Sound Mix')]/..//text()",
|
|
postprocess=makeSplitter('Sound Mix:')),
|
|
# Collects akas not encosed in <i> tags.
|
|
Attribute(key='other akas',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Also Known As')]/../div//text()",
|
|
postprocess=makeSplitter(sep='::',
|
|
origNotesSep='" - ',
|
|
newNotesSep='::',
|
|
strip='"')),
|
|
Attribute(key='runtimes',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Runtime')]/../div/text()",
|
|
postprocess=makeSplitter()),
|
|
Attribute(key='certificates',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Certificat')]/..//text()",
|
|
postprocess=makeSplitter('Certification:')),
|
|
Attribute(key='number of seasons',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Seasons')]/..//text()",
|
|
postprocess=lambda x: x.count('|') + 1),
|
|
Attribute(key='original air date',
|
|
path="./h5[starts-with(text(), " \
|
|
"'Original Air Date')]/../div/text()"),
|
|
Attribute(key='tv series link',
|
|
path="./h5[starts-with(text(), " \
|
|
"'TV Series')]/..//a/@href"),
|
|
Attribute(key='tv series title',
|
|
path="./h5[starts-with(text(), " \
|
|
"'TV Series')]/..//a/text()")
|
|
]),
|
|
|
|
Extractor(label='language codes',
|
|
path="//h5[starts-with(text(), 'Language')]/..//a[starts-with(@href, '/language/')]",
|
|
attrs=Attribute(key='language codes', multi=True,
|
|
path="./@href",
|
|
postprocess=lambda x: x.split('/')[2].strip()
|
|
)),
|
|
|
|
Extractor(label='country codes',
|
|
path="//h5[starts-with(text(), 'Country')]/..//a[starts-with(@href, '/country/')]",
|
|
attrs=Attribute(key='country codes', multi=True,
|
|
path="./@href",
|
|
postprocess=lambda x: x.split('/')[2].strip()
|
|
)),
|
|
|
|
Extractor(label='creator',
|
|
path="//h5[starts-with(text(), 'Creator')]/..//a",
|
|
attrs=Attribute(key='creator', multi=True,
|
|
path={'name': "./text()",
|
|
'link': "./@href"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('name') or u'',
|
|
personID=analyze_imdbid(x.get('link')))
|
|
)),
|
|
|
|
Extractor(label='thin writer',
|
|
path="//h5[starts-with(text(), 'Writer')]/..//a",
|
|
attrs=Attribute(key='thin writer', multi=True,
|
|
path={'name': "./text()",
|
|
'link': "./@href"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('name') or u'',
|
|
personID=analyze_imdbid(x.get('link')))
|
|
)),
|
|
|
|
Extractor(label='thin director',
|
|
path="//h5[starts-with(text(), 'Director')]/..//a",
|
|
attrs=Attribute(key='thin director', multi=True,
|
|
path={'name': "./text()",
|
|
'link': "@href"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('name') or u'',
|
|
personID=analyze_imdbid(x.get('link')))
|
|
)),
|
|
|
|
Extractor(label='top 250/bottom 100',
|
|
path="//div[@class='starbar-special']/" \
|
|
"a[starts-with(@href, '/chart/')]",
|
|
attrs=Attribute(key='top/bottom rank',
|
|
path="./text()")),
|
|
|
|
Extractor(label='series years',
|
|
path="//div[@id='tn15title']//span" \
|
|
"[starts-with(text(), 'TV series')]",
|
|
attrs=Attribute(key='series years',
|
|
path="./text()",
|
|
postprocess=lambda x: \
|
|
x.replace('TV series','').strip())),
|
|
|
|
Extractor(label='number of episodes',
|
|
path="//a[@title='Full Episode List']",
|
|
attrs=Attribute(key='number of episodes',
|
|
path="./text()",
|
|
postprocess=lambda x: \
|
|
_toInt(x, [(' Episodes', '')]))),
|
|
|
|
Extractor(label='akas',
|
|
path="//i[@class='transl']",
|
|
attrs=Attribute(key='akas', multi=True, path='text()',
|
|
postprocess=lambda x:
|
|
x.replace(' ', ' ').rstrip('-').replace('" - ',
|
|
'"::', 1).strip('"').replace(' ', ' '))),
|
|
|
|
Extractor(label='production notes/status',
|
|
path="//h5[starts-with(text(), 'Status:')]/..//div[@class='info-content']",
|
|
attrs=Attribute(key='production status',
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip().split('|')[0].strip().lower())),
|
|
|
|
Extractor(label='production notes/status updated',
|
|
path="//h5[starts-with(text(), 'Status Updated:')]/..//div[@class='info-content']",
|
|
attrs=Attribute(key='production status updated',
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip())),
|
|
|
|
Extractor(label='production notes/comments',
|
|
path="//h5[starts-with(text(), 'Comments:')]/..//div[@class='info-content']",
|
|
attrs=Attribute(key='production comments',
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip())),
|
|
|
|
Extractor(label='production notes/note',
|
|
path="//h5[starts-with(text(), 'Note:')]/..//div[@class='info-content']",
|
|
attrs=Attribute(key='production note',
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip())),
|
|
|
|
Extractor(label='blackcatheader',
|
|
group="//b[@class='blackcatheader']",
|
|
group_key="./text()",
|
|
group_key_normalize=lambda x: x.lower(),
|
|
path="../ul/li",
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={'name': "./a//text()",
|
|
'comp-link': "./a/@href",
|
|
'notes': "./text()"},
|
|
postprocess=lambda x: \
|
|
Company(name=x.get('name') or u'',
|
|
companyID=analyze_imdbid(x.get('comp-link')),
|
|
notes=(x.get('notes') or u'').strip())
|
|
)),
|
|
|
|
Extractor(label='rating',
|
|
path="//div[@class='starbar-meta']/b",
|
|
attrs=Attribute(key='rating',
|
|
path=".//text()")),
|
|
|
|
Extractor(label='votes',
|
|
path="//div[@class='starbar-meta']/a[@href]",
|
|
attrs=Attribute(key='votes',
|
|
path=".//text()")),
|
|
|
|
Extractor(label='cover url',
|
|
path="//a[@name='poster']",
|
|
attrs=Attribute(key='cover url',
|
|
path="./img/@src"))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile(r'(<b class="blackcatheader">.+?</b>)', re.I),
|
|
r'</div><div>\1'),
|
|
('<small>Full cast and crew for<br>', ''),
|
|
('<td> </td>', '<td>...</td>'),
|
|
('<span class="tv-extra">TV mini-series</span>',
|
|
'<span class="tv-extra">(mini)</span>'),
|
|
(_reRolesMovie, _manageRoles),
|
|
(_reAkas, _replaceBR)]
|
|
|
|
def preprocess_dom(self, dom):
|
|
# Handle series information.
|
|
xpath = self.xpath(dom, "//b[text()='Series Crew']")
|
|
if xpath:
|
|
b = xpath[-1] # In doubt, take the last one.
|
|
for a in self.xpath(b, "./following::h5/a[@class='glossary']"):
|
|
name = a.get('name')
|
|
if name:
|
|
a.set('name', 'series %s' % name)
|
|
# Remove links to IMDbPro.
|
|
for proLink in self.xpath(dom, "//span[@class='pro-link']"):
|
|
proLink.drop_tree()
|
|
# Remove some 'more' links (keep others, like the one around
|
|
# the number of votes).
|
|
for tn15more in self.xpath(dom,
|
|
"//a[@class='tn15more'][starts-with(@href, '/title/')]"):
|
|
tn15more.drop_tree()
|
|
return dom
|
|
|
|
re_space = re.compile(r'\s+')
|
|
re_airdate = re.compile(r'(.*)\s*\(season (\d+), episode (\d+)\)', re.I)
|
|
def postprocess_data(self, data):
|
|
# Convert section names.
|
|
for sect in data.keys():
|
|
if sect in _SECT_CONV:
|
|
data[_SECT_CONV[sect]] = data[sect]
|
|
del data[sect]
|
|
sect = _SECT_CONV[sect]
|
|
# Filter out fake values.
|
|
for key in data:
|
|
value = data[key]
|
|
if isinstance(value, list) and value:
|
|
if isinstance(value[0], Person):
|
|
data[key] = filter(lambda x: x.personID is not None, value)
|
|
if isinstance(value[0], _Container):
|
|
for obj in data[key]:
|
|
obj.accessSystem = self._as
|
|
obj.modFunct = self._modFunct
|
|
if 'akas' in data or 'other akas' in data:
|
|
akas = data.get('akas') or []
|
|
other_akas = data.get('other akas') or []
|
|
akas += other_akas
|
|
nakas = []
|
|
for aka in akas:
|
|
aka = aka.strip()
|
|
if aka.endswith('" -'):
|
|
aka = aka[:-3].rstrip()
|
|
nakas.append(aka)
|
|
if 'akas' in data:
|
|
del data['akas']
|
|
if 'other akas' in data:
|
|
del data['other akas']
|
|
if nakas:
|
|
data['akas'] = nakas
|
|
if 'runtimes' in data:
|
|
data['runtimes'] = [x.replace(' min', u'')
|
|
for x in data['runtimes']]
|
|
if 'original air date' in data:
|
|
oid = self.re_space.sub(' ', data['original air date']).strip()
|
|
data['original air date'] = oid
|
|
aid = self.re_airdate.findall(oid)
|
|
if aid and len(aid[0]) == 3:
|
|
date, season, episode = aid[0]
|
|
date = date.strip()
|
|
try: season = int(season)
|
|
except: pass
|
|
try: episode = int(episode)
|
|
except: pass
|
|
if date and date != '????':
|
|
data['original air date'] = date
|
|
else:
|
|
del data['original air date']
|
|
# Handle also "episode 0".
|
|
if season or type(season) is type(0):
|
|
data['season'] = season
|
|
if episode or type(season) is type(0):
|
|
data['episode'] = episode
|
|
for k in ('writer', 'director'):
|
|
t_k = 'thin %s' % k
|
|
if t_k not in data:
|
|
continue
|
|
if k not in data:
|
|
data[k] = data[t_k]
|
|
del data[t_k]
|
|
if 'top/bottom rank' in data:
|
|
tbVal = data['top/bottom rank'].lower()
|
|
if tbVal.startswith('top'):
|
|
tbKey = 'top 250 rank'
|
|
tbVal = _toInt(tbVal, [('top 250: #', '')])
|
|
else:
|
|
tbKey = 'bottom 100 rank'
|
|
tbVal = _toInt(tbVal, [('bottom 100: #', '')])
|
|
if tbVal:
|
|
data[tbKey] = tbVal
|
|
del data['top/bottom rank']
|
|
if 'year' in data and data['year'] == '????':
|
|
del data['year']
|
|
if 'tv series link' in data:
|
|
if 'tv series title' in data:
|
|
data['episode of'] = Movie(title=data['tv series title'],
|
|
movieID=analyze_imdbid(
|
|
data['tv series link']),
|
|
accessSystem=self._as,
|
|
modFunct=self._modFunct)
|
|
del data['tv series title']
|
|
del data['tv series link']
|
|
if 'rating' in data:
|
|
try:
|
|
data['rating'] = float(data['rating'].replace('/10', ''))
|
|
except (TypeError, ValueError):
|
|
pass
|
|
if 'votes' in data:
|
|
try:
|
|
votes = data['votes'].replace(',', '').replace('votes', '')
|
|
data['votes'] = int(votes)
|
|
except (TypeError, ValueError):
|
|
pass
|
|
return data
|
|
|
|
|
|
def _process_plotsummary(x):
|
|
"""Process a plot (contributed by Rdian06)."""
|
|
xauthor = x.get('author')
|
|
if xauthor:
|
|
xauthor = xauthor.replace('{', '<').replace('}', '>').replace('(',
|
|
'<').replace(')', '>').strip()
|
|
xplot = x.get('plot', u'').strip()
|
|
if xauthor:
|
|
xplot += u'::%s' % xauthor
|
|
return xplot
|
|
|
|
class DOMHTMLPlotParser(DOMParserBase):
|
|
"""Parser for the "plot summary" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a 'plot' key, containing a list
|
|
of string with the structure: 'summary::summary_author <author@email>'.
|
|
|
|
Example:
|
|
pparser = HTMLPlotParser()
|
|
result = pparser.parse(plot_summary_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
# Notice that recently IMDb started to put the email of the
|
|
# author only in the link, that we're not collecting, here.
|
|
extractors = [Extractor(label='plot',
|
|
path="//p[@class='plotpar']",
|
|
attrs=Attribute(key='plot',
|
|
multi=True,
|
|
path={'plot': './text()',
|
|
'author': './i/a/text()'},
|
|
postprocess=_process_plotsummary))]
|
|
|
|
|
|
def _process_award(x):
|
|
award = {}
|
|
award['award'] = x.get('award').strip()
|
|
if not award['award']:
|
|
return {}
|
|
award['year'] = x.get('year').strip()
|
|
if award['year'] and award['year'].isdigit():
|
|
award['year'] = int(award['year'])
|
|
award['result'] = x.get('result').strip()
|
|
category = x.get('category').strip()
|
|
if category:
|
|
award['category'] = category
|
|
received_with = x.get('with')
|
|
if received_with is not None:
|
|
award['with'] = received_with.strip()
|
|
notes = x.get('notes')
|
|
if notes is not None:
|
|
notes = notes.strip()
|
|
if notes:
|
|
award['notes'] = notes
|
|
award['anchor'] = x.get('anchor')
|
|
return award
|
|
|
|
|
|
|
|
class DOMHTMLAwardsParser(DOMParserBase):
|
|
"""Parser for the "awards" page of a given person or movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
awparser = HTMLAwardsParser()
|
|
result = awparser.parse(awards_html_string)
|
|
"""
|
|
subject = 'title'
|
|
_containsObjects = True
|
|
|
|
extractors = [
|
|
Extractor(label='awards',
|
|
group="//table//big",
|
|
group_key="./a",
|
|
path="./ancestor::tr[1]/following-sibling::tr/" \
|
|
"td[last()][not(@colspan)]",
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={
|
|
'year': "../td[1]/a/text()",
|
|
'result': "../td[2]/b/text()",
|
|
'award': "../td[3]/text()",
|
|
'category': "./text()[1]",
|
|
# FIXME: takes only the first co-recipient
|
|
'with': "./small[starts-with(text()," \
|
|
" 'Shared with:')]/following-sibling::a[1]/text()",
|
|
'notes': "./small[last()]//text()",
|
|
'anchor': ".//text()"
|
|
},
|
|
postprocess=_process_award
|
|
)),
|
|
Extractor(label='recipients',
|
|
group="//table//big",
|
|
group_key="./a",
|
|
path="./ancestor::tr[1]/following-sibling::tr/" \
|
|
"td[last()]/small[1]/preceding-sibling::a",
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={
|
|
'name': "./text()",
|
|
'link': "./@href",
|
|
'anchor': "..//text()"
|
|
}
|
|
))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<tr><td[^>]*>.*?</td></tr>\n\n</table>)', re.I),
|
|
r'\1</table>'),
|
|
(re.compile('(<tr><td[^>]*>\n\n<big>.*?</big></td></tr>)', re.I),
|
|
r'</table><table class="_imdbpy">\1'),
|
|
(re.compile('(<table[^>]*>\n\n)</table>(<table)', re.I), r'\1\2'),
|
|
(re.compile('(<small>.*?)<br>(.*?</small)', re.I), r'\1 \2'),
|
|
(re.compile('(</tr>\n\n)(<td)', re.I), r'\1<tr>\2')
|
|
]
|
|
|
|
def preprocess_dom(self, dom):
|
|
"""Repeat td elements according to their rowspan attributes
|
|
in subsequent tr elements.
|
|
"""
|
|
cols = self.xpath(dom, "//td[@rowspan]")
|
|
for col in cols:
|
|
span = int(col.get('rowspan'))
|
|
del col.attrib['rowspan']
|
|
position = len(self.xpath(col, "./preceding-sibling::td"))
|
|
row = col.getparent()
|
|
for tr in self.xpath(row, "./following-sibling::tr")[:span-1]:
|
|
# if not cloned, child will be moved to new parent
|
|
clone = self.clone(col)
|
|
# XXX: beware that here we don't use an "adapted" function,
|
|
# because both BeautifulSoup and lxml uses the same
|
|
# "insert" method.
|
|
tr.insert(position, clone)
|
|
return dom
|
|
|
|
def postprocess_data(self, data):
|
|
if len(data) == 0:
|
|
return {}
|
|
nd = []
|
|
for key in data.keys():
|
|
dom = self.get_dom(key)
|
|
assigner = self.xpath(dom, "//a/text()")[0]
|
|
for entry in data[key]:
|
|
if not entry.has_key('name'):
|
|
if not entry:
|
|
continue
|
|
# this is an award, not a recipient
|
|
entry['assigner'] = assigner.strip()
|
|
# find the recipients
|
|
matches = [p for p in data[key]
|
|
if p.has_key('name') and (entry['anchor'] ==
|
|
p['anchor'])]
|
|
if self.subject == 'title':
|
|
recipients = [Person(name=recipient['name'],
|
|
personID=analyze_imdbid(recipient['link']))
|
|
for recipient in matches]
|
|
entry['to'] = recipients
|
|
elif self.subject == 'name':
|
|
recipients = [Movie(title=recipient['name'],
|
|
movieID=analyze_imdbid(recipient['link']))
|
|
for recipient in matches]
|
|
entry['for'] = recipients
|
|
nd.append(entry)
|
|
del entry['anchor']
|
|
return {'awards': nd}
|
|
|
|
|
|
class DOMHTMLTaglinesParser(DOMParserBase):
|
|
"""Parser for the "taglines" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
tparser = DOMHTMLTaglinesParser()
|
|
result = tparser.parse(taglines_html_string)
|
|
"""
|
|
extractors = [Extractor(label='taglines',
|
|
path="//div[@id='tn15content']/p",
|
|
attrs=Attribute(key='taglines', multi=True,
|
|
path="./text()"))]
|
|
|
|
|
|
class DOMHTMLKeywordsParser(DOMParserBase):
|
|
"""Parser for the "keywords" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
kwparser = DOMHTMLKeywordsParser()
|
|
result = kwparser.parse(keywords_html_string)
|
|
"""
|
|
extractors = [Extractor(label='keywords',
|
|
path="//a[starts-with(@href, '/keyword/')]",
|
|
attrs=Attribute(key='keywords',
|
|
path="./text()", multi=True,
|
|
postprocess=lambda x: \
|
|
x.lower().replace(' ', '-')))]
|
|
|
|
|
|
class DOMHTMLAlternateVersionsParser(DOMParserBase):
|
|
"""Parser for the "alternate versions" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
avparser = HTMLAlternateVersionsParser()
|
|
result = avparser.parse(alternateversions_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
extractors = [Extractor(label='alternate versions',
|
|
path="//ul[@class='trivia']/li",
|
|
attrs=Attribute(key='alternate versions',
|
|
multi=True,
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip()))]
|
|
|
|
|
|
class DOMHTMLTriviaParser(DOMParserBase):
|
|
"""Parser for the "trivia" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
avparser = HTMLAlternateVersionsParser()
|
|
result = avparser.parse(alternateversions_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
extractors = [Extractor(label='alternate versions',
|
|
path="//div[@class='sodatext']",
|
|
attrs=Attribute(key='trivia',
|
|
multi=True,
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip()))]
|
|
|
|
def preprocess_dom(self, dom):
|
|
# Remove "link this quote" links.
|
|
for qLink in self.xpath(dom, "//span[@class='linksoda']"):
|
|
qLink.drop_tree()
|
|
return dom
|
|
|
|
|
|
|
|
class DOMHTMLSoundtrackParser(DOMHTMLAlternateVersionsParser):
|
|
kind = 'soundtrack'
|
|
|
|
preprocessors = [
|
|
('<br>', '\n')
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
if 'soundtrack' in data:
|
|
nd = []
|
|
for x in data['soundtrack']:
|
|
ds = x.split('\n')
|
|
title = ds[0]
|
|
if title[0] == '"' and title[-1] == '"':
|
|
title = title[1:-1]
|
|
nds = []
|
|
newData = {}
|
|
for l in ds[1:]:
|
|
if ' with ' in l or ' by ' in l or ' from ' in l \
|
|
or ' of ' in l or l.startswith('From '):
|
|
nds.append(l)
|
|
else:
|
|
if nds:
|
|
nds[-1] += l
|
|
else:
|
|
nds.append(l)
|
|
newData[title] = {}
|
|
for l in nds:
|
|
skip = False
|
|
for sep in ('From ',):
|
|
if l.startswith(sep):
|
|
fdix = len(sep)
|
|
kind = l[:fdix].rstrip().lower()
|
|
info = l[fdix:].lstrip()
|
|
newData[title][kind] = info
|
|
skip = True
|
|
if not skip:
|
|
for sep in ' with ', ' by ', ' from ', ' of ':
|
|
fdix = l.find(sep)
|
|
if fdix != -1:
|
|
fdix = fdix+len(sep)
|
|
kind = l[:fdix].rstrip().lower()
|
|
info = l[fdix:].lstrip()
|
|
newData[title][kind] = info
|
|
break
|
|
nd.append(newData)
|
|
data['soundtrack'] = nd
|
|
return data
|
|
|
|
|
|
class DOMHTMLCrazyCreditsParser(DOMParserBase):
|
|
"""Parser for the "crazy credits" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
ccparser = DOMHTMLCrazyCreditsParser()
|
|
result = ccparser.parse(crazycredits_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
extractors = [Extractor(label='crazy credits', path="//ul/li/tt",
|
|
attrs=Attribute(key='crazy credits', multi=True,
|
|
path=".//text()",
|
|
postprocess=lambda x: \
|
|
x.replace('\n', ' ').replace(' ', ' ')))]
|
|
|
|
|
|
class DOMHTMLGoofsParser(DOMParserBase):
|
|
"""Parser for the "goofs" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
gparser = DOMHTMLGoofsParser()
|
|
result = gparser.parse(goofs_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
extractors = [Extractor(label='goofs', path="//ul[@class='trivia']/li",
|
|
attrs=Attribute(key='goofs', multi=True, path=".//text()",
|
|
postprocess=lambda x: (x or u'').strip()))]
|
|
|
|
|
|
class DOMHTMLQuotesParser(DOMParserBase):
|
|
"""Parser for the "memorable quotes" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
qparser = DOMHTMLQuotesParser()
|
|
result = qparser.parse(quotes_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
extractors = [
|
|
Extractor(label='quotes',
|
|
path="//div[@class='_imdbpy']",
|
|
attrs=Attribute(key='quotes',
|
|
multi=True,
|
|
path=".//text()",
|
|
postprocess=lambda x: x.strip().replace(' \n',
|
|
'::').replace('::\n', '::').replace('\n', ' ')))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<a name="?qt[0-9]{7}"?></a>)', re.I),
|
|
r'\1<div class="_imdbpy">'),
|
|
(re.compile('<hr width="30%">', re.I), '</div>'),
|
|
(re.compile('<hr/>', re.I), '</div>'),
|
|
(re.compile('<script.*?</script>', re.I|re.S), ''),
|
|
# For BeautifulSoup.
|
|
(re.compile('<!-- sid: t-channel : MIDDLE_CENTER -->', re.I), '</div>')
|
|
]
|
|
|
|
def preprocess_dom(self, dom):
|
|
# Remove "link this quote" links.
|
|
for qLink in self.xpath(dom, "//p[@class='linksoda']"):
|
|
qLink.drop_tree()
|
|
return dom
|
|
|
|
def postprocess_data(self, data):
|
|
if 'quotes' not in data:
|
|
return {}
|
|
for idx, quote in enumerate(data['quotes']):
|
|
data['quotes'][idx] = quote.split('::')
|
|
return data
|
|
|
|
|
|
class DOMHTMLReleaseinfoParser(DOMParserBase):
|
|
"""Parser for the "release dates" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
rdparser = DOMHTMLReleaseinfoParser()
|
|
result = rdparser.parse(releaseinfo_html_string)
|
|
"""
|
|
extractors = [Extractor(label='release dates',
|
|
path="//th[@class='xxxx']/../../tr",
|
|
attrs=Attribute(key='release dates', multi=True,
|
|
path={'country': ".//td[1]//text()",
|
|
'date': ".//td[2]//text()",
|
|
'notes': ".//td[3]//text()"})),
|
|
Extractor(label='akas',
|
|
path="//div[@class='_imdbpy_akas']/table/tr",
|
|
attrs=Attribute(key='akas', multi=True,
|
|
path={'title': "./td[1]/text()",
|
|
'countries': "./td[2]/text()"}))]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<h5><a name="?akas"?.*</table>)', re.I | re.M | re.S),
|
|
r'<div class="_imdbpy_akas">\1</div>')]
|
|
|
|
def postprocess_data(self, data):
|
|
if not ('release dates' in data or 'akas' in data): return data
|
|
releases = data.get('release dates') or []
|
|
rl = []
|
|
for i in releases:
|
|
country = i.get('country')
|
|
date = i.get('date')
|
|
if not (country and date): continue
|
|
country = country.strip()
|
|
date = date.strip()
|
|
if not (country and date): continue
|
|
notes = i['notes']
|
|
info = u'%s::%s' % (country, date)
|
|
if notes:
|
|
info += notes
|
|
rl.append(info)
|
|
if releases:
|
|
del data['release dates']
|
|
if rl:
|
|
data['release dates'] = rl
|
|
akas = data.get('akas') or []
|
|
nakas = []
|
|
for aka in akas:
|
|
title = (aka.get('title') or '').strip()
|
|
if not title:
|
|
continue
|
|
countries = (aka.get('countries') or '').split('/')
|
|
if not countries:
|
|
nakas.append(title)
|
|
else:
|
|
for country in countries:
|
|
nakas.append('%s::%s' % (title, country.strip()))
|
|
if akas:
|
|
del data['akas']
|
|
if nakas:
|
|
data['akas from release info'] = nakas
|
|
return data
|
|
|
|
|
|
class DOMHTMLRatingsParser(DOMParserBase):
|
|
"""Parser for the "user ratings" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
rparser = DOMHTMLRatingsParser()
|
|
result = rparser.parse(userratings_html_string)
|
|
"""
|
|
re_means = re.compile('mean\s*=\s*([0-9]\.[0-9])\.\s*median\s*=\s*([0-9])',
|
|
re.I)
|
|
extractors = [
|
|
Extractor(label='number of votes',
|
|
path="//td[b='Percentage']/../../tr",
|
|
attrs=[Attribute(key='votes',
|
|
multi=True,
|
|
path={
|
|
'votes': "td[1]//text()",
|
|
'ordinal': "td[3]//text()"
|
|
})]),
|
|
Extractor(label='mean and median',
|
|
path="//p[starts-with(text(), 'Arithmetic mean')]",
|
|
attrs=Attribute(key='mean and median',
|
|
path="text()")),
|
|
Extractor(label='rating',
|
|
path="//a[starts-with(@href, '/search/title?user_rating=')]",
|
|
attrs=Attribute(key='rating',
|
|
path="text()")),
|
|
Extractor(label='demographic voters',
|
|
path="//td[b='Average']/../../tr",
|
|
attrs=Attribute(key='demographic voters',
|
|
multi=True,
|
|
path={
|
|
'voters': "td[1]//text()",
|
|
'votes': "td[2]//text()",
|
|
'average': "td[3]//text()"
|
|
})),
|
|
Extractor(label='top 250',
|
|
path="//a[text()='top 250']",
|
|
attrs=Attribute(key='top 250',
|
|
path="./preceding-sibling::text()[1]"))
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
nd = {}
|
|
votes = data.get('votes', [])
|
|
if votes:
|
|
nd['number of votes'] = {}
|
|
for i in xrange(1, 11):
|
|
_ordinal = int(votes[i]['ordinal'])
|
|
_strvts = votes[i]['votes'] or '0'
|
|
nd['number of votes'][_ordinal] = \
|
|
int(_strvts.replace(',', ''))
|
|
mean = data.get('mean and median', '')
|
|
if mean:
|
|
means = self.re_means.findall(mean)
|
|
if means and len(means[0]) == 2:
|
|
am, med = means[0]
|
|
try: am = float(am)
|
|
except (ValueError, OverflowError): pass
|
|
if type(am) is type(1.0):
|
|
nd['arithmetic mean'] = am
|
|
try: med = int(med)
|
|
except (ValueError, OverflowError): pass
|
|
if type(med) is type(0):
|
|
nd['median'] = med
|
|
if 'rating' in data:
|
|
nd['rating'] = float(data['rating'])
|
|
dem_voters = data.get('demographic voters')
|
|
if dem_voters:
|
|
nd['demographic'] = {}
|
|
for i in xrange(1, len(dem_voters)):
|
|
if (dem_voters[i]['votes'] is not None) \
|
|
and (dem_voters[i]['votes'].strip()):
|
|
nd['demographic'][dem_voters[i]['voters'].strip().lower()] \
|
|
= (int(dem_voters[i]['votes'].replace(',', '')),
|
|
float(dem_voters[i]['average']))
|
|
if 'imdb users' in nd.get('demographic', {}):
|
|
nd['votes'] = nd['demographic']['imdb users'][0]
|
|
nd['demographic']['all votes'] = nd['demographic']['imdb users']
|
|
del nd['demographic']['imdb users']
|
|
top250 = data.get('top 250')
|
|
if top250:
|
|
sd = top250[9:]
|
|
i = sd.find(' ')
|
|
if i != -1:
|
|
sd = sd[:i]
|
|
try: sd = int(sd)
|
|
except (ValueError, OverflowError): pass
|
|
if type(sd) is type(0):
|
|
nd['top 250 rank'] = sd
|
|
return nd
|
|
|
|
|
|
class DOMHTMLEpisodesRatings(DOMParserBase):
|
|
"""Parser for the "episode ratings ... by date" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
erparser = DOMHTMLEpisodesRatings()
|
|
result = erparser.parse(eprating_html_string)
|
|
"""
|
|
_containsObjects = True
|
|
|
|
extractors = [Extractor(label='title', path="//title",
|
|
attrs=Attribute(key='title', path="./text()")),
|
|
Extractor(label='ep ratings',
|
|
path="//th/../..//tr",
|
|
attrs=Attribute(key='episodes', multi=True,
|
|
path={'nr': ".//td[1]/text()",
|
|
'ep title': ".//td[2]//text()",
|
|
'movieID': ".//td[2]/a/@href",
|
|
'rating': ".//td[3]/text()",
|
|
'votes': ".//td[4]/text()"}))]
|
|
|
|
def postprocess_data(self, data):
|
|
if 'title' not in data or 'episodes' not in data: return {}
|
|
nd = []
|
|
title = data['title']
|
|
for i in data['episodes']:
|
|
ept = i['ep title']
|
|
movieID = analyze_imdbid(i['movieID'])
|
|
votes = i['votes']
|
|
rating = i['rating']
|
|
if not (ept and movieID and votes and rating): continue
|
|
try:
|
|
votes = int(votes.replace(',', '').replace('.', ''))
|
|
except:
|
|
pass
|
|
try:
|
|
rating = float(rating)
|
|
except:
|
|
pass
|
|
ept = ept.strip()
|
|
ept = u'%s {%s' % (title, ept)
|
|
nr = i['nr']
|
|
if nr:
|
|
ept += u' (#%s)' % nr.strip()
|
|
ept += '}'
|
|
if movieID is not None:
|
|
movieID = str(movieID)
|
|
m = Movie(title=ept, movieID=movieID, accessSystem=self._as,
|
|
modFunct=self._modFunct)
|
|
epofdict = m.get('episode of')
|
|
if epofdict is not None:
|
|
m['episode of'] = Movie(data=epofdict, accessSystem=self._as,
|
|
modFunct=self._modFunct)
|
|
nd.append({'episode': m, 'votes': votes, 'rating': rating})
|
|
return {'episodes rating': nd}
|
|
|
|
|
|
def _normalize_href(href):
|
|
if (href is not None) and (not href.lower().startswith('http://')):
|
|
if href.startswith('/'): href = href[1:]
|
|
# TODO: imdbURL_base may be set by the user!
|
|
href = '%s%s' % (imdbURL_base, href)
|
|
return href
|
|
|
|
|
|
class DOMHTMLOfficialsitesParser(DOMParserBase):
|
|
"""Parser for the "official sites", "external reviews", "newsgroup
|
|
reviews", "miscellaneous links", "sound clips", "video clips" and
|
|
"photographs" pages of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
osparser = DOMHTMLOfficialsitesParser()
|
|
result = osparser.parse(officialsites_html_string)
|
|
"""
|
|
kind = 'official sites'
|
|
|
|
extractors = [
|
|
Extractor(label='site',
|
|
path="//ol/li/a",
|
|
attrs=Attribute(key='self.kind',
|
|
multi=True,
|
|
path={
|
|
'link': "./@href",
|
|
'info': "./text()"
|
|
},
|
|
postprocess=lambda x: (x.get('info').strip(),
|
|
urllib.unquote(_normalize_href(x.get('link'))))))
|
|
]
|
|
|
|
|
|
class DOMHTMLConnectionParser(DOMParserBase):
|
|
"""Parser for the "connections" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
connparser = DOMHTMLConnectionParser()
|
|
result = connparser.parse(connections_html_string)
|
|
"""
|
|
_containsObjects = True
|
|
|
|
extractors = [Extractor(label='connection',
|
|
group="//div[@class='_imdbpy']",
|
|
group_key="./h5/text()",
|
|
group_key_normalize=lambda x: x.lower(),
|
|
path="./a",
|
|
attrs=Attribute(key=None,
|
|
path={'title': "./text()",
|
|
'movieID': "./@href"},
|
|
multi=True))]
|
|
|
|
preprocessors = [
|
|
('<h5>', '</div><div class="_imdbpy"><h5>'),
|
|
# To get the movie's year.
|
|
('</a> (', ' ('),
|
|
('\n<br/>', '</a>'),
|
|
('<br/> - ', '::')
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
for key in data.keys():
|
|
nl = []
|
|
for v in data[key]:
|
|
title = v['title']
|
|
ts = title.split('::', 1)
|
|
title = ts[0].strip()
|
|
notes = u''
|
|
if len(ts) == 2:
|
|
notes = ts[1].strip()
|
|
m = Movie(title=title,
|
|
movieID=analyze_imdbid(v['movieID']),
|
|
accessSystem=self._as, notes=notes,
|
|
modFunct=self._modFunct)
|
|
nl.append(m)
|
|
data[key] = nl
|
|
if not data: return {}
|
|
return {'connections': data}
|
|
|
|
|
|
class DOMHTMLLocationsParser(DOMParserBase):
|
|
"""Parser for the "locations" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
lparser = DOMHTMLLocationsParser()
|
|
result = lparser.parse(locations_html_string)
|
|
"""
|
|
extractors = [Extractor(label='locations', path="//dt",
|
|
attrs=Attribute(key='locations', multi=True,
|
|
path={'place': ".//text()",
|
|
'note': "./following-sibling::dd[1]" \
|
|
"//text()"},
|
|
postprocess=lambda x: (u'%s::%s' % (
|
|
x['place'].strip(),
|
|
(x['note'] or u'').strip())).strip(':')))]
|
|
|
|
|
|
class DOMHTMLTechParser(DOMParserBase):
|
|
"""Parser for the "technical", "business", "literature",
|
|
"publicity" (for people) and "contacts (for people) pages of
|
|
a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
tparser = HTMLTechParser()
|
|
result = tparser.parse(technical_html_string)
|
|
"""
|
|
kind = 'tech'
|
|
|
|
extractors = [Extractor(label='tech',
|
|
group="//h5",
|
|
group_key="./text()",
|
|
group_key_normalize=lambda x: x.lower(),
|
|
path="./following-sibling::div[1]",
|
|
attrs=Attribute(key=None,
|
|
path=".//text()",
|
|
postprocess=lambda x: [t.strip()
|
|
for t in x.split('\n') if t.strip()]))]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<h5>.*?</h5>)', re.I), r'</div>\1<div class="_imdbpy">'),
|
|
(re.compile('((<br/>|</p>|</table>))\n?<br/>(?!<a)', re.I),
|
|
r'\1</div>'),
|
|
# the ones below are for the publicity parser
|
|
(re.compile('<p>(.*?)</p>', re.I), r'\1<br/>'),
|
|
(re.compile('(</td><td valign="top">)', re.I), r'\1::'),
|
|
(re.compile('(</tr><tr>)', re.I), r'\n\1'),
|
|
# this is for splitting individual entries
|
|
(re.compile('<br/>', re.I), r'\n'),
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
for key in data:
|
|
data[key] = filter(None, data[key])
|
|
if self.kind in ('literature', 'business', 'contacts') and data:
|
|
if 'screenplay/teleplay' in data:
|
|
data['screenplay-teleplay'] = data['screenplay/teleplay']
|
|
del data['screenplay/teleplay']
|
|
data = {self.kind: data}
|
|
else:
|
|
if self.kind == 'publicity':
|
|
if 'biography (print)' in data:
|
|
data['biography-print'] = data['biography (print)']
|
|
del data['biography (print)']
|
|
# Tech info.
|
|
for key in data.keys():
|
|
if key.startswith('film negative format'):
|
|
data['film negative format'] = data[key]
|
|
del data[key]
|
|
elif key.startswith('film length'):
|
|
data['film length'] = data[key]
|
|
del data[key]
|
|
return data
|
|
|
|
|
|
class DOMHTMLRecParser(DOMParserBase):
|
|
"""Parser for the "recommendations" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
rparser = HTMLRecParser()
|
|
result = rparser.parse(recommendations_html_string)
|
|
"""
|
|
_containsObjects = True
|
|
|
|
extractors = [Extractor(label='recommendations',
|
|
path="//td[@valign='middle'][1]",
|
|
attrs=Attribute(key='../../tr/td[1]//text()',
|
|
multi=True,
|
|
path={'title': ".//text()",
|
|
'movieID': ".//a/@href"}))]
|
|
def postprocess_data(self, data):
|
|
for key in data.keys():
|
|
n_key = key
|
|
n_keyl = n_key.lower()
|
|
if n_keyl == 'suggested by the database':
|
|
n_key = 'database'
|
|
elif n_keyl == 'imdb users recommend':
|
|
n_key = 'users'
|
|
data[n_key] = [Movie(title=x['title'],
|
|
movieID=analyze_imdbid(x['movieID']),
|
|
accessSystem=self._as, modFunct=self._modFunct)
|
|
for x in data[key]]
|
|
del data[key]
|
|
if data: return {'recommendations': data}
|
|
return data
|
|
|
|
|
|
class DOMHTMLNewsParser(DOMParserBase):
|
|
"""Parser for the "news" page of a given movie or person.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
nwparser = DOMHTMLNewsParser()
|
|
result = nwparser.parse(news_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
extractors = [
|
|
Extractor(label='news',
|
|
path="//h2",
|
|
attrs=Attribute(key='news',
|
|
multi=True,
|
|
path={
|
|
'title': "./text()",
|
|
'fromdate': "../following-sibling::p[1]/small//text()",
|
|
# FIXME: sometimes (see The Matrix (1999)) <p> is found
|
|
# inside news text.
|
|
'body': "../following-sibling::p[2]//text()",
|
|
'link': "../..//a[text()='Permalink']/@href",
|
|
'fulllink': "../..//a[starts-with(text(), " \
|
|
"'See full article at')]/@href"
|
|
},
|
|
postprocess=lambda x: {
|
|
'title': x.get('title').strip(),
|
|
'date': x.get('fromdate').split('|')[0].strip(),
|
|
'from': x.get('fromdate').split('|')[1].replace('From ',
|
|
'').strip(),
|
|
'body': (x.get('body') or u'').strip(),
|
|
'link': _normalize_href(x.get('link')),
|
|
'full article link': _normalize_href(x.get('fulllink'))
|
|
}))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<a name=[^>]+><h2>)', re.I), r'<div class="_imdbpy">\1'),
|
|
(re.compile('(<hr/>)', re.I), r'</div>\1'),
|
|
(re.compile('<p></p>', re.I), r'')
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
if not data.has_key('news'):
|
|
return {}
|
|
for news in data['news']:
|
|
if news.has_key('full article link'):
|
|
if news['full article link'] is None:
|
|
del news['full article link']
|
|
return data
|
|
|
|
|
|
def _parse_review(x):
|
|
result = {}
|
|
title = x.get('title').strip()
|
|
if title[-1] == ':': title = title[:-1]
|
|
result['title'] = title
|
|
result['link'] = _normalize_href(x.get('link'))
|
|
kind = x.get('kind').strip()
|
|
if kind[-1] == ':': kind = kind[:-1]
|
|
result['review kind'] = kind
|
|
text = x.get('review').replace('\n\n', '||').replace('\n', ' ').split('||')
|
|
review = '\n'.join(text)
|
|
if x.get('author') is not None:
|
|
author = x.get('author').strip()
|
|
review = review.split(author)[0].strip()
|
|
result['review author'] = author[2:]
|
|
if x.get('item') is not None:
|
|
item = x.get('item').strip()
|
|
review = review[len(item):].strip()
|
|
review = "%s: %s" % (item, review)
|
|
result['review'] = review
|
|
return result
|
|
|
|
|
|
class DOMHTMLSeasonEpisodesParser(DOMParserBase):
|
|
"""Parser for the "episode list" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
sparser = DOMHTMLSeasonEpisodesParser()
|
|
result = sparser.parse(episodes_html_string)
|
|
"""
|
|
extractors = [
|
|
Extractor(label='series link',
|
|
path="//div[@class='parent']",
|
|
attrs=[Attribute(key='series link',
|
|
path=".//a/@href")]
|
|
),
|
|
|
|
Extractor(label='series title',
|
|
path="//head/meta[@property='og:title']",
|
|
attrs=[Attribute(key='series title',
|
|
path="./@content")]
|
|
),
|
|
|
|
Extractor(label='seasons list',
|
|
path="//select[@id='bySeason']//option",
|
|
attrs=[Attribute(key='_seasons',
|
|
multi=True,
|
|
path="./@value")]),
|
|
|
|
Extractor(label='selected season',
|
|
path="//select[@id='bySeason']//option[@selected]",
|
|
attrs=[Attribute(key='_current_season',
|
|
path='./@value')]),
|
|
|
|
Extractor(label='episodes',
|
|
path=".",
|
|
group="//div[@class='info']",
|
|
group_key=".//meta/@content",
|
|
group_key_normalize=lambda x: 'episode %s' % x,
|
|
attrs=[Attribute(key=None,
|
|
multi=True,
|
|
path={
|
|
"link": ".//strong//a[@href][1]/@href",
|
|
"original air date": ".//div[@class='airdate']/text()",
|
|
"title": ".//strong//text()",
|
|
"plot": ".//div[@class='item_description']//text()"
|
|
}
|
|
)]
|
|
)
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
series_id = analyze_imdbid(data.get('series link'))
|
|
series_title = data.get('series title', '').strip()
|
|
selected_season = data.get('_current_season',
|
|
'unknown season').strip()
|
|
if not (series_id and series_title):
|
|
return {}
|
|
series = Movie(title=series_title, movieID=str(series_id),
|
|
accessSystem=self._as, modFunct=self._modFunct)
|
|
if series.get('kind') == 'movie':
|
|
series['kind'] = u'tv series'
|
|
try: selected_season = int(selected_season)
|
|
except: pass
|
|
nd = {selected_season: {}}
|
|
for episode_nr, episode in data.iteritems():
|
|
if not (episode and episode[0] and
|
|
episode_nr.startswith('episode ')):
|
|
continue
|
|
episode = episode[0]
|
|
episode_nr = episode_nr[8:].rstrip()
|
|
try: episode_nr = int(episode_nr)
|
|
except: pass
|
|
episode_id = analyze_imdbid(episode.get('link' ''))
|
|
episode_air_date = episode.get('original air date',
|
|
'').strip()
|
|
episode_title = episode.get('title', '').strip()
|
|
episode_plot = episode.get('plot', '')
|
|
if not (episode_nr and episode_id and episode_title):
|
|
continue
|
|
ep_obj = Movie(movieID=episode_id, title=episode_title,
|
|
accessSystem=self._as, modFunct=self._modFunct)
|
|
ep_obj['kind'] = u'episode'
|
|
ep_obj['episode of'] = series
|
|
ep_obj['season'] = selected_season
|
|
ep_obj['episode'] = episode_nr
|
|
if episode_air_date:
|
|
ep_obj['original air date'] = episode_air_date
|
|
if episode_air_date[-4:].isdigit():
|
|
ep_obj['year'] = episode_air_date[-4:]
|
|
if episode_plot:
|
|
ep_obj['plot'] = episode_plot
|
|
nd[selected_season][episode_nr] = ep_obj
|
|
_seasons = data.get('_seasons') or []
|
|
for idx, season in enumerate(_seasons):
|
|
try: _seasons[idx] = int(season)
|
|
except: pass
|
|
return {'episodes': nd, '_seasons': _seasons,
|
|
'_current_season': selected_season}
|
|
|
|
|
|
def _build_episode(x):
|
|
"""Create a Movie object for a given series' episode."""
|
|
episode_id = analyze_imdbid(x.get('link'))
|
|
episode_title = x.get('title')
|
|
e = Movie(movieID=episode_id, title=episode_title)
|
|
e['kind'] = u'episode'
|
|
oad = x.get('oad')
|
|
if oad:
|
|
e['original air date'] = oad.strip()
|
|
year = x.get('year')
|
|
if year is not None:
|
|
year = year[5:]
|
|
if year == 'unknown': year = u'????'
|
|
if year and year.isdigit():
|
|
year = int(year)
|
|
e['year'] = year
|
|
else:
|
|
if oad and oad[-4:].isdigit():
|
|
e['year'] = int(oad[-4:])
|
|
epinfo = x.get('episode')
|
|
if epinfo is not None:
|
|
season, episode = epinfo.split(':')[0].split(',')
|
|
e['season'] = int(season[7:])
|
|
e['episode'] = int(episode[8:])
|
|
else:
|
|
e['season'] = 'unknown'
|
|
e['episode'] = 'unknown'
|
|
plot = x.get('plot')
|
|
if plot:
|
|
e['plot'] = plot.strip()
|
|
return e
|
|
|
|
|
|
class DOMHTMLEpisodesParser(DOMParserBase):
|
|
"""Parser for the "episode list" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
eparser = DOMHTMLEpisodesParser()
|
|
result = eparser.parse(episodes_html_string)
|
|
"""
|
|
# XXX: no more used for the list of episodes parser,
|
|
# but only for the episodes cast parser (see below).
|
|
_containsObjects = True
|
|
|
|
kind = 'episodes list'
|
|
_episodes_path = "..//h4"
|
|
_oad_path = "./following-sibling::span/strong[1]/text()"
|
|
|
|
def _init(self):
|
|
self.extractors = [
|
|
Extractor(label='series',
|
|
path="//html",
|
|
attrs=[Attribute(key='series title',
|
|
path=".//title/text()"),
|
|
Attribute(key='series movieID',
|
|
path=".//h1/a[@class='main']/@href",
|
|
postprocess=analyze_imdbid)
|
|
]),
|
|
Extractor(label='episodes',
|
|
group="//div[@class='_imdbpy']/h3",
|
|
group_key="./a/@name",
|
|
path=self._episodes_path,
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={
|
|
'link': "./a/@href",
|
|
'title': "./a/text()",
|
|
'year': "./preceding-sibling::a[1]/@name",
|
|
'episode': "./text()[1]",
|
|
'oad': self._oad_path,
|
|
'plot': "./following-sibling::text()[1]"
|
|
},
|
|
postprocess=_build_episode))]
|
|
if self.kind == 'episodes cast':
|
|
self.extractors += [
|
|
Extractor(label='cast',
|
|
group="//h4",
|
|
group_key="./text()[1]",
|
|
group_key_normalize=lambda x: x.strip(),
|
|
path="./following-sibling::table[1]//td[@class='nm']",
|
|
attrs=Attribute(key=None,
|
|
multi=True,
|
|
path={'person': "..//text()",
|
|
'link': "./a/@href",
|
|
'roleID': \
|
|
"../td[4]/div[@class='_imdbpyrole']/@roleid"},
|
|
postprocess=lambda x: \
|
|
build_person(x.get('person') or u'',
|
|
personID=analyze_imdbid(x.get('link')),
|
|
roleID=(x.get('roleID') or u'').split('/'),
|
|
accessSystem=self._as,
|
|
modFunct=self._modFunct)))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('(<hr/>\n)(<h3>)', re.I),
|
|
r'</div>\1<div class="_imdbpy">\2'),
|
|
(re.compile('(</p>\n\n)</div>', re.I), r'\1'),
|
|
(re.compile('<h3>(.*?)</h3>', re.I), r'<h4>\1</h4>'),
|
|
(_reRolesMovie, _manageRoles),
|
|
(re.compile('(<br/> <br/>\n)(<hr/>)', re.I), r'\1</div>\2')
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
# A bit extreme?
|
|
if not 'series title' in data: return {}
|
|
if not 'series movieID' in data: return {}
|
|
stitle = data['series title'].replace('- Episode list', '')
|
|
stitle = stitle.replace('- Episodes list', '')
|
|
stitle = stitle.replace('- Episode cast', '')
|
|
stitle = stitle.replace('- Episodes cast', '')
|
|
stitle = stitle.strip()
|
|
if not stitle: return {}
|
|
seriesID = data['series movieID']
|
|
if seriesID is None: return {}
|
|
series = Movie(title=stitle, movieID=str(seriesID),
|
|
accessSystem=self._as, modFunct=self._modFunct)
|
|
nd = {}
|
|
for key in data.keys():
|
|
if key.startswith('filter-season-') or key.startswith('season-'):
|
|
season_key = key.replace('filter-season-', '').replace('season-', '')
|
|
try: season_key = int(season_key)
|
|
except: pass
|
|
nd[season_key] = {}
|
|
ep_counter = 1
|
|
for episode in data[key]:
|
|
if not episode: continue
|
|
episode_key = episode.get('episode')
|
|
if episode_key is None: continue
|
|
if not isinstance(episode_key, int):
|
|
episode_key = ep_counter
|
|
ep_counter += 1
|
|
cast_key = 'Season %s, Episode %s:' % (season_key,
|
|
episode_key)
|
|
if data.has_key(cast_key):
|
|
cast = data[cast_key]
|
|
for i in xrange(len(cast)):
|
|
cast[i].billingPos = i + 1
|
|
episode['cast'] = cast
|
|
episode['episode of'] = series
|
|
nd[season_key][episode_key] = episode
|
|
if len(nd) == 0:
|
|
return {}
|
|
return {'episodes': nd}
|
|
|
|
|
|
class DOMHTMLEpisodesCastParser(DOMHTMLEpisodesParser):
|
|
"""Parser for the "episodes cast" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
eparser = DOMHTMLEpisodesParser()
|
|
result = eparser.parse(episodes_html_string)
|
|
"""
|
|
kind = 'episodes cast'
|
|
_episodes_path = "..//h4"
|
|
_oad_path = "./following-sibling::b[1]/text()"
|
|
|
|
|
|
class DOMHTMLFaqsParser(DOMParserBase):
|
|
"""Parser for the "FAQ" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
fparser = DOMHTMLFaqsParser()
|
|
result = fparser.parse(faqs_html_string)
|
|
"""
|
|
_defGetRefs = True
|
|
|
|
# XXX: bsoup and lxml don't match (looks like a minor issue, anyway).
|
|
|
|
extractors = [
|
|
Extractor(label='faqs',
|
|
path="//div[@class='section']",
|
|
attrs=Attribute(key='faqs',
|
|
multi=True,
|
|
path={
|
|
'question': "./h3/a/span/text()",
|
|
'answer': "../following-sibling::div[1]//text()"
|
|
},
|
|
postprocess=lambda x: u'%s::%s' % (x.get('question').strip(),
|
|
'\n\n'.join(x.get('answer').replace(
|
|
'\n\n', '\n').strip().split('||')))))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('<br/><br/>', re.I), r'||'),
|
|
(re.compile('<h4>(.*?)</h4>\n', re.I), r'||\1--'),
|
|
(re.compile('<span class="spoiler"><span>(.*?)</span></span>', re.I),
|
|
r'[spoiler]\1[/spoiler]')
|
|
]
|
|
|
|
|
|
class DOMHTMLAiringParser(DOMParserBase):
|
|
"""Parser for the "airing" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
aparser = DOMHTMLAiringParser()
|
|
result = aparser.parse(airing_html_string)
|
|
"""
|
|
_containsObjects = True
|
|
|
|
extractors = [
|
|
Extractor(label='series title',
|
|
path="//title",
|
|
attrs=Attribute(key='series title', path="./text()",
|
|
postprocess=lambda x: \
|
|
x.replace(' - TV schedule', u''))),
|
|
Extractor(label='series id',
|
|
path="//h1/a[@href]",
|
|
attrs=Attribute(key='series id', path="./@href")),
|
|
|
|
Extractor(label='tv airings',
|
|
path="//tr[@class]",
|
|
attrs=Attribute(key='airing',
|
|
multi=True,
|
|
path={
|
|
'date': "./td[1]//text()",
|
|
'time': "./td[2]//text()",
|
|
'channel': "./td[3]//text()",
|
|
'link': "./td[4]/a[1]/@href",
|
|
'title': "./td[4]//text()",
|
|
'season': "./td[5]//text()",
|
|
},
|
|
postprocess=lambda x: {
|
|
'date': x.get('date'),
|
|
'time': x.get('time'),
|
|
'channel': x.get('channel').strip(),
|
|
'link': x.get('link'),
|
|
'title': x.get('title'),
|
|
'season': (x.get('season') or '').strip()
|
|
}
|
|
))
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
if len(data) == 0:
|
|
return {}
|
|
seriesTitle = data['series title']
|
|
seriesID = analyze_imdbid(data['series id'])
|
|
if data.has_key('airing'):
|
|
for airing in data['airing']:
|
|
title = airing.get('title', '').strip()
|
|
if not title:
|
|
epsTitle = seriesTitle
|
|
if seriesID is None:
|
|
continue
|
|
epsID = seriesID
|
|
else:
|
|
epsTitle = '%s {%s}' % (data['series title'],
|
|
airing['title'])
|
|
epsID = analyze_imdbid(airing['link'])
|
|
e = Movie(title=epsTitle, movieID=epsID)
|
|
airing['episode'] = e
|
|
del airing['link']
|
|
del airing['title']
|
|
if not airing['season']:
|
|
del airing['season']
|
|
if 'series title' in data:
|
|
del data['series title']
|
|
if 'series id' in data:
|
|
del data['series id']
|
|
if 'airing' in data:
|
|
data['airing'] = filter(None, data['airing'])
|
|
if 'airing' not in data or not data['airing']:
|
|
return {}
|
|
return data
|
|
|
|
|
|
class DOMHTMLSynopsisParser(DOMParserBase):
|
|
"""Parser for the "synopsis" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
sparser = HTMLSynopsisParser()
|
|
result = sparser.parse(synopsis_html_string)
|
|
"""
|
|
extractors = [
|
|
Extractor(label='synopsis',
|
|
path="//div[@class='display'][not(@style)]",
|
|
attrs=Attribute(key='synopsis',
|
|
path=".//text()",
|
|
postprocess=lambda x: '\n\n'.join(x.strip().split('||'))))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('<br/><br/>', re.I), r'||')
|
|
]
|
|
|
|
|
|
class DOMHTMLParentsGuideParser(DOMParserBase):
|
|
"""Parser for the "parents guide" page of a given movie.
|
|
The page should be provided as a string, as taken from
|
|
the akas.imdb.com server. The final result will be a
|
|
dictionary, with a key for every relevant section.
|
|
|
|
Example:
|
|
pgparser = HTMLParentsGuideParser()
|
|
result = pgparser.parse(parentsguide_html_string)
|
|
"""
|
|
extractors = [
|
|
Extractor(label='parents guide',
|
|
group="//div[@class='section']",
|
|
group_key="./h3/a/span/text()",
|
|
group_key_normalize=lambda x: x.lower(),
|
|
path="../following-sibling::div[1]/p",
|
|
attrs=Attribute(key=None,
|
|
path=".//text()",
|
|
postprocess=lambda x: [t.strip().replace('\n', ' ')
|
|
for t in x.split('||') if t.strip()]))
|
|
]
|
|
|
|
preprocessors = [
|
|
(re.compile('<br/><br/>', re.I), r'||')
|
|
]
|
|
|
|
def postprocess_data(self, data):
|
|
data2 = {}
|
|
for key in data:
|
|
if data[key]:
|
|
data2[key] = data[key]
|
|
if not data2:
|
|
return {}
|
|
return {'parents guide': data2}
|
|
|
|
|
|
_OBJECTS = {
|
|
'movie_parser': ((DOMHTMLMovieParser,), None),
|
|
'plot_parser': ((DOMHTMLPlotParser,), None),
|
|
'movie_awards_parser': ((DOMHTMLAwardsParser,), None),
|
|
'taglines_parser': ((DOMHTMLTaglinesParser,), None),
|
|
'keywords_parser': ((DOMHTMLKeywordsParser,), None),
|
|
'crazycredits_parser': ((DOMHTMLCrazyCreditsParser,), None),
|
|
'goofs_parser': ((DOMHTMLGoofsParser,), None),
|
|
'alternateversions_parser': ((DOMHTMLAlternateVersionsParser,), None),
|
|
'trivia_parser': ((DOMHTMLTriviaParser,), None),
|
|
'soundtrack_parser': ((DOMHTMLSoundtrackParser,), {'kind': 'soundtrack'}),
|
|
'quotes_parser': ((DOMHTMLQuotesParser,), None),
|
|
'releasedates_parser': ((DOMHTMLReleaseinfoParser,), None),
|
|
'ratings_parser': ((DOMHTMLRatingsParser,), None),
|
|
'officialsites_parser': ((DOMHTMLOfficialsitesParser,), None),
|
|
'externalrev_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'external reviews'}),
|
|
'newsgrouprev_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'newsgroup reviews'}),
|
|
'misclinks_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'misc links'}),
|
|
'soundclips_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'sound clips'}),
|
|
'videoclips_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'video clips'}),
|
|
'photosites_parser': ((DOMHTMLOfficialsitesParser,),
|
|
{'kind': 'photo sites'}),
|
|
'connections_parser': ((DOMHTMLConnectionParser,), None),
|
|
'tech_parser': ((DOMHTMLTechParser,), None),
|
|
'business_parser': ((DOMHTMLTechParser,),
|
|
{'kind': 'business', '_defGetRefs': 1}),
|
|
'literature_parser': ((DOMHTMLTechParser,), {'kind': 'literature'}),
|
|
'locations_parser': ((DOMHTMLLocationsParser,), None),
|
|
'rec_parser': ((DOMHTMLRecParser,), None),
|
|
'news_parser': ((DOMHTMLNewsParser,), None),
|
|
'episodes_parser': ((DOMHTMLEpisodesParser,), None),
|
|
'season_episodes_parser': ((DOMHTMLSeasonEpisodesParser,), None),
|
|
'episodes_cast_parser': ((DOMHTMLEpisodesCastParser,), None),
|
|
'eprating_parser': ((DOMHTMLEpisodesRatings,), None),
|
|
'movie_faqs_parser': ((DOMHTMLFaqsParser,), None),
|
|
'airing_parser': ((DOMHTMLAiringParser,), None),
|
|
'synopsis_parser': ((DOMHTMLSynopsisParser,), None),
|
|
'parentsguide_parser': ((DOMHTMLParentsGuideParser,), None)
|
|
}
|
|
|