mirror of
https://github.com/moparisthebest/SickRage
synced 2024-10-31 23:45:02 -04:00
0d9fbc1ad7
This version of SickBeard uses both TVDB and TVRage to search and gather it's series data from allowing you to now have access to and download shows that you couldn't before because of being locked into only what TheTVDB had to offer. Also this edition is based off the code we used in our XEM editon so it does come with scene numbering support as well as all the other features our XEM edition has to offer. Please before using this with your existing database (sickbeard.db) please make a backup copy of it and delete any other database files such as cache.db and failed.db if present, we HIGHLY recommend starting out with no database files at all to make this a fresh start but the choice is at your own risk! Enjoy!
290 lines
11 KiB
Python
290 lines
11 KiB
Python
from lib.hachoir_metadata.metadata import RootMetadata, registerExtractor
|
|
from lib.hachoir_metadata.image import computeComprRate
|
|
from lib.hachoir_parser.image.exif import ExifEntry
|
|
from lib.hachoir_parser.image.jpeg import (
|
|
JpegFile, JpegChunk,
|
|
QUALITY_HASH_COLOR, QUALITY_SUM_COLOR,
|
|
QUALITY_HASH_GRAY, QUALITY_SUM_GRAY)
|
|
from lib.hachoir_core.field import MissingField
|
|
from lib.hachoir_core.i18n import _
|
|
from lib.hachoir_core.tools import makeUnicode
|
|
from lib.hachoir_metadata.safe import fault_tolerant
|
|
from datetime import datetime
|
|
|
|
def deg2float(degree, minute, second):
|
|
return degree + (float(minute) + float(second) / 60.0) / 60.0
|
|
|
|
class JpegMetadata(RootMetadata):
|
|
EXIF_KEY = {
|
|
# Exif metadatas
|
|
ExifEntry.TAG_CAMERA_MANUFACTURER: "camera_manufacturer",
|
|
ExifEntry.TAG_CAMERA_MODEL: "camera_model",
|
|
ExifEntry.TAG_ORIENTATION: "image_orientation",
|
|
ExifEntry.TAG_EXPOSURE: "camera_exposure",
|
|
ExifEntry.TAG_FOCAL: "camera_focal",
|
|
ExifEntry.TAG_BRIGHTNESS: "camera_brightness",
|
|
ExifEntry.TAG_APERTURE: "camera_aperture",
|
|
|
|
# Generic metadatas
|
|
ExifEntry.TAG_IMG_TITLE: "title",
|
|
ExifEntry.TAG_SOFTWARE: "producer",
|
|
ExifEntry.TAG_FILE_TIMESTAMP: "creation_date",
|
|
ExifEntry.TAG_WIDTH: "width",
|
|
ExifEntry.TAG_HEIGHT: "height",
|
|
ExifEntry.TAG_USER_COMMENT: "comment",
|
|
}
|
|
|
|
IPTC_KEY = {
|
|
80: "author",
|
|
90: "city",
|
|
101: "country",
|
|
116: "copyright",
|
|
120: "title",
|
|
231: "comment",
|
|
}
|
|
|
|
orientation_name = {
|
|
1: _('Horizontal (normal)'),
|
|
2: _('Mirrored horizontal'),
|
|
3: _('Rotated 180'),
|
|
4: _('Mirrored vertical'),
|
|
5: _('Mirrored horizontal then rotated 90 counter-clock-wise'),
|
|
6: _('Rotated 90 clock-wise'),
|
|
7: _('Mirrored horizontal then rotated 90 clock-wise'),
|
|
8: _('Rotated 90 counter clock-wise'),
|
|
}
|
|
|
|
def extract(self, jpeg):
|
|
if "start_frame/content" in jpeg:
|
|
self.startOfFrame(jpeg["start_frame/content"])
|
|
elif "start_scan/content/nr_components" in jpeg:
|
|
self.bits_per_pixel = 8 * jpeg["start_scan/content/nr_components"].value
|
|
if "app0/content" in jpeg:
|
|
self.extractAPP0(jpeg["app0/content"])
|
|
|
|
if "exif/content" in jpeg:
|
|
for ifd in jpeg.array("exif/content/ifd"):
|
|
for entry in ifd.array("entry"):
|
|
self.processIfdEntry(ifd, entry)
|
|
self.readGPS(ifd)
|
|
if "photoshop/content" in jpeg:
|
|
psd = jpeg["photoshop/content"]
|
|
if "version/content/reader_name" in psd:
|
|
self.producer = psd["version/content/reader_name"].value
|
|
if "iptc/content" in psd:
|
|
self.parseIPTC(psd["iptc/content"])
|
|
for field in jpeg.array("comment"):
|
|
if "content/comment" in field:
|
|
self.comment = field["content/comment"].value
|
|
self.computeQuality(jpeg)
|
|
if "data" in jpeg:
|
|
computeComprRate(self, jpeg["data"].size)
|
|
if not self.has("producer") and "photoshop" in jpeg:
|
|
self.producer = u"Adobe Photoshop"
|
|
if self.has("compression"):
|
|
self.compression = "JPEG"
|
|
|
|
@fault_tolerant
|
|
def startOfFrame(self, sof):
|
|
# Set compression method
|
|
key = sof["../type"].value
|
|
self.compression = "JPEG (%s)" % JpegChunk.START_OF_FRAME[key]
|
|
|
|
# Read image size and bits/pixel
|
|
self.width = sof["width"].value
|
|
self.height = sof["height"].value
|
|
nb_components = sof["nr_components"].value
|
|
self.bits_per_pixel = 8 * nb_components
|
|
if nb_components == 3:
|
|
self.pixel_format = _("YCbCr")
|
|
elif nb_components == 1:
|
|
self.pixel_format = _("Grayscale")
|
|
self.nb_colors = 256
|
|
|
|
@fault_tolerant
|
|
def computeQuality(self, jpeg):
|
|
# This function is an adaption to Python of ImageMagick code
|
|
# to compute JPEG quality using quantization tables
|
|
|
|
# Read quantization tables
|
|
qtlist = []
|
|
for dqt in jpeg.array("quantization"):
|
|
for qt in dqt.array("content/qt"):
|
|
# TODO: Take care of qt["index"].value?
|
|
qtlist.append(qt)
|
|
if not qtlist:
|
|
return
|
|
|
|
# Compute sum of all coefficients
|
|
sumcoeff = 0
|
|
for qt in qtlist:
|
|
coeff = qt.array("coeff")
|
|
for index in xrange(64):
|
|
sumcoeff += coeff[index].value
|
|
|
|
# Choose the right quality table and compute hash value
|
|
try:
|
|
hashval= qtlist[0]["coeff[2]"].value + qtlist[0]["coeff[53]"].value
|
|
if 2 <= len(qtlist):
|
|
hashval += qtlist[1]["coeff[0]"].value + qtlist[1]["coeff[63]"].value
|
|
hashtable = QUALITY_HASH_COLOR
|
|
sumtable = QUALITY_SUM_COLOR
|
|
else:
|
|
hashtable = QUALITY_HASH_GRAY
|
|
sumtable = QUALITY_SUM_GRAY
|
|
except (MissingField, IndexError):
|
|
# A coefficient is missing, so don't compute JPEG quality
|
|
return
|
|
|
|
# Find the JPEG quality
|
|
for index in xrange(100):
|
|
if (hashval >= hashtable[index]) or (sumcoeff >= sumtable[index]):
|
|
quality = "%s%%" % (index + 1)
|
|
if (hashval > hashtable[index]) or (sumcoeff > sumtable[index]):
|
|
quality += " " + _("(approximate)")
|
|
self.comment = "JPEG quality: %s" % quality
|
|
return
|
|
|
|
@fault_tolerant
|
|
def extractAPP0(self, app0):
|
|
self.format_version = u"JFIF %u.%02u" \
|
|
% (app0["ver_maj"].value, app0["ver_min"].value)
|
|
if "y_density" in app0:
|
|
self.width_dpi = app0["x_density"].value
|
|
self.height_dpi = app0["y_density"].value
|
|
|
|
@fault_tolerant
|
|
def processIfdEntry(self, ifd, entry):
|
|
# Skip unknown tags
|
|
tag = entry["tag"].value
|
|
if tag not in self.EXIF_KEY:
|
|
return
|
|
key = self.EXIF_KEY[tag]
|
|
if key in ("width", "height") and self.has(key):
|
|
# EXIF "valid size" are sometimes not updated when the image is scaled
|
|
# so we just ignore it
|
|
return
|
|
|
|
# Read value
|
|
if "value" in entry:
|
|
value = entry["value"].value
|
|
else:
|
|
value = ifd["value_%s" % entry.name].value
|
|
|
|
# Convert value to string
|
|
if tag == ExifEntry.TAG_ORIENTATION:
|
|
value = self.orientation_name.get(value, value)
|
|
elif tag == ExifEntry.TAG_EXPOSURE:
|
|
if not value:
|
|
return
|
|
if isinstance(value, float):
|
|
value = (value, u"1/%g" % (1/value))
|
|
elif entry["type"].value in (ExifEntry.TYPE_RATIONAL, ExifEntry.TYPE_SIGNED_RATIONAL):
|
|
value = (value, u"%.3g" % value)
|
|
|
|
# Store information
|
|
setattr(self, key, value)
|
|
|
|
@fault_tolerant
|
|
def readGPS(self, ifd):
|
|
# Read latitude and longitude
|
|
latitude_ref = None
|
|
longitude_ref = None
|
|
latitude = None
|
|
longitude = None
|
|
altitude_ref = 1
|
|
altitude = None
|
|
timestamp = None
|
|
datestamp = None
|
|
for entry in ifd.array("entry"):
|
|
tag = entry["tag"].value
|
|
if tag == ExifEntry.TAG_GPS_LATITUDE_REF:
|
|
if entry["value"].value == "N":
|
|
latitude_ref = 1
|
|
else:
|
|
latitude_ref = -1
|
|
elif tag == ExifEntry.TAG_GPS_LONGITUDE_REF:
|
|
if entry["value"].value == "E":
|
|
longitude_ref = 1
|
|
else:
|
|
longitude_ref = -1
|
|
elif tag == ExifEntry.TAG_GPS_ALTITUDE_REF:
|
|
if entry["value"].value == 1:
|
|
altitude_ref = -1
|
|
else:
|
|
altitude_ref = 1
|
|
elif tag == ExifEntry.TAG_GPS_LATITUDE:
|
|
latitude = [ifd["value_%s[%u]" % (entry.name, index)].value for index in xrange(3)]
|
|
elif tag == ExifEntry.TAG_GPS_LONGITUDE:
|
|
longitude = [ifd["value_%s[%u]" % (entry.name, index)].value for index in xrange(3)]
|
|
elif tag == ExifEntry.TAG_GPS_ALTITUDE:
|
|
altitude = ifd["value_%s" % entry.name].value
|
|
elif tag == ExifEntry.TAG_GPS_DATESTAMP:
|
|
datestamp = ifd["value_%s" % entry.name].value
|
|
elif tag == ExifEntry.TAG_GPS_TIMESTAMP:
|
|
items = [ifd["value_%s[%u]" % (entry.name, index)].value for index in xrange(3)]
|
|
items = map(int, items)
|
|
items = map(str, items)
|
|
timestamp = ":".join(items)
|
|
if latitude_ref and latitude:
|
|
value = deg2float(*latitude)
|
|
if latitude_ref < 0:
|
|
value = -value
|
|
self.latitude = value
|
|
if longitude and longitude_ref:
|
|
value = deg2float(*longitude)
|
|
if longitude_ref < 0:
|
|
value = -value
|
|
self.longitude = value
|
|
if altitude:
|
|
value = altitude
|
|
if altitude_ref < 0:
|
|
value = -value
|
|
self.altitude = value
|
|
if datestamp:
|
|
if timestamp:
|
|
datestamp += " " + timestamp
|
|
self.creation_date = datestamp
|
|
|
|
def parseIPTC(self, iptc):
|
|
datestr = hourstr = None
|
|
for field in iptc:
|
|
# Skip incomplete field
|
|
if "tag" not in field or "content" not in field:
|
|
continue
|
|
|
|
# Get value
|
|
value = field["content"].value
|
|
if isinstance(value, (str, unicode)):
|
|
value = value.replace("\r", " ")
|
|
value = value.replace("\n", " ")
|
|
|
|
# Skip unknown tag
|
|
tag = field["tag"].value
|
|
if tag == 55:
|
|
datestr = value
|
|
continue
|
|
if tag == 60:
|
|
hourstr = value
|
|
continue
|
|
if tag not in self.IPTC_KEY:
|
|
if tag != 0:
|
|
self.warning("Skip IPTC key %s: %s" % (
|
|
field["tag"].display, makeUnicode(value)))
|
|
continue
|
|
setattr(self, self.IPTC_KEY[tag], value)
|
|
if datestr and hourstr:
|
|
try:
|
|
year = int(datestr[0:4])
|
|
month = int(datestr[4:6])
|
|
day = int(datestr[6:8])
|
|
hour = int(hourstr[0:2])
|
|
min = int(hourstr[2:4])
|
|
sec = int(hourstr[4:6])
|
|
self.creation_date = datetime(year, month, day, hour, min, sec)
|
|
except ValueError:
|
|
pass
|
|
|
|
registerExtractor(JpegFile, JpegMetadata)
|
|
|