First attempt at merging video-patch.
This commit is contained in:
commit
ac52f76025
@ -24,14 +24,16 @@ def trim_base_custom(path, base):
|
|||||||
return path
|
return path
|
||||||
def trim_base(path):
|
def trim_base(path):
|
||||||
return trim_base_custom(path, trim_base.base)
|
return trim_base_custom(path, trim_base.base)
|
||||||
def cache_base(path):
|
def cache_base(path, filepath=False):
|
||||||
|
if len(path) == 0:
|
||||||
|
return "root"
|
||||||
|
elif filepath and len(path.split(os.sep)) < 2:
|
||||||
|
path = "root-" + path
|
||||||
path = trim_base(path).replace('/', '-').replace(' ', '_').replace('(', '').replace('&', '').replace(',', '').replace(')', '').replace('#', '').replace('[', '').replace(']', '').replace('"', '').replace("'", '').replace('_-_', '-').lower()
|
path = trim_base(path).replace('/', '-').replace(' ', '_').replace('(', '').replace('&', '').replace(',', '').replace(')', '').replace('#', '').replace('[', '').replace(']', '').replace('"', '').replace("'", '').replace('_-_', '-').lower()
|
||||||
while path.find("--") != -1:
|
while path.find("--") != -1:
|
||||||
path = path.replace("--", "-")
|
path = path.replace("--", "-")
|
||||||
while path.find("__") != -1:
|
while path.find("__") != -1:
|
||||||
path = path.replace("__", "_")
|
path = path.replace("__", "_")
|
||||||
if len(path) == 0:
|
|
||||||
path = "root"
|
|
||||||
return path
|
return path
|
||||||
def json_cache(path):
|
def json_cache(path):
|
||||||
return cache_base(path) + ".json"
|
return cache_base(path) + ".json"
|
||||||
@ -40,6 +42,8 @@ def image_cache(path, size, square=False):
|
|||||||
suffix = str(size) + "s"
|
suffix = str(size) + "s"
|
||||||
else:
|
else:
|
||||||
suffix = str(size)
|
suffix = str(size)
|
||||||
return cache_base(path) + "_" + suffix + ".jpg"
|
return cache_base(path, True) + "_" + suffix + ".jpg"
|
||||||
|
def video_cache(path):
|
||||||
|
return cache_base(path, True) + ".webm"
|
||||||
def file_mtime(path):
|
def file_mtime(path):
|
||||||
return datetime.fromtimestamp(int(os.path.getmtime(path)))
|
return datetime.fromtimestamp(int(os.path.getmtime(path)))
|
||||||
|
@ -7,13 +7,15 @@ from PIL import Image
|
|||||||
from PIL.ExifTags import TAGS
|
from PIL.ExifTags import TAGS
|
||||||
from multiprocessing import Pool
|
from multiprocessing import Pool
|
||||||
import gc
|
import gc
|
||||||
|
import tempfile
|
||||||
|
from VideoToolWrapper import *
|
||||||
|
|
||||||
def make_thumbs(self, original_path, thumb_path, size):
|
def make_photo_thumbs(self, original_path, thumb_path, size):
|
||||||
# The pool methods use a queue.Queue to pass tasks to the worker processes.
|
# The pool methods use a queue.Queue to pass tasks to the worker processes.
|
||||||
# Everything that goes through the queue.Queue must be pickable, and since
|
# Everything that goes through the queue.Queue must be pickable, and since
|
||||||
# self._thumbnail is not defined at the top level, it's not pickable.
|
# self._photo_thumbnail is not defined at the top level, it's not pickable.
|
||||||
# This is why we have this "dummy" function, so that it's pickable.
|
# This is why we have this "dummy" function, so that it's pickable.
|
||||||
self._thumbnail(original_path, thumb_path, size[0], size[1])
|
self._photo_thumbnail(original_path, thumb_path, size[0], size[1])
|
||||||
|
|
||||||
class Album(object):
|
class Album(object):
|
||||||
def __init__(self, path):
|
def __init__(self, path):
|
||||||
@ -116,6 +118,7 @@ class Photo(object):
|
|||||||
def __init__(self, path, thumb_path=None, attributes=None):
|
def __init__(self, path, thumb_path=None, attributes=None):
|
||||||
self._path = trim_base(path)
|
self._path = trim_base(path)
|
||||||
self.is_valid = True
|
self.is_valid = True
|
||||||
|
image = None
|
||||||
try:
|
try:
|
||||||
mtime = file_mtime(path)
|
mtime = file_mtime(path)
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
@ -128,17 +131,26 @@ class Photo(object):
|
|||||||
return
|
return
|
||||||
self._attributes = {}
|
self._attributes = {}
|
||||||
self._attributes["dateTimeFile"] = mtime
|
self._attributes["dateTimeFile"] = mtime
|
||||||
|
self._attributes["mediaType"] = "photo"
|
||||||
|
|
||||||
try:
|
try:
|
||||||
image = Image.open(path)
|
image = Image.open(path)
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
raise
|
raise
|
||||||
except:
|
except:
|
||||||
|
self._video_metadata(path)
|
||||||
|
|
||||||
|
if isinstance(image, Image.Image):
|
||||||
|
self._photo_metadata(image)
|
||||||
|
self._photo_thumbnails(image, thumb_path)
|
||||||
|
elif self._attributes["mediaType"] == "video":
|
||||||
|
self._video_thumbnails(thumb_path, path)
|
||||||
|
self._video_transcode(thumb_path, path)
|
||||||
|
else:
|
||||||
self.is_valid = False
|
self.is_valid = False
|
||||||
return
|
return
|
||||||
self._metadata(path)
|
|
||||||
self._thumbnails(path, thumb_path)
|
def _photo_metadata(self, image):
|
||||||
def _metadata(self, path):
|
|
||||||
self._attributes["size"] = image.size
|
self._attributes["size"] = image.size
|
||||||
self._orientation = 1
|
self._orientation = 1
|
||||||
try:
|
try:
|
||||||
@ -168,8 +180,8 @@ class Photo(object):
|
|||||||
self._orientation = exif["Orientation"];
|
self._orientation = exif["Orientation"];
|
||||||
if self._orientation in range(5, 9):
|
if self._orientation in range(5, 9):
|
||||||
self._attributes["size"] = (self._attributes["size"][1], self._attributes["size"][0])
|
self._attributes["size"] = (self._attributes["size"][1], self._attributes["size"][0])
|
||||||
if self._orientation - 1 < len(self._metadata.orientation_list):
|
if self._orientation - 1 < len(self._photo_metadata.orientation_list):
|
||||||
self._attributes["orientation"] = self._metadata.orientation_list[self._orientation - 1]
|
self._attributes["orientation"] = self._photo_metadata.orientation_list[self._orientation - 1]
|
||||||
if "Make" in exif:
|
if "Make" in exif:
|
||||||
self._attributes["make"] = exif["Make"]
|
self._attributes["make"] = exif["Make"]
|
||||||
if "Model" in exif:
|
if "Model" in exif:
|
||||||
@ -188,51 +200,72 @@ class Photo(object):
|
|||||||
self._attributes["iso"] = exif["PhotographicSensitivity"]
|
self._attributes["iso"] = exif["PhotographicSensitivity"]
|
||||||
if "ExposureTime" in exif:
|
if "ExposureTime" in exif:
|
||||||
self._attributes["exposureTime"] = exif["ExposureTime"]
|
self._attributes["exposureTime"] = exif["ExposureTime"]
|
||||||
if "Flash" in exif and exif["Flash"] in self._metadata.flash_dictionary:
|
if "Flash" in exif and exif["Flash"] in self._photo_metadata.flash_dictionary:
|
||||||
try:
|
try:
|
||||||
self._attributes["flash"] = self._metadata.flash_dictionary[exif["Flash"]]
|
self._attributes["flash"] = self._photo_metadata.flash_dictionary[exif["Flash"]]
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
raise
|
raise
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
if "LightSource" in exif and exif["LightSource"] in self._metadata.light_source_dictionary:
|
if "LightSource" in exif and exif["LightSource"] in self._photo_metadata.light_source_dictionary:
|
||||||
try:
|
try:
|
||||||
self._attributes["lightSource"] = self._metadata.light_source_dictionary[exif["LightSource"]]
|
self._attributes["lightSource"] = self._photo_metadata.light_source_dictionary[exif["LightSource"]]
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
raise
|
raise
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
if "ExposureProgram" in exif and exif["ExposureProgram"] < len(self._metadata.exposure_list):
|
if "ExposureProgram" in exif and exif["ExposureProgram"] < len(self._photo_metadata.exposure_list):
|
||||||
self._attributes["exposureProgram"] = self._metadata.exposure_list[exif["ExposureProgram"]]
|
self._attributes["exposureProgram"] = self._photo_metadata.exposure_list[exif["ExposureProgram"]]
|
||||||
if "SpectralSensitivity" in exif:
|
if "SpectralSensitivity" in exif:
|
||||||
self._attributes["spectralSensitivity"] = exif["SpectralSensitivity"]
|
self._attributes["spectralSensitivity"] = exif["SpectralSensitivity"]
|
||||||
if "MeteringMode" in exif and exif["MeteringMode"] < len(self._metadata.metering_list):
|
if "MeteringMode" in exif and exif["MeteringMode"] < len(self._photo_metadata.metering_list):
|
||||||
self._attributes["meteringMode"] = self._metadata.metering_list[exif["MeteringMode"]]
|
self._attributes["meteringMode"] = self._photo_metadata.metering_list[exif["MeteringMode"]]
|
||||||
if "SensingMethod" in exif and exif["SensingMethod"] < len(self._metadata.sensing_method_list):
|
if "SensingMethod" in exif and exif["SensingMethod"] < len(self._photo_metadata.sensing_method_list):
|
||||||
self._attributes["sensingMethod"] = self._metadata.sensing_method_list[exif["SensingMethod"]]
|
self._attributes["sensingMethod"] = self._photo_metadata.sensing_method_list[exif["SensingMethod"]]
|
||||||
if "SceneCaptureType" in exif and exif["SceneCaptureType"] < len(self._metadata.scene_capture_type_list):
|
if "SceneCaptureType" in exif and exif["SceneCaptureType"] < len(self._photo_metadata.scene_capture_type_list):
|
||||||
self._attributes["sceneCaptureType"] = self._metadata.scene_capture_type_list[exif["SceneCaptureType"]]
|
self._attributes["sceneCaptureType"] = self._photo_metadata.scene_capture_type_list[exif["SceneCaptureType"]]
|
||||||
if "SubjectDistanceRange" in exif and exif["SubjectDistanceRange"] < len(self._metadata.subject_distance_range_list):
|
if "SubjectDistanceRange" in exif and exif["SubjectDistanceRange"] < len(self._photo_metadata.subject_distance_range_list):
|
||||||
self._attributes["subjectDistanceRange"] = self._metadata.subject_distance_range_list[exif["SubjectDistanceRange"]]
|
self._attributes["subjectDistanceRange"] = self._photo_metadata.subject_distance_range_list[exif["SubjectDistanceRange"]]
|
||||||
if "ExposureCompensation" in exif:
|
if "ExposureCompensation" in exif:
|
||||||
self._attributes["exposureCompensation"] = exif["ExposureCompensation"]
|
self._attributes["exposureCompensation"] = exif["ExposureCompensation"]
|
||||||
if "ExposureBiasValue" in exif:
|
if "ExposureBiasValue" in exif:
|
||||||
self._attributes["exposureCompensation"] = exif["ExposureBiasValue"]
|
self._attributes["exposureCompensation"] = exif["ExposureBiasValue"]
|
||||||
if "DateTimeOriginal" in exif:
|
if "DateTimeOriginal" in exif:
|
||||||
self._attributes["dateTimeOriginal"] = exif["DateTimeOriginal"]
|
self._attributes["dateTimeOriginal"] = datetime.strptime(exif["DateTimeOriginal"], '%Y:%m:%d %H:%M:%S')
|
||||||
if "DateTime" in exif:
|
if "DateTime" in exif:
|
||||||
self._attributes["dateTime"] = exif["DateTime"]
|
self._attributes["dateTime"] = datetime.strptime(exif["DateTime"], '%Y:%m:%d %H:%M:%S')
|
||||||
|
|
||||||
_metadata.flash_dictionary = {0x0: "No Flash", 0x1: "Fired",0x5: "Fired, Return not detected",0x7: "Fired, Return detected",0x8: "On, Did not fire",0x9: "On, Fired",0xd: "On, Return not detected",0xf: "On, Return detected",0x10: "Off, Did not fire",0x14: "Off, Did not fire, Return not detected",0x18: "Auto, Did not fire",0x19: "Auto, Fired",0x1d: "Auto, Fired, Return not detected",0x1f: "Auto, Fired, Return detected",0x20: "No flash function",0x30: "Off, No flash function",0x41: "Fired, Red-eye reduction",0x45: "Fired, Red-eye reduction, Return not detected",0x47: "Fired, Red-eye reduction, Return detected",0x49: "On, Red-eye reduction",0x4d: "On, Red-eye reduction, Return not detected",0x4f: "On, Red-eye reduction, Return detected",0x50: "Off, Red-eye reduction",0x58: "Auto, Did not fire, Red-eye reduction",0x59: "Auto, Fired, Red-eye reduction",0x5d: "Auto, Fired, Red-eye reduction, Return not detected",0x5f: "Auto, Fired, Red-eye reduction, Return detected"}
|
|
||||||
_metadata.light_source_dictionary = {0: "Unknown", 1: "Daylight", 2: "Fluorescent", 3: "Tungsten (incandescent light)", 4: "Flash", 9: "Fine weather", 10: "Cloudy weather", 11: "Shade", 12: "Daylight fluorescent (D 5700 - 7100K)", 13: "Day white fluorescent (N 4600 - 5400K)", 14: "Cool white fluorescent (W 3900 - 4500K)", 15: "White fluorescent (WW 3200 - 3700K)", 17: "Standard light A", 18: "Standard light B", 19: "Standard light C", 20: "D55", 21: "D65", 22: "D75", 23: "D50", 24: "ISO studio tungsten"}
|
|
||||||
_metadata.metering_list = ["Unknown", "Average", "Center-weighted average", "Spot", "Multi-spot", "Multi-segment", "Partial"]
|
|
||||||
_metadata.exposure_list = ["Not Defined", "Manual", "Program AE", "Aperture-priority AE", "Shutter speed priority AE", "Creative (Slow speed)", "Action (High speed)", "Portrait", "Landscape", "Bulb"]
|
|
||||||
_metadata.orientation_list = ["Horizontal (normal)", "Mirror horizontal", "Rotate 180", "Mirror vertical", "Mirror horizontal and rotate 270 CW", "Rotate 90 CW", "Mirror horizontal and rotate 90 CW", "Rotate 270 CW"]
|
|
||||||
_metadata.sensing_method_list = ["Not defined", "One-chip color area sensor", "Two-chip color area sensor", "Three-chip color area sensor", "Color sequential area sensor", "Trilinear sensor", "Color sequential linear sensor"]
|
|
||||||
_metadata.scene_capture_type_list = ["Standard", "Landscape", "Portrait", "Night scene"]
|
|
||||||
_metadata.subject_distance_range_list = ["Unknown", "Macro", "Close view", "Distant view"]
|
|
||||||
|
|
||||||
def _thumbnail(self, original_path, thumb_path, size, square=False):
|
_photo_metadata.flash_dictionary = {0x0: "No Flash", 0x1: "Fired",0x5: "Fired, Return not detected",0x7: "Fired, Return detected",0x8: "On, Did not fire",0x9: "On, Fired",0xd: "On, Return not detected",0xf: "On, Return detected",0x10: "Off, Did not fire",0x14: "Off, Did not fire, Return not detected",0x18: "Auto, Did not fire",0x19: "Auto, Fired",0x1d: "Auto, Fired, Return not detected",0x1f: "Auto, Fired, Return detected",0x20: "No flash function",0x30: "Off, No flash function",0x41: "Fired, Red-eye reduction",0x45: "Fired, Red-eye reduction, Return not detected",0x47: "Fired, Red-eye reduction, Return detected",0x49: "On, Red-eye reduction",0x4d: "On, Red-eye reduction, Return not detected",0x4f: "On, Red-eye reduction, Return detected",0x50: "Off, Red-eye reduction",0x58: "Auto, Did not fire, Red-eye reduction",0x59: "Auto, Fired, Red-eye reduction",0x5d: "Auto, Fired, Red-eye reduction, Return not detected",0x5f: "Auto, Fired, Red-eye reduction, Return detected"}
|
||||||
|
_photo_metadata.light_source_dictionary = {0: "Unknown", 1: "Daylight", 2: "Fluorescent", 3: "Tungsten (incandescent light)", 4: "Flash", 9: "Fine weather", 10: "Cloudy weather", 11: "Shade", 12: "Daylight fluorescent (D 5700 - 7100K)", 13: "Day white fluorescent (N 4600 - 5400K)", 14: "Cool white fluorescent (W 3900 - 4500K)", 15: "White fluorescent (WW 3200 - 3700K)", 17: "Standard light A", 18: "Standard light B", 19: "Standard light C", 20: "D55", 21: "D65", 22: "D75", 23: "D50", 24: "ISO studio tungsten"}
|
||||||
|
_photo_metadata.metering_list = ["Unknown", "Average", "Center-weighted average", "Spot", "Multi-spot", "Multi-segment", "Partial"]
|
||||||
|
_photo_metadata.exposure_list = ["Not Defined", "Manual", "Program AE", "Aperture-priority AE", "Shutter speed priority AE", "Creative (Slow speed)", "Action (High speed)", "Portrait", "Landscape", "Bulb"]
|
||||||
|
_photo_metadata.orientation_list = ["Horizontal (normal)", "Mirror horizontal", "Rotate 180", "Mirror vertical", "Mirror horizontal and rotate 270 CW", "Rotate 90 CW", "Mirror horizontal and rotate 90 CW", "Rotate 270 CW"]
|
||||||
|
_photo_metadata.sensing_method_list = ["Not defined", "One-chip color area sensor", "Two-chip color area sensor", "Three-chip color area sensor", "Color sequential area sensor", "Trilinear sensor", "Color sequential linear sensor"]
|
||||||
|
_photo_metadata.scene_capture_type_list = ["Standard", "Landscape", "Portrait", "Night scene"]
|
||||||
|
_photo_metadata.subject_distance_range_list = ["Unknown", "Macro", "Close view", "Distant view"]
|
||||||
|
|
||||||
|
|
||||||
|
def _video_metadata(self, path, original=True):
|
||||||
|
p = VideoProbeWrapper().call('-show_format', '-show_streams', '-of', 'json', '-loglevel', '0', path)
|
||||||
|
if p == False:
|
||||||
|
self.is_valid = False
|
||||||
|
return
|
||||||
|
info = json.loads(p)
|
||||||
|
for s in info["streams"]:
|
||||||
|
if 'codec_type' in s and s['codec_type'] == 'video':
|
||||||
|
self._attributes["mediaType"] = "video"
|
||||||
|
self._attributes["size"] = (int(s["width"]), int(s["height"]))
|
||||||
|
if "duration" in s:
|
||||||
|
self._attributes["duration"] = s["duration"]
|
||||||
|
if "tags" in s and "rotate" in s["tags"]:
|
||||||
|
self._attributes["rotate"] = s["tags"]["rotate"]
|
||||||
|
if original:
|
||||||
|
self._attributes["originalSize"] = (int(s["width"]), int(s["height"]))
|
||||||
|
break
|
||||||
|
|
||||||
|
|
||||||
|
def _photo_thumbnail(self, original_path, thumb_path, size, square=False):
|
||||||
try:
|
try:
|
||||||
image = Image.open(original_path)
|
image = Image.open(original_path)
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
@ -284,6 +317,7 @@ class Photo(object):
|
|||||||
raise
|
raise
|
||||||
except:
|
except:
|
||||||
message("corrupt image", os.path.basename(original_path))
|
message("corrupt image", os.path.basename(original_path))
|
||||||
|
self.is_valid = False
|
||||||
return
|
return
|
||||||
if square:
|
if square:
|
||||||
if image.size[0] > image.size[1]:
|
if image.size[0] > image.size[1]:
|
||||||
@ -314,15 +348,80 @@ class Photo(object):
|
|||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def _thumbnails(self, original_path, thumb_path):
|
def _photo_thumbnails(self, original_path, thumb_path):
|
||||||
# get number of cores on the system, and use all minus one
|
# get number of cores on the system, and use all minus one
|
||||||
num_of_cores = os.sysconf('SC_NPROCESSORS_ONLN') - 1
|
num_of_cores = os.sysconf('SC_NPROCESSORS_ONLN') - 1
|
||||||
pool = Pool(processes=num_of_cores)
|
pool = Pool(processes=num_of_cores)
|
||||||
for size in Photo.thumb_sizes:
|
for size in Photo.thumb_sizes:
|
||||||
pool.apply_async(make_thumbs, args = (self, original_path, thumb_path, size))
|
pool.apply_async(make_photo_thumbs, args = (self, original_path, thumb_path, size))
|
||||||
pool.close()
|
pool.close()
|
||||||
pool.join()
|
pool.join()
|
||||||
|
|
||||||
|
def _video_thumbnails(self, thumb_path, original_path):
|
||||||
|
(tfd, tfn) = tempfile.mkstemp();
|
||||||
|
p = VideoTranscodeWrapper().call('-i', original_path, '-f', 'image2', '-vsync', '1', '-vframes', '1', '-an', '-loglevel', 'quiet', tfn)
|
||||||
|
if p == False:
|
||||||
|
message("couldn't extract video frame", os.path.basename(original_path))
|
||||||
|
os.unlink(tfn)
|
||||||
|
self.is_valid = False
|
||||||
|
return
|
||||||
|
try:
|
||||||
|
image = Image.open(tfn)
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
raise
|
||||||
|
except:
|
||||||
|
message("couldn't open video thumbnail", tfn)
|
||||||
|
os.unlink(tfn)
|
||||||
|
self.is_valid = False
|
||||||
|
return
|
||||||
|
mirror = image
|
||||||
|
if "rotate" in self._attributes:
|
||||||
|
if self._attributes["rotate"] == "90":
|
||||||
|
mirror = image.transpose(Image.ROTATE_270)
|
||||||
|
elif self._attributes["rotate"] == "180":
|
||||||
|
mirror = image.transpose(Image.ROTATE_180)
|
||||||
|
elif self._attributes["rotate"] == "270":
|
||||||
|
mirror = image.transpose(Image.ROTATE_90)
|
||||||
|
for size in Photo.thumb_sizes:
|
||||||
|
if size[1]:
|
||||||
|
self._thumbnail(mirror, thumb_path, original_path, size[0], size[1])
|
||||||
|
os.unlink(tfn)
|
||||||
|
|
||||||
|
def _video_transcode(self, transcode_path, original_path):
|
||||||
|
transcode_path = os.path.join(transcode_path, video_cache(self._path))
|
||||||
|
# get number of cores on the system, and use all minus one
|
||||||
|
num_of_cores = os.sysconf('SC_NPROCESSORS_ONLN') - 1
|
||||||
|
transcode_cmd = ['-i', original_path, '-c:v', 'libvpx', '-crf', '10', '-b:v', '800k', '-c:a', 'libvorbis', '-f', 'webm', '-threads', num_of_cores, '-loglevel', '0', '-y']
|
||||||
|
filters = []
|
||||||
|
info_string = "%s -> webm" % (os.path.basename(original_path))
|
||||||
|
message("transcoding", info_string)
|
||||||
|
if os.path.exists(transcode_path) and file_mtime(transcode_path) >= self._attributes["dateTimeFile"]:
|
||||||
|
self._video_metadata(transcode_path, False)
|
||||||
|
return
|
||||||
|
if "originalSize" in self._attributes and self._attributes["originalSize"][1] > 720:
|
||||||
|
filters.append("scale=trunc(oh*a/2)*2:min(720\,iw)")
|
||||||
|
if "rotate" in self._attributes:
|
||||||
|
if self._attributes["rotate"] == "90":
|
||||||
|
filters.append('transpose=1')
|
||||||
|
elif self._attributes["rotate"] == "180":
|
||||||
|
filters.append('vflip,hflip')
|
||||||
|
elif self._attributes["rotate"] == "270":
|
||||||
|
filters.append('transpose=2')
|
||||||
|
if len(filters):
|
||||||
|
transcode_cmd.append('-vf')
|
||||||
|
transcode_cmd.append(','.join(filters))
|
||||||
|
transcode_cmd.append(transcode_path)
|
||||||
|
p = VideoTranscodeWrapper().call(*transcode_cmd)
|
||||||
|
if p == False:
|
||||||
|
message("transcoding failure", os.path.basename(original_path))
|
||||||
|
try:
|
||||||
|
os.unlink(transcode_path)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
self.is_valid = False
|
||||||
|
return
|
||||||
|
self._video_metadata(transcode_path, False)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def name(self):
|
def name(self):
|
||||||
return os.path.basename(self._path)
|
return os.path.basename(self._path)
|
||||||
@ -333,7 +432,15 @@ class Photo(object):
|
|||||||
return self._path
|
return self._path
|
||||||
@property
|
@property
|
||||||
def image_caches(self):
|
def image_caches(self):
|
||||||
return [image_cache(self._path, size[0], size[1]) for size in Photo.thumb_sizes]
|
caches = []
|
||||||
|
if "mediaType" in self._attributes and self._attributes["mediaType"] == "video":
|
||||||
|
for size in Photo.thumb_sizes:
|
||||||
|
if size[1]:
|
||||||
|
caches.append(image_cache(self._path, size[0], size[1]))
|
||||||
|
caches.append(video_cache(self._path))
|
||||||
|
else:
|
||||||
|
caches = [image_cache(self._path, size[0], size[1]) for size in Photo.thumb_sizes]
|
||||||
|
return caches
|
||||||
@property
|
@property
|
||||||
def date(self):
|
def date(self):
|
||||||
correct_date = None;
|
correct_date = None;
|
||||||
|
38
scanner/VideoToolWrapper.py
Normal file
38
scanner/VideoToolWrapper.py
Normal file
@ -0,0 +1,38 @@
|
|||||||
|
from CachePath import message
|
||||||
|
import os
|
||||||
|
import subprocess
|
||||||
|
|
||||||
|
class VideoToolWrapper(object):
|
||||||
|
def call(self, *args):
|
||||||
|
path = args[-1]
|
||||||
|
for tool in self.wrappers:
|
||||||
|
try:
|
||||||
|
p = subprocess.check_output((tool,) + args)
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
if self.cleanup:
|
||||||
|
self.remove(path)
|
||||||
|
raise
|
||||||
|
except OSError:
|
||||||
|
continue
|
||||||
|
except:
|
||||||
|
if self.cleanup:
|
||||||
|
self.remove(path)
|
||||||
|
return False
|
||||||
|
return p
|
||||||
|
return False
|
||||||
|
|
||||||
|
def remove(self, path):
|
||||||
|
try:
|
||||||
|
os.unlink(path)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
class VideoTranscodeWrapper(VideoToolWrapper):
|
||||||
|
def __init__(self):
|
||||||
|
self.wrappers = ['avconv', 'ffmpeg']
|
||||||
|
self.cleanup = True
|
||||||
|
|
||||||
|
class VideoProbeWrapper(VideoToolWrapper):
|
||||||
|
def __init__(self):
|
||||||
|
self.wrappers = ['avprobe', 'ffprobe']
|
||||||
|
self.cleanup = False
|
@ -91,9 +91,14 @@ a:hover {
|
|||||||
right: 0;
|
right: 0;
|
||||||
text-align: center;
|
text-align: center;
|
||||||
}
|
}
|
||||||
#photo-box {
|
#photo-box, #video-box {
|
||||||
display: inline;
|
display: inline;
|
||||||
}
|
}
|
||||||
|
#video-box-inner {
|
||||||
|
position: absolute;
|
||||||
|
top: 50%;
|
||||||
|
width: 100%;
|
||||||
|
}
|
||||||
#photo-links {
|
#photo-links {
|
||||||
background-color: #000000;
|
background-color: #000000;
|
||||||
font-weight: bold;
|
font-weight: bold;
|
||||||
@ -194,6 +199,14 @@ a:hover {
|
|||||||
font-weight: bold;
|
font-weight: bold;
|
||||||
font-style: italic;
|
font-style: italic;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#video-unsupported {
|
||||||
|
background-image: url(../img/video-unsupported.png);
|
||||||
|
background-position: top center;
|
||||||
|
background-repeat: no-repeat;
|
||||||
|
padding-top: 96px;
|
||||||
|
}
|
||||||
|
|
||||||
#auth-text input {
|
#auth-text input {
|
||||||
color: rgb(0, 0, 0);
|
color: rgb(0, 0, 0);
|
||||||
background-color: rgb(200, 200, 200);
|
background-color: rgb(200, 200, 200);
|
||||||
|
BIN
web/img/video-icon.png
Normal file
BIN
web/img/video-icon.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 1.1 KiB |
BIN
web/img/video-unsupported.png
Normal file
BIN
web/img/video-unsupported.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 5.0 KiB |
@ -21,6 +21,10 @@
|
|||||||
<div id="metadata"></div>
|
<div id="metadata"></div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
<div id="video-box">
|
||||||
|
<div id="video-box-inner">
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
<a id="back">‹</a>
|
<a id="back">‹</a>
|
||||||
<a id="next">›</a>
|
<a id="next">›</a>
|
||||||
|
1393
web/js/005-modernizr.js
Normal file
1393
web/js/005-modernizr.js
Normal file
File diff suppressed because it is too large
Load Diff
@ -147,6 +147,9 @@
|
|||||||
hash = hash.substring(5);
|
hash = hash.substring(5);
|
||||||
return "cache/" + hash;
|
return "cache/" + hash;
|
||||||
};
|
};
|
||||||
|
PhotoFloat.videoPath = function(album, video) {
|
||||||
|
return "cache/" + PhotoFloat.cachePath(PhotoFloat.photoHash(album, video) + ".webm");
|
||||||
|
};
|
||||||
PhotoFloat.originalPhotoPath = function(album, photo) {
|
PhotoFloat.originalPhotoPath = function(album, photo) {
|
||||||
return "albums/" + album.path + "/" + photo.name;
|
return "albums/" + album.path + "/" + photo.name;
|
||||||
};
|
};
|
||||||
@ -179,6 +182,7 @@
|
|||||||
PhotoFloat.prototype.photoHash = PhotoFloat.photoHash;
|
PhotoFloat.prototype.photoHash = PhotoFloat.photoHash;
|
||||||
PhotoFloat.prototype.albumHash = PhotoFloat.albumHash;
|
PhotoFloat.prototype.albumHash = PhotoFloat.albumHash;
|
||||||
PhotoFloat.prototype.photoPath = PhotoFloat.photoPath;
|
PhotoFloat.prototype.photoPath = PhotoFloat.photoPath;
|
||||||
|
PhotoFloat.prototype.videoPath = PhotoFloat.videoPath;
|
||||||
PhotoFloat.prototype.originalPhotoPath = PhotoFloat.originalPhotoPath;
|
PhotoFloat.prototype.originalPhotoPath = PhotoFloat.originalPhotoPath;
|
||||||
PhotoFloat.prototype.trimExtension = PhotoFloat.trimExtension;
|
PhotoFloat.prototype.trimExtension = PhotoFloat.trimExtension;
|
||||||
PhotoFloat.prototype.cleanHash = PhotoFloat.cleanHash;
|
PhotoFloat.prototype.cleanHash = PhotoFloat.cleanHash;
|
||||||
|
@ -98,6 +98,8 @@ $(document).ready(function() {
|
|||||||
for (i = 0; i < currentAlbum.photos.length; ++i) {
|
for (i = 0; i < currentAlbum.photos.length; ++i) {
|
||||||
link = $("<a href=\"#!/" + photoFloat.photoHash(currentAlbum, currentAlbum.photos[i]) + "\"></a>");
|
link = $("<a href=\"#!/" + photoFloat.photoHash(currentAlbum, currentAlbum.photos[i]) + "\"></a>");
|
||||||
image = $("<img title=\"" + photoFloat.trimExtension(currentAlbum.photos[i].name) + "\" alt=\"" + photoFloat.trimExtension(currentAlbum.photos[i].name) + "\" src=\"" + photoFloat.photoPath(currentAlbum, currentAlbum.photos[i], 150, true) + "\" height=\"150\" width=\"150\" />");
|
image = $("<img title=\"" + photoFloat.trimExtension(currentAlbum.photos[i].name) + "\" alt=\"" + photoFloat.trimExtension(currentAlbum.photos[i].name) + "\" src=\"" + photoFloat.photoPath(currentAlbum, currentAlbum.photos[i], 150, true) + "\" height=\"150\" width=\"150\" />");
|
||||||
|
if (currentAlbum.photos[i].mediaType == "video")
|
||||||
|
image.css("background-image", "url(" + image.attr("src") + ")").attr("src", "img/video-icon.png");
|
||||||
image.get(0).photo = currentAlbum.photos[i];
|
image.get(0).photo = currentAlbum.photos[i];
|
||||||
link.append(image);
|
link.append(image);
|
||||||
photos.push(link);
|
photos.push(link);
|
||||||
@ -145,6 +147,8 @@ $(document).ready(function() {
|
|||||||
$("#album-view").removeClass("photo-view-container");
|
$("#album-view").removeClass("photo-view-container");
|
||||||
$("#subalbums").show();
|
$("#subalbums").show();
|
||||||
$("#photo-view").hide();
|
$("#photo-view").hide();
|
||||||
|
$("#video-box-inner").empty();
|
||||||
|
$("#video-box").hide();
|
||||||
}
|
}
|
||||||
setTimeout(scrollToThumb, 1);
|
setTimeout(scrollToThumb, 1);
|
||||||
}
|
}
|
||||||
@ -164,26 +168,70 @@ $(document).ready(function() {
|
|||||||
else if (image.css("height") !== "100%")
|
else if (image.css("height") !== "100%")
|
||||||
image.css("height", "100%").css("width", "auto").css("position", "").css("bottom", "");
|
image.css("height", "100%").css("width", "auto").css("position", "").css("bottom", "");
|
||||||
}
|
}
|
||||||
|
function scaleVideo() {
|
||||||
|
var video, container;
|
||||||
|
video = $("#video");
|
||||||
|
if (video.get(0) === this)
|
||||||
|
$(window).bind("resize", scaleVideo);
|
||||||
|
container = $("#photo-view");
|
||||||
|
if (video.attr("width") > container.width() && container.height() * video.attr("ratio") > container.width())
|
||||||
|
video.css("width", container.width()).css("height", container.width() / video.attr("ratio")).parent().css("height", container.width() / video.attr("ratio")).css("margin-top", - container.width() / video.attr("ratio") / 2).css("top", "50%");
|
||||||
|
else if (video.attr("height") > container.height() && container.height() * video.attr("ratio") < container.width())
|
||||||
|
video.css("height", container.height()).css("width", container.height() * video.attr("ratio")).parent().css("height", "100%").css("margin-top", "0").css("top", "0");
|
||||||
|
else
|
||||||
|
video.css("height", "").css("width", "").parent().css("height", video.attr("height")).css("margin-top", - video.attr("height") / 2).css("top", "50%");
|
||||||
|
}
|
||||||
function showPhoto() {
|
function showPhoto() {
|
||||||
var width, height, photoSrc, previousPhoto, nextPhoto, nextLink, text;
|
var width, height, photoSrc, videoSrc, previousPhoto, nextPhoto, nextLink, text;
|
||||||
width = currentPhoto.size[0];
|
width = currentPhoto.size[0];
|
||||||
height = currentPhoto.size[1];
|
height = currentPhoto.size[1];
|
||||||
if (width > height) {
|
|
||||||
height = height / width * maxSize;
|
if (currentPhoto.mediaType == "video") {
|
||||||
width = maxSize;
|
if (!Modernizr.video) {
|
||||||
} else {
|
$('<div id="video-unsupported"><p>Sorry, your browser doesn\'t support the HTML5 <video> element!</p><p>Here\'s a <a href="http://caniuse.com/video">list of which browsers do</a>.</p></div>').appendTo('#video-box-inner');
|
||||||
width = width / height * maxSize;
|
}
|
||||||
height = maxSize;
|
else if (!Modernizr.video.webm) {
|
||||||
|
$('<div id="video-unsupported"><p>Sorry, your browser doesn\'t support the WebM video format!</p></div>').appendTo('#video-box-inner');
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
$(window).unbind("resize", scaleVideo);
|
||||||
|
$(window).unbind("resize", scaleImage);
|
||||||
|
videoSrc = photoFloat.videoPath(currentAlbum, currentPhoto);
|
||||||
|
$('<video/>', { id: 'video', controls: true }).appendTo('#video-box-inner')
|
||||||
|
.attr("width", width).attr("height", height).attr("ratio", currentPhoto.size[0] / currentPhoto.size[1])
|
||||||
|
.attr("src", videoSrc)
|
||||||
|
.attr("alt", currentPhoto.name)
|
||||||
|
.on('loadstart', scaleVideo);
|
||||||
|
}
|
||||||
|
$("head").append("<link rel=\"video_src\" href=\"" + videoSrc + "\" />");
|
||||||
|
$("#video-box-inner").css('height', height + 'px').css('margin-top', - height / 2);
|
||||||
|
$("#photo-box").hide();
|
||||||
|
$("#video-box").show();
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
width = currentPhoto.size[0];
|
||||||
|
height = currentPhoto.size[1];
|
||||||
|
if (width > height) {
|
||||||
|
height = height / width * maxSize;
|
||||||
|
width = maxSize;
|
||||||
|
} else {
|
||||||
|
width = width / height * maxSize;
|
||||||
|
height = maxSize;
|
||||||
|
}
|
||||||
|
$(window).unbind("resize", scaleVideo);
|
||||||
|
$(window).unbind("resize", scaleImage);
|
||||||
|
photoSrc = photoFloat.photoPath(currentAlbum, currentPhoto, maxSize, false);
|
||||||
|
$("#photo")
|
||||||
|
.attr("width", width).attr("height", height).attr("ratio", currentPhoto.size[0] / currentPhoto.size[1])
|
||||||
|
.attr("src", photoSrc)
|
||||||
|
.attr("alt", currentPhoto.name)
|
||||||
|
.attr("title", currentPhoto.date)
|
||||||
|
.load(scaleImage);
|
||||||
|
$("head").append("<link rel=\"image_src\" href=\"" + photoSrc + "\" />");
|
||||||
|
$("#video-box-inner").empty();
|
||||||
|
$("#video-box").hide();
|
||||||
|
$("#photo-box").show();
|
||||||
}
|
}
|
||||||
$(window).unbind("resize", scaleImage);
|
|
||||||
photoSrc = photoFloat.photoPath(currentAlbum, currentPhoto, maxSize, false);
|
|
||||||
$("#photo")
|
|
||||||
.attr("width", width).attr("height", height).attr("ratio", currentPhoto.size[0] / currentPhoto.size[1])
|
|
||||||
.attr("src", photoSrc)
|
|
||||||
.attr("alt", currentPhoto.name)
|
|
||||||
.attr("title", currentPhoto.date)
|
|
||||||
.load(scaleImage);
|
|
||||||
$("head").append("<link rel=\"image_src\" href=\"" + photoSrc + "\" />");
|
|
||||||
|
|
||||||
previousPhoto = currentAlbum.photos[
|
previousPhoto = currentAlbum.photos[
|
||||||
(currentPhotoIndex - 1 < 0) ? (currentAlbum.photos.length - 1) : (currentPhotoIndex - 1)
|
(currentPhotoIndex - 1 < 0) ? (currentAlbum.photos.length - 1) : (currentPhotoIndex - 1)
|
||||||
@ -267,6 +315,7 @@ $(document).ready(function() {
|
|||||||
$(window).hashchange(function() {
|
$(window).hashchange(function() {
|
||||||
$("#loading").show();
|
$("#loading").show();
|
||||||
$("link[rel=image_src]").remove();
|
$("link[rel=image_src]").remove();
|
||||||
|
$("link[rel=video_src]").remove();
|
||||||
if (location.search.indexOf("?_escaped_fragment_=") === 0) {
|
if (location.search.indexOf("?_escaped_fragment_=") === 0) {
|
||||||
location.hash = location.search.substring(20);
|
location.hash = location.search.substring(20);
|
||||||
location.search = "";
|
location.search = "";
|
||||||
|
Loading…
Reference in New Issue
Block a user