2008-04-27 16:54:37 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
2008-06-19 09:21:21 +00:00
|
|
|
# vi:si:et:sw=4:sts=4:ts=4
|
2008-07-06 13:00:06 +00:00
|
|
|
# GPL 2008
|
2009-05-28 17:00:30 +00:00
|
|
|
from __future__ import division
|
2008-04-27 16:54:37 +00:00
|
|
|
import os
|
2009-03-16 17:15:14 +00:00
|
|
|
import hashlib
|
2010-11-15 16:35:07 +00:00
|
|
|
import re
|
2009-05-28 17:00:30 +00:00
|
|
|
import sys
|
2009-06-14 19:22:47 +00:00
|
|
|
import struct
|
2009-08-07 11:35:28 +00:00
|
|
|
import subprocess
|
|
|
|
|
2010-07-28 13:08:06 +00:00
|
|
|
from ox.utils import json
|
2009-05-28 17:00:30 +00:00
|
|
|
|
2011-04-06 13:24:32 +00:00
|
|
|
__all__ = ['sha1sum', 'oshash', 'avinfo', 'makedirs']
|
2008-04-27 16:54:37 +00:00
|
|
|
|
|
|
|
def sha1sum(filename):
|
2009-03-16 17:15:14 +00:00
|
|
|
sha1 = hashlib.sha1()
|
2008-06-19 09:21:21 +00:00
|
|
|
file=open(filename)
|
2008-04-27 16:54:37 +00:00
|
|
|
buffer=file.read(4096)
|
2008-06-19 09:21:21 +00:00
|
|
|
while buffer:
|
|
|
|
sha1.update(buffer)
|
|
|
|
buffer=file.read(4096)
|
|
|
|
file.close()
|
|
|
|
return sha1.hexdigest()
|
2008-04-27 16:54:37 +00:00
|
|
|
|
2009-05-28 17:00:30 +00:00
|
|
|
'''
|
|
|
|
os hash - http://trac.opensubtitles.org/projects/opensubtitles/wiki/HashSourceCodes
|
2009-06-14 19:22:47 +00:00
|
|
|
plus modification for files < 64k, buffer is filled with file data and padded with 0
|
2009-05-28 17:00:30 +00:00
|
|
|
'''
|
2009-06-14 19:22:47 +00:00
|
|
|
def oshash(filename):
|
|
|
|
try:
|
|
|
|
longlongformat = 'q' # long long
|
|
|
|
bytesize = struct.calcsize(longlongformat)
|
2009-05-28 17:00:30 +00:00
|
|
|
|
2009-06-14 19:22:47 +00:00
|
|
|
f = open(filename, "rb")
|
|
|
|
|
|
|
|
filesize = os.path.getsize(filename)
|
|
|
|
hash = filesize
|
|
|
|
if filesize < 65536:
|
|
|
|
for x in range(int(filesize/bytesize)):
|
|
|
|
buffer = f.read(bytesize)
|
|
|
|
(l_value,)= struct.unpack(longlongformat, buffer)
|
|
|
|
hash += l_value
|
|
|
|
hash = hash & 0xFFFFFFFFFFFFFFFF #to remain as 64bit number
|
|
|
|
else:
|
|
|
|
for x in range(int(65536/bytesize)):
|
|
|
|
buffer = f.read(bytesize)
|
|
|
|
(l_value,)= struct.unpack(longlongformat, buffer)
|
|
|
|
hash += l_value
|
|
|
|
hash = hash & 0xFFFFFFFFFFFFFFFF #to remain as 64bit number
|
|
|
|
f.seek(max(0,filesize-65536),0)
|
|
|
|
for x in range(int(65536/bytesize)):
|
|
|
|
buffer = f.read(bytesize)
|
|
|
|
(l_value,)= struct.unpack(longlongformat, buffer)
|
|
|
|
hash += l_value
|
|
|
|
hash = hash & 0xFFFFFFFFFFFFFFFF
|
|
|
|
f.close()
|
|
|
|
returnedhash = "%016x" % hash
|
|
|
|
return returnedhash
|
|
|
|
except(IOError):
|
2009-05-28 17:00:30 +00:00
|
|
|
return "IOError"
|
2008-04-27 16:54:37 +00:00
|
|
|
|
2009-08-07 11:35:28 +00:00
|
|
|
def avinfo(filename):
|
2010-02-27 11:47:13 +00:00
|
|
|
if os.path.getsize(filename):
|
2011-03-26 11:12:56 +00:00
|
|
|
p = subprocess.Popen(['ffmpeg2theora', '--info', filename],
|
|
|
|
stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
2010-02-27 11:47:13 +00:00
|
|
|
info, error = p.communicate()
|
2010-11-15 16:35:07 +00:00
|
|
|
try:
|
|
|
|
info = json.loads(info)
|
|
|
|
except:
|
|
|
|
#remove metadata, can be broken
|
|
|
|
reg = re.compile('"metadata": {.*?},', re.DOTALL)
|
|
|
|
info = re.sub(reg, '', info)
|
|
|
|
info = json.loads(info)
|
2011-03-26 11:12:56 +00:00
|
|
|
if 'video' in info:
|
|
|
|
for v in info['video']:
|
2011-04-19 13:42:04 +00:00
|
|
|
if not 'display_aspect_ratio' in v and 'width' in v:
|
2011-03-26 11:12:56 +00:00
|
|
|
v['display_aspect_ratio'] = '%d:%d' % (v['width'], v['height'])
|
|
|
|
v['pixel_aspect_ratio'] = '1:1'
|
2010-11-15 16:35:07 +00:00
|
|
|
return info
|
|
|
|
|
2010-02-27 11:47:13 +00:00
|
|
|
return {'path': filename, 'size': 0}
|
2011-04-06 13:24:32 +00:00
|
|
|
|
|
|
|
def makedirs(path):
|
|
|
|
if not os.path.exists(path):
|
|
|
|
try:
|
|
|
|
os.makedirs(path)
|
|
|
|
except OSError, e:
|
|
|
|
if e.errno != 17:
|
|
|
|
raise
|
2011-10-11 14:14:29 +00:00
|
|
|
|
|
|
|
def copy_file(source, target, verbose=False):
|
|
|
|
if verbose:
|
|
|
|
print 'copying', source, 'to', target
|
|
|
|
write_file(target, read_file(source))
|
|
|
|
|
|
|
|
def read_file(file, verbose=False):
|
|
|
|
if verbose:
|
|
|
|
print 'reading', file
|
|
|
|
f = open(file)
|
|
|
|
data = f.read()
|
|
|
|
f.close()
|
|
|
|
return data
|
|
|
|
|
|
|
|
def read_json(file, verbose=False):
|
|
|
|
return json.loads(read_file(file, verbose=verbose))
|
|
|
|
|
|
|
|
def write_file(file, data, verbose=False):
|
|
|
|
if verbose:
|
|
|
|
print 'writing', file
|
|
|
|
write_path(file)
|
|
|
|
f = open(file, 'w')
|
|
|
|
f.write(data)
|
|
|
|
f.close()
|
|
|
|
return len(data)
|
|
|
|
|
|
|
|
def write_json(file, data, indent=0, sort_keys=False, verbose=False):
|
|
|
|
data = json.dumps(data, indent=indent, sort_keys=sort_keys)
|
|
|
|
write_file(file, data, verbose=verbose)
|
|
|
|
|
|
|
|
def write_link(source, target, verbose=False):
|
|
|
|
if verbose:
|
|
|
|
print 'linking', source, 'to', target
|
|
|
|
write_path(target)
|
|
|
|
if os.path.exists(target):
|
|
|
|
os.unlink(target)
|
|
|
|
os.symlink(source, target)
|
|
|
|
|
|
|
|
def write_path(file):
|
|
|
|
path = os.path.split(file)[0]
|
|
|
|
if path and not os.path.exists(path):
|
|
|
|
os.makedirs(path)
|