2014-05-04 17:26:43 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# vi:si:et:sw=4:sts=4:ts=4
|
2014-09-02 22:32:44 +00:00
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2016-01-19 10:05:16 +00:00
|
|
|
from datetime import datetime
|
2014-09-02 22:32:44 +00:00
|
|
|
from io import StringIO, BytesIO
|
2016-01-19 10:05:16 +00:00
|
|
|
from PIL import Image
|
|
|
|
import base64
|
|
|
|
import hashlib
|
|
|
|
import json
|
|
|
|
import os
|
2014-05-04 17:26:43 +00:00
|
|
|
import re
|
2014-05-17 22:18:32 +00:00
|
|
|
import socket
|
2016-01-19 10:05:16 +00:00
|
|
|
import stdnum.isbn
|
2014-05-22 14:20:40 +00:00
|
|
|
import subprocess
|
2016-01-19 10:05:16 +00:00
|
|
|
import sys
|
|
|
|
import time
|
2014-05-04 17:26:43 +00:00
|
|
|
|
|
|
|
import ox
|
2015-11-26 00:26:10 +00:00
|
|
|
from OpenSSL.crypto import (
|
|
|
|
load_privatekey, load_certificate,
|
|
|
|
dump_privatekey, dump_certificate,
|
|
|
|
FILETYPE_ASN1, FILETYPE_PEM, PKey, TYPE_RSA,
|
|
|
|
X509, X509Extension
|
|
|
|
)
|
|
|
|
from Crypto.PublicKey import RSA
|
|
|
|
from Crypto.Util.asn1 import DerSequence
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2016-01-11 13:43:54 +00:00
|
|
|
from meta.utils import normalize_isbn, find_isbns, get_language, to_isbn13
|
2016-01-31 17:28:53 +00:00
|
|
|
from win32utils import get_short_path_name
|
2014-05-14 09:57:11 +00:00
|
|
|
|
2014-05-18 23:24:04 +00:00
|
|
|
import logging
|
2016-01-23 13:26:13 +00:00
|
|
|
logging.getLogger('PIL').setLevel(logging.ERROR)
|
2015-11-29 14:56:38 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
2014-05-17 22:18:32 +00:00
|
|
|
|
2016-12-26 15:55:07 +00:00
|
|
|
ENCODING = 'base64'
|
2014-05-17 22:18:32 +00:00
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def valid_olid(id):
|
|
|
|
return id.startswith('OL') and id.endswith('M')
|
|
|
|
|
|
|
|
def get_positions(ids, pos):
|
|
|
|
'''
|
|
|
|
>>> get_positions([1,2,3,4], [2,4])
|
|
|
|
{2: 1, 4: 3}
|
|
|
|
'''
|
|
|
|
positions = {}
|
|
|
|
for i in pos:
|
|
|
|
try:
|
|
|
|
positions[i] = ids.index(i)
|
|
|
|
except:
|
|
|
|
pass
|
|
|
|
return positions
|
|
|
|
|
|
|
|
def get_by_key(objects, key, value):
|
2014-09-02 22:32:44 +00:00
|
|
|
obj = [o for o in objects if o.get(key) == value]
|
2014-05-04 17:26:43 +00:00
|
|
|
return obj and obj[0] or None
|
|
|
|
|
|
|
|
def get_by_id(objects, id):
|
|
|
|
return get_by_key(objects, 'id', id)
|
|
|
|
|
2016-02-18 13:53:49 +00:00
|
|
|
def is_svg(data):
|
|
|
|
return data and b'<svg' in data[:256]
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def resize_image(data, width=None, size=None):
|
2014-09-02 22:32:44 +00:00
|
|
|
if isinstance(data, bytes):
|
|
|
|
data = BytesIO(data)
|
|
|
|
else:
|
|
|
|
data = StringIO(data)
|
|
|
|
source = Image.open(data)
|
2016-02-18 13:53:49 +00:00
|
|
|
#if source.mode not in ('1', 'CMYK', 'L', 'RGB', 'RGBA', 'RGBX', 'YCbCr'):
|
|
|
|
source = source.convert('RGB')
|
2014-05-04 17:26:43 +00:00
|
|
|
source_width = source.size[0]
|
|
|
|
source_height = source.size[1]
|
|
|
|
if size:
|
|
|
|
if source_width > source_height:
|
|
|
|
width = size
|
|
|
|
height = int(width / (float(source_width) / source_height))
|
|
|
|
height = height - height % 2
|
|
|
|
else:
|
|
|
|
height = size
|
|
|
|
width = int(height * (float(source_width) / source_height))
|
|
|
|
width = width - width % 2
|
|
|
|
|
|
|
|
else:
|
|
|
|
height = int(width / (float(source_width) / source_height))
|
|
|
|
height = height - height % 2
|
|
|
|
|
|
|
|
width = max(width, 1)
|
|
|
|
height = max(height, 1)
|
|
|
|
|
|
|
|
if width < source_width:
|
|
|
|
resize_method = Image.ANTIALIAS
|
|
|
|
else:
|
|
|
|
resize_method = Image.BICUBIC
|
|
|
|
output = source.resize((width, height), resize_method)
|
2014-09-02 22:32:44 +00:00
|
|
|
o = BytesIO()
|
2014-05-04 17:26:43 +00:00
|
|
|
output.save(o, format='jpeg')
|
|
|
|
data = o.getvalue()
|
|
|
|
o.close()
|
|
|
|
return data
|
|
|
|
|
|
|
|
def sort_title(title):
|
|
|
|
|
2014-09-02 22:32:44 +00:00
|
|
|
title = title.replace('Æ', 'Ae')
|
2014-05-04 17:26:43 +00:00
|
|
|
if isinstance(title, str):
|
2014-09-02 22:32:44 +00:00
|
|
|
title = str(title)
|
2014-05-04 17:26:43 +00:00
|
|
|
title = ox.sort_string(title)
|
|
|
|
|
|
|
|
#title
|
2014-09-02 22:32:44 +00:00
|
|
|
title = re.sub('[\'!¿¡,\.;\-"\:\*\[\]]', '', title)
|
2014-05-04 17:26:43 +00:00
|
|
|
return title.strip()
|
|
|
|
|
|
|
|
def get_position_by_id(list, key):
|
|
|
|
for i in range(0, len(list)):
|
|
|
|
if list[i]['id'] == key:
|
|
|
|
return i
|
|
|
|
return -1
|
|
|
|
|
2015-11-26 00:26:10 +00:00
|
|
|
def get_user_id(private_key, cert_path):
|
|
|
|
if os.path.exists(private_key):
|
|
|
|
with open(private_key) as fd:
|
|
|
|
key = load_privatekey(FILETYPE_PEM, fd.read())
|
|
|
|
if key.bits() != 1024:
|
|
|
|
os.unlink(private_key)
|
|
|
|
else:
|
|
|
|
user_id = get_service_id(private_key)
|
|
|
|
if not os.path.exists(private_key):
|
|
|
|
if os.path.exists(cert_path):
|
|
|
|
os.unlink(cert_path)
|
|
|
|
folder = os.path.dirname(private_key)
|
|
|
|
if not os.path.exists(folder):
|
|
|
|
os.makedirs(folder)
|
|
|
|
os.chmod(folder, 0o700)
|
|
|
|
key = PKey()
|
|
|
|
key.generate_key(TYPE_RSA, 1024)
|
|
|
|
with open(private_key, 'wb') as fd:
|
|
|
|
os.chmod(private_key, 0o600)
|
|
|
|
fd.write(dump_privatekey(FILETYPE_PEM, key))
|
|
|
|
os.chmod(private_key, 0o400)
|
|
|
|
user_id = get_service_id(private_key)
|
|
|
|
if not os.path.exists(cert_path):
|
|
|
|
ca = X509()
|
|
|
|
ca.set_version(2)
|
|
|
|
ca.set_serial_number(1)
|
|
|
|
ca.get_subject().CN = user_id
|
|
|
|
ca.gmtime_adj_notBefore(0)
|
|
|
|
ca.gmtime_adj_notAfter(24 * 60 * 60)
|
|
|
|
ca.set_issuer(ca.get_subject())
|
|
|
|
ca.set_pubkey(key)
|
|
|
|
ca.add_extensions([
|
|
|
|
X509Extension(b"basicConstraints", True, b"CA:TRUE, pathlen:0"),
|
|
|
|
X509Extension(b"nsCertType", True, b"sslCA"),
|
|
|
|
X509Extension(b"extendedKeyUsage", True,
|
|
|
|
b"serverAuth,clientAuth,emailProtection,timeStamping,msCodeInd,msCodeCom,msCTLSign,msSGC,msEFS,nsSGC"),
|
|
|
|
X509Extension(b"keyUsage", False, b"keyCertSign, cRLSign"),
|
|
|
|
X509Extension(b"subjectKeyIdentifier", False, b"hash", subject=ca),
|
|
|
|
])
|
|
|
|
ca.sign(key, "sha256")
|
|
|
|
with open(cert_path, 'wb') as fd:
|
|
|
|
fd.write(dump_certificate(FILETYPE_PEM, ca))
|
|
|
|
return user_id
|
|
|
|
|
|
|
|
def get_service_id(private_key_file=None, cert=None):
|
|
|
|
'''
|
|
|
|
service_id is the first half of the sha1 of the rsa public key encoded in base32
|
|
|
|
'''
|
|
|
|
if private_key_file:
|
|
|
|
with open(private_key_file, 'rb') as fd:
|
|
|
|
private_key = fd.read()
|
|
|
|
public_key = RSA.importKey(private_key).publickey().exportKey('DER')[22:]
|
|
|
|
# compute sha1 of public key and encode first half in base32
|
|
|
|
service_id = base64.b32encode(hashlib.sha1(public_key).digest()[:10]).lower().decode()
|
|
|
|
'''
|
|
|
|
# compute public key from priate key and export in DER format
|
|
|
|
# ignoring the SPKI header(22 bytes)
|
|
|
|
key = load_privatekey(FILETYPE_PEM, private_key)
|
|
|
|
cert = X509()
|
|
|
|
cert.set_pubkey(key)
|
|
|
|
public_key = dump_privatekey(FILETYPE_ASN1, cert.get_pubkey())[22:]
|
|
|
|
# compute sha1 of public key and encode first half in base32
|
|
|
|
service_id = base64.b32encode(hashlib.sha1(public_key).digest()[:10]).lower().decode()
|
|
|
|
'''
|
|
|
|
elif cert:
|
|
|
|
# compute sha1 of public key and encode first half in base32
|
|
|
|
key = load_certificate(FILETYPE_ASN1, cert).get_pubkey()
|
|
|
|
pub_der = DerSequence()
|
|
|
|
pub_der.decode(dump_privatekey(FILETYPE_ASN1, key))
|
|
|
|
public_key = RSA.construct((pub_der._seq[1], pub_der._seq[2])).exportKey('DER')[22:]
|
|
|
|
service_id = base64.b32encode(hashlib.sha1(public_key).digest()[:10]).lower().decode()
|
|
|
|
return service_id
|
|
|
|
|
2014-05-18 23:24:04 +00:00
|
|
|
def update_dict(root, data):
|
|
|
|
for key in data:
|
2016-01-05 15:11:59 +00:00
|
|
|
keys = [part.replace('\0', '.') for part in key.replace('\\.', '\0').split('.')]
|
2014-05-18 23:24:04 +00:00
|
|
|
value = data[key]
|
|
|
|
p = root
|
|
|
|
while len(keys)>1:
|
|
|
|
key = keys.pop(0)
|
|
|
|
if isinstance(p, list):
|
|
|
|
p = p[get_position_by_id(p, key)]
|
|
|
|
else:
|
|
|
|
if key not in p:
|
|
|
|
p[key] = {}
|
|
|
|
p = p[key]
|
|
|
|
if value == None and keys[0] in p:
|
|
|
|
del p[keys[0]]
|
|
|
|
else:
|
|
|
|
p[keys[0]] = value
|
2016-01-05 15:11:59 +00:00
|
|
|
if hasattr(root, '_save'):
|
|
|
|
root._save()
|
2014-05-18 23:24:04 +00:00
|
|
|
|
2016-01-18 12:02:42 +00:00
|
|
|
def remove_empty_folders(prefix, keep_root=False):
|
2014-05-18 23:24:04 +00:00
|
|
|
empty = []
|
|
|
|
for root, folders, files in os.walk(prefix):
|
2016-01-16 06:44:54 +00:00
|
|
|
if len(files) == 1 and files[0] == '.DS_Store':
|
|
|
|
os.unlink(os.path.join(root, files[0]))
|
|
|
|
files = []
|
2014-05-18 23:24:04 +00:00
|
|
|
if not folders and not files:
|
2016-01-18 12:02:42 +00:00
|
|
|
if root != prefix or not keep_root:
|
|
|
|
empty.append(root)
|
2014-05-18 23:24:04 +00:00
|
|
|
for folder in empty:
|
|
|
|
remove_empty_tree(folder)
|
|
|
|
|
|
|
|
def remove_empty_tree(leaf):
|
|
|
|
while leaf:
|
|
|
|
if not os.path.exists(leaf):
|
|
|
|
leaf = os.path.dirname(leaf)
|
|
|
|
elif os.path.isdir(leaf) and not os.listdir(leaf):
|
|
|
|
logger.debug('rmdir %s', leaf)
|
|
|
|
os.rmdir(leaf)
|
|
|
|
else:
|
|
|
|
break
|
2014-05-21 00:02:21 +00:00
|
|
|
|
2016-02-07 13:53:22 +00:00
|
|
|
try:
|
|
|
|
utc_0 = int(time.mktime(datetime(1970, 1, 1).timetuple()))
|
|
|
|
except:
|
|
|
|
utc_0 = int(time.mktime(time.gmtime()) - time.mktime(time.localtime()))
|
2014-05-21 00:02:21 +00:00
|
|
|
|
|
|
|
def datetime2ts(dt):
|
|
|
|
return int(time.mktime(dt.utctimetuple())) - utc_0
|
|
|
|
|
|
|
|
def ts2datetime(ts):
|
|
|
|
return datetime.utcfromtimestamp(float(ts))
|
2015-03-31 18:24:14 +00:00
|
|
|
|
|
|
|
def run(*cmd):
|
|
|
|
p = subprocess.Popen(cmd, close_fds=True)
|
|
|
|
p.wait()
|
|
|
|
return p.returncode
|
|
|
|
|
|
|
|
def get(*cmd):
|
2016-01-31 13:23:11 +00:00
|
|
|
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
2015-03-31 18:24:14 +00:00
|
|
|
stdout, error = p.communicate()
|
|
|
|
return stdout.decode()
|
|
|
|
|
|
|
|
def makefolder(path):
|
|
|
|
dirname = os.path.dirname(path)
|
|
|
|
if not os.path.exists(dirname):
|
|
|
|
os.makedirs(dirname)
|
2015-11-30 16:50:03 +00:00
|
|
|
|
|
|
|
|
2016-01-19 05:21:43 +00:00
|
|
|
def open_file(path=None):
|
|
|
|
cmd = []
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
cmd += ['open', path]
|
|
|
|
elif sys.platform.startswith('linux'):
|
|
|
|
cmd += ['xdg-open', path]
|
2016-01-31 13:23:11 +00:00
|
|
|
elif sys.platform == 'win32':
|
|
|
|
path = '\\'.join(path.split('/'))
|
|
|
|
os.startfile(path)
|
|
|
|
cmd = []
|
2016-01-19 05:21:43 +00:00
|
|
|
else:
|
|
|
|
logger.debug('unsupported platform %s', sys.platform)
|
2016-01-31 13:23:11 +00:00
|
|
|
if cmd:
|
|
|
|
subprocess.Popen(cmd, close_fds=True)
|
2016-01-19 05:21:43 +00:00
|
|
|
|
2015-11-30 17:07:07 +00:00
|
|
|
def open_folder(folder=None, path=None):
|
|
|
|
cmd = []
|
|
|
|
if path and not folder:
|
|
|
|
folder = os.path.dirname(path)
|
|
|
|
if folder and not path:
|
|
|
|
path = folder
|
|
|
|
if sys.platform == 'darwin':
|
|
|
|
if folder and not path:
|
|
|
|
path = folder
|
|
|
|
cmd += ['open', '-R', path]
|
|
|
|
elif sys.platform.startswith('linux'):
|
|
|
|
cmd += ['xdg-open', folder]
|
2016-01-31 13:23:11 +00:00
|
|
|
elif sys.platform == 'win32':
|
2016-01-31 19:19:25 +00:00
|
|
|
path = '\\'.join(path.split('/'))
|
2016-01-31 19:32:02 +00:00
|
|
|
cmd = 'explorer.exe /select,"%s"' % path
|
2015-11-30 17:07:07 +00:00
|
|
|
else:
|
|
|
|
logger.debug('unsupported platform %s', sys.platform)
|
2016-01-31 13:23:11 +00:00
|
|
|
if cmd:
|
|
|
|
subprocess.Popen(cmd, close_fds=True)
|
2015-12-02 21:05:23 +00:00
|
|
|
|
|
|
|
def can_connect_dns(host="8.8.8.8", port=53):
|
|
|
|
"""
|
|
|
|
host: 8.8.8.8 (google-public-dns-a.google.com)
|
|
|
|
port: 53/tcp
|
|
|
|
"""
|
2016-02-25 11:43:22 +00:00
|
|
|
import socks
|
|
|
|
import state
|
2015-12-02 21:05:23 +00:00
|
|
|
try:
|
2016-02-25 11:43:22 +00:00
|
|
|
sock = socks.socksocket(socket.AF_INET, socket.SOCK_STREAM, 6)
|
2016-03-18 10:29:18 +00:00
|
|
|
sock.settimeout(2)
|
2016-02-25 11:43:22 +00:00
|
|
|
socks_port = state.tor.socks_port if state.tor else 9150
|
|
|
|
sock.set_proxy(socks.SOCKS5, "localhost", socks_port, True)
|
|
|
|
sock.connect((host, port))
|
2015-12-02 21:05:23 +00:00
|
|
|
return True
|
|
|
|
except:
|
2016-03-18 10:29:18 +00:00
|
|
|
#logger.debug('failed to connect', exc_info=True)
|
2015-12-02 21:05:23 +00:00
|
|
|
pass
|
|
|
|
return False
|
2016-01-15 07:59:35 +00:00
|
|
|
|
|
|
|
def _to_json(python_object):
|
|
|
|
if isinstance(python_object, datetime):
|
|
|
|
if python_object.year < 1900:
|
|
|
|
tt = python_object.timetuple()
|
|
|
|
return '%d-%02d-%02dT%02d:%02d%02dZ' % tuple(list(tt)[:6])
|
|
|
|
return python_object.strftime('%Y-%m-%dT%H:%M:%SZ')
|
|
|
|
raise TypeError(u'%s %s is not JSON serializable' % (repr(python_object), type(python_object)))
|
2016-01-16 05:17:52 +00:00
|
|
|
|
|
|
|
def get_ratio(data):
|
|
|
|
try:
|
|
|
|
img = Image.open(BytesIO(data))
|
|
|
|
return img.size[0]/img.size[1]
|
|
|
|
except:
|
|
|
|
return 1
|
2016-01-19 10:05:16 +00:00
|
|
|
|
|
|
|
|
|
|
|
def get_meta_hash(data):
|
2016-02-10 14:02:32 +00:00
|
|
|
data = data.copy()
|
2016-01-19 10:05:16 +00:00
|
|
|
if 'sharemetadata' in data:
|
|
|
|
del data['sharemetadata']
|
|
|
|
for key in list(data):
|
|
|
|
if not data[key]:
|
|
|
|
del data[key]
|
|
|
|
return hashlib.sha1(json.dumps(data,
|
|
|
|
ensure_ascii=False, sort_keys=True).encode()).hexdigest()
|
2016-01-31 07:28:15 +00:00
|
|
|
|
|
|
|
|
|
|
|
def update_static():
|
|
|
|
import settings
|
|
|
|
import os
|
|
|
|
import ox
|
|
|
|
path = os.path.join(settings.static_path, 'js')
|
|
|
|
files = sorted([
|
|
|
|
file for file in os.listdir(path)
|
|
|
|
if not file.startswith('.')
|
|
|
|
and not file.startswith('oml.')
|
|
|
|
])
|
|
|
|
ox.file.write_json(os.path.join(settings.static_path, 'json', 'js.json'), files, indent=4)
|
|
|
|
ox.file.write_file(
|
|
|
|
os.path.join(path, 'oml.min.js'),
|
|
|
|
'\n'.join([
|
|
|
|
ox.js.minify(ox.file.read_file(os.path.join(path, file)).decode('utf-8'))
|
|
|
|
for file in files
|
|
|
|
])
|
|
|
|
)
|
2016-02-01 07:45:34 +00:00
|
|
|
|
|
|
|
def check_pid(pid):
|
|
|
|
try:
|
|
|
|
os.kill(pid, 0)
|
|
|
|
except:
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
return True
|
|
|
|
|
|
|
|
def check_pidfile(pid):
|
|
|
|
try:
|
|
|
|
with open(pid) as fd:
|
|
|
|
pid = int(fd.read())
|
|
|
|
except:
|
|
|
|
return False
|
|
|
|
return check_pid(pid)
|
|
|
|
|
|
|
|
def ctl(*args):
|
2016-02-01 08:06:48 +00:00
|
|
|
import settings
|
2016-02-01 07:45:34 +00:00
|
|
|
if sys.platform == 'win32':
|
|
|
|
platform_win32 = os.path.normpath(os.path.join(settings.base_dir, '..', 'platform_win32'))
|
|
|
|
python = os.path.join(platform_win32, 'pythonw.exe')
|
|
|
|
cmd = [python, 'oml'] + list(args)
|
|
|
|
startupinfo = subprocess.STARTUPINFO()
|
|
|
|
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
|
|
|
|
startupinfo.wShowWindow = subprocess.SW_HIDE
|
|
|
|
subprocess.Popen(cmd, cwd=settings.base_dir, start_new_session=True, startupinfo=startupinfo)
|
|
|
|
else:
|
|
|
|
subprocess.Popen([os.path.join(settings.base_dir, 'ctl')] + list(args),
|
|
|
|
close_fds=True, start_new_session=True)
|
2016-02-08 09:03:21 +00:00
|
|
|
|
|
|
|
def user_sort_key(u):
|
|
|
|
return ox.sort_string(str(u.get('index', '')) + 'Z' + (u.get('name') or ''))
|
2016-02-10 14:02:32 +00:00
|
|
|
|
|
|
|
def get_peer(peerid):
|
|
|
|
import state
|
|
|
|
import library
|
|
|
|
if peerid not in state.peers:
|
|
|
|
state.peers[peerid] = library.Peer(peerid)
|
|
|
|
return state.peers[peerid]
|
2016-02-23 08:17:10 +00:00
|
|
|
|
|
|
|
def send_debug():
|
|
|
|
import settings
|
|
|
|
import tor_request
|
|
|
|
import gzip
|
|
|
|
import io
|
|
|
|
url = 'http://rnogx24drkbnrxa3.onion/debug'
|
|
|
|
headers = {
|
|
|
|
'User-Agent': settings.USER_AGENT,
|
|
|
|
}
|
|
|
|
debug_log = os.path.join(settings.data_path, 'debug.log')
|
2016-02-27 07:01:12 +00:00
|
|
|
last_debug = settings.server.get('last_debug')
|
|
|
|
old = last_debug is not None
|
2016-02-23 08:17:10 +00:00
|
|
|
try:
|
|
|
|
if os.path.exists(debug_log):
|
|
|
|
data = []
|
2016-02-27 07:01:12 +00:00
|
|
|
with open(debug_log, 'r') as fd:
|
2016-02-23 08:17:10 +00:00
|
|
|
for line in fd:
|
2016-02-27 07:01:12 +00:00
|
|
|
t = line.split(':DEBUG')[0]
|
|
|
|
if t.count('-') == 2:
|
|
|
|
timestamp = t
|
|
|
|
if old and timestamp > last_debug:
|
|
|
|
old = False
|
2016-02-23 08:17:10 +00:00
|
|
|
if not old:
|
|
|
|
data.append(line)
|
2016-02-27 07:06:58 +00:00
|
|
|
data = ''.join(data)
|
2016-02-23 08:17:10 +00:00
|
|
|
if data:
|
|
|
|
bytes_io = io.BytesIO()
|
|
|
|
gzip_file = gzip.GzipFile(fileobj=bytes_io, mode='wb')
|
2016-02-27 07:01:12 +00:00
|
|
|
gzip_file.write(data.encode())
|
2016-02-23 08:17:10 +00:00
|
|
|
gzip_file.close()
|
|
|
|
result = bytes_io.getvalue()
|
|
|
|
bytes_io.close()
|
|
|
|
opener = tor_request.get_opener()
|
|
|
|
opener.addheaders = list(zip(headers.keys(), headers.values()))
|
|
|
|
r = opener.open(url, result)
|
|
|
|
if r.status != 200:
|
|
|
|
logger.debug('failed to send debug information')
|
2016-02-27 07:01:12 +00:00
|
|
|
else:
|
|
|
|
settings.server['last_debug'] = timestamp
|
2016-02-23 08:17:10 +00:00
|
|
|
except:
|
|
|
|
logger.debug('failed to send debug information')
|