2014-05-04 17:26:43 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
2019-02-02 08:51:25 +00:00
|
|
|
from io import BytesIO
|
2014-09-02 22:32:44 +00:00
|
|
|
from queue import Queue
|
2014-05-04 17:26:43 +00:00
|
|
|
from threading import Thread
|
2014-05-18 23:24:04 +00:00
|
|
|
import gzip
|
2019-02-02 08:51:25 +00:00
|
|
|
import json
|
2014-05-04 17:26:43 +00:00
|
|
|
import os
|
2016-02-12 09:00:59 +00:00
|
|
|
import socket
|
2016-02-18 11:38:41 +00:00
|
|
|
import socks
|
2019-02-02 08:51:25 +00:00
|
|
|
import time
|
|
|
|
import urllib.error
|
|
|
|
import urllib.parse
|
|
|
|
import urllib.request
|
2014-05-04 17:26:43 +00:00
|
|
|
|
|
|
|
import ox
|
2014-05-18 23:24:04 +00:00
|
|
|
from tornado.ioloop import PeriodicCallback
|
2014-05-04 17:26:43 +00:00
|
|
|
|
|
|
|
import settings
|
|
|
|
import user.models
|
|
|
|
|
|
|
|
from websocket import trigger_event
|
2014-05-12 12:57:47 +00:00
|
|
|
from localnodes import LocalNodes
|
2015-11-26 00:26:10 +00:00
|
|
|
from tor_request import get_opener
|
2016-02-10 14:02:32 +00:00
|
|
|
from utils import user_sort_key, get_peer
|
2014-05-21 22:41:29 +00:00
|
|
|
import state
|
2014-08-09 16:14:14 +00:00
|
|
|
import db
|
2016-02-10 14:02:32 +00:00
|
|
|
import library
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2014-05-17 14:26:59 +00:00
|
|
|
import logging
|
2015-11-29 14:56:38 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
2014-05-17 14:26:59 +00:00
|
|
|
|
2017-06-03 20:50:14 +00:00
|
|
|
DEBUG_NODES = False
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2014-05-19 15:00:33 +00:00
|
|
|
class Node(Thread):
|
2014-09-02 23:09:42 +00:00
|
|
|
host = None
|
2015-12-01 08:59:52 +00:00
|
|
|
local = None
|
2015-12-02 21:05:23 +00:00
|
|
|
_online = None
|
2014-05-18 23:50:05 +00:00
|
|
|
TIMEOUT = 5
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2019-02-02 07:13:37 +00:00
|
|
|
def __init__(self, nodes, user_id):
|
2014-05-12 12:57:47 +00:00
|
|
|
self._nodes = nodes
|
2019-02-02 07:13:37 +00:00
|
|
|
self.user_id = user_id
|
2015-11-26 00:26:10 +00:00
|
|
|
self._opener = get_opener(self.user_id)
|
2014-05-19 15:00:33 +00:00
|
|
|
self._q = Queue()
|
2019-02-02 08:51:25 +00:00
|
|
|
self._pingcb = PeriodicCallback(self.ping, 10 * settings.server['pull_interval'])
|
2019-02-02 09:12:26 +00:00
|
|
|
state.main.add_callback(self._pingcb.start)
|
2014-05-19 15:00:33 +00:00
|
|
|
Thread.__init__(self)
|
|
|
|
self.daemon = True
|
2016-01-31 16:45:14 +00:00
|
|
|
self.start()
|
2014-05-19 15:00:33 +00:00
|
|
|
|
|
|
|
def run(self):
|
2016-02-14 05:53:55 +00:00
|
|
|
self.ping()
|
2016-01-31 16:45:14 +00:00
|
|
|
while not state.shutdown:
|
2014-08-09 16:14:14 +00:00
|
|
|
action = self._q.get()
|
2016-01-31 16:45:14 +00:00
|
|
|
if state.shutdown:
|
2014-08-09 16:14:14 +00:00
|
|
|
break
|
2019-02-02 08:51:25 +00:00
|
|
|
if action == 'ping':
|
2015-12-01 08:59:52 +00:00
|
|
|
self.online = self.can_connect()
|
2019-02-02 08:51:25 +00:00
|
|
|
elif action == 'send_response':
|
|
|
|
if self.online:
|
|
|
|
self._send_response()
|
|
|
|
else:
|
|
|
|
if not self._q.qsize():
|
|
|
|
time.sleep(5)
|
|
|
|
self.send_response()
|
2019-02-02 07:13:37 +00:00
|
|
|
elif isinstance(action, list) and len(action) == 2:
|
|
|
|
if self.online:
|
|
|
|
self._call(action[0], *action[1])
|
|
|
|
else:
|
|
|
|
if not self._q.qsize():
|
|
|
|
time.sleep(5)
|
2019-02-20 12:58:25 +00:00
|
|
|
else:
|
|
|
|
time.sleep(0.1)
|
2019-02-02 07:13:37 +00:00
|
|
|
self.queue(action[0], *action[1])
|
2015-12-01 08:59:52 +00:00
|
|
|
else:
|
|
|
|
logger.debug('unknown action %s', action)
|
2014-05-19 15:00:33 +00:00
|
|
|
|
|
|
|
def join(self):
|
2015-12-01 08:59:52 +00:00
|
|
|
self._q.put('')
|
2014-08-09 18:32:41 +00:00
|
|
|
#return Thread.join(self)
|
2014-05-19 15:00:33 +00:00
|
|
|
|
|
|
|
def ping(self):
|
2016-09-29 08:35:47 +00:00
|
|
|
if state.online or self.get_local():
|
2015-12-02 21:05:23 +00:00
|
|
|
self._q.put('ping')
|
2014-05-19 15:00:33 +00:00
|
|
|
|
2019-02-02 07:13:37 +00:00
|
|
|
def queue(self, action, *args):
|
2019-02-20 12:58:25 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('queue node action %s->%s%s', self.user_id, action, args)
|
2019-02-02 07:13:37 +00:00
|
|
|
self._q.put([action, args])
|
|
|
|
|
|
|
|
def _call(self, action, *args):
|
|
|
|
r = getattr(self, action)(*args)
|
|
|
|
logger.debug('call node api %s->%s%s = %s', self.user_id, action, args, r)
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
@property
|
|
|
|
def url(self):
|
2015-11-26 00:26:10 +00:00
|
|
|
if self.local:
|
|
|
|
if ':' in self.local:
|
|
|
|
url = 'https://[%s]:%s' % (self.local, self.port)
|
2014-05-12 12:57:47 +00:00
|
|
|
else:
|
2015-11-26 00:26:10 +00:00
|
|
|
url = 'https://%s:%s' % (self.local, self.port)
|
2016-04-04 23:28:55 +00:00
|
|
|
else:
|
2015-11-26 00:26:10 +00:00
|
|
|
url = 'https://%s.onion:9851' % self.user_id
|
2014-05-04 17:26:43 +00:00
|
|
|
return url
|
|
|
|
|
2015-12-02 21:05:23 +00:00
|
|
|
@property
|
|
|
|
def online(self):
|
|
|
|
return self._online
|
|
|
|
|
|
|
|
@online.setter
|
|
|
|
def online(self, online):
|
|
|
|
if self._online != online:
|
|
|
|
self._online = online
|
|
|
|
self.trigger_status()
|
|
|
|
else:
|
|
|
|
self._online = online
|
|
|
|
|
2014-05-13 10:36:02 +00:00
|
|
|
def resolve(self):
|
2015-12-01 10:51:58 +00:00
|
|
|
#logger.debug('resolve node %s', self.user_id)
|
2014-05-22 14:20:40 +00:00
|
|
|
r = self.get_local()
|
2014-05-04 17:26:43 +00:00
|
|
|
if r:
|
2015-11-26 00:26:10 +00:00
|
|
|
self.local = r['host']
|
2014-05-04 17:26:43 +00:00
|
|
|
if 'port' in r:
|
|
|
|
self.port = r['port']
|
|
|
|
else:
|
2015-11-26 00:26:10 +00:00
|
|
|
self.local = None
|
2014-05-04 17:26:43 +00:00
|
|
|
self.port = 9851
|
|
|
|
|
2019-02-10 12:16:35 +00:00
|
|
|
def is_local(self):
|
|
|
|
return self._nodes and self.user_id in self._nodes.local
|
|
|
|
|
2014-05-12 12:57:47 +00:00
|
|
|
def get_local(self):
|
2016-03-14 13:31:56 +00:00
|
|
|
if self._nodes and self._nodes.local:
|
2019-02-10 12:16:35 +00:00
|
|
|
return self._nodes.local.get_data(self.user_id)
|
2014-05-12 12:57:47 +00:00
|
|
|
return None
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def request(self, action, *args):
|
2015-11-26 00:26:10 +00:00
|
|
|
self.resolve()
|
2014-05-13 10:36:02 +00:00
|
|
|
url = self.url
|
2016-04-04 23:28:55 +00:00
|
|
|
if self.local:
|
|
|
|
logger.debug('request:%s(%s:%s): %s%s', self.user_id, self.local, self.port, action, list(args))
|
|
|
|
else:
|
|
|
|
logger.debug('request:%s: %s%s', self.user_id, action, list(args))
|
2014-09-09 10:08:04 +00:00
|
|
|
content = json.dumps([action, args]).encode()
|
2014-05-04 17:26:43 +00:00
|
|
|
headers = {
|
|
|
|
'User-Agent': settings.USER_AGENT,
|
2014-05-21 22:41:29 +00:00
|
|
|
'X-Node-Protocol': settings.NODE_PROTOCOL,
|
2014-05-04 17:26:43 +00:00
|
|
|
'Accept': 'text/plain',
|
|
|
|
'Accept-Encoding': 'gzip',
|
|
|
|
'Content-Type': 'application/json',
|
|
|
|
}
|
2015-11-26 00:26:10 +00:00
|
|
|
self._opener.addheaders = list(zip(headers.keys(), headers.values()))
|
2015-12-02 15:30:37 +00:00
|
|
|
#logger.debug('headers: %s', self._opener.addheaders)
|
2014-05-14 09:57:11 +00:00
|
|
|
try:
|
2016-04-04 23:29:49 +00:00
|
|
|
r = self._opener.open(url, data=content, timeout=self.TIMEOUT*12)
|
2014-09-02 22:32:44 +00:00
|
|
|
except urllib.error.HTTPError as e:
|
2014-05-14 09:57:11 +00:00
|
|
|
if e.code == 403:
|
2016-01-14 12:26:01 +00:00
|
|
|
logger.debug('403: %s (%s)', url, self.user_id)
|
|
|
|
if state.tasks:
|
|
|
|
state.tasks.queue('peering', (self.user_id, False))
|
2019-01-17 18:19:23 +00:00
|
|
|
del self._nodes._nodes[self.user_id]
|
2016-01-14 12:26:01 +00:00
|
|
|
self.online = False
|
|
|
|
return None
|
2014-05-17 14:26:59 +00:00
|
|
|
logger.debug('urllib2.HTTPError %s %s', e, e.code)
|
2014-05-14 09:57:11 +00:00
|
|
|
self.online = False
|
|
|
|
return None
|
2014-09-02 22:32:44 +00:00
|
|
|
except urllib.error.URLError as e:
|
2014-05-17 14:26:59 +00:00
|
|
|
logger.debug('urllib2.URLError %s', e)
|
2014-05-14 09:57:11 +00:00
|
|
|
self.online = False
|
|
|
|
return None
|
2016-07-04 10:06:50 +00:00
|
|
|
except socket.timeout:
|
|
|
|
logger.debug('timeout %s', url)
|
|
|
|
self.online = False
|
|
|
|
return None
|
2014-05-14 09:57:11 +00:00
|
|
|
except:
|
2016-01-24 09:13:03 +00:00
|
|
|
logger.debug('unknown url error', exc_info=True)
|
2014-05-14 09:57:11 +00:00
|
|
|
self.online = False
|
|
|
|
return None
|
|
|
|
data = r.read()
|
2014-05-18 23:24:04 +00:00
|
|
|
if r.headers.get('content-encoding', None) == 'gzip':
|
2014-10-31 14:47:54 +00:00
|
|
|
data = gzip.GzipFile(fileobj=BytesIO(data)).read()
|
2014-05-21 22:41:29 +00:00
|
|
|
|
|
|
|
version = r.headers.get('X-Node-Protocol', None)
|
|
|
|
if version != settings.NODE_PROTOCOL:
|
2016-01-17 15:00:19 +00:00
|
|
|
logger.debug('version does not match local: %s remote %s (%s)', settings.NODE_PROTOCOL, version, self.user_id)
|
2014-05-21 22:41:29 +00:00
|
|
|
self.online = False
|
|
|
|
if version > settings.NODE_PROTOCOL:
|
|
|
|
state.update_required = True
|
|
|
|
return None
|
|
|
|
|
2015-11-26 00:26:10 +00:00
|
|
|
response = json.loads(data.decode('utf-8'))
|
2014-05-04 17:26:43 +00:00
|
|
|
return response
|
|
|
|
|
2014-05-18 03:01:24 +00:00
|
|
|
def can_connect(self):
|
2015-12-01 08:59:52 +00:00
|
|
|
self.resolve()
|
|
|
|
url = self.url
|
2016-02-24 07:19:00 +00:00
|
|
|
if not state.online and not self.local:
|
|
|
|
return False
|
2014-05-18 03:01:24 +00:00
|
|
|
try:
|
2014-05-22 14:20:40 +00:00
|
|
|
if url:
|
|
|
|
headers = {
|
|
|
|
'User-Agent': settings.USER_AGENT,
|
|
|
|
'X-Node-Protocol': settings.NODE_PROTOCOL,
|
|
|
|
'Accept-Encoding': 'gzip',
|
|
|
|
}
|
2015-11-26 00:26:10 +00:00
|
|
|
self._opener.addheaders = list(zip(headers.keys(), headers.values()))
|
2015-12-01 10:51:58 +00:00
|
|
|
self._opener.timeout = 2
|
2014-09-02 23:09:42 +00:00
|
|
|
r = self._opener.open(url)
|
2014-05-22 14:20:40 +00:00
|
|
|
version = r.headers.get('X-Node-Protocol', None)
|
|
|
|
if version != settings.NODE_PROTOCOL:
|
|
|
|
logger.debug('version does not match local: %s remote %s', settings.NODE_PROTOCOL, version)
|
|
|
|
return False
|
|
|
|
c = r.read()
|
2016-02-23 09:07:06 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('can connect to: %s', url)
|
2014-05-22 14:20:40 +00:00
|
|
|
return True
|
2014-05-18 03:01:24 +00:00
|
|
|
except:
|
2016-02-23 09:07:06 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('can not connect to: %s', url)
|
2014-05-18 03:01:24 +00:00
|
|
|
pass
|
|
|
|
return False
|
|
|
|
|
2015-12-01 08:59:52 +00:00
|
|
|
def is_online(self):
|
2019-02-10 12:16:35 +00:00
|
|
|
return self.online or self.is_local()
|
2015-12-01 08:59:52 +00:00
|
|
|
|
2016-01-17 13:12:56 +00:00
|
|
|
def send_response(self):
|
|
|
|
self._q.put('send_response')
|
|
|
|
|
|
|
|
def _send_response(self):
|
2016-01-12 07:32:39 +00:00
|
|
|
with db.session():
|
2016-03-01 11:38:58 +00:00
|
|
|
u = user.models.User.get(self.user_id)
|
2019-01-31 08:41:54 +00:00
|
|
|
if u:
|
|
|
|
user_pending = u.pending
|
|
|
|
user_peered = u.peered
|
|
|
|
user_queued = u.queued
|
2019-02-02 08:51:25 +00:00
|
|
|
else:
|
|
|
|
user_queued = False
|
2019-01-31 08:41:54 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('go online peered=%s queued=%s %s (%s)', u.peered, u.queued, u.id, u.nickname)
|
|
|
|
|
2019-02-02 08:51:25 +00:00
|
|
|
if user_queued:
|
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('connected to %s', self.url)
|
|
|
|
logger.debug('queued peering event pending=%s peered=%s', user_pending, user_peered)
|
|
|
|
if user_pending == 'sent':
|
|
|
|
self.peering('requestPeering')
|
|
|
|
elif user_pending == '' and user_peered:
|
|
|
|
self.peering('acceptPeering')
|
|
|
|
else:
|
|
|
|
#fixme, what about cancel/reject peering here?
|
|
|
|
self.peering('removePeering')
|
2014-05-19 15:00:33 +00:00
|
|
|
|
|
|
|
def trigger_status(self):
|
2015-12-02 21:05:23 +00:00
|
|
|
if self.online is not None:
|
|
|
|
trigger_event('status', {
|
|
|
|
'id': self.user_id,
|
|
|
|
'online': self.online
|
|
|
|
})
|
2014-05-04 17:26:43 +00:00
|
|
|
|
|
|
|
def pullChanges(self):
|
2016-02-24 07:19:00 +00:00
|
|
|
if state.shutdown:
|
|
|
|
return
|
|
|
|
self.online = self.can_connect()
|
|
|
|
if not self.online or state.shutdown:
|
|
|
|
return
|
2017-06-03 20:50:14 +00:00
|
|
|
self.resolve()
|
2016-02-10 14:02:32 +00:00
|
|
|
peer = get_peer(self.user_id)
|
2017-06-03 20:50:14 +00:00
|
|
|
path = peer._logpath
|
|
|
|
if os.path.exists(path):
|
|
|
|
size = os.path.getsize(path)
|
|
|
|
else:
|
|
|
|
size = 0
|
|
|
|
url = '%s/log' % self.url
|
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('pullChanges: %s [%s]', self.user_id, url)
|
|
|
|
headers = self.headers.copy()
|
|
|
|
if size:
|
|
|
|
headers['Range'] = '%s-' % size
|
|
|
|
self._opener.addheaders = list(zip(headers.keys(), headers.values()))
|
2016-02-10 14:02:32 +00:00
|
|
|
try:
|
2017-06-03 20:50:14 +00:00
|
|
|
r = self._opener.open(url, timeout=self.TIMEOUT*60)
|
|
|
|
except urllib.error.HTTPError as e:
|
|
|
|
if e.code == 403:
|
|
|
|
logger.debug('pullChanges 403: %s (%s)', url, self.user_id)
|
|
|
|
if state.tasks:
|
|
|
|
state.tasks.queue('peering', (self.user_id, False))
|
2019-01-17 18:19:23 +00:00
|
|
|
del self._nodes._nodes[self.user_id]
|
2017-06-03 20:50:14 +00:00
|
|
|
self.online = False
|
|
|
|
else:
|
|
|
|
logger.debug('unknown http errpr %s %s (%s)', e.code, url, self.user_id)
|
|
|
|
return False
|
|
|
|
except socket.timeout:
|
|
|
|
logger.debug('timeout %s', url)
|
|
|
|
return False
|
|
|
|
except socks.GeneralProxyError:
|
|
|
|
logger.debug('openurl failed %s', url)
|
|
|
|
return False
|
|
|
|
except urllib.error.URLError as e:
|
|
|
|
logger.debug('openurl failed urllib2.URLError %s', e.reason)
|
|
|
|
return False
|
2016-02-10 14:02:32 +00:00
|
|
|
except:
|
2017-06-03 20:50:14 +00:00
|
|
|
logger.debug('openurl failed %s', url, exc_info=True)
|
2016-02-10 14:02:32 +00:00
|
|
|
return False
|
2017-06-03 20:50:14 +00:00
|
|
|
if r.getcode() in (200, 206):
|
|
|
|
changed = False
|
|
|
|
chunk_size = 16 * 1024
|
|
|
|
mode = 'ab' if r.getcode() == 206 else 'wb'
|
|
|
|
content = b''
|
|
|
|
|
|
|
|
try:
|
|
|
|
if r.headers.get('content-encoding', None) == 'gzip':
|
|
|
|
fileobj = gzip.GzipFile(fileobj=r)
|
|
|
|
else:
|
|
|
|
fileobj = r
|
|
|
|
for chunk in iter(lambda: fileobj.read(chunk_size), b''):
|
|
|
|
content += chunk
|
|
|
|
eol = content.rfind(b'\n') + 1
|
|
|
|
if eol > 0:
|
|
|
|
with open(path, mode) as fd:
|
|
|
|
fd.write(content[:eol])
|
|
|
|
content = content[eol:]
|
|
|
|
mode = 'ab'
|
|
|
|
changed = True
|
|
|
|
if state.shutdown:
|
|
|
|
return False
|
|
|
|
if state.bandwidth:
|
|
|
|
while not state.bandwidth.download(chunk_size) and not state.shutdown:
|
|
|
|
time.sleep(0.1)
|
|
|
|
if content:
|
|
|
|
with open(path, mode) as fd:
|
|
|
|
fd.write(content)
|
|
|
|
changed = True
|
|
|
|
if changed:
|
|
|
|
peer.apply_log()
|
|
|
|
except:
|
|
|
|
logger.debug('download failed %s', url, exc_info=True)
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
logger.debug('FAILED %s', url)
|
2016-02-10 14:02:32 +00:00
|
|
|
return False
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2014-05-18 03:01:24 +00:00
|
|
|
def peering(self, action):
|
2019-02-02 07:13:37 +00:00
|
|
|
pull_changes = False
|
2016-01-12 07:32:39 +00:00
|
|
|
with db.session():
|
|
|
|
u = user.models.User.get_or_create(self.user_id)
|
2019-01-31 08:41:54 +00:00
|
|
|
user_info = u.info
|
|
|
|
if action in ('requestPeering', 'acceptPeering'):
|
|
|
|
r = self.request(action, settings.preferences['username'], user_info.get('message'))
|
|
|
|
else:
|
|
|
|
r = self.request(action, user_info.get('message'))
|
|
|
|
if r is not None:
|
|
|
|
with db.session():
|
|
|
|
u = user.models.User.get(self.user_id)
|
2016-01-12 07:32:39 +00:00
|
|
|
u.queued = False
|
|
|
|
if 'message' in u.info:
|
|
|
|
del u.info['message']
|
|
|
|
u.save()
|
2019-02-02 07:13:37 +00:00
|
|
|
if action == 'acceptPeering':
|
|
|
|
pull_changes = True
|
2019-01-31 08:41:54 +00:00
|
|
|
else:
|
|
|
|
logger.debug('peering failed? %s %s', action, r)
|
|
|
|
if action in ('cancelPeering', 'rejectPeering', 'removePeering'):
|
|
|
|
self.online = False
|
|
|
|
with db.session():
|
|
|
|
u = user.models.User.get(self.user_id)
|
|
|
|
trigger_event('peering.%s' % action.replace('Peering', ''), u.json())
|
2019-02-02 07:13:37 +00:00
|
|
|
if pull_changes:
|
|
|
|
self.pullChanges()
|
2019-01-31 08:41:54 +00:00
|
|
|
return True
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2016-01-15 07:28:01 +00:00
|
|
|
headers = {
|
|
|
|
'X-Node-Protocol': settings.NODE_PROTOCOL,
|
|
|
|
'User-Agent': settings.USER_AGENT,
|
|
|
|
'Accept-Encoding': 'gzip',
|
|
|
|
}
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def download(self, item):
|
2015-12-01 08:59:52 +00:00
|
|
|
self.resolve()
|
2014-05-04 17:26:43 +00:00
|
|
|
url = '%s/get/%s' % (self.url, item.id)
|
2016-02-23 09:07:06 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('download %s', url)
|
2016-01-15 07:28:01 +00:00
|
|
|
self._opener.addheaders = list(zip(self.headers.keys(), self.headers.values()))
|
2014-09-05 17:10:47 +00:00
|
|
|
try:
|
2016-01-18 15:45:59 +00:00
|
|
|
r = self._opener.open(url, timeout=self.TIMEOUT*5)
|
2016-02-12 09:00:59 +00:00
|
|
|
except socket.timeout:
|
|
|
|
logger.debug('timeout %s', url)
|
|
|
|
return False
|
2016-02-18 11:38:41 +00:00
|
|
|
except socks.GeneralProxyError:
|
|
|
|
logger.debug('openurl failed %s', url)
|
|
|
|
return False
|
2016-02-20 14:50:09 +00:00
|
|
|
except urllib.error.URLError as e:
|
|
|
|
logger.debug('openurl failed urllib2.URLError %s', e.reason)
|
|
|
|
return False
|
2016-02-20 18:25:06 +00:00
|
|
|
except:
|
|
|
|
logger.debug('openurl failed %s', url, exc_info=True)
|
|
|
|
return False
|
2014-05-14 09:57:11 +00:00
|
|
|
if r.getcode() == 200:
|
2014-09-05 17:10:47 +00:00
|
|
|
try:
|
|
|
|
if r.headers.get('content-encoding', None) == 'gzip':
|
2015-11-30 23:26:35 +00:00
|
|
|
fileobj = gzip.GzipFile(fileobj=r)
|
2016-02-25 07:19:53 +00:00
|
|
|
else:
|
|
|
|
fileobj = r
|
|
|
|
content = []
|
|
|
|
ct = time.time()
|
|
|
|
size = item.info['size']
|
|
|
|
received = 0
|
2016-01-18 15:45:59 +00:00
|
|
|
chunk_size = 16*1024
|
|
|
|
for chunk in iter(lambda: fileobj.read(chunk_size), b''):
|
2016-02-25 07:19:53 +00:00
|
|
|
content.append(chunk)
|
|
|
|
received += len(chunk)
|
|
|
|
if time.time() - ct > 1:
|
|
|
|
ct = time.time()
|
2016-02-11 15:55:41 +00:00
|
|
|
if state.shutdown:
|
|
|
|
return False
|
|
|
|
t = state.downloads.transfers.get(item.id)
|
2016-02-25 07:19:53 +00:00
|
|
|
if not t: # transfer was canceled
|
2016-01-20 07:35:44 +00:00
|
|
|
trigger_event('transfer', {
|
|
|
|
'id': item.id, 'progress': -1
|
|
|
|
})
|
2016-01-18 07:00:40 +00:00
|
|
|
return False
|
|
|
|
else:
|
2016-02-25 07:19:53 +00:00
|
|
|
t['progress'] = received / size
|
2016-01-18 07:00:40 +00:00
|
|
|
trigger_event('transfer', {
|
2016-02-11 15:55:41 +00:00
|
|
|
'id': item.id, 'progress': t['progress']
|
2016-01-18 07:00:40 +00:00
|
|
|
})
|
2016-02-11 15:55:41 +00:00
|
|
|
state.downloads.transfers[item.id] = t
|
2016-01-18 15:45:59 +00:00
|
|
|
if state.bandwidth:
|
2016-01-31 16:45:14 +00:00
|
|
|
while not state.bandwidth.download(chunk_size) and not state.shutdown:
|
2016-01-18 15:45:59 +00:00
|
|
|
time.sleep(0.1)
|
2016-02-25 07:19:53 +00:00
|
|
|
return item.save_file(b''.join(content))
|
2014-09-05 17:10:47 +00:00
|
|
|
except:
|
2016-01-24 09:13:03 +00:00
|
|
|
logger.debug('download failed %s', url, exc_info=True)
|
2014-09-05 17:10:47 +00:00
|
|
|
return False
|
2014-05-04 17:26:43 +00:00
|
|
|
else:
|
2014-05-17 14:26:59 +00:00
|
|
|
logger.debug('FAILED %s', url)
|
2014-05-04 17:26:43 +00:00
|
|
|
return False
|
|
|
|
|
2016-01-21 07:05:49 +00:00
|
|
|
def download_preview(self, item_id):
|
2016-01-15 07:28:01 +00:00
|
|
|
from item.icons import icons
|
|
|
|
self.resolve()
|
2016-02-23 09:07:06 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('download preview for %s from %s', item_id, self.url)
|
2016-01-21 07:05:49 +00:00
|
|
|
url = '%s/preview/%s' % (self.url, item_id)
|
2016-01-15 07:28:01 +00:00
|
|
|
self._opener.addheaders = list(zip(self.headers.keys(), self.headers.values()))
|
|
|
|
try:
|
|
|
|
r = self._opener.open(url, timeout=self.TIMEOUT*2)
|
2016-02-18 11:38:41 +00:00
|
|
|
except socket.timeout:
|
|
|
|
logger.debug('timeout %s', url)
|
|
|
|
return False
|
|
|
|
except socks.GeneralProxyError:
|
|
|
|
logger.debug('download failed %s', url)
|
|
|
|
return False
|
2016-01-15 07:28:01 +00:00
|
|
|
except:
|
2016-01-24 10:05:43 +00:00
|
|
|
logger.debug('download failed %s', url, exc_info=True)
|
2016-02-04 13:07:19 +00:00
|
|
|
self.online = False
|
2016-01-15 07:28:01 +00:00
|
|
|
return False
|
|
|
|
code = r.getcode()
|
|
|
|
if code == 200:
|
|
|
|
try:
|
|
|
|
if r.headers.get('content-encoding', None) == 'gzip':
|
|
|
|
fileobj = gzip.GzipFile(fileobj=r)
|
2016-02-25 07:19:53 +00:00
|
|
|
else:
|
|
|
|
fileobj = r
|
2016-01-15 07:28:01 +00:00
|
|
|
content = fileobj.read()
|
2016-01-21 07:05:49 +00:00
|
|
|
key = 'preview:' + item_id
|
2016-01-15 07:28:01 +00:00
|
|
|
icons[key] = content
|
2016-01-21 07:05:49 +00:00
|
|
|
icons.clear(key+':')
|
2016-01-15 07:28:01 +00:00
|
|
|
return True
|
|
|
|
except:
|
2016-01-24 09:13:03 +00:00
|
|
|
logger.debug('preview download failed %s', url, exc_info=True)
|
2016-01-15 07:28:01 +00:00
|
|
|
elif code == 404:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
logger.debug('FAILED %s', url)
|
2016-01-24 07:44:43 +00:00
|
|
|
return False
|
2016-01-15 07:28:01 +00:00
|
|
|
|
2014-05-14 09:57:11 +00:00
|
|
|
def download_upgrade(self, release):
|
|
|
|
for module in release['modules']:
|
|
|
|
path = os.path.join(settings.update_path, release['modules'][module]['name'])
|
2014-05-04 17:26:43 +00:00
|
|
|
if not os.path.exists(path):
|
2014-05-14 09:57:11 +00:00
|
|
|
url = '%s/oml/%s' % (self.url, release['modules'][module]['name'])
|
|
|
|
sha1 = release['modules'][module]['sha1']
|
2014-05-04 17:26:43 +00:00
|
|
|
headers = {
|
|
|
|
'User-Agent': settings.USER_AGENT,
|
|
|
|
}
|
2015-11-26 00:26:10 +00:00
|
|
|
self._opener.addheaders = list(zip(headers.keys(), headers.values()))
|
2014-05-14 09:57:11 +00:00
|
|
|
r = self._opener.open(url)
|
|
|
|
if r.getcode() == 200:
|
2014-05-04 17:26:43 +00:00
|
|
|
with open(path, 'w') as fd:
|
2014-05-14 09:57:11 +00:00
|
|
|
fd.write(r.read())
|
2014-05-04 17:26:43 +00:00
|
|
|
if (ox.sha1sum(path) != sha1):
|
2014-05-17 14:26:59 +00:00
|
|
|
logger.error('invalid update!')
|
2014-05-04 17:26:43 +00:00
|
|
|
os.unlink(path)
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
return False
|
|
|
|
|
2019-01-17 18:19:23 +00:00
|
|
|
def upload(self, items):
|
2019-01-28 10:09:00 +00:00
|
|
|
logger.debug('add items to %s\'s inbox: %s', self.user_id, items)
|
2019-01-17 18:19:23 +00:00
|
|
|
r = self.request('upload', items)
|
|
|
|
return bool(r)
|
|
|
|
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
class Nodes(Thread):
|
|
|
|
_nodes = {}
|
2016-03-14 13:31:56 +00:00
|
|
|
local = None
|
2016-01-23 16:49:34 +00:00
|
|
|
_pulling = False
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2014-08-09 16:33:59 +00:00
|
|
|
def __init__(self):
|
2014-05-04 17:26:43 +00:00
|
|
|
self._q = Queue()
|
2015-12-02 21:05:23 +00:00
|
|
|
with db.session():
|
|
|
|
for u in user.models.User.query.filter_by(peered=True):
|
|
|
|
self.queue('add', u.id)
|
2017-06-03 20:50:14 +00:00
|
|
|
get_peer(u.id)
|
2015-12-02 21:05:23 +00:00
|
|
|
for u in user.models.User.query.filter_by(queued=True):
|
|
|
|
logger.debug('adding queued node... %s', u.id)
|
2016-01-17 13:12:56 +00:00
|
|
|
self.queue('add', u.id, True)
|
2016-03-14 13:31:56 +00:00
|
|
|
self.local = LocalNodes()
|
2016-02-10 14:02:32 +00:00
|
|
|
self._pullcb = PeriodicCallback(self.pull, settings.server['pull_interval'])
|
2019-02-02 09:28:10 +00:00
|
|
|
state.main.add_callback(self._pullcb.start)
|
2014-05-04 17:26:43 +00:00
|
|
|
Thread.__init__(self)
|
|
|
|
self.daemon = True
|
|
|
|
self.start()
|
|
|
|
|
2016-02-25 07:19:53 +00:00
|
|
|
def run(self):
|
|
|
|
library.sync_db()
|
|
|
|
self.queue('pull')
|
|
|
|
while not state.shutdown:
|
|
|
|
args = self._q.get()
|
|
|
|
if args:
|
2019-02-02 18:00:29 +00:00
|
|
|
if DEBUG_NODES:
|
|
|
|
logger.debug('processing nodes queue: next: "%s", %s entries in queue', args[0], self._q.qsize())
|
2016-03-14 13:31:56 +00:00
|
|
|
if args[0] == 'add':
|
2016-02-25 07:19:53 +00:00
|
|
|
self._add(*args[1:])
|
|
|
|
elif args[0] == 'pull':
|
|
|
|
self._pull()
|
|
|
|
else:
|
|
|
|
self._call(*args)
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def queue(self, *args):
|
2019-02-02 18:00:29 +00:00
|
|
|
if args and DEBUG_NODES:
|
2019-02-02 07:13:37 +00:00
|
|
|
logger.debug('queue "%s", %s entries in queue', args, self._q.qsize())
|
2014-05-04 17:26:43 +00:00
|
|
|
self._q.put(list(args))
|
|
|
|
|
2019-02-02 07:13:37 +00:00
|
|
|
def peer_queue(self, peer, action, *args):
|
|
|
|
if peer not in self._nodes:
|
|
|
|
self._add(peer)
|
2019-02-02 12:06:11 +00:00
|
|
|
elif not self._nodes[peer].is_online():
|
|
|
|
self._nodes[peer].ping()
|
2019-02-02 07:13:37 +00:00
|
|
|
self._nodes[peer].queue(action, *args)
|
|
|
|
|
2014-05-21 22:41:29 +00:00
|
|
|
def is_online(self, id):
|
2015-12-01 08:59:52 +00:00
|
|
|
return id in self._nodes and self._nodes[id].is_online()
|
2014-05-04 17:26:43 +00:00
|
|
|
|
|
|
|
def download(self, id, item):
|
|
|
|
return id in self._nodes and self._nodes[id].download(item)
|
|
|
|
|
2016-01-21 07:05:49 +00:00
|
|
|
def download_preview(self, id, item):
|
|
|
|
return id in self._nodes and \
|
|
|
|
self._nodes[id].is_online() and \
|
|
|
|
self._nodes[id].download_preview(item)
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def _call(self, target, action, *args):
|
|
|
|
if target == 'all':
|
2014-09-02 22:32:44 +00:00
|
|
|
nodes = list(self._nodes.values())
|
2014-05-14 09:57:11 +00:00
|
|
|
elif target == 'peered':
|
2019-02-02 12:06:11 +00:00
|
|
|
ids = []
|
|
|
|
with db.session():
|
|
|
|
from user.models import User
|
|
|
|
for u in User.query.filter(User.id != settings.USER_ID).filter_by(peered=True).all():
|
|
|
|
ids.append(u.id)
|
|
|
|
nodes = [n for n in list(self._nodes.values()) if n.user_id in ids]
|
2014-05-04 17:26:43 +00:00
|
|
|
elif target == 'online':
|
2014-09-02 22:32:44 +00:00
|
|
|
nodes = [n for n in list(self._nodes.values()) if n.online]
|
2014-05-04 17:26:43 +00:00
|
|
|
else:
|
2019-02-02 07:13:37 +00:00
|
|
|
if target not in self._nodes:
|
2016-01-17 13:12:56 +00:00
|
|
|
self._add(target)
|
2014-05-04 17:26:43 +00:00
|
|
|
nodes = [self._nodes[target]]
|
|
|
|
for node in nodes:
|
2019-02-02 07:13:37 +00:00
|
|
|
node._call(action, *args)
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2016-01-17 13:12:56 +00:00
|
|
|
def _add(self, user_id, send_response=False):
|
2014-05-04 17:26:43 +00:00
|
|
|
if user_id not in self._nodes:
|
|
|
|
from user.models import User
|
2014-08-09 16:14:14 +00:00
|
|
|
with db.session():
|
2019-02-02 07:13:37 +00:00
|
|
|
User.get_or_create(user_id)
|
|
|
|
self._nodes[user_id] = Node(self, user_id)
|
2014-05-04 17:26:43 +00:00
|
|
|
else:
|
2016-02-14 05:53:55 +00:00
|
|
|
self._nodes[user_id].ping()
|
2016-01-17 13:12:56 +00:00
|
|
|
if send_response:
|
|
|
|
self._nodes[user_id].send_response()
|
2014-05-04 17:26:43 +00:00
|
|
|
|
2016-02-25 07:39:59 +00:00
|
|
|
def pull(self):
|
|
|
|
if not self._pulling:
|
|
|
|
self.queue('pull')
|
|
|
|
|
2016-01-23 16:49:34 +00:00
|
|
|
def _pull(self):
|
2019-01-13 07:21:38 +00:00
|
|
|
if not state.sync_enabled or settings.preferences.get('downloadRate') == 0:
|
|
|
|
return
|
2016-01-23 16:49:34 +00:00
|
|
|
if state.activity and state.activity.get('activity') == 'import':
|
|
|
|
return
|
|
|
|
self._pulling = True
|
2016-02-24 07:19:00 +00:00
|
|
|
if state.shutdown:
|
|
|
|
return
|
2016-02-08 09:03:21 +00:00
|
|
|
users = []
|
|
|
|
with db.session():
|
|
|
|
from user.models import User
|
2019-02-01 09:26:33 +00:00
|
|
|
for u in User.query.filter(User.id != settings.USER_ID).filter_by(peered=True).all():
|
2016-02-25 07:39:59 +00:00
|
|
|
users.append(u.json(['id', 'index', 'name']))
|
2016-02-08 09:03:21 +00:00
|
|
|
users.sort(key=user_sort_key)
|
|
|
|
for u in users:
|
|
|
|
if state.shutdown:
|
|
|
|
break
|
|
|
|
node = self._nodes.get(u['id'])
|
2019-02-02 08:51:25 +00:00
|
|
|
if node and node.is_online():
|
2016-02-24 07:19:00 +00:00
|
|
|
node.pullChanges()
|
2016-01-23 16:49:34 +00:00
|
|
|
self._pulling = False
|
|
|
|
|
2014-05-04 17:26:43 +00:00
|
|
|
def join(self):
|
|
|
|
self._q.put(None)
|
2014-09-02 22:32:44 +00:00
|
|
|
for node in list(self._nodes.values()):
|
2014-05-19 15:00:33 +00:00
|
|
|
node.join()
|
2016-03-14 13:31:56 +00:00
|
|
|
if self.local:
|
|
|
|
self.local.close()
|
2019-01-12 17:41:14 +00:00
|
|
|
return super().join(1)
|
2015-11-26 00:26:10 +00:00
|
|
|
|
|
|
|
def publish_node():
|
|
|
|
update_online()
|
|
|
|
state._online = PeriodicCallback(update_online, 60000)
|
|
|
|
state._online.start()
|
|
|
|
|
|
|
|
def update_online():
|
|
|
|
online = state.tor and state.tor.is_online()
|
|
|
|
if online != state.online:
|
|
|
|
state.online = online
|
|
|
|
trigger_event('status', {
|
|
|
|
'id': settings.USER_ID,
|
|
|
|
'online': state.online
|
|
|
|
})
|
2016-02-13 09:27:55 +00:00
|
|
|
if state.online:
|
|
|
|
for node in list(state.nodes._nodes.values()):
|
|
|
|
node.trigger_status()
|