remove unused scrape table
This commit is contained in:
parent
5f8094bba3
commit
97fe8731b8
6 changed files with 1 additions and 93 deletions
|
@ -67,46 +67,3 @@ class Downloads(Thread):
|
||||||
time.sleep(step)
|
time.sleep(step)
|
||||||
timeout -= step
|
timeout -= step
|
||||||
|
|
||||||
class ScrapeThread(Thread):
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
self._running = True
|
|
||||||
Thread.__init__(self)
|
|
||||||
self.daemon = True
|
|
||||||
self.start()
|
|
||||||
|
|
||||||
def scrape_queue(self):
|
|
||||||
import item.models
|
|
||||||
scraped = False
|
|
||||||
for s in item.models.Scrape.query.filter(
|
|
||||||
item.models.Scrape.added!=None,
|
|
||||||
).order_by(item.models.Scrape.added):
|
|
||||||
if not self._running:
|
|
||||||
return True
|
|
||||||
logger.debug('scrape %s', s.item)
|
|
||||||
try:
|
|
||||||
if s.item.scrape():
|
|
||||||
for f in s.item.files:
|
|
||||||
f.move()
|
|
||||||
s.item.update_icons()
|
|
||||||
s.item.save()
|
|
||||||
s.remove()
|
|
||||||
trigger_event('change', {})
|
|
||||||
scraped = True
|
|
||||||
except:
|
|
||||||
logger.debug('scrape failed %s', s.item, exc_info=1)
|
|
||||||
return scraped
|
|
||||||
|
|
||||||
def run(self):
|
|
||||||
time.sleep(2)
|
|
||||||
while self._running:
|
|
||||||
wait = False
|
|
||||||
with db.session():
|
|
||||||
if not self.scrape_queue():
|
|
||||||
wait = True
|
|
||||||
if wait:
|
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
def join(self):
|
|
||||||
self._running = False
|
|
||||||
return Thread.join(self)
|
|
||||||
|
|
|
@ -264,7 +264,6 @@ class Item(db.Model):
|
||||||
def delete(self, commit=True):
|
def delete(self, commit=True):
|
||||||
Sort.query.filter_by(item_id=self.id).delete()
|
Sort.query.filter_by(item_id=self.id).delete()
|
||||||
Transfer.query.filter_by(item_id=self.id).delete()
|
Transfer.query.filter_by(item_id=self.id).delete()
|
||||||
Scrape.query.filter_by(item_id=self.id).delete()
|
|
||||||
state.db.session.delete(self)
|
state.db.session.delete(self)
|
||||||
icons.clear('cover:%s' % self.id)
|
icons.clear('cover:%s' % self.id)
|
||||||
icons.clear('preview:%s' % self.id)
|
icons.clear('preview:%s' % self.id)
|
||||||
|
@ -314,7 +313,6 @@ class Item(db.Model):
|
||||||
self.sync_metadata()
|
self.sync_metadata()
|
||||||
|
|
||||||
def edit(self, data, modified=None):
|
def edit(self, data, modified=None):
|
||||||
Scrape.query.filter_by(item_id=self.id).delete()
|
|
||||||
self.update_metadata(data, modified)
|
self.update_metadata(data, modified)
|
||||||
for f in self.files.all():
|
for f in self.files.all():
|
||||||
f.move()
|
f.move()
|
||||||
|
@ -710,38 +708,6 @@ class File(db.Model):
|
||||||
state.db.session.add(self)
|
state.db.session.add(self)
|
||||||
state.db.session.commit()
|
state.db.session.commit()
|
||||||
|
|
||||||
class Scrape(db.Model):
|
|
||||||
|
|
||||||
__tablename__ = 'scrape'
|
|
||||||
|
|
||||||
item_id = sa.Column(sa.String(32), sa.ForeignKey('item.id'), primary_key=True)
|
|
||||||
item = sa.orm.relationship('Item', backref=sa.orm.backref('scraping', lazy='dynamic'))
|
|
||||||
|
|
||||||
added = sa.Column(sa.DateTime())
|
|
||||||
|
|
||||||
def __repr__(self):
|
|
||||||
return '='.join(map(str, [self.item_id, self.added]))
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def get(cls, item_id):
|
|
||||||
return cls.query.filter_by(item_id=item_id).first()
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def get_or_create(cls, item_id):
|
|
||||||
t = cls.get(item_id)
|
|
||||||
if not t:
|
|
||||||
t = cls(item_id=item_id)
|
|
||||||
t.added = datetime.utcnow()
|
|
||||||
t.save()
|
|
||||||
return t
|
|
||||||
|
|
||||||
def save(self):
|
|
||||||
state.db.session.add(self)
|
|
||||||
state.db.session.commit()
|
|
||||||
|
|
||||||
def remove(self):
|
|
||||||
state.db.session.delete(self)
|
|
||||||
state.db.session.commit()
|
|
||||||
|
|
||||||
class Transfer(db.Model):
|
class Transfer(db.Model):
|
||||||
__tablename__ = 'transfer'
|
__tablename__ = 'transfer'
|
||||||
|
|
|
@ -65,9 +65,6 @@ def shutdown():
|
||||||
if state.downloads:
|
if state.downloads:
|
||||||
logger.debug('shutdown downloads')
|
logger.debug('shutdown downloads')
|
||||||
state.downloads.join()
|
state.downloads.join()
|
||||||
if state.scraping:
|
|
||||||
logger.debug('shutdown scraping')
|
|
||||||
state.scraping.join()
|
|
||||||
logger.debug('shutdown http_server')
|
logger.debug('shutdown http_server')
|
||||||
state.http_server.stop()
|
state.http_server.stop()
|
||||||
if state.tasks:
|
if state.tasks:
|
||||||
|
@ -152,7 +149,6 @@ def run():
|
||||||
state.tor = tor.Tor()
|
state.tor = tor.Tor()
|
||||||
state.node = node.server.start()
|
state.node = node.server.start()
|
||||||
state.downloads = downloads.Downloads()
|
state.downloads = downloads.Downloads()
|
||||||
#state.scraping = downloads.ScrapeThread()
|
|
||||||
state.nodes = nodes.Nodes()
|
state.nodes = nodes.Nodes()
|
||||||
def publish():
|
def publish():
|
||||||
if not state.tor.is_online():
|
if not state.tor.is_online():
|
||||||
|
|
11
oml/setup.py
11
oml/setup.py
|
@ -345,15 +345,6 @@ def upgrade_db(old, new=None):
|
||||||
if old <= '20140527-120-3cb9819':
|
if old <= '20140527-120-3cb9819':
|
||||||
run_sql('CREATE INDEX ix_find_findvalue ON find (findvalue)')
|
run_sql('CREATE INDEX ix_find_findvalue ON find (findvalue)')
|
||||||
|
|
||||||
if old <= '20150307-272-557f4d3':
|
|
||||||
if not db.table_exists('scrape'):
|
|
||||||
run_sql('''CREATE TABLE scrape (
|
|
||||||
item_id VARCHAR(32) NOT NULL,
|
|
||||||
added DATETIME,
|
|
||||||
PRIMARY KEY (item_id),
|
|
||||||
FOREIGN KEY(item_id) REFERENCES item (id)
|
|
||||||
)''')
|
|
||||||
run_sql('CREATE INDEX idx_scrape_added ON scrape (added)')
|
|
||||||
if old <= '20151118-346-7e86e68':
|
if old <= '20151118-346-7e86e68':
|
||||||
old_key = os.path.join(settings.data_path, 'node.ssl.key')
|
old_key = os.path.join(settings.data_path, 'node.ssl.key')
|
||||||
if os.path.exists(old_key):
|
if os.path.exists(old_key):
|
||||||
|
@ -446,7 +437,6 @@ def update_database():
|
||||||
item.models.File,
|
item.models.File,
|
||||||
item.models.Find,
|
item.models.Find,
|
||||||
item.models.Item,
|
item.models.Item,
|
||||||
item.models.Scrape,
|
|
||||||
item.models.Sort,
|
item.models.Sort,
|
||||||
item.models.Transfer,
|
item.models.Transfer,
|
||||||
item.models.user_items,
|
item.models.user_items,
|
||||||
|
@ -458,7 +448,6 @@ def update_database():
|
||||||
user.models.User,
|
user.models.User,
|
||||||
]
|
]
|
||||||
indexes = [
|
indexes = [
|
||||||
'CREATE INDEX idx_scrape_added ON scrape (added)',
|
|
||||||
'CREATE INDEX ix_find_findvalue ON find (findvalue)',
|
'CREATE INDEX ix_find_findvalue ON find (findvalue)',
|
||||||
'CREATE INDEX ix_find_key ON find ("key")',
|
'CREATE INDEX ix_find_key ON find ("key")',
|
||||||
'CREATE INDEX ix_useritem_user ON useritem ("user_id")',
|
'CREATE INDEX ix_useritem_user ON useritem ("user_id")',
|
||||||
|
|
|
@ -5,7 +5,6 @@ nodes = False
|
||||||
node = False
|
node = False
|
||||||
online = False
|
online = False
|
||||||
tasks = False
|
tasks = False
|
||||||
scraping = False
|
|
||||||
downloads = False
|
downloads = False
|
||||||
tor = False
|
tor = False
|
||||||
update = False
|
update = False
|
||||||
|
|
|
@ -417,5 +417,6 @@ def migrate_7():
|
||||||
session.delete(c)
|
session.delete(c)
|
||||||
session.commit()
|
session.commit()
|
||||||
db.run_sql('DROP TABLE IF EXISTS metadata')
|
db.run_sql('DROP TABLE IF EXISTS metadata')
|
||||||
|
db.run_sql('DROP TABLE IF EXISTS scrape')
|
||||||
db.run_sql('VACUUM')
|
db.run_sql('VACUUM')
|
||||||
return 7
|
return 7
|
||||||
|
|
Loading…
Reference in a new issue