cache tpb search only for 24 hours
This commit is contained in:
parent
30afb98645
commit
5e224d3cdb
2 changed files with 5 additions and 4 deletions
2
README
2
README
|
@ -2,7 +2,7 @@ python-oxweb the internet is a dict
|
||||||
|
|
||||||
Depends:
|
Depends:
|
||||||
python2.5
|
python2.5
|
||||||
python-oxutils
|
python-oxutils (bzr branch http//code.0xdb.org/python-oxutils)
|
||||||
python-beautifulsoup (http://www.crummy.com/software/BeautifulSoup/)
|
python-beautifulsoup (http://www.crummy.com/software/BeautifulSoup/)
|
||||||
python-feedparser (http://www.feedparser.org/)
|
python-feedparser (http://www.feedparser.org/)
|
||||||
(there seam to be some issues if not using the one from ubuntu/debian)
|
(there seam to be some issues if not using the one from ubuntu/debian)
|
||||||
|
|
|
@ -13,6 +13,7 @@ import oxlib
|
||||||
|
|
||||||
from torrent import Torrent
|
from torrent import Torrent
|
||||||
|
|
||||||
|
cache_timeout = 24*60*60 # cache search only for 24 hours
|
||||||
|
|
||||||
season_episode = re.compile("S..E..", re.IGNORECASE)
|
season_episode = re.compile("S..E..", re.IGNORECASE)
|
||||||
|
|
||||||
|
@ -22,8 +23,8 @@ def _getUrl(url, data=None, headers=cache.DEFAULT_HEADERS, timeout=cache.cache_t
|
||||||
headers['Cookie'] = 'language=en_EN'
|
headers['Cookie'] = 'language=en_EN'
|
||||||
return cache.getUrl(url, data, headers, timeout)
|
return cache.getUrl(url, data, headers, timeout)
|
||||||
|
|
||||||
def _getUrlUnicode(url):
|
def _getUrlUnicode(url, timeout=cache.cache_timeout):
|
||||||
return cache.getUrlUnicode(url, _getUrl=_getUrl)
|
return cache.getUrlUnicode(url, _getUrl=_getUrl, timeout=timeout)
|
||||||
|
|
||||||
def findMovies(query, max_results=10):
|
def findMovies(query, max_results=10):
|
||||||
results = []
|
results = []
|
||||||
|
@ -36,7 +37,7 @@ def findMovies(query, max_results=10):
|
||||||
if not url.startswith('/'):
|
if not url.startswith('/'):
|
||||||
url = "/" + url
|
url = "/" + url
|
||||||
url = "http://thepiratebay.org" + url
|
url = "http://thepiratebay.org" + url
|
||||||
data = _getUrlUnicode(url)
|
data = _getUrlUnicode(url, timeout=cache_timeout)
|
||||||
regexp = '''<tr.*?<td class="vertTh"><a href="/browse/(.*?)".*?<td><a href="(/torrent/.*?)" class="detLink".*?>(.*?)</a>.*?</tr>'''
|
regexp = '''<tr.*?<td class="vertTh"><a href="/browse/(.*?)".*?<td><a href="(/torrent/.*?)" class="detLink".*?>(.*?)</a>.*?</tr>'''
|
||||||
for row in re.compile(regexp, re.DOTALL).findall(data):
|
for row in re.compile(regexp, re.DOTALL).findall(data):
|
||||||
torrentType = row[0]
|
torrentType = row[0]
|
||||||
|
|
Loading…
Reference in a new issue