Browse Source
Used RSS, Passkey - No need for login (although the site is a private tracker)pull/2722/head
2 changed files with 119 additions and 0 deletions
@ -0,0 +1,49 @@ |
|||||
|
from .main import DownRev |
||||
|
|
||||
|
def start(): |
||||
|
return DownRev() |
||||
|
|
||||
|
config = [{ |
||||
|
'name': 'downrev', |
||||
|
'groups': [ |
||||
|
{ |
||||
|
'tab': 'searcher', |
||||
|
'list': 'torrent_providers', |
||||
|
'name': 'DownRev', |
||||
|
'description': 'See <a href="http://www.downrev.net">DownRev</a>', |
||||
|
'options': [ |
||||
|
{ |
||||
|
'name': 'enabled', |
||||
|
'type': 'enabler', |
||||
|
'default': False, |
||||
|
}, |
||||
|
{ |
||||
|
'name': 'passkey', |
||||
|
'default': '', |
||||
|
}, |
||||
|
{ |
||||
|
'name': 'seed_ratio', |
||||
|
'label': 'Seed ratio', |
||||
|
'type': 'float', |
||||
|
'default': 1, |
||||
|
'description': 'Will not be (re)moved until this seed ratio is met.', |
||||
|
}, |
||||
|
{ |
||||
|
'name': 'seed_time', |
||||
|
'label': 'Seed time', |
||||
|
'type': 'int', |
||||
|
'default': 40, |
||||
|
'description': 'Will not be (re)moved until this seed time (in hours) is met.', |
||||
|
}, |
||||
|
{ |
||||
|
'name': 'extra_score', |
||||
|
'advanced': True, |
||||
|
'label': 'Extra Score', |
||||
|
'type': 'int', |
||||
|
'default': 0, |
||||
|
'description': 'Starting score for each release found via this provider.', |
||||
|
}, |
||||
|
], |
||||
|
}, |
||||
|
], |
||||
|
}] |
@ -0,0 +1,70 @@ |
|||||
|
from bs4 import BeautifulSoup |
||||
|
from couchpotato.core.helpers.rss import RSS |
||||
|
from couchpotato.core.helpers.encoding import tryUrlencode |
||||
|
from couchpotato.core.helpers.variable import tryInt, possibleTitles, getTitle |
||||
|
from couchpotato.core.logger import CPLog |
||||
|
from couchpotato.core.providers.torrent.base import TorrentProvider |
||||
|
from urlparse import urlparse, parse_qs |
||||
|
import re |
||||
|
import traceback |
||||
|
|
||||
|
log = CPLog(__name__) |
||||
|
|
||||
|
|
||||
|
class DownRev(TorrentProvider, RSS): |
||||
|
|
||||
|
urls = { |
||||
|
'test' : 'https://www.downrev.net/', |
||||
|
'login' : 'https://www.downrev.net/takelogin.php', |
||||
|
'detail' : 'https://www.downrev.net/torrent/%s', |
||||
|
'search' : 'https://www.downrev.net/rss2.php?cats=%d&type=dl&passkey=%s&like=%s', |
||||
|
'download' : 'https://www.downrev.net/down.php?id=%s&passkey=%s', |
||||
|
'login_check': 'http://www.downrev.net/inbox', |
||||
|
} |
||||
|
|
||||
|
cat_ids = [ |
||||
|
([68], ['1080p']), |
||||
|
([45], ['720p']), |
||||
|
([36], ['dvdrip', 'scr', 'r5']), |
||||
|
([65], ['brrip']), |
||||
|
([34], ['dvdr']), |
||||
|
] |
||||
|
|
||||
|
http_time_between_calls = 1 #seconds |
||||
|
|
||||
|
def _search(self, movie, quality, results): |
||||
|
|
||||
|
url = self.urls['search'] % (self.getCatId(quality['identifier'])[0], self.conf('passkey') , possibleTitles(getTitle(movie['library']))[0].replace(' ', '.')) |
||||
|
data = self.getRSSData(url, opener = self.login_opener) |
||||
|
|
||||
|
if data: |
||||
|
try: |
||||
|
for result in data: |
||||
|
title = self.getTextElement(result, "title") |
||||
|
desc = self.getTextElement(result, "description") |
||||
|
link = self.getTextElement(result, "link") |
||||
|
|
||||
|
# Extract from link |
||||
|
o = urlparse(link) |
||||
|
ID = parse_qs(o.query)['id'][0] |
||||
|
|
||||
|
p = re.compile(r'\W*Size[^: ]*:\s*(\d+.\d+\s\w+)\D*Leechers[^: ]*:\s(\d+)\D*Seeders[^: ]*:\s(\d+)') |
||||
|
m2 = p.findall(desc.replace('/', '').replace('<br/>', '')) |
||||
|
|
||||
|
size = m2[0][0] |
||||
|
leechers = m2[0][1] |
||||
|
seeders = m2[0][2] |
||||
|
|
||||
|
results.append({ |
||||
|
'id': ID, |
||||
|
'name': title, |
||||
|
'url': self.urls['download'] % (ID, self.conf('passkey')), |
||||
|
'detail_url': self.urls['detail'] % ID, |
||||
|
'size': self.parseSize(size), |
||||
|
'seeders': tryInt(seeders), |
||||
|
'leechers': tryInt(leechers) |
||||
|
}) |
||||
|
|
||||
|
except: |
||||
|
log.error('Failed getting results from %s: %s', (self.getName(), traceback.format_exc())) |
||||
|
|
Loading…
Reference in new issue