本文整理汇总了Python中resources.lib.modules.common.random_agent函数的典型用法代码示例。如果您正苦于以下问题:Python random_agent函数的具体用法?Python random_agent怎么用?Python random_agent使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。
在下文中一共展示了random_agent函数的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。
示例1: episode
def episode(self, url, imdb, tvdb, title, premiered, season, episode):
try:
headers = {'User-Agent': random_agent()}
data = urlparse.parse_qs(url)
data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])
title = data['tvshowtitle'] if 'tvshowtitle' in data else data['title']
data['season'], data['episode'] = season, episode
year = data['year']
cleanmovie = cleantitle.get(title)
title = cleantitle.getsearch(title)
query = self.search_link % (urllib.quote_plus(title))
query = urlparse.urljoin(self.base_link, query)
seasoncheck = "s%02d" % int(season)
html = requests.get(query, headers=headers, timeout=15).content
containers = re.compile('<span class="year">(.+?)</span><a class="play" href="(.+?)" title="(.+?)">').findall(html)
for r_year, r_href, r_title in containers:
if cleanmovie in cleantitle.get(r_title):
if seasoncheck in cleantitle.get(r_title):
if year == r_year:
url = r_href.encode('utf-8') + "?p=" + episode + "&s=11"
print ("SOCKSHARE PASSED EPISODE", url)
return url
except:
pass
示例2: movie
def movie(self, imdb, title, year):
self.elysium_url = []
try:
if not debridstatus == 'true': raise Exception()
self.elysium_url = []
headers = {'Accept-Language': 'en-US,en;q=0.5', 'User-Agent': random_agent()}
cleanmovie = cleantitle.get(title)
title = cleantitle.getsearch(title)
titlecheck = cleanmovie+year
query = self.search_link % (urllib.quote_plus(title), ep_search)
query = urlparse.urljoin(self.base_link, query)
html = BeautifulSoup(requests.get(query, headers=headers, timeout=10).content)
containers = html.findAll('h2', attrs={'class': 'title'})
for result in containers:
r_title = result.findAll('a')[0]
r_title = r_title.string
r_href = result.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
r_title = r_title.encode('utf-8')
c_title = cleantitle_get_2(r_title)
if titlecheck in c_title:
self.elysium_url.append([r_href,r_title])
return self.elysium_url
except:
return
示例3: sources
def sources(self, url, hostDict, hostprDict):
try:
sources = []
headers = {'User-Agent': random_agent()}
if url == None: return sources
# print("ANIMETOON SOURCES", url)
r = BeautifulSoup(requests.get(url, headers=headers).content)
r = r.findAll('iframe')
# print ("ANIMETOON s1", r)
for u in r:
try:
u = u['src'].encode('utf-8')
# print ("ANIMETOON s2", u)
html = requests.get(u, headers=headers).text
r_src = re.compile("url:\s*'(.+?)'").findall(html)
for src in r_src:
# print ("ANIMETOON s3", src)
vid_url = src.encode('utf-8')
sources.append({'source': 'cdn', 'quality': 'SD', 'provider': 'Animetoon', 'url': vid_url, 'direct': True, 'debridonly': False})
except:
pass
return sources
except:
return sources
示例4: movie
def movie(self, imdb, title, year):
self.zen_url = []
try:
headers = {'User-Agent': random_agent()}
query = self.search_link % (urllib.quote_plus(title),year)
query = urlparse.urljoin(self.base_link, query)
cleaned_title = cleantitle.get(title)
html = BeautifulSoup(requests.get(query, headers=headers, timeout=30).content)
containers = html.findAll('div', attrs={'class': 'name'})
for container in containers:
# print ("MOVIEXK r1", container)
r_href = container.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
# print ("MOVIEXK r2", r_href)
r_title = re.findall('</span>(.*?)</a>', str(container))[0]
# print ("MOVIEXK r3", r_title)
r_title = r_title.encode('utf-8')
# print ("MOVIEXK RESULTS", r_title, r_href)
if year in r_title:
if cleaned_title == cleantitle.get(r_title):
redirect = requests.get(r_href, headers=headers, timeout=30).text
try:
r_url_trailer = re.search('<dd>[Tt]railer</dd>', redirect)
if r_url_trailer: continue
except:
pass
r_url = re.findall('<a href="(.*?)" class="btn-watch"',redirect)[0]
r_url = r_url.encode('utf-8')
print ("MOVIEXK PLAY URL", r_url)
self.zen_url.append(r_url)
return self.zen_url
except:
return
示例5: sources
def sources(self, url, hostDict, hostprDict):
try:
sources = []
headers = {'Accept-Language': 'en-US,en;q=0.5', 'User-Agent': random_agent()}
for movielink,title in self.genesisreborn_url:
quality = quality_tag(title)
html = BeautifulSoup(requests.get(movielink, headers=headers, timeout=10).content)
containers = html.findAll('div', attrs={'class': 'txt-block'})
for result in containers:
print("THREEMOVIES LINKS ",result)
links = result.findAll('a')
for r_href in links:
url = r_href['href']
myurl = str(url)
if any (value in myurl for value in hostprDict):
try:host = re.findall('([\w]+[.][\w]+)$', urlparse.urlparse(url.strip().lower()).netloc)[0]
except: host = 'Threemovies'
url = client.replaceHTMLCodes(url)
url = url.encode('utf-8')
host = client.replaceHTMLCodes(host)
host = host.encode('utf-8')
sources.append({'source': host, 'quality': quality, 'provider': 'Threemovies', 'url': url, 'direct': False, 'debridonly': True})
return sources
except:
return sources
示例6: episode
def episode(self, url, imdb, tvdb, title, premiered, season, episode):
try:
self.elysium_url = []
if not debridstatus == 'true': raise Exception()
headers = {'Accept-Language': 'en-US,en;q=0.5', 'User-Agent': random_agent()}
data = urlparse.parse_qs(url)
data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])
title = data['tvshowtitle'] if 'tvshowtitle' in data else data['title']
year = data['year']
cleanmovie = cleantitle.get(title)
data['season'], data['episode'] = season, episode
ep_search = 'S%02dE%02d' % (int(data['season']), int(data['episode']))
episodecheck = str(ep_search).lower()
titlecheck = cleanmovie+episodecheck
query = self.search_link % (urllib.quote_plus(title), ep_search)
query = urlparse.urljoin(self.base_link, query)
print("HEVC query", query)
html = BeautifulSoup(rq.get(query, headers=headers, timeout=10).content)
containers = html.findAll('div', attrs={'class': 'postcontent'})
for result in containers:
print("HEVC containers", result)
r_title = result.findAll('a')[0]["title"]
r_href = result.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
r_title = r_title.encode('utf-8')
check = cleantitle.get(r_title)
if titlecheck in check:
self.elysium_url.append([r_href,r_title])
print("HEVC PASSED MOVIE ", r_title, r_href)
return self.url
except:
return
示例7: episode
def episode(self, url, imdb, tvdb, title, premiered, season, episode):
self.zen_url = []
try:
headers = {'User-Agent': random_agent()}
data = urlparse.parse_qs(url)
data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])
title = data['tvshowtitle'] if 'tvshowtitle' in data else data['title']
data['season'], data['episode'] = season, episode
self.zen_url = []
title = cleantitle.getsearch(title)
title = title.replace(' ','-')
query = title + "-" + season + "x" + episode
query= self.ep_link % query
query = urlparse.urljoin(self.base_link, query)
r = BeautifulSoup(requests.get(query, headers=headers, timeout=10).content)
r = r.findAll('iframe')
# print ("ANIMETOON s1", r)
for u in r:
u = u['src'].encode('utf-8')
if u.startswith("//"): u = "http:" + u
self.zen_url.append(u)
return self.zen_url
except:
return
示例8: sources
def sources(self, url, hostDict, hostprDict):
try:
sources = []
headers = {'User-Agent': random_agent()}
if url == None: return sources
url = urlparse.urljoin(self.base_link, url)
r = BeautifulSoup(requests.get(url, headers=headers).content)
r = r.findAll('iframe')
# print ("GOGOANIME s1", r)
for u in r:
try:
u = u['src'].encode('utf-8')
# print ("GOGOANIME s2", u)
if not 'vidstreaming' in u: raise Exception()
html = BeautifulSoup(requests.get(u, headers=headers).content)
r_src = html.findAll('source')
for src in r_src:
vid_url = src['src'].encode('utf-8')
try: sources.append({'source': 'gvideo', 'quality': directstream.googletag(vid_url)[0]['quality'], 'provider': 'Gogoanime', 'url': vid_url, 'direct': True, 'debridonly': False})
except: pass
except:
pass
return sources
except:
return sources
示例9: sources
def sources(self, url, hostDict, hostprDict):
try:
sources = []
if url == None: return sources
referer = url
headers = {'User-Agent': random_agent(), 'X-Requested-With': 'XMLHttpRequest', 'Referer': referer}
url_plugin = urlparse.urljoin(self.base_link, '/ip.file/swf/plugins/ipplugins.php')
html = BeautifulSoup(requests.get(referer, headers=headers, timeout=15).content)
# print ("SOCKSHARE NEW SOURCES", html)
r = html.findAll('div', attrs={'class': 'new_player'})
for container in r:
block = container.findAll('a')
for items in block:
p1 = items['data-film'].encode('utf-8')
p2 = items['data-name'].encode('utf-8')
p3 = items['data-server'].encode('utf-8')
post = {'ipplugins': '1', 'ip_film': p1, 'ip_name': p2 , 'ip_server': p3}
req = requests.post(url_plugin, data=post, headers=headers).json()
token = req['s'].encode('utf-8')
server = req['v'].encode('utf-8')
url = urlparse.urljoin(self.base_link, '/ip.file/swf/ipplayer/ipplayer.php')
post = {'u': token, 'w': '100%', 'h': '360' , 's': server, 'n':'0'}
req_player = requests.post(url, data=post, headers=headers).json()
# print ("SOCKSHARE SOURCES", req_player)
result = req_player['data']
result = [i['files'] for i in result]
for i in result:
try: sources.append({'source': 'gvideo', 'quality': directstream.googletag(i)[0]['quality'], 'provider': 'Sockshare', 'url': i, 'direct': True, 'debridonly': False})
except: pass
return sources
except:
return sources
示例10: movie
def movie(self, imdb, title, year):
try:
self.elysium_url = []
if not debridstatus == 'true': raise Exception()
headers = {'User-Agent': random_agent()}
cleanmovie = cleantitle.get(title)
title = cleantitle.getsearch(title)
titlecheck = cleanmovie+year
query = self.search_link % (urllib.quote_plus(title), year)
query = urlparse.urljoin(self.base_link, query)
print("HEVC query", query)
html = BeautifulSoup(rq.get(query, headers=headers, timeout=10).content)
containers = html.findAll('div', attrs={'class': 'postcontent'})
for result in containers:
print("HEVC containers", result)
r_title = result.findAll('a')[0]["title"]
r_href = result.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
r_title = r_title.encode('utf-8')
c_title = cleantitle.get(r_title)
if year in r_title and cleanmovie in c_title:
self.elysium_url.append([r_href,r_title])
print("HEVC PASSED MOVIE ", r_title, r_href)
return self.elysium_url
except:
return
示例11: sources
def sources(self, url, hostDict, hostprDict):
try:
sources = []
for movielink,title in self.elysium_url:
headers = {'User-Agent': random_agent()}
html = BeautifulSoup(requests.get(movielink, headers=headers, timeout=15).content)
result = html.findAll('div', attrs={'class': 'comm_content'})[:3]
for r in result:
r_href = r.findAll('a')
for items in r_href:
url = items['href'].encode('utf-8')
if "1080" in url: quality = "1080p"
elif "720" in url: quality = "HD"
else: quality = "SD"
info = ''
if "hevc" in url.lower(): info = "HEVC"
if not any(value in url for value in ['sample','uploadkadeh','wordpress','crazy4tv','imdb.com','youtube','userboard','kumpulbagi','mexashare','myvideolink.xyz', 'myvideolinks.xyz' , 'costaction', 'crazydl','.rar', '.RAR', 'ul.to', 'safelinking','linx.2ddl.ag','upload.so','.zip', 'go4up', 'adf.ly','.jpg','.jpeg']):
if any(value in url for value in hostprDict):
try:host = re.findall('([\w]+[.][\w]+)$', urlparse.urlparse(url.strip().lower()).netloc)[0]
except: host = 'Videomega'
url = client.replaceHTMLCodes(url)
url = url.encode('utf-8')
sources.append({'source': host, 'quality': quality, 'provider': 'Scnsrc', 'url': url, 'info': info,'direct': False, 'debridonly': True})
return sources
except:
return sources
示例12: sources
def sources(self, url, hostDict, hostprDict):
sources = []
try:
headers = {'User-Agent': random_agent()}
for url in self.genesisreborn_url:
if url == None: return
html = requests.get(url, headers=headers, timeout=10).text
match = re.compile('<a href="[^"]+go.php\?url=([^"]+)" target="_blank">').findall(html)
for url in match:
try:
# print("SOLAR SOURCE", url)
host = re.findall('([\w]+[.][\w]+)$', urlparse.urlparse(url.strip().lower()).netloc)[0]
host = host.encode('utf-8')
if not host in hostDict: raise Exception()
quality = "SD"
# print("OpenMovies SOURCE", stream_url, label)
sources.append({'source': host, 'quality':quality, 'provider': 'Solar', 'url': url, 'direct': False, 'debridonly': False})
except:
pass
return sources
except:
return sources
示例13: movie
def movie(self, imdb, title, year):
self.elysium_url = []
try:
if not alluc_status == 'true': raise Exception()
print ("ALLUC STARTED" , alluc_user, alluc_pw, max_items)
headers = {'User-Agent': random_agent()}
search_title = cleantitle.getsearch(title)
cleanmovie = cleantitle.get(title) + year
query = "%s+%s" % (urllib.quote_plus(search_title),year)
print ("ALLUC r1", query)
query = self.api_link % (alluc_user, alluc_pw, query)
if alluc_debrid == 'true': query = query + max_result_string
else: query = query + '+%23newlinks' + max_result_string
print ("ALLUC r2", query)
html = requests.get(query, headers=headers, timeout=15).json()
for result in html['result']:
if len(result['hosterurls']) > 1: continue
if result['extension'] == 'rar': continue
stream_url = result['hosterurls'][0]['url'].encode('utf-8')
stream_title = result['title'].encode('utf-8')
stream_title = cleantitle.getsearch(stream_title)
if cleanmovie in cleantitle.get(stream_title):
self.elysium_url.append([stream_url,stream_title])
print ("ALLUC r3", self.elysium_url)
return self.elysium_url
except:
return
示例14: movie
def movie(self, imdb, title, year):
self.elysium_url = []
try:
if not debridstatus == 'true': raise Exception()
self.elysium_url = []
cleanmovie = cleantitle.get(title)
title = cleantitle.getsearch(title)
titlecheck = cleanmovie+year
query = self.search_link % (urllib.quote_plus(title),year)
query = urlparse.urljoin(self.base_link, query)
query = query + "&x=0&y=0"
headers = {'User-Agent': random_agent()}
html = BeautifulSoup(requests.get(query, headers=headers, timeout=30).content)
result = html.findAll('div', attrs={'class': 'post'})
for r in result:
r_href = r.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
# print ("MOVIEXK r2", r_href)
r_title = r.findAll('a')[0]["title"]
# print ("MOVIEXK r3", r_title)
r_title = r_title.encode('utf-8')
c_title = cleantitle_get_2(r_title)
if year in r_title:
if titlecheck in c_title:
self.elysium_url.append([r_href,r_title])
# print "SCNSRC MOVIES %s %s" % (r_title , r_href)
return self.elysium_url
except:
return
示例15: episode
def episode(self, url, imdb, tvdb, title, premiered, season, episode):
try:
if not debridstatus == 'true': raise Exception()
self.zen_url = []
headers = {'Accept-Language': 'en-US,en;q=0.5', 'User-Agent': random_agent()}
data = urlparse.parse_qs(url)
data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])
title = data['tvshowtitle'] if 'tvshowtitle' in data else data['title']
cleanmovie = cleantitle.get(title)
data['season'], data['episode'] = season, episode
ep_search = 'S%02dE%02d' % (int(data['season']), int(data['episode']))
episodecheck = str(ep_search).lower()
query = self.search_link % (urllib.quote_plus(title), ep_search)
query = urlparse.urljoin(self.base_link, query)
titlecheck = cleanmovie + episodecheck
html = BeautifulSoup(OPEN_URL(query).content)
containers = html.findAll('h1', attrs={'class': 'entry-title'})
for result in containers:
r_title = result.findAll('a')[0]
r_title = r_title.string
r_href = result.findAll('a')[0]["href"]
r_href = r_href.encode('utf-8')
r_title = r_title.encode('utf-8')
r_title = cleantitle.get(r_title)
if titlecheck in r_title:
self.zen_url.append([r_href,r_title])
return self.url
except:
return