当前位置: 首页>>代码示例>>Python>>正文


Python pygoogle.pygoogle函数代码示例

本文整理汇总了Python中pygoogle.pygoogle函数的典型用法代码示例。如果您正苦于以下问题:Python pygoogle函数的具体用法?Python pygoogle怎么用?Python pygoogle使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。


在下文中一共展示了pygoogle函数的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。

示例1: _get_goog_urls

 def _get_goog_urls(query):
     g = pygoogle.pygoogle(query)
     g.pages = 1 
     g.hl = "en"
     self.l.info("Google search result count: %s" % g.get_result_count())
     if g.get_result_count() > 0:
         return g.search_page_wise()[0]
     else:
         g = pygoogle.pygoogle("site:blogspot.com groot")
         g.pages = 1
         g.hl = "en"
         self.l.info("No results for original query, retrying with 'groot'")
         return g.search_page_wise()[0]
开发者ID:larsendt,项目名称:questionbot,代码行数:13,代码来源:corpus.py

示例2: get_impact_factor_from_issn

def get_impact_factor_from_issn(issn="1475-7516", debug=False):
    """
      For the input ISSN in the format NNNN-NNNN obtain
      the headers and the datasets in a nested list
      equivalent to an array of (# headers)*[4 (years)]
    """
    g = pygoogle("site:http://www.bioxbio.com/if/html " + issn)
    g.pages = 1
    if g.get_urls():
        if_file = urllib.urlopen(g.get_urls()[0])
        html = if_file.read()
        if_file.close()
    else:
        return [], []

    if debug:
        print (html)
    soup = BeautifulSoup(html)
    table = soup.find("table")

    # The first tr contains the field names.
    headings = [th.get_text().strip() for th in table.find("tr").find_all("td")]

    datasets = []
    for row in table.find_all("tr")[1:]:
        dataset = [eval(td.get_text().replace("-", "0")) for td in row.find_all("td") if td.get_text().strip()]
        datasets.append(dataset)

    return headings, datasets
开发者ID:restrepo,项目名称:gssis,代码行数:29,代码来源:csvreader.py

示例3: searchGoogle

def searchGoogle(searchTerm): 

    print ''
    print 'Searching Google...'
    print ''

    googler = pygoogle(searchTerm) # initialize pygoogle object with search term
    googler.pages = 3 # set max pages

    print '*********************************'
    print 'Google Results'
    print '*********************************'
    print ''

    # display google results in a formatted way
    for keys, values in googler.search().items():
        theKey=keys.replace("'","'")
        theKey=theKey.replace("&","&")
        theValue=values.replace("'","'")
        theValue=theValue.replace("&","&")
        print 'Title: ' + (theKey.encode('ascii', 'ignore'))
        print 'URL: ' + (theValue.encode('ascii', 'ignore'))
        print ''
    print ''
    print '*********************************'
    print ''
开发者ID:hawkbluedevil,项目名称:threatHawk,代码行数:26,代码来源:threatHawk.py

示例4: dork

    def dork(self,search_term,p,output):
        print YELLOW + "[+] " + END + WHITE + "Searching for " + END + "%s " % search_term
        gs = pygoogle(search_term)
        gs.pages = p
        print YELLOW + "[+] " + END + WHITE + "Results Found : " + END + "%s " % (gs.get_result_count())
        if gs.get_result_count() == 0: print RED + "[-] " + END + WHITE + "No Results Found" + END; time.sleep(1); sys.exit()

        print YELLOW + "[+] " + END + WHITE + "Fetching " + END + "[%s] Results " % (gs.get_result_count())
        url_list = gs.get_urls()

        if len(url_list) == 0:
            print YELLOW + "[!] " + END + WHITE + "Got 0 URLs" + END
            print RED + "[!] " + END + WHITE + "Nothing to save" + END
            time.sleep(1)
            sys.exit()
            
        elif len(url_list) > 1:
            print YELLOW + "[+] " + END + WHITE + "Got " + END + "[%s] URLs" % (len(url_list))
            print YELLOW + "[+] " + END + WHITE + "Writing URLs to " + END + "[%s] " % (output)

            with open(output,'w') as w_file:
                for i in url_list: w_file.write(i+'\n')
            print YELLOW + "[+] " + END + WHITE + "URLs saved to " + END + "[%s] " % (output)

            time.sleep(2)
开发者ID:tofuseng,项目名称:sqlivulscan,代码行数:25,代码来源:pydorker.py

示例5: googleIt

def googleIt(url):
		db = database()
		source = url
		en = ["it","zh-Hans","fr","nl","es","pt-BR","ca","pa","qu","mr","mo","mn","ne","pcm","nn","or","qu"]
		random.shuffle(en)

		search = pygoogle.pygoogle(hl=en[0],query="site:"+source)
		urlList = search.get_urls()

		print urlList
		sha1 = hashlib.sha1()

		for eachUrl in urlList:
			#Generate hash for url - used as primary key for database
			try:
				eachUrl = u"".join(eachUrl).encode('utf-8').strip()
				sha1.update(eachUrl)
				hash = sha1.hexdigest()
				numTLD = db.countTLD(eachUrl)
				#Persist item in database 
				db.addGoodUrl(source,hash, eachUrl, numTLD)
			except:
				doNothing =0
		print 'Done'
		db.close()
开发者ID:sunnyneo,项目名称:scrapy-script,代码行数:25,代码来源:getgoodurl.py

示例6: do_action

def do_action(lcars, string, case):
	target = string.split(' ', 1)[1]

	g = pygoogle("site:youtube.com " + target)
	g.pages = 1
	urls = g.get_urls()
	if len(urls) == 0:
		reply = "No results found for" + target
		lcars.reply_with(reply)
		return reply
	link = urls[0]

	if "user" in link or "store" in link or "feed" in link or "playlist" in link or "channel" in link:
		for url in urls:
			if "user" in url:
				link = "http://youtube.nestharion.de/" + url.split('/')[-1]
				break
		if not link:
			for url in urls:
				if "store" in url or "feed" in url or "playlist" in url or "channel" in url:
					continue
				else:
					link = url
					break
	if not link:
		link = urls[randint(0, len(urls) - 1)]

	lcars.background(["google-chrome", link])
	reply = "Playing " + target
	lcars.reply_with(reply)
	os.system("sleep 1")
	os.system("xdotool windowactivate --sync $(xdotool search --class Chrome | head -n 1) & sleep 0.3; pkill xdotool")
	os.system("xdotool windowactivate --sync $(xdotool search --class Chrome | tail -n 1) & sleep 0.3; pkill xdotool")

	return reply
开发者ID:soreau,项目名称:lcars,代码行数:35,代码来源:lcars_youtube.py

示例7: fresh_google_check

def fresh_google_check(link: str, attempt=5, debug=False):
    """Проверяет, индексировался ли уже ресурс гуглом раньше.

    чем за 2 недели до сегодня.
    :param link:
    :param attempt:
    :return:

    """
    if debug:
        return False
    try:
        assert isinstance(link, str)
        today = datetime.date.today()
        date_s = _date_to_julian_day(today - datetime.timedelta(days=365 * 8))
        date_e = _date_to_julian_day(today - datetime.timedelta(days=7 * 2))
        query = u'site:%s daterange:%s-%s' % (link, date_s, date_e, )

        result = False
        for i in range(0, attempt):
            g = pygoogle(query.encode('utf-8'),
                         raise_http_exceptions=True,
                         proxies=settings.PROXIES_FOR_GOOGLING)

            try:
                result = bool(g.get_result_count())
            except PyGoogleHttpException as e:
                renew_connection()
                continue
            break
    except (AssertionError, PyGoogleHttpException, stem.SocketError):
        result = False

    return result
开发者ID:cheshirski,项目名称:pythondigest,代码行数:34,代码来源:__init__.py

示例8: LinkedIn

def LinkedIn(linkedinusername):
	from pygoogle import pygoogle
	g = pygoogle("linkedin "+linkedinusername)
	g.pages = 5
	g.get_result_count()
	myURLs = g.get_urls()
	return myURLs
开发者ID:Erik-k,项目名称:maltego_project,代码行数:7,代码来源:linkedin_xform.py

示例9: google_search

def google_search(search_string):
    g = pygoogle(search_string)
    g.pages = 1
    results = g.get_urls()
    try:
        return results[0]
    except:
        return "That was not the word you're looking for"
开发者ID:Murklan,项目名称:Ircbot,代码行数:8,代码来源:googlewiki.py

示例10: google_first_result

def google_first_result(googlestring):
	pygoog = pygoogle(googlestring)
	pygoog.pages = 1
	urls = pygoog.get_urls()
	try:		
		return urls[0]
	except IndexError:
		return "http://www.google.com"
开发者ID:EvanJRichter,项目名称:Stalkr,代码行数:8,代码来源:crawler.py

示例11: https_search

def https_search(url):
   
    string_search = "inurl:https site:"+str(url)

    g = pygoogle(string_search)
    g.pages = 5
    g.hl = "br"

    print string_search 

    results_numbers = 0
    count = 0
    temp = 6 # segundos

    while results_numbers == 0 :
        results_numbers = g.get_result_count()
        print "Resultados:",results_numbers
        print
        if results_numbers == 0:
            time.sleep( temp ) 
            count += temp 
            if count > 60: # segundos
                count = -1
                print "Desisto!"
                break

    desired_results = 5
    search_sites = {}

    if count == -1:
        print "Sem estima dos resultados da pesquisa"
        return 0

    elif results_numbers < desired_results:
        print "Poucos sites!"
        desired_results = results_numbers
    
    while len(search_sites) == 0:
        search_sites = g.search()
        print search_sites
        print
        for key in search_sites.keys():
            #print key, search_sites[key]
            print unicode(key).encode('cp850'), unicode(search_sites[key]).encode('cp850')
        if len(search_sites) == 0 or len(search_sites) < desired_results:
            time.sleep( temp ) 
            count += temp 
            if count > 60: # segundos
                count = -1
                print "Desisto!"
                break
    
    if count == -1:
        print "Possivel bloqueio do Google"
        return 0
   
    
    print "Fim"
开发者ID:douglasiuri,项目名称:https-check,代码行数:58,代码来源:https_search.py

示例12: googleSearch

def googleSearch ( searchString ):
    g = pygoogle(searchString)
    g.pages = 2
    urls = g.get_urls()
    urls = urls[:10]
    for i in range(len(urls)):
        urls[i]=unicode(urlnorm.norm(urls[i]))

    return urls
开发者ID:Walliee,项目名称:FocusedCrawler,代码行数:9,代码来源:gQuery.py

示例13: getSearchAndDownloadPaper

def getSearchAndDownloadPaper(textToSearch, fileNameToSave):
	g = pygoogle(textToSearch + ' filetype:pdf')
	g.pages = 1
	try:
		pdfUrl = g.get_urls()[0]
		urllib.urlretrieve(pdfUrl, "../pdfdownloaded/" + fileNameToSave)
		time.sleep(random.randint(30,60))
	except IndexError:
		print fileNameToSave + " " + textToSearch
		time.sleep(180);
开发者ID:nmonath,项目名称:NLPProject,代码行数:10,代码来源:pdfFromGoogle.py

示例14: crackedonpygoogle

def crackedonpygoogle(passhash, plaintext):  #trying this approach
    from pygoogle import pygoogle
    googleresult = pygoogle(passhash)  #default is for moderate safe search.  Probably OK to let this be, since we won't find porn while googling a password hash.  Probably throwing caution (and Rule 34) to the wind here.
    googleresult.pages = 1
    resulturls = googleresult.get_urls()
    for i in range(0,len(resulturls)):
        resulturls[i] = str(resulturls[i])
    if crackedonweb(passhash, plaintext, resulturls):
        return True
    else:
        return False
开发者ID:michael-weinstein,项目名称:Hash-Sambrada,代码行数:11,代码来源:brokenhash0.2.py

示例15: searchtext

	def searchtext(self, user_input):
		searchterms = user_input + ' site:stackoverflow.com'
		print "Searching:", searchterms
		g = pygoogle(searchterms)
		g.pages = 1
		urls = g.get_urls()
		#go through search results
		for url in urls[:int(len(urls)/4+0.5)]:
			req = urllib2.Request(url, headers = hdr)
			try:
				self.myParser.answers = 0
				page = urllib2.urlopen(req)
				html = page.read()
				#print html
				html_fixed = html.replace('&gt;', '3cmr93iwm0c9ri3w0')
				html_fixed = html_fixed.replace('&lt;','98jdsf98j3oisdf')
				html_fixed = html_fixed.replace('&amp;','dksljf9w8ejfosidjf')

				#html_fixed = html_fixed.replace('...',' ')
				self.myParser.feed(html_fixed)
				self.snips = self.myParser.snips
				#print self.snips
				for x in self.snips:
					for y in x[0]:
						print url
						answer = sublime.ok_cancel_dialog(y.replace('98jdsf98j3oisdf','<').replace('3cmr93iwm0c9ri3w0','>').replace('dksljf9w8ejfosidjf','&'))
						if answer == 1:
							self.view.insert(self.editor,
								self.view.sel()[0].begin(),y.replace('98jdsf98j3oisdf','<').replace('3cmr93iwm0c9ri3w0','>').replace('dksljf9w8ejfosidjf','&'))
							if self.language in starter:
								self.view.insert(self.editor,
									self.view.sel()[0].begin(),"\n\n"+starter[self.language]+'\n'+x[1].replace('98jdsf98j3oisdf','<').replace('3cmr93iwm0c9ri3w0','>').replace('\t',' ').replace('\n','').replace(starter[self.language],' ').replace(ender[self.language],' ').replace('dksljf9w8ejfosidjf','&')+'\n'+\
									ender[self.language]+"\n\n")
							else:
								self.view.insert(self.editor,
									self.view.sel()[0].begin(),"/*"+x[1].replace('98jdsf98j3oisdf','<').replace('3cmr93iwm0c9ri3w0','>').replace('\t',' ').replace('\n','').replace('dksljf9w8ejfosidjf','&')+\
									'*/'+"\n\n")
							self.myParser.snips = []
							self.myParser.curr_snips = []
							self.myParser.curr_snip = ''
							self.myParser.curr_comment = ''
							self.snips = []
							break
					else: 
						continue
					break
				else:
						self.myParser.snips = []
						continue
				break
			except urllib2.HTTPError,e:
				print e.fp.read()
开发者ID:MilesCranmer,项目名称:gso,代码行数:52,代码来源:stack.py


注:本文中的pygoogle.pygoogle函数示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。