本文整理汇总了Python中logs.LogManager.LogManager.error方法的典型用法代码示例。如果您正苦于以下问题:Python LogManager.error方法的具体用法?Python LogManager.error怎么用?Python LogManager.error使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在类logs.LogManager.LogManager
的用法示例。
在下文中一共展示了LogManager.error方法的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。
示例1: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Spider:
def __init__(self):
self.logger = LogManager(__name__)
self.opener = None
self.mycookie = None
def login(self, url, loginInfo, retry=0):
"""
Login request for user
url = '' Ex. http://www.example.com/login
loginInfo = {} Ex. {'user': 'user', 'pass': 'pass'}
"""
host = ("Host", "www.vizury.com")
conn = ("Connection", "keep-alive")
enc = ("Accept-Encoding", "gzip, deflate")
ac = ("Accept", "text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8")
ln = ("Accept-Language", "en-us,en;q=0.5")
self.opener = self.createOpener([config.USER_AGENT, conn, enc, ac, ln, host], self.createCookieJarHandler())
urllib2.install_opener(self.opener)
try:
return self.opener.open(url, urllib.urlencode(loginInfo)).read()
except Exception, x:
self.logger.error(x.message)
if retry < config.RETRY_COUNT:
self.login(url, loginInfo, retry + 1)
return None
示例2: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Browser:
def __init__(self):
self.logger = LogManager(__name__)
self.browser = None
self.browserCookieJar = None
def browserLogin(self, url, loginParams, formId=None, saveCookie=False, retry=0):
"""
Login page just like web browser
url = '' Ex. http://www.example.com
loginInfo = {} Ex. {'user': 'user', 'pass': 'pass'}
"""
try:
self.browser = self.createBrowser([config.USER_AGENT])
self.browser.open(url, timeout=config.TIMEOUT)
if formId is not None:
self.browser.select_form(predicate=lambda f: 'id' in f.attrs and f.attrs['id'] == formId)
else:
self.browser.select_form(nr=0)
for key in loginParams:
self.browser.form[key] = loginParams[key]
self.browser.submit()
if saveCookie:
self.browserCookieJar.save(config.COOKIE_FILE)
return self.browser.response().read()
except Exception, x:
self.logger.error(x)
if retry < config.RETRY_COUNT:
self.browserLogin(url, loginParams, formId, saveCookie, retry + 1)
return None
示例3: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class GoogleFinanceScrapper:
isFinished = False
def __init__(self, filename):
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
self.utils = Utils()
self.filename = filename
self.url = 'https://www.google.com/finance?'
self.main_url = 'https://www.google.com'
self.csvWriter = Csv('google_finance.csv')
csvDataHeader = ['Ticker Symbol', 'Quarter End', 'Revenue', 'Total Revenue', 'Date of Scrape']
self.csvWriter.writeCsvRow(csvDataHeader)
def run(self):
self.scrapData()
self.csvWriter.closeWriter()
def scrapData(self):
try:
file = open(self.filename, 'rb')
for line in file.readlines():
if self.isFinished: return
line = self.regex.replaceData('\r+', '', line)
line = self.regex.reduceNewLine(line)
line = self.regex.reduceBlankSpace(line)
line = line.strip()
params = urllib.urlencode({'q': line})
url = self.url + params
self.scrapBykeyword(url, line)
except Exception, x:
print x
self.logger.error('Error: ' + x.message)
示例4: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Spider:
def __init__(self):
self.logger = LogManager(__name__)
self.opener = None
self.mycookie = None
def login(self, url, loginInfo, retry=0, proxy=None):
"""
Login request for user
url = '' Ex. http://www.example.com/login
loginInfo = {} Ex. {'user': 'user', 'pass': 'pass'}
"""
conn = ('Connection', 'keep-alive')
ac = ('Accept', 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8')
ln = ('Accept-Language', 'en-us,en;q=0.5')
if proxy is None:
self.opener = self.createOpener([config.USER_AGENT, conn, ac, ln], self.createCookieJarHandler())
else:
self.opener = self.createOpener([config.USER_AGENT, conn, ac, ln], self.createCookieJarHandler(), proxy)
urllib2.install_opener(self.opener)
try:
return self.opener.open(url, urllib.urlencode(loginInfo)).read()
except Exception, x:
print x.message
self.logger.error(x.message)
if retry < config.RETRY_COUNT:
self.login(url, loginInfo, retry + 1)
return None
示例5: AmazonScrapper
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class AmazonScrapper(QThread):
notifyAmazon = pyqtSignal(object)
def __init__(self, urlList, category):
QThread.__init__(self)
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
self.utils = Utils()
self.urlList = urlList
self.category = category
dupCsvReader = Csv()
self.dupCsvRows = dupCsvReader.readCsvRow(category + '.csv')
self.csvWriter = Csv(category + '.csv')
csvDataHeader = ['SKU', 'Title', 'Sub Title', 'Price', 'Shipping Weight', 'Image URL']
if csvDataHeader not in self.dupCsvRows:
self.dupCsvRows.append(csvDataHeader)
self.csvWriter.writeCsvRow(csvDataHeader)
self.mainUrl = 'http://www.amazon.com'
self.scrapUrl = None
self.dbHelper = DbHelper('amazon.db')
self.dbHelper.createTable(category)
self.total = self.dbHelper.getTotalProduct(category)
def run(self, retry=0):
try:
# self.scrapProductDetail(
# 'http://www.amazon.com/Casio-MRW-S300H-8BVCF-Solar-Powered-Analog/dp/B00ELALKH2/ref=sr_1_544/184-7248556-2619812?s=watches&ie=UTF8&qid=1397580509&sr=1-544')
# return
if self.urlList is not None and len(self.urlList):
for url in self.urlList:
if len(url) > 0:
url = self.regex.replaceData('(?i)\r', '', url)
url = self.regex.replaceData('(?i)\n', '', url)
self.notifyAmazon.emit('<font color=green><b>Amazon Main URL: %s</b></font>' % url)
imUrl = None
retry = 0
while imUrl is None and retry < 4:
imUrl = self.reformatUrl(url)
retry += 1
if imUrl is None:
imUrl = url
self.total = 0
print 'URL: ' + str(imUrl)
sortList = ['relevance-fs-browse-rank', 'price', '-price', 'reviewrank_authority',
'date-desc-rank']
for sort in sortList:
self.scrapReformatData(imUrl, sort)
self.notifyAmazon.emit(
'<font color=red><b>Finish data for Amazon Main URL: %s</b></font><br /><br />' % url)
self.notifyAmazon.emit('<font color=red><b>Amazon Data Scraping finished.</b></font>')
except Exception, x:
print x.message
self.logger.error('Exception at run: ', x.message)
if retry < 5:
self.run(retry + 1)
示例6: NisbetProduct
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class NisbetProduct(QtCore.QThread):
scrapProductData = QtCore.pyqtSignal(object)
stopThread = QtCore.pyqtSignal(int)
def __init__(self):
QtCore.QThread.__init__(self)
self.isExiting = False
self.totalProducts = 0
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
dupCsvReader = Csv()
self.dupCsvRows = dupCsvReader.readCsvRow('nisbets.csv', 0)
self.csvWriter = Csv('nisbets.csv')
self.mainUrl = 'http://www.nisbets.co.uk'
csvHeaderList = ['URL', 'Product Code', 'Product Technical Specifications', 'Product Name', 'Brand',
'Product Price', 'Product Short Description',
'Product Long Description', 'Image File Name', 'User Manual File Name',
'Exploded View File Name', 'Spares Code', 'Accessories', 'Product Status' 'Category1',
'Category2', 'Category3',
'Category4']
if 'URL' not in self.dupCsvRows:
self.csvWriter.writeCsvRow(csvHeaderList)
self.dupCsvRows.append(csvHeaderList[0])
self.utils = Utils()
def run(self):
self.scrapData()
def stop(self):
self.isExiting = True
def scrapData(self):
if self.isExiting: return
self.scrapProductData.emit('<font color=green><b>Main URL: </b>%s</font>' % self.mainUrl)
self.logger.debug('===== URL [' + self.mainUrl + '] =====')
data = self.spider.fetchData(self.mainUrl)
if data and len(str(data).strip()) > 0:
data = self.regex.reduceNewLine(data)
data = self.regex.reduceBlankSpace(data)
category1Chunk = self.regex.getAllSearchedData('(?i)<li id="li-id-\d+">(.*?)</ul> </li>', data)
if category1Chunk and len(str(category1Chunk).strip()) > 0:
i = 0
for category1Data in category1Chunk:
category1 = self.regex.getSearchedData('(?i)<a href="[^"]*">([^<]*)</a>', category1Data)
category2Chunk = self.regex.getAllSearchedData('(?i)<li><a href="([^"]*)">([^<]*)</a>',
category1Data)
if category2Chunk and len(str(category2Chunk).strip()) > 0:
for category2Data in category2Chunk:
try:
self.scrapCategory2Data(self.mainUrl + category2Data[0], category1, category2Data[1])
except Exception, x:
self.logger.error(x)
self.scrapProductData.emit('<font color=red><b>Finish Scraping Product data from %s</b></font>' % self.mainUrl)
示例7: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Csv:
def __init__(self, fileName=None):
self.logger = LogManager(__name__)
if fileName is not None:
self.writer = csv.writer(open(fileName, 'ab'))
def writeCsvRow(self, data):
try:
self.writer.writerow(data)
except Exception, x:
self.logger.error(x)
示例8: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Csv:
def __init__(self, fileName=None):
self.logger = LogManager(__name__)
if fileName is not None:
self.writer = UnicodeWriter(open(fileName, 'wb'), quoting=csv.QUOTE_ALL)
def writeCsvRow(self, data):
try:
self.writer.writerow(data)
except Exception, x:
self.logger.error(x)
示例9: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Utils:
def __init__(self):
self.logger = LogManager(__name__)
def downloadFile(self, url, savePath):
try:
directory = os.path.dirname(savePath)
if not os.path.exists(directory):
os.makedirs(directory)
webFile = urllib2.urlopen(url)
localFile = open(savePath, 'wb')
localFile.write(webFile.read())
except Exception, x:
self.logger.error(x)
示例10: SaraivaScrapper
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class SaraivaScrapper(QThread):
notifySaraiva = pyqtSignal(object)
def __init__(self, urlList, category, htmlTag, replaceTag):
QThread.__init__(self)
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
self.utils = Utils()
self.urlList = urlList
self.category = category
self.htmlTag = self.regex.replaceData('\r+', '', htmlTag)
self.htmlTag = self.regex.replaceData('\n+', ' ', self.htmlTag)
self.htmlTag = self.regex.replaceData('\s+', ' ', self.htmlTag)
self.htmlTag = self.regex.replaceData(r'\"+', '\"', self.htmlTag)
self.replaceTag = replaceTag
self.csvWriter = Csv(category + '.csv')
csvDataHeader = ['Link', 'Name', 'Subtitle', 'Price', 'Synopsis and Characteristics', 'Picture']
self.csvWriter.writeCsvRow(csvDataHeader)
self.mainUrl = 'http://busca.livrariasaraiva.com.br'
self.scrapUrl = None
self.dbHelper = DbHelper('saraiva.db')
self.dbHelper.createTable(category)
self.total = self.dbHelper.getTotalProduct(category)
def run(self, retry=0):
try:
if self.urlList is not None and len(self.urlList):
for url in self.urlList:
if len(url) > 0:
url = self.regex.replaceData('(?i)\r', '', url)
url = self.regex.replaceData('(?i)\n', '', url)
self.notifySaraiva.emit('<font color=green><b>Saraiva Main URL: %s</b></font>' % url)
paginationUrl, self.maxRecords = self.reformatUrl(url)
self.notifySaraiva.emit(
'<font color=black><b>Total Records: %s</b></font>' % str(self.maxRecords))
print 'Max records: ', self.maxRecords
print 'URL: ' + str(paginationUrl)
sortList = ['&isort=globalpop', '&isort=best', '&isort=title', '&isort=title+rev',
'&isort=price+rev',
'&isort=price', '&isort=date+rev']
for sort in sortList:
self.scrapResults(paginationUrl, sort)
self.notifySaraiva.emit('<font color=red><b>Saraiva Data Scraping finished.</b></font>')
except Exception, x:
print x.message
self.logger.error('Exception at run: ', x.message)
if retry < 5:
self.run(retry + 1)
示例11: PaodeacucarScrapper
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class PaodeacucarScrapper(QThread):
notifyPaode = pyqtSignal(object)
def __init__(self):
QThread.__init__(self)
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
self.utils = Utils()
self.mainUrl = 'http://www.paodeacucar.com.br/'
self.url = 'http://www.paodeacucar.com.br/'
dupCsvReader = Csv()
self.dupCsvRows = dupCsvReader.readCsvRow('paodeacucar.csv', 4)
self.csvWriter = Csv('paodeacucar.csv')
csvDataHeader = ['SKU', 'Category', 'Subcategory', 'Name', 'URL', 'URL Image', 'Details',
'Nutrients Table html code', 'Price from, 28/abr/14', '28/abr/14']
if 'URL' not in self.dupCsvRows:
self.dupCsvRows.append(csvDataHeader)
self.csvWriter.writeCsvRow(csvDataHeader)
def run(self):
self.scrapData()
def scrapData(self):
try:
print 'Main URL: ', self.url
self.notifyPaode.emit(('<font color=green><b>Main URL: %s</b></font>' % self.url))
data = self.spider.fetchData(self.url)
if data and len(data) > 0:
data = self.regex.reduceNewLine(data)
data = self.regex.reduceBlankSpace(data)
soup = BeautifulSoup(data)
categories = soup.find('nav', class_='items-wrapper').find_all('li', class_=re.compile('\s*item\s*'))
print 'Total Categories: ', len(categories)
self.notifyPaode.emit(('<font color=black><b>Total Categories: %s</b></font>' % str(len(categories))))
for category in categories:
if category.a is not None:
submenu_target = self.regex.replaceData('#', '', category.a.get('data-target'))
sub_categories = soup.find('ul', id=submenu_target).find_all('li', class_='item')
print 'Total Sub Categories: ', len(sub_categories)
self.notifyPaode.emit(('<font color=black><b>Total Subcategories: %s</b></font>' % str(len(sub_categories))))
for sub_category in sub_categories:
sub_category_label = sub_category.find('span', class_='label').text
sub_category_url = sub_category.a.get('href') if sub_category.a is not None else 'N/A'
self.scrapItems(sub_category_url, category.text, sub_category_label)
except Exception, x:
self.logger.error(x.message)
print x
示例12: CsBrands
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class CsBrands(QThread):
notifyBrand = pyqtSignal(object)
def __init__(self):
QThread.__init__(self)
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
self.utils = Utils()
dupCsvReader = Csv()
self.dupCsvRows = dupCsvReader.readCsvRow("cs_Brands.csv")
self.csvWriter = Csv("cs_Brands.csv")
self.mainUrl = "http://www.cs-catering-equipment.co.uk/brands"
self.isExiting = False
headerData = [
"URL",
"Parent Category",
"Brand Category",
"Brand Description",
"Image File",
"Product Codes in this category",
]
if headerData not in self.dupCsvRows:
self.csvWriter.writeCsvRow(headerData)
def run(self):
self.scrapBrands()
self.notifyBrand.emit("<font color=red><b>Finished Scraping All Brands.</b></font>")
def scrapBrands(self):
self.notifyBrand.emit("<font color=green><b>Main URL: %s<b></font>" % self.mainUrl)
self.notifyBrand.emit("<b>Try To scrap All Brands.<b>")
data = self.spider.fetchData(self.mainUrl)
if data and len(data) > 0:
data = self.regex.reduceNewLine(data)
data = self.regex.reduceBlankSpace(data)
brandChunks = self.regex.getAllSearchedData('(?i)<div class="man-group man-group-[a-z]">(.*?)</div>', data)
if brandChunks and len(brandChunks) > 0:
for brandChunk in brandChunks:
brands = self.regex.getAllSearchedData('(?i)<a href="([^"]*)"[^>]*?>([^<]*)</a>', brandChunk)
self.notifyBrand.emit("<b>Total Brands Found: %s<b>" % str(len(brands)))
if brands and len(brands) > 0:
for brand in brands:
try:
self.scrapBrandInfo(brand[0], "Shop By Brand", brand[1])
except Exception, x:
self.logger.error(x)
示例13: CsTest
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class CsTest(QThread):
notifyProduct = pyqtSignal(object)
def __init__(self):
QThread.__init__(self)
self.logger = LogManager(__name__)
self.spider = Spider()
self.regex = Regex()
dupCsvReader = Csv()
self.dupCsvRows0 = dupCsvReader.readCsvRow('cs_product.csv', 0)
self.dupCsvRows = dupCsvReader.readCsvRow('cs_product.csv', 1)
self.csvWriter = Csv('cs_product.csv')
# self.mainUrl = 'http://www.cs-catering-equipment.co.uk/'
self.mainUrl = 'http://www.cs-catering-equipment.co.uk/brands'
self.utils = Utils()
if 'Product Code' not in self.dupCsvRows:
self.csvWriter.writeCsvRow(
['URL', 'Product Code', 'Product Name', 'Manufacturer', 'List Price', 'Product Price', 'Discount',
'Product Short Description', 'Product Long Description', 'Product Technical Specifications', 'Warranty'
,
'Delivery',
'Product Image',
'Category 1', 'Category 2', 'Category 3', 'Category 4', 'Brand Image'])
self.totalProducts = len(self.dupCsvRows)
def run(self):
self.scrapBrands()
self.notifyProduct.emit('<font color=red><b>Finished Scraping All Brands.</b></font>')
def scrapBrands(self):
self.notifyProduct.emit('<font color=green><b>Main URL: %s<b></font>' % self.mainUrl)
self.notifyProduct.emit('<b>Try To scrap All Brands.<b>')
data = self.spider.fetchData(self.mainUrl)
if data and len(data) > 0:
data = self.regex.reduceNewLine(data)
data = self.regex.reduceBlankSpace(data)
brandChunks = self.regex.getAllSearchedData('(?i)<div class="man-group man-group-[a-z]">(.*?)</div>', data)
if brandChunks and len(brandChunks) > 0:
for brandChunk in brandChunks:
brands = self.regex.getAllSearchedData('(?i)<a href="([^"]*)"[^>]*?>([^<]*)</a>', brandChunk)
self.notifyProduct.emit('<b>Total Brands Found: %s<b>' % str(len(brands)))
if brands and len(brands) > 0:
for brand in brands:
try:
self.scrapBrandInfo(brand[0], 'Shop By Brand', brand[1])
except Exception, x:
self.logger.error(x)
示例14: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Spider:
def __init__(self):
self.logger = LogManager(__name__)
self.opener = None
def login(self, url, loginInfo, retry=0):
"""
Login request for user
url = '' Ex. http://www.example.com/login
loginInfo = {} Ex. {'user': 'user', 'pass': 'pass'}
"""
self.opener = self.createOpener([config.USER_AGENT], self.createCookieJarHandler())
urllib2.install_opener(self.opener)
try:
return self.opener.open(url, urllib.urlencode(loginInfo)).read()
except Exception, x:
self.logger.error(x.message)
if retry < config.RETRY_COUNT:
self.login(url, loginInfo, retry + 1)
return None
示例15: __init__
# 需要导入模块: from logs.LogManager import LogManager [as 别名]
# 或者: from logs.LogManager.LogManager import error [as 别名]
class Spider:
def __init__(self):
self.logger = LogManager(__name__)
self.opener = None
self.mycookie = None
def login(self, url, loginInfo, retry=0, proxy=None):
"""
Login request for user
url = '' Ex. http://www.example.com/login
loginInfo = {} Ex. {'user': 'user', 'pass': 'pass'}
"""
conn = ('Connection', 'keep-alive')
ac = ('Accept', 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8')
ln = ('Accept-Language', 'en-us,en;q=0.5')
if proxy is None:
self.opener = self.createOpener([config.USER_AGENT, conn, ac, ln], self.createCookieJarHandler())
else:
self.opener = self.createOpener([config.USER_AGENT, conn, ac, ln], self.createCookieJarHandler(), proxy)
urllib2.install_opener(self.opener)
try:
response = self.opener.open(url, urllib.urlencode(loginInfo))
print 'Response from Server:'
print 'Status: ', response.getcode()
print response.info()
self.logger.debug('Response from Server:')
self.logger.debug('Status: ' + str(response.getcode()))
self.logger.debug(response.info())
redirected_url = response.url
return redirected_url, response.read()
except Exception, x:
print x
self.logger.error(x.message)
if retry < config.RETRY_COUNT:
print 'Retry again. Please wait 5 seconds...'
time.sleep(5)
self.login(url, loginInfo, retry + 1)
else:
print 'Failed to retrieve data after maximum %d retry!' % config.RETRY_COUNT
return None, None