本文整理汇总了Python中scrapy.http.request.Request.meta['symptom_questions']方法的典型用法代码示例。如果您正苦于以下问题:Python Request.meta['symptom_questions']方法的具体用法?Python Request.meta['symptom_questions']怎么用?Python Request.meta['symptom_questions']使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在类scrapy.http.request.Request
的用法示例。
在下文中一共展示了Request.meta['symptom_questions']方法的1个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。
示例1: _parse_symptom_question
# 需要导入模块: from scrapy.http.request import Request [as 别名]
# 或者: from scrapy.http.request.Request import meta['symptom_questions'] [as 别名]
def _parse_symptom_question(self, response):
symptom_question_item = response.meta.get('symptom_questions')
# print response.url
if not symptom_question_item:
symptom_question_item = SymptomQuestionItem()
symptom_question_item['symptom_name'] = response.meta['symptom_item']['name']
symptom_question_item['qids'] = []
# parse
urls = response.xpath('//div[@class="p_list_li"]/div[@class="p_list_cent"]/div[@class="p_list_centt"]/dl/dt/a/@href').extract()
symptom_question_item['qids'] += [u.split('/')[-1].split('.')[0] for u in urls]
# last_url = response.xpath('//div[@class="portldet-content"]/a/@href').extract()[-1]
next_url = response.xpath('//div[@class="portlet-content"]/a[text()="下一页 >"]/@href').extract()
if not next_url:
# 所有页都处理完了
print symptom_question_item
yield symptom_question_item
else:
url = next_url[0]
# print url
# print symptom_question_item['qids']
request = Request(url, dont_filter=True, callback=self._parse_symptom_question)
request.meta['symptom_questions'] = symptom_question_item
# print request
yield request