繁体   English   中英

无法从 scrapy.CrawlerProcess 获取 Scrapy Stats

[英]Can't get Scrapy Stats from scrapy.CrawlerProcess

我正在从另一个脚本运行爬虫,我需要从 Crawler 检索并保存到变量统计信息中。 我查看了文档和其他 StackOverflow 问题,但我无法解决这个问题。

这是我运行爬网的脚本:

import scrapy
from scrapy.crawler import CrawlerProcess


process = CrawlerProcess({})
process.crawl(spiders.MySpider)
process.start()

stats = CrawlerProcess.stats.getstats() # I need something like this

我希望统计信息包含这段数据(scrapy.statscollectors):

     {'downloader/request_bytes': 44216,
     'downloader/request_count': 36,
     'downloader/request_method_count/GET': 36,
     'downloader/response_bytes': 1061929,
     'downloader/response_count': 36,
     'downloader/response_status_count/200': 36,
     'finish_reason': 'finished',
     'finish_time': datetime.datetime(2018, 11, 9, 16, 31, 2, 382546),
     'log_count/DEBUG': 37,
     'log_count/ERROR': 35,
     'log_count/INFO': 9,
     'memusage/max': 62623744,
     'memusage/startup': 62623744,
     'request_depth_max': 1,
     'response_received_count': 36,
     'scheduler/dequeued': 36,
     'scheduler/dequeued/memory': 36,
     'scheduler/enqueued': 36,
     'scheduler/enqueued/memory': 36,
     'start_time': datetime.datetime(2018, 11, 9, 16, 30, 38, 140469)}

我已经检查了 CrawlerProcess,它会在抓取过程完成后返回延迟并从其“爬虫”字段中删除爬虫。

有没有办法解决这个问题?

最好的,彼得

根据文档CrawlerProcess.crawl接受爬虫或蜘蛛类,您可以通过CrawlerProcess.create_crawler从蜘蛛类创建爬虫。

因此,您可以在开始爬取过程之前创建爬虫实例,然后检索预期的属性。

下面我通过编辑几行原始代码为您提供了一个示例:

import scrapy
from scrapy.crawler import CrawlerProcess


class TestSpider(scrapy.Spider):
    name = 'test'
    start_urls = ['http://httpbin.org/get']

    def parse(self, response):
        self.crawler.stats.inc_value('foo')


process = CrawlerProcess({})
crawler = process.create_crawler(TestSpider)
process.crawl(crawler)
process.start()


stats_obj = crawler.stats
stats_dict = crawler.stats.get_stats()
# perform the actions you want with the stats object or dict

如果您想通过信号获取脚本中的统计信息。 这将有助于 -

 def spider_results(spider): results = [] stats = [] def crawler_results(signal, sender, item, response, spider): results.append(item) def crawler_stats(*args, **kwargs): # runs when spider closed stats.append(kwargs['sender'].stats.get_stats()) dispatcher.connect(crawler_results, signal=signals.item_scraped) dispatcher.connect(crawler_stats, signal=signals.spider_closed) process = CrawlerProcess() process.crawl(spider) process.start() # the script will block here until the crawling is finished return results, stats

我希望它有帮助!

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM