简体   繁体   English

Scrapy教程错误:“不再支持使用多个蜘蛛运行'Scrapy Crawl'”

[英]Scrapy tutorial error: “Running 'Scrapy Crawl' with more than one spider is no longer supported”

Windows 10, Command Line Terminal, Python 3.6.3 |Anaconda, Inc.| Windows 10,命令行终端,Python 3.6.3 | Anaconda,Inc. | (default, Oct 15 2017, 03:27:45) [MSC v.1900 64 bit (AMD64)] on win32 (默认值,2017年10月15日,03:27:45)[win32上的[MSC v.1900 64位(AMD64)]]

New to Python & following tutorial instructions here: https://docs.scrapy.org/en/latest/intro/tutorial.html Python的新手,并按照此处的指导说明进行操作: https : //docs.scrapy.org/en/latest/intro/tutorial.html

On the step where I run my first crawl, I get the following error...I've only created one file quotes_spider.py as the tutorial suggests. 在运行第一次爬网的步骤中,出现以下错误...按照教程的建议,我仅创建了一个文件quotes_spider.py。 I've also tried changing the directory to c:\\Anaconda\\projects\\tutorial\\startproject and c:\\Anaconda\\projects\\tutorial\\startproject\\spiders but get the same error. 我也尝试过将目录更改为c:\\Anaconda\\projects\\tutorial\\startprojectc:\\Anaconda\\projects\\tutorial\\startproject\\spiders但遇到相同的错误。

c:\Anaconda\projects\tutorial>c:\Anaconda\Scripts\scrapy.exe -scrapy crawl quotes  

2017-11-09 16:08:26 [scrapy.utils.log] INFO: Scrapy 1.4.0 started (bot: startproject)
2017-11-09 16:08:26 [scrapy.utils.log] INFO: Overridden settings: {'BOT_NAME': 'startproject', 'NEWSPIDER_MODULE': 'startproject.spiders', 'ROBOTSTXT_OBEY': True, 'SPIDER_MODULES': ['startproject.spiders']}
Usage
=====
  scrapy crawl [options] <spider>

scrapy: error: running 'scrapy crawl' with more than one spider is no longer supported

What's wrong 怎么了

....scrapy.exe -scrapy crawl quotes

What it should be 应该是什么

....scrapy.exe crawl quotes

Why 为什么

This is scrapy's all available commands, there is no -scrapy option: 这是scrapy的所有可用命令,没有-scrapy选项: 在此处输入图片说明

Suggestions 意见建议

Scrapy Official Tutorial PDF Version Scrapy官方教程PDF版本

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM