[英]scrapy newbie: tutorial. error when running scrapy crawl dmoz
I've set up my PATH variables and I think I'm configuring everything right. 我已经设置了PATH变量,我认为我正在配置一切正确。 But when I run the "scrapy crawl dmoz" within my startproject folder, I get these error messages:
但是当我在startproject文件夹中运行“scrapy crawl dmoz”时,我收到以下错误消息:
c:\matt\testing\dmoz>scrapy crawl dmoz
2012-04-24 18:12:56-0400 [scrapy] INFO: Scrapy 0.14.0.2841 started (bot: dmoz)
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole,
CloseSpider, WebService, CoreStats, SpiderState
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled downloader middlewares:
HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware,
faultHeadersMiddleware, RedirectMiddleware, CookiesMiddleware,
HttpCompressionMiddleware, ChunkedTransferMiddleware, DownloaderStats
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled spider middlewares:
HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddware
2012-04-24 18:12:56-0400 [scrapy] DEBUG: Enabled item pipelines:
Traceback (most recent call last):
File "c:\Python27\Scripts\scrapy", line 4, in <module>
execute()
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7-
win32.egg\scrapy\cmdline.py", line 132, in execute
_run_print_help(parser, _run_command, cmd, args, opts)
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7-
win32.egg\scrapy\cmdline.py", line 97, in _run_print_help
func(*a, **kw)
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7-
win32.egg\scrapy\cmdline.py", line 139, in _run_command
cmd.run(args, opts)
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7-
win32.egg\scrapy\commands\crawl.py", line 43, in run
spider = self.crawler.spiders.create(spname, **opts.spargs)
File "c:\Python27\lib\site-packages\scrapy-0.14.0.2841-py2.7-
win32.egg\scrapy\spidermanager.py", line 43, in create
raise KeyError("Spider not found: %s" % spider_name)
KeyError: 'Spider not found: dmoz'
Does anyone know what might be going on? 有谁知道可能会发生什么?
I had this issue too. 我也有这个问题。
It is because the scrapy tutorial asks you to place the spider you create in /dmoz/spiders/
but scrapy is looking in tutorial/tutorial/spiders
. 这是因为scrapy教程要求你将你创建的蜘蛛放在
/dmoz/spiders/
但scrapy正在查看tutorial/tutorial/spiders
。
Save the dmoz_spider.py
in tutorial/tutorial/spiders
and the crawl should work. 将
dmoz_spider.py
保存在tutorial/tutorial/spiders
,爬行应该可以正常工作。
try in a command line 尝试在命令行中
C:\Users\Akhtar Wahab> python # any any directory path
if it works 如果有效
try 尝试
scrapy version
if its also working 如果它也有效
then make sure you made scrapy project with 然后确保你做了scrapy项目
scrapy startproject name
if above everything is Positive for you 如果以上一切都对你有利
then make sure you are Running scraper command in a directory where scrapy.cfg
is present . 然后确保在
scrapy.cfg
所在的目录中运行scraper命令。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.