簡體   English   中英

Scrapy:Python無法找到蜘蛛

[英]Scrapy: Python cannot find the spider

我正在嘗試遵循Scrapy教程,但是我陷入了第一步。 我想我已經正確創建了蜘蛛:

class dmoz(BaseSpider):
    name = "dmoz"
    allowed_domains = ["dmoz.org"]
    start_urls = [
        "http://www.dmoz.org/Computers/Programming/Languages/Python/Books/",
        "http://www.dmoz.org/Computers/Programming/Languages/Python/Resources/"
    ]

def parse(self, response):
    filename = response.url.split("/")[-2]
    open(filename, 'wb').write(response.body)

我已經從IDLE外殼程序(在與終端窗口目錄相對應的給定文件夾中鍵入.py擴展名)保存了該文件(作為dmoz_spider.py)。

但是,當我輸入scrapy crawl dmoz我得到了:

2013-08-09 19:18:06+0200 [scrapy] INFO: Scrapy 0.16.5 started (bot: dmoz)
2013-08-09 19:18:07+0200 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole, CloseSpider, WebService, CoreStats, SpiderState
2013-08-09 19:18:08+0200 [scrapy] DEBUG: Enabled downloader middlewares: HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware, DefaultHeadersMiddleware, RedirectMiddleware, CookiesMiddleware, HttpCompressionMiddleware, ChunkedTransferMiddleware, DownloaderStats
2013-08-09 19:18:08+0200 [scrapy] DEBUG: Enabled spider middlewares: HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddleware
2013-08-09 19:18:08+0200 [scrapy] DEBUG: Enabled item pipelines: 
Traceback (most recent call last):
  File "/Library/Frameworks/Python.framework/Versions/2.7/bin/scrapy", line 5, in <module>
    pkg_resources.run_script('Scrapy==0.16.5', 'scrapy')
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/pkg_resources.py", line 499, in run_script
    self.require(requires)[0].run_script(script_name, ns)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/pkg_resources.py", line 1235, in run_script
    execfile(script_filename, namespace, namespace)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/EGG-INFO/scripts/scrapy", line 4, in <module>
    execute()
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/scrapy/cmdline.py", line 131, in execute
    _run_print_help(parser, _run_command, cmd, args, opts)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/scrapy/cmdline.py", line 76, in _run_print_help
    func(*a, **kw)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/scrapy/cmdline.py", line 138, in _run_command
    cmd.run(args, opts)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/scrapy/commands/crawl.py", line 43, in run
    spider = self.crawler.spiders.create(spname, **opts.spargs)
  File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/Scrapy-0.16.5-py2.7.egg/scrapy/spidermanager.py", line 43, in create
    raise KeyError("Spider not found: %s" % spider_name)
KeyError: 'Spider not found: dmoz'

我不明白哪里出了問題,但是鑒於我是編程的新手,這可能很容易。

您需要位於包含scrapy.cfg的目錄中:

stav@maia:/srv/scrapy/tutorial$ ls
scrapy.cfg  tutorial/

這是我系統上項目中文件的樹狀列表:

stav@maia:/srv/scrapy/tutorial$ tree
.
├── scrapy.cfg
└── tutorial
    ├── __init__.py
    ├── items.py
    ├── pipelines.py
    ├── settings.py
    └── spiders
        ├── dmoz_spider.py
        └── __init__.py

2 directories, 13 files

您應該向我們展示用於執行命令的整個命令行,包括工作目錄:

stav@maia:/srv/scrapy/tutorial$ scrapy crawl dmoz
2013-08-11 11:00:23-0500 [scrapy] INFO: Scrapy 0.17.0 started (bot: tutorial)
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Optional features available: ssl, django, http11, boto, libxml2
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Overridden settings: {'NEWSPIDER_MODULE': 'tutorial.spiders', 'SPIDER_MODULES': ['tutorial.spiders'], 'USER_AGENT': 'tutorial/1.0', 'BOT_NAME': 'tutorial'}
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole, CloseSpider, WebService, CoreStats, SpiderState
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Enabled downloader middlewares: HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware, DefaultHeadersMiddleware, MetaRefreshMiddleware, HttpCompressionMiddleware, RedirectMiddleware, CookiesMiddleware, ChunkedTransferMiddleware, DownloaderStats
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Enabled spider middlewares: HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddleware
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Enabled item pipelines:
2013-08-11 11:00:23-0500 [dmoz] INFO: Spider opened
2013-08-11 11:00:23-0500 [dmoz] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min)
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Telnet console listening on 0.0.0.0:6023
2013-08-11 11:00:23-0500 [scrapy] DEBUG: Web service listening on 0.0.0.0:6080
2013-08-11 11:00:24-0500 [dmoz] DEBUG: Crawled (200) <GET http://www.dmoz.org/Computers/Programming/Languages/Python/Books/> (referer: None)
2013-08-11 11:00:24-0500 [dmoz] DEBUG: Crawled (200) <GET http://www.dmoz.org/Computers/Programming/Languages/Python/Resources/> (referer: None)
2013-08-11 11:00:24-0500 [dmoz] INFO: Closing spider (finished)
2013-08-11 11:00:24-0500 [dmoz] INFO: Dumping Scrapy stats:
    {'downloader/request_bytes': 486,
     'downloader/request_count': 2,
     'downloader/request_method_count/GET': 2,
     'downloader/response_bytes': 12980,
     'downloader/response_count': 2,
     'downloader/response_status_count/200': 2,
     'finish_reason': 'finished',
     'finish_time': datetime.datetime(2013, 8, 11, 16, 0, 24, 101947),
     'log_count/DEBUG': 10,
     'log_count/INFO': 4,
     'response_received_count': 2,
     'scheduler/dequeued': 2,
     'scheduler/dequeued/memory': 2,
     'scheduler/enqueued': 2,
     'scheduler/enqueued/memory': 2,
     'start_time': datetime.datetime(2013, 8, 11, 16, 0, 23, 408890)}
2013-08-11 11:00:24-0500 [dmoz] INFO: Spider closed (finished)

如果上述解決方案不起作用,則::

打開教程文件夾中的settings.py並進行以下更改

BOT_NAME = 'dmoz'

使用您在dmoz_spider.py文件中明確定義的名稱,將“ BOT_NAME 'tutorial'的名稱從'tutorial'更改為'tutorial'

您在Virtualenv中運行嗎? 如果是這樣,請pip freeze一個pip freeze並向我們​​顯示是否已安裝所有scrapy依賴項

代碼很好,我只是復制粘貼了您的代碼,然后就可以毫無問題地運行它。 同樣,您應該能夠在scrapy項目文件夾中的任何位置運行spider。

請確保dmoz_spider.py位於子目錄“ spiders”中

mv dmoz_spider.py蜘蛛/。

暫無
暫無

聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.

 
粵ICP備18138465號  © 2020-2024 STACKOOM.COM