How To Build Scrapy Environment
YetToCome
wjw15129 at gmail.com
Mon Sep 23 04:56:27 EDT 2013
在 2013年9月23日星期一UTC+8下午4时37分22秒,Peter Otten写道:
> YetToCome wrote:
>
>
>
> > I have already install Twisted, zope.interface, w3lib, libxml2, etc,but it
>
> > still can not be built,here is the error message:
>
> >
>
> > Traceback (most recent call last):
>
> > File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
>
> > "__main__", fname, loader, pkg_name)
>
> > File "C:\Python27\lib\runpy.py", line 72, in _run_code
>
> > exec code in run_globals
>
> > File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 167, in
>
> > <module>
>
> > execute()
>
> > File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 142, in
>
> > execute
>
> > _run_print_help(parser, _run_command, cmd, args, opts)
>
> > File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 88, in
>
> > _run_print
>
> > _help
>
> > func(*a, **kw)
>
> > File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 149, in
>
> > _run_comm
>
> > and
>
> > cmd.run(args, opts)
>
> > File "C:\Python27\lib\site-packages\scrapy\commands\crawl.py", line 47,
>
> > in run
>
> > crawler = self.crawler_process.create_crawler()
>
> > File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 142, in
>
> > create_cr
>
> > awler
>
> > self.crawlers[name] = Crawler(self.settings)
>
> > File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 23, in
>
> > __init__
>
> > self.spiders = spman_cls.from_crawler(self)
>
> > File "C:\Python27\lib\site-packages\scrapy\spidermanager.py", line 35,
>
> > in from
>
> > _crawler
>
> > sm = cls.from_settings(crawler.settings)
>
> > File "C:\Python27\lib\site-packages\scrapy\spidermanager.py", line 31,
>
> > in from
>
> > _settings
>
> > return cls(settings.getlist('SPIDER_MODULES'))
>
> > File "C:\Python27\lib\site-packages\scrapy\spidermanager.py", line 22,
>
> > in __in
>
> > it__
>
> > for module in walk_modules(name):
>
> > File "C:\Python27\lib\site-packages\scrapy\utils\misc.py", line 66, in
>
> > walk_mo
>
> > dules
>
> > submod = __import__(fullpath, {}, {}, [''])
>
> > File "blog_crawl\spiders\dmoz_spider.py", line 1, in <module>
>
> > class DmozSpider(BaseSpider):
>
> > NameError: name 'BaseSpider' is not defined
>
> >
>
> > someone tells me why, pls
>
>
>
> Assuming you are working your way through the tutorial at
>
>
>
> http://doc.scrapy.org/en/latest/intro/tutorial.html#our-first-spider
>
>
>
> you probably forgot to import the BaseSpider class with
>
>
>
> from scrapy.spider import BaseSpider
>
>
>
> as shown in the code snippet in the "Our first Spider" section of the
>
> tutorial.
it had an another error: No module named queuelib, but i have installed all the libs mentioned in that passage...
2013-09-23 16:44:17+0800 [scrapy] INFO: Scrapy 0.18.2 started (bot: tutorial)
2013-09-23 16:44:17+0800 [scrapy] DEBUG: Optional features available: ssl, http1
1, libxml2
2013-09-23 16:44:17+0800 [scrapy] DEBUG: Overridden settings: {'NEWSPIDER_MODULE
': 'tutorial.spiders', 'SPIDER_MODULES': ['tutorial.spiders'], 'BOT_NAME': 'tuto
rial'}
2013-09-23 16:44:17+0800 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetCon
sole, CloseSpider, WebService, CoreStats, SpiderState
Traceback (most recent call last):
File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "C:\Python27\lib\runpy.py", line 72, in _run_code
exec code in run_globals
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 167, in <module>
execute()
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 142, in execute
_run_print_help(parser, _run_command, cmd, args, opts)
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 88, in _run_print
_help
func(*a, **kw)
File "C:\Python27\lib\site-packages\scrapy\cmdline.py", line 149, in _run_comm
and
cmd.run(args, opts)
File "C:\Python27\lib\site-packages\scrapy\commands\crawl.py", line 50, in run
self.crawler_process.start()
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 93, in start
if self.start_crawling():
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 168, in start_cra
wling
return self.start_crawler() is not None
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 151, in start_cra
wler
crawler.configure()
File "C:\Python27\lib\site-packages\scrapy\crawler.py", line 45, in configure
self.engine = ExecutionEngine(self, self._spider_closed)
File "C:\Python27\lib\site-packages\scrapy\core\engine.py", line 61, in __init
__
self.scheduler_cls = load_object(self.settings['SCHEDULER'])
File "C:\Python27\lib\site-packages\scrapy\utils\misc.py", line 40, in load_ob
ject
raise ImportError, "Error loading object '%s': %s" % (path, e)
ImportError: Error loading object 'scrapy.core.scheduler.Scheduler': No module n
amed queuelib
More information about the Python-list
mailing list