python

关注公众号 jb51net

关闭
首页 > 脚本专栏 > python > Scrapy启动报错invalid syntax

Scrapy启动报错invalid syntax的解决

作者:紫陌幽茗

这篇文章主要介绍了Scrapy启动报错invalid syntax的解决方案,具有很好的参考价值,希望对大家有所帮助。如有错误或未考虑完全的地方,望不吝赐教

Scrapy启动报错invalid syntax

报错信息

Traceback (most recent call last): File “d:\python\lib\runpy.py”,
line 193, in _run_module_as_main
“main”, mod_spec) File “d:\python\lib\runpy.py”, line 85, in run_code
exec(code, run_globals) File "D:\python\Scripts\scrapy.exe_main.py", line 9, in File
“d:\python\lib\site-packages\scrapy\cmdline.py”, line 150, in execute
_run_print_help(parser, _run_command, cmd, args, opts) File “d:\python\lib\site-packages\scrapy\cmdline.py”, line 90, in
_run_print_help
func(*a, **kw) File “d:\python\lib\site-packages\scrapy\cmdline.py”, line 157, in
_run_command
cmd.run(args, opts) File “d:\python\lib\site-packages\scrapy\commands\crawl.py”, line 57, in
run
self.crawler_process.crawl(spname, **opts.spargs) File “d:\python\lib\site-packages\scrapy\crawler.py”, line 170, in crawl
crawler = self.create_crawler(crawler_or_spidercls) File “d:\python\lib\site-packages\scrapy\crawler.py”, line 198, in
create_crawler
return self._create_crawler(crawler_or_spidercls) File “d:\python\lib\site-packages\scrapy\crawler.py”, line 203, in
create_crawler
return Crawler(spidercls, self.settings) File “d:\python\lib\site-packages\scrapy\crawler.py”, line 55, in init
self.extensions = ExtensionManager.from_crawler(self) File “d:\python\lib\site-packages\scrapy\middleware.py”, line 58, in
from_crawler
return cls.from_settings(crawler.settings, crawler) File “d:\python\lib\site-packages\scrapy\middleware.py”, line 34, in
from_settings
mwcls = load_object(clspath) File “d:\python\lib\site-packages\scrapy\utils\misc.py”, line 44, in
load_object
mod = import_module(module) File "d:\python\lib\importlib_init.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level) File “

启动命令如下:

scrapy crawl first_spider -o first_spider.csv

解决方法

经过查询发现async在python3.7中是保留字,因此将twisted\conch\manhole.py中的async修改为isasync即可解决问题

使用scrapy做爬虫遇到的一些坑

使用xpath和CSS遇到的一些 invalid syntax

1.引号的歧义:最外面应该使用单引号

错误:

正确:把定位的xpath外加单引号

2.使用了关键字做变量

错误:

正确:注意关键字不能做变量

3.进入shell窗口时,url添加了“”

错误:

正确:去掉url的引号

4.使用xpath匹配相关信息,返回为空列表

错误:因为一开始网址就是错误的,所以不管匹配的xpath是不是对的,都会返回【】

正确:

以上为个人经验,希望能给大家一个参考,也希望大家多多支持脚本之家。

您可能感兴趣的文章:
阅读全文