scrapy [boto] ERROR: Caught exception reading instance data URLError: <urlopen error [Errno 10051] >
2015-10-22 18:29
1526 查看
执行过程中出现错误:
在setting.py中禁用s3 download就可以了
解决方法:
在stackoverflow上有些解释,有时间可以看看这篇错误分析错误
下面是我贴的部分内容。
That particular error message is being generated by boto (boto 2.38.0 py27_0), which is used to connect to Amazon S3. Scrapy doesn’t have this enabled by default.
EDIT: In reply to the comments, this appears to be a bug with Scrapy when boto is present (bug here).
In response “how to disable the Download handler”, add the following to your settings.py file:
Your settings.py file should be in the root of your Scrapy project folder, (one level deeper than your scrapy.cfg file).
If you’ve already got DOWNLOAD_HANDLERS in your settings.py file, just add a new entry for ‘s3’ with a None value.
EDIT 2: I’d highly recommend looking at setting up virtual environments for your projects. Look into virtualenv, and it’s usage. I’d make this recommendation regardless of packages used for this project, but doubly so with your extreme number of packages.
2015-09-09 11:13:26 [boto] DEBUG: Retrieving credentials from metadata server. 2015-09-09 11:13:27 [boto] ERROR: Caught exception reading instance data Traceback (most recent call last): File "D:\anzhuang\Anaconda\lib\site-packages\boto\utils.py", line 210, in retry_url r = opener.open(req, timeout=timeout) File "D:\anzhuang\Anaconda\lib\urllib2.py", line 431, in open response = self._open(req, data) File "D:\anzhuang\Anaconda\lib\urllib2.py", line 449, in _open '_open', req) File "D:\anzhuang\Anaconda\lib\urllib2.py", line 409, in _call_chain result = func(*args) File "D:\anzhuang\Anaconda\lib\urllib2.py", line 1227, in http_open return self.do_open(httplib.HTTPConnection, req) File "D:\anzhuang\Anaconda\lib\urllib2.py", line 1197, in do_open raise URLError(err) URLError: <urlopen error timed out> 2015-09-09 11:13:27 [boto] ERROR: Unable to read instance data, giving up
在setting.py中禁用s3 download就可以了
解决方法:
DOWNLOAD_HANDLERS = {'S3': None,}
在stackoverflow上有些解释,有时间可以看看这篇错误分析错误
下面是我贴的部分内容。
That particular error message is being generated by boto (boto 2.38.0 py27_0), which is used to connect to Amazon S3. Scrapy doesn’t have this enabled by default.
EDIT: In reply to the comments, this appears to be a bug with Scrapy when boto is present (bug here).
In response “how to disable the Download handler”, add the following to your settings.py file:
DOWNLOAD_HANDLERS = { 'S3':None, }
Your settings.py file should be in the root of your Scrapy project folder, (one level deeper than your scrapy.cfg file).
If you’ve already got DOWNLOAD_HANDLERS in your settings.py file, just add a new entry for ‘s3’ with a None value.
EDIT 2: I’d highly recommend looking at setting up virtual environments for your projects. Look into virtualenv, and it’s usage. I’d make this recommendation regardless of packages used for this project, but doubly so with your extreme number of packages.
相关文章推荐
- linux mount nobody nogroup
- opencv2感兴趣区域以及图像尺寸求解
- odoo8编辑视图中sheet边距过宽问题调整
- Tomcat 并发优化(收藏)
- linux下踢出已登录用户
- CentOS6.7 常用操作命令
- linux python升级过程
- 常用Linux操作命令总结
- Linux Shell命令行快捷键
- Linux修改SSH连接数 重启SSH服务
- 十大好用的Linux实用工具推荐
- vs编译项目报错:The OutputPath property is not set for this project
- CentOS7下LVS+Keepalived实现高性能高可用负载均衡
- Hadoop单机伪分布式配置
- linux中sd卡驱动(2)
- Linux Ubuntu Nginx负载均衡 安装 配置
- 垂直门户网站和电商的SEO思路
- linux中sd卡驱动(1)
- Linux进程间通信(五) - 信号灯(史上最全)及其经典应用案例
- 为什么linux物理内存还有很多,却开始使用swap? NUMA - 罪魁祸首