几个因为hadoop配置文件不当造成的错误
2013-11-21 23:55
316 查看
192.168.1.20: Exception in thread "main" java.lang.IllegalArgumentException
192.168.1.20: at java.net.URI.create(URI.java:841)
192.168.1.20: at org.apache.hadoop.fs.FileSystem.getDefaultUri(FileSystem.java:132)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:236)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:194)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:150)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:676)
192.168.1.20: Caused by: java.net.URISyntaxException: Illegal character in authority at index 7: hdfs://192.168.1.20:9000
192.168.1.20: at java.net.URI$Parser.fail(URI.java:2810)
The reason is in core-site.xml hdfs://master:9000 </value>多了一个空格。
[hadoop@Master bin]$ start-all.sh
Warning: $HADOOP_HOME is deprecated.
starting namenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-namenode-Master.Hadoop.out
192.168.1.23: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave3.Hadoop.out
192.168.1.21: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave1.Hadoop.out
192.168.1.22: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave2.Hadoop.out
192.168.1.21: [Fatal Error] core-site.xml:24:1: Content is not allowed in trailing section.
192.168.1.20: starting secondarynamenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-secondarynamenode-Master.Hadoop.out
starting jobtracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-jobtracker-Master.Hadoop.out
192.168.1.23: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave3.Hadoop.out
192.168.1.22: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave2.Hadoop.out
192.168.1.21: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave1.Hadoop.out
192.168.1.21: [Fatal Error] core-site.xml:24:1: Content is not allowed in trailing section.
core-site.xml 后边有空格或者空行
[hadoop@Master bin]$ start-all.sh
Warning: $HADOOP_HOME is deprecated.
starting namenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-namenode-Master.Hadoop.out
192.168.1.23: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave3.Hadoop.out
192.168.1.21: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave1.Hadoop.out
192.168.1.22: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave2.Hadoop.out
192.168.1.20: starting secondarynamenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-secondarynamenode-Master.Hadoop.out
starting jobtracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-jobtracker-Master.Hadoop.out
192.168.1.22: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave2.Hadoop.out
192.168.1.23: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave3.Hadoop.out
192.168.1.21: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave1.Hadoop.out
[hadoop@Master bin]$ cd
[hadoop@Master ~]$ ll /usr/hadoop/tmp
total 4
drwxrwxr-x 4 hadoop hadoop 4096 Oct 29 13:31 dfs
succeed in Master!
BUT not open MapReduce in Slaves
Last login: Tue Oct 29 11:26:27 2013 from superserver.home
[hadoop@Slave3 ~]$ ll /usr/hadoop/tmp/
total 4
drwxrwxr-x 3 hadoop hadoop 4096 Oct 29 13:31 dfs
[hadoop@Slave3 ~]$
Problable Solution:
The reason is in mapred-site.xml http://192.168.1.20:9000 </value>多了一个空格
192.168.1.20: at java.net.URI.create(URI.java:841)
192.168.1.20: at org.apache.hadoop.fs.FileSystem.getDefaultUri(FileSystem.java:132)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:236)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:194)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:150)
192.168.1.20: at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:676)
192.168.1.20: Caused by: java.net.URISyntaxException: Illegal character in authority at index 7: hdfs://192.168.1.20:9000
192.168.1.20: at java.net.URI$Parser.fail(URI.java:2810)
The reason is in core-site.xml hdfs://master:9000 </value>多了一个空格。
[hadoop@Master bin]$ start-all.sh
Warning: $HADOOP_HOME is deprecated.
starting namenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-namenode-Master.Hadoop.out
192.168.1.23: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave3.Hadoop.out
192.168.1.21: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave1.Hadoop.out
192.168.1.22: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave2.Hadoop.out
192.168.1.21: [Fatal Error] core-site.xml:24:1: Content is not allowed in trailing section.
192.168.1.20: starting secondarynamenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-secondarynamenode-Master.Hadoop.out
starting jobtracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-jobtracker-Master.Hadoop.out
192.168.1.23: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave3.Hadoop.out
192.168.1.22: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave2.Hadoop.out
192.168.1.21: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave1.Hadoop.out
192.168.1.21: [Fatal Error] core-site.xml:24:1: Content is not allowed in trailing section.
core-site.xml 后边有空格或者空行
[hadoop@Master bin]$ start-all.sh
Warning: $HADOOP_HOME is deprecated.
starting namenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-namenode-Master.Hadoop.out
192.168.1.23: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave3.Hadoop.out
192.168.1.21: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave1.Hadoop.out
192.168.1.22: starting datanode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-datanode-Slave2.Hadoop.out
192.168.1.20: starting secondarynamenode, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-secondarynamenode-Master.Hadoop.out
starting jobtracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-jobtracker-Master.Hadoop.out
192.168.1.22: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave2.Hadoop.out
192.168.1.23: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave3.Hadoop.out
192.168.1.21: starting tasktracker, logging to /usr/hadoop/libexec/../logs/hadoop-hadoop-tasktracker-Slave1.Hadoop.out
[hadoop@Master bin]$ cd
[hadoop@Master ~]$ ll /usr/hadoop/tmp
total 4
drwxrwxr-x 4 hadoop hadoop 4096 Oct 29 13:31 dfs
succeed in Master!
BUT not open MapReduce in Slaves
Last login: Tue Oct 29 11:26:27 2013 from superserver.home
[hadoop@Slave3 ~]$ ll /usr/hadoop/tmp/
total 4
drwxrwxr-x 3 hadoop hadoop 4096 Oct 29 13:31 dfs
[hadoop@Slave3 ~]$
Problable Solution:
The reason is in mapred-site.xml http://192.168.1.20:9000 </value>多了一个空格
相关文章推荐
- 安装Hadoop需要配置的几个文件
- Hadoop-2.8.0集群搭建、hadoop源码编译和安装、host配置、ssh免密登录、hadoop配置文件中的参数配置参数总结、hadoop集群测试,安装过程中的常见错误
- 我用myeclipse导入了一个用eclipse开发的项目,项目上有个差报错,是里面几个页面报的假错,部署到tomcat后,tomcat配置里这个项目的running值是false,访问就报404错误,而且不编译class文件
- hadoop2.2.0安装的几个配置文件
- 在电脑上找到这个路径:D:\jakarta-tomcat-6\conf\Catalina,将localhost目录直接删除掉,再一运行,就没事了! 因为配置文件中配置了启动程序,而webapps文件夹下却没有此应用程序,所以出现了上述错误。
- IIS:错误: 无法提交配置更改,因为文件已在磁盘上更改
- hibernate经常报的几个不起眼的错误, 都是因为不细心或者手滑了输入有误造成了
- HTTP错误404.13 - Not Found 请求筛选模块被配置为拒绝超过请求内容长度的请求,原因是Web服务器上的请求筛选被配置为拒绝该请求,因为内容长度超过配置的值(IIS 7 默认文件上传大
- Detailed Error Information: Module IIS Web Core Notification 未知 Handler 尚未确定 Error Code 0x80070032 配置错误 无法读取配置节“spring”,因为它缺少节声明 配置文件 \\?\d:\用户目录\我的文档\visual studio 2015\Projects\LiS
- 解决整合spring和mybatis后数据源配置文件读取错误,错误提示数据库用户名密码错误。
- 【阿里聚安全·安全周刊】阿里双11技术十二讲直播预约|AWS S3配置错误曝光NSA陆军机密文件
- Urlrewrite配置不当造成死循环问题解析
- php-fpm nginx配置文件 处理 Permission denied 错误
- hadoop fs -put 上传文件遇到错误2 hadoop fs -put /opt/exercise/ds.txt ds.txt
- hadoop2配置文件详细解释
- hadoop配置文件加载顺序
- PS常见错误-无法完成请求,因为文件格式模块不能解析该文件
- 文件内码不同造成的错误
- 用VS2005在Release配置下编译出现清单文件的错误的解决方法
- C访问hadoop程序终端显示运行正确,因为连接参数错误,使得通过网页查看就是没有成功原因分析和解决方案