3,数据挖掘环境搭建-Hadoop集群搭建与测试
2017-11-08 23:09
441 查看
安装hadoop-2.6.5.tar.gz
hdp10 namenode/datanode
hdp11 datanode
hdp12 datanode
0 安装前准备
修改hosts,修改主机名,关闭防火墙,关闭selinux,安装ssh,配置免密码,安装JDK
1,添加hadoop用户名
useradd hadoop
passwd hadoop
vi /etc/sudoers
2,修改/home/hadoop/apps/hadoop-2.6.5/etc/hadoop下的配置文件
2.1,修改hadoop-env.sh
vi hadoop-env.sh
2.2 修改core-site.xml
vi core-site.xml
2.3,修改hdfs-site.xml
vi hdfs-site.xml
2.4 修改mapred-site.xml
vi mapred-site.xml
2.5,修改yarn-site.xml
vi yarn-site.xml
3,将hadoop文件添加到系统环境变量
vi /etc/profile
4,修改slaves,设置datanode与nodemanager
vi /home/hadoop/apps/hadoop-2.6.5/etc/hadoop/slaves
hdp10
hdp11
hdp12
5,将/etc/profile和hadoop文件夹拷贝至hdp11,hdp12相应的位置
scp /etc/profile root@hdp11:/etc
scp /etc/profile root@hdp12:/etc
scp -r /home/hadoop/apps/hadoop-2.6.5 root@hdp11:/home/hadoop/apps
scp -r /home/hadoop/apps/hadoop-2.6.5 root@hdp12:/home/hadoop/apps
chown -R hadoop:hadoop /home/hadoop/apps/hadoop-2.6.5
6,去除HDFS WARN util.NativeCodeLoader
vi /home/hadoop/apps/hadoop-2.5.2/etc/hadoop/log4j.properties
log4j.logger.org.apache.hadoop.util.NativeCodeLoader=ERROR
7,格式化hadoop文件系统
hdfs namenode -format
8,启动hadoop
cd /home/hadoop/apps/hadoop-2.6.5/sbin
start-dfs.sh,启动HDFS
stop-dfs.sh,关闭HDFS
start-yarn.sh,启动YARN
stop-yarn.sh, 关闭YARN
9,查看启动状态
jps
10,范例
vi 1.txt
hadoop fs -mkdir -p /wordcount/input
hadoop fs -put 1.txt /wordcount/input
cd /home/hadoop/apps/hadoop-2.6.5/share/hadoop/mapreduce
hadoop jar hadoop-mapreduce-examples-2.6.5.jar wordcount /wordcount/input /wordcount/output0
hdp10 namenode/datanode
hdp11 datanode
hdp12 datanode
0 安装前准备
修改hosts,修改主机名,关闭防火墙,关闭selinux,安装ssh,配置免密码,安装JDK
1,添加hadoop用户名
useradd hadoop
passwd hadoop
vi /etc/sudoers
root ALL=(ALL) ALL hadoop ALL=(ALL) ALL
2,修改/home/hadoop/apps/hadoop-2.6.5/etc/hadoop下的配置文件
2.1,修改hadoop-env.sh
vi hadoop-env.sh
export JAVA_HOME=/usr/java/jdk1.8.0_144
2.2 修改core-site.xml
vi core-site.xml
<configuration> <property> <name>fs.defaultFS</name> <value>hdfs://hdp10:9000</value> </property> <property> <name>hadoop.tmp.dir</name> <value>/home/hadoop/hdpdata</value> </property> </configuration>
2.3,修改hdfs-site.xml
vi hdfs-site.xml
<configuration> <property> <name>dfs.replication</name> <value>3</value> </property> <property> <name>hadoop.tmp.dir</name> <value>/home/hadoop/hdpdata</value> </property> <property> <name>dfs.secondary.http.address</name> <value>hdp10:50070</value> </property> <property> <name>dfs.http.address</name> <value>hdp10:50090</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>/home/hadoop/datanode</value> </property> <property> <name>dfs.namenode.data.dir</name> <value>/home/hadoop/namenode</value> </property> <property> <name>dfs.permissions</name> <value>false</value> </property> </configuration>
2.4 修改mapred-site.xml
vi mapred-site.xml
<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> </configuration>
2.5,修改yarn-site.xml
vi yarn-site.xml
<configuration> <property> <name>yarn.resourcemanager.hostname</name> <value>hdp10</value> </property> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> </configuration>
3,将hadoop文件添加到系统环境变量
vi /etc/profile
export HADOOP_HOME=/home/hadoop/apps/hadoop-2.6.5 export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
4,修改slaves,设置datanode与nodemanager
vi /home/hadoop/apps/hadoop-2.6.5/etc/hadoop/slaves
hdp10
hdp11
hdp12
5,将/etc/profile和hadoop文件夹拷贝至hdp11,hdp12相应的位置
scp /etc/profile root@hdp11:/etc
scp /etc/profile root@hdp12:/etc
scp -r /home/hadoop/apps/hadoop-2.6.5 root@hdp11:/home/hadoop/apps
scp -r /home/hadoop/apps/hadoop-2.6.5 root@hdp12:/home/hadoop/apps
chown -R hadoop:hadoop /home/hadoop/apps/hadoop-2.6.5
6,去除HDFS WARN util.NativeCodeLoader
vi /home/hadoop/apps/hadoop-2.5.2/etc/hadoop/log4j.properties
log4j.logger.org.apache.hadoop.util.NativeCodeLoader=ERROR
7,格式化hadoop文件系统
hdfs namenode -format
8,启动hadoop
cd /home/hadoop/apps/hadoop-2.6.5/sbin
start-dfs.sh,启动HDFS
stop-dfs.sh,关闭HDFS
start-yarn.sh,启动YARN
stop-yarn.sh, 关闭YARN
9,查看启动状态
jps
10,范例
vi 1.txt
hello world hello count hello class hello rose hello jerry hello class hello rose
hadoop fs -mkdir -p /wordcount/input
hadoop fs -put 1.txt /wordcount/input
cd /home/hadoop/apps/hadoop-2.6.5/share/hadoop/mapreduce
hadoop jar hadoop-mapreduce-examples-2.6.5.jar wordcount /wordcount/input /wordcount/output0
相关文章推荐
- (大数据之hadoop)Hadoop2.2.0版本多节点集群及eclipse开发环境搭建
- Java+大数据开发——Hadoop集群环境搭建(一)
- 5,数据挖掘环境搭建-Zookeeper集群搭建
- [大数据]连载No1之Hadoop概念和伪分布式集群环境搭建
- 大数据环境搭建-之-hadoop 2.x分布式部署-集群配置
- 4,数据挖掘环境搭建-hadoop-2.6.5-src.tar.gz编译
- Java+大数据开发——Hadoop集群环境搭建(二)
- 大数据测试之hadoop单机环境搭建(超级详细版)
- hadoop spark 大数据集群环境搭建(一)
- 6,数据挖掘环境搭建-Spark集群搭建
- 大数据测试之hadoop单机环境搭建(超级详细版)
- Hadoop完全分布式集群环境搭建及测试
- MongoDB分片集群环境搭建好后简单的创建和测试数据库表
- DayDayUP_大数据学习课程[1]_hadoop2.6.0完全分布式集群环境和伪分布式集群搭建
- 大数据1-hadoop、zookeeper、hbase、spark集群环境搭建
- 大数据平台 Hadoop 的分布式集群环境搭建
- 搭建Hadoop分布式集群------测试Hadoop分布式集群环境
- 大数据hadoop入门学习之集群环境搭建集合
- 【大数据】开发环境搭建(三):hadoop伪分布式集群环境搭建(下)
- 大数据测试之hadoop单机环境搭建(超级详细版)