快速搭建hadoop环境
2014-11-21 14:37
148 查看
1,准备三台linux机器,配置好Java环境,最好JAVA_HOME的值都配成一样的。
2,在三台机器上分别解压Hadoop安装包,解压路径也需要一样。
vi /etc/profile
export JAVA_HOME=/opt/jdk1.7.0_55
export HADOOP_HOME=/usr/local/hadoop-2.5.1
export CLASSPATH=.:$JAVA_HOME/lib
export PATH=$PATH:$JAVA_HOME/jre/bin:$HADOOP_HOME/bin
source /etc/profile
2,给三台机器建Hadoop的用户
sudo addgroup hadoop
sudo adduser --ingroup hadoop hadoop
ls -l /etc/sudoers
chmod u+w /etc/sudoers
add ---> hadoop ALL=(ALL) ALL
chmod u-w /etc/sudoers
sudo chown -R hadoop:hadoop hadoop2.5.1
3,让三台机器可以在Hadoop用户下免密SSH互通
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
4, vi /etc/hosts 添加三台机器的Host Name和IP
192.168.0.2 master
192.168.0.3 slave1
192.168.0.4 slave2
5,同步三台机器的时间
crontab -e */5 * * * * /usr/sbin/ntpdate ntp.api.bz &> /dev/null
6,修改Hadoop的配置文件
=======================================================
slave
192.168.0.3
192.168.0.4
==================================================
core-site.xml
<configuration>
<!-- file system properties -->
<property>
<name>fs.default.name</name>
<value>hdfs://123.57.46.179:9000</value>
</property>
<!--- global properties -->
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop-2.5.1/tmp</value>
</property>
</configuration>
====================================================
hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>2</value>
</property>
</configuration>
=====================================================
mapred-site.xml
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>123.57.46.179:9001</value>
</property>
</configuration>
===============================
hadoop namenode -format
start-all.sh
./start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
Starting namenodes on [master]
master: starting namenode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-namenode-master.out
slave1: starting datanode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-datanode-slave1.out
slave2: starting datanode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-datanode-slave2.out
starting yarn daemons
starting resourcemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-resourcemanager-master.out
slave1: starting nodemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-nodemanager-slave1.out
slave2: starting nodemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-nodemanager-slave2.out
验证启动如下:
JPS
hadoop dfsadmin -report
hadoop dfs -mkdir testIn
hadoop dfs -ls
hdfs://localhost:9000/user/hadoop/testIn
hadoop dfs -copyFromLocal /tmp/test.* tesIn
hadoop jar hadooptest.jar wordcount testIn testOut
2,在三台机器上分别解压Hadoop安装包,解压路径也需要一样。
vi /etc/profile
export JAVA_HOME=/opt/jdk1.7.0_55
export HADOOP_HOME=/usr/local/hadoop-2.5.1
export CLASSPATH=.:$JAVA_HOME/lib
export PATH=$PATH:$JAVA_HOME/jre/bin:$HADOOP_HOME/bin
source /etc/profile
2,给三台机器建Hadoop的用户
sudo addgroup hadoop
sudo adduser --ingroup hadoop hadoop
ls -l /etc/sudoers
chmod u+w /etc/sudoers
add ---> hadoop ALL=(ALL) ALL
chmod u-w /etc/sudoers
sudo chown -R hadoop:hadoop hadoop2.5.1
3,让三台机器可以在Hadoop用户下免密SSH互通
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
4, vi /etc/hosts 添加三台机器的Host Name和IP
192.168.0.2 master
192.168.0.3 slave1
192.168.0.4 slave2
5,同步三台机器的时间
crontab -e */5 * * * * /usr/sbin/ntpdate ntp.api.bz &> /dev/null
6,修改Hadoop的配置文件
=======================================================
slave
192.168.0.3
192.168.0.4
==================================================
core-site.xml
<configuration>
<!-- file system properties -->
<property>
<name>fs.default.name</name>
<value>hdfs://123.57.46.179:9000</value>
</property>
<!--- global properties -->
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop-2.5.1/tmp</value>
</property>
</configuration>
====================================================
hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>2</value>
</property>
</configuration>
=====================================================
mapred-site.xml
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>123.57.46.179:9001</value>
</property>
</configuration>
===============================
hadoop namenode -format
start-all.sh
./start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
Starting namenodes on [master]
master: starting namenode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-namenode-master.out
slave1: starting datanode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-datanode-slave1.out
slave2: starting datanode, logging to /usr/local/hadoop-2.5.1/logs/hadoop-hadoop-datanode-slave2.out
starting yarn daemons
starting resourcemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-resourcemanager-master.out
slave1: starting nodemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-nodemanager-slave1.out
slave2: starting nodemanager, logging to /usr/local/hadoop-2.5.1/logs/yarn-hadoop-nodemanager-slave2.out
验证启动如下:
JPS
hadoop dfsadmin -report
hadoop dfs -mkdir testIn
hadoop dfs -ls
hdfs://localhost:9000/user/hadoop/testIn
hadoop dfs -copyFromLocal /tmp/test.* tesIn
hadoop jar hadooptest.jar wordcount testIn testOut
相关文章推荐
- docker + ambari(hortonworks ) 快速搭建hadoop 环境
- hadoop环境快速搭建
- 高效5步走,快速搭建Hadoop2伪分布环境
- VMWare下快速搭建Hadoop完全分布式环境
- 详解从 0 开始使用 Docker 快速搭建 Hadoop 集群环境
- Hadoop初体验:快速搭建Hadoop伪分布式环境
- Win7下快速搭建hadoopEclipse运行环境--方案
- Win7下快速搭建hadoopEclipse运行环境--方案
- 快速搭建Hadoop及HBase分布式环境
- 【hadoop】hadoop环境快速搭建
- 在不到1G内存机器上快速搭建一个hadoop环境
- 快速搭建 Hadoop 分布式环境
- 快速搭建 Hadoop 环境 推荐
- hadoop3.1.0 HDFS快速搭建伪分布式环境
- Eclipse3.4+Maven2+jetty6快速搭建web快速开发环境
- 搭建本地hadoop测试环境
- 主题:推荐一个eclipse插件:一键式快速搭建Struts2.0,SSH2环境
- PHPnow - 快速搭建 PHP 运行环境
- 关于Cygwin环境下快速搭建GTK+2.0开发环境的方法(原创)
- 快速搭建PHP开发环境WAMP+ZendStudio+ZendDebugger