Centos6.5中编译hadoop2.x 并安装 运行wordCount
2018-03-06 17:10
579 查看
下载安装包:
http://mirror.bit.edu.cn/apache/maven/maven-3/3.0.5/binaries/
解压:
tar -zxvf apache-maven-3.0.5-bin.tar.gz
设置环境变量
安装svn
yun install svn
安装libtool cmake
yum install autoconf automake libtool cmake
安装ncurses-devel
yum install ncurses-devel
安装openssl-devel
yum install openssl-devel
安装gcc
yum install gcc*
安装并设置protobuf
下载 http://pan.baidu.com/s/1pJlZubT
下载hadoop源包
svn checkout http://svn.apache.org/repos/asf/hadoop/common/tags/release-2.2.0
编译包
mvn package -Pdist,native -DskipTests -Dtar
file ./libhadoop.so.1.0.0
验证是否为64位
hadoop fs -mkdir -p /class3/input
准备测试数据
hadoop fs -copyFromLocal ../etc/hadoop/* /class3/input
运行wordCount
hadoop jar ../share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar wordcount /class3/input /class3/output
查看运行结果
hadoop fs -cat /class3/output/part-r-00000 | less
http://mirror.bit.edu.cn/apache/maven/maven-3/3.0.5/binaries/
解压:
tar -zxvf apache-maven-3.0.5-bin.tar.gz
设置环境变量
export MAVEN_HOME=/app/lib/apache-maven-3.0.5 export PATH=$PATH:$MAVEN_HOME/bin
安装svn
yun install svn
安装libtool cmake
yum install autoconf automake libtool cmake
安装ncurses-devel
yum install ncurses-devel
安装openssl-devel
yum install openssl-devel
安装gcc
yum install gcc*
安装并设置protobuf
下载 http://pan.baidu.com/s/1pJlZubT
./configure make make check make install //验证 protoc
下载hadoop源包
svn checkout http://svn.apache.org/repos/asf/hadoop/common/tags/release-2.2.0
编译包
mvn package -Pdist,native -DskipTests -Dtar
file ./libhadoop.so.1.0.0
验证是否为64位
安装hadoop 2.x
解压安装包 mkdir tmp mkdir hdfs mkdir hdfs/name mkdir hdfs/data
配置hadoop-env.sh
export HADOOP_CONF_DIR=/app/hadoop-2.2.0/etc/hadoop export HADOOP_HOME=/app/hadoop-2.2.0 export PATH=$PATH:$HADOOP_HOME/bin export JAVA_HOME=/app/lib/jdk-8
配置/etc/profile
export HADOOP_HOME=/app/hadoop-2.2.0 export PATH=$PATH:$HADOOP_HOME/bin
配置yarn-env.sh
export JAVA_HOME=/app/lib/jdk-8
配置core-site.xml
<configuration> <property> <name>fs.default.name</name> <value>hdfs://hadoop:9000</value> </property> <property> <name>fs.defaultFS</name> </property> <property> <name>io.file.buffer.size</name> <value>131072</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/app/hadoop-2.2.0/tmp</value> <description>A base for other temporary directories.</description> </property> <property> <name>hadoop.proxyuser.hduser.hosts</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.hduser.groups</name> <value>*</value> </property> </configuration>
配置hdfs-site.xml
<configuration> <property> <name>dfs.namenode.secondary.http-address</name> <value>hadoop:9001</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/app/hadoop-2.2.0/hdfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/app/hadoop-2.2.0/hdfs/data</value> </property> <property> <name>dfs.replication</name> <value>1</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property> </configuration>
配置mapred-site.xml
<configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>hadoop:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>hadoop:19888</value> </property> </configuration>
配置yarn-site.xml
<configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.address</name> <value>hadoop:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>hadoop:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>hadoop:8031</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>hadoop:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>hadoop:8088</value> </property> </configuration>
设置slaves
vim slaves hadoop
设置slaves
./hadoop namenode -format ./start-dfs.sh ./start-yarn.sh jps: namenode datanode secondarynamenode resourcemanager nodemanager
测试wordCount
hdfs中创建目录hadoop fs -mkdir -p /class3/input
准备测试数据
hadoop fs -copyFromLocal ../etc/hadoop/* /class3/input
运行wordCount
hadoop jar ../share/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar wordcount /class3/input /class3/output
查看运行结果
hadoop fs -cat /class3/output/part-r-00000 | less
相关文章推荐
- 大数据之Hadoop平台(二)Centos6.5(64bit)Hadoop2.5.1伪分布式安装记录,wordcount运行测试
- 大数据之Hadoop平台(二)Centos6.5(64bit)Hadoop2.5.1伪分布式安装记录,wordcount运行测试
- CentOS安装Hadoop并运行WordCount实例
- centos6.5配置Hadoop环境,运行wordcount例子
- CentOS上安装Hadoop2.7,添加数据节点,运行wordcount
- CentOS6.8下Hadoop2.7.2怎么运行自带的wordcount程序
- win7下安装hadoop 2.6.0 的eclipse插件并编写运行WordCount程序
- Centos6.5源码编译安装Hadoop2.5.1
- Hadoop的安装与配置及示例程序wordcount的运行
- hadoop wordcount demo 编译、打包、运行(自己的经历)
- hadoop2.7.3 编译运行WordCount.java
- Hadoop在Linux下伪分布式的安装 wordcount实例的运行
- hadoop2.7.3 编译运行WordCount.java
- 运行hadoop的WordCount程序——编译,打包,运行
- Windows下Cygwin环境的Hadoop安装(3)- 运行hadoop中的wordcount实例遇到的问题和解决方法
- Ubuntu14.04安装配置Hadoop2.6.0(完全分布式)与 wordcount实例运行
- hadoop2.3安装和wordcount运行验证
- hadoop2.7.3 编译运行WordCount.java
- CentOS6.5+HADOOP2.7.1安装配置测试编译详细教程