Ubuntu下安装及配置单点hadoop
2012-02-21 14:49
387 查看
环境:Ubuntu10.10,Hadoop 0.21.0首先准备工作:
1,我是在deskTop版本的Ubuntu下安装的所以,需要先安装ssh server。这个很好找,直接到到新立得里搜索ssh,第一个就是。
2,安装sun jdk6,切忌一定要是java6及其以上版本。先到更新管理器里把canonical的源加进去。
之后sudo apt-get update.
3, sudo apt-get sun-java6-jdk
4, sudo update-java-alternatives -s java-6-sun
5,增加一个用户组用户,用于hadoop运行及访问。
sudo addgroup hadoop
sudo adduser --ingroup hadoop hadoop
复制代码
6,生成SSH证书,配置SSH加密key
user@ubuntu:~$ su - hadoop
hadoop@ubuntu:~$ ssh-keygen -t rsa -P ""
Generating public/private rsa key pair.
Enter file in which to save the key (/home/hadoop/.ssh/id_rsa):
Created directory '/home/hadoop/.ssh'.
Your identification has been saved in /home/hadoop/.ssh/id_rsa.
Your public key has been saved in /home/hadoop/.ssh/id_rsa.pub.
The key fingerprint is:
9b:82:ea:58:b4:e0:35:d7:ff:19:66:a6:ef:ae:0e:d2 hadoop@ubuntu
The key's randomart image is:
[...snipp...]
hadoop@ubuntu:~$
复制代码
hadoop@ubuntu:~$ cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
复制代码
配置完成,测试一下:
hadoop@ubuntu:~$ ssh localhost
The authenticity of host 'localhost (::1)' can't be established.
RSA key fingerprint is d7:87:25:47:ae:02:00:eb:1d:75:4f:bb:44:f9:36:26.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added 'localhost' (RSA) to the list of known hosts.
Linux ubuntu 2.6.32-22-generic #33-Ubuntu SMP Wed Apr 28 13:27:30 UTC 2010 i686 GNU/Linux
Ubuntu 10.04 LTS
[...snipp...]
hadoop@ubuntu:~$
复制代码
8,禁用ipV6配置:
打开/etc/sysctl.conf, 此文件需要root权限。
再次文件中,追加如下:
net.ipv6.conf.all.disable_ipv6 = 1
net.ipv6.conf.default.disable_ipv6 = 1
net.ipv6.conf.lo.disable_ipv6 = 1
复制代码
重启,测试是否配置成功:
$ cat /proc/sys/net/ipv6/conf/all/disable_ipv6
复制代码
如果是1就ok了。
下面就是安装Hadoop了。
首先是:下载,解压缩,分配权限。
下载就不说了。
下载后运行如下:
$ cd /usr/local
$ sudo tar xzf hadoop-0.20.2.tar.gz
$ sudo mv hadoop-0.20.2 hadoop
$ sudo chown -R hadoop:hadoop hadoop
复制代码
ook,现在安装完毕。
下面说说如何配置和启动:
基本思路是,配置JDK地址,配置core-site.xml,配置mapred-site.xml,hdfs-site.xml.
首先建立一个用来存放数据的目录:mkdir /usr/local/hadoop-datastore
打开conf/core-site.xml,配置如下
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop-datastore/</value>
<description>A base for other temporary directories.</description>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:54310</value>
<description>The name of the default file system. A URI whose
scheme and authority determine the FileSystem implementation. The
uri's scheme determines the config property (fs.SCHEME.impl) naming
the FileSystem implementation class. The uri's authority is used to
determine the host, port, etc. for a filesystem.</description>
</property>
</configuration>
复制代码
mapred-site.xml如下:
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>localhost:54311</value>
<description>The host and port that the MapReduce job tracker runs
at. If "local", then jobs are run in-process as a single map
and reduce task.
</description>
</property>
</configuration>
复制代码
hdfs-site.xml如下:
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
<description>Default block replication.
The actual number of replications can be specified when the file is created.
The default is used if replication is not specified in create time.
</description>
</property>
</configuration>
复制代码
ok,配置完毕
格式化HDFS:
hadoop@ubuntu:~$ /hadoop/bin/hadoop namenode -format
复制代码
启动HDFS和MapReduce
hadoop@ubuntu:~$ /bin/start-all.sh
复制代码
停止服务的脚本是:
hadoop@ubuntu:~$ /bin/stop-all.sh
http://www.hadoopor.com/thread-2674-1-1.html
1,我是在deskTop版本的Ubuntu下安装的所以,需要先安装ssh server。这个很好找,直接到到新立得里搜索ssh,第一个就是。
2,安装sun jdk6,切忌一定要是java6及其以上版本。先到更新管理器里把canonical的源加进去。
之后sudo apt-get update.
3, sudo apt-get sun-java6-jdk
4, sudo update-java-alternatives -s java-6-sun
5,增加一个用户组用户,用于hadoop运行及访问。
sudo addgroup hadoop
sudo adduser --ingroup hadoop hadoop
复制代码
6,生成SSH证书,配置SSH加密key
user@ubuntu:~$ su - hadoop
hadoop@ubuntu:~$ ssh-keygen -t rsa -P ""
Generating public/private rsa key pair.
Enter file in which to save the key (/home/hadoop/.ssh/id_rsa):
Created directory '/home/hadoop/.ssh'.
Your identification has been saved in /home/hadoop/.ssh/id_rsa.
Your public key has been saved in /home/hadoop/.ssh/id_rsa.pub.
The key fingerprint is:
9b:82:ea:58:b4:e0:35:d7:ff:19:66:a6:ef:ae:0e:d2 hadoop@ubuntu
The key's randomart image is:
[...snipp...]
hadoop@ubuntu:~$
复制代码
hadoop@ubuntu:~$ cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
复制代码
配置完成,测试一下:
hadoop@ubuntu:~$ ssh localhost
The authenticity of host 'localhost (::1)' can't be established.
RSA key fingerprint is d7:87:25:47:ae:02:00:eb:1d:75:4f:bb:44:f9:36:26.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added 'localhost' (RSA) to the list of known hosts.
Linux ubuntu 2.6.32-22-generic #33-Ubuntu SMP Wed Apr 28 13:27:30 UTC 2010 i686 GNU/Linux
Ubuntu 10.04 LTS
[...snipp...]
hadoop@ubuntu:~$
复制代码
8,禁用ipV6配置:
打开/etc/sysctl.conf, 此文件需要root权限。
再次文件中,追加如下:
net.ipv6.conf.all.disable_ipv6 = 1
net.ipv6.conf.default.disable_ipv6 = 1
net.ipv6.conf.lo.disable_ipv6 = 1
复制代码
重启,测试是否配置成功:
$ cat /proc/sys/net/ipv6/conf/all/disable_ipv6
复制代码
如果是1就ok了。
下面就是安装Hadoop了。
首先是:下载,解压缩,分配权限。
下载就不说了。
下载后运行如下:
$ cd /usr/local
$ sudo tar xzf hadoop-0.20.2.tar.gz
$ sudo mv hadoop-0.20.2 hadoop
$ sudo chown -R hadoop:hadoop hadoop
复制代码
ook,现在安装完毕。
下面说说如何配置和启动:
基本思路是,配置JDK地址,配置core-site.xml,配置mapred-site.xml,hdfs-site.xml.
首先建立一个用来存放数据的目录:mkdir /usr/local/hadoop-datastore
打开conf/core-site.xml,配置如下
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop-datastore/</value>
<description>A base for other temporary directories.</description>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:54310</value>
<description>The name of the default file system. A URI whose
scheme and authority determine the FileSystem implementation. The
uri's scheme determines the config property (fs.SCHEME.impl) naming
the FileSystem implementation class. The uri's authority is used to
determine the host, port, etc. for a filesystem.</description>
</property>
</configuration>
复制代码
mapred-site.xml如下:
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>localhost:54311</value>
<description>The host and port that the MapReduce job tracker runs
at. If "local", then jobs are run in-process as a single map
and reduce task.
</description>
</property>
</configuration>
复制代码
hdfs-site.xml如下:
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
<description>Default block replication.
The actual number of replications can be specified when the file is created.
The default is used if replication is not specified in create time.
</description>
</property>
</configuration>
复制代码
ok,配置完毕
格式化HDFS:
hadoop@ubuntu:~$ /hadoop/bin/hadoop namenode -format
复制代码
启动HDFS和MapReduce
hadoop@ubuntu:~$ /bin/start-all.sh
复制代码
停止服务的脚本是:
hadoop@ubuntu:~$ /bin/stop-all.sh
http://www.hadoopor.com/thread-2674-1-1.html
相关文章推荐
- Ubuntu与Centos的Hadoop安装以及编译运行MapReduce,Hadoop集群安装配置教程
- Ubuntu下伪分布式模式Hadoop的安装及配置
- Hadoop安装教程_单机/伪分布式配置_Hadoop2.6.0/Ubuntu14.04
- Hadoop安装教程_单机/伪分布式配置_Hadoop2.6.0/Ubuntu14.04
- Hadoop集群安装配置教程_Hadoop2.6.0_Ubuntu/CentOS
- Hadoop集群安装配置教程_Hadoop2.6.0_Ubuntu/CentOS
- Hadoop安装教程_单机/伪分布式配置_Hadoop2.6.0/Ubuntu14.04
- [转] hadoop学习---3.hadoop在ubuntu下的安装配置
- Ubuntu下面安装和配置Hadoop的dfs-fuse
- Hadoop集群安装配置教程_Hadoop2.6.0_Ubuntu/CentOS
- Ubuntu下 Hadoop 1.2.1 配置安装
- Ubuntu配置hadoop单机+伪分布式环境+eclipse--安装hadoop(一)
- hadoop在ubuntu下的安装配置
- 详解Ubuntu16.04下Hadoop 2.7.3的安装与配置
- hadoop在ubuntu下的安装配置
- Ubuntu下安装单机版Hadoop的配置
- hadoop学习---3.hadoop在ubuntu下的安装配置(转)
- Hadoop安装教程_单机/伪分布式配置_hadoop-2.7.5/Ubuntu14.04
- Hadoop 2.x安装教程_单机/伪分布式配置_Ubuntu14.04 64bitx
- Hadoop安装教程_单机配置_Hadoop1.2.1/Ubuntu16.04