Ubuntu安装Spark和Hadoop集群
2017-08-02 22:38
465 查看
一、JDK安装
手动解压JDK的压缩包,然后设置环境变量
1.1在/usr/目录下创建java目录
root@ubuntu:~# mkdir/usr/java
root@ubuntu:~# cd /usr/java
1.2 下载jdk,然后解压
http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html root@ubuntu:~# tar -zxvf jdk-8u144-linux-x64.tar.gz
1.3 设置环境变量
root@ubuntu:~# vi /etc/profile
在profile中添加如下内容:
#set java environment
JAVA_HOME=/usr/java/jdk1.8.0_144
JRE_HOME=/usr/java/jdk1.8.0_144/jre
CLASS_PATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar:$JRE_HOME/lib
PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
export JAVA_HOME JRE_HOME CLASS_PATH PATH
让修改生效:
root@ubuntu:~# source /etc/profile
1.4 验证JDK有效性
root@ubuntu:~# java -version
java version "1.8.0_144"
Java(TM) SE Runtime Environment (build 1.8.0_144-b01)
Java HotSpot(TM) 64-Bit Server VM (build 25.144-b01, mixed mode)
root@ubuntu:~#
二、Scala安装
下载scala:http://www.scala-lang.org/download/2.11.6.html
root@ubuntu:~# cd /usr
root@ubuntu:/usr# mkdir scala
root@ubuntu:/usr# cd scala/
root@ubuntu:/usr/scala# cp /home/baihu/pragrom/scala-2.11.6.tgz .
root@ubuntu:/usr/scala# ls
scala-2.11.6.tgz
root@ubuntu:/usr/scala# tar -zxf scala-2.11.6.tgz
root@ubuntu:/usr/scala# ls
scala-2.11.6 scala-2.11.6.tgz
root@ubuntu:/usr/scala#
编辑配置文件添加scala的配置
[root@master scala-2.11.6]# vi /etc/profile
export SCALA_HOME=/usr/scala/scala-2.11.6
export PATH=$PATH:$SCALA_HOME/bin
root@ubuntu:/usr/scala# source /etc/profile
root@ubuntu:/usr/scala# scala -version
Scala code runner version 2.11.6 -- Copyright 2002-2013, LAMP/EPFL
三、安装Spark
下载spark最新版spark-2.2.0-bin-hadoop2.7.tgz
http://spark.apache.org/downloads.html
root@ubuntu:/usr# mkdir spark
root@ubuntu:/usr# cd spark/
root@ubuntu:/usr/spark# mv /home/baihu/pragrom/spark-2.2.0-bin-hadoop2.7.tgz .
root@ubuntu:/usr/spark# ls
spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# tar -zxf spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# ls
spark-2.2.0-bin-hadoop2.7 spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# cd spark-2.2.0-bin-hadoop2.7/
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7# ls
bin data jars licenses python README.md sbin
conf examples LICENSE NOTICE R RELEASE yarn
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7# cd bin/
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7/bin# ./pyspark
Python 2.7.6 (default, Jun 22 2015, 17:58:13)
[GCC 4.8.2] on linux2
Type "help", "copyright", "credits" or "license" for more information.
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
17/08/02 07:34:23 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
17/08/02 07:34:23 WARN Utils: Your hostname, ubuntu resolves to a loopback address: 127.0.1.1; using 192.168.75.130 instead (on interface eth0)
17/08/02 07:34:23 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
17/08/02 07:34:52 WARN ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0
17/08/02 07:34:53 WARN ObjectStore: Failed to get database default, returning NoSuchObjectException
17/08/02 07:34:55 WARN ObjectStore: Failed to get database global_temp, returning NoSuchObjectException
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/__ / .__/\_,_/_/ /_/\_\ version 2.2.0
/_/
Using Python version 2.7.6 (default, Jun 22 2015 17:58:13)
SparkSession available as 'spark'.
>>>
手动解压JDK的压缩包,然后设置环境变量
1.1在/usr/目录下创建java目录
root@ubuntu:~# mkdir/usr/java
root@ubuntu:~# cd /usr/java
1.2 下载jdk,然后解压
http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html root@ubuntu:~# tar -zxvf jdk-8u144-linux-x64.tar.gz
1.3 设置环境变量
root@ubuntu:~# vi /etc/profile
在profile中添加如下内容:
#set java environment
JAVA_HOME=/usr/java/jdk1.8.0_144
JRE_HOME=/usr/java/jdk1.8.0_144/jre
CLASS_PATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar:$JRE_HOME/lib
PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
export JAVA_HOME JRE_HOME CLASS_PATH PATH
让修改生效:
root@ubuntu:~# source /etc/profile
1.4 验证JDK有效性
root@ubuntu:~# java -version
java version "1.8.0_144"
Java(TM) SE Runtime Environment (build 1.8.0_144-b01)
Java HotSpot(TM) 64-Bit Server VM (build 25.144-b01, mixed mode)
root@ubuntu:~#
二、Scala安装
下载scala:http://www.scala-lang.org/download/2.11.6.html
root@ubuntu:~# cd /usr
root@ubuntu:/usr# mkdir scala
root@ubuntu:/usr# cd scala/
root@ubuntu:/usr/scala# cp /home/baihu/pragrom/scala-2.11.6.tgz .
root@ubuntu:/usr/scala# ls
scala-2.11.6.tgz
root@ubuntu:/usr/scala# tar -zxf scala-2.11.6.tgz
root@ubuntu:/usr/scala# ls
scala-2.11.6 scala-2.11.6.tgz
root@ubuntu:/usr/scala#
编辑配置文件添加scala的配置
[root@master scala-2.11.6]# vi /etc/profile
export SCALA_HOME=/usr/scala/scala-2.11.6
export PATH=$PATH:$SCALA_HOME/bin
root@ubuntu:/usr/scala# source /etc/profile
root@ubuntu:/usr/scala# scala -version
Scala code runner version 2.11.6 -- Copyright 2002-2013, LAMP/EPFL
三、安装Spark
下载spark最新版spark-2.2.0-bin-hadoop2.7.tgz
http://spark.apache.org/downloads.html
root@ubuntu:/usr# mkdir spark
root@ubuntu:/usr# cd spark/
root@ubuntu:/usr/spark# mv /home/baihu/pragrom/spark-2.2.0-bin-hadoop2.7.tgz .
root@ubuntu:/usr/spark# ls
spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# tar -zxf spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# ls
spark-2.2.0-bin-hadoop2.7 spark-2.2.0-bin-hadoop2.7.tgz
root@ubuntu:/usr/spark# cd spark-2.2.0-bin-hadoop2.7/
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7# ls
bin data jars licenses python README.md sbin
conf examples LICENSE NOTICE R RELEASE yarn
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7# cd bin/
root@ubuntu:/usr/spark/spark-2.2.0-bin-hadoop2.7/bin# ./pyspark
Python 2.7.6 (default, Jun 22 2015, 17:58:13)
[GCC 4.8.2] on linux2
Type "help", "copyright", "credits" or "license" for more information.
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
17/08/02 07:34:23 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
17/08/02 07:34:23 WARN Utils: Your hostname, ubuntu resolves to a loopback address: 127.0.1.1; using 192.168.75.130 instead (on interface eth0)
17/08/02 07:34:23 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
17/08/02 07:34:52 WARN ObjectStore: Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0
17/08/02 07:34:53 WARN ObjectStore: Failed to get database default, returning NoSuchObjectException
17/08/02 07:34:55 WARN ObjectStore: Failed to get database global_temp, returning NoSuchObjectException
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/__ / .__/\_,_/_/ /_/\_\ version 2.2.0
/_/
Using Python version 2.7.6 (default, Jun 22 2015 17:58:13)
SparkSession available as 'spark'.
>>>
相关文章推荐
- 联想ThinkPad S3-S440虚拟机安装,ubuntu安装,Hadoop(2.7.1)详解及WordCount运行,spark集群搭建
- Ubuntu1604安装ganglia并监控Hadoop/HBase/Spark集群
- CentOS 6.5离线安装cloudera manager hadoop spark集群
- ubuntu安装spark2.1 hadoop2.7.3集群
- Spark最佳实践中一键安装Spark集群的一些问题---Ubuntu如何root登录安装
- Ubuntu 16.04安装Hadoop及Spark
- 安装Hadoop及Spark(Ubuntu 16.04)
- Hadoop+Hbase+Spark集群配置—Zookeeper安装
- hadoop集群下spark安装
- 【Spark亚太研究院系列丛书】Spark实战高手之路-第一章 构建Spark集群-安装Ubuntu系统(3)
- Hadoop集群安装配置教程_Hadoop2.6.0_Ubuntu/CentOS
- Hadoop集群安装配置教程_Hadoop2.6.0_Ubuntu/CentOS
- Hadoop Spark 集群简便安装总结
- 64位Ubuntu1404集群安装配置hadoop-2.2.0
- docker for mac 安装spark 2.2.0 + hadoop 2.8.0 集群
- 使用 Shell 脚本进行 Hadoop Spark 集群的批量安装
- Linux ->> UBuntu 14.04 LTE下安装Hadoop 1.2.1(集群分布式模式)
- hadoop集群上面安装 spark 1.1.0
- Ubuntu系统(64位)下安装并配置Hadoop-2.2.0集群
- Ubuntu16.4-Hadoop2.7.5分布式集群搭建(三)--- Hbase的安装与配置