您的位置:首页 > 运维架构 > Docker

windows 基于docker下的 spark 开发环境搭建

2017-02-08 02:23 465 查看
docker toolbox
https://www.docker.com/products/docker-toolbox

spark
https://hub.docker.com/r/singularities/spark/~/dockerfile/

# start-hadoop-namenode
# hadoop fs -mkdir /user
# hadoop fs -mkdir /user/root/
# hadoop fs -put ./README.md /user/root

# start-spark
# start-spark worker [master]
# spark-shell
# spark-shell --master spark://a60b8c8f9653:7077

scala> val lines = sc.textFile("file:///usr/local/spark-2.1.0/README.md")
scala> val lines = sc.textFile("hdfs:///usr/local/spark-2.1.0/README.md")
lines: org.apache.spark.rdd.RDD[String] = file:///usr/local/spark-2.1.0/README.md MapPartitionsRDD[1] at textFile at <console>:24
scala> lines.count()
res0: Long = 104
scala> lines.saveAsTextFile("hdfs:///user/root/README2.md") // 保存到hdfs
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签:  spark