用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
2017-01-05 11:05
309 查看
一、将pyspark放入:
该目录位置(我的是mac):
/Library/Python/2.7/site-packages
二、env配置:
步骤1:
步骤2:
步骤3:
SPARK_CLASSPATH
/Users/Chaves/workspace/spark/hbase-0.98.3/lib/:/Users/Chaves/workspace/spark/spark-2.0.1-bin-hadoop2.6/lib/:
SPARK_HOME
/Users/Chaves/workspace/spark/spark-2.0.1-bin-hadoop2.6
更多其他pycharm 的详细设计见:http://blog.csdn.net/u013660881/article/details/46777415
三、运行命令:
1,spark 终端运行命令
如本机spark包位置
/Users/个人目录/workspace/spark/spark-2.0.1-bin-hadoop2.6/conf
spark-env.sh
2,修改SPARK_CLASSPATH
2.1,在以下spark的bin目录下运行:
/Users/个人目录/workspace/spark/spark-2.0.1-bin-hadoop2.6/bin
2.2,启动命令(./spark-submit –jars包)
./spark-submit —jars jar包地址 —py-files 工具包 算法文件地址 参数1 参数2 参数3 …nt
2.0与1.0的区别:
该目录位置(我的是mac):
/Library/Python/2.7/site-packages
二、env配置:
步骤1:
步骤2:
步骤3:
SPARK_CLASSPATH
/Users/Chaves/workspace/spark/hbase-0.98.3/lib/:/Users/Chaves/workspace/spark/spark-2.0.1-bin-hadoop2.6/lib/:
SPARK_HOME
/Users/Chaves/workspace/spark/spark-2.0.1-bin-hadoop2.6
更多其他pycharm 的详细设计见:http://blog.csdn.net/u013660881/article/details/46777415
三、运行命令:
1,spark 终端运行命令
如本机spark包位置
/Users/个人目录/workspace/spark/spark-2.0.1-bin-hadoop2.6/conf
spark-env.sh
2,修改SPARK_CLASSPATH
2.1,在以下spark的bin目录下运行:
/Users/个人目录/workspace/spark/spark-2.0.1-bin-hadoop2.6/bin
2.2,启动命令(./spark-submit –jars包)
./spark-submit —jars jar包地址 —py-files 工具包 算法文件地址 参数1 参数2 参数3 …nt
2.0与1.0的区别:
spark = SparkSession.builder.master("local").appName("pyspark2_0_1_test").getOrCreate() sc = self.spark.sparkContext hc = HiveContext(sc) ....
相关文章推荐
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- 用pycharm + python写spark(spark-2.0.1-bin-hadoop2.6)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之Basic包下的JavaPageRank.java(图文详解)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之Basic包下的JavaTC.java(图文详解)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之Basic包下的SparkTC.scala(图文详解)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之Basic包下的JavaSparkPi.java(图文详解)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之Basic包下的SparkPi.scala(图文详解)
- spark-2.2.0-bin-hadoop2.6和spark-1.6.1-bin-hadoop2.6发行包自带案例全面详解(java、python、r和scala)之环境准备(图文详解)
- ps -aux | sort -k4nr /opt/models/jdk//bin/java -cp /opt/models/spark-1.6.1-bin-hadoop2.6/conf/
- 使用PyCharm配置Spark的Python开发环境(基础)
- spark运行过程中报错Could not locate executable null\bin\winutils.exe in the Hadoop binaries.解决