之前已经在(),并正常启动Hadoop集群。
master
安装Scala
[root@master ~]# wget http://downloads.lightbend.com/scala/2.10.6/scala-2.10.6.tgz[root@master ~]# tar -zxvf scala-2.10.6.tgz[root@master ~]# mv scala-2.10.6 /usr/local/
Scala环境变量
[root@master ~]# vi /etc/profile.d/scala.sh
添加
export SCALA_HOME=/usr/local/scala-2.10.6export PATH=$PATH:$SCALA_HOME/bin
保存退出
[root@master ~]# source /etc/profile.d/scala.sh[root@master ~]# scala -versionScala code runner version 2.10.6 -- Copyright 2002-2013, LAMP/EPFL
安装Spark
[root@master ~]# wget http://archive.apache.org/dist/spark/spark-1.5.2/spark-1.5.2-bin-hadoop2.6.tgz[root@master ~]# tar -zxvf spark-1.5.2-bin-hadoop2.6.tgz[root@master ~]# mv spark-1.5.2-bin-hadoop2.6 /usr/local/
Spark环境变量
[root@master ~]# vi /etc/profile.d/spark.sh
添加
export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoopexport HDFS_CONF_DIR=$HADOOP_HOME/etc/hadoopexport YARN_CONF_DIR==$HADOOP_HOME/etc/hadoopexport SPARK_HOME=/usr/local/spark-1.5.2-bin-hadoop2.6export PATH=$PATH:$SPARK_HOME/bin
保存退出
[root@master ~]# source /etc/profile.d/spark.sh
[root@master ~]# cp /usr/local/spark-1.5.2-bin-hadoop2.6/conf/spark-env.sh.template /usr/local/spark-1.5.2-bin-hadoop2.6/conf/spark-env.sh[root@master ~]# vi /usr/local/spark-1.5.2-bin-hadoop2.6/conf/spark-env.sh
添加
export JAVA_HOME=/usr/local/jdk1.7.0_79export SCALA_HOME=/usr/local/scala-2.10.6export HADOOP_HOME=/usr/local/hadoop-2.7.2export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
保存退出
[root@master ~]# cp /usr/local/spark-1.5.2-bin-hadoop2.6/conf/slaves.template /usr/local/spark-1.5.2-bin-hadoop2.6/conf/slaves[root@master ~]# vi /usr/local/spark-1.5.2-bin-hadoop2.6/conf/slaves
删除localhost
,将内容修改为
masterslave
保存退出
验证Spark安装
[root@master spark-1.5.2-bin-hadoop2.6]# cd /usr/local/spark-1.5.2-bin-hadoop2.6/ [root@master spark-1.5.2-bin-hadoop2.6]# ./bin/spark-submit --class org.apache.spark.examples.SparkPi --master yarn-cluster lib/spark-examples*.jar 1016/11/26 03:35:07 INFO yarn.Client: Application report for application_1480118541212_0002 (state: FINISHED)16/11/26 03:35:07 INFO yarn.Client: client token: N/A diagnostics: N/A ApplicationMaster host: 192.168.229.131 ApplicationMaster RPC port: 0 queue: default start time: 1480160059605 final status: SUCCEEDED tracking URL: http://master:8088/proxy/application_1480118541212_0002/ user: root16/11/26 03:35:07 INFO util.ShutdownHookManager: Shutdown hook called16/11/26 03:35:07 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-e29afacf-58e9-4805-a59b-6fb5223ec872
[root@master spark-1.5.2-bin-hadoop2.6]# ssh slaveLast login: Sat Nov 26 02:30:33 2016 from 192.168.229.1[root@slave ~]# cd /usr/local/hadoop-2.7.2/logs/userlogs/application_1480118541212_0002/[root@slave application_1480118541212_0002]# cat container_1480118541212_0002_01_000001/stdout Pi is roughly 3.142008
Spark安装正常
浏览器中访问,可看到Spark作业界面。
[root@master spark-1.5.2-bin-hadoop2.6]# vi sbin/start-master.sh
发现
SPARK_MASTER_WEBUI_PORT=8080
[root@master spark-1.5.2-bin-hadoop2.6]# sbin/start-master.sh
浏览器可访问master:8080,查看Spark Master
[root@master ~]# ls /usr/local/spark-1.5.2-bin-hadoop2.6/bin/beeline pyspark.cmd spark-class.cmd spark-shell.cmdbeeline.cmd run-example sparkR spark-sqlload-spark-env.cmd run-example2.cmd sparkR2.cmd spark-submitload-spark-env.sh run-example.cmd sparkR.cmd spark-submit2.cmdpyspark spark-class spark-shell spark-submit.cmdpyspark2.cmd spark-class2.cmd spark-shell2.cmd[root@master ~]# /usr/local/spark-1.5.2-bin-hadoop2.6/bin/spark-shell
浏览器中访问master:4040
scala> exit[root@master ~]#