spark集群安装部署(spark on yarn)
2015-01-17 17:35
633 查看
spark集群安装部署(spark on yarn)
1.前提
已经部署了Hadoop 2.2集群
2.下载并安装scala
2.1下载scala
http://www.scala-lang.org/download/
2.2安装scala
mkdir -p /usr/local/myspark/scala
cd /opt
tar -zxvf scala-2.11.2.tgz
cp -r scala-2.11.2 /usr/local/myspark/scala/
echo "export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2" >> /etc/profile
echo "export PATH=$SCALA_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
3.下载spark
http://spark.apache.org/downloads.html
4.安装spark
mkdir -p /usr/local/myspark/spark
cd /opt
tar -zxvf spark-1.0.2-bin-hadoop2.tgz
cp -r spark-1.0.2-bin-hadoop2 /usr/local/myspark/spark/
5.配置
5.2配置spark-env.sh
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp spark-env.sh.template spark-env.sh
vi spark-env.sh
export JAVA_HOME=/usr/local/java/jdk1.7.0_25
export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2
export HADOOP_CONF_DIR=/usr/local/hadoop/hadoop-2.2.0/etc/hadoop
export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2
![](http://img.blog.csdn.net/20150117172517375?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
5.2修改spark-defaults.conf
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp spark-defaults.conf.template spark-defaults.conf
vi spark-defaults.conf
spark.master yarn-cluster
spark.eventLog.enabled true
spark.eventLog.dir hdfs://master:9000/sparkeventlog
spark.serializer org.apache.spark.serializer.KryoSerializer
![](http://img.blog.csdn.net/20150117172531625?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
5.3修改slaves
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
vi slaves
master
master2
slave1
slave2
slave3
![](http://img.blog.csdn.net/20150117172637368?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/00356b207fa1444abde2720080f3a167/clipboard.png)
5.4修改log4j.properties
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp log4j.properties.template log4j.properties
vi log4j.properties
# Set everything to be logged to the console
spark.log=/var/log
log4j.rootCategory=INFO, console,file
log4j.appender.console=org.apache.log4j.ConsoleAppender
log4j.appender.console.target=System.err
log4j.appender.console.layout=org.apache.log4j.PatternLayout
log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n
log4j.appender.file=org.apache.log4j.RollingFileAppender
log4j.appender.file.MaxFileSize=5MB
log4j.appender.file.MaxBackupIndex=10
log4j.appender.file.layout=org.apache.log4j.PatternLayout
log4j.appender.file.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n
log4j.appender.file.File=${spark.log}/spark.log
# Settings to quiet third party logs that are too verbose
log4j.logger.org.eclipse.jetty=WARN
log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO
log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO
![](http://img.blog.csdn.net/20150117172648546?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
5.5修改/etc/profile
vi /etc/profile
echo "export SPARK_EXAMPLES_JAR=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/lib/spark-examples-1.0.2-hadoop2.2.0.jar" >> /etc/profile
echo "export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2" >> /etc/profile
echo "export PATH=$SPARK_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
6.拷贝到其他节点
mkdir -p /usr/local/myspark/scala
mkdir -p /usr/local/myspark/spark
scp -r 10.41.2.82:/usr/local/myspark/scala/scala-2.11.2 /usr/local/myspark/scala/
scp -r 10.41.2.82:/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2 /usr/local/myspark/spark/
echo "export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2" >> /etc/profile
echo "export PATH=$SCALA_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
echo "export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2" >> /etc/profile
echo "export PATH=$SPARK_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
7.启动停止
7.1启动
在master(10.41.2.82)上运行
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/sbin/start-all.sh
![](http://img.blog.csdn.net/20150117172802887?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
7.2停止
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/sbin/stop-all.sh
8.测试
8.1Web UI http://10.41.2.82:8080 http://master:8080
![](http://img.blog.csdn.net/20150117172805046?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117172820953?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
8.2运行demo
在10.41.2.82上执行以下命令:
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/bin/spark-submit --class org.apache.spark.examples.SparkPi --master yarn-cluster --num-executors 3 --driver-memory 4g --executor-memory 2g --executor-cores
1 /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/lib/spark-examples-1.0.2-hadoop2.2.0.jar 10
![](http://img.blog.csdn.net/20150117172917518?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117172934101?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117172947938?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117172959388?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/f98db7f012d84f478fc8e9697a719fc1/clipboard.png)
![](http://img.blog.csdn.net/20150117173016143?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117173032570?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117173028062?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/e02194514901428c932e634eaadfa1ef/clipboard.png)
![](http://img.blog.csdn.net/20150117173044406?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117173122906?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
![](http://img.blog.csdn.net/20150117173247681?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
查看结果:
访问: http://master:8088/proxy/application_1409622175934_0001/A
![](http://img.blog.csdn.net/20150117173305044?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
点击logs
![](http://img.blog.csdn.net/20150117173349317?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
点击stdout : Total
file length is 23 bytes.
![](http://img.blog.csdn.net/20150117173417241?watermark/2/text/aHR0cDovL2Jsb2cuY3Nkbi5uZXQvZGluZ3pmYW5n/font/5a6L5L2T/fontsize/400/fill/I0JBQkFCMA==/dissolve/70/gravity/Center)
结果为:
1.前提
已经部署了Hadoop 2.2集群
2.下载并安装scala
2.1下载scala
http://www.scala-lang.org/download/
2.2安装scala
mkdir -p /usr/local/myspark/scala
cd /opt
tar -zxvf scala-2.11.2.tgz
cp -r scala-2.11.2 /usr/local/myspark/scala/
echo "export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2" >> /etc/profile
echo "export PATH=$SCALA_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
3.下载spark
http://spark.apache.org/downloads.html
4.安装spark
mkdir -p /usr/local/myspark/spark
cd /opt
tar -zxvf spark-1.0.2-bin-hadoop2.tgz
cp -r spark-1.0.2-bin-hadoop2 /usr/local/myspark/spark/
5.配置
5.2配置spark-env.sh
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp spark-env.sh.template spark-env.sh
vi spark-env.sh
export JAVA_HOME=/usr/local/java/jdk1.7.0_25
export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2
export HADOOP_CONF_DIR=/usr/local/hadoop/hadoop-2.2.0/etc/hadoop
export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2
5.2修改spark-defaults.conf
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp spark-defaults.conf.template spark-defaults.conf
vi spark-defaults.conf
spark.master yarn-cluster
spark.eventLog.enabled true
spark.eventLog.dir hdfs://master:9000/sparkeventlog
spark.serializer org.apache.spark.serializer.KryoSerializer
5.3修改slaves
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
vi slaves
master
master2
slave1
slave2
slave3
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/00356b207fa1444abde2720080f3a167/clipboard.png)
5.4修改log4j.properties
cd /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/conf
cp log4j.properties.template log4j.properties
vi log4j.properties
# Set everything to be logged to the console
spark.log=/var/log
log4j.rootCategory=INFO, console,file
log4j.appender.console=org.apache.log4j.ConsoleAppender
log4j.appender.console.target=System.err
log4j.appender.console.layout=org.apache.log4j.PatternLayout
log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n
log4j.appender.file=org.apache.log4j.RollingFileAppender
log4j.appender.file.MaxFileSize=5MB
log4j.appender.file.MaxBackupIndex=10
log4j.appender.file.layout=org.apache.log4j.PatternLayout
log4j.appender.file.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{1}: %m%n
log4j.appender.file.File=${spark.log}/spark.log
# Settings to quiet third party logs that are too verbose
log4j.logger.org.eclipse.jetty=WARN
log4j.logger.org.apache.spark.repl.SparkIMain$exprTyper=INFO
log4j.logger.org.apache.spark.repl.SparkILoop$SparkILoopInterpreter=INFO
5.5修改/etc/profile
vi /etc/profile
echo "export SPARK_EXAMPLES_JAR=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/lib/spark-examples-1.0.2-hadoop2.2.0.jar" >> /etc/profile
echo "export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2" >> /etc/profile
echo "export PATH=$SPARK_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
6.拷贝到其他节点
mkdir -p /usr/local/myspark/scala
mkdir -p /usr/local/myspark/spark
scp -r 10.41.2.82:/usr/local/myspark/scala/scala-2.11.2 /usr/local/myspark/scala/
scp -r 10.41.2.82:/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2 /usr/local/myspark/spark/
echo "export SCALA_HOME=/usr/local/myspark/scala/scala-2.11.2" >> /etc/profile
echo "export PATH=$SCALA_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
echo "export SPARK_HOME=/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2" >> /etc/profile
echo "export PATH=$SPARK_HOME/bin:$PATH" >> /etc/profile
source /etc/profile
7.启动停止
7.1启动
在master(10.41.2.82)上运行
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/sbin/start-all.sh
7.2停止
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/sbin/stop-all.sh
8.测试
8.1Web UI http://10.41.2.82:8080 http://master:8080
8.2运行demo
在10.41.2.82上执行以下命令:
/usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/bin/spark-submit --class org.apache.spark.examples.SparkPi --master yarn-cluster --num-executors 3 --driver-memory 4g --executor-memory 2g --executor-cores
1 /usr/local/myspark/spark/spark-1.0.2-bin-hadoop2/lib/spark-examples-1.0.2-hadoop2.2.0.jar 10
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/f98db7f012d84f478fc8e9697a719fc1/clipboard.png)
![](file:///C:/Users/dingzhf/AppData/Local/YNote/data/dingzfang@163.com/e02194514901428c932e634eaadfa1ef/clipboard.png)
查看结果:
访问: http://master:8088/proxy/application_1409622175934_0001/A
点击logs
点击stdout : Total
file length is 23 bytes.
结果为:
Pi is roughly 3.145044
相关文章推荐
- Spark2.0.1 on yarn with hue 集群搭建部署(五)hue安装支持hadoop
- Spark2.0.1 on yarn with hue 集群安装部署(八)hue+livy+spark整合
- Spark2.0.1 on yarn with hue 集群安装部署(九)启动脚本编写
- Spark2.0.1 on yarn with hue 集群安装部署(六)livy安装测试
- Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- spark2.0.1 on yarn with hue 集群搭建部署(三)msyql安装
- Spark On YARN 集群安装部署
- 【转载】Spark On YARN 集群安装部署
- Spark On YARN 集群安装部署
- Spark2.0.1 on yarn with hue 集群安装部署(六)hue+hive+sparksql
- spark on yarn 集群 安装
- Spark2.2.1 on yarn 模式集群安装
- Spark2.0.1 on yarn with hue 集群搭建部署(二)spark on yarn搭建
- spark2.0.1 on yarn with hue 集群搭建部署(四)hive、hive metastore安装
- spark on yarn 集群部署