赞
踩
链接:https://pan.baidu.com/s/1RyJ2I4wUlVxgaGJXtIsBNw
提取码:rrzz
复制这段内容后打开百度网盘手机App,操作更方便哦
1. 上传并解压缩spark
tar -xvf spark-2.4.4-bin-hadoop2.7.tgz
移动到 /usr/local/目录下
mv spark-2.4.4-bin-hadoop2.7 /usr/local/spark
2. 配置spark-env.sh
进入 /usr/local/spark/conf
cp spark-env.sh.template spark-env.sh
修改 spark-env.sh,尾部添加:
export JAVA_HOME= /usr/local/jdk
export HADOOP_HOME=/usr/local/hadoop
export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
export SPARK_MASTER_IP=master
export SPARK_MASTER_PORT=7077
3.配置slaves
修改 cp slaves.template slaves
末尾添加:
master
slave1
slave2
4. 配置slave1和slave2的spark
scp -r /usr/local/spark root@slave1: /usr/local/spark
scp -r /usr/local/spark root@slave2: /usr/local/spark
启动spark:
cd /usr/local/spark/sbin
执行 ./start-all.sh (注意:不加./的话,但有start-all.sh会执行到hadoop的启动命令,加./是启动的是spark的启动命令)
注意: 启动spark前要先启动hadoop
分别是hadoop和启动spark后jps显示
spark的task监控web, hadoop的MapReduce是8088端口,spark的是8080端口
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。