scala-2.13.0.tgz
spark-2.4.3-bin-hadoop2.7 .tgz
链接:https://pan.baidu.com/s/1WaohSAwWaOhWURthWx8KHg
提取码:hkd6
由于电脑配置问题,这里我只做了两台虚拟机进行实验
主 hadp01
从 hadp06
ssh-keygen #生成密钥对
ssh-copy-id hadp01 #将公钥拷贝给本机
ssh-copy-id hadp06 #将公钥拷贝给其他机
cd /opt/install
rz
tar -zxvf scala-2.13.1.tgz
tar -zxvf spark-2.4.4-bin-hadoop2.7.tgz
移动到 /opt/software 目录 mv scala-2.13.1 /opt/software
mv spark-2.4.4-bin-hadoop2.7 /opt/software
vi /etc/profile
SCALA_HOME=/opt/software/scala-2.13.1
SPARK_HOME=/opt/software/spark-2.4.4-bin-hadoop2.7
PATH一行直接加 :$SPARK_HOME/bin:$PATH:$SCALA_HOME/bin
cd /opt/software/spark-2.4.4-bin-hadoop2.7/conf
将 slaves.template spark-env.sh.template .template去掉
mv spark-env.sh.template spark-env.sh
mv slaves.template slaves
vim spark-env.sh
添加
export JAVA_HOME=/opt/software/jdk1.8
export SCALA_HOME=/opt/software/scala-2.13.1
export HADOOP_HOME=/opt/software/hadoop-2.7.7
export HADOOP_CONF_DIR=/opt/software/hadoop-2.7.7/etc/hadoop
export SPARK_MASTER_HOST=hadp01 主机名
export SPARK_WORKER_MEMORY=1g 虚拟机内存
export SPARK_WORKER_CORES=1 虚拟机核数
export SPARK_HOME=/opt/software/spark-2.4.4-bin-hadoop2.7
export SPARK_DIST_CLASSPATH=$(/opt/software/hadoop-2.7.7/bin/hadoop classpath)
vim slaves
将localhost改为 主从主机名
start-all.sh
再在主节点启动spark
cd /opt/software/spark-2.4.4-bin-hadoop2.7
sbin/start-all.sh
主:
从:
原文:https://www.cnblogs.com/the-roc/p/12718078.html