Ubuntu Hadoop 集群搭建指南
一 环境准备与规划
二 基础环境配置
sudo apt update && sudo apt install -y openjdk-8-jdkjava -version 应返回 1.8.x。sudo adduser hadoop,sudo usermod -aG sudo hadoopsudo mkdir -p /usr/local/hadoop /opt/hadoop && sudo chown -R hadoop:hadoop /usr/local/hadoop /opt/hadoop/etc/hostname 分别设为 master/slave1/slave2;/etc/hosts(示例):192.168.1.10 master
192.168.1.11 slave1
192.168.1.12 slave2
sudo apt install -y openssh-serverssh localhost 可免密登录本机(用于后续分发公钥)。三 SSH 免密登录
ssh-keygen -t rsa -P "" -f ~/.ssh/id_rsassh-copy-id hadoop@master、ssh-copy-id hadoop@slave1、ssh-copy-id hadoop@slave2ssh slave1、ssh slave2 无需密码即可登录;如需 Worker 免密回连 Master,可在对应节点重复上述分发步骤。四 Hadoop 安装与配置
wget https://archive.apache.org/dist/hadoop/common/hadoop-3.3.4/hadoop-3.3.4.tar.gzsudo tar -xvf hadoop-3.3.4.tar.gz -C /usr/local
sudo mv /usr/local/hadoop-3.3.4 /usr/local/hadoop
sudo chown -R hadoop:hadoop /usr/local/hadoop
~/.bashrc 或 /etc/profile.d/hadoop.sh 添加:export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
source ~/.bashrc<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://master:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/hadoop/tmp</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.replication</name>
<value>2</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/opt/hadoop/hdfs/namenode</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/opt/hadoop/hdfs/datanode</value>
</property>
</configuration>
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
<configuration>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>master</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.resource.memory-mb</name>
<value>4096</value>
</property>
</configuration>
slave1
slave2
sudo mkdir -p /opt/hadoop/tmp /opt/hadoop/hdfs/{namenode,datanode}
sudo chown -R hadoop:hadoop /opt/hadoop
五 启动与验证
hdfs namenode -formatstart-dfs.shstart-yarn.shjps,应看到:NameNode、ResourceManagerjps,应看到:DataNode、NodeManagerworkers 是否配置正确;dfs.datanode.data.dir 目录存在且 hadoop 用户可写;~/.ssh/authorized_keys 权限为 600,.ssh 为 700。