在 CentOS 上部署 Hadoop 的完整步骤
一 环境准备与系统要求
二 单机模式部署步骤(适合开发测试)
sudo yum install -y java-11-openjdk-devel~/.bashrc 或 /etc/profile):
echo 'export JAVA_HOME=/usr/lib/jvm/java-11-openjdk' >> ~/.bashrcecho 'export PATH=$JAVA_HOME/bin:$PATH' >> ~/.bashrcsource ~/.bashrcsudo yum install -y openssh-serversudo systemctl start sshd && sudo systemctl enable sshdssh-keygen -t rsa(回车跳过密码)cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys && chmod 600 ~/.ssh/authorized_keysssh localhost 验证wget https://dlcdn.apache.org/hadoop/common/hadoop-3.3.6/hadoop-3.3.6.tar.gztar -xzvf hadoop-3.3.6.tar.gz -C /opt/sudo chown -R hadoopuser:hadoopuser /opt/hadoop-3.3.6~/.bashrc)
echo 'export HADOOP_HOME=/opt/hadoop-3.3.6' >> ~/.bashrcecho 'export PATH=$HADOOP_HOME/bin:$PATH' >> ~/.bashrcecho 'export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop' >> ~/.bashrcsource ~/.bashrc$HADOOP_HOME/etc/hadoop/)
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>
</configuration>
hdfs namenode -format(仅首次)start-dfs.sh && start-yarn.shjps 应看到 NameNode、DataNode、ResourceManager、NodeManager三 伪分布式与多节点集群部署要点
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://namenode:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop/tmp</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/usr/local/hadoop/data/namenode</value>
</property>
<property>
<name>dfs.datanode.data.dir</value>
<value>/usr/local/hadoop/data/datanode</value>
</property>
</configuration>
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>resourcemanager</value>
</property>
</configuration>
hdfs namenode -formatstart-dfs.sh && start-yarn.shjps 检查进程,访问 9870/8088 验证 Web UI。四 常见问题与排查
~/.ssh/authorized_keys 权限为 600,并确保 sshd 已启动。java -version 校验,必要时切换 JAVA_HOME。source ~/.bashrc 或重启会话。