在 Debian 上搭建 HDFS 集群的完整步骤
一 环境准备与规划
二 安装与基础配置
sudo apt update && sudo apt upgrade -ysudo apt install openjdk-11-jdk -y 或 sudo apt install openjdk-8-jdk -yjava -versionsudo adduser hadoop && sudo usermod -aG sudo hadoopsudo mkdir -p /opt/hadoop/hdfs/{name,data}wget https://downloads.apache.org/hadoop/common/hadoop-3.3.6/hadoop-3.3.6.tar.gzsudo tar -xzvf hadoop-3.3.6.tar.gz -C /usr/local/sudo ln -sfn /usr/local/hadoop-3.3.6 /usr/local/hadoopexport JAVA_HOME=/usr/lib/jvm/java-11-openjdk-amd64
export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
source /etc/profile 或 source /etc/profile.d/hadoop.shssh-keygen -t rsa -b 4096ssh-copy-id hadoop@node1、ssh-copy-id hadoop@node2 …ssh hadoop@node1 无需密码三 HDFS 核心配置
$HADOOP_HOME/etc/hadoop/export JAVA_HOME=/usr/lib/jvm/java-11-openjdk-amd64
export HADOOP_HOME=/usr/local/hadoop
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://master:9000</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///opt/hadoop/hdfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///opt/hadoop/hdfs/data</value>
</property>
</configuration>
192.168.1.100 master
192.168.1.101 node1
192.168.1.102 node2
sudo chown -R hadoop:hadoop /opt/hadoop/hdfs四 启动与验证
su - hadoophdfs namenode -formatstart-dfs.shjps(应看到 NameNode、DataNode 等)hdfs dfs -mkdir -p /user/hadoop
hdfs dfs -put $HADOOP_HOME/README.txt /user/hadoop/
hdfs dfs -ls /user/hadoop
~/.ssh/authorized_keys 权限为 600五 扩展与高可用(可选)
<configuration>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>master</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>
</configuration>
start-yarn.sh